A stochastic global optimization method for (urban?) multi-parameter systems. Rodolphe Le Riche CNRS and Ecole des Mines de St-Etienne
|
|
- Neal Chapman
- 5 years ago
- Views:
Transcription
1 A stochastic global optimization method for (urban?) multi-parameter systems Rodolphe Le Riche CNRS and Ecole des Mines de St-Etienne
2 Purpose of this presentation Discussion object : cities as an entity for heating consumption and related public policies. Cities are analyzed, modeled and optimized for heat consumption (including the effects of solar passive input). Cities are objects that are described at different parameterization levels : entire city (coarser level), districts, buildings (finer level). In mechanical engineering, multi-levels of parameters are often associated to composite materials. Talk : describe an idea for multi-level optimization that emerged in composite design. Any use for cities?
3 Notation and application example Composite design at various scales min f ( y ( x)) macro x S y(x) numerical simulator of the structure (stress, strains, strength, mass,...) meso, v ( plate stiffnesses ) v(x) micro, x ( fiber positions ) v(x) is numerically inexpensive. x(v) doesn't exist : there are many fiber positions for one choice of plate stiffnesses.
4 Notation and application example Heating consumption at various scales min f ( y ( x)) macro city x S y(x) numerical simulator of the structure (solar energy, thermal simulation) meso, v «IRIS» (district) micro, x building IRIS : Ilôt regroupé pour l'information statistique
5 Scope of application : global, derivative free problems Here, we focus on f min f x x S S ℝ n or ℕn or { ℝ n1 ℕn2 } i.e., continuous or integer or mixte optimization. local optimum x l V xl x * global optimum
6 Flow chart of a general stochastic optimizer Initialize t and pt(x) Sample Calculate f(xt+1) Update the distribution xt+1 ~ pt(x) pt+1(x) = Update( x1, f(x1),, xt+1, f(xt+1) ) or more often pt+1(x) = Update( pt(x), xt+1, f(xt+1) ) Stop or [ t = t+1 and go back to Sample ] with different p's if x is continuous or discrete or mixed.
7 A simple example in Rn : ES-(1+1) Initializations : x, f(x), m, C, While tmax. t < tmax do, Sample N(m,C) --> x' Calculate f(x'), t = t+1 If f(x')<f(x), x = x', f(x) = f(x') Endif Update m (e.g., m=x) and C End while Normal law N(m,C) [ 2 σ C = 0 σ σ ] for variables seen as independent.
8 Illustration : adaptation of 2D Gaussian with ES-(1+1)
9 Discrete variables : The Univariate Marginal Density Algorithm (UMDA) ( Baluja 1994 as PBIL and Mühlenbein 1996) x S {1,2,, A}n (alphabet of cardinality A ) o o o o n e.g. { 45, 0, 45, 90 } (fiber orientations) n e.g. {matl1,, matla } (material choice) The algorithm is that of a population based stochastic optimization (see before + many x's at each iteration) with different sampling and updating of pt. pt assumes that the variables are independent (drop t ), n p x = p i x i i=1 0.6 pi p2i pia p1i 1 A pij = 1 j= A xi
10 UMDA (2) Sampling : For i=1, n ui ~ U [0,1] If 0 ui p1i x i =1 u 1 0 k 2 p 1 A p +p k If 1 k 1 p ij ui pij j=1 A 1 If j=1 j=1 x i =k j p i ui 1 x i =A Learning : Select the μ best points out of λ, f x 1 : f x 2 : f x : pij is the frequency of j at position i in the bests : j pi = k =1 : k I x i = j 1 p 1 j i, k k I x i = j =1 if x i = j, =0 otherwise (minimum frequency for ergodicity)
11 Application to composite design for frequency density learned by UMDA (2D) contour lines of the penalized objective function Independent densities can neither represent curvatures nor variables' couplings. ( from Grosset, L., Le Riche, R. and Haftka, R.T., A double-distribution statistical algorithm for composite laminate optimization, SMO, 2006 )
12 Stochastic discrete optimization : learning the variables dependencies More sophisticated discrete optimization methods attempt to learn the couplings between variables. For example, with pairwise dependencies : X1:n X2:n... Xn:n p( x ) = p ( x 1 : n ) p( x 2: n x 1: n ) p ( x n : n x n 1: n ) Trade-off : richer probabilistic structures better capture the objective function landscape but they also have more parameters need more f evaluations to be learned. MIMIC ( Mutual Information Maximizing Input Clustering ) algorithm : De Bonnet, Isbell and Viola, BMDA ( Bivariate Marginal Distribution Algorithm ) : Pelikan and Muehlenbein, 1999.
13 Multi-level parameter optimization with DDOA ( from Grosset, L., Le Riche, R. and Haftka, R.T., A double-distribution statistical algorithm for composite laminate optimization, SMO, 2006 ) Mathematical motivation : create couplings between variables using many independent distributions (in x and v spaces). Numerical motivation : take into account expert knowledge in the optimization to improve efficiency. E.g. in composites, the lamination parameters v (the plate stiffnesses) make physical sense.
14 Example in composites Use of the lamination parameters v = lamination parameters = geometric contribution of the plies to the stiffness. Inexpensive to calculate from x (fiber angles). Simplifications : fewer v's than fiber angles. Often, the v's are taken as continuous. But f(v) typically does not exist (e.g., ply failure criterion).
15 Past examples in composites ( Liu, Haftka, and Akgün, «Two-level composite wing structural optimization using response surfaces», Merval, Samuelides and Grihon, «Lagrange-Kuhn-Tucker coordination for multilevel optimization of aeronautical structures», ) Initial problem : Optimize a composite structure made of several assembled panels by changing each ply orientation many discrete variables Decomposed problem : Structure level Optimize a composite structure made of several assembled panels by changing the lamination parameters of each panel few continuous variables optimal v's Laminate level Minimize the distance to target lamination parameters by changing the ply orientations few discrete variables BUT for such a sequential approach to make sense, f (v ) must exist and guide to optimal regions (i.e., prohibits emergence of solutions at finer scales).
16 The DDOA stochastic optimization algorithm objective function If v(x) is costless, it does not cost to learn densities in the x AND v spaces at the same time. does not exist p(x) pddoa ( x) = p X v ( X )=V (x ) x v v(x)
17 The DDOA algorithm : X v(x)=v? Simple mathematical illustration : 1 1 T p X (x ) = exp x x 2π 2 ( ) v = x1+ x pv (v ) = exp (v 1) 2 2 π ( ) Intermediate step for a given v : p X v( X )= v=1 ( x) is a degenerated Gaussian along x1+ x 2 = 1 (cross-section of the 2D bell curve along the blue line + normalization)
18 The DDOA algorithm : X v(x)=v? p X v ( X )=V (x ) = p X v( X )=v pv (v (x )) =... = exp ( x1 x 2 )2 ( x 1 + x 2 1)2 2π 4 2 ( ) p X v( X )=V ( x) is a coupled distribution that merges the effects of X and V Analytical calculation in the Gaussian case. In practice, use simulations...
19 The DDOA algorithm (flow chart) Choose λ, μ, ρ such that ρ>>1 and λ>μ Initialize pv(v)and px(x) For i=1,λ do Sample vtarget from pv(v) Sample ρ>>1 x's from px(x) x(i) = the closest x to vtarget Calculate f(x(i)) sampling of X v(x)=v end For Rank x(1:λ),, x(λ:λ) the proposed points Update pv(v) and px(x) from x(1:λ),, x(μ:λ) Stop? If no, go back to top...
20 Application of DDOA to composite design for frequency px(x) and pv(v) can be simple densities, without variables couplings ( easy to learn), yet pddoa(x) is a coupled density. f(x) and selected points n p X x = i =1 pi x i pddoa(x) One half of the algorithm searches in a low dimension space.
21 additional slides
22 Introduction to stochastic optimization Random numbers are versatile search engines (work both in Rn and / or Nn ). They can also yield efficient methods. Let pt(x) denote the probability density function of x at iteration t (e.g., after t evaluations of f). It represents the belief at t that the optimum x* is at x. How to «sample pt(x)» once (Scilab notation)? if x is uniform between m and M, X ~ U[m,M], call x = m + rand(n,1).*(m-m) if x is (multi-)gaussian with mean m and covariance matrix C, X ~ N(m,C),call x = m + grand(1,'mn',0,c)
23 Application to a laminate frequency problem (1) ( from Grosset, L., Le Riche, R. and Haftka, R.T., A double-distribution statistical algorithm for composite laminate optimization, SMO, 2006 ) max x f 1 x 1,, x 15, the first eigenfreq. of a simply supported plate such that 0.48 eff x 0.52 where x i {0 o, 15o,, 90o } the constraint is enforced by penalty and creates a narrow ridge in the design space
24 Application to a laminate frequency problem (2) Optimum : [90o4 /±75o /±60o2 /±45o5 /±305o ]s Compare UMDA to a GA (genetic algorithm) and SHC (Stochastic Hill Climber) Reliability = probability of finding the optimum at a given cost. UMDA performs fairly well on this problem.
25 Example in composites Use of the lamination parameters Simplifications : fewer v's than fiber angles. Often, the v's are taken as continuous. But f(v) typically does not exist (e.g., ply failure criterion).
Discrete stochastic optimization with continuous auxiliary variables. Rodolphe Le Riche*, Alexis Lasseigne**, François-Xavier Irisarri**
Discrete stochastic optimization with continuous auxiliary variables Rodolphe Le Riche*, Alexis Lasseigne**, François-Xavier Irisarri** *CNRS and Ecole des Mines de St-Etienne, Fr. ** ONERA, Composite
More informationSTATISTICAL OPTIMIZATION OF COMPOSITE LAMINATES: INTRODUCTION OF COUPLING VIA A CHANGE OF VARIABLES
STATISTICAL OPTIMIZATION OF COMPOSITE LAMINATES: INTRODUCTION OF COUPLING VIA A CHANGE OF VARIABLES Laurent Grosset,2 (AIAA member), Rodolphe Le Riche and Raphael T. Haftka 2 (AIAA fellow) CNRS URA 884
More informationEstimation-of-Distribution Algorithms. Discrete Domain.
Estimation-of-Distribution Algorithms. Discrete Domain. Petr Pošík Introduction to EDAs 2 Genetic Algorithms and Epistasis.....................................................................................
More informationSequential approaches to reliability estimation and optimization based on kriging
Sequential approaches to reliability estimation and optimization based on kriging Rodolphe Le Riche1,2 and Olivier Roustant2 1 CNRS ; 2 Ecole des Mines de Saint-Etienne JSO 2012, ONERA Palaiseau 1 Intro
More informationBehaviour of the UMDA c algorithm with truncation selection on monotone functions
Mannheim Business School Dept. of Logistics Technical Report 01/2005 Behaviour of the UMDA c algorithm with truncation selection on monotone functions Jörn Grahl, Stefan Minner, Franz Rothlauf Technical
More informationSVM optimization and Kernel methods
Announcements SVM optimization and Kernel methods w 4 is up. Due in a week. Kaggle is up 4/13/17 1 4/13/17 2 Outline Review SVM optimization Non-linear transformations in SVM Soft-margin SVM Goal: Find
More informationLinear & nonlinear classifiers
Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1396 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1396 1 / 44 Table
More informationEE613 Machine Learning for Engineers. Kernel methods Support Vector Machines. jean-marc odobez 2015
EE613 Machine Learning for Engineers Kernel methods Support Vector Machines jean-marc odobez 2015 overview Kernel methods introductions and main elements defining kernels Kernelization of k-nn, K-Means,
More informationLinear & nonlinear classifiers
Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1394 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1394 1 / 34 Table
More informationQUESTION BANK Composite Materials
QUESTION BANK Composite Materials 1. Define composite material. 2. What is the need for composite material? 3. Mention important characterits of composite material 4. Give examples for fiber material 5.
More informationLinear Regression. CSL603 - Fall 2017 Narayanan C Krishnan
Linear Regression CSL603 - Fall 2017 Narayanan C Krishnan ckn@iitrpr.ac.in Outline Univariate regression Multivariate regression Probabilistic view of regression Loss functions Bias-Variance analysis Regularization
More informationLinear Regression. CSL465/603 - Fall 2016 Narayanan C Krishnan
Linear Regression CSL465/603 - Fall 2016 Narayanan C Krishnan ckn@iitrpr.ac.in Outline Univariate regression Multivariate regression Probabilistic view of regression Loss functions Bias-Variance analysis
More informationL5 Support Vector Classification
L5 Support Vector Classification Support Vector Machine Problem definition Geometrical picture Optimization problem Optimization Problem Hard margin Convexity Dual problem Soft margin problem Alexander
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 295-P, Spring 213 Prof. Erik Sudderth Lecture 11: Inference & Learning Overview, Gaussian Graphical Models Some figures courtesy Michael Jordan s draft
More informationMixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate
Mixture Models & EM icholas Ruozzi University of Texas at Dallas based on the slides of Vibhav Gogate Previously We looed at -means and hierarchical clustering as mechanisms for unsupervised learning -means
More informationProbabilistic Model-Building Genetic Algorithms
Probabilistic Model-Building Genetic Algorithms Martin Pelikan Dept. of Math. and Computer Science University of Missouri at St. Louis St. Louis, Missouri pelikan@cs.umsl.edu Foreword! Motivation Genetic
More informationStructured Variational Inference
Structured Variational Inference Sargur srihari@cedar.buffalo.edu 1 Topics 1. Structured Variational Approximations 1. The Mean Field Approximation 1. The Mean Field Energy 2. Maximizing the energy functional:
More informationMixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate
Mixture Models & EM icholas Ruozzi University of Texas at Dallas based on the slides of Vibhav Gogate Previously We looed at -means and hierarchical clustering as mechanisms for unsupervised learning -means
More informationClustering by Mixture Models. General background on clustering Example method: k-means Mixture model based clustering Model estimation
Clustering by Mixture Models General bacground on clustering Example method: -means Mixture model based clustering Model estimation 1 Clustering A basic tool in data mining/pattern recognition: Divide
More informationMachine Learning Lecture 5
Machine Learning Lecture 5 Linear Discriminant Functions 26.10.2017 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Course Outline Fundamentals Bayes Decision Theory
More informationChris Bishop s PRML Ch. 8: Graphical Models
Chris Bishop s PRML Ch. 8: Graphical Models January 24, 2008 Introduction Visualize the structure of a probabilistic model Design and motivate new models Insights into the model s properties, in particular
More informationPreliminary statistics
1 Preliminary statistics The solution of a geophysical inverse problem can be obtained by a combination of information from observed data, the theoretical relation between data and earth parameters (models),
More informationMultivariate Gaussians. Sargur Srihari
Multivariate Gaussians Sargur srihari@cedar.buffalo.edu 1 Topics 1. Multivariate Gaussian: Basic Parameterization 2. Covariance and Information Form 3. Operations on Gaussians 4. Independencies in Gaussians
More informationLinear Models for Regression. Sargur Srihari
Linear Models for Regression Sargur srihari@cedar.buffalo.edu 1 Topics in Linear Regression What is regression? Polynomial Curve Fitting with Scalar input Linear Basis Function Models Maximum Likelihood
More informationParametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a
Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a Some slides are due to Christopher Bishop Limitations of K-means Hard assignments of data points to clusters small shift of a
More informationPerformance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project
Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Devin Cornell & Sushruth Sastry May 2015 1 Abstract In this article, we explore
More informationMaterials and Structures. Indian Institute of Technology Kanpur
Introduction to Composite Materials and Structures Nachiketa Tiwari Indian Institute of Technology Kanpur Lecture 15 Behavior of Unidirectional Composites Lecture Overview Mt Material ilaxes in unidirectional
More informationForecasting Data Streams: Next Generation Flow Field Forecasting
Forecasting Data Streams: Next Generation Flow Field Forecasting Kyle Caudle South Dakota School of Mines & Technology (SDSMT) kyle.caudle@sdsmt.edu Joint work with Michael Frey (Bucknell University) and
More informationA New Trust Region Algorithm Using Radial Basis Function Models
A New Trust Region Algorithm Using Radial Basis Function Models Seppo Pulkkinen University of Turku Department of Mathematics July 14, 2010 Outline 1 Introduction 2 Background Taylor series approximations
More informationSupport Vector Machines: Maximum Margin Classifiers
Support Vector Machines: Maximum Margin Classifiers Machine Learning and Pattern Recognition: September 16, 2008 Piotr Mirowski Based on slides by Sumit Chopra and Fu-Jie Huang 1 Outline What is behind
More informationSupport Vector Machine (SVM) and Kernel Methods
Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2014 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationStatistical Machine Learning from Data
Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Support Vector Machines Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole Polytechnique
More informationHidden Markov Models and Gaussian Mixture Models
Hidden Markov Models and Gaussian Mixture Models Hiroshi Shimodaira and Steve Renals Automatic Speech Recognition ASR Lectures 4&5 23&27 January 2014 ASR Lectures 4&5 Hidden Markov Models and Gaussian
More informationOPTIMISATION OF LAMINATED COMPOSITE PLATES CONSIDERING DIFFERENT FAILURE CRITERIA
OPTIMIATION OF LAMINATED COMPOITE PLATE CONIDERING DIFFERENT FAILURE CRITERIA R.H. Lopez 1, M.A. Luersen, J.E.ouza de Cursi 1 1 Laboratoire de Mécanique de Rouen, Institut National des ciences Appliquées
More informationStatistical Pattern Recognition
Statistical Pattern Recognition Expectation Maximization (EM) and Mixture Models Hamid R. Rabiee Jafar Muhammadi, Mohammad J. Hosseini Spring 2014 http://ce.sharif.edu/courses/92-93/2/ce725-2 Agenda Expectation-maximization
More informationLecture 13 : Variational Inference: Mean Field Approximation
10-708: Probabilistic Graphical Models 10-708, Spring 2017 Lecture 13 : Variational Inference: Mean Field Approximation Lecturer: Willie Neiswanger Scribes: Xupeng Tong, Minxing Liu 1 Problem Setup 1.1
More informationProbability and Information Theory. Sargur N. Srihari
Probability and Information Theory Sargur N. srihari@cedar.buffalo.edu 1 Topics in Probability and Information Theory Overview 1. Why Probability? 2. Random Variables 3. Probability Distributions 4. Marginal
More informationSupport Vector Machine (SVM) and Kernel Methods
Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2016 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationData Mining. Linear & nonlinear classifiers. Hamid Beigy. Sharif University of Technology. Fall 1396
Data Mining Linear & nonlinear classifiers Hamid Beigy Sharif University of Technology Fall 1396 Hamid Beigy (Sharif University of Technology) Data Mining Fall 1396 1 / 31 Table of contents 1 Introduction
More informationGradient Descent. Dr. Xiaowei Huang
Gradient Descent Dr. Xiaowei Huang https://cgi.csc.liv.ac.uk/~xiaowei/ Up to now, Three machine learning algorithms: decision tree learning k-nn linear regression only optimization objectives are discussed,
More informationDensity Estimation. Seungjin Choi
Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/
More informationLecture Note 1: Probability Theory and Statistics
Univ. of Michigan - NAME 568/EECS 568/ROB 530 Winter 2018 Lecture Note 1: Probability Theory and Statistics Lecturer: Maani Ghaffari Jadidi Date: April 6, 2018 For this and all future notes, if you would
More informationCSC2515 Winter 2015 Introduction to Machine Learning. Lecture 2: Linear regression
CSC2515 Winter 2015 Introduction to Machine Learning Lecture 2: Linear regression All lecture slides will be available as.pdf on the course website: http://www.cs.toronto.edu/~urtasun/courses/csc2515/csc2515_winter15.html
More informationPredicting Failure of Multiangle Composite Laminates
Predicting Failure of Multiangle Composite Laminates Preliminary discussion (not in textbook): Micromechanics failure analyses vs Macromechanics failure analyses Fiber Architecture of Some Common Composite
More informationExpectation propagation for signal detection in flat-fading channels
Expectation propagation for signal detection in flat-fading channels Yuan Qi MIT Media Lab Cambridge, MA, 02139 USA yuanqi@media.mit.edu Thomas Minka CMU Statistics Department Pittsburgh, PA 15213 USA
More informationS.Nagendra, D.Jestin, Z.Gurdal, R.T.Haftka and L.T.Watson Computers & Structures, Vol. 58, No. 3, pp , 1996.
S.Nagendra, D.Jestin, Z.Gurdal, R.T.Haftka and L.T.Watson Computers & Structures, Vol. 58, No. 3, pp. 543-555, 1996. Presented by Vignesh Solai Rameshbabu Introduction Simple programming technique which
More informationPartially Observable Markov Decision Processes (POMDPs)
Partially Observable Markov Decision Processes (POMDPs) Sachin Patil Guest Lecture: CS287 Advanced Robotics Slides adapted from Pieter Abbeel, Alex Lee Outline Introduction to POMDPs Locally Optimal Solutions
More informationCS-E4830 Kernel Methods in Machine Learning
CS-E4830 Kernel Methods in Machine Learning Lecture 3: Convex optimization and duality Juho Rousu 27. September, 2017 Juho Rousu 27. September, 2017 1 / 45 Convex optimization Convex optimisation This
More informationLog-Linear Models, MEMMs, and CRFs
Log-Linear Models, MEMMs, and CRFs Michael Collins 1 Notation Throughout this note I ll use underline to denote vectors. For example, w R d will be a vector with components w 1, w 2,... w d. We use expx
More informationRecita,on: Loss, Regulariza,on, and Dual*
10-701 Recita,on: Loss, Regulariza,on, and Dual* Jay- Yoon Lee 02/26/2015 *Adopted figures from 10725 lecture slides and from the book Elements of Sta,s,cal Learning Loss and Regulariza,on Op,miza,on problem
More informationSupport Vector Machine (SVM) and Kernel Methods
Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2015 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationGaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008
Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:
More informationToday. Calculus. Linear Regression. Lagrange Multipliers
Today Calculus Lagrange Multipliers Linear Regression 1 Optimization with constraints What if I want to constrain the parameters of the model. The mean is less than 10 Find the best likelihood, subject
More information13: Variational inference II
10-708: Probabilistic Graphical Models, Spring 2015 13: Variational inference II Lecturer: Eric P. Xing Scribes: Ronghuo Zheng, Zhiting Hu, Yuntian Deng 1 Introduction We started to talk about variational
More informationSupport Vector Machines
Support Vector Machines Le Song Machine Learning I CSE 6740, Fall 2013 Naïve Bayes classifier Still use Bayes decision rule for classification P y x = P x y P y P x But assume p x y = 1 is fully factorized
More informationCSE 254: MAP estimation via agreement on (hyper)trees: Message-passing and linear programming approaches
CSE 254: MAP estimation via agreement on (hyper)trees: Message-passing and linear programming approaches A presentation by Evan Ettinger November 11, 2005 Outline Introduction Motivation and Background
More informationClustering and Gaussian Mixture Models
Clustering and Gaussian Mixture Models Piyush Rai IIT Kanpur Probabilistic Machine Learning (CS772A) Jan 25, 2016 Probabilistic Machine Learning (CS772A) Clustering and Gaussian Mixture Models 1 Recap
More informationMultiple Random Variables
Multiple Random Variables This Version: July 30, 2015 Multiple Random Variables 2 Now we consider models with more than one r.v. These are called multivariate models For instance: height and weight An
More informationMathematical Preliminaries
Mathematical Preliminaries Economics 3307 - Intermediate Macroeconomics Aaron Hedlund Baylor University Fall 2013 Econ 3307 (Baylor University) Mathematical Preliminaries Fall 2013 1 / 25 Outline I: Sequences
More informationSensitivity and Reliability Analysis of Nonlinear Frame Structures
Sensitivity and Reliability Analysis of Nonlinear Frame Structures Michael H. Scott Associate Professor School of Civil and Construction Engineering Applied Mathematics and Computation Seminar April 8,
More informationCPSC 340: Machine Learning and Data Mining. More PCA Fall 2017
CPSC 340: Machine Learning and Data Mining More PCA Fall 2017 Admin Assignment 4: Due Friday of next week. No class Monday due to holiday. There will be tutorials next week on MAP/PCA (except Monday).
More informationCS839: Probabilistic Graphical Models. Lecture 7: Learning Fully Observed BNs. Theo Rekatsinas
CS839: Probabilistic Graphical Models Lecture 7: Learning Fully Observed BNs Theo Rekatsinas 1 Exponential family: a basic building block For a numeric random variable X p(x ) =h(x)exp T T (x) A( ) = 1
More informationComposite Structural Mechanics using MATLAB
Session 2520 Composite Structural Mechanics using MATLAB Oscar Barton, Jr., Jacob B. Wallace United States Naval Academy Annapolis, Md 21402 Abstract In this paper MATLAB is adopted as the programming
More informationKernel Methods. Machine Learning A W VO
Kernel Methods Machine Learning A 708.063 07W VO Outline 1. Dual representation 2. The kernel concept 3. Properties of kernels 4. Examples of kernel machines Kernel PCA Support vector regression (Relevance
More informationAdvanced Optimization
Advanced Optimization Lecture 3: 1: Randomized Algorithms for for Continuous Discrete Problems Problems November 22, 2016 Master AIC Université Paris-Saclay, Orsay, France Anne Auger INRIA Saclay Ile-de-France
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 12: Gaussian Belief Propagation, State Space Models and Kalman Filters Guest Kalman Filter Lecture by
More informationWritten Examination
Division of Scientific Computing Department of Information Technology Uppsala University Optimization Written Examination 202-2-20 Time: 4:00-9:00 Allowed Tools: Pocket Calculator, one A4 paper with notes
More informationProbabilistic Model-Building Genetic Algorithms
Probabilistic Model-Building Genetic Algorithms a.k.a. Estimation of Distribution Algorithms a.k.a. Iterated Density Estimation Algorithms Martin Pelikan Foreword Motivation Genetic and evolutionary computation
More informationData Mining and Analysis: Fundamental Concepts and Algorithms
Data Mining and Analysis: Fundamental Concepts and Algorithms dataminingbook.info Mohammed J. Zaki 1 Wagner Meira Jr. 2 1 Department of Computer Science Rensselaer Polytechnic Institute, Troy, NY, USA
More informationLinear Classification. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington
Linear Classification CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 Example of Linear Classification Red points: patterns belonging
More informationECE521 lecture 4: 19 January Optimization, MLE, regularization
ECE521 lecture 4: 19 January 2017 Optimization, MLE, regularization First four lectures Lectures 1 and 2: Intro to ML Probability review Types of loss functions and algorithms Lecture 3: KNN Convexity
More informationSupport Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012
Support Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Linear classifier Which classifier? x 2 x 1 2 Linear classifier Margin concept x 2
More informationCh 4. Linear Models for Classification
Ch 4. Linear Models for Classification Pattern Recognition and Machine Learning, C. M. Bishop, 2006. Department of Computer Science and Engineering Pohang University of Science and echnology 77 Cheongam-ro,
More informationPILCO: A Model-Based and Data-Efficient Approach to Policy Search
PILCO: A Model-Based and Data-Efficient Approach to Policy Search (M.P. Deisenroth and C.E. Rasmussen) CSC2541 November 4, 2016 PILCO Graphical Model PILCO Probabilistic Inference for Learning COntrol
More informationGaussian processes and bayesian optimization Stanisław Jastrzębski. kudkudak.github.io kudkudak
Gaussian processes and bayesian optimization Stanisław Jastrzębski kudkudak.github.io kudkudak Plan Goal: talk about modern hyperparameter optimization algorithms Bayes reminder: equivalent linear regression
More informationIntroduction Linear system Nonlinear equation Interpolation
Interpolation Interpolation is the process of estimating an intermediate value from a set of discrete or tabulated values. Suppose we have the following tabulated values: y y 0 y 1 y 2?? y 3 y 4 y 5 x
More informationMultivariate Distributions CIVL 7012/8012
Multivariate Distributions CIVL 7012/8012 Multivariate Distributions Engineers often are interested in more than one measurement from a single item. Multivariate distributions describe the probability
More informationDiscrete Mathematics and Probability Theory Fall 2015 Lecture 21
CS 70 Discrete Mathematics and Probability Theory Fall 205 Lecture 2 Inference In this note we revisit the problem of inference: Given some data or observations from the world, what can we infer about
More informationJoint Gaussian Graphical Model Review Series I
Joint Gaussian Graphical Model Review Series I Probability Foundations Beilun Wang Advisor: Yanjun Qi 1 Department of Computer Science, University of Virginia http://jointggm.org/ June 23rd, 2017 Beilun
More informationCSE446: Clustering and EM Spring 2017
CSE446: Clustering and EM Spring 2017 Ali Farhadi Slides adapted from Carlos Guestrin, Dan Klein, and Luke Zettlemoyer Clustering systems: Unsupervised learning Clustering Detect patterns in unlabeled
More information{ p if x = 1 1 p if x = 0
Discrete random variables Probability mass function Given a discrete random variable X taking values in X = {v 1,..., v m }, its probability mass function P : X [0, 1] is defined as: P (v i ) = Pr[X =
More informationLearning Bayesian networks
1 Lecture topics: Learning Bayesian networks from data maximum likelihood, BIC Bayesian, marginal likelihood Learning Bayesian networks There are two problems we have to solve in order to estimate Bayesian
More information1 EM algorithm: updating the mixing proportions {π k } ik are the posterior probabilities at the qth iteration of EM.
Université du Sud Toulon - Var Master Informatique Probabilistic Learning and Data Analysis TD: Model-based clustering by Faicel CHAMROUKHI Solution The aim of this practical wor is to show how the Classification
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear
More informationVariational Inference (11/04/13)
STA561: Probabilistic machine learning Variational Inference (11/04/13) Lecturer: Barbara Engelhardt Scribes: Matt Dickenson, Alireza Samany, Tracy Schifeling 1 Introduction In this lecture we will further
More informationComputational statistics
Computational statistics Combinatorial optimization Thierry Denœux February 2017 Thierry Denœux Computational statistics February 2017 1 / 37 Combinatorial optimization Assume we seek the maximum of f
More informationA Progressive Hedging Approach to Multistage Stochastic Generation and Transmission Investment Planning
A Progressive Hedging Approach to Multistage Stochastic Generation and Transmission Investment Planning Yixian Liu Ramteen Sioshansi Integrated Systems Engineering Department The Ohio State University
More informationChapter 5 continued. Chapter 5 sections
Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationAn Information Geometry Perspective on Estimation of Distribution Algorithms: Boundary Analysis
An Information Geometry Perspective on Estimation of Distribution Algorithms: Boundary Analysis Luigi Malagò Department of Electronics and Information Politecnico di Milano Via Ponzio, 34/5 20133 Milan,
More informationGeneralized Method of Moments Estimation
Generalized Method of Moments Estimation Lars Peter Hansen March 0, 2007 Introduction Generalized methods of moments (GMM) refers to a class of estimators which are constructed from exploiting the sample
More informationLet X and Y be two real valued stochastic variables defined on (Ω, F, P). Theorem: If X and Y are independent then. . p.1/21
Multivariate transformations The remaining part of the probability course is centered around transformations t : R k R m and how they transform probability measures. For instance tx 1,...,x k = x 1 +...
More informationMulti-Objective Optimization of Carbon Fibre Reinforced Plastic (CFRP) Circular Hollow Section Using Genetic Algorithm for Engineering Structures
International Refereed Journal of Engineering and Science (IRJES) ISSN (Online) 2319-183X, (rint) 2319-1821 Volume 4, Issue 12 (December 2015),.24-28 Multi-Objective Optimization of Carbon Fibre Reinforced
More informationNONLINEAR. (Hillier & Lieberman Introduction to Operations Research, 8 th edition)
NONLINEAR PROGRAMMING (Hillier & Lieberman Introduction to Operations Research, 8 th edition) Nonlinear Programming g Linear programming has a fundamental role in OR. In linear programming all its functions
More information3 : Representation of Undirected GM
10-708: Probabilistic Graphical Models 10-708, Spring 2016 3 : Representation of Undirected GM Lecturer: Eric P. Xing Scribes: Longqi Cai, Man-Chia Chang 1 MRF vs BN There are two types of graphical models:
More informationSupport Vector Machine (continued)
Support Vector Machine continued) Overlapping class distribution: In practice the class-conditional distributions may overlap, so that the training data points are no longer linearly separable. We need
More informationPrincipal Components Analysis. Sargur Srihari University at Buffalo
Principal Components Analysis Sargur Srihari University at Buffalo 1 Topics Projection Pursuit Methods Principal Components Examples of using PCA Graphical use of PCA Multidimensional Scaling Srihari 2
More informationMaximum Likelihood Estimation. only training data is available to design a classifier
Introduction to Pattern Recognition [ Part 5 ] Mahdi Vasighi Introduction Bayesian Decision Theory shows that we could design an optimal classifier if we knew: P( i ) : priors p(x i ) : class-conditional
More informationNeutron inverse kinetics via Gaussian Processes
Neutron inverse kinetics via Gaussian Processes P. Picca Politecnico di Torino, Torino, Italy R. Furfaro University of Arizona, Tucson, Arizona Outline Introduction Review of inverse kinetics techniques
More informationChapter 6 - Random Processes
EE385 Class Notes //04 John Stensby Chapter 6 - Random Processes Recall that a random variable X is a mapping between the sample space S and the extended real line R +. That is, X : S R +. A random process
More information