A stochastic global optimization method for (urban?) multi-parameter systems. Rodolphe Le Riche CNRS and Ecole des Mines de St-Etienne

Size: px
Start display at page:

Download "A stochastic global optimization method for (urban?) multi-parameter systems. Rodolphe Le Riche CNRS and Ecole des Mines de St-Etienne"

Transcription

1 A stochastic global optimization method for (urban?) multi-parameter systems Rodolphe Le Riche CNRS and Ecole des Mines de St-Etienne

2 Purpose of this presentation Discussion object : cities as an entity for heating consumption and related public policies. Cities are analyzed, modeled and optimized for heat consumption (including the effects of solar passive input). Cities are objects that are described at different parameterization levels : entire city (coarser level), districts, buildings (finer level). In mechanical engineering, multi-levels of parameters are often associated to composite materials. Talk : describe an idea for multi-level optimization that emerged in composite design. Any use for cities?

3 Notation and application example Composite design at various scales min f ( y ( x)) macro x S y(x) numerical simulator of the structure (stress, strains, strength, mass,...) meso, v ( plate stiffnesses ) v(x) micro, x ( fiber positions ) v(x) is numerically inexpensive. x(v) doesn't exist : there are many fiber positions for one choice of plate stiffnesses.

4 Notation and application example Heating consumption at various scales min f ( y ( x)) macro city x S y(x) numerical simulator of the structure (solar energy, thermal simulation) meso, v «IRIS» (district) micro, x building IRIS : Ilôt regroupé pour l'information statistique

5 Scope of application : global, derivative free problems Here, we focus on f min f x x S S ℝ n or ℕn or { ℝ n1 ℕn2 } i.e., continuous or integer or mixte optimization. local optimum x l V xl x * global optimum

6 Flow chart of a general stochastic optimizer Initialize t and pt(x) Sample Calculate f(xt+1) Update the distribution xt+1 ~ pt(x) pt+1(x) = Update( x1, f(x1),, xt+1, f(xt+1) ) or more often pt+1(x) = Update( pt(x), xt+1, f(xt+1) ) Stop or [ t = t+1 and go back to Sample ] with different p's if x is continuous or discrete or mixed.

7 A simple example in Rn : ES-(1+1) Initializations : x, f(x), m, C, While tmax. t < tmax do, Sample N(m,C) --> x' Calculate f(x'), t = t+1 If f(x')<f(x), x = x', f(x) = f(x') Endif Update m (e.g., m=x) and C End while Normal law N(m,C) [ 2 σ C = 0 σ σ ] for variables seen as independent.

8 Illustration : adaptation of 2D Gaussian with ES-(1+1)

9 Discrete variables : The Univariate Marginal Density Algorithm (UMDA) ( Baluja 1994 as PBIL and Mühlenbein 1996) x S {1,2,, A}n (alphabet of cardinality A ) o o o o n e.g. { 45, 0, 45, 90 } (fiber orientations) n e.g. {matl1,, matla } (material choice) The algorithm is that of a population based stochastic optimization (see before + many x's at each iteration) with different sampling and updating of pt. pt assumes that the variables are independent (drop t ), n p x = p i x i i=1 0.6 pi p2i pia p1i 1 A pij = 1 j= A xi

10 UMDA (2) Sampling : For i=1, n ui ~ U [0,1] If 0 ui p1i x i =1 u 1 0 k 2 p 1 A p +p k If 1 k 1 p ij ui pij j=1 A 1 If j=1 j=1 x i =k j p i ui 1 x i =A Learning : Select the μ best points out of λ, f x 1 : f x 2 : f x : pij is the frequency of j at position i in the bests : j pi = k =1 : k I x i = j 1 p 1 j i, k k I x i = j =1 if x i = j, =0 otherwise (minimum frequency for ergodicity)

11 Application to composite design for frequency density learned by UMDA (2D) contour lines of the penalized objective function Independent densities can neither represent curvatures nor variables' couplings. ( from Grosset, L., Le Riche, R. and Haftka, R.T., A double-distribution statistical algorithm for composite laminate optimization, SMO, 2006 )

12 Stochastic discrete optimization : learning the variables dependencies More sophisticated discrete optimization methods attempt to learn the couplings between variables. For example, with pairwise dependencies : X1:n X2:n... Xn:n p( x ) = p ( x 1 : n ) p( x 2: n x 1: n ) p ( x n : n x n 1: n ) Trade-off : richer probabilistic structures better capture the objective function landscape but they also have more parameters need more f evaluations to be learned. MIMIC ( Mutual Information Maximizing Input Clustering ) algorithm : De Bonnet, Isbell and Viola, BMDA ( Bivariate Marginal Distribution Algorithm ) : Pelikan and Muehlenbein, 1999.

13 Multi-level parameter optimization with DDOA ( from Grosset, L., Le Riche, R. and Haftka, R.T., A double-distribution statistical algorithm for composite laminate optimization, SMO, 2006 ) Mathematical motivation : create couplings between variables using many independent distributions (in x and v spaces). Numerical motivation : take into account expert knowledge in the optimization to improve efficiency. E.g. in composites, the lamination parameters v (the plate stiffnesses) make physical sense.

14 Example in composites Use of the lamination parameters v = lamination parameters = geometric contribution of the plies to the stiffness. Inexpensive to calculate from x (fiber angles). Simplifications : fewer v's than fiber angles. Often, the v's are taken as continuous. But f(v) typically does not exist (e.g., ply failure criterion).

15 Past examples in composites ( Liu, Haftka, and Akgün, «Two-level composite wing structural optimization using response surfaces», Merval, Samuelides and Grihon, «Lagrange-Kuhn-Tucker coordination for multilevel optimization of aeronautical structures», ) Initial problem : Optimize a composite structure made of several assembled panels by changing each ply orientation many discrete variables Decomposed problem : Structure level Optimize a composite structure made of several assembled panels by changing the lamination parameters of each panel few continuous variables optimal v's Laminate level Minimize the distance to target lamination parameters by changing the ply orientations few discrete variables BUT for such a sequential approach to make sense, f (v ) must exist and guide to optimal regions (i.e., prohibits emergence of solutions at finer scales).

16 The DDOA stochastic optimization algorithm objective function If v(x) is costless, it does not cost to learn densities in the x AND v spaces at the same time. does not exist p(x) pddoa ( x) = p X v ( X )=V (x ) x v v(x)

17 The DDOA algorithm : X v(x)=v? Simple mathematical illustration : 1 1 T p X (x ) = exp x x 2π 2 ( ) v = x1+ x pv (v ) = exp (v 1) 2 2 π ( ) Intermediate step for a given v : p X v( X )= v=1 ( x) is a degenerated Gaussian along x1+ x 2 = 1 (cross-section of the 2D bell curve along the blue line + normalization)

18 The DDOA algorithm : X v(x)=v? p X v ( X )=V (x ) = p X v( X )=v pv (v (x )) =... = exp ( x1 x 2 )2 ( x 1 + x 2 1)2 2π 4 2 ( ) p X v( X )=V ( x) is a coupled distribution that merges the effects of X and V Analytical calculation in the Gaussian case. In practice, use simulations...

19 The DDOA algorithm (flow chart) Choose λ, μ, ρ such that ρ>>1 and λ>μ Initialize pv(v)and px(x) For i=1,λ do Sample vtarget from pv(v) Sample ρ>>1 x's from px(x) x(i) = the closest x to vtarget Calculate f(x(i)) sampling of X v(x)=v end For Rank x(1:λ),, x(λ:λ) the proposed points Update pv(v) and px(x) from x(1:λ),, x(μ:λ) Stop? If no, go back to top...

20 Application of DDOA to composite design for frequency px(x) and pv(v) can be simple densities, without variables couplings ( easy to learn), yet pddoa(x) is a coupled density. f(x) and selected points n p X x = i =1 pi x i pddoa(x) One half of the algorithm searches in a low dimension space.

21 additional slides

22 Introduction to stochastic optimization Random numbers are versatile search engines (work both in Rn and / or Nn ). They can also yield efficient methods. Let pt(x) denote the probability density function of x at iteration t (e.g., after t evaluations of f). It represents the belief at t that the optimum x* is at x. How to «sample pt(x)» once (Scilab notation)? if x is uniform between m and M, X ~ U[m,M], call x = m + rand(n,1).*(m-m) if x is (multi-)gaussian with mean m and covariance matrix C, X ~ N(m,C),call x = m + grand(1,'mn',0,c)

23 Application to a laminate frequency problem (1) ( from Grosset, L., Le Riche, R. and Haftka, R.T., A double-distribution statistical algorithm for composite laminate optimization, SMO, 2006 ) max x f 1 x 1,, x 15, the first eigenfreq. of a simply supported plate such that 0.48 eff x 0.52 where x i {0 o, 15o,, 90o } the constraint is enforced by penalty and creates a narrow ridge in the design space

24 Application to a laminate frequency problem (2) Optimum : [90o4 /±75o /±60o2 /±45o5 /±305o ]s Compare UMDA to a GA (genetic algorithm) and SHC (Stochastic Hill Climber) Reliability = probability of finding the optimum at a given cost. UMDA performs fairly well on this problem.

25 Example in composites Use of the lamination parameters Simplifications : fewer v's than fiber angles. Often, the v's are taken as continuous. But f(v) typically does not exist (e.g., ply failure criterion).

Discrete stochastic optimization with continuous auxiliary variables. Rodolphe Le Riche*, Alexis Lasseigne**, François-Xavier Irisarri**

Discrete stochastic optimization with continuous auxiliary variables. Rodolphe Le Riche*, Alexis Lasseigne**, François-Xavier Irisarri** Discrete stochastic optimization with continuous auxiliary variables Rodolphe Le Riche*, Alexis Lasseigne**, François-Xavier Irisarri** *CNRS and Ecole des Mines de St-Etienne, Fr. ** ONERA, Composite

More information

STATISTICAL OPTIMIZATION OF COMPOSITE LAMINATES: INTRODUCTION OF COUPLING VIA A CHANGE OF VARIABLES

STATISTICAL OPTIMIZATION OF COMPOSITE LAMINATES: INTRODUCTION OF COUPLING VIA A CHANGE OF VARIABLES STATISTICAL OPTIMIZATION OF COMPOSITE LAMINATES: INTRODUCTION OF COUPLING VIA A CHANGE OF VARIABLES Laurent Grosset,2 (AIAA member), Rodolphe Le Riche and Raphael T. Haftka 2 (AIAA fellow) CNRS URA 884

More information

Estimation-of-Distribution Algorithms. Discrete Domain.

Estimation-of-Distribution Algorithms. Discrete Domain. Estimation-of-Distribution Algorithms. Discrete Domain. Petr Pošík Introduction to EDAs 2 Genetic Algorithms and Epistasis.....................................................................................

More information

Sequential approaches to reliability estimation and optimization based on kriging

Sequential approaches to reliability estimation and optimization based on kriging Sequential approaches to reliability estimation and optimization based on kriging Rodolphe Le Riche1,2 and Olivier Roustant2 1 CNRS ; 2 Ecole des Mines de Saint-Etienne JSO 2012, ONERA Palaiseau 1 Intro

More information

Behaviour of the UMDA c algorithm with truncation selection on monotone functions

Behaviour of the UMDA c algorithm with truncation selection on monotone functions Mannheim Business School Dept. of Logistics Technical Report 01/2005 Behaviour of the UMDA c algorithm with truncation selection on monotone functions Jörn Grahl, Stefan Minner, Franz Rothlauf Technical

More information

SVM optimization and Kernel methods

SVM optimization and Kernel methods Announcements SVM optimization and Kernel methods w 4 is up. Due in a week. Kaggle is up 4/13/17 1 4/13/17 2 Outline Review SVM optimization Non-linear transformations in SVM Soft-margin SVM Goal: Find

More information

Linear & nonlinear classifiers

Linear & nonlinear classifiers Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1396 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1396 1 / 44 Table

More information

EE613 Machine Learning for Engineers. Kernel methods Support Vector Machines. jean-marc odobez 2015

EE613 Machine Learning for Engineers. Kernel methods Support Vector Machines. jean-marc odobez 2015 EE613 Machine Learning for Engineers Kernel methods Support Vector Machines jean-marc odobez 2015 overview Kernel methods introductions and main elements defining kernels Kernelization of k-nn, K-Means,

More information

Linear & nonlinear classifiers

Linear & nonlinear classifiers Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1394 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1394 1 / 34 Table

More information

QUESTION BANK Composite Materials

QUESTION BANK Composite Materials QUESTION BANK Composite Materials 1. Define composite material. 2. What is the need for composite material? 3. Mention important characterits of composite material 4. Give examples for fiber material 5.

More information

Linear Regression. CSL603 - Fall 2017 Narayanan C Krishnan

Linear Regression. CSL603 - Fall 2017 Narayanan C Krishnan Linear Regression CSL603 - Fall 2017 Narayanan C Krishnan ckn@iitrpr.ac.in Outline Univariate regression Multivariate regression Probabilistic view of regression Loss functions Bias-Variance analysis Regularization

More information

Linear Regression. CSL465/603 - Fall 2016 Narayanan C Krishnan

Linear Regression. CSL465/603 - Fall 2016 Narayanan C Krishnan Linear Regression CSL465/603 - Fall 2016 Narayanan C Krishnan ckn@iitrpr.ac.in Outline Univariate regression Multivariate regression Probabilistic view of regression Loss functions Bias-Variance analysis

More information

L5 Support Vector Classification

L5 Support Vector Classification L5 Support Vector Classification Support Vector Machine Problem definition Geometrical picture Optimization problem Optimization Problem Hard margin Convexity Dual problem Soft margin problem Alexander

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Brown University CSCI 295-P, Spring 213 Prof. Erik Sudderth Lecture 11: Inference & Learning Overview, Gaussian Graphical Models Some figures courtesy Michael Jordan s draft

More information

Mixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate

Mixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate Mixture Models & EM icholas Ruozzi University of Texas at Dallas based on the slides of Vibhav Gogate Previously We looed at -means and hierarchical clustering as mechanisms for unsupervised learning -means

More information

Probabilistic Model-Building Genetic Algorithms

Probabilistic Model-Building Genetic Algorithms Probabilistic Model-Building Genetic Algorithms Martin Pelikan Dept. of Math. and Computer Science University of Missouri at St. Louis St. Louis, Missouri pelikan@cs.umsl.edu Foreword! Motivation Genetic

More information

Structured Variational Inference

Structured Variational Inference Structured Variational Inference Sargur srihari@cedar.buffalo.edu 1 Topics 1. Structured Variational Approximations 1. The Mean Field Approximation 1. The Mean Field Energy 2. Maximizing the energy functional:

More information

Mixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate

Mixture Models & EM. Nicholas Ruozzi University of Texas at Dallas. based on the slides of Vibhav Gogate Mixture Models & EM icholas Ruozzi University of Texas at Dallas based on the slides of Vibhav Gogate Previously We looed at -means and hierarchical clustering as mechanisms for unsupervised learning -means

More information

Clustering by Mixture Models. General background on clustering Example method: k-means Mixture model based clustering Model estimation

Clustering by Mixture Models. General background on clustering Example method: k-means Mixture model based clustering Model estimation Clustering by Mixture Models General bacground on clustering Example method: -means Mixture model based clustering Model estimation 1 Clustering A basic tool in data mining/pattern recognition: Divide

More information

Machine Learning Lecture 5

Machine Learning Lecture 5 Machine Learning Lecture 5 Linear Discriminant Functions 26.10.2017 Bastian Leibe RWTH Aachen http://www.vision.rwth-aachen.de leibe@vision.rwth-aachen.de Course Outline Fundamentals Bayes Decision Theory

More information

Chris Bishop s PRML Ch. 8: Graphical Models

Chris Bishop s PRML Ch. 8: Graphical Models Chris Bishop s PRML Ch. 8: Graphical Models January 24, 2008 Introduction Visualize the structure of a probabilistic model Design and motivate new models Insights into the model s properties, in particular

More information

Preliminary statistics

Preliminary statistics 1 Preliminary statistics The solution of a geophysical inverse problem can be obtained by a combination of information from observed data, the theoretical relation between data and earth parameters (models),

More information

Multivariate Gaussians. Sargur Srihari

Multivariate Gaussians. Sargur Srihari Multivariate Gaussians Sargur srihari@cedar.buffalo.edu 1 Topics 1. Multivariate Gaussian: Basic Parameterization 2. Covariance and Information Form 3. Operations on Gaussians 4. Independencies in Gaussians

More information

Linear Models for Regression. Sargur Srihari

Linear Models for Regression. Sargur Srihari Linear Models for Regression Sargur srihari@cedar.buffalo.edu 1 Topics in Linear Regression What is regression? Polynomial Curve Fitting with Scalar input Linear Basis Function Models Maximum Likelihood

More information

Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a

Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a Some slides are due to Christopher Bishop Limitations of K-means Hard assignments of data points to clusters small shift of a

More information

Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project

Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Devin Cornell & Sushruth Sastry May 2015 1 Abstract In this article, we explore

More information

Materials and Structures. Indian Institute of Technology Kanpur

Materials and Structures. Indian Institute of Technology Kanpur Introduction to Composite Materials and Structures Nachiketa Tiwari Indian Institute of Technology Kanpur Lecture 15 Behavior of Unidirectional Composites Lecture Overview Mt Material ilaxes in unidirectional

More information

Forecasting Data Streams: Next Generation Flow Field Forecasting

Forecasting Data Streams: Next Generation Flow Field Forecasting Forecasting Data Streams: Next Generation Flow Field Forecasting Kyle Caudle South Dakota School of Mines & Technology (SDSMT) kyle.caudle@sdsmt.edu Joint work with Michael Frey (Bucknell University) and

More information

A New Trust Region Algorithm Using Radial Basis Function Models

A New Trust Region Algorithm Using Radial Basis Function Models A New Trust Region Algorithm Using Radial Basis Function Models Seppo Pulkkinen University of Turku Department of Mathematics July 14, 2010 Outline 1 Introduction 2 Background Taylor series approximations

More information

Support Vector Machines: Maximum Margin Classifiers

Support Vector Machines: Maximum Margin Classifiers Support Vector Machines: Maximum Margin Classifiers Machine Learning and Pattern Recognition: September 16, 2008 Piotr Mirowski Based on slides by Sumit Chopra and Fu-Jie Huang 1 Outline What is behind

More information

Support Vector Machine (SVM) and Kernel Methods

Support Vector Machine (SVM) and Kernel Methods Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2014 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin

More information

Statistical Machine Learning from Data

Statistical Machine Learning from Data Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Support Vector Machines Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole Polytechnique

More information

Hidden Markov Models and Gaussian Mixture Models

Hidden Markov Models and Gaussian Mixture Models Hidden Markov Models and Gaussian Mixture Models Hiroshi Shimodaira and Steve Renals Automatic Speech Recognition ASR Lectures 4&5 23&27 January 2014 ASR Lectures 4&5 Hidden Markov Models and Gaussian

More information

OPTIMISATION OF LAMINATED COMPOSITE PLATES CONSIDERING DIFFERENT FAILURE CRITERIA

OPTIMISATION OF LAMINATED COMPOSITE PLATES CONSIDERING DIFFERENT FAILURE CRITERIA OPTIMIATION OF LAMINATED COMPOITE PLATE CONIDERING DIFFERENT FAILURE CRITERIA R.H. Lopez 1, M.A. Luersen, J.E.ouza de Cursi 1 1 Laboratoire de Mécanique de Rouen, Institut National des ciences Appliquées

More information

Statistical Pattern Recognition

Statistical Pattern Recognition Statistical Pattern Recognition Expectation Maximization (EM) and Mixture Models Hamid R. Rabiee Jafar Muhammadi, Mohammad J. Hosseini Spring 2014 http://ce.sharif.edu/courses/92-93/2/ce725-2 Agenda Expectation-maximization

More information

Lecture 13 : Variational Inference: Mean Field Approximation

Lecture 13 : Variational Inference: Mean Field Approximation 10-708: Probabilistic Graphical Models 10-708, Spring 2017 Lecture 13 : Variational Inference: Mean Field Approximation Lecturer: Willie Neiswanger Scribes: Xupeng Tong, Minxing Liu 1 Problem Setup 1.1

More information

Probability and Information Theory. Sargur N. Srihari

Probability and Information Theory. Sargur N. Srihari Probability and Information Theory Sargur N. srihari@cedar.buffalo.edu 1 Topics in Probability and Information Theory Overview 1. Why Probability? 2. Random Variables 3. Probability Distributions 4. Marginal

More information

Support Vector Machine (SVM) and Kernel Methods

Support Vector Machine (SVM) and Kernel Methods Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2016 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin

More information

Data Mining. Linear & nonlinear classifiers. Hamid Beigy. Sharif University of Technology. Fall 1396

Data Mining. Linear & nonlinear classifiers. Hamid Beigy. Sharif University of Technology. Fall 1396 Data Mining Linear & nonlinear classifiers Hamid Beigy Sharif University of Technology Fall 1396 Hamid Beigy (Sharif University of Technology) Data Mining Fall 1396 1 / 31 Table of contents 1 Introduction

More information

Gradient Descent. Dr. Xiaowei Huang

Gradient Descent. Dr. Xiaowei Huang Gradient Descent Dr. Xiaowei Huang https://cgi.csc.liv.ac.uk/~xiaowei/ Up to now, Three machine learning algorithms: decision tree learning k-nn linear regression only optimization objectives are discussed,

More information

Density Estimation. Seungjin Choi

Density Estimation. Seungjin Choi Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/

More information

Lecture Note 1: Probability Theory and Statistics

Lecture Note 1: Probability Theory and Statistics Univ. of Michigan - NAME 568/EECS 568/ROB 530 Winter 2018 Lecture Note 1: Probability Theory and Statistics Lecturer: Maani Ghaffari Jadidi Date: April 6, 2018 For this and all future notes, if you would

More information

CSC2515 Winter 2015 Introduction to Machine Learning. Lecture 2: Linear regression

CSC2515 Winter 2015 Introduction to Machine Learning. Lecture 2: Linear regression CSC2515 Winter 2015 Introduction to Machine Learning Lecture 2: Linear regression All lecture slides will be available as.pdf on the course website: http://www.cs.toronto.edu/~urtasun/courses/csc2515/csc2515_winter15.html

More information

Predicting Failure of Multiangle Composite Laminates

Predicting Failure of Multiangle Composite Laminates Predicting Failure of Multiangle Composite Laminates Preliminary discussion (not in textbook): Micromechanics failure analyses vs Macromechanics failure analyses Fiber Architecture of Some Common Composite

More information

Expectation propagation for signal detection in flat-fading channels

Expectation propagation for signal detection in flat-fading channels Expectation propagation for signal detection in flat-fading channels Yuan Qi MIT Media Lab Cambridge, MA, 02139 USA yuanqi@media.mit.edu Thomas Minka CMU Statistics Department Pittsburgh, PA 15213 USA

More information

S.Nagendra, D.Jestin, Z.Gurdal, R.T.Haftka and L.T.Watson Computers & Structures, Vol. 58, No. 3, pp , 1996.

S.Nagendra, D.Jestin, Z.Gurdal, R.T.Haftka and L.T.Watson Computers & Structures, Vol. 58, No. 3, pp , 1996. S.Nagendra, D.Jestin, Z.Gurdal, R.T.Haftka and L.T.Watson Computers & Structures, Vol. 58, No. 3, pp. 543-555, 1996. Presented by Vignesh Solai Rameshbabu Introduction Simple programming technique which

More information

Partially Observable Markov Decision Processes (POMDPs)

Partially Observable Markov Decision Processes (POMDPs) Partially Observable Markov Decision Processes (POMDPs) Sachin Patil Guest Lecture: CS287 Advanced Robotics Slides adapted from Pieter Abbeel, Alex Lee Outline Introduction to POMDPs Locally Optimal Solutions

More information

CS-E4830 Kernel Methods in Machine Learning

CS-E4830 Kernel Methods in Machine Learning CS-E4830 Kernel Methods in Machine Learning Lecture 3: Convex optimization and duality Juho Rousu 27. September, 2017 Juho Rousu 27. September, 2017 1 / 45 Convex optimization Convex optimisation This

More information

Log-Linear Models, MEMMs, and CRFs

Log-Linear Models, MEMMs, and CRFs Log-Linear Models, MEMMs, and CRFs Michael Collins 1 Notation Throughout this note I ll use underline to denote vectors. For example, w R d will be a vector with components w 1, w 2,... w d. We use expx

More information

Recita,on: Loss, Regulariza,on, and Dual*

Recita,on: Loss, Regulariza,on, and Dual* 10-701 Recita,on: Loss, Regulariza,on, and Dual* Jay- Yoon Lee 02/26/2015 *Adopted figures from 10725 lecture slides and from the book Elements of Sta,s,cal Learning Loss and Regulariza,on Op,miza,on problem

More information

Support Vector Machine (SVM) and Kernel Methods

Support Vector Machine (SVM) and Kernel Methods Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2015 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin

More information

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008 Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:

More information

Today. Calculus. Linear Regression. Lagrange Multipliers

Today. Calculus. Linear Regression. Lagrange Multipliers Today Calculus Lagrange Multipliers Linear Regression 1 Optimization with constraints What if I want to constrain the parameters of the model. The mean is less than 10 Find the best likelihood, subject

More information

13: Variational inference II

13: Variational inference II 10-708: Probabilistic Graphical Models, Spring 2015 13: Variational inference II Lecturer: Eric P. Xing Scribes: Ronghuo Zheng, Zhiting Hu, Yuntian Deng 1 Introduction We started to talk about variational

More information

Support Vector Machines

Support Vector Machines Support Vector Machines Le Song Machine Learning I CSE 6740, Fall 2013 Naïve Bayes classifier Still use Bayes decision rule for classification P y x = P x y P y P x But assume p x y = 1 is fully factorized

More information

CSE 254: MAP estimation via agreement on (hyper)trees: Message-passing and linear programming approaches

CSE 254: MAP estimation via agreement on (hyper)trees: Message-passing and linear programming approaches CSE 254: MAP estimation via agreement on (hyper)trees: Message-passing and linear programming approaches A presentation by Evan Ettinger November 11, 2005 Outline Introduction Motivation and Background

More information

Clustering and Gaussian Mixture Models

Clustering and Gaussian Mixture Models Clustering and Gaussian Mixture Models Piyush Rai IIT Kanpur Probabilistic Machine Learning (CS772A) Jan 25, 2016 Probabilistic Machine Learning (CS772A) Clustering and Gaussian Mixture Models 1 Recap

More information

Multiple Random Variables

Multiple Random Variables Multiple Random Variables This Version: July 30, 2015 Multiple Random Variables 2 Now we consider models with more than one r.v. These are called multivariate models For instance: height and weight An

More information

Mathematical Preliminaries

Mathematical Preliminaries Mathematical Preliminaries Economics 3307 - Intermediate Macroeconomics Aaron Hedlund Baylor University Fall 2013 Econ 3307 (Baylor University) Mathematical Preliminaries Fall 2013 1 / 25 Outline I: Sequences

More information

Sensitivity and Reliability Analysis of Nonlinear Frame Structures

Sensitivity and Reliability Analysis of Nonlinear Frame Structures Sensitivity and Reliability Analysis of Nonlinear Frame Structures Michael H. Scott Associate Professor School of Civil and Construction Engineering Applied Mathematics and Computation Seminar April 8,

More information

CPSC 340: Machine Learning and Data Mining. More PCA Fall 2017

CPSC 340: Machine Learning and Data Mining. More PCA Fall 2017 CPSC 340: Machine Learning and Data Mining More PCA Fall 2017 Admin Assignment 4: Due Friday of next week. No class Monday due to holiday. There will be tutorials next week on MAP/PCA (except Monday).

More information

CS839: Probabilistic Graphical Models. Lecture 7: Learning Fully Observed BNs. Theo Rekatsinas

CS839: Probabilistic Graphical Models. Lecture 7: Learning Fully Observed BNs. Theo Rekatsinas CS839: Probabilistic Graphical Models Lecture 7: Learning Fully Observed BNs Theo Rekatsinas 1 Exponential family: a basic building block For a numeric random variable X p(x ) =h(x)exp T T (x) A( ) = 1

More information

Composite Structural Mechanics using MATLAB

Composite Structural Mechanics using MATLAB Session 2520 Composite Structural Mechanics using MATLAB Oscar Barton, Jr., Jacob B. Wallace United States Naval Academy Annapolis, Md 21402 Abstract In this paper MATLAB is adopted as the programming

More information

Kernel Methods. Machine Learning A W VO

Kernel Methods. Machine Learning A W VO Kernel Methods Machine Learning A 708.063 07W VO Outline 1. Dual representation 2. The kernel concept 3. Properties of kernels 4. Examples of kernel machines Kernel PCA Support vector regression (Relevance

More information

Advanced Optimization

Advanced Optimization Advanced Optimization Lecture 3: 1: Randomized Algorithms for for Continuous Discrete Problems Problems November 22, 2016 Master AIC Université Paris-Saclay, Orsay, France Anne Auger INRIA Saclay Ile-de-France

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 12: Gaussian Belief Propagation, State Space Models and Kalman Filters Guest Kalman Filter Lecture by

More information

Written Examination

Written Examination Division of Scientific Computing Department of Information Technology Uppsala University Optimization Written Examination 202-2-20 Time: 4:00-9:00 Allowed Tools: Pocket Calculator, one A4 paper with notes

More information

Probabilistic Model-Building Genetic Algorithms

Probabilistic Model-Building Genetic Algorithms Probabilistic Model-Building Genetic Algorithms a.k.a. Estimation of Distribution Algorithms a.k.a. Iterated Density Estimation Algorithms Martin Pelikan Foreword Motivation Genetic and evolutionary computation

More information

Data Mining and Analysis: Fundamental Concepts and Algorithms

Data Mining and Analysis: Fundamental Concepts and Algorithms Data Mining and Analysis: Fundamental Concepts and Algorithms dataminingbook.info Mohammed J. Zaki 1 Wagner Meira Jr. 2 1 Department of Computer Science Rensselaer Polytechnic Institute, Troy, NY, USA

More information

Linear Classification. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington

Linear Classification. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington Linear Classification CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 Example of Linear Classification Red points: patterns belonging

More information

ECE521 lecture 4: 19 January Optimization, MLE, regularization

ECE521 lecture 4: 19 January Optimization, MLE, regularization ECE521 lecture 4: 19 January 2017 Optimization, MLE, regularization First four lectures Lectures 1 and 2: Intro to ML Probability review Types of loss functions and algorithms Lecture 3: KNN Convexity

More information

Support Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012

Support Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012 Support Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Linear classifier Which classifier? x 2 x 1 2 Linear classifier Margin concept x 2

More information

Ch 4. Linear Models for Classification

Ch 4. Linear Models for Classification Ch 4. Linear Models for Classification Pattern Recognition and Machine Learning, C. M. Bishop, 2006. Department of Computer Science and Engineering Pohang University of Science and echnology 77 Cheongam-ro,

More information

PILCO: A Model-Based and Data-Efficient Approach to Policy Search

PILCO: A Model-Based and Data-Efficient Approach to Policy Search PILCO: A Model-Based and Data-Efficient Approach to Policy Search (M.P. Deisenroth and C.E. Rasmussen) CSC2541 November 4, 2016 PILCO Graphical Model PILCO Probabilistic Inference for Learning COntrol

More information

Gaussian processes and bayesian optimization Stanisław Jastrzębski. kudkudak.github.io kudkudak

Gaussian processes and bayesian optimization Stanisław Jastrzębski. kudkudak.github.io kudkudak Gaussian processes and bayesian optimization Stanisław Jastrzębski kudkudak.github.io kudkudak Plan Goal: talk about modern hyperparameter optimization algorithms Bayes reminder: equivalent linear regression

More information

Introduction Linear system Nonlinear equation Interpolation

Introduction Linear system Nonlinear equation Interpolation Interpolation Interpolation is the process of estimating an intermediate value from a set of discrete or tabulated values. Suppose we have the following tabulated values: y y 0 y 1 y 2?? y 3 y 4 y 5 x

More information

Multivariate Distributions CIVL 7012/8012

Multivariate Distributions CIVL 7012/8012 Multivariate Distributions CIVL 7012/8012 Multivariate Distributions Engineers often are interested in more than one measurement from a single item. Multivariate distributions describe the probability

More information

Discrete Mathematics and Probability Theory Fall 2015 Lecture 21

Discrete Mathematics and Probability Theory Fall 2015 Lecture 21 CS 70 Discrete Mathematics and Probability Theory Fall 205 Lecture 2 Inference In this note we revisit the problem of inference: Given some data or observations from the world, what can we infer about

More information

Joint Gaussian Graphical Model Review Series I

Joint Gaussian Graphical Model Review Series I Joint Gaussian Graphical Model Review Series I Probability Foundations Beilun Wang Advisor: Yanjun Qi 1 Department of Computer Science, University of Virginia http://jointggm.org/ June 23rd, 2017 Beilun

More information

CSE446: Clustering and EM Spring 2017

CSE446: Clustering and EM Spring 2017 CSE446: Clustering and EM Spring 2017 Ali Farhadi Slides adapted from Carlos Guestrin, Dan Klein, and Luke Zettlemoyer Clustering systems: Unsupervised learning Clustering Detect patterns in unlabeled

More information

{ p if x = 1 1 p if x = 0

{ p if x = 1 1 p if x = 0 Discrete random variables Probability mass function Given a discrete random variable X taking values in X = {v 1,..., v m }, its probability mass function P : X [0, 1] is defined as: P (v i ) = Pr[X =

More information

Learning Bayesian networks

Learning Bayesian networks 1 Lecture topics: Learning Bayesian networks from data maximum likelihood, BIC Bayesian, marginal likelihood Learning Bayesian networks There are two problems we have to solve in order to estimate Bayesian

More information

1 EM algorithm: updating the mixing proportions {π k } ik are the posterior probabilities at the qth iteration of EM.

1 EM algorithm: updating the mixing proportions {π k } ik are the posterior probabilities at the qth iteration of EM. Université du Sud Toulon - Var Master Informatique Probabilistic Learning and Data Analysis TD: Model-based clustering by Faicel CHAMROUKHI Solution The aim of this practical wor is to show how the Classification

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

Variational Inference (11/04/13)

Variational Inference (11/04/13) STA561: Probabilistic machine learning Variational Inference (11/04/13) Lecturer: Barbara Engelhardt Scribes: Matt Dickenson, Alireza Samany, Tracy Schifeling 1 Introduction In this lecture we will further

More information

Computational statistics

Computational statistics Computational statistics Combinatorial optimization Thierry Denœux February 2017 Thierry Denœux Computational statistics February 2017 1 / 37 Combinatorial optimization Assume we seek the maximum of f

More information

A Progressive Hedging Approach to Multistage Stochastic Generation and Transmission Investment Planning

A Progressive Hedging Approach to Multistage Stochastic Generation and Transmission Investment Planning A Progressive Hedging Approach to Multistage Stochastic Generation and Transmission Investment Planning Yixian Liu Ramteen Sioshansi Integrated Systems Engineering Department The Ohio State University

More information

Chapter 5 continued. Chapter 5 sections

Chapter 5 continued. Chapter 5 sections Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions

More information

An Information Geometry Perspective on Estimation of Distribution Algorithms: Boundary Analysis

An Information Geometry Perspective on Estimation of Distribution Algorithms: Boundary Analysis An Information Geometry Perspective on Estimation of Distribution Algorithms: Boundary Analysis Luigi Malagò Department of Electronics and Information Politecnico di Milano Via Ponzio, 34/5 20133 Milan,

More information

Generalized Method of Moments Estimation

Generalized Method of Moments Estimation Generalized Method of Moments Estimation Lars Peter Hansen March 0, 2007 Introduction Generalized methods of moments (GMM) refers to a class of estimators which are constructed from exploiting the sample

More information

Let X and Y be two real valued stochastic variables defined on (Ω, F, P). Theorem: If X and Y are independent then. . p.1/21

Let X and Y be two real valued stochastic variables defined on (Ω, F, P). Theorem: If X and Y are independent then. . p.1/21 Multivariate transformations The remaining part of the probability course is centered around transformations t : R k R m and how they transform probability measures. For instance tx 1,...,x k = x 1 +...

More information

Multi-Objective Optimization of Carbon Fibre Reinforced Plastic (CFRP) Circular Hollow Section Using Genetic Algorithm for Engineering Structures

Multi-Objective Optimization of Carbon Fibre Reinforced Plastic (CFRP) Circular Hollow Section Using Genetic Algorithm for Engineering Structures International Refereed Journal of Engineering and Science (IRJES) ISSN (Online) 2319-183X, (rint) 2319-1821 Volume 4, Issue 12 (December 2015),.24-28 Multi-Objective Optimization of Carbon Fibre Reinforced

More information

NONLINEAR. (Hillier & Lieberman Introduction to Operations Research, 8 th edition)

NONLINEAR. (Hillier & Lieberman Introduction to Operations Research, 8 th edition) NONLINEAR PROGRAMMING (Hillier & Lieberman Introduction to Operations Research, 8 th edition) Nonlinear Programming g Linear programming has a fundamental role in OR. In linear programming all its functions

More information

3 : Representation of Undirected GM

3 : Representation of Undirected GM 10-708: Probabilistic Graphical Models 10-708, Spring 2016 3 : Representation of Undirected GM Lecturer: Eric P. Xing Scribes: Longqi Cai, Man-Chia Chang 1 MRF vs BN There are two types of graphical models:

More information

Support Vector Machine (continued)

Support Vector Machine (continued) Support Vector Machine continued) Overlapping class distribution: In practice the class-conditional distributions may overlap, so that the training data points are no longer linearly separable. We need

More information

Principal Components Analysis. Sargur Srihari University at Buffalo

Principal Components Analysis. Sargur Srihari University at Buffalo Principal Components Analysis Sargur Srihari University at Buffalo 1 Topics Projection Pursuit Methods Principal Components Examples of using PCA Graphical use of PCA Multidimensional Scaling Srihari 2

More information

Maximum Likelihood Estimation. only training data is available to design a classifier

Maximum Likelihood Estimation. only training data is available to design a classifier Introduction to Pattern Recognition [ Part 5 ] Mahdi Vasighi Introduction Bayesian Decision Theory shows that we could design an optimal classifier if we knew: P( i ) : priors p(x i ) : class-conditional

More information

Neutron inverse kinetics via Gaussian Processes

Neutron inverse kinetics via Gaussian Processes Neutron inverse kinetics via Gaussian Processes P. Picca Politecnico di Torino, Torino, Italy R. Furfaro University of Arizona, Tucson, Arizona Outline Introduction Review of inverse kinetics techniques

More information

Chapter 6 - Random Processes

Chapter 6 - Random Processes EE385 Class Notes //04 John Stensby Chapter 6 - Random Processes Recall that a random variable X is a mapping between the sample space S and the extended real line R +. That is, X : S R +. A random process

More information