Model reduction of Multiscale Models using Kernel Methods

Size: px
Start display at page:

Download "Model reduction of Multiscale Models using Kernel Methods"

Transcription

1 Model reduction of Multiscale Models using Kernel Methods at the WORM 213 Haus der Kirche, Bad Herrenalb Daniel Wirtz August 23, 213 Dipl. Math. Dipl. Inf. Daniel Wirtz, IANS, University of Stuttgart Jun.-Prof. Dr. Bernard Haasdonk, IANS, University of Stuttgart 1/52

2 Introduction Contents Introduction Kernel Expansion construction Support Vector Machines Vectorial Greedy Algorithms Experimental results Human spine simulation (N. Karajan) Shock waves in porous media (F. Kissling) 2/52

3 Introduction Up Next Introduction 3/52

4 Introduction Reduction setting C Macro- and D microscale model. Interaction parameters from domains P C, P D One macrolevel step requires many microlevel solves Each simulation of D is actually a function evaluation x P D y start C D y P C Figure: Communication of macro and microscale models f : P D P C x P D (x). Goal: Find a surrogate D, i.e. f f Figure: Example multibody system C with many microscale models D i 4/52

5 Introduction Buzzwords! Response Surface Methodology [Myers et al.(29)myers, Montgomery & Anderson-Cook, Khuri & Mukhopadhyay(21)] Surrogate Modeling Figure: Illustrations of response surfaces 5/52

6 Introduction Building surfaces How to build a surrogate or response surface? Classically via polynomials Interpolation, Regression Multivariate variants Polynomial Chaos Expansions [Wiener(1938), Karajan et al.(213)karajan, Ehlers, Oladyshkin & Otto.] Alternative: Kernels! Symmetric function K : R d R d R Positive definite (analysis!) E.g. Gaussian (for γ > ) K g (x, y) = exp( x y 2 /γ 2 ) 6/52

7 Introduction More famous kernel examples Compactly supported Wendland kernels Kw d,k (x, y) = φ d,k w ( ) x y for d, k N and polynomials γ, φ d,k w (r) = (1 r) l+k + p d,k (r). p d,1 (r) = (l + 1)r + 1, l = d/2 + k + 1 p d,2 (r) = 1 3 (l2 + 4l + 3)r 2 + (l + 2)r + 1, [...] γ is a dilation hyperparameter for the kernel width. d, k determines the smoothness in the sense of K d,k w C 2k (P D ) if P D R s, s d. 7/52

8 Introduction Plots for different kernels Figure: Examples. 1D: K g(γ =.33), Kw 2,, K1,1 w.2d: Kg(γ =.49), K2, w, K3,3 w. γ = 1 for all Kw. 8/52

9 Introduction From kernels to kernel expansions f = Kernel expansion! f(x) = N c i K(x i, x) i=1 with centers x i R d and coefficients c i R. Figure: Illustration for a 1D Gaussian kernel expansion with γ = 1 and x i = [ 1,.5, 1.2], c i = [.5,.2,.8] 9/52

10 Expansion construction Up Next Expansion construction 1/52

11 Expansion construction How to obtain kernel expansions? Various methods available Mostly using only training data {x 1,..., x N },{y 1,..., y N } Support Vector Machines / Regression [Schölkopf & Smola(22)] 1 f = arg min g H N N i=1 V ( y i g(x i) ) + λ g 2 H, V (e) = max{, e ɛ} f = /52

12 Expansion construction ɛ-svr Support Vector Regression with ɛ-loss (Dual problem) For C := 1 2λN and y = (y 1,..., y N ) T R N we solve max α +,α R N W (α +, α ) (1) If α +, α s.t. α +, i C, i = 1... N, = α + i α i, i = 1... N, W (α +, α ) := 1 ( α + α ) T ( K α + α ) 2 N ɛ (α + i + α i ) + y T ( α + α ). i=1 solve (1), then N f(x) := c ik(x, x i), i=1 c i := (α α + ) i Computation via Sequential Minimal Optimization (SMO) 12/52

13 Expansion construction ɛ-svr illustration: 1D SMO.5 Iteration 1 f(x) and approximation x 13/52 Figure: ɛ-svr 1D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers

14 Expansion construction ɛ-svr illustration: 1D SMO.5 Iteration 2 f(x) and approximation x 14/52 Figure: ɛ-svr 1D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers

15 Expansion construction ɛ-svr illustration: 1D SMO.5 Iteration 3 f(x) and approximation x 15/52 Figure: ɛ-svr 1D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers

16 Expansion construction ɛ-svr illustration: 1D SMO.5 Iteration 4 f(x) and approximation x 16/52 Figure: ɛ-svr 1D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers

17 Expansion construction ɛ-svr illustration: 1D SMO.5 Iteration 5 f(x) and approximation x 17/52 Figure: ɛ-svr 1D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers

18 Expansion construction ɛ-svr illustration: 1D SMO.5 Iteration 6 f(x) and approximation x 18/52 Figure: ɛ-svr 1D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers

19 Expansion construction ɛ-svr illustration: 1D SMO.5 Iteration 7 f(x) and approximation x Figure: ɛ-svr 1D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers 19/52

20 Expansion construction ɛ-svr illustration: 1D SMO.5 Iteration 8 f(x) and approximation x Figure: ɛ-svr 1D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers 2/52

21 Expansion construction ɛ-svr illustration: 2D SMO.5 Iteration 1 f(x) and approximation x 21/52 Figure: ɛ-svr 2D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers

22 Expansion construction ɛ-svr illustration: 2D SMO.5 Iteration 2 f(x) and approximation x 22/52 Figure: ɛ-svr 2D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers

23 Expansion construction ɛ-svr illustration: 2D SMO.5 Iteration 3 f(x) and approximation x 23/52 Figure: ɛ-svr 2D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers

24 Expansion construction ɛ-svr illustration: 2D SMO.5 Iteration 4 f(x) and approximation x 24/52 Figure: ɛ-svr 2D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers

25 Expansion construction ɛ-svr illustration: 2D SMO.5 Iteration 5 f(x) and approximation x 25/52 Figure: ɛ-svr 2D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers

26 Expansion construction ɛ-svr illustration: 2D SMO.5 Iteration 6 f(x) and approximation x 26/52 Figure: ɛ-svr 2D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers

27 Expansion construction ɛ-svr illustration: 2D SMO.5 Iteration 7 f(x) and approximation x Figure: ɛ-svr 2D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers 27/52

28 Expansion construction ɛ-svr illustration: 2D SMO.5 Iteration 8 f(x) and approximation x Figure: ɛ-svr 2D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers 28/52

29 Expansion construction ɛ-svr illustration: 2D SMO full solution.5 It: 3523, #SV=2, eps=.1 f(x) x Figure: ɛ-svr 2D SMO solution. Red: Target function with margin ɛ =.1, blue: approximation, black: centers 29/52

30 Expansion construction The VKOGA Vectorial kernel orthogonal greedy algorithm Let m N and f H m, define X :=, f := and for k >, Ω k := { x P D K(x, ) H X k 1}, the sequences x k := arg max x P D \Ω k X k := X k 1 {x k }, f k := I m X k [f]. f(x) f k 1 (x) 2 2 K(x, ) IXk 1 [K(x, )], H Here, I m X k [f] denotes component-wise interpolation on X k. In/Exclusion of the denominator yields different variants. Details in [Wirtz & Haasdonk(213)] 3/52

31 Expansion construction VKOGA illustration 1.5 Iteration f(x) x 31/52 Figure: VKOGA iterations one to four. Green/Blue: Components, Dashed: Approximations, Red: Total error, Diamonds: Centers, Red circle: Max overall error, Black lines: Max component errors

32 Expansion construction VKOGA illustration 1.5 Iteration f(x) x 32/52 Figure: VKOGA iterations one to four. Green/Blue: Components, Dashed: Approximations, Red: Total error, Diamonds: Centers, Red circle: Max overall error, Black lines: Max component errors

33 Expansion construction VKOGA illustration 1.5 Iteration f(x) x 33/52 Figure: VKOGA iterations one to four. Green/Blue: Components, Dashed: Approximations, Red: Total error, Diamonds: Centers, Red circle: Max overall error, Black lines: Max component errors

34 Expansion construction VKOGA illustration 1.5 Iteration f(x) x 34/52 Figure: VKOGA iterations one to four. Green/Blue: Components, Dashed: Approximations, Red: Total error, Diamonds: Centers, Red circle: Max overall error, Black lines: Max component errors

35 Experimental results Up Next Experimental results 35/52

36 Experimental results Multibody spine simulation: Introduction Cooperation with N. Karajan, D. Otto Linking MBS spine simulation with FE intervertebral model Figure: Illustration from [Karajan et al.(211)karajan, Röhrle, Ehlers & Schmitt] 36/52

37 Experimental results Multibody spine simulation: Animations for bone and disk interactions Figure: Animations for MBS vertebra interaction and ϕ 1, u 2, u 3 disk response cases 37/52

38 Experimental results Multibody spine simulation: Combined load cases for IVDs Figure: Different load case animations for general and geometry-fitted intervertebral disks 38/52

39 Experimental results Multibody spine simulation: Combined load cases for IVDs Computation with quasi-static timesteps from equillibrium state Figure: Joystick like computation of training data 39/52

40 Experimental results Multibody spine simulation: Approximation Plot of M1 against u2 and u3 FEM inputs ϕ 1, u 2, u 3 P D R 3 Disc response M 1, R 2, R 3 P C R 3 Precomputation of 274/1433 disc responses About 3min average simulation time each dot Future work: Actual input is ϕ 1, u 2, u 3, M 1, R 2, R 3 from last (time)step M u3 Plot of M1 against u2 and phi1 x phi1 u2 M1 2 2 u2 3 2 Figure: Plots of training data 4/52

41 Experimental results Multibody spine simulation: Results with Gauss kernels 41/52 Figure: Plots of approximation, u2, u3 M1, ϕ1 = [ 5.7,, 5.3] (top) and u2, ϕ1 M1, u3 = [ 3.1,, 3.1] (bottom)

42 Experimental results Multibody spine simulation: Results with Wendland kernels Figure: Plots of approximation using Wendland kernels and best configuration γ : , k : 1, d : 2 42/52

43 Experimental results Multibody spine simulation: Outlook Simulation using kernel approximations! So far: Using Polynomial Chaos Expansions Figure: Simulation of full multibody spine model using suggorates 43/52

44 Experimental results Two-phase flow in porous media: Intro Cooperation with F. Kissling (IANS) Two-phase flow in porous media with nonclassical shock waves Reg. Buckley-Leverett problem S t + (vf(s)) = ɛs xx + ɛ 2 τs xxt On front: Solve Riemann problems for overshoot Figure: Illustrations from [Kissling & Rhode(212)] 44/52

45 Experimental results Two-phase flow in porous media: Approximation Riemann problem initial data Left-side saturation s l Scaling parameter τ Result: Plateau value S ɛ Precomputation of 5 Rieman problems Suitable transformation of training data Scaling Logarithmic values Figure: Plots of training data 45/52

46 Experimental results Two-phase flow in porous media: Results 46/52 Figure: Plots of approximations. Top: SVR (Gauss), Bottom: Greedy algorithm (Wendland)

47 Experimental results Two-phase flow in porous media: Results Time-reduction from 2h to 24min for full simulation N S S SV R L Figure: Full simulation plots and error table. Full model (top) and SVR reduced (bottom) Details in [Kissling & Rhode(212)] 47/52

48 Summary What did we see? Setting: Reduction for Microscale models Use of cheap surrogates for micro-scale level Kernel methods can provide such replacements! Support vector regression (SVR) Vectorial greedy algorithms (VKOGA) Examples and results available: Human Spine Simulations Shock-waves in porous media 48/52

49 -THE END-... any questions? Thank you for your attention! 49/52

50 References I [Karajan et al.(213)karajan, Ehlers, Oladyshkin & Otto.] Karajan, N.; Ehlers, W.; Oladyshkin, S. & Otto., D.: Application of the Polynomial Chaos Expansion to Approximate the Homogenised Behaviour of the Intervertebral Disc. Preprint, University of Stuttgart, SRC SimTech (213), in preparation. [Karajan et al.(211)karajan, Röhrle, Ehlers & Schmitt] Karajan, N.; Röhrle, O.; Ehlers, W. & Schmitt, S.: Linking Continuous and Discrete Intervertebral Disc Models through Homogenisation. Preprint, SRC SimTech (211), URL uni-stuttgart.de/publikationen/prints.php?id=657, submitted to Biomechanics and Modeling in Mechanobiology. [Khuri & Mukhopadhyay(21)] Khuri, A. I. & Mukhopadhyay, S.: Response surface methodology. Wiley Interdisciplinary Reviews: Computational Statistics 2 (21), , ISSN /52

51 References II [Kissling & Rhode(212)] Kissling, F. & Rhode, C.: The Computation of Nonclassical Shock Waves in Porous Media with a Heterogeneous Multiscale Method: The Multidimensional Case. Preprint, Stuttgart Research Centre for Simulation Technology (212), URL publikationen/prints.php?id=734, submitted to SIAM Multiscale Modeling and Simulation. [Myers et al.(29)myers, Montgomery & Anderson-Cook] Myers, R. H.; Montgomery, D. C. & Anderson-Cook, C. M.: Response Surface Methodology - Process and Product Optimization Using Designed Experiments. Wiley Series in Probability and Statistics, Wiley 29, 3 edn. [Schölkopf & Smola(22)] Schölkopf, B. & Smola, A. J.: Learning with Kernels. Adaptive Computation and Machine Learning, The MIT Press /52

52 References III [Wiener(1938)] Wiener, N.: The homogeneous chaos. Amer. J. Math. 6 (1938), , ISSN 29327, URL [Wirtz & Haasdonk(213)] Wirtz, D. & Haasdonk, B.: An improved vectorial kernel orthogonal greedy algorithm. Dolomites Research Notes on Approximation 6 (213), /52

CIS 520: Machine Learning Oct 09, Kernel Methods

CIS 520: Machine Learning Oct 09, Kernel Methods CIS 520: Machine Learning Oct 09, 207 Kernel Methods Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture They may or may not cover all the material discussed

More information

A-Posteriori Error Estimation for Parameterized Kernel-Based Systems

A-Posteriori Error Estimation for Parameterized Kernel-Based Systems A-Posteriori Error Estimation for Parameterized Kernel-Based Systems Daniel Wirtz Bernard Haasdonk Institute for Applied Analysis and Numerical Simulation University of Stuttgart 70569 Stuttgart, Pfaffenwaldring

More information

Support Vector Machine for Classification and Regression

Support Vector Machine for Classification and Regression Support Vector Machine for Classification and Regression Ahlame Douzal AMA-LIG, Université Joseph Fourier Master 2R - MOSIG (2013) November 25, 2013 Loss function, Separating Hyperplanes, Canonical Hyperplan

More information

Stochastic representation of random positive-definite tensor-valued properties: application to 3D anisotropic permeability random fields

Stochastic representation of random positive-definite tensor-valued properties: application to 3D anisotropic permeability random fields Sous la co-tutelle de : LABORATOIRE DE MODÉLISATION ET SIMULATION MULTI ÉCHELLE CNRS UPEC UNIVERSITÉ PARIS-EST CRÉTEIL UPEM UNIVERSITÉ PARIS-EST MARNE-LA-VALLÉE Stochastic representation of random positive-definite

More information

Kernel Methods. Outline

Kernel Methods. Outline Kernel Methods Quang Nguyen University of Pittsburgh CS 3750, Fall 2011 Outline Motivation Examples Kernels Definitions Kernel trick Basic properties Mercer condition Constructing feature space Hilbert

More information

SMO Algorithms for Support Vector Machines without Bias Term

SMO Algorithms for Support Vector Machines without Bias Term Institute of Automatic Control Laboratory for Control Systems and Process Automation Prof. Dr.-Ing. Dr. h. c. Rolf Isermann SMO Algorithms for Support Vector Machines without Bias Term Michael Vogt, 18-Jul-2002

More information

A GENERAL FORMULATION FOR SUPPORT VECTOR MACHINES. Wei Chu, S. Sathiya Keerthi, Chong Jin Ong

A GENERAL FORMULATION FOR SUPPORT VECTOR MACHINES. Wei Chu, S. Sathiya Keerthi, Chong Jin Ong A GENERAL FORMULATION FOR SUPPORT VECTOR MACHINES Wei Chu, S. Sathiya Keerthi, Chong Jin Ong Control Division, Department of Mechanical Engineering, National University of Singapore 0 Kent Ridge Crescent,

More information

An Improved Conjugate Gradient Scheme to the Solution of Least Squares SVM

An Improved Conjugate Gradient Scheme to the Solution of Least Squares SVM An Improved Conjugate Gradient Scheme to the Solution of Least Squares SVM Wei Chu Chong Jin Ong chuwei@gatsby.ucl.ac.uk mpeongcj@nus.edu.sg S. Sathiya Keerthi mpessk@nus.edu.sg Control Division, Department

More information

Computer Vision Group Prof. Daniel Cremers. 4. Gaussian Processes - Regression

Computer Vision Group Prof. Daniel Cremers. 4. Gaussian Processes - Regression Group Prof. Daniel Cremers 4. Gaussian Processes - Regression Definition (Rep.) Definition: A Gaussian process is a collection of random variables, any finite number of which have a joint Gaussian distribution.

More information

ν =.1 a max. of 10% of training set can be margin errors ν =.8 a max. of 80% of training can be margin errors

ν =.1 a max. of 10% of training set can be margin errors ν =.8 a max. of 80% of training can be margin errors p.1/1 ν-svms In the traditional softmargin classification SVM formulation we have a penalty constant C such that 1 C size of margin. Furthermore, there is no a priori guidance as to what C should be set

More information

Advanced Introduction to Machine Learning

Advanced Introduction to Machine Learning 10-715 Advanced Introduction to Machine Learning Homework Due Oct 15, 10.30 am Rules Please follow these guidelines. Failure to do so, will result in loss of credit. 1. Homework is due on the due date

More information

Statistical Machine Learning from Data

Statistical Machine Learning from Data Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Support Vector Machines Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole Polytechnique

More information

Approximation Theoretical Questions for SVMs

Approximation Theoretical Questions for SVMs Ingo Steinwart LA-UR 07-7056 October 20, 2007 Statistical Learning Theory: an Overview Support Vector Machines Informal Description of the Learning Goal X space of input samples Y space of labels, usually

More information

Review: Support vector machines. Machine learning techniques and image analysis

Review: Support vector machines. Machine learning techniques and image analysis Review: Support vector machines Review: Support vector machines Margin optimization min (w,w 0 ) 1 2 w 2 subject to y i (w 0 + w T x i ) 1 0, i = 1,..., n. Review: Support vector machines Margin optimization

More information

Computer Vision Group Prof. Daniel Cremers. 9. Gaussian Processes - Regression

Computer Vision Group Prof. Daniel Cremers. 9. Gaussian Processes - Regression Group Prof. Daniel Cremers 9. Gaussian Processes - Regression Repetition: Regularized Regression Before, we solved for w using the pseudoinverse. But: we can kernelize this problem as well! First step:

More information

Recent Results for Moving Least Squares Approximation

Recent Results for Moving Least Squares Approximation Recent Results for Moving Least Squares Approximation Gregory E. Fasshauer and Jack G. Zhang Abstract. We describe two experiments recently conducted with the approximate moving least squares (MLS) approximation

More information

A model order reduction technique for speeding up computational homogenisation

A model order reduction technique for speeding up computational homogenisation A model order reduction technique for speeding up computational homogenisation Olivier Goury, Pierre Kerfriden, Wing Kam Liu, Stéphane Bordas Cardiff University Outline Introduction Heterogeneous materials

More information

Support Vector Machines II. CAP 5610: Machine Learning Instructor: Guo-Jun QI

Support Vector Machines II. CAP 5610: Machine Learning Instructor: Guo-Jun QI Support Vector Machines II CAP 5610: Machine Learning Instructor: Guo-Jun QI 1 Outline Linear SVM hard margin Linear SVM soft margin Non-linear SVM Application Linear Support Vector Machine An optimization

More information

Lecture 10: A brief introduction to Support Vector Machine

Lecture 10: A brief introduction to Support Vector Machine Lecture 10: A brief introduction to Support Vector Machine Advanced Applied Multivariate Analysis STAT 2221, Fall 2013 Sungkyu Jung Department of Statistics, University of Pittsburgh Xingye Qiao Department

More information

Introduction to SVM and RVM

Introduction to SVM and RVM Introduction to SVM and RVM Machine Learning Seminar HUS HVL UIB Yushu Li, UIB Overview Support vector machine SVM First introduced by Vapnik, et al. 1992 Several literature and wide applications Relevance

More information

Solving the SVM Optimization Problem

Solving the SVM Optimization Problem Solving the SVM Optimization Problem Kernel-based Learning Methods Christian Igel Institut für Neuroinformatik Ruhr-Universität Bochum, Germany http://www.neuroinformatik.rub.de July 16, 2009 Christian

More information

Employing Model Reduction for Uncertainty Visualization in the Context of CO 2 Storage Simulation

Employing Model Reduction for Uncertainty Visualization in the Context of CO 2 Storage Simulation Employing Model Reduction for Uncertainty Visualization in the Context of CO 2 Storage Simulation Marcel Hlawatsch, Sergey Oladyshkin, Daniel Weiskopf University of Stuttgart Problem setting - underground

More information

Support Vector Machines. Maximizing the Margin

Support Vector Machines. Maximizing the Margin Support Vector Machines Support vector achines (SVMs) learn a hypothesis: h(x) = b + Σ i= y i α i k(x, x i ) (x, y ),..., (x, y ) are the training exs., y i {, } b is the bias weight. α,..., α are the

More information

MLCC 2018 Variable Selection and Sparsity. Lorenzo Rosasco UNIGE-MIT-IIT

MLCC 2018 Variable Selection and Sparsity. Lorenzo Rosasco UNIGE-MIT-IIT MLCC 2018 Variable Selection and Sparsity Lorenzo Rosasco UNIGE-MIT-IIT Outline Variable Selection Subset Selection Greedy Methods: (Orthogonal) Matching Pursuit Convex Relaxation: LASSO & Elastic Net

More information

Non-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines

Non-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines Non-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Fall 2018 CS 551, Fall

More information

Smooth Bayesian Kernel Machines

Smooth Bayesian Kernel Machines Smooth Bayesian Kernel Machines Rutger W. ter Borg 1 and Léon J.M. Rothkrantz 2 1 Nuon NV, Applied Research & Technology Spaklerweg 20, 1096 BA Amsterdam, the Netherlands rutger@terborg.net 2 Delft University

More information

COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017

COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017 COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University FEATURE EXPANSIONS FEATURE EXPANSIONS

More information

Introduction to Support Vector Machines

Introduction to Support Vector Machines Introduction to Support Vector Machines Andreas Maletti Technische Universität Dresden Fakultät Informatik June 15, 2006 1 The Problem 2 The Basics 3 The Proposed Solution Learning by Machines Learning

More information

Support Vector Machines Explained

Support Vector Machines Explained December 23, 2008 Support Vector Machines Explained Tristan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introduction This document has been written in an attempt to make the Support Vector Machines (SVM),

More information

Introduction to Support Vector Machines

Introduction to Support Vector Machines Introduction to Support Vector Machines Shivani Agarwal Support Vector Machines (SVMs) Algorithm for learning linear classifiers Motivated by idea of maximizing margin Efficient extension to non-linear

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 Outlines Overview Introduction Linear Algebra Probability Linear Regression

More information

Machine Learning and Data Mining. Support Vector Machines. Kalev Kask

Machine Learning and Data Mining. Support Vector Machines. Kalev Kask Machine Learning and Data Mining Support Vector Machines Kalev Kask Linear classifiers Which decision boundary is better? Both have zero training error (perfect training accuracy) But, one of them seems

More information

Least Squares SVM Regression

Least Squares SVM Regression Least Squares SVM Regression Consider changing SVM to LS SVM by making following modifications: min (w,e) ½ w 2 + ½C Σ e(i) 2 subject to d(i) (w T Φ( x(i))+ b) = e(i), i, and C>0. Note that e(i) is error

More information

Support Vector Machines: Maximum Margin Classifiers

Support Vector Machines: Maximum Margin Classifiers Support Vector Machines: Maximum Margin Classifiers Machine Learning and Pattern Recognition: September 16, 2008 Piotr Mirowski Based on slides by Sumit Chopra and Fu-Jie Huang 1 Outline What is behind

More information

Hilbert Space Methods in Learning

Hilbert Space Methods in Learning Hilbert Space Methods in Learning guest lecturer: Risi Kondor 6772 Advanced Machine Learning and Perception (Jebara), Columbia University, October 15, 2003. 1 1. A general formulation of the learning problem

More information

Computational tractability of machine learning algorithms for tall fat data

Computational tractability of machine learning algorithms for tall fat data Computational tractability of machine learning algorithms for tall fat data Getting good enough solutions as fast as possible Vikas Chandrakant Raykar vikas@cs.umd.edu University of Maryland, CollegePark

More information

Modeling Dependence of Daily Stock Prices and Making Predictions of Future Movements

Modeling Dependence of Daily Stock Prices and Making Predictions of Future Movements Modeling Dependence of Daily Stock Prices and Making Predictions of Future Movements Taavi Tamkivi, prof Tõnu Kollo Institute of Mathematical Statistics University of Tartu 29. June 2007 Taavi Tamkivi,

More information

BANA 7046 Data Mining I Lecture 6. Other Data Mining Algorithms 1

BANA 7046 Data Mining I Lecture 6. Other Data Mining Algorithms 1 BANA 7046 Data Mining I Lecture 6. Other Data Mining Algorithms 1 Shaobo Li University of Cincinnati 1 Partially based on Hastie, et al. (2009) ESL, and James, et al. (2013) ISLR Data Mining I Lecture

More information

Reduced Basis Method for Parametric

Reduced Basis Method for Parametric 1/24 Reduced Basis Method for Parametric H 2 -Optimal Control Problems NUMOC2017 Andreas Schmidt, Bernard Haasdonk Institute for Applied Analysis and Numerical Simulation, University of Stuttgart andreas.schmidt@mathematik.uni-stuttgart.de

More information

State Space Representation of Gaussian Processes

State Space Representation of Gaussian Processes State Space Representation of Gaussian Processes Simo Särkkä Department of Biomedical Engineering and Computational Science (BECS) Aalto University, Espoo, Finland June 12th, 2013 Simo Särkkä (Aalto University)

More information

Lecture 18: Kernels Risk and Loss Support Vector Regression. Aykut Erdem December 2016 Hacettepe University

Lecture 18: Kernels Risk and Loss Support Vector Regression. Aykut Erdem December 2016 Hacettepe University Lecture 18: Kernels Risk and Loss Support Vector Regression Aykut Erdem December 2016 Hacettepe University Administrative We will have a make-up lecture on next Saturday December 24, 2016 Presentations

More information

SUPPORT VECTOR MACHINE FOR THE SIMULTANEOUS APPROXIMATION OF A FUNCTION AND ITS DERIVATIVE

SUPPORT VECTOR MACHINE FOR THE SIMULTANEOUS APPROXIMATION OF A FUNCTION AND ITS DERIVATIVE SUPPORT VECTOR MACHINE FOR THE SIMULTANEOUS APPROXIMATION OF A FUNCTION AND ITS DERIVATIVE M. Lázaro 1, I. Santamaría 2, F. Pérez-Cruz 1, A. Artés-Rodríguez 1 1 Departamento de Teoría de la Señal y Comunicaciones

More information

LINEAR CLASSIFICATION, PERCEPTRON, LOGISTIC REGRESSION, SVC, NAÏVE BAYES. Supervised Learning

LINEAR CLASSIFICATION, PERCEPTRON, LOGISTIC REGRESSION, SVC, NAÏVE BAYES. Supervised Learning LINEAR CLASSIFICATION, PERCEPTRON, LOGISTIC REGRESSION, SVC, NAÏVE BAYES Supervised Learning Linear vs non linear classifiers In K-NN we saw an example of a non-linear classifier: the decision boundary

More information

Kernel Methods. Konstantin Tretyakov MTAT Machine Learning

Kernel Methods. Konstantin Tretyakov MTAT Machine Learning Kernel Methods Konstantin Tretyakov (kt@ut.ee) MTAT.03.227 Machine Learning So far Supervised machine learning Linear models Non-linear models Unsupervised machine learning Generic scaffolding So far Supervised

More information

Kernel Methods. Konstantin Tretyakov MTAT Machine Learning

Kernel Methods. Konstantin Tretyakov MTAT Machine Learning Kernel Methods Konstantin Tretyakov (kt@ut.ee) MTAT.03.227 Machine Learning So far Supervised machine learning Linear models Least squares regression, SVR Fisher s discriminant, Perceptron, Logistic model,

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Kernel Methods Varun Chandola Computer Science & Engineering State University of New York at Buffalo Buffalo, NY, USA chandola@buffalo.edu Chandola@UB CSE 474/574 1 / 21

More information

On-line Support Vector Machines and Optimization Strategies

On-line Support Vector Machines and Optimization Strategies On-line Support Vector Machines and Optimization Strategies J.M. Górriz, C.G. Puntonet, M. Salmerón, R. Martin-Clemente and S. Hornillo-Mellado and J. Ortega E.S.I., Informática, Universidad de Granada

More information

Chemometrics: Classification of spectra

Chemometrics: Classification of spectra Chemometrics: Classification of spectra Vladimir Bochko Jarmo Alander University of Vaasa November 1, 2010 Vladimir Bochko Chemometrics: Classification 1/36 Contents Terminology Introduction Big picture

More information

Kernel-based Approximation. Methods using MATLAB. Gregory Fasshauer. Interdisciplinary Mathematical Sciences. Michael McCourt.

Kernel-based Approximation. Methods using MATLAB. Gregory Fasshauer. Interdisciplinary Mathematical Sciences. Michael McCourt. SINGAPORE SHANGHAI Vol TAIPEI - Interdisciplinary Mathematical Sciences 19 Kernel-based Approximation Methods using MATLAB Gregory Fasshauer Illinois Institute of Technology, USA Michael McCourt University

More information

GAUSSIAN PROCESS REGRESSION

GAUSSIAN PROCESS REGRESSION GAUSSIAN PROCESS REGRESSION CSE 515T Spring 2015 1. BACKGROUND The kernel trick again... The Kernel Trick Consider again the linear regression model: y(x) = φ(x) w + ε, with prior p(w) = N (w; 0, Σ). The

More information

Nonlinear Support Vector Machines through Iterative Majorization and I-Splines

Nonlinear Support Vector Machines through Iterative Majorization and I-Splines Nonlinear Support Vector Machines through Iterative Majorization and I-Splines P.J.F. Groenen G. Nalbantov J.C. Bioch July 9, 26 Econometric Institute Report EI 26-25 Abstract To minimize the primal support

More information

Classification with Kernel Mahalanobis Distance Classifiers

Classification with Kernel Mahalanobis Distance Classifiers Classification with Kernel Mahalanobis Distance Classifiers Bernard Haasdonk and Elżbieta P ekalska 2 Institute of Numerical and Applied Mathematics, University of Münster, Germany, haasdonk@math.uni-muenster.de

More information

About this class. Maximizing the Margin. Maximum margin classifiers. Picture of large and small margin hyperplanes

About this class. Maximizing the Margin. Maximum margin classifiers. Picture of large and small margin hyperplanes About this class Maximum margin classifiers SVMs: geometric derivation of the primal problem Statement of the dual problem The kernel trick SVMs as the solution to a regularization problem Maximizing the

More information

A Magiv CV Theory for Large-Margin Classifiers

A Magiv CV Theory for Large-Margin Classifiers A Magiv CV Theory for Large-Margin Classifiers Hui Zou School of Statistics, University of Minnesota June 30, 2018 Joint work with Boxiang Wang Outline 1 Background 2 Magic CV formula 3 Magic support vector

More information

EE613 Machine Learning for Engineers. Kernel methods Support Vector Machines. jean-marc odobez 2015

EE613 Machine Learning for Engineers. Kernel methods Support Vector Machines. jean-marc odobez 2015 EE613 Machine Learning for Engineers Kernel methods Support Vector Machines jean-marc odobez 2015 overview Kernel methods introductions and main elements defining kernels Kernelization of k-nn, K-Means,

More information

MATH 590: Meshfree Methods

MATH 590: Meshfree Methods MATH 590: Meshfree Methods Chapter 33: Adaptive Iteration Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Fall 2010 fasshauer@iit.edu MATH 590 Chapter 33 1 Outline 1 A

More information

Computer Vision Group Prof. Daniel Cremers. 2. Regression (cont.)

Computer Vision Group Prof. Daniel Cremers. 2. Regression (cont.) Prof. Daniel Cremers 2. Regression (cont.) Regression with MLE (Rep.) Assume that y is affected by Gaussian noise : t = f(x, w)+ where Thus, we have p(t x, w, )=N (t; f(x, w), 2 ) 2 Maximum A-Posteriori

More information

... SPARROW. SPARse approximation Weighted regression. Pardis Noorzad. Department of Computer Engineering and IT Amirkabir University of Technology

... SPARROW. SPARse approximation Weighted regression. Pardis Noorzad. Department of Computer Engineering and IT Amirkabir University of Technology ..... SPARROW SPARse approximation Weighted regression Pardis Noorzad Department of Computer Engineering and IT Amirkabir University of Technology Université de Montréal March 12, 2012 SPARROW 1/47 .....

More information

Support Vector Machine (SVM) and Kernel Methods

Support Vector Machine (SVM) and Kernel Methods Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2014 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin

More information

Computation of operators in wavelet coordinates

Computation of operators in wavelet coordinates Computation of operators in wavelet coordinates Tsogtgerel Gantumur and Rob Stevenson Department of Mathematics Utrecht University Tsogtgerel Gantumur - Computation of operators in wavelet coordinates

More information

Machine Learning, Fall 2011: Homework 5

Machine Learning, Fall 2011: Homework 5 0-60 Machine Learning, Fall 0: Homework 5 Machine Learning Department Carnegie Mellon University Due:??? Instructions There are 3 questions on this assignment. Please submit your completed homework to

More information

ECS289: Scalable Machine Learning

ECS289: Scalable Machine Learning ECS289: Scalable Machine Learning Cho-Jui Hsieh UC Davis Sept 29, 2016 Outline Convex vs Nonconvex Functions Coordinate Descent Gradient Descent Newton s method Stochastic Gradient Descent Numerical Optimization

More information

MATH 590: Meshfree Methods

MATH 590: Meshfree Methods MATH 590: Meshfree Methods Chapter 33: Adaptive Iteration Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Fall 2010 fasshauer@iit.edu MATH 590 Chapter 33 1 Outline 1 A

More information

TUM 2016 Class 3 Large scale learning by regularization

TUM 2016 Class 3 Large scale learning by regularization TUM 2016 Class 3 Large scale learning by regularization Lorenzo Rosasco UNIGE-MIT-IIT July 25, 2016 Learning problem Solve min w E(w), E(w) = dρ(x, y)l(w x, y) given (x 1, y 1 ),..., (x n, y n ) Beyond

More information

COMP 551 Applied Machine Learning Lecture 20: Gaussian processes

COMP 551 Applied Machine Learning Lecture 20: Gaussian processes COMP 55 Applied Machine Learning Lecture 2: Gaussian processes Instructor: Ryan Lowe (ryan.lowe@cs.mcgill.ca) Slides mostly by: (herke.vanhoof@mcgill.ca) Class web page: www.cs.mcgill.ca/~hvanho2/comp55

More information

MATH 590: Meshfree Methods

MATH 590: Meshfree Methods MATH 590: Meshfree Methods Chapter 14: The Power Function and Native Space Error Estimates Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Fall 2010 fasshauer@iit.edu

More information

Introduction to emulators - the what, the when, the why

Introduction to emulators - the what, the when, the why School of Earth and Environment INSTITUTE FOR CLIMATE & ATMOSPHERIC SCIENCE Introduction to emulators - the what, the when, the why Dr Lindsay Lee 1 What is a simulator? A simulator is a computer code

More information

Gaussian processes for inference in stochastic differential equations

Gaussian processes for inference in stochastic differential equations Gaussian processes for inference in stochastic differential equations Manfred Opper, AI group, TU Berlin November 6, 2017 Manfred Opper, AI group, TU Berlin (TU Berlin) inference in SDE November 6, 2017

More information

Support Vector Machine (continued)

Support Vector Machine (continued) Support Vector Machine continued) Overlapping class distribution: In practice the class-conditional distributions may overlap, so that the training data points are no longer linearly separable. We need

More information

Scale-Invariance of Support Vector Machines based on the Triangular Kernel. Abstract

Scale-Invariance of Support Vector Machines based on the Triangular Kernel. Abstract Scale-Invariance of Support Vector Machines based on the Triangular Kernel François Fleuret Hichem Sahbi IMEDIA Research Group INRIA Domaine de Voluceau 78150 Le Chesnay, France Abstract This paper focuses

More information

Support Vector Machines

Support Vector Machines Support Vector Machines Here we approach the two-class classification problem in a direct way: We try and find a plane that separates the classes in feature space. If we cannot, we get creative in two

More information

Kernel methods and the exponential family

Kernel methods and the exponential family Kernel methods and the exponential family Stéphane Canu 1 and Alex J. Smola 2 1- PSI - FRE CNRS 2645 INSA de Rouen, France St Etienne du Rouvray, France Stephane.Canu@insa-rouen.fr 2- Statistical Machine

More information

MACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA

MACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA 1 MACHINE LEARNING Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA 2 Practicals Next Week Next Week, Practical Session on Computer Takes Place in Room GR

More information

Max Margin-Classifier

Max Margin-Classifier Max Margin-Classifier Oliver Schulte - CMPT 726 Bishop PRML Ch. 7 Outline Maximum Margin Criterion Math Maximizing the Margin Non-Separable Data Kernels and Non-linear Mappings Where does the maximization

More information

Support Vector Regression with Automatic Accuracy Control B. Scholkopf y, P. Bartlett, A. Smola y,r.williamson FEIT/RSISE, Australian National University, Canberra, Australia y GMD FIRST, Rudower Chaussee

More information

Machine Learning : Support Vector Machines

Machine Learning : Support Vector Machines Machine Learning Support Vector Machines 05/01/2014 Machine Learning : Support Vector Machines Linear Classifiers (recap) A building block for almost all a mapping, a partitioning of the input space into

More information

Support Vector Machine (SVM) and Kernel Methods

Support Vector Machine (SVM) and Kernel Methods Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2016 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin

More information

Statistical Methods for NLP

Statistical Methods for NLP Statistical Methods for NLP Text Categorization, Support Vector Machines Sameer Maskey Announcement Reading Assignments Will be posted online tonight Homework 1 Assigned and available from the course website

More information

Support Vector Machines and Kernel Methods

Support Vector Machines and Kernel Methods 2018 CS420 Machine Learning, Lecture 3 Hangout from Prof. Andrew Ng. http://cs229.stanford.edu/notes/cs229-notes3.pdf Support Vector Machines and Kernel Methods Weinan Zhang Shanghai Jiao Tong University

More information

A direct formulation for sparse PCA using semidefinite programming

A direct formulation for sparse PCA using semidefinite programming A direct formulation for sparse PCA using semidefinite programming A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley Available online at www.princeton.edu/~aspremon

More information

Pattern Recognition and Machine Learning. Bishop Chapter 6: Kernel Methods

Pattern Recognition and Machine Learning. Bishop Chapter 6: Kernel Methods Pattern Recognition and Machine Learning Chapter 6: Kernel Methods Vasil Khalidov Alex Kläser December 13, 2007 Training Data: Keep or Discard? Parametric methods (linear/nonlinear) so far: learn parameter

More information

Support Vector Machine Regression for Volatile Stock Market Prediction

Support Vector Machine Regression for Volatile Stock Market Prediction Support Vector Machine Regression for Volatile Stock Market Prediction Haiqin Yang, Laiwan Chan, and Irwin King Department of Computer Science and Engineering The Chinese University of Hong Kong Shatin,

More information

(Kernels +) Support Vector Machines

(Kernels +) Support Vector Machines (Kernels +) Support Vector Machines Machine Learning Torsten Möller Reading Chapter 5 of Machine Learning An Algorithmic Perspective by Marsland Chapter 6+7 of Pattern Recognition and Machine Learning

More information

Learning gradients: prescriptive models

Learning gradients: prescriptive models Department of Statistical Science Institute for Genome Sciences & Policy Department of Computer Science Duke University May 11, 2007 Relevant papers Learning Coordinate Covariances via Gradients. Sayan

More information

Pattern Recognition and Machine Learning. Perceptrons and Support Vector machines

Pattern Recognition and Machine Learning. Perceptrons and Support Vector machines Pattern Recognition and Machine Learning James L. Crowley ENSIMAG 3 - MMIS Fall Semester 2016 Lessons 6 10 Jan 2017 Outline Perceptrons and Support Vector machines Notation... 2 Perceptrons... 3 History...3

More information

CS798: Selected topics in Machine Learning

CS798: Selected topics in Machine Learning CS798: Selected topics in Machine Learning Support Vector Machine Jakramate Bootkrajang Department of Computer Science Chiang Mai University Jakramate Bootkrajang CS798: Selected topics in Machine Learning

More information

A Polynomial Chaos Approach to Robust Multiobjective Optimization

A Polynomial Chaos Approach to Robust Multiobjective Optimization A Polynomial Chaos Approach to Robust Multiobjective Optimization Silvia Poles 1, Alberto Lovison 2 1 EnginSoft S.p.A., Optimization Consulting Via Giambellino, 7 35129 Padova, Italy s.poles@enginsoft.it

More information

arxiv: v1 [math.na] 3 Apr 2019

arxiv: v1 [math.na] 3 Apr 2019 arxiv:1904.02017v1 [math.na] 3 Apr 2019 Poly-Sinc Solution of Stochastic Elliptic Differential Equations Maha Youssef and Roland Pulch Institute of Mathematics and Computer Science, University of Greifswald,

More information

below, kernel PCA Eigenvectors, and linear combinations thereof. For the cases where the pre-image does exist, we can provide a means of constructing

below, kernel PCA Eigenvectors, and linear combinations thereof. For the cases where the pre-image does exist, we can provide a means of constructing Kernel PCA Pattern Reconstruction via Approximate Pre-Images Bernhard Scholkopf, Sebastian Mika, Alex Smola, Gunnar Ratsch, & Klaus-Robert Muller GMD FIRST, Rudower Chaussee 5, 12489 Berlin, Germany fbs,

More information

Lecture 5: GPs and Streaming regression

Lecture 5: GPs and Streaming regression Lecture 5: GPs and Streaming regression Gaussian Processes Information gain Confidence intervals COMP-652 and ECSE-608, Lecture 5 - September 19, 2017 1 Recall: Non-parametric regression Input space X

More information

Kernel Methods and Support Vector Machines

Kernel Methods and Support Vector Machines Kernel Methods and Support Vector Machines Oliver Schulte - CMPT 726 Bishop PRML Ch. 6 Support Vector Machines Defining Characteristics Like logistic regression, good for continuous input features, discrete

More information

Reproducing Kernel Hilbert Spaces

Reproducing Kernel Hilbert Spaces Reproducing Kernel Hilbert Spaces Lorenzo Rosasco 9.520 Class 03 February 12, 2007 About this class Goal To introduce a particularly useful family of hypothesis spaces called Reproducing Kernel Hilbert

More information

Reproducing Kernel Hilbert Spaces

Reproducing Kernel Hilbert Spaces 9.520: Statistical Learning Theory and Applications February 10th, 2010 Reproducing Kernel Hilbert Spaces Lecturer: Lorenzo Rosasco Scribe: Greg Durrett 1 Introduction In the previous two lectures, we

More information

Tutorial on Gaussian Processes and the Gaussian Process Latent Variable Model

Tutorial on Gaussian Processes and the Gaussian Process Latent Variable Model Tutorial on Gaussian Processes and the Gaussian Process Latent Variable Model (& discussion on the GPLVM tech. report by Prof. N. Lawrence, 06) Andreas Damianou Department of Neuro- and Computer Science,

More information

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition NONLINEAR CLASSIFICATION AND REGRESSION Nonlinear Classification and Regression: Outline 2 Multi-Layer Perceptrons The Back-Propagation Learning Algorithm Generalized Linear Models Radial Basis Function

More information

Discussion of Some Problems About Nonlinear Time Series Prediction Using ν-support Vector Machine

Discussion of Some Problems About Nonlinear Time Series Prediction Using ν-support Vector Machine Commun. Theor. Phys. (Beijing, China) 48 (2007) pp. 117 124 c International Academic Publishers Vol. 48, No. 1, July 15, 2007 Discussion of Some Problems About Nonlinear Time Series Prediction Using ν-support

More information

Formulation with slack variables

Formulation with slack variables Formulation with slack variables Optimal margin classifier with slack variables and kernel functions described by Support Vector Machine (SVM). min (w,ξ) ½ w 2 + γσξ(i) subject to ξ(i) 0 i, d(i) (w T x(i)

More information

Empirical Interpolation of Nonlinear Parametrized Evolution Operators

Empirical Interpolation of Nonlinear Parametrized Evolution Operators MÜNSTER Empirical Interpolation of Nonlinear Parametrized Evolution Operators Martin Drohmann, Bernard Haasdonk, Mario Ohlberger 03/12/2010 MÜNSTER 2/20 > Motivation: Reduced Basis Method RB Scenario:

More information

Support Vector Machines

Support Vector Machines Support Vector Machines Le Song Machine Learning I CSE 6740, Fall 2013 Naïve Bayes classifier Still use Bayes decision rule for classification P y x = P x y P y P x But assume p x y = 1 is fully factorized

More information

Neural Networks and Deep Learning

Neural Networks and Deep Learning Neural Networks and Deep Learning Professor Ameet Talwalkar November 12, 2015 Professor Ameet Talwalkar Neural Networks and Deep Learning November 12, 2015 1 / 16 Outline 1 Review of last lecture AdaBoost

More information