Model reduction of Multiscale Models using Kernel Methods
|
|
- Lester Higgins
- 6 years ago
- Views:
Transcription
1 Model reduction of Multiscale Models using Kernel Methods at the WORM 213 Haus der Kirche, Bad Herrenalb Daniel Wirtz August 23, 213 Dipl. Math. Dipl. Inf. Daniel Wirtz, IANS, University of Stuttgart Jun.-Prof. Dr. Bernard Haasdonk, IANS, University of Stuttgart 1/52
2 Introduction Contents Introduction Kernel Expansion construction Support Vector Machines Vectorial Greedy Algorithms Experimental results Human spine simulation (N. Karajan) Shock waves in porous media (F. Kissling) 2/52
3 Introduction Up Next Introduction 3/52
4 Introduction Reduction setting C Macro- and D microscale model. Interaction parameters from domains P C, P D One macrolevel step requires many microlevel solves Each simulation of D is actually a function evaluation x P D y start C D y P C Figure: Communication of macro and microscale models f : P D P C x P D (x). Goal: Find a surrogate D, i.e. f f Figure: Example multibody system C with many microscale models D i 4/52
5 Introduction Buzzwords! Response Surface Methodology [Myers et al.(29)myers, Montgomery & Anderson-Cook, Khuri & Mukhopadhyay(21)] Surrogate Modeling Figure: Illustrations of response surfaces 5/52
6 Introduction Building surfaces How to build a surrogate or response surface? Classically via polynomials Interpolation, Regression Multivariate variants Polynomial Chaos Expansions [Wiener(1938), Karajan et al.(213)karajan, Ehlers, Oladyshkin & Otto.] Alternative: Kernels! Symmetric function K : R d R d R Positive definite (analysis!) E.g. Gaussian (for γ > ) K g (x, y) = exp( x y 2 /γ 2 ) 6/52
7 Introduction More famous kernel examples Compactly supported Wendland kernels Kw d,k (x, y) = φ d,k w ( ) x y for d, k N and polynomials γ, φ d,k w (r) = (1 r) l+k + p d,k (r). p d,1 (r) = (l + 1)r + 1, l = d/2 + k + 1 p d,2 (r) = 1 3 (l2 + 4l + 3)r 2 + (l + 2)r + 1, [...] γ is a dilation hyperparameter for the kernel width. d, k determines the smoothness in the sense of K d,k w C 2k (P D ) if P D R s, s d. 7/52
8 Introduction Plots for different kernels Figure: Examples. 1D: K g(γ =.33), Kw 2,, K1,1 w.2d: Kg(γ =.49), K2, w, K3,3 w. γ = 1 for all Kw. 8/52
9 Introduction From kernels to kernel expansions f = Kernel expansion! f(x) = N c i K(x i, x) i=1 with centers x i R d and coefficients c i R. Figure: Illustration for a 1D Gaussian kernel expansion with γ = 1 and x i = [ 1,.5, 1.2], c i = [.5,.2,.8] 9/52
10 Expansion construction Up Next Expansion construction 1/52
11 Expansion construction How to obtain kernel expansions? Various methods available Mostly using only training data {x 1,..., x N },{y 1,..., y N } Support Vector Machines / Regression [Schölkopf & Smola(22)] 1 f = arg min g H N N i=1 V ( y i g(x i) ) + λ g 2 H, V (e) = max{, e ɛ} f = /52
12 Expansion construction ɛ-svr Support Vector Regression with ɛ-loss (Dual problem) For C := 1 2λN and y = (y 1,..., y N ) T R N we solve max α +,α R N W (α +, α ) (1) If α +, α s.t. α +, i C, i = 1... N, = α + i α i, i = 1... N, W (α +, α ) := 1 ( α + α ) T ( K α + α ) 2 N ɛ (α + i + α i ) + y T ( α + α ). i=1 solve (1), then N f(x) := c ik(x, x i), i=1 c i := (α α + ) i Computation via Sequential Minimal Optimization (SMO) 12/52
13 Expansion construction ɛ-svr illustration: 1D SMO.5 Iteration 1 f(x) and approximation x 13/52 Figure: ɛ-svr 1D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers
14 Expansion construction ɛ-svr illustration: 1D SMO.5 Iteration 2 f(x) and approximation x 14/52 Figure: ɛ-svr 1D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers
15 Expansion construction ɛ-svr illustration: 1D SMO.5 Iteration 3 f(x) and approximation x 15/52 Figure: ɛ-svr 1D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers
16 Expansion construction ɛ-svr illustration: 1D SMO.5 Iteration 4 f(x) and approximation x 16/52 Figure: ɛ-svr 1D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers
17 Expansion construction ɛ-svr illustration: 1D SMO.5 Iteration 5 f(x) and approximation x 17/52 Figure: ɛ-svr 1D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers
18 Expansion construction ɛ-svr illustration: 1D SMO.5 Iteration 6 f(x) and approximation x 18/52 Figure: ɛ-svr 1D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers
19 Expansion construction ɛ-svr illustration: 1D SMO.5 Iteration 7 f(x) and approximation x Figure: ɛ-svr 1D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers 19/52
20 Expansion construction ɛ-svr illustration: 1D SMO.5 Iteration 8 f(x) and approximation x Figure: ɛ-svr 1D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers 2/52
21 Expansion construction ɛ-svr illustration: 2D SMO.5 Iteration 1 f(x) and approximation x 21/52 Figure: ɛ-svr 2D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers
22 Expansion construction ɛ-svr illustration: 2D SMO.5 Iteration 2 f(x) and approximation x 22/52 Figure: ɛ-svr 2D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers
23 Expansion construction ɛ-svr illustration: 2D SMO.5 Iteration 3 f(x) and approximation x 23/52 Figure: ɛ-svr 2D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers
24 Expansion construction ɛ-svr illustration: 2D SMO.5 Iteration 4 f(x) and approximation x 24/52 Figure: ɛ-svr 2D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers
25 Expansion construction ɛ-svr illustration: 2D SMO.5 Iteration 5 f(x) and approximation x 25/52 Figure: ɛ-svr 2D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers
26 Expansion construction ɛ-svr illustration: 2D SMO.5 Iteration 6 f(x) and approximation x 26/52 Figure: ɛ-svr 2D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers
27 Expansion construction ɛ-svr illustration: 2D SMO.5 Iteration 7 f(x) and approximation x Figure: ɛ-svr 2D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers 27/52
28 Expansion construction ɛ-svr illustration: 2D SMO.5 Iteration 8 f(x) and approximation x Figure: ɛ-svr 2D SMO iterations. Red: Target function with margin ɛ =.1, blue: approximation, black: centers 28/52
29 Expansion construction ɛ-svr illustration: 2D SMO full solution.5 It: 3523, #SV=2, eps=.1 f(x) x Figure: ɛ-svr 2D SMO solution. Red: Target function with margin ɛ =.1, blue: approximation, black: centers 29/52
30 Expansion construction The VKOGA Vectorial kernel orthogonal greedy algorithm Let m N and f H m, define X :=, f := and for k >, Ω k := { x P D K(x, ) H X k 1}, the sequences x k := arg max x P D \Ω k X k := X k 1 {x k }, f k := I m X k [f]. f(x) f k 1 (x) 2 2 K(x, ) IXk 1 [K(x, )], H Here, I m X k [f] denotes component-wise interpolation on X k. In/Exclusion of the denominator yields different variants. Details in [Wirtz & Haasdonk(213)] 3/52
31 Expansion construction VKOGA illustration 1.5 Iteration f(x) x 31/52 Figure: VKOGA iterations one to four. Green/Blue: Components, Dashed: Approximations, Red: Total error, Diamonds: Centers, Red circle: Max overall error, Black lines: Max component errors
32 Expansion construction VKOGA illustration 1.5 Iteration f(x) x 32/52 Figure: VKOGA iterations one to four. Green/Blue: Components, Dashed: Approximations, Red: Total error, Diamonds: Centers, Red circle: Max overall error, Black lines: Max component errors
33 Expansion construction VKOGA illustration 1.5 Iteration f(x) x 33/52 Figure: VKOGA iterations one to four. Green/Blue: Components, Dashed: Approximations, Red: Total error, Diamonds: Centers, Red circle: Max overall error, Black lines: Max component errors
34 Expansion construction VKOGA illustration 1.5 Iteration f(x) x 34/52 Figure: VKOGA iterations one to four. Green/Blue: Components, Dashed: Approximations, Red: Total error, Diamonds: Centers, Red circle: Max overall error, Black lines: Max component errors
35 Experimental results Up Next Experimental results 35/52
36 Experimental results Multibody spine simulation: Introduction Cooperation with N. Karajan, D. Otto Linking MBS spine simulation with FE intervertebral model Figure: Illustration from [Karajan et al.(211)karajan, Röhrle, Ehlers & Schmitt] 36/52
37 Experimental results Multibody spine simulation: Animations for bone and disk interactions Figure: Animations for MBS vertebra interaction and ϕ 1, u 2, u 3 disk response cases 37/52
38 Experimental results Multibody spine simulation: Combined load cases for IVDs Figure: Different load case animations for general and geometry-fitted intervertebral disks 38/52
39 Experimental results Multibody spine simulation: Combined load cases for IVDs Computation with quasi-static timesteps from equillibrium state Figure: Joystick like computation of training data 39/52
40 Experimental results Multibody spine simulation: Approximation Plot of M1 against u2 and u3 FEM inputs ϕ 1, u 2, u 3 P D R 3 Disc response M 1, R 2, R 3 P C R 3 Precomputation of 274/1433 disc responses About 3min average simulation time each dot Future work: Actual input is ϕ 1, u 2, u 3, M 1, R 2, R 3 from last (time)step M u3 Plot of M1 against u2 and phi1 x phi1 u2 M1 2 2 u2 3 2 Figure: Plots of training data 4/52
41 Experimental results Multibody spine simulation: Results with Gauss kernels 41/52 Figure: Plots of approximation, u2, u3 M1, ϕ1 = [ 5.7,, 5.3] (top) and u2, ϕ1 M1, u3 = [ 3.1,, 3.1] (bottom)
42 Experimental results Multibody spine simulation: Results with Wendland kernels Figure: Plots of approximation using Wendland kernels and best configuration γ : , k : 1, d : 2 42/52
43 Experimental results Multibody spine simulation: Outlook Simulation using kernel approximations! So far: Using Polynomial Chaos Expansions Figure: Simulation of full multibody spine model using suggorates 43/52
44 Experimental results Two-phase flow in porous media: Intro Cooperation with F. Kissling (IANS) Two-phase flow in porous media with nonclassical shock waves Reg. Buckley-Leverett problem S t + (vf(s)) = ɛs xx + ɛ 2 τs xxt On front: Solve Riemann problems for overshoot Figure: Illustrations from [Kissling & Rhode(212)] 44/52
45 Experimental results Two-phase flow in porous media: Approximation Riemann problem initial data Left-side saturation s l Scaling parameter τ Result: Plateau value S ɛ Precomputation of 5 Rieman problems Suitable transformation of training data Scaling Logarithmic values Figure: Plots of training data 45/52
46 Experimental results Two-phase flow in porous media: Results 46/52 Figure: Plots of approximations. Top: SVR (Gauss), Bottom: Greedy algorithm (Wendland)
47 Experimental results Two-phase flow in porous media: Results Time-reduction from 2h to 24min for full simulation N S S SV R L Figure: Full simulation plots and error table. Full model (top) and SVR reduced (bottom) Details in [Kissling & Rhode(212)] 47/52
48 Summary What did we see? Setting: Reduction for Microscale models Use of cheap surrogates for micro-scale level Kernel methods can provide such replacements! Support vector regression (SVR) Vectorial greedy algorithms (VKOGA) Examples and results available: Human Spine Simulations Shock-waves in porous media 48/52
49 -THE END-... any questions? Thank you for your attention! 49/52
50 References I [Karajan et al.(213)karajan, Ehlers, Oladyshkin & Otto.] Karajan, N.; Ehlers, W.; Oladyshkin, S. & Otto., D.: Application of the Polynomial Chaos Expansion to Approximate the Homogenised Behaviour of the Intervertebral Disc. Preprint, University of Stuttgart, SRC SimTech (213), in preparation. [Karajan et al.(211)karajan, Röhrle, Ehlers & Schmitt] Karajan, N.; Röhrle, O.; Ehlers, W. & Schmitt, S.: Linking Continuous and Discrete Intervertebral Disc Models through Homogenisation. Preprint, SRC SimTech (211), URL uni-stuttgart.de/publikationen/prints.php?id=657, submitted to Biomechanics and Modeling in Mechanobiology. [Khuri & Mukhopadhyay(21)] Khuri, A. I. & Mukhopadhyay, S.: Response surface methodology. Wiley Interdisciplinary Reviews: Computational Statistics 2 (21), , ISSN /52
51 References II [Kissling & Rhode(212)] Kissling, F. & Rhode, C.: The Computation of Nonclassical Shock Waves in Porous Media with a Heterogeneous Multiscale Method: The Multidimensional Case. Preprint, Stuttgart Research Centre for Simulation Technology (212), URL publikationen/prints.php?id=734, submitted to SIAM Multiscale Modeling and Simulation. [Myers et al.(29)myers, Montgomery & Anderson-Cook] Myers, R. H.; Montgomery, D. C. & Anderson-Cook, C. M.: Response Surface Methodology - Process and Product Optimization Using Designed Experiments. Wiley Series in Probability and Statistics, Wiley 29, 3 edn. [Schölkopf & Smola(22)] Schölkopf, B. & Smola, A. J.: Learning with Kernels. Adaptive Computation and Machine Learning, The MIT Press /52
52 References III [Wiener(1938)] Wiener, N.: The homogeneous chaos. Amer. J. Math. 6 (1938), , ISSN 29327, URL [Wirtz & Haasdonk(213)] Wirtz, D. & Haasdonk, B.: An improved vectorial kernel orthogonal greedy algorithm. Dolomites Research Notes on Approximation 6 (213), /52
CIS 520: Machine Learning Oct 09, Kernel Methods
CIS 520: Machine Learning Oct 09, 207 Kernel Methods Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture They may or may not cover all the material discussed
More informationA-Posteriori Error Estimation for Parameterized Kernel-Based Systems
A-Posteriori Error Estimation for Parameterized Kernel-Based Systems Daniel Wirtz Bernard Haasdonk Institute for Applied Analysis and Numerical Simulation University of Stuttgart 70569 Stuttgart, Pfaffenwaldring
More informationSupport Vector Machine for Classification and Regression
Support Vector Machine for Classification and Regression Ahlame Douzal AMA-LIG, Université Joseph Fourier Master 2R - MOSIG (2013) November 25, 2013 Loss function, Separating Hyperplanes, Canonical Hyperplan
More informationStochastic representation of random positive-definite tensor-valued properties: application to 3D anisotropic permeability random fields
Sous la co-tutelle de : LABORATOIRE DE MODÉLISATION ET SIMULATION MULTI ÉCHELLE CNRS UPEC UNIVERSITÉ PARIS-EST CRÉTEIL UPEM UNIVERSITÉ PARIS-EST MARNE-LA-VALLÉE Stochastic representation of random positive-definite
More informationKernel Methods. Outline
Kernel Methods Quang Nguyen University of Pittsburgh CS 3750, Fall 2011 Outline Motivation Examples Kernels Definitions Kernel trick Basic properties Mercer condition Constructing feature space Hilbert
More informationSMO Algorithms for Support Vector Machines without Bias Term
Institute of Automatic Control Laboratory for Control Systems and Process Automation Prof. Dr.-Ing. Dr. h. c. Rolf Isermann SMO Algorithms for Support Vector Machines without Bias Term Michael Vogt, 18-Jul-2002
More informationA GENERAL FORMULATION FOR SUPPORT VECTOR MACHINES. Wei Chu, S. Sathiya Keerthi, Chong Jin Ong
A GENERAL FORMULATION FOR SUPPORT VECTOR MACHINES Wei Chu, S. Sathiya Keerthi, Chong Jin Ong Control Division, Department of Mechanical Engineering, National University of Singapore 0 Kent Ridge Crescent,
More informationAn Improved Conjugate Gradient Scheme to the Solution of Least Squares SVM
An Improved Conjugate Gradient Scheme to the Solution of Least Squares SVM Wei Chu Chong Jin Ong chuwei@gatsby.ucl.ac.uk mpeongcj@nus.edu.sg S. Sathiya Keerthi mpessk@nus.edu.sg Control Division, Department
More informationComputer Vision Group Prof. Daniel Cremers. 4. Gaussian Processes - Regression
Group Prof. Daniel Cremers 4. Gaussian Processes - Regression Definition (Rep.) Definition: A Gaussian process is a collection of random variables, any finite number of which have a joint Gaussian distribution.
More informationν =.1 a max. of 10% of training set can be margin errors ν =.8 a max. of 80% of training can be margin errors
p.1/1 ν-svms In the traditional softmargin classification SVM formulation we have a penalty constant C such that 1 C size of margin. Furthermore, there is no a priori guidance as to what C should be set
More informationAdvanced Introduction to Machine Learning
10-715 Advanced Introduction to Machine Learning Homework Due Oct 15, 10.30 am Rules Please follow these guidelines. Failure to do so, will result in loss of credit. 1. Homework is due on the due date
More informationStatistical Machine Learning from Data
Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Support Vector Machines Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole Polytechnique
More informationApproximation Theoretical Questions for SVMs
Ingo Steinwart LA-UR 07-7056 October 20, 2007 Statistical Learning Theory: an Overview Support Vector Machines Informal Description of the Learning Goal X space of input samples Y space of labels, usually
More informationReview: Support vector machines. Machine learning techniques and image analysis
Review: Support vector machines Review: Support vector machines Margin optimization min (w,w 0 ) 1 2 w 2 subject to y i (w 0 + w T x i ) 1 0, i = 1,..., n. Review: Support vector machines Margin optimization
More informationComputer Vision Group Prof. Daniel Cremers. 9. Gaussian Processes - Regression
Group Prof. Daniel Cremers 9. Gaussian Processes - Regression Repetition: Regularized Regression Before, we solved for w using the pseudoinverse. But: we can kernelize this problem as well! First step:
More informationRecent Results for Moving Least Squares Approximation
Recent Results for Moving Least Squares Approximation Gregory E. Fasshauer and Jack G. Zhang Abstract. We describe two experiments recently conducted with the approximate moving least squares (MLS) approximation
More informationA model order reduction technique for speeding up computational homogenisation
A model order reduction technique for speeding up computational homogenisation Olivier Goury, Pierre Kerfriden, Wing Kam Liu, Stéphane Bordas Cardiff University Outline Introduction Heterogeneous materials
More informationSupport Vector Machines II. CAP 5610: Machine Learning Instructor: Guo-Jun QI
Support Vector Machines II CAP 5610: Machine Learning Instructor: Guo-Jun QI 1 Outline Linear SVM hard margin Linear SVM soft margin Non-linear SVM Application Linear Support Vector Machine An optimization
More informationLecture 10: A brief introduction to Support Vector Machine
Lecture 10: A brief introduction to Support Vector Machine Advanced Applied Multivariate Analysis STAT 2221, Fall 2013 Sungkyu Jung Department of Statistics, University of Pittsburgh Xingye Qiao Department
More informationIntroduction to SVM and RVM
Introduction to SVM and RVM Machine Learning Seminar HUS HVL UIB Yushu Li, UIB Overview Support vector machine SVM First introduced by Vapnik, et al. 1992 Several literature and wide applications Relevance
More informationSolving the SVM Optimization Problem
Solving the SVM Optimization Problem Kernel-based Learning Methods Christian Igel Institut für Neuroinformatik Ruhr-Universität Bochum, Germany http://www.neuroinformatik.rub.de July 16, 2009 Christian
More informationEmploying Model Reduction for Uncertainty Visualization in the Context of CO 2 Storage Simulation
Employing Model Reduction for Uncertainty Visualization in the Context of CO 2 Storage Simulation Marcel Hlawatsch, Sergey Oladyshkin, Daniel Weiskopf University of Stuttgart Problem setting - underground
More informationSupport Vector Machines. Maximizing the Margin
Support Vector Machines Support vector achines (SVMs) learn a hypothesis: h(x) = b + Σ i= y i α i k(x, x i ) (x, y ),..., (x, y ) are the training exs., y i {, } b is the bias weight. α,..., α are the
More informationMLCC 2018 Variable Selection and Sparsity. Lorenzo Rosasco UNIGE-MIT-IIT
MLCC 2018 Variable Selection and Sparsity Lorenzo Rosasco UNIGE-MIT-IIT Outline Variable Selection Subset Selection Greedy Methods: (Orthogonal) Matching Pursuit Convex Relaxation: LASSO & Elastic Net
More informationNon-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines
Non-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Fall 2018 CS 551, Fall
More informationSmooth Bayesian Kernel Machines
Smooth Bayesian Kernel Machines Rutger W. ter Borg 1 and Léon J.M. Rothkrantz 2 1 Nuon NV, Applied Research & Technology Spaklerweg 20, 1096 BA Amsterdam, the Netherlands rutger@terborg.net 2 Delft University
More informationCOMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017
COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University FEATURE EXPANSIONS FEATURE EXPANSIONS
More informationIntroduction to Support Vector Machines
Introduction to Support Vector Machines Andreas Maletti Technische Universität Dresden Fakultät Informatik June 15, 2006 1 The Problem 2 The Basics 3 The Proposed Solution Learning by Machines Learning
More informationSupport Vector Machines Explained
December 23, 2008 Support Vector Machines Explained Tristan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introduction This document has been written in an attempt to make the Support Vector Machines (SVM),
More informationIntroduction to Support Vector Machines
Introduction to Support Vector Machines Shivani Agarwal Support Vector Machines (SVMs) Algorithm for learning linear classifiers Motivated by idea of maximizing margin Efficient extension to non-linear
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 Outlines Overview Introduction Linear Algebra Probability Linear Regression
More informationMachine Learning and Data Mining. Support Vector Machines. Kalev Kask
Machine Learning and Data Mining Support Vector Machines Kalev Kask Linear classifiers Which decision boundary is better? Both have zero training error (perfect training accuracy) But, one of them seems
More informationLeast Squares SVM Regression
Least Squares SVM Regression Consider changing SVM to LS SVM by making following modifications: min (w,e) ½ w 2 + ½C Σ e(i) 2 subject to d(i) (w T Φ( x(i))+ b) = e(i), i, and C>0. Note that e(i) is error
More informationSupport Vector Machines: Maximum Margin Classifiers
Support Vector Machines: Maximum Margin Classifiers Machine Learning and Pattern Recognition: September 16, 2008 Piotr Mirowski Based on slides by Sumit Chopra and Fu-Jie Huang 1 Outline What is behind
More informationHilbert Space Methods in Learning
Hilbert Space Methods in Learning guest lecturer: Risi Kondor 6772 Advanced Machine Learning and Perception (Jebara), Columbia University, October 15, 2003. 1 1. A general formulation of the learning problem
More informationComputational tractability of machine learning algorithms for tall fat data
Computational tractability of machine learning algorithms for tall fat data Getting good enough solutions as fast as possible Vikas Chandrakant Raykar vikas@cs.umd.edu University of Maryland, CollegePark
More informationModeling Dependence of Daily Stock Prices and Making Predictions of Future Movements
Modeling Dependence of Daily Stock Prices and Making Predictions of Future Movements Taavi Tamkivi, prof Tõnu Kollo Institute of Mathematical Statistics University of Tartu 29. June 2007 Taavi Tamkivi,
More informationBANA 7046 Data Mining I Lecture 6. Other Data Mining Algorithms 1
BANA 7046 Data Mining I Lecture 6. Other Data Mining Algorithms 1 Shaobo Li University of Cincinnati 1 Partially based on Hastie, et al. (2009) ESL, and James, et al. (2013) ISLR Data Mining I Lecture
More informationReduced Basis Method for Parametric
1/24 Reduced Basis Method for Parametric H 2 -Optimal Control Problems NUMOC2017 Andreas Schmidt, Bernard Haasdonk Institute for Applied Analysis and Numerical Simulation, University of Stuttgart andreas.schmidt@mathematik.uni-stuttgart.de
More informationState Space Representation of Gaussian Processes
State Space Representation of Gaussian Processes Simo Särkkä Department of Biomedical Engineering and Computational Science (BECS) Aalto University, Espoo, Finland June 12th, 2013 Simo Särkkä (Aalto University)
More informationLecture 18: Kernels Risk and Loss Support Vector Regression. Aykut Erdem December 2016 Hacettepe University
Lecture 18: Kernels Risk and Loss Support Vector Regression Aykut Erdem December 2016 Hacettepe University Administrative We will have a make-up lecture on next Saturday December 24, 2016 Presentations
More informationSUPPORT VECTOR MACHINE FOR THE SIMULTANEOUS APPROXIMATION OF A FUNCTION AND ITS DERIVATIVE
SUPPORT VECTOR MACHINE FOR THE SIMULTANEOUS APPROXIMATION OF A FUNCTION AND ITS DERIVATIVE M. Lázaro 1, I. Santamaría 2, F. Pérez-Cruz 1, A. Artés-Rodríguez 1 1 Departamento de Teoría de la Señal y Comunicaciones
More informationLINEAR CLASSIFICATION, PERCEPTRON, LOGISTIC REGRESSION, SVC, NAÏVE BAYES. Supervised Learning
LINEAR CLASSIFICATION, PERCEPTRON, LOGISTIC REGRESSION, SVC, NAÏVE BAYES Supervised Learning Linear vs non linear classifiers In K-NN we saw an example of a non-linear classifier: the decision boundary
More informationKernel Methods. Konstantin Tretyakov MTAT Machine Learning
Kernel Methods Konstantin Tretyakov (kt@ut.ee) MTAT.03.227 Machine Learning So far Supervised machine learning Linear models Non-linear models Unsupervised machine learning Generic scaffolding So far Supervised
More informationKernel Methods. Konstantin Tretyakov MTAT Machine Learning
Kernel Methods Konstantin Tretyakov (kt@ut.ee) MTAT.03.227 Machine Learning So far Supervised machine learning Linear models Least squares regression, SVR Fisher s discriminant, Perceptron, Logistic model,
More informationIntroduction to Machine Learning
Introduction to Machine Learning Kernel Methods Varun Chandola Computer Science & Engineering State University of New York at Buffalo Buffalo, NY, USA chandola@buffalo.edu Chandola@UB CSE 474/574 1 / 21
More informationOn-line Support Vector Machines and Optimization Strategies
On-line Support Vector Machines and Optimization Strategies J.M. Górriz, C.G. Puntonet, M. Salmerón, R. Martin-Clemente and S. Hornillo-Mellado and J. Ortega E.S.I., Informática, Universidad de Granada
More informationChemometrics: Classification of spectra
Chemometrics: Classification of spectra Vladimir Bochko Jarmo Alander University of Vaasa November 1, 2010 Vladimir Bochko Chemometrics: Classification 1/36 Contents Terminology Introduction Big picture
More informationKernel-based Approximation. Methods using MATLAB. Gregory Fasshauer. Interdisciplinary Mathematical Sciences. Michael McCourt.
SINGAPORE SHANGHAI Vol TAIPEI - Interdisciplinary Mathematical Sciences 19 Kernel-based Approximation Methods using MATLAB Gregory Fasshauer Illinois Institute of Technology, USA Michael McCourt University
More informationGAUSSIAN PROCESS REGRESSION
GAUSSIAN PROCESS REGRESSION CSE 515T Spring 2015 1. BACKGROUND The kernel trick again... The Kernel Trick Consider again the linear regression model: y(x) = φ(x) w + ε, with prior p(w) = N (w; 0, Σ). The
More informationNonlinear Support Vector Machines through Iterative Majorization and I-Splines
Nonlinear Support Vector Machines through Iterative Majorization and I-Splines P.J.F. Groenen G. Nalbantov J.C. Bioch July 9, 26 Econometric Institute Report EI 26-25 Abstract To minimize the primal support
More informationClassification with Kernel Mahalanobis Distance Classifiers
Classification with Kernel Mahalanobis Distance Classifiers Bernard Haasdonk and Elżbieta P ekalska 2 Institute of Numerical and Applied Mathematics, University of Münster, Germany, haasdonk@math.uni-muenster.de
More informationAbout this class. Maximizing the Margin. Maximum margin classifiers. Picture of large and small margin hyperplanes
About this class Maximum margin classifiers SVMs: geometric derivation of the primal problem Statement of the dual problem The kernel trick SVMs as the solution to a regularization problem Maximizing the
More informationA Magiv CV Theory for Large-Margin Classifiers
A Magiv CV Theory for Large-Margin Classifiers Hui Zou School of Statistics, University of Minnesota June 30, 2018 Joint work with Boxiang Wang Outline 1 Background 2 Magic CV formula 3 Magic support vector
More informationEE613 Machine Learning for Engineers. Kernel methods Support Vector Machines. jean-marc odobez 2015
EE613 Machine Learning for Engineers Kernel methods Support Vector Machines jean-marc odobez 2015 overview Kernel methods introductions and main elements defining kernels Kernelization of k-nn, K-Means,
More informationMATH 590: Meshfree Methods
MATH 590: Meshfree Methods Chapter 33: Adaptive Iteration Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Fall 2010 fasshauer@iit.edu MATH 590 Chapter 33 1 Outline 1 A
More informationComputer Vision Group Prof. Daniel Cremers. 2. Regression (cont.)
Prof. Daniel Cremers 2. Regression (cont.) Regression with MLE (Rep.) Assume that y is affected by Gaussian noise : t = f(x, w)+ where Thus, we have p(t x, w, )=N (t; f(x, w), 2 ) 2 Maximum A-Posteriori
More information... SPARROW. SPARse approximation Weighted regression. Pardis Noorzad. Department of Computer Engineering and IT Amirkabir University of Technology
..... SPARROW SPARse approximation Weighted regression Pardis Noorzad Department of Computer Engineering and IT Amirkabir University of Technology Université de Montréal March 12, 2012 SPARROW 1/47 .....
More informationSupport Vector Machine (SVM) and Kernel Methods
Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2014 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationComputation of operators in wavelet coordinates
Computation of operators in wavelet coordinates Tsogtgerel Gantumur and Rob Stevenson Department of Mathematics Utrecht University Tsogtgerel Gantumur - Computation of operators in wavelet coordinates
More informationMachine Learning, Fall 2011: Homework 5
0-60 Machine Learning, Fall 0: Homework 5 Machine Learning Department Carnegie Mellon University Due:??? Instructions There are 3 questions on this assignment. Please submit your completed homework to
More informationECS289: Scalable Machine Learning
ECS289: Scalable Machine Learning Cho-Jui Hsieh UC Davis Sept 29, 2016 Outline Convex vs Nonconvex Functions Coordinate Descent Gradient Descent Newton s method Stochastic Gradient Descent Numerical Optimization
More informationMATH 590: Meshfree Methods
MATH 590: Meshfree Methods Chapter 33: Adaptive Iteration Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Fall 2010 fasshauer@iit.edu MATH 590 Chapter 33 1 Outline 1 A
More informationTUM 2016 Class 3 Large scale learning by regularization
TUM 2016 Class 3 Large scale learning by regularization Lorenzo Rosasco UNIGE-MIT-IIT July 25, 2016 Learning problem Solve min w E(w), E(w) = dρ(x, y)l(w x, y) given (x 1, y 1 ),..., (x n, y n ) Beyond
More informationCOMP 551 Applied Machine Learning Lecture 20: Gaussian processes
COMP 55 Applied Machine Learning Lecture 2: Gaussian processes Instructor: Ryan Lowe (ryan.lowe@cs.mcgill.ca) Slides mostly by: (herke.vanhoof@mcgill.ca) Class web page: www.cs.mcgill.ca/~hvanho2/comp55
More informationMATH 590: Meshfree Methods
MATH 590: Meshfree Methods Chapter 14: The Power Function and Native Space Error Estimates Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Fall 2010 fasshauer@iit.edu
More informationIntroduction to emulators - the what, the when, the why
School of Earth and Environment INSTITUTE FOR CLIMATE & ATMOSPHERIC SCIENCE Introduction to emulators - the what, the when, the why Dr Lindsay Lee 1 What is a simulator? A simulator is a computer code
More informationGaussian processes for inference in stochastic differential equations
Gaussian processes for inference in stochastic differential equations Manfred Opper, AI group, TU Berlin November 6, 2017 Manfred Opper, AI group, TU Berlin (TU Berlin) inference in SDE November 6, 2017
More informationSupport Vector Machine (continued)
Support Vector Machine continued) Overlapping class distribution: In practice the class-conditional distributions may overlap, so that the training data points are no longer linearly separable. We need
More informationScale-Invariance of Support Vector Machines based on the Triangular Kernel. Abstract
Scale-Invariance of Support Vector Machines based on the Triangular Kernel François Fleuret Hichem Sahbi IMEDIA Research Group INRIA Domaine de Voluceau 78150 Le Chesnay, France Abstract This paper focuses
More informationSupport Vector Machines
Support Vector Machines Here we approach the two-class classification problem in a direct way: We try and find a plane that separates the classes in feature space. If we cannot, we get creative in two
More informationKernel methods and the exponential family
Kernel methods and the exponential family Stéphane Canu 1 and Alex J. Smola 2 1- PSI - FRE CNRS 2645 INSA de Rouen, France St Etienne du Rouvray, France Stephane.Canu@insa-rouen.fr 2- Statistical Machine
More informationMACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA
1 MACHINE LEARNING Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA 2 Practicals Next Week Next Week, Practical Session on Computer Takes Place in Room GR
More informationMax Margin-Classifier
Max Margin-Classifier Oliver Schulte - CMPT 726 Bishop PRML Ch. 7 Outline Maximum Margin Criterion Math Maximizing the Margin Non-Separable Data Kernels and Non-linear Mappings Where does the maximization
More informationSupport Vector Regression with Automatic Accuracy Control B. Scholkopf y, P. Bartlett, A. Smola y,r.williamson FEIT/RSISE, Australian National University, Canberra, Australia y GMD FIRST, Rudower Chaussee
More informationMachine Learning : Support Vector Machines
Machine Learning Support Vector Machines 05/01/2014 Machine Learning : Support Vector Machines Linear Classifiers (recap) A building block for almost all a mapping, a partitioning of the input space into
More informationSupport Vector Machine (SVM) and Kernel Methods
Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2016 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationStatistical Methods for NLP
Statistical Methods for NLP Text Categorization, Support Vector Machines Sameer Maskey Announcement Reading Assignments Will be posted online tonight Homework 1 Assigned and available from the course website
More informationSupport Vector Machines and Kernel Methods
2018 CS420 Machine Learning, Lecture 3 Hangout from Prof. Andrew Ng. http://cs229.stanford.edu/notes/cs229-notes3.pdf Support Vector Machines and Kernel Methods Weinan Zhang Shanghai Jiao Tong University
More informationA direct formulation for sparse PCA using semidefinite programming
A direct formulation for sparse PCA using semidefinite programming A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley Available online at www.princeton.edu/~aspremon
More informationPattern Recognition and Machine Learning. Bishop Chapter 6: Kernel Methods
Pattern Recognition and Machine Learning Chapter 6: Kernel Methods Vasil Khalidov Alex Kläser December 13, 2007 Training Data: Keep or Discard? Parametric methods (linear/nonlinear) so far: learn parameter
More informationSupport Vector Machine Regression for Volatile Stock Market Prediction
Support Vector Machine Regression for Volatile Stock Market Prediction Haiqin Yang, Laiwan Chan, and Irwin King Department of Computer Science and Engineering The Chinese University of Hong Kong Shatin,
More information(Kernels +) Support Vector Machines
(Kernels +) Support Vector Machines Machine Learning Torsten Möller Reading Chapter 5 of Machine Learning An Algorithmic Perspective by Marsland Chapter 6+7 of Pattern Recognition and Machine Learning
More informationLearning gradients: prescriptive models
Department of Statistical Science Institute for Genome Sciences & Policy Department of Computer Science Duke University May 11, 2007 Relevant papers Learning Coordinate Covariances via Gradients. Sayan
More informationPattern Recognition and Machine Learning. Perceptrons and Support Vector machines
Pattern Recognition and Machine Learning James L. Crowley ENSIMAG 3 - MMIS Fall Semester 2016 Lessons 6 10 Jan 2017 Outline Perceptrons and Support Vector machines Notation... 2 Perceptrons... 3 History...3
More informationCS798: Selected topics in Machine Learning
CS798: Selected topics in Machine Learning Support Vector Machine Jakramate Bootkrajang Department of Computer Science Chiang Mai University Jakramate Bootkrajang CS798: Selected topics in Machine Learning
More informationA Polynomial Chaos Approach to Robust Multiobjective Optimization
A Polynomial Chaos Approach to Robust Multiobjective Optimization Silvia Poles 1, Alberto Lovison 2 1 EnginSoft S.p.A., Optimization Consulting Via Giambellino, 7 35129 Padova, Italy s.poles@enginsoft.it
More informationarxiv: v1 [math.na] 3 Apr 2019
arxiv:1904.02017v1 [math.na] 3 Apr 2019 Poly-Sinc Solution of Stochastic Elliptic Differential Equations Maha Youssef and Roland Pulch Institute of Mathematics and Computer Science, University of Greifswald,
More informationbelow, kernel PCA Eigenvectors, and linear combinations thereof. For the cases where the pre-image does exist, we can provide a means of constructing
Kernel PCA Pattern Reconstruction via Approximate Pre-Images Bernhard Scholkopf, Sebastian Mika, Alex Smola, Gunnar Ratsch, & Klaus-Robert Muller GMD FIRST, Rudower Chaussee 5, 12489 Berlin, Germany fbs,
More informationLecture 5: GPs and Streaming regression
Lecture 5: GPs and Streaming regression Gaussian Processes Information gain Confidence intervals COMP-652 and ECSE-608, Lecture 5 - September 19, 2017 1 Recall: Non-parametric regression Input space X
More informationKernel Methods and Support Vector Machines
Kernel Methods and Support Vector Machines Oliver Schulte - CMPT 726 Bishop PRML Ch. 6 Support Vector Machines Defining Characteristics Like logistic regression, good for continuous input features, discrete
More informationReproducing Kernel Hilbert Spaces
Reproducing Kernel Hilbert Spaces Lorenzo Rosasco 9.520 Class 03 February 12, 2007 About this class Goal To introduce a particularly useful family of hypothesis spaces called Reproducing Kernel Hilbert
More informationReproducing Kernel Hilbert Spaces
9.520: Statistical Learning Theory and Applications February 10th, 2010 Reproducing Kernel Hilbert Spaces Lecturer: Lorenzo Rosasco Scribe: Greg Durrett 1 Introduction In the previous two lectures, we
More informationTutorial on Gaussian Processes and the Gaussian Process Latent Variable Model
Tutorial on Gaussian Processes and the Gaussian Process Latent Variable Model (& discussion on the GPLVM tech. report by Prof. N. Lawrence, 06) Andreas Damianou Department of Neuro- and Computer Science,
More informationNONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition
NONLINEAR CLASSIFICATION AND REGRESSION Nonlinear Classification and Regression: Outline 2 Multi-Layer Perceptrons The Back-Propagation Learning Algorithm Generalized Linear Models Radial Basis Function
More informationDiscussion of Some Problems About Nonlinear Time Series Prediction Using ν-support Vector Machine
Commun. Theor. Phys. (Beijing, China) 48 (2007) pp. 117 124 c International Academic Publishers Vol. 48, No. 1, July 15, 2007 Discussion of Some Problems About Nonlinear Time Series Prediction Using ν-support
More informationFormulation with slack variables
Formulation with slack variables Optimal margin classifier with slack variables and kernel functions described by Support Vector Machine (SVM). min (w,ξ) ½ w 2 + γσξ(i) subject to ξ(i) 0 i, d(i) (w T x(i)
More informationEmpirical Interpolation of Nonlinear Parametrized Evolution Operators
MÜNSTER Empirical Interpolation of Nonlinear Parametrized Evolution Operators Martin Drohmann, Bernard Haasdonk, Mario Ohlberger 03/12/2010 MÜNSTER 2/20 > Motivation: Reduced Basis Method RB Scenario:
More informationSupport Vector Machines
Support Vector Machines Le Song Machine Learning I CSE 6740, Fall 2013 Naïve Bayes classifier Still use Bayes decision rule for classification P y x = P x y P y P x But assume p x y = 1 is fully factorized
More informationNeural Networks and Deep Learning
Neural Networks and Deep Learning Professor Ameet Talwalkar November 12, 2015 Professor Ameet Talwalkar Neural Networks and Deep Learning November 12, 2015 1 / 16 Outline 1 Review of last lecture AdaBoost
More information