Generalized Sobol indices for dependent variables
|
|
- Gary Cole
- 5 years ago
- Views:
Transcription
1 Generalized Sobol indices for dependent variables Gaelle Chastaing Fabrice Gamboa Clémentine Prieur July 1st, /28 Gaelle Chastaing Sensitivity analysis and dependent variables
2 Contents 1 Context and motivation 2 Generalized functional decomposition 3 Generalized sensitivity indices 4 Procedure of estimation 5 Numerical example 2/28 Gaelle Chastaing Sensitivity analysis and dependent variables
3 1 Context and motivation Contents 2 Generalized functional decomposition Construction Illustration 3 Generalized sensitivity indices 4 Procedure of estimation Construction of the approximation spaces Least-squared estimation Estimation of sensitivity indices 5 Numerical example 3/28 Gaelle Chastaing Sensitivity analysis and dependent variables
4 Mathematical framework Disturbance Input variables X = (X 1,,X p ) Model η(x) Output Y Sensitivity analysis X is a random vector of distribution P X, not necessarily independent, η L 2 R (Rp ) i.e., for all h 1,h 2 L 2 R (Rp ), h 1,h 2 = h 1 (x)h 2 (x)dp X (x) = E[h 1 (X)h 2 (X)] R p Y is in value in R, and V (Y ) 0. 4/28 Gaelle Chastaing Sensitivity analysis and dependent variables
5 Use of functional decomposition When X independent, the Hoeffding decomposition gives p Y = η(x) = η + η i (X i ) + + η 1,,p (X) (1) i=1 (1) exists and is unique iff E[η u (X u )η v (X v )] = 0 u, v {1,, p}, u v with X u = (X u1,, X ut ) if u = (u 1,, u t ) {1,, p} 5/28 Gaelle Chastaing Sensitivity analysis and dependent variables
6 Moreover η = E(Y ) When X independent, Then, The Sobol index V (Y ) = i V [η i (X i )]+ i<j V [η ij (X i, X j )]+ +V [η 1,,p (X)] The Sobol index of a group of variables Xu is S u = V (η u) V (Y ) η i = E(Y X i ) η η ij = E(Y X i, X j ) η i η j η S u = V [E(Y X u)] v u ( 1) u v V [E(Y X v )] V (Y ) 6/28 Gaelle Chastaing Sensitivity analysis and dependent variables
7 Functional decomposition When X no more independent, The usual approach [XG08, LRY + 10, Can12] is Y u {1,,p} η u (X u ) + ε, max u = d 2, 3 Use of surrogate models to approximate each components (splines, polynomial chaos,...); ANOVA decomposition V (Y ) u d Sensitivity index [ V (η u (X u )) + v ucov(η ] u (X u ), η v (X v )) S u = V (η u(x u )) + v u Cov(η u(x u ), η v (X v )) V (Y ) Degree of implication of meta models? 7/28 Gaelle Chastaing Sensitivity analysis and dependent variables
8 1 Context and motivation Contents 2 Generalized functional decomposition Construction Illustration 3 Generalized sensitivity indices 4 Procedure of estimation Construction of the approximation spaces Least-squared estimation Estimation of sensitivity indices 5 Numerical example 8/28 Gaelle Chastaing Sensitivity analysis and dependent variables
9 Generalized functional decomposition Our objective : Propose a decomposition for the initial model η ; Ensure the existence and uniqueness of such decomposition Context Y = η(x), X P X p X the density function of X p Xu et p X u marginal densities of X u and X u = X \ X u 9/28 Gaelle Chastaing Sensitivity analysis and dependent variables
10 Generalized functional decomposition Condition 0 < M 1, u {1,..,p}, p X M p Xu p X u (C.1) Theorem Under (C.1), we have η(x) = η + i η i (X i ) + i<j η ij (X i,x j ) + + η 1,,p (X) Moreover, this decomposition is unique iff η u Hu, 0 with Hu 0 = { h u (X u ), E[h u (X u )h v (X v )] = 0, v u,h v Hv 0 } 0/28 Gaelle Chastaing Sensitivity analysis and dependent variables
11 Illustration with p = 3 inputs Functional decomposition η(x) = η + η 1 (X 1 ) + η 2 (X 2 ) + η 3 (X 3 ) + η 12 (X 1,X 2 ) +η 13 (X 1,X 3 ) + η 23 (X 2,X 3 ) + η 123 (X) Orthogonality relation representation for P X = P X1 P Xp (Hoeffding decomposition) For general P X (Generalized decomposition) 11/28 Gaelle Chastaing Sensitivity analysis and dependent variables
12 Illustration Hoeffding decomposition η 123 Generalized decomposition η 123 η 12 η 23 η 13 η 12 η 23 η 13 η 1 η 2 η 3 η 1 η 2 η 3 η η 12/28 Gaelle Chastaing Sensitivity analysis and dependent variables
13 1 Context and motivation Contents 2 Generalized functional decomposition Construction Illustration 3 Generalized sensitivity indices 4 Procedure of estimation Construction of the approximation spaces Least-squared estimation Estimation of sensitivity indices 5 Numerical example 13/28 Gaelle Chastaing Sensitivity analysis and dependent variables
14 Generalized Sobol sensitivity indices We have Y = η(x) = u η u(x u ), E[η u (X u )η v (X v )] = 0 whether u v or v u As a consequence, V (Y ) = u [V (η u ) + u v u,v Sensitivity index for the group X u is Cov(η u,η v )] S u = V (η u(x u )) + u v u,v Cov(η u(x u ), η v (X v )) V (Y ) and S u = 1 u 14/28 Gaelle Chastaing Sensitivity analysis and dependent variables
15 1 Context and motivation Contents 2 Generalized functional decomposition Construction Illustration 3 Generalized sensitivity indices 4 Procedure of estimation Construction of the approximation spaces Least-squared estimation Estimation of sensitivity indices 5 Numerical example 15/28 Gaelle Chastaing Sensitivity analysis and dependent variables
16 Procedure Remind that Y = u η u(x u ) with Idea : η u Hu 0 := {h u (X u ), E[h u (X u )h v (X v )] = 0, v u} Hu 0 Hv, 0 v u 1 Construct H 0,L u = Span { φ u 1,,φu L u } H 0 u, u; 2 Estimate components (η u ) u by the LSE (η u ) u Arg min η u H 0 u E[(Y u {1,,p} η u (X u )) 2 ]; 3 Estimate empirically the generalized sensitivity indices S u = V (η u(x u )) + u v u,v Cov(η u(x u ),η v (X v )). V (Y ) 6/28 Gaelle Chastaing Sensitivity analysis and dependent variables
17 Step 1 : Construction of the approximation spaces H 0 u = { h u (X u ), E[h u (X u )h v (X v )] = 0, v u,h v H 0 v}, u Definition of the finite approximation spaces H 0,L u use of truncated orthonormal basis (φ i l i ) L l i =0 of L 2 (R, B(R),P Xi ), i = 1,,p; inductive procedure on u ; exploitation of extended basis to reconstitute the hierarchical orthogonality. H 0 u, u by 17/28 Gaelle Chastaing Sensitivity analysis and dependent variables
18 Procedure 1 Initialization : for any i {1,,p}, construct Proceed as H 0,L i H 0 i := {h i (X i ), E[h i (X i )] = 0} choose a truncated orthonormal basis (φ i li ) L l of i=0 L 2 (R, B(R), P Xi ) with φ i 0 = 1. Then, Set Finally, H 0,L i E(φ i l i ) = 0, l i = 1,, L = Span { φ i 1,,φ i } L L i h i (X i ) = βl i i φ i l i (X i ), and E[h i (X i )] = 0 l i =1 8/28 Gaelle Chastaing Sensitivity analysis and dependent variables
19 Procedure(2) 2 For any i j {1,,p}, construct H 0,L ij Hij 0 := {h ij (X i,x j ), E[h ij h i ] = E[h ij h j ] = E[h ij ] = 0} Proceed recursively as set L L φ ij l ij (X i, X j ) = φ i l i φ j l j (X i, X j )+ λ i kφ i k(x i )+ λ j k φj k (X j)+c k=1 k=1 compute the coefficients (C, λ i 1,, λ i L, λj 1,, λj L ) by solving Set H 0,L ij = Span E[φ ij l ij φ i k ] = 0, k {1,, L} E[φ ij l ij φ j k ] = 0, k {1,, L} E[φ ij l ij ] = 0 { φ ij 1, φij L ij }, L ij = L L. 3 The same idea for any u = k 9/28 Gaelle Chastaing Sensitivity analysis and dependent variables (2)
20 Suppose that Construction in practice we get a i.i.d. n-sample (y l,x l ) l=1,,n define the empirical expected value E n ( ) as n L E n (g) = 1 n n g(x l ) The same procedure can be applied by replacing E( ) by E n ( ). Finally, H 0,L u,n = Span {ˆφu 1,n,, ˆφ } u L u,n, u l=1 0/28 Gaelle Chastaing Sensitivity analysis and dependent variables
21 We get Step 2 : Least-squared estimation η u (X u ) L u l u=1 β u l u ˆφu lu,n (X u), u Estimate components (η u ) from the minimization min β u lu 1 n n s=1[y l u L u l u=1 The total number of components (β u l u ) is pl + ( ) p L Unfeasible scheme for large p! β u l u ˆφu lu,n (xl u )]2 ( ) p L /28 Gaelle Chastaing Sensitivity analysis and dependent variables
22 Solution Order of ANOVA d p, i.e. Y η u (X u ) u d Use of a penalization to select a small number of (ˆφ u l u,n ), min β u lu 1 n n [y l s=1 L u u d l u=1 β u l u ˆφu lu,n (xl u)] 2 + λj(β 1 1,,β u l u, ) J(β) = β 0 := j 1(β j 0) by greedy algorithm ; Discrete process ; sparse selection from a dictionary ; J(β) = β 1 := j β j by LARS algorithm. Convex relaxation of the l 0 penalty; build up the regression function by successive steps; 2/28 Gaelle Chastaing Sensitivity analysis and dependent variables
23 Step 3 : Estimation of sensitivity indices We get ˆη u = L u l u=1 ˆβ u l u ˆφu lu,n, u, u d Empirical estimation of sensitivity indices S u with Ŝ u = V (ˆη u (X u )) + u v u,v Ĉov(ˆη u(x u ), ˆη v (X v )). V (Y ) Ĉov(ˆη u (X u ), ˆη v (X v )) = 1 n V (Y ) = 1 n n ˆη u (x l u)ˆη v (x l v), u,v l=1 n (y l ȳ) 2, ȳ = 1 n l=1 23/28 Gaelle Chastaing Sensitivity analysis and dependent variables n l=1 y l
24 1 Context and motivation Contents 2 Generalized functional decomposition Construction Illustration 3 Generalized sensitivity indices 4 Procedure of estimation Construction of the approximation spaces Least-squared estimation Estimation of sensitivity indices 5 Numerical example 24/28 Gaelle Chastaing Sensitivity analysis and dependent variables
25 River flood inundation Aim : prevent from inundation The maximal overflow is modelized by eight parameters Why randomness and uncertainty? spatio-temporal variability ; errors of measurements. 25/28 Gaelle Chastaing Sensitivity analysis and dependent variables
26 River flood inundation Test with n = 500 observations, d = 2; pairwise correlations ; Hermite basis of degree 10, i.e. P = 2880 parameters Greedy algorithm LARS algorithm Quantiles Quantiles Q K s Z v Z m C b H d L B Sensitivity indices Q K s Z v Z m C b H d L B 26/28 Gaelle Chastaing Sensitivity analysis and dependent variables Sensitivity indices
27 Thank you for your attention! 27/28 Gaelle Chastaing Sensitivity analysis and dependent variables
28 Y. Caniou. Analyse de sensibilité globale pour les modèles imbriqués et multiéchelles. PhD thesis, Université Blaise Pascal - Clermont II, G. Chastaing, F. Gamboa, and C. Prieur. Generalized hoeffding-sobol decomposition for dependent variables -Application to sensitivity analysis. Electronic Journal of Statistics, 6 : , G. Chastaing, F. Gamboa, and C. Prieur. Generalized sobol sensitivity indices for dependent variables : Numerical methods. Available at http ://arxiv.org/abs/ , G. Li, H. Rabitz, P.E. Yelvington, O. Oluwole, F. Bacon, Kolb C.E., and J. Schoendorf. Global sensitivity analysis with independent and/or correlated inputs. Journal of Physical Chemistry A, 114 : , I.M. Sobol. Sensitivity estimates for nonlinear mathematical models. Mathematical Modeling and Computational Experiment, 1(4) : , C. Xu and G. Gertner. Uncertainty and sensitivity analysis for models with correlated parameters. Reliability Engineering & System Safety, 93 : , /28 Gaelle Chastaing Sensitivity analysis and dependent variables
The Sobol decomposition
Introduction Generalized decomposition Generalized sensitivity indices Examples of distribution The Sobol decomposition Suppose x =(x1,,x p ) [0, 1] p η : [0, 1] p R adeterministicfunction ANOVA decomposition
More informationAnalysis of covariance (ANCOVA) using polynomial chaos expansions
Research Collection Conference Paper Analysis of covariance (ANCOVA) using polynomial chaos expansions Author(s): Sudret, Bruno; Caniou, Yves Publication Date: 013 Permanent Link: https://doi.org/10.399/ethz-a-0100633
More informationarxiv: v3 [math.st] 11 Mar 2012
Electronic Journal of Statistics Vol. 0 (0000) ISSN: 1935-754 DOI: 10.114/154957804100000000 arxiv:111.1788v3 [math.st] 11 Mar 01 Generalized Hoeffding-Sobol Decomposition for Dependent Variables - Application
More informationComputational methods for uncertainty quantification and sensitivity analysis of complex systems
DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Computational methods for uncertainty quantification and sensitivity analysis of complex systems
More informationPolynomial chaos expansions for sensitivity analysis
c DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Polynomial chaos expansions for sensitivity analysis B. Sudret Chair of Risk, Safety & Uncertainty
More informationPolynomial chaos expansions for structural reliability analysis
DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Polynomial chaos expansions for structural reliability analysis B. Sudret & S. Marelli Incl.
More informationarxiv: v1 [math.st] 14 Oct 2013
ANOVA decomposition of conditional Gaussian processes for sensitivity analysis with dependent inputs Gaelle Chastaing Loic Le Gratiet arxiv:1310.3578v1 [math.st] 14 Oct 2013 Abstract October 15, 2013 Complex
More informationGaussian process regression for Sensitivity analysis
Gaussian process regression for Sensitivity analysis GPSS Workshop on UQ, Sheffield, September 2016 Nicolas Durrande, Mines St-Étienne, durrande@emse.fr GPSS workshop on UQ GPs for sensitivity analysis
More informationModeling Real Estate Data using Quantile Regression
Modeling Real Estate Data using Semiparametric Quantile Regression Department of Statistics University of Innsbruck September 9th, 2011 Overview 1 Application: 2 3 4 Hedonic regression data for house prices
More informationSobol-Hoeffding Decomposition with Application to Global Sensitivity Analysis
Sobol-Hoeffding decomposition Application to Global SA Computation of the SI Sobol-Hoeffding Decomposition with Application to Global Sensitivity Analysis Olivier Le Maître with Colleague & Friend Omar
More informationVariance-based sensitivity indices for models with dependent inputs
Variance-based sensitivity indices for models with dependent inputs Thierry A. Mara, Stefano Tarantola To cite this version: Thierry A. Mara, Stefano Tarantola. Variance-based sensitivity indices for models
More informationLecture 14: Variable Selection - Beyond LASSO
Fall, 2017 Extension of LASSO To achieve oracle properties, L q penalty with 0 < q < 1, SCAD penalty (Fan and Li 2001; Zhang et al. 2007). Adaptive LASSO (Zou 2006; Zhang and Lu 2007; Wang et al. 2007)
More informationLocal and Global Sensitivity Analysis
Omar 1,2 1 Duke University Department of Mechanical Engineering & Materials Science omar.knio@duke.edu 2 KAUST Division of Computer, Electrical, Mathematical Science & Engineering omar.knio@kaust.edu.sa
More informationLectures. Variance-based sensitivity analysis in the presence of correlated input variables. Thomas Most. Source:
Lectures Variance-based sensitivity analysis in the presence of correlated input variables Thomas Most Source: www.dynardo.de/en/library Variance-based sensitivity analysis in the presence of correlated
More informationRegularization and Variable Selection via the Elastic Net
p. 1/1 Regularization and Variable Selection via the Elastic Net Hui Zou and Trevor Hastie Journal of Royal Statistical Society, B, 2005 Presenter: Minhua Chen, Nov. 07, 2008 p. 2/1 Agenda Introduction
More informationPenalized Squared Error and Likelihood: Risk Bounds and Fast Algorithms
university-logo Penalized Squared Error and Likelihood: Risk Bounds and Fast Algorithms Andrew Barron Cong Huang Xi Luo Department of Statistics Yale University 2008 Workshop on Sparsity in High Dimensional
More informationLinear Regression Linear Regression with Shrinkage
Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle
More informationStratified Random Sampling for Dependent Inputs
Stratified Random Sampling for Dependent Inputs Anirban Mondal Case Western Reserve University, Cleveland, OH 44106, USA arxiv:1904.00555v1 [stat.me] 1 Apr 2019 Abhijit Mandal Wayne State University, Detroit,
More informationEcon 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines
Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Maximilian Kasy Department of Economics, Harvard University 1 / 37 Agenda 6 equivalent representations of the
More informationInterpolation. 1. Judd, K. Numerical Methods in Economics, Cambridge: MIT Press. Chapter
Key References: Interpolation 1. Judd, K. Numerical Methods in Economics, Cambridge: MIT Press. Chapter 6. 2. Press, W. et. al. Numerical Recipes in C, Cambridge: Cambridge University Press. Chapter 3
More informationStochastic optimization - how to improve computational efficiency?
Stochastic optimization - how to improve computational efficiency? Christian Bucher Center of Mechanics and Structural Dynamics Vienna University of Technology & DYNARDO GmbH, Vienna Presentation at Czech
More informationFast learning rates for plug-in classifiers under the margin condition
Fast learning rates for plug-in classifiers under the margin condition Jean-Yves Audibert 1 Alexandre B. Tsybakov 2 1 Certis ParisTech - Ecole des Ponts, France 2 LPMA Université Pierre et Marie Curie,
More informationLinear Models Review
Linear Models Review Vectors in IR n will be written as ordered n-tuples which are understood to be column vectors, or n 1 matrices. A vector variable will be indicted with bold face, and the prime sign
More informationMultivariate Regression
Multivariate Regression The so-called supervised learning problem is the following: we want to approximate the random variable Y with an appropriate function of the random variables X 1,..., X p with the
More informationapproximation algorithms I
SUM-OF-SQUARES method and approximation algorithms I David Steurer Cornell Cargese Workshop, 201 meta-task encoded as low-degree polynomial in R x example: f(x) = i,j n w ij x i x j 2 given: functions
More informationDirect Learning: Linear Regression. Donglin Zeng, Department of Biostatistics, University of North Carolina
Direct Learning: Linear Regression Parametric learning We consider the core function in the prediction rule to be a parametric function. The most commonly used function is a linear function: squared loss:
More informationLecture 6: Linear models and Gauss-Markov theorem
Lecture 6: Linear models and Gauss-Markov theorem Linear model setting Results in simple linear regression can be extended to the following general linear model with independently observed response variables
More informationNew results for the functional ANOVA and. Sobol indices
Generalized Sobol Indices 1 New results for the functional ANOVA and Sobol indices Art B. Owen Stanford University Generalized Sobol Indices 2 Ilya Meerovich Sobol At MCM 2001, Salzburg Known for Sobol
More informationAdditive Isotonic Regression
Additive Isotonic Regression Enno Mammen and Kyusang Yu 11. July 2006 INTRODUCTION: We have i.i.d. random vectors (Y 1, X 1 ),..., (Y n, X n ) with X i = (X1 i,..., X d i ) and we consider the additive
More informationEE 381V: Large Scale Optimization Fall Lecture 24 April 11
EE 381V: Large Scale Optimization Fall 2012 Lecture 24 April 11 Lecturer: Caramanis & Sanghavi Scribe: Tao Huang 24.1 Review In past classes, we studied the problem of sparsity. Sparsity problem is that
More informationINTRODUCING LINEAR REGRESSION MODELS Response or Dependent variable y
INTRODUCING LINEAR REGRESSION MODELS Response or Dependent variable y Predictor or Independent variable x Model with error: for i = 1,..., n, y i = α + βx i + ε i ε i : independent errors (sampling, measurement,
More informationLinear regression methods
Linear regression methods Most of our intuition about statistical methods stem from linear regression. For observations i = 1,..., n, the model is Y i = p X ij β j + ε i, j=1 where Y i is the response
More informationMachine Learning. Part 1. Linear Regression. Machine Learning: Regression Case. .. Dennis Sun DATA 401 Data Science Alex Dekhtyar..
.. Dennis Sun DATA 401 Data Science Alex Dekhtyar.. Machine Learning. Part 1. Linear Regression Machine Learning: Regression Case. Dataset. Consider a collection of features X = {X 1,...,X n }, such that
More informationSTAT5044: Regression and Anova. Inyoung Kim
STAT5044: Regression and Anova Inyoung Kim 2 / 51 Outline 1 Matrix Expression 2 Linear and quadratic forms 3 Properties of quadratic form 4 Properties of estimates 5 Distributional properties 3 / 51 Matrix
More informationThe deterministic Lasso
The deterministic Lasso Sara van de Geer Seminar für Statistik, ETH Zürich Abstract We study high-dimensional generalized linear models and empirical risk minimization using the Lasso An oracle inequality
More informationIntroduction to Smoothing spline ANOVA models (metamodelling)
Introduction to Smoothing spline ANOVA models (metamodelling) M. Ratto DYNARE Summer School, Paris, June 215. Joint Research Centre www.jrc.ec.europa.eu Serving society Stimulating innovation Supporting
More informationQuantifying Stochastic Model Errors via Robust Optimization
Quantifying Stochastic Model Errors via Robust Optimization IPAM Workshop on Uncertainty Quantification for Multiscale Stochastic Systems and Applications Jan 19, 2016 Henry Lam Industrial & Operations
More informationEfficient Complex Output Prediction
Efficient Complex Output Prediction Florence d Alché-Buc Joint work with Romain Brault, Alex Lambert, Maxime Sangnier October 12, 2017 LTCI, Télécom ParisTech, Institut-Mines Télécom, Université Paris-Saclay
More informationNeural networks COMS 4771
Neural networks COMS 4771 1. Logistic regression Logistic regression Suppose X = R d and Y = {0, 1}. A logistic regression model is a statistical model where the conditional probability function has a
More informationA Unified Framework for Uncertainty and Sensitivity Analysis of Computational Models with Many Input Parameters
A Unified Framework for Uncertainty and Sensitivity Analysis of Computational Models with Many Input Parameters C. F. Jeff Wu H. Milton Stewart School of Industrial and Systems Engineering Georgia Institute
More informationLeast Squares Regression
E0 70 Machine Learning Lecture 4 Jan 7, 03) Least Squares Regression Lecturer: Shivani Agarwal Disclaimer: These notes are a brief summary of the topics covered in the lecture. They are not a substitute
More informationNonparametric Regression. Badr Missaoui
Badr Missaoui Outline Kernel and local polynomial regression. Penalized regression. We are given n pairs of observations (X 1, Y 1 ),...,(X n, Y n ) where Y i = r(x i ) + ε i, i = 1,..., n and r(x) = E(Y
More informationMotivation Sparse Signal Recovery is an interesting area with many potential applications. Methods developed for solving sparse signal recovery proble
Bayesian Methods for Sparse Signal Recovery Bhaskar D Rao 1 University of California, San Diego 1 Thanks to David Wipf, Zhilin Zhang and Ritwik Giri Motivation Sparse Signal Recovery is an interesting
More informationMachine Learning. Regression Case Part 1. Linear Regression. Machine Learning: Regression Case.
.. Cal Poly CSC 566 Advanced Data Mining Alexander Dekhtyar.. Machine Learning. Regression Case Part 1. Linear Regression Machine Learning: Regression Case. Dataset. Consider a collection of features X
More informationLinear regression COMS 4771
Linear regression COMS 4771 1. Old Faithful and prediction functions Prediction problem: Old Faithful geyser (Yellowstone) Task: Predict time of next eruption. 1 / 40 Statistical model for time between
More information11. Learning graphical models
Learning graphical models 11-1 11. Learning graphical models Maximum likelihood Parameter learning Structural learning Learning partially observed graphical models Learning graphical models 11-2 statistical
More informationLeast Squares Regression
CIS 50: Machine Learning Spring 08: Lecture 4 Least Squares Regression Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may not cover all the
More informationLinear Regression Linear Regression with Shrinkage
Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle
More informationSparse polynomial chaos expansions in engineering applications
DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Sparse polynomial chaos expansions in engineering applications B. Sudret G. Blatman (EDF R&D,
More informationFunctional Estimation in Systems Defined by Differential Equation using Bayesian Smoothing Methods
Université Catholique de Louvain Institut de Statistique, Biostatistique et Sciences Actuarielles Functional Estimation in Systems Defined by Differential Equation using Bayesian Smoothing Methods 19th
More informationLinear Regression. Aarti Singh. Machine Learning / Sept 27, 2010
Linear Regression Aarti Singh Machine Learning 10-701/15-781 Sept 27, 2010 Discrete to Continuous Labels Classification Sports Science News Anemic cell Healthy cell Regression X = Document Y = Topic X
More informationChapter 3. Linear Models for Regression
Chapter 3. Linear Models for Regression Wei Pan Division of Biostatistics, School of Public Health, University of Minnesota, Minneapolis, MN 55455 Email: weip@biostat.umn.edu PubH 7475/8475 c Wei Pan Linear
More informationLecture 3: Statistical Decision Theory (Part II)
Lecture 3: Statistical Decision Theory (Part II) Hao Helen Zhang Hao Helen Zhang Lecture 3: Statistical Decision Theory (Part II) 1 / 27 Outline of This Note Part I: Statistics Decision Theory (Classical
More informationParcimonie en apprentissage statistique
Parcimonie en apprentissage statistique Guillaume Obozinski Ecole des Ponts - ParisTech Journée Parcimonie Fédération Charles Hermite, 23 Juin 2014 Parcimonie en apprentissage 1/44 Classical supervised
More informationSupport Vector Machine
Support Vector Machine Fabrice Rossi SAMM Université Paris 1 Panthéon Sorbonne 2018 Outline Linear Support Vector Machine Kernelized SVM Kernels 2 From ERM to RLM Empirical Risk Minimization in the binary
More informationLinear Models in Machine Learning
CS540 Intro to AI Linear Models in Machine Learning Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu We briefly go over two linear models frequently used in machine learning: linear regression for, well, regression,
More informationECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference
ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Sparse Recovery using L1 minimization - algorithms Yuejie Chi Department of Electrical and Computer Engineering Spring
More informationConvex relaxation for Combinatorial Penalties
Convex relaxation for Combinatorial Penalties Guillaume Obozinski Equipe Imagine Laboratoire d Informatique Gaspard Monge Ecole des Ponts - ParisTech Joint work with Francis Bach Fête Parisienne in Computation,
More informationSVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels
SVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels Karl Stratos June 21, 2018 1 / 33 Tangent: Some Loose Ends in Logistic Regression Polynomial feature expansion in logistic
More informationModel Selection and Geometry
Model Selection and Geometry Pascal Massart Université Paris-Sud, Orsay Leipzig, February Purpose of the talk! Concentration of measure plays a fundamental role in the theory of model selection! Model
More informationLecture 18: Kernels Risk and Loss Support Vector Regression. Aykut Erdem December 2016 Hacettepe University
Lecture 18: Kernels Risk and Loss Support Vector Regression Aykut Erdem December 2016 Hacettepe University Administrative We will have a make-up lecture on next Saturday December 24, 2016 Presentations
More informationWell-developed and understood properties
1 INTRODUCTION TO LINEAR MODELS 1 THE CLASSICAL LINEAR MODEL Most commonly used statistical models Flexible models Well-developed and understood properties Ease of interpretation Building block for more
More information1 Probability Model. 1.1 Types of models to be discussed in the course
Sufficiency January 11, 2016 Debdeep Pati 1 Probability Model Model: A family of distributions {P θ : θ Θ}. P θ (B) is the probability of the event B when the parameter takes the value θ. P θ is described
More informationOslo Class 6 Sparsity based regularization
RegML2017@SIMULA Oslo Class 6 Sparsity based regularization Lorenzo Rosasco UNIGE-MIT-IIT May 4, 2017 Learning from data Possible only under assumptions regularization min Ê(w) + λr(w) w Smoothness Sparsity
More informationChapter 2. Section Section 2.9. J. Kim (ISU) Chapter 2 1 / 26. Design-optimal estimator under stratified random sampling
Chapter 2 Section 2.4 - Section 2.9 J. Kim (ISU) Chapter 2 1 / 26 2.4 Regression and stratification Design-optimal estimator under stratified random sampling where (Ŝxxh, Ŝxyh) ˆβ opt = ( x st, ȳ st )
More informationRegression Shrinkage and Selection via the Lasso
Regression Shrinkage and Selection via the Lasso ROBERT TIBSHIRANI, 1996 Presenter: Guiyun Feng April 27 () 1 / 20 Motivation Estimation in Linear Models: y = β T x + ɛ. data (x i, y i ), i = 1, 2,...,
More informationEffect of outliers on the variable selection by the regularized regression
Communications for Statistical Applications and Methods 2018, Vol. 25, No. 2, 235 243 https://doi.org/10.29220/csam.2018.25.2.235 Print ISSN 2287-7843 / Online ISSN 2383-4757 Effect of outliers on the
More informationMessage passing and approximate message passing
Message passing and approximate message passing Arian Maleki Columbia University 1 / 47 What is the problem? Given pdf µ(x 1, x 2,..., x n ) we are interested in arg maxx1,x 2,...,x n µ(x 1, x 2,..., x
More information(Part 1) High-dimensional statistics May / 41
Theory for the Lasso Recall the linear model Y i = p j=1 β j X (j) i + ɛ i, i = 1,..., n, or, in matrix notation, Y = Xβ + ɛ, To simplify, we assume that the design X is fixed, and that ɛ is N (0, σ 2
More informationUnsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto
Unsupervised Learning Techniques 9.520 Class 07, 1 March 2006 Andrea Caponnetto About this class Goal To introduce some methods for unsupervised learning: Gaussian Mixtures, K-Means, ISOMAP, HLLE, Laplacian
More informationNow consider the case where E(Y) = µ = Xβ and V (Y) = σ 2 G, where G is diagonal, but unknown.
Weighting We have seen that if E(Y) = Xβ and V (Y) = σ 2 G, where G is known, the model can be rewritten as a linear model. This is known as generalized least squares or, if G is diagonal, with trace(g)
More informationAddressing high dimensionality in reliability analysis using low-rank tensor approximations
Addressing high dimensionality in reliability analysis using low-rank tensor approximations K Konakli, Bruno Sudret To cite this version: K Konakli, Bruno Sudret. Addressing high dimensionality in reliability
More informationvariability of the model, represented by σ 2 and not accounted for by Xβ
Posterior Predictive Distribution Suppose we have observed a new set of explanatory variables X and we want to predict the outcomes ỹ using the regression model. Components of uncertainty in p(ỹ y) variability
More informationA new estimator for quantile-oriented sensitivity indices
A new estimator for quantile-oriented sensitivity indices Thomas Browne Supervisors : J-C. Fort (Paris 5) & T. Klein (IMT-Toulouse) Advisors : B. Iooss & L. Le Gratiet (EDF R&D-MRI Chatou) EDF R&D-MRI
More informationA regularized least-squares method for sparse low-rank approximation of multivariate functions
Workshop Numerical methods for high-dimensional problems April 18, 2014 A regularized least-squares method for sparse low-rank approximation of multivariate functions Mathilde Chevreuil joint work with
More informationSENSITIVITY ANALYSIS IN NUMERICAL SIMULATION OF MULTIPHASE FLOW FOR CO 2 STORAGE IN SALINE AQUIFERS USING THE PROBABILISTIC COLLOCATION APPROACH
XIX International Conference on Water Resources CMWR 2012 University of Illinois at Urbana-Champaign June 17-22,2012 SENSITIVITY ANALYSIS IN NUMERICAL SIMULATION OF MULTIPHASE FLOW FOR CO 2 STORAGE IN
More informationMachine Learning CSE546 Carlos Guestrin University of Washington. October 7, Efficiency: If size(w) = 100B, each prediction is expensive:
Simple Variable Selection LASSO: Sparse Regression Machine Learning CSE546 Carlos Guestrin University of Washington October 7, 2013 1 Sparsity Vector w is sparse, if many entries are zero: Very useful
More informationAdaptive Piecewise Polynomial Estimation via Trend Filtering
Adaptive Piecewise Polynomial Estimation via Trend Filtering Liubo Li, ShanShan Tu The Ohio State University li.2201@osu.edu, tu.162@osu.edu October 1, 2015 Liubo Li, ShanShan Tu (OSU) Trend Filtering
More informationClassification with Reject Option
Classification with Reject Option Bartlett and Wegkamp (2008) Wegkamp and Yuan (2010) February 17, 2012 Outline. Introduction.. Classification with reject option. Spirit of the papers BW2008.. Infinite
More informationThe Adaptive Lasso and Its Oracle Properties Hui Zou (2006), JASA
The Adaptive Lasso and Its Oracle Properties Hui Zou (2006), JASA Presented by Dongjun Chung March 12, 2010 Introduction Definition Oracle Properties Computations Relationship: Nonnegative Garrote Extensions:
More informationScientific Computing: An Introductory Survey
Scientific Computing: An Introductory Survey Chapter 7 Interpolation Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction permitted
More informationStatistics II. Management Degree Management Statistics IIDegree. Statistics II. 2 nd Sem. 2013/2014. Management Degree. Simple Linear Regression
Model 1 2 Ordinary Least Squares 3 4 Non-linearities 5 of the coefficients and their to the model We saw that econometrics studies E (Y x). More generally, we shall study regression analysis. : The regression
More informationApproximate Kernel PCA with Random Features
Approximate Kernel PCA with Random Features (Computational vs. Statistical Tradeoff) Bharath K. Sriperumbudur Department of Statistics, Pennsylvania State University Journées de Statistique Paris May 28,
More informationBINARY CLASSIFICATION
BINARY CLASSIFICATION MAXIM RAGINSY The problem of binary classification can be stated as follows. We have a random couple Z = X, Y ), where X R d is called the feature vector and Y {, } is called the
More informationCOMS 4721: Machine Learning for Data Science Lecture 6, 2/2/2017
COMS 4721: Machine Learning for Data Science Lecture 6, 2/2/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University UNDERDETERMINED LINEAR EQUATIONS We
More informationCurve learning. p.1/35
Curve learning Gérard Biau UNIVERSITÉ MONTPELLIER II p.1/35 Summary The problem The mathematical model Functional classification 1. Fourier filtering 2. Wavelet filtering Applications p.2/35 The problem
More informationrecent developments of approximation theory and greedy algorithms
recent developments of approximation theory and greedy algorithms Peter Binev Department of Mathematics and Interdisciplinary Mathematics Institute University of South Carolina Reduced Order Modeling in
More informationA Bootstrap Lasso + Partial Ridge Method to Construct Confidence Intervals for Parameters in High-dimensional Sparse Linear Models
A Bootstrap Lasso + Partial Ridge Method to Construct Confidence Intervals for Parameters in High-dimensional Sparse Linear Models Jingyi Jessica Li Department of Statistics University of California, Los
More informationMinimizing the Number of Function Evaluations to Estimate Sobol Indices Using Quasi-Monte Carlo
Minimizing the Number of Function Evaluations to Estimate Sobol Indices Using Quasi-Monte Carlo Lluís Antoni Jiménez Rugama Joint work with: Fred J Hickernell (IIT), Clémentine Prieur (Univ Grenoble),
More informationChap. 3. Controlled Systems, Controllability
Chap. 3. Controlled Systems, Controllability 1. Controllability of Linear Systems 1.1. Kalman s Criterion Consider the linear system ẋ = Ax + Bu where x R n : state vector and u R m : input vector. A :
More informationRecap from previous lecture
Recap from previous lecture Learning is using past experience to improve future performance. Different types of learning: supervised unsupervised reinforcement active online... For a machine, experience
More informationMA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1
MA 575 Linear Models: Cedric E Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1 1 Within-group Correlation Let us recall the simple two-level hierarchical
More informationSTAT5044: Regression and Anova
STAT5044: Regression and Anova Inyoung Kim 1 / 25 Outline 1 Multiple Linear Regression 2 / 25 Basic Idea An extra sum of squares: the marginal reduction in the error sum of squares when one or several
More informationIndirect Rule Learning: Support Vector Machines. Donglin Zeng, Department of Biostatistics, University of North Carolina
Indirect Rule Learning: Support Vector Machines Indirect learning: loss optimization It doesn t estimate the prediction rule f (x) directly, since most loss functions do not have explicit optimizers. Indirection
More informationEE613 Machine Learning for Engineers. Kernel methods Support Vector Machines. jean-marc odobez 2015
EE613 Machine Learning for Engineers Kernel methods Support Vector Machines jean-marc odobez 2015 overview Kernel methods introductions and main elements defining kernels Kernelization of k-nn, K-Means,
More informationModel Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao
Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics Jiti Gao Department of Statistics School of Mathematics and Statistics The University of Western Australia Crawley
More informationA Least Squares Formulation for Canonical Correlation Analysis
A Least Squares Formulation for Canonical Correlation Analysis Liang Sun, Shuiwang Ji, and Jieping Ye Department of Computer Science and Engineering Arizona State University Motivation Canonical Correlation
More informationBayesian variable selection via. Penalized credible regions. Brian Reich, NCSU. Joint work with. Howard Bondell and Ander Wilson
Bayesian variable selection via penalized credible regions Brian Reich, NC State Joint work with Howard Bondell and Ander Wilson Brian Reich, NCSU Penalized credible regions 1 Motivation big p, small n
More informationSpectral Regularization
Spectral Regularization Lorenzo Rosasco 9.520 Class 07 February 27, 2008 About this class Goal To discuss how a class of regularization methods originally designed for solving ill-posed inverse problems,
More informationUncertainty Quantification in Computational Science
DTU 2010 - Lecture I Uncertainty Quantification in Computational Science Jan S Hesthaven Brown University Jan.Hesthaven@Brown.edu Objective of lectures The main objective of these lectures are To offer
More information