Generalized Sobol indices for dependent variables

Size: px
Start display at page:

Download "Generalized Sobol indices for dependent variables"

Transcription

1 Generalized Sobol indices for dependent variables Gaelle Chastaing Fabrice Gamboa Clémentine Prieur July 1st, /28 Gaelle Chastaing Sensitivity analysis and dependent variables

2 Contents 1 Context and motivation 2 Generalized functional decomposition 3 Generalized sensitivity indices 4 Procedure of estimation 5 Numerical example 2/28 Gaelle Chastaing Sensitivity analysis and dependent variables

3 1 Context and motivation Contents 2 Generalized functional decomposition Construction Illustration 3 Generalized sensitivity indices 4 Procedure of estimation Construction of the approximation spaces Least-squared estimation Estimation of sensitivity indices 5 Numerical example 3/28 Gaelle Chastaing Sensitivity analysis and dependent variables

4 Mathematical framework Disturbance Input variables X = (X 1,,X p ) Model η(x) Output Y Sensitivity analysis X is a random vector of distribution P X, not necessarily independent, η L 2 R (Rp ) i.e., for all h 1,h 2 L 2 R (Rp ), h 1,h 2 = h 1 (x)h 2 (x)dp X (x) = E[h 1 (X)h 2 (X)] R p Y is in value in R, and V (Y ) 0. 4/28 Gaelle Chastaing Sensitivity analysis and dependent variables

5 Use of functional decomposition When X independent, the Hoeffding decomposition gives p Y = η(x) = η + η i (X i ) + + η 1,,p (X) (1) i=1 (1) exists and is unique iff E[η u (X u )η v (X v )] = 0 u, v {1,, p}, u v with X u = (X u1,, X ut ) if u = (u 1,, u t ) {1,, p} 5/28 Gaelle Chastaing Sensitivity analysis and dependent variables

6 Moreover η = E(Y ) When X independent, Then, The Sobol index V (Y ) = i V [η i (X i )]+ i<j V [η ij (X i, X j )]+ +V [η 1,,p (X)] The Sobol index of a group of variables Xu is S u = V (η u) V (Y ) η i = E(Y X i ) η η ij = E(Y X i, X j ) η i η j η S u = V [E(Y X u)] v u ( 1) u v V [E(Y X v )] V (Y ) 6/28 Gaelle Chastaing Sensitivity analysis and dependent variables

7 Functional decomposition When X no more independent, The usual approach [XG08, LRY + 10, Can12] is Y u {1,,p} η u (X u ) + ε, max u = d 2, 3 Use of surrogate models to approximate each components (splines, polynomial chaos,...); ANOVA decomposition V (Y ) u d Sensitivity index [ V (η u (X u )) + v ucov(η ] u (X u ), η v (X v )) S u = V (η u(x u )) + v u Cov(η u(x u ), η v (X v )) V (Y ) Degree of implication of meta models? 7/28 Gaelle Chastaing Sensitivity analysis and dependent variables

8 1 Context and motivation Contents 2 Generalized functional decomposition Construction Illustration 3 Generalized sensitivity indices 4 Procedure of estimation Construction of the approximation spaces Least-squared estimation Estimation of sensitivity indices 5 Numerical example 8/28 Gaelle Chastaing Sensitivity analysis and dependent variables

9 Generalized functional decomposition Our objective : Propose a decomposition for the initial model η ; Ensure the existence and uniqueness of such decomposition Context Y = η(x), X P X p X the density function of X p Xu et p X u marginal densities of X u and X u = X \ X u 9/28 Gaelle Chastaing Sensitivity analysis and dependent variables

10 Generalized functional decomposition Condition 0 < M 1, u {1,..,p}, p X M p Xu p X u (C.1) Theorem Under (C.1), we have η(x) = η + i η i (X i ) + i<j η ij (X i,x j ) + + η 1,,p (X) Moreover, this decomposition is unique iff η u Hu, 0 with Hu 0 = { h u (X u ), E[h u (X u )h v (X v )] = 0, v u,h v Hv 0 } 0/28 Gaelle Chastaing Sensitivity analysis and dependent variables

11 Illustration with p = 3 inputs Functional decomposition η(x) = η + η 1 (X 1 ) + η 2 (X 2 ) + η 3 (X 3 ) + η 12 (X 1,X 2 ) +η 13 (X 1,X 3 ) + η 23 (X 2,X 3 ) + η 123 (X) Orthogonality relation representation for P X = P X1 P Xp (Hoeffding decomposition) For general P X (Generalized decomposition) 11/28 Gaelle Chastaing Sensitivity analysis and dependent variables

12 Illustration Hoeffding decomposition η 123 Generalized decomposition η 123 η 12 η 23 η 13 η 12 η 23 η 13 η 1 η 2 η 3 η 1 η 2 η 3 η η 12/28 Gaelle Chastaing Sensitivity analysis and dependent variables

13 1 Context and motivation Contents 2 Generalized functional decomposition Construction Illustration 3 Generalized sensitivity indices 4 Procedure of estimation Construction of the approximation spaces Least-squared estimation Estimation of sensitivity indices 5 Numerical example 13/28 Gaelle Chastaing Sensitivity analysis and dependent variables

14 Generalized Sobol sensitivity indices We have Y = η(x) = u η u(x u ), E[η u (X u )η v (X v )] = 0 whether u v or v u As a consequence, V (Y ) = u [V (η u ) + u v u,v Sensitivity index for the group X u is Cov(η u,η v )] S u = V (η u(x u )) + u v u,v Cov(η u(x u ), η v (X v )) V (Y ) and S u = 1 u 14/28 Gaelle Chastaing Sensitivity analysis and dependent variables

15 1 Context and motivation Contents 2 Generalized functional decomposition Construction Illustration 3 Generalized sensitivity indices 4 Procedure of estimation Construction of the approximation spaces Least-squared estimation Estimation of sensitivity indices 5 Numerical example 15/28 Gaelle Chastaing Sensitivity analysis and dependent variables

16 Procedure Remind that Y = u η u(x u ) with Idea : η u Hu 0 := {h u (X u ), E[h u (X u )h v (X v )] = 0, v u} Hu 0 Hv, 0 v u 1 Construct H 0,L u = Span { φ u 1,,φu L u } H 0 u, u; 2 Estimate components (η u ) u by the LSE (η u ) u Arg min η u H 0 u E[(Y u {1,,p} η u (X u )) 2 ]; 3 Estimate empirically the generalized sensitivity indices S u = V (η u(x u )) + u v u,v Cov(η u(x u ),η v (X v )). V (Y ) 6/28 Gaelle Chastaing Sensitivity analysis and dependent variables

17 Step 1 : Construction of the approximation spaces H 0 u = { h u (X u ), E[h u (X u )h v (X v )] = 0, v u,h v H 0 v}, u Definition of the finite approximation spaces H 0,L u use of truncated orthonormal basis (φ i l i ) L l i =0 of L 2 (R, B(R),P Xi ), i = 1,,p; inductive procedure on u ; exploitation of extended basis to reconstitute the hierarchical orthogonality. H 0 u, u by 17/28 Gaelle Chastaing Sensitivity analysis and dependent variables

18 Procedure 1 Initialization : for any i {1,,p}, construct Proceed as H 0,L i H 0 i := {h i (X i ), E[h i (X i )] = 0} choose a truncated orthonormal basis (φ i li ) L l of i=0 L 2 (R, B(R), P Xi ) with φ i 0 = 1. Then, Set Finally, H 0,L i E(φ i l i ) = 0, l i = 1,, L = Span { φ i 1,,φ i } L L i h i (X i ) = βl i i φ i l i (X i ), and E[h i (X i )] = 0 l i =1 8/28 Gaelle Chastaing Sensitivity analysis and dependent variables

19 Procedure(2) 2 For any i j {1,,p}, construct H 0,L ij Hij 0 := {h ij (X i,x j ), E[h ij h i ] = E[h ij h j ] = E[h ij ] = 0} Proceed recursively as set L L φ ij l ij (X i, X j ) = φ i l i φ j l j (X i, X j )+ λ i kφ i k(x i )+ λ j k φj k (X j)+c k=1 k=1 compute the coefficients (C, λ i 1,, λ i L, λj 1,, λj L ) by solving Set H 0,L ij = Span E[φ ij l ij φ i k ] = 0, k {1,, L} E[φ ij l ij φ j k ] = 0, k {1,, L} E[φ ij l ij ] = 0 { φ ij 1, φij L ij }, L ij = L L. 3 The same idea for any u = k 9/28 Gaelle Chastaing Sensitivity analysis and dependent variables (2)

20 Suppose that Construction in practice we get a i.i.d. n-sample (y l,x l ) l=1,,n define the empirical expected value E n ( ) as n L E n (g) = 1 n n g(x l ) The same procedure can be applied by replacing E( ) by E n ( ). Finally, H 0,L u,n = Span {ˆφu 1,n,, ˆφ } u L u,n, u l=1 0/28 Gaelle Chastaing Sensitivity analysis and dependent variables

21 We get Step 2 : Least-squared estimation η u (X u ) L u l u=1 β u l u ˆφu lu,n (X u), u Estimate components (η u ) from the minimization min β u lu 1 n n s=1[y l u L u l u=1 The total number of components (β u l u ) is pl + ( ) p L Unfeasible scheme for large p! β u l u ˆφu lu,n (xl u )]2 ( ) p L /28 Gaelle Chastaing Sensitivity analysis and dependent variables

22 Solution Order of ANOVA d p, i.e. Y η u (X u ) u d Use of a penalization to select a small number of (ˆφ u l u,n ), min β u lu 1 n n [y l s=1 L u u d l u=1 β u l u ˆφu lu,n (xl u)] 2 + λj(β 1 1,,β u l u, ) J(β) = β 0 := j 1(β j 0) by greedy algorithm ; Discrete process ; sparse selection from a dictionary ; J(β) = β 1 := j β j by LARS algorithm. Convex relaxation of the l 0 penalty; build up the regression function by successive steps; 2/28 Gaelle Chastaing Sensitivity analysis and dependent variables

23 Step 3 : Estimation of sensitivity indices We get ˆη u = L u l u=1 ˆβ u l u ˆφu lu,n, u, u d Empirical estimation of sensitivity indices S u with Ŝ u = V (ˆη u (X u )) + u v u,v Ĉov(ˆη u(x u ), ˆη v (X v )). V (Y ) Ĉov(ˆη u (X u ), ˆη v (X v )) = 1 n V (Y ) = 1 n n ˆη u (x l u)ˆη v (x l v), u,v l=1 n (y l ȳ) 2, ȳ = 1 n l=1 23/28 Gaelle Chastaing Sensitivity analysis and dependent variables n l=1 y l

24 1 Context and motivation Contents 2 Generalized functional decomposition Construction Illustration 3 Generalized sensitivity indices 4 Procedure of estimation Construction of the approximation spaces Least-squared estimation Estimation of sensitivity indices 5 Numerical example 24/28 Gaelle Chastaing Sensitivity analysis and dependent variables

25 River flood inundation Aim : prevent from inundation The maximal overflow is modelized by eight parameters Why randomness and uncertainty? spatio-temporal variability ; errors of measurements. 25/28 Gaelle Chastaing Sensitivity analysis and dependent variables

26 River flood inundation Test with n = 500 observations, d = 2; pairwise correlations ; Hermite basis of degree 10, i.e. P = 2880 parameters Greedy algorithm LARS algorithm Quantiles Quantiles Q K s Z v Z m C b H d L B Sensitivity indices Q K s Z v Z m C b H d L B 26/28 Gaelle Chastaing Sensitivity analysis and dependent variables Sensitivity indices

27 Thank you for your attention! 27/28 Gaelle Chastaing Sensitivity analysis and dependent variables

28 Y. Caniou. Analyse de sensibilité globale pour les modèles imbriqués et multiéchelles. PhD thesis, Université Blaise Pascal - Clermont II, G. Chastaing, F. Gamboa, and C. Prieur. Generalized hoeffding-sobol decomposition for dependent variables -Application to sensitivity analysis. Electronic Journal of Statistics, 6 : , G. Chastaing, F. Gamboa, and C. Prieur. Generalized sobol sensitivity indices for dependent variables : Numerical methods. Available at http ://arxiv.org/abs/ , G. Li, H. Rabitz, P.E. Yelvington, O. Oluwole, F. Bacon, Kolb C.E., and J. Schoendorf. Global sensitivity analysis with independent and/or correlated inputs. Journal of Physical Chemistry A, 114 : , I.M. Sobol. Sensitivity estimates for nonlinear mathematical models. Mathematical Modeling and Computational Experiment, 1(4) : , C. Xu and G. Gertner. Uncertainty and sensitivity analysis for models with correlated parameters. Reliability Engineering & System Safety, 93 : , /28 Gaelle Chastaing Sensitivity analysis and dependent variables

The Sobol decomposition

The Sobol decomposition Introduction Generalized decomposition Generalized sensitivity indices Examples of distribution The Sobol decomposition Suppose x =(x1,,x p ) [0, 1] p η : [0, 1] p R adeterministicfunction ANOVA decomposition

More information

Analysis of covariance (ANCOVA) using polynomial chaos expansions

Analysis of covariance (ANCOVA) using polynomial chaos expansions Research Collection Conference Paper Analysis of covariance (ANCOVA) using polynomial chaos expansions Author(s): Sudret, Bruno; Caniou, Yves Publication Date: 013 Permanent Link: https://doi.org/10.399/ethz-a-0100633

More information

arxiv: v3 [math.st] 11 Mar 2012

arxiv: v3 [math.st] 11 Mar 2012 Electronic Journal of Statistics Vol. 0 (0000) ISSN: 1935-754 DOI: 10.114/154957804100000000 arxiv:111.1788v3 [math.st] 11 Mar 01 Generalized Hoeffding-Sobol Decomposition for Dependent Variables - Application

More information

Computational methods for uncertainty quantification and sensitivity analysis of complex systems

Computational methods for uncertainty quantification and sensitivity analysis of complex systems DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Computational methods for uncertainty quantification and sensitivity analysis of complex systems

More information

Polynomial chaos expansions for sensitivity analysis

Polynomial chaos expansions for sensitivity analysis c DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Polynomial chaos expansions for sensitivity analysis B. Sudret Chair of Risk, Safety & Uncertainty

More information

Polynomial chaos expansions for structural reliability analysis

Polynomial chaos expansions for structural reliability analysis DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Polynomial chaos expansions for structural reliability analysis B. Sudret & S. Marelli Incl.

More information

arxiv: v1 [math.st] 14 Oct 2013

arxiv: v1 [math.st] 14 Oct 2013 ANOVA decomposition of conditional Gaussian processes for sensitivity analysis with dependent inputs Gaelle Chastaing Loic Le Gratiet arxiv:1310.3578v1 [math.st] 14 Oct 2013 Abstract October 15, 2013 Complex

More information

Gaussian process regression for Sensitivity analysis

Gaussian process regression for Sensitivity analysis Gaussian process regression for Sensitivity analysis GPSS Workshop on UQ, Sheffield, September 2016 Nicolas Durrande, Mines St-Étienne, durrande@emse.fr GPSS workshop on UQ GPs for sensitivity analysis

More information

Modeling Real Estate Data using Quantile Regression

Modeling Real Estate Data using Quantile Regression Modeling Real Estate Data using Semiparametric Quantile Regression Department of Statistics University of Innsbruck September 9th, 2011 Overview 1 Application: 2 3 4 Hedonic regression data for house prices

More information

Sobol-Hoeffding Decomposition with Application to Global Sensitivity Analysis

Sobol-Hoeffding Decomposition with Application to Global Sensitivity Analysis Sobol-Hoeffding decomposition Application to Global SA Computation of the SI Sobol-Hoeffding Decomposition with Application to Global Sensitivity Analysis Olivier Le Maître with Colleague & Friend Omar

More information

Variance-based sensitivity indices for models with dependent inputs

Variance-based sensitivity indices for models with dependent inputs Variance-based sensitivity indices for models with dependent inputs Thierry A. Mara, Stefano Tarantola To cite this version: Thierry A. Mara, Stefano Tarantola. Variance-based sensitivity indices for models

More information

Lecture 14: Variable Selection - Beyond LASSO

Lecture 14: Variable Selection - Beyond LASSO Fall, 2017 Extension of LASSO To achieve oracle properties, L q penalty with 0 < q < 1, SCAD penalty (Fan and Li 2001; Zhang et al. 2007). Adaptive LASSO (Zou 2006; Zhang and Lu 2007; Wang et al. 2007)

More information

Local and Global Sensitivity Analysis

Local and Global Sensitivity Analysis Omar 1,2 1 Duke University Department of Mechanical Engineering & Materials Science omar.knio@duke.edu 2 KAUST Division of Computer, Electrical, Mathematical Science & Engineering omar.knio@kaust.edu.sa

More information

Lectures. Variance-based sensitivity analysis in the presence of correlated input variables. Thomas Most. Source:

Lectures. Variance-based sensitivity analysis in the presence of correlated input variables. Thomas Most. Source: Lectures Variance-based sensitivity analysis in the presence of correlated input variables Thomas Most Source: www.dynardo.de/en/library Variance-based sensitivity analysis in the presence of correlated

More information

Regularization and Variable Selection via the Elastic Net

Regularization and Variable Selection via the Elastic Net p. 1/1 Regularization and Variable Selection via the Elastic Net Hui Zou and Trevor Hastie Journal of Royal Statistical Society, B, 2005 Presenter: Minhua Chen, Nov. 07, 2008 p. 2/1 Agenda Introduction

More information

Penalized Squared Error and Likelihood: Risk Bounds and Fast Algorithms

Penalized Squared Error and Likelihood: Risk Bounds and Fast Algorithms university-logo Penalized Squared Error and Likelihood: Risk Bounds and Fast Algorithms Andrew Barron Cong Huang Xi Luo Department of Statistics Yale University 2008 Workshop on Sparsity in High Dimensional

More information

Linear Regression Linear Regression with Shrinkage

Linear Regression Linear Regression with Shrinkage Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle

More information

Stratified Random Sampling for Dependent Inputs

Stratified Random Sampling for Dependent Inputs Stratified Random Sampling for Dependent Inputs Anirban Mondal Case Western Reserve University, Cleveland, OH 44106, USA arxiv:1904.00555v1 [stat.me] 1 Apr 2019 Abhijit Mandal Wayne State University, Detroit,

More information

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Maximilian Kasy Department of Economics, Harvard University 1 / 37 Agenda 6 equivalent representations of the

More information

Interpolation. 1. Judd, K. Numerical Methods in Economics, Cambridge: MIT Press. Chapter

Interpolation. 1. Judd, K. Numerical Methods in Economics, Cambridge: MIT Press. Chapter Key References: Interpolation 1. Judd, K. Numerical Methods in Economics, Cambridge: MIT Press. Chapter 6. 2. Press, W. et. al. Numerical Recipes in C, Cambridge: Cambridge University Press. Chapter 3

More information

Stochastic optimization - how to improve computational efficiency?

Stochastic optimization - how to improve computational efficiency? Stochastic optimization - how to improve computational efficiency? Christian Bucher Center of Mechanics and Structural Dynamics Vienna University of Technology & DYNARDO GmbH, Vienna Presentation at Czech

More information

Fast learning rates for plug-in classifiers under the margin condition

Fast learning rates for plug-in classifiers under the margin condition Fast learning rates for plug-in classifiers under the margin condition Jean-Yves Audibert 1 Alexandre B. Tsybakov 2 1 Certis ParisTech - Ecole des Ponts, France 2 LPMA Université Pierre et Marie Curie,

More information

Linear Models Review

Linear Models Review Linear Models Review Vectors in IR n will be written as ordered n-tuples which are understood to be column vectors, or n 1 matrices. A vector variable will be indicted with bold face, and the prime sign

More information

Multivariate Regression

Multivariate Regression Multivariate Regression The so-called supervised learning problem is the following: we want to approximate the random variable Y with an appropriate function of the random variables X 1,..., X p with the

More information

approximation algorithms I

approximation algorithms I SUM-OF-SQUARES method and approximation algorithms I David Steurer Cornell Cargese Workshop, 201 meta-task encoded as low-degree polynomial in R x example: f(x) = i,j n w ij x i x j 2 given: functions

More information

Direct Learning: Linear Regression. Donglin Zeng, Department of Biostatistics, University of North Carolina

Direct Learning: Linear Regression. Donglin Zeng, Department of Biostatistics, University of North Carolina Direct Learning: Linear Regression Parametric learning We consider the core function in the prediction rule to be a parametric function. The most commonly used function is a linear function: squared loss:

More information

Lecture 6: Linear models and Gauss-Markov theorem

Lecture 6: Linear models and Gauss-Markov theorem Lecture 6: Linear models and Gauss-Markov theorem Linear model setting Results in simple linear regression can be extended to the following general linear model with independently observed response variables

More information

New results for the functional ANOVA and. Sobol indices

New results for the functional ANOVA and. Sobol indices Generalized Sobol Indices 1 New results for the functional ANOVA and Sobol indices Art B. Owen Stanford University Generalized Sobol Indices 2 Ilya Meerovich Sobol At MCM 2001, Salzburg Known for Sobol

More information

Additive Isotonic Regression

Additive Isotonic Regression Additive Isotonic Regression Enno Mammen and Kyusang Yu 11. July 2006 INTRODUCTION: We have i.i.d. random vectors (Y 1, X 1 ),..., (Y n, X n ) with X i = (X1 i,..., X d i ) and we consider the additive

More information

EE 381V: Large Scale Optimization Fall Lecture 24 April 11

EE 381V: Large Scale Optimization Fall Lecture 24 April 11 EE 381V: Large Scale Optimization Fall 2012 Lecture 24 April 11 Lecturer: Caramanis & Sanghavi Scribe: Tao Huang 24.1 Review In past classes, we studied the problem of sparsity. Sparsity problem is that

More information

INTRODUCING LINEAR REGRESSION MODELS Response or Dependent variable y

INTRODUCING LINEAR REGRESSION MODELS Response or Dependent variable y INTRODUCING LINEAR REGRESSION MODELS Response or Dependent variable y Predictor or Independent variable x Model with error: for i = 1,..., n, y i = α + βx i + ε i ε i : independent errors (sampling, measurement,

More information

Linear regression methods

Linear regression methods Linear regression methods Most of our intuition about statistical methods stem from linear regression. For observations i = 1,..., n, the model is Y i = p X ij β j + ε i, j=1 where Y i is the response

More information

Machine Learning. Part 1. Linear Regression. Machine Learning: Regression Case. .. Dennis Sun DATA 401 Data Science Alex Dekhtyar..

Machine Learning. Part 1. Linear Regression. Machine Learning: Regression Case. .. Dennis Sun DATA 401 Data Science Alex Dekhtyar.. .. Dennis Sun DATA 401 Data Science Alex Dekhtyar.. Machine Learning. Part 1. Linear Regression Machine Learning: Regression Case. Dataset. Consider a collection of features X = {X 1,...,X n }, such that

More information

STAT5044: Regression and Anova. Inyoung Kim

STAT5044: Regression and Anova. Inyoung Kim STAT5044: Regression and Anova Inyoung Kim 2 / 51 Outline 1 Matrix Expression 2 Linear and quadratic forms 3 Properties of quadratic form 4 Properties of estimates 5 Distributional properties 3 / 51 Matrix

More information

The deterministic Lasso

The deterministic Lasso The deterministic Lasso Sara van de Geer Seminar für Statistik, ETH Zürich Abstract We study high-dimensional generalized linear models and empirical risk minimization using the Lasso An oracle inequality

More information

Introduction to Smoothing spline ANOVA models (metamodelling)

Introduction to Smoothing spline ANOVA models (metamodelling) Introduction to Smoothing spline ANOVA models (metamodelling) M. Ratto DYNARE Summer School, Paris, June 215. Joint Research Centre www.jrc.ec.europa.eu Serving society Stimulating innovation Supporting

More information

Quantifying Stochastic Model Errors via Robust Optimization

Quantifying Stochastic Model Errors via Robust Optimization Quantifying Stochastic Model Errors via Robust Optimization IPAM Workshop on Uncertainty Quantification for Multiscale Stochastic Systems and Applications Jan 19, 2016 Henry Lam Industrial & Operations

More information

Efficient Complex Output Prediction

Efficient Complex Output Prediction Efficient Complex Output Prediction Florence d Alché-Buc Joint work with Romain Brault, Alex Lambert, Maxime Sangnier October 12, 2017 LTCI, Télécom ParisTech, Institut-Mines Télécom, Université Paris-Saclay

More information

Neural networks COMS 4771

Neural networks COMS 4771 Neural networks COMS 4771 1. Logistic regression Logistic regression Suppose X = R d and Y = {0, 1}. A logistic regression model is a statistical model where the conditional probability function has a

More information

A Unified Framework for Uncertainty and Sensitivity Analysis of Computational Models with Many Input Parameters

A Unified Framework for Uncertainty and Sensitivity Analysis of Computational Models with Many Input Parameters A Unified Framework for Uncertainty and Sensitivity Analysis of Computational Models with Many Input Parameters C. F. Jeff Wu H. Milton Stewart School of Industrial and Systems Engineering Georgia Institute

More information

Least Squares Regression

Least Squares Regression E0 70 Machine Learning Lecture 4 Jan 7, 03) Least Squares Regression Lecturer: Shivani Agarwal Disclaimer: These notes are a brief summary of the topics covered in the lecture. They are not a substitute

More information

Nonparametric Regression. Badr Missaoui

Nonparametric Regression. Badr Missaoui Badr Missaoui Outline Kernel and local polynomial regression. Penalized regression. We are given n pairs of observations (X 1, Y 1 ),...,(X n, Y n ) where Y i = r(x i ) + ε i, i = 1,..., n and r(x) = E(Y

More information

Motivation Sparse Signal Recovery is an interesting area with many potential applications. Methods developed for solving sparse signal recovery proble

Motivation Sparse Signal Recovery is an interesting area with many potential applications. Methods developed for solving sparse signal recovery proble Bayesian Methods for Sparse Signal Recovery Bhaskar D Rao 1 University of California, San Diego 1 Thanks to David Wipf, Zhilin Zhang and Ritwik Giri Motivation Sparse Signal Recovery is an interesting

More information

Machine Learning. Regression Case Part 1. Linear Regression. Machine Learning: Regression Case.

Machine Learning. Regression Case Part 1. Linear Regression. Machine Learning: Regression Case. .. Cal Poly CSC 566 Advanced Data Mining Alexander Dekhtyar.. Machine Learning. Regression Case Part 1. Linear Regression Machine Learning: Regression Case. Dataset. Consider a collection of features X

More information

Linear regression COMS 4771

Linear regression COMS 4771 Linear regression COMS 4771 1. Old Faithful and prediction functions Prediction problem: Old Faithful geyser (Yellowstone) Task: Predict time of next eruption. 1 / 40 Statistical model for time between

More information

11. Learning graphical models

11. Learning graphical models Learning graphical models 11-1 11. Learning graphical models Maximum likelihood Parameter learning Structural learning Learning partially observed graphical models Learning graphical models 11-2 statistical

More information

Least Squares Regression

Least Squares Regression CIS 50: Machine Learning Spring 08: Lecture 4 Least Squares Regression Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may not cover all the

More information

Linear Regression Linear Regression with Shrinkage

Linear Regression Linear Regression with Shrinkage Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle

More information

Sparse polynomial chaos expansions in engineering applications

Sparse polynomial chaos expansions in engineering applications DEPARTMENT OF CIVIL, ENVIRONMENTAL AND GEOMATIC ENGINEERING CHAIR OF RISK, SAFETY & UNCERTAINTY QUANTIFICATION Sparse polynomial chaos expansions in engineering applications B. Sudret G. Blatman (EDF R&D,

More information

Functional Estimation in Systems Defined by Differential Equation using Bayesian Smoothing Methods

Functional Estimation in Systems Defined by Differential Equation using Bayesian Smoothing Methods Université Catholique de Louvain Institut de Statistique, Biostatistique et Sciences Actuarielles Functional Estimation in Systems Defined by Differential Equation using Bayesian Smoothing Methods 19th

More information

Linear Regression. Aarti Singh. Machine Learning / Sept 27, 2010

Linear Regression. Aarti Singh. Machine Learning / Sept 27, 2010 Linear Regression Aarti Singh Machine Learning 10-701/15-781 Sept 27, 2010 Discrete to Continuous Labels Classification Sports Science News Anemic cell Healthy cell Regression X = Document Y = Topic X

More information

Chapter 3. Linear Models for Regression

Chapter 3. Linear Models for Regression Chapter 3. Linear Models for Regression Wei Pan Division of Biostatistics, School of Public Health, University of Minnesota, Minneapolis, MN 55455 Email: weip@biostat.umn.edu PubH 7475/8475 c Wei Pan Linear

More information

Lecture 3: Statistical Decision Theory (Part II)

Lecture 3: Statistical Decision Theory (Part II) Lecture 3: Statistical Decision Theory (Part II) Hao Helen Zhang Hao Helen Zhang Lecture 3: Statistical Decision Theory (Part II) 1 / 27 Outline of This Note Part I: Statistics Decision Theory (Classical

More information

Parcimonie en apprentissage statistique

Parcimonie en apprentissage statistique Parcimonie en apprentissage statistique Guillaume Obozinski Ecole des Ponts - ParisTech Journée Parcimonie Fédération Charles Hermite, 23 Juin 2014 Parcimonie en apprentissage 1/44 Classical supervised

More information

Support Vector Machine

Support Vector Machine Support Vector Machine Fabrice Rossi SAMM Université Paris 1 Panthéon Sorbonne 2018 Outline Linear Support Vector Machine Kernelized SVM Kernels 2 From ERM to RLM Empirical Risk Minimization in the binary

More information

Linear Models in Machine Learning

Linear Models in Machine Learning CS540 Intro to AI Linear Models in Machine Learning Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu We briefly go over two linear models frequently used in machine learning: linear regression for, well, regression,

More information

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Sparse Recovery using L1 minimization - algorithms Yuejie Chi Department of Electrical and Computer Engineering Spring

More information

Convex relaxation for Combinatorial Penalties

Convex relaxation for Combinatorial Penalties Convex relaxation for Combinatorial Penalties Guillaume Obozinski Equipe Imagine Laboratoire d Informatique Gaspard Monge Ecole des Ponts - ParisTech Joint work with Francis Bach Fête Parisienne in Computation,

More information

SVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels

SVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels SVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels Karl Stratos June 21, 2018 1 / 33 Tangent: Some Loose Ends in Logistic Regression Polynomial feature expansion in logistic

More information

Model Selection and Geometry

Model Selection and Geometry Model Selection and Geometry Pascal Massart Université Paris-Sud, Orsay Leipzig, February Purpose of the talk! Concentration of measure plays a fundamental role in the theory of model selection! Model

More information

Lecture 18: Kernels Risk and Loss Support Vector Regression. Aykut Erdem December 2016 Hacettepe University

Lecture 18: Kernels Risk and Loss Support Vector Regression. Aykut Erdem December 2016 Hacettepe University Lecture 18: Kernels Risk and Loss Support Vector Regression Aykut Erdem December 2016 Hacettepe University Administrative We will have a make-up lecture on next Saturday December 24, 2016 Presentations

More information

Well-developed and understood properties

Well-developed and understood properties 1 INTRODUCTION TO LINEAR MODELS 1 THE CLASSICAL LINEAR MODEL Most commonly used statistical models Flexible models Well-developed and understood properties Ease of interpretation Building block for more

More information

1 Probability Model. 1.1 Types of models to be discussed in the course

1 Probability Model. 1.1 Types of models to be discussed in the course Sufficiency January 11, 2016 Debdeep Pati 1 Probability Model Model: A family of distributions {P θ : θ Θ}. P θ (B) is the probability of the event B when the parameter takes the value θ. P θ is described

More information

Oslo Class 6 Sparsity based regularization

Oslo Class 6 Sparsity based regularization RegML2017@SIMULA Oslo Class 6 Sparsity based regularization Lorenzo Rosasco UNIGE-MIT-IIT May 4, 2017 Learning from data Possible only under assumptions regularization min Ê(w) + λr(w) w Smoothness Sparsity

More information

Chapter 2. Section Section 2.9. J. Kim (ISU) Chapter 2 1 / 26. Design-optimal estimator under stratified random sampling

Chapter 2. Section Section 2.9. J. Kim (ISU) Chapter 2 1 / 26. Design-optimal estimator under stratified random sampling Chapter 2 Section 2.4 - Section 2.9 J. Kim (ISU) Chapter 2 1 / 26 2.4 Regression and stratification Design-optimal estimator under stratified random sampling where (Ŝxxh, Ŝxyh) ˆβ opt = ( x st, ȳ st )

More information

Regression Shrinkage and Selection via the Lasso

Regression Shrinkage and Selection via the Lasso Regression Shrinkage and Selection via the Lasso ROBERT TIBSHIRANI, 1996 Presenter: Guiyun Feng April 27 () 1 / 20 Motivation Estimation in Linear Models: y = β T x + ɛ. data (x i, y i ), i = 1, 2,...,

More information

Effect of outliers on the variable selection by the regularized regression

Effect of outliers on the variable selection by the regularized regression Communications for Statistical Applications and Methods 2018, Vol. 25, No. 2, 235 243 https://doi.org/10.29220/csam.2018.25.2.235 Print ISSN 2287-7843 / Online ISSN 2383-4757 Effect of outliers on the

More information

Message passing and approximate message passing

Message passing and approximate message passing Message passing and approximate message passing Arian Maleki Columbia University 1 / 47 What is the problem? Given pdf µ(x 1, x 2,..., x n ) we are interested in arg maxx1,x 2,...,x n µ(x 1, x 2,..., x

More information

(Part 1) High-dimensional statistics May / 41

(Part 1) High-dimensional statistics May / 41 Theory for the Lasso Recall the linear model Y i = p j=1 β j X (j) i + ɛ i, i = 1,..., n, or, in matrix notation, Y = Xβ + ɛ, To simplify, we assume that the design X is fixed, and that ɛ is N (0, σ 2

More information

Unsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto

Unsupervised Learning Techniques Class 07, 1 March 2006 Andrea Caponnetto Unsupervised Learning Techniques 9.520 Class 07, 1 March 2006 Andrea Caponnetto About this class Goal To introduce some methods for unsupervised learning: Gaussian Mixtures, K-Means, ISOMAP, HLLE, Laplacian

More information

Now consider the case where E(Y) = µ = Xβ and V (Y) = σ 2 G, where G is diagonal, but unknown.

Now consider the case where E(Y) = µ = Xβ and V (Y) = σ 2 G, where G is diagonal, but unknown. Weighting We have seen that if E(Y) = Xβ and V (Y) = σ 2 G, where G is known, the model can be rewritten as a linear model. This is known as generalized least squares or, if G is diagonal, with trace(g)

More information

Addressing high dimensionality in reliability analysis using low-rank tensor approximations

Addressing high dimensionality in reliability analysis using low-rank tensor approximations Addressing high dimensionality in reliability analysis using low-rank tensor approximations K Konakli, Bruno Sudret To cite this version: K Konakli, Bruno Sudret. Addressing high dimensionality in reliability

More information

variability of the model, represented by σ 2 and not accounted for by Xβ

variability of the model, represented by σ 2 and not accounted for by Xβ Posterior Predictive Distribution Suppose we have observed a new set of explanatory variables X and we want to predict the outcomes ỹ using the regression model. Components of uncertainty in p(ỹ y) variability

More information

A new estimator for quantile-oriented sensitivity indices

A new estimator for quantile-oriented sensitivity indices A new estimator for quantile-oriented sensitivity indices Thomas Browne Supervisors : J-C. Fort (Paris 5) & T. Klein (IMT-Toulouse) Advisors : B. Iooss & L. Le Gratiet (EDF R&D-MRI Chatou) EDF R&D-MRI

More information

A regularized least-squares method for sparse low-rank approximation of multivariate functions

A regularized least-squares method for sparse low-rank approximation of multivariate functions Workshop Numerical methods for high-dimensional problems April 18, 2014 A regularized least-squares method for sparse low-rank approximation of multivariate functions Mathilde Chevreuil joint work with

More information

SENSITIVITY ANALYSIS IN NUMERICAL SIMULATION OF MULTIPHASE FLOW FOR CO 2 STORAGE IN SALINE AQUIFERS USING THE PROBABILISTIC COLLOCATION APPROACH

SENSITIVITY ANALYSIS IN NUMERICAL SIMULATION OF MULTIPHASE FLOW FOR CO 2 STORAGE IN SALINE AQUIFERS USING THE PROBABILISTIC COLLOCATION APPROACH XIX International Conference on Water Resources CMWR 2012 University of Illinois at Urbana-Champaign June 17-22,2012 SENSITIVITY ANALYSIS IN NUMERICAL SIMULATION OF MULTIPHASE FLOW FOR CO 2 STORAGE IN

More information

Machine Learning CSE546 Carlos Guestrin University of Washington. October 7, Efficiency: If size(w) = 100B, each prediction is expensive:

Machine Learning CSE546 Carlos Guestrin University of Washington. October 7, Efficiency: If size(w) = 100B, each prediction is expensive: Simple Variable Selection LASSO: Sparse Regression Machine Learning CSE546 Carlos Guestrin University of Washington October 7, 2013 1 Sparsity Vector w is sparse, if many entries are zero: Very useful

More information

Adaptive Piecewise Polynomial Estimation via Trend Filtering

Adaptive Piecewise Polynomial Estimation via Trend Filtering Adaptive Piecewise Polynomial Estimation via Trend Filtering Liubo Li, ShanShan Tu The Ohio State University li.2201@osu.edu, tu.162@osu.edu October 1, 2015 Liubo Li, ShanShan Tu (OSU) Trend Filtering

More information

Classification with Reject Option

Classification with Reject Option Classification with Reject Option Bartlett and Wegkamp (2008) Wegkamp and Yuan (2010) February 17, 2012 Outline. Introduction.. Classification with reject option. Spirit of the papers BW2008.. Infinite

More information

The Adaptive Lasso and Its Oracle Properties Hui Zou (2006), JASA

The Adaptive Lasso and Its Oracle Properties Hui Zou (2006), JASA The Adaptive Lasso and Its Oracle Properties Hui Zou (2006), JASA Presented by Dongjun Chung March 12, 2010 Introduction Definition Oracle Properties Computations Relationship: Nonnegative Garrote Extensions:

More information

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Scientific Computing: An Introductory Survey Chapter 7 Interpolation Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction permitted

More information

Statistics II. Management Degree Management Statistics IIDegree. Statistics II. 2 nd Sem. 2013/2014. Management Degree. Simple Linear Regression

Statistics II. Management Degree Management Statistics IIDegree. Statistics II. 2 nd Sem. 2013/2014. Management Degree. Simple Linear Regression Model 1 2 Ordinary Least Squares 3 4 Non-linearities 5 of the coefficients and their to the model We saw that econometrics studies E (Y x). More generally, we shall study regression analysis. : The regression

More information

Approximate Kernel PCA with Random Features

Approximate Kernel PCA with Random Features Approximate Kernel PCA with Random Features (Computational vs. Statistical Tradeoff) Bharath K. Sriperumbudur Department of Statistics, Pennsylvania State University Journées de Statistique Paris May 28,

More information

BINARY CLASSIFICATION

BINARY CLASSIFICATION BINARY CLASSIFICATION MAXIM RAGINSY The problem of binary classification can be stated as follows. We have a random couple Z = X, Y ), where X R d is called the feature vector and Y {, } is called the

More information

COMS 4721: Machine Learning for Data Science Lecture 6, 2/2/2017

COMS 4721: Machine Learning for Data Science Lecture 6, 2/2/2017 COMS 4721: Machine Learning for Data Science Lecture 6, 2/2/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University UNDERDETERMINED LINEAR EQUATIONS We

More information

Curve learning. p.1/35

Curve learning. p.1/35 Curve learning Gérard Biau UNIVERSITÉ MONTPELLIER II p.1/35 Summary The problem The mathematical model Functional classification 1. Fourier filtering 2. Wavelet filtering Applications p.2/35 The problem

More information

recent developments of approximation theory and greedy algorithms

recent developments of approximation theory and greedy algorithms recent developments of approximation theory and greedy algorithms Peter Binev Department of Mathematics and Interdisciplinary Mathematics Institute University of South Carolina Reduced Order Modeling in

More information

A Bootstrap Lasso + Partial Ridge Method to Construct Confidence Intervals for Parameters in High-dimensional Sparse Linear Models

A Bootstrap Lasso + Partial Ridge Method to Construct Confidence Intervals for Parameters in High-dimensional Sparse Linear Models A Bootstrap Lasso + Partial Ridge Method to Construct Confidence Intervals for Parameters in High-dimensional Sparse Linear Models Jingyi Jessica Li Department of Statistics University of California, Los

More information

Minimizing the Number of Function Evaluations to Estimate Sobol Indices Using Quasi-Monte Carlo

Minimizing the Number of Function Evaluations to Estimate Sobol Indices Using Quasi-Monte Carlo Minimizing the Number of Function Evaluations to Estimate Sobol Indices Using Quasi-Monte Carlo Lluís Antoni Jiménez Rugama Joint work with: Fred J Hickernell (IIT), Clémentine Prieur (Univ Grenoble),

More information

Chap. 3. Controlled Systems, Controllability

Chap. 3. Controlled Systems, Controllability Chap. 3. Controlled Systems, Controllability 1. Controllability of Linear Systems 1.1. Kalman s Criterion Consider the linear system ẋ = Ax + Bu where x R n : state vector and u R m : input vector. A :

More information

Recap from previous lecture

Recap from previous lecture Recap from previous lecture Learning is using past experience to improve future performance. Different types of learning: supervised unsupervised reinforcement active online... For a machine, experience

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1 MA 575 Linear Models: Cedric E Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1 1 Within-group Correlation Let us recall the simple two-level hierarchical

More information

STAT5044: Regression and Anova

STAT5044: Regression and Anova STAT5044: Regression and Anova Inyoung Kim 1 / 25 Outline 1 Multiple Linear Regression 2 / 25 Basic Idea An extra sum of squares: the marginal reduction in the error sum of squares when one or several

More information

Indirect Rule Learning: Support Vector Machines. Donglin Zeng, Department of Biostatistics, University of North Carolina

Indirect Rule Learning: Support Vector Machines. Donglin Zeng, Department of Biostatistics, University of North Carolina Indirect Rule Learning: Support Vector Machines Indirect learning: loss optimization It doesn t estimate the prediction rule f (x) directly, since most loss functions do not have explicit optimizers. Indirection

More information

EE613 Machine Learning for Engineers. Kernel methods Support Vector Machines. jean-marc odobez 2015

EE613 Machine Learning for Engineers. Kernel methods Support Vector Machines. jean-marc odobez 2015 EE613 Machine Learning for Engineers Kernel methods Support Vector Machines jean-marc odobez 2015 overview Kernel methods introductions and main elements defining kernels Kernelization of k-nn, K-Means,

More information

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics Jiti Gao Department of Statistics School of Mathematics and Statistics The University of Western Australia Crawley

More information

A Least Squares Formulation for Canonical Correlation Analysis

A Least Squares Formulation for Canonical Correlation Analysis A Least Squares Formulation for Canonical Correlation Analysis Liang Sun, Shuiwang Ji, and Jieping Ye Department of Computer Science and Engineering Arizona State University Motivation Canonical Correlation

More information

Bayesian variable selection via. Penalized credible regions. Brian Reich, NCSU. Joint work with. Howard Bondell and Ander Wilson

Bayesian variable selection via. Penalized credible regions. Brian Reich, NCSU. Joint work with. Howard Bondell and Ander Wilson Bayesian variable selection via penalized credible regions Brian Reich, NC State Joint work with Howard Bondell and Ander Wilson Brian Reich, NCSU Penalized credible regions 1 Motivation big p, small n

More information

Spectral Regularization

Spectral Regularization Spectral Regularization Lorenzo Rosasco 9.520 Class 07 February 27, 2008 About this class Goal To discuss how a class of regularization methods originally designed for solving ill-posed inverse problems,

More information

Uncertainty Quantification in Computational Science

Uncertainty Quantification in Computational Science DTU 2010 - Lecture I Uncertainty Quantification in Computational Science Jan S Hesthaven Brown University Jan.Hesthaven@Brown.edu Objective of lectures The main objective of these lectures are To offer

More information