Model selection theory: a tutorial with applications to learning

Similar documents
Model Selection and Geometry

Kernel change-point detection

Learning Theory. Ingo Steinwart University of Stuttgart. September 4, 2013

Fast learning rates for plug-in classifiers under the margin condition

Kernels to detect abrupt changes in time series

High-dimensional regression with unknown variance

Risk bounds for model selection via penalization

Adaptive estimation of the intensity of inhomogeneous Poisson processes via concentration inequalities

arxiv:math/ v1 [math.st] 23 Feb 2007

Empirical Risk Minimization

The Behaviour of the Akaike Information Criterion when Applied to Non-nested Sequences of Models

arxiv: v1 [math.st] 10 May 2009

Concentration, self-bounding functions

Gaussian model selection

Least squares under convex constraint

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013

ISyE 691 Data mining and analytics

Lecture 3: Statistical Decision Theory (Part II)

Risk Bounds for CART Classifiers under a Margin Condition

Variable selection for model-based clustering

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures

The deterministic Lasso

Model Selection Tutorial 2: Problems With Using AIC to Select a Subset of Exposures in a Regression Model

Seminar über Statistik FS2008: Model Selection

A Tight Excess Risk Bound via a Unified PAC-Bayesian- Rademacher-Shtarkov-MDL Complexity

ASYMPTOTIC EQUIVALENCE OF DENSITY ESTIMATION AND GAUSSIAN WHITE NOISE. By Michael Nussbaum Weierstrass Institute, Berlin

Model comparison and selection

Indirect Rule Learning: Support Vector Machines. Donglin Zeng, Department of Biostatistics, University of North Carolina

Regression I: Mean Squared Error and Measuring Quality of Fit

A talk on Oracle inequalities and regularization. by Sara van de Geer

The sample complexity of agnostic learning with deterministic labels

ECE521 lecture 4: 19 January Optimization, MLE, regularization

Generalization theory

GAUSSIAN MODEL SELECTION WITH AN UNKNOWN VARIANCE. By Yannick Baraud, Christophe Giraud and Sylvie Huet Université de Nice Sophia Antipolis and INRA

Lecture 7 Introduction to Statistical Decision Theory

Plug-in Approach to Active Learning

Simultaneous estimation of the mean and the variance in heteroscedastic Gaussian regression

Sparse Linear Models (10/7/13)

Fast Rates for Estimation Error and Oracle Inequalities for Model Selection

Lecture 35: December The fundamental statistical distances

FORMULATION OF THE LEARNING PROBLEM

Inverse problems in statistics

Inverse Statistical Learning

Linear Regression. Aarti Singh. Machine Learning / Sept 27, 2010

Exact Minimax Strategies for Predictive Density Estimation, Data Compression, and Model Selection

D I S C U S S I O N P A P E R

Bayesian Regularization

Minimax lower bounds

Penalized Squared Error and Likelihood: Risk Bounds and Fast Algorithms

Inverse problems in statistics

Adaptive Sampling Under Low Noise Conditions 1

The Codimension of the Zeros of a Stable Process in Random Scenery

21.2 Example 1 : Non-parametric regression in Mean Integrated Square Error Density Estimation (L 2 2 risk)

A non asymptotic penalized criterion for Gaussian mixture model selection

Nonparametric regression with martingale increment errors

STAT 535 Lecture 5 November, 2018 Brief overview of Model Selection and Regularization c Marina Meilă

Bayesian Adaptation. Aad van der Vaart. Vrije Universiteit Amsterdam. aad. Bayesian Adaptation p. 1/4

COMPARING LEARNING METHODS FOR CLASSIFICATION

Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½

Material presented. Direct Models for Classification. Agenda. Classification. Classification (2) Classification by machines 6/16/2010.

Perceptron Revisited: Linear Separators. Support Vector Machines

Active Learning: Disagreement Coefficient

Minimax Estimation of Kernel Mean Embeddings

Least Squares Regression

Curve learning. p.1/35

Convex relaxation for Combinatorial Penalties

Testing Restrictions and Comparing Models

A General Overview of Parametric Estimation and Inference Techniques.

An overview of a nonparametric estimation method for Lévy processes

10708 Graphical Models: Homework 2

Least Squares Regression

Linear & nonlinear classifiers

STA414/2104 Statistical Methods for Machine Learning II

3.0.1 Multivariate version and tensor product of experiments

Asymptotic efficiency of simple decisions for the compound decision problem

PCA with random noise. Van Ha Vu. Department of Mathematics Yale University

UNIVERSITÄT POTSDAM Institut für Mathematik

Parametric Techniques

Uniform concentration inequalities, martingales, Rademacher complexity and symmetrization

10-701/ Machine Learning - Midterm Exam, Fall 2010

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Machine Learning. Computational Learning Theory. Eric Xing , Fall Lecture 9, October 5, 2016

Direct Learning: Linear Regression. Donglin Zeng, Department of Biostatistics, University of North Carolina

Using CART to Detect Multiple Change Points in the Mean for large samples

Today. Calculus. Linear Regression. Lagrange Multipliers

Introduction to Statistical modeling: handout for Math 489/583

Approximation Theoretical Questions for SVMs

ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering

Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a

Entropy and Ergodic Theory Lecture 15: A first look at concentration

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators

Bayesian Paradigm. Maximum A Posteriori Estimation

Goodness-of-Fit Tests for Time Series Models: A Score-Marked Empirical Process Approach

Gradient-Based Learning. Sargur N. Srihari

Statistics 203: Introduction to Regression and Analysis of Variance Course review

Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D.

Statistics for high-dimensional data: Group Lasso and additive models

Parametric Techniques Lecture 3

Introduction to Empirical Processes and Semiparametric Inference Lecture 12: Glivenko-Cantelli and Donsker Results

Regression, Ridge Regression, Lasso

Transcription:

Model selection theory: a tutorial with applications to learning Pascal Massart Université Paris-Sud, Orsay ALT 2012, October 29

Asymptotic approach to model selection - Idea of using some penalized empirical criterion goes back to the seminal works of Akaike ( 70). - Akaike celebrated criterion (AIC) suggests to penalize the log-likelihood by the number of parameters of the parametric model. - This criterion is based on some asymptotic approximation that essentially relies on Wilks Theorem

Wilks Theorem: under some proper regularity conditions the log-likelihood L ( n θ) based on n i.i.d. observations with distribution belonging to a parametric model with D parameters obeys to the following weak convergence result ( ( ) L ( θ )) χ 2 D n 0 2 L n θ ( ) where denotes the MLE and θ 0 is the true value of the parameter.

Non asymptotic Theory In many situations, it is usefull to make the size of the models tend to infinity or make the list of models depend on n. In these situations, classical asymptotic analysis breaks down and one needs to introduce an alternative approach that we call non asymptotic. We still like Large values of n! But the size of the models as well the size of the list of models should be authorized to be large too.

Functional estimation The basic problem Construct estimators of some function s, using as few prior information on s as possible. Some typical frameworks are the following. Density estimation ( ) X 1,..., X n i.i.d. sample with unknown density s with respect to some given measure. Regression framework One observes With The explanatory variables The errors are i.i.d. with are fixed or i.i.d.

Binary classification We consider an i.i.d. regression framework where the response variable Y is a «label» :0 or 1. A basic problem in statistical learning is to estimate the best classifier, where η denotes the regression function Gaussian white noise Let s be a numerical function on 0,1. One observes the process on 0,1 defined by ( dy n ) ( x) = s( x)dx + 1 n db ( x ( ),Y n ) Where B is a Brownian motion. The level of noise is written as by allow an easy comparison. ( 0) = 0

Empirical Risk Minimization (ERM) A classical strategy to estimate s consists of taking a set of functions S (a «model») and consider some empirical criterion (based on the data) such that achieves a minimum at point. The ERM estimator of minimizes over S. One can hope that ŝ ŝ is close to, if the target belongs to model S (or at least is not far from S). This approach is most popular in the parametric case (i.e. when S is defined by a finite number of parameters and one assumes that ).

Maximum likelihood estimation (MLE) Context:density estimation (i.i.d. setting to be simple) i.i.d. sample with distribution with ( ) X 1,..., X n Kullback Leibler information

Least squares Regression with White noise with Density with

Exact calculations in the linear case In the white noise or the density frameworks, when S is a finite dimensional subspace of (where denotes the Lebesgue measure in the white noise case), the LSE can be explicitly computed. Let basis of S, then ˆβ λ = φ ( ( λ x)dy n ) ŝ = ( x) or be some orthonormal ˆβλ φ λ λ Λ ˆβ λ = 1 n n φ λ i=1 ( ) X i White noise Density

The model choice paradigm If a model S is defined by a «small» number of parameters (as compared to n), then the target s can happen to be far from the model. If the number of parameters is taken too large then ŝ will be a poor estimator of s even if s truly belongs to S. Illustration (white noise) One takes S as a linear space with dimension D, the expected quadratic risk of the LSE can be easily computed E ŝ s 2 = d 2 ( s,s) + D n Of course, since we do not know the quadratic risk cannot be used as a model choice criterion but just as a benchmark.

First Conclusions It is safer to play with several possible models rather than with a single one given in advance. The notion of expected risk allows to compare the candidates and can serve as a benchmark. According to the risk minimization criterion, S is a «good» model does not mean that the target s belongs to S. Since the minimization of the risk cannot be used as a selection criterion, one needs to introduce some empirical version of it.

Model selection via penalization Consider some empirical criterion. Framework: Consider some (at most countable) collection of models. Represent each model by the ERM ŝ m on. Purpose: select the «best» estimator among the collection ( ŝ ) m m M. Procedure: Given some penalty function, we take ˆm minimizing over γ n ŝ m and define ( ) + pen( m) s = ŝ ˆm.

The classical asymptotic approach Origin: Akaike (log-likelihood), Mallows (least squares) 1 The penalty function is proportional to the number of parameters of the model S. m Akaike : D m / n Mallows : 2D m / n, where the variance of the errors of the regression framework is assumed to be equal to 1 by the sake of simplicity. 2 The heuristics (Akaike ( 73)) leading to the choice of the penalty function D m / n relies on the assumption: the dimensions and the number of the models are bounded w.r.t. n and n tends to infinity.

BIC (log-likelihood) criterion Schwartz ( 78) : - aims at selecting a «true» model rather than mimicking an oracle - also asymptotic, with a penalty which is proportional to the number of parameters: ln n ( )D m / n The non asymptotic approach Barron,Cover ( 91) for discrete models, Birgé, Massart ( 97) and Barron, Birgé, Massart ( 99)) for general models. Differs from the asymptotic approach on the following points

The number as well as the dimensions of the models may depend on n. One can choose a list of models because of its approximation properties: wavelet expansions, trigonometric or piecewise polynomials, artificial neural networks etc It may perfectly happen that many models of the list have the same dimension and in our view, the «complexity» of the list of models is typically taken into account. Shape of the penalty with e x m Σ. m M C 1 D m n + C 2 x m n

Data driven penalization Practical implementation requires some datadriven calibration of the penalty. «Recipe» 1. Compute the ERM ŝ D on the union of models with D parameters 2. Use theory to guess the shape of the penalty pen(d), typically pen(d)=ad (but ad(2+ln(n/d)) is another possibility) 3. Estimate a from the data by multiplying by 2 the smallest value for which the penalized criterion explodes. Implemented first by Lebarbier ( 05) for multiple change points detection

Celeux, Martin, Maugis 07 Gene expression data: 1020 genes and 20 experiments Mixture models Choice of K? Slope heuristics: K=17 BIC: K=17 ICL: K=15 Adjustment of the slope Comparison

Akaike s heuristics revisited The main issue is to remove the asymptotic approximation argument in Akaike s heuristics γ n ( ŝ ) D = γ ( n s ) D γ ( n s ) D γ ( n ŝ ) D variance term minimizing minimizing γ n ( ŝ ) D + pen( D), is equivalent to γ n ( s ) D γ ( n s) ˆv D + pen( D) Fair estimate of (s,s D )

( ) = ˆv D + ( s D,ŝ ) D Ideally: pen id D In order to (approximately) minimize (s,ŝ D ) = (s,s D ) + ( s D,ŝ ) D The key : Evaluate the excess risks ˆv D = γ n ( s ) D γ ( n ŝ ) D ( s D,ŝ ) D This the very point where the various approaches diverge. Akaike s criterion relies on the asymptotic approximation (s D,ŝ D ) ˆv D D 2n

The method initiated in Birgé, Massart ( 97) relies on upper bounds for the sum of the excess risks which can be written as ( ) = γ ( n s ) D γ ( n ŝ ) D ˆv D + s D,ŝ D γ n where denotes the empirical process γ ( n t) = γ ( n t) E γ ( n t) These bounds derive from concentration inequalities for the supremum of the appropriately weighted empirical process γ n ( t) γ n u ( ) ( ) ω t,u,t S D The prototype being Talagrand s inequality ( 96) for empirical processes.

This approach has been fruitfully used in several works. Among others: Baraud ( 00) and ( 03) for least squares in the regression framework, Castellan ( 03) for log-splines density estimation, Patricia Reynaud ( 03) for poisson processes, etc

Main drawback: typically involve some unkown multiplicative constante which may depend on the unknown distribution (variance of the regression errors, supremum of the density, classification noise etc ). Needs to be calibrated Slope heuristics : one looks for some approximation of (typically) of the form ad with a unknown. When D is large, γ ( n s ) D is almost constant, it suffices to «read» a as a slope on the graph of γ ( n ŝ ) D. On chooses the final penalty as pen( D) = 2 ad

In fact pen ( min D) = ˆv D is a minimal penalty and the slope heuristics provides a way of approximating it. The factor 2 which is finally used reflects our hope that the excess risks ˆv D = γ n ( s ) D γ ( n ŝ ) ( s,ŝ ) D D D are of the same order of magnitude. If this the case then «optimal» penalty=2 * «minimal» penalty

Recent advances Justification of the slope heuristics: Arlot and Massart (JMLR 08) for histograms in the regression framework. Phd of Saumard (2010) regular parametric models. Boucheron and Massart (PTRF 11) for concentration of the empirical excess risk (Wilks phenomenon) Calibration of regularization Linear estimators Arlot and Bach (2010) Lasso type algorithms. Thesis: Connault (2010) and Meynet (work in progress )

High dimensional Wilks phenomenon Wilks Theorem: under some proper regularity conditions the log-likelihood L ( n θ) based on n i.i.d. observations with distribution belonging to a parametric model with D parameters obeys to the following weak convergence result ( ( ) L ( θ )) χ 2 D n 0 2 L n θ ( ) where denotes the MLE and θ 0 is the true value of the parameter.

Question: what s left if we consider possibly irregular empirical risk minimization procedures and let the dimension of the model tend to infinity? Obviously one cannot expect similar asymptotic results. However it is still possible to exhibit some kind of Wilks phenomenon. Motivation: modification of Akaike s heuristics for model selection Data-driven penalties

A statistical learning framework We consider the i.i.d. framework where one observes independent copies of a random variable with distribution. We have in mind the regression framework for which. is an explanatory variable and Let s is the response variable. be some target function to be estimated. For instance, if denotes the regression function The function of interest function itself. s may be the regression

In the binary classification case where the response variable takes only the two values 0 and 1, it may be the Bayes classifier We consider some criterion target function over some set S s s = 1Ι η 1/ 2 { }, such that the achieves the minimum of ( ) t Pγ t,.. For example with S = L 2 leads to the regression function as a minimizer { { }} with S = t :X 0,1 classifier leads to the Bayes

Introducing the empirical criterion in order to estimate one considers some S subset of (a «model») and defines the empirical risk minimizer of over. as a minimizer This commonly used procedure includes LSE and also MLE for density estimation. In this presentation we shall assume that (makes life simpler but not necessary) and also that 0 γ 1 S S s (boundedness is necessary). ŝ s S

Introducing the natural loss function s,t We are dealing with two «dual» estimation errors the excess loss ( ) = Pγ ( t,. ) Pγ ( s,. ) ( ) = Pγ ( ŝ,. ) Pγ ( s,. ) s,ŝ the empirical excess loss n ( s,ŝ) = P n γ ( s,. ) P n γ ( ŝ,. ) Note that for MLE and Wilks theorem provides the asymptotic behavior of s,ŝ when is a regular parametric model. n ( ) S ( ) = logt (.) γ t,.

Crucial issue: Concentration of the empirical excess loss: connected to empirical processes theory because Difficult problem: Talagrand s inequality does not make directly the job (the Let us begin with the related but easier question: n ( s,ŝ) = sup P n t S rate is hard to gain). What is the order of magnitude of the excess loss and the empirical excess loss? ( γ ( s,. ) γ ( t,. )) 1/ n

Risk bounds for the excess loss We need to relate the variance of with the excess loss Introducing some pseudo-metric d such that We assume that for some convenient function In the regression or the classification case d is simply the distance and is either identity for regression or is related to a margin condition for classification. ( ) = P γ ( t,. ) γ ( s,. ) s,t P γ t,. ( ) ( ( ) γ ( s,. )) 2 d 2 ( s,t) ( ) w ( s,t) d s,t ( ) w γ ( t,. ) γ ( s,. ) w

Tsybakov s margin condition (AOS 2004) where and with d 2 ( s,t) = E s X ( ) t( X ) Since for binary classification this condition is closely related to the behavior of around. For example margin condition κ = 1 is achieved whenever 2η 1 h

Heuristics Let us introduce Then γ n ( t) = ( P n P)γ ( t,. ) ( s,ŝ) + ( n s,ŝ) = γ ( n s) γ ( n ŝ) ( s,ŝ ) γ ( n s) γ ( n ŝ) Now the variance of is bounded by hence empirical process theory tells you that the uniform fluctuation of remains under control in the ball

(Massart, Nédélec AOS 2006) Theorem : Let, such that, with and. Assume that and E sup t S,d ( s,t) σ ( ( )) n γ ( n s) γ n t φ σ ( ) for every such that. Then, defining as one has where ( ) E s,s Cε * is an absolute constant. 2

Application to classification Tsybakov s framework Tsybakov s margin condition means that An entropy with bracketing condition implies that one can take and we recover Tsybakov s rate ε 2 κ /( 2κ + ρ 1) * n VC-classes under margin condition 2η 1 h one has w ε. If is a VC-class with VC-dimension ( ) = ε / h φ σ so that (whenever ) S D ( ) ( ) σ D 1+ log( 1 / σ ) nh 2 D ε * 2 = C nh nh2 D 1+ log D

Main points Local behavior of the empirical process Connection between and. Better rates than usual in VC-theory These rates are optimal (minimax)

Concentration of the empirical excess loss Joint work with Boucheron (PTRF 2011). Since s,ŝ the proof of the above Theorem also leads to an upper bound for the empirical excess risk for the same price. In other words is also bounded by ( ) + ( n s,ŝ) = γ ( n s) γ ( n ŝ) E γ n ( s) γ n ŝ up to some absolute constant. Concentration: start from identity ( ) n ( s,ŝ) = sup P n t S ( γ ( s,. ) γ ( t,. ))

Concentration tools Let ξ ' ' 1,...ξ n be some independent copy of ξ 1,...ξ n. Defining ' Z i and setting ( ) = ζ ξ 1,...,ξ i 1,ξ i ',ξ i+1,...ξ n V + = E n i=1 ( ' Z Z ) 2 i ξ + Efron-Stein s inequality asserts that A Burkholder-type inequality (BBLM, AOP2005) For every such that is integrable, one has ( Z E Z ) + q 3q V + q/ 2

Comments This inequality can be compared to Burkholder s martingale inequality where denotes the quadratic variation w.r.t. Doob s filtration, and trivial -field. It can also be compared with Marcinkiewicz Zygmund s inequality which asserts that in the case where

Note that the constant in Burkhölder s inequality cannot generally be improved. Our inequality is therefore somewhere «between» Burkholder s and Marcinkiewicz-Zygmund s inequalities. We always get the factor instead of which turns out to be a crucial gain if one wants to derive sub-gaussian inequalities. The price is to make the substitution which is absolutely painless in the Marcinkiewicz-Zygmund case. More generally, for the examples that we have in view, will turn to be a quite manageable quantity and explicit moment inequalities will be obtained by applying iteratively the preceding one.

Fundamental example The supremum of an empirical process Z = sup t S n i=1 provides an important example, both for theory and applications. Assuming that, Talagrand s inequality (Inventiones 1996) ensures the existence of some absolute positive contants and, such that f t ( ), ξ i where W = sup t S n i=1 f t 2 ( ξ ) i

Moment inequalities in action Why? Example: for empirical processes, one can prove that for some absolute positive constant Optimize Markov s inequality w.r.t. q Talagrand s concentration inequality

Refining Talagrand s inequality The key: In the process of recovering Talagrand s inequality via the moment method above, we may improve on the variance factor. Indeed, setting Z = sup t S we see that and therefore V + = E n n f ( t ξ ) i = fŝ ξ i i=1 i=1 Z Z i ' n i=1 fŝ ( ) and ( ) = n n s,ŝ ( ξ ) i ( ' fŝ ξ ) i n ( ' Z Z ) 2 i ξ + 2 Pfŝ2 + fŝ2 ξ i i=1 ( ( ))

at this stage instead of using the crude bound V + n 2 sup t S Pf t 2 + sup t S P n f t 2 we can use the refined bound V + n 2Pf 2 + 2P fŝ2 ŝ n ( ) 4Pfŝ2 + 2 P n P ( 2 ) ( ) fŝ Now the point is that on the one hand ( ) w 2 s,s 2 P f s ( )

and on the other hand we can handle the second term root trick. ( ) ( P n P) 2 fŝ P n P ( ) fŝ behave not worse than So finally it can be proved that by using some kind of square ( 2 ) can indeed shown to ( P n P) ( ) fŝ = ( n s,ŝ) + ( s,ŝ) Var Z E V + ( Cnw2 ε ) * and similar results for higher moments.

Illustration 1 In the (bounded) regression case. If we consider the regressogram estimator on some partition with pieces, it can be proved that In this case ne n s,ŝ can be shown to be approximately proportional to D. This exemplifies the high dimensional Wilks phenomenon. Application to model selection with adaptive penalties: Arlot and Massart, JMLR 2009.. D n n ( s,ŝ) E n s,ŝ ( ) ( ) C q qd + q

Illustration 2 It can be shown that in the classification case, If S is a VC-class with VC-dimension D, under the margin condition 2η 1 h nh ( n s,ŝ) E ( n s,ŝ) C qd 1+ log nh2 q D + q provided that nh 2 D. Application to model selection: work in progress with Saumard and Boucheron.

Thanks for your attention!