Combining estimates in regression and. classication. Michael LeBlanc. and. Robert Tibshirani. and. Department of Statistics. cuniversity of Toronto

Similar documents
A Bias Correction for the Minimum Error Rate in Cross-validation

1/sqrt(B) convergence 1/B convergence B

Linear Regression and Its Applications

Inversion Base Height. Daggot Pressure Gradient Visibility (miles)

Least Absolute Shrinkage is Equivalent to Quadratic Penalization

COS 424: Interacting with Data

High-dimensional regression

Machine Learning for OR & FE

A Modern Look at Classical Multivariate Techniques

Plan of Class 4. Radial Basis Functions with moving centers. Projection Pursuit Regression and ridge. Principal Component Analysis: basic ideas

Business Statistics. Tommaso Proietti. Model Evaluation and Selection. DEF - Università di Roma 'Tor Vergata'

Widths. Center Fluctuations. Centers. Centers. Widths

Proteomics and Variable Selection

SF2930 Regression Analysis

INRIA Rh^one-Alpes. Abstract. Friedman (1989) has proposed a regularization technique (RDA) of discriminant analysis

Lecture 7: Modeling Krak(en)

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines

* * * * * * * * * * * * * * * ** * **

MS-C1620 Statistical inference

Regression Shrinkage and Selection via the Lasso

RANDOMIZING OUTPUTS TO INCREASE PREDICTION ACCURACY

COS513: FOUNDATIONS OF PROBABILISTIC MODELS LECTURE 10

Motivating the Covariance Matrix

Recap from previous lecture

Comments on \Wavelets in Statistics: A Review" by. A. Antoniadis. Jianqing Fan. University of North Carolina, Chapel Hill

Non-linear Supervised High Frequency Trading Strategies with Applications in US Equity Markets

Lecture 6: Methods for high-dimensional problems

1. Introduction Over the last three decades a number of model selection criteria have been proposed, including AIC (Akaike, 1973), AICC (Hurvich & Tsa

Gaussian Process Regression: Active Data Selection and Test Point. Rejection. Sambu Seo Marko Wallat Thore Graepel Klaus Obermayer

EXAM IN STATISTICAL MACHINE LEARNING STATISTISK MASKININLÄRNING

Managing Uncertainty

CMSC858P Supervised Learning Methods

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

Variable Selection and Weighting by Nearest Neighbor Ensembles

Multivariate statistical methods and data mining in particle physics

Chap 1. Overview of Statistical Learning (HTF, , 2.9) Yongdai Kim Seoul National University

Machine Learning. Regression basics. Marc Toussaint University of Stuttgart Summer 2015

CS 195-5: Machine Learning Problem Set 1

Gaussian Processes for Regression. Carl Edward Rasmussen. Department of Computer Science. Toronto, ONT, M5S 1A4, Canada.

Learning with Ensembles: How. over-tting can be useful. Anders Krogh Copenhagen, Denmark. Abstract

Machine Learning. Nathalie Villa-Vialaneix - Formation INRA, Niveau 3

Linear Models for Regression

Averaging Estimators for Regressions with a Possible Structural Break

Performance of Cross Validation in Tree-Based Models

output dimension input dimension Gaussian evidence Gaussian Gaussian evidence evidence from t +1 inputs and outputs at time t x t+2 x t-1 x t+1

An Unbiased C p Criterion for Multivariate Ridge Regression

Final Overview. Introduction to ML. Marek Petrik 4/25/2017

COS513: FOUNDATIONS OF PROBABILISTIC MODELS LECTURE 9: LINEAR REGRESSION

Linear model selection and regularization

Announcements. Proposals graded

below, kernel PCA Eigenvectors, and linear combinations thereof. For the cases where the pre-image does exist, we can provide a means of constructing

Iterative Selection Using Orthogonal Regression Techniques

Business Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata'

COMP 551 Applied Machine Learning Lecture 3: Linear regression (cont d)

Math for Machine Learning Open Doors to Data Science and Artificial Intelligence. Richard Han

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function.

Generalized Elastic Net Regression

Classification and Regression Trees

Regularization: Ridge Regression and the LASSO

Chapter 14 Stein-Rule Estimation

Biostatistics Advanced Methods in Biostatistics IV

Module 17: Bayesian Statistics for Genetics Lecture 4: Linear regression

ISyE 691 Data mining and analytics

Tutorial on Machine Learning for Advanced Electronics

STAT 535 Lecture 5 November, 2018 Brief overview of Model Selection and Regularization c Marina Meilă

The lasso, persistence, and cross-validation

Nearest Neighbor. Machine Learning CSE546 Kevin Jamieson University of Washington. October 26, Kevin Jamieson 2

Lossless Online Bayesian Bagging

Data Mining Stat 588

Classification Methods II: Linear and Quadratic Discrimminant Analysis

ON THE FORWARD AND BACKWARD ALGORITHMS OF PROJECTION PURSUIT 1. BY MU ZHU University of Waterloo

STATISTICAL LEARNING SYSTEMS

PDEEC Machine Learning 2016/17

A Note on Bootstraps and Robustness. Tony Lancaster, Brown University, December 2003.

data lam=36.9 lam=6.69 lam=4.18 lam=2.92 lam=2.21 time max wavelength modulus of max wavelength cycle

MODULE -4 BAYEIAN LEARNING

MSA220 Statistical Learning for Big Data

MINIMUM EXPECTED RISK PROBABILITY ESTIMATES FOR NONPARAMETRIC NEIGHBORHOOD CLASSIFIERS. Maya Gupta, Luca Cazzanti, and Santosh Srivastava

Machine Learning for Microeconometrics

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

UNIVERSITETET I OSLO

Zero-One Loss Functions

Boosting. Ryan Tibshirani Data Mining: / April Optional reading: ISL 8.2, ESL , 10.7, 10.13

Lecture 3. Linear Regression II Bastian Leibe RWTH Aachen

This model of the conditional expectation is linear in the parameters. A more practical and relaxed attitude towards linear regression is to say that

c Copyright 998 by Gareth James All Rights Reserved ii

Boosting Methods: Why They Can Be Useful for High-Dimensional Data

Generalized Boosted Models: A guide to the gbm package

The Poisson transform for unnormalised statistical models. Nicolas Chopin (ENSAE) joint work with Simon Barthelmé (CNRS, Gipsa-LAB)

Introduction to machine learning and pattern recognition Lecture 2 Coryn Bailer-Jones


Estimating subgroup specific treatment effects via concave fusion

Regularized Discriminant Analysis. Part I. Linear and Quadratic Discriminant Analysis. Discriminant Analysis. Example. Example. Class distribution

Naive Bayesian classifiers for multinomial features: a theoretical analysis

Tracking time-varying-coe$cient functions

The exam is closed book, closed notes except your one-page (two sides) or two-page (one side) crib sheet.

Statistical Data Mining and Machine Learning Hilary Term 2016

Support Vector Machines for Classification: A Statistical Portrait

Feature selection and classifier performance in computer-aided diagnosis: The effect of finite sample size

Prediction & Feature Selection in GLM

Transcription:

Combining estimates in regression and classication Michael LeBlanc and Robert Tibshirani Department of Preventive Medicine and Biostatistics and Department of Statistics University of Toronto December 3, 993 cuniversity of Toronto Abstract We consider the problem of how to combine a collection of general regression t vectors in order to obtain a better predictive model. The individual ts may be from subset linear regression, ridge regression, or something more complex like a neural network. We develop a general framework for this problem and examine a recent cross-validation-based proposal called \stacking" in this context. Combination methods based on the bootstrap and analytic methods are also derived and compared in a number of examples, including best subsets regression and regression trees. Finally, we apply these ideas to classication problems where the estimated combination weights can yield insight into the structure of the problem. Introduction Consider a standard regression setup: we have predictor measurements x i = (x i ;...x ip ) T and a response measurement y i on N independent training cases. Let z represent the entire training sample. Our goal is derive a function cz(x) that accurately predicts future y values. Suppose we have available K dierent regression tted values from these data, denoted by c k z(x), for k = ; 2;...K. For example, c k z(x) might be a least squares t for some subset of the variables, or a ridge regression t, or

something more complicated like the result of a projection-pursuit regression or neural network model. Or the collection of c k z(x)s might correspond to a single procedure run with K dierent values of an adjustable parameter. In this paper we consider the problem of how to best combine these estimates in order to obtain an estimator that is better than any of the individual ts. The class that we consider has the form of a simple linear combination: KX k= k c k z(x): () One way to obtain estimates of ;... K is by least squares regression of y on c z(x);...c K z (x). But this might produce poor estimates because it doesn't take into account the relative amount of tting that is present in each of the c k z(x)s, or the correlation of the c k z(x)s induced by the fact that all of the regression ts are estimated from the data z. For example, if the c k z(x)s represent a nested set of linear models, ^ k will equal one for the largest model and zero for the others, and hence will simply reproduce the largest model. In a paper in the neural network literature, Wolpert (992) presented an interesting idea known as \stacked generalization" for combining estimators. His proposal was translated into statistical language by Breiman (993); he applied and studied it the regression setting, calling it \stacked regression". Here is how stacking works. We let c k z (?i) (x i ) denote the leave-one out crossvalidated t for c k z(x), evaluated at x = x i. The stacking method minimizes h i= y i? KX k= k c k z (?i) (x i )i 2 (2) producing estimates ^ ;... ^ K. The nal predictor function is vz(x) = P ^k c k z(x). Notice how this diers from a more standard use of cross-validation. Usually for each method k one constructs the prediction error estimate cpe(k) = N i= [y i? c k z (?i) (x i )] 2 : (3) Then we choose c k z(x) that minimizes PE(k). c In stacking we are estimating a linear combination of models rather than choosing just one. In the P particular cases that he tried, Breiman found that the linear combination ^ k= k c k z(x) did not exhibit good prediction performance. However K when the coecients in (2) were constrained to be non-negative, vz(x) showed better prediction error than any of the individual c k z(x). In some cases the improvement was substantial. This paper grew out of our attempt to understand how and why stacking works. Cross-validation is usually used to estimate the prediction error of an 2

estimator. At rst glance, stacking seems to use cross-validation for a fundamentally dierent purpose, namely to construct a new estimator from the data. In this paper we derive a framework for the problem of combining estimators and as a result we cast the stacking method into more familiar terms. We also derive combination estimators based on the bootstrap, and analytic methods in some simple cases. 2 A framework for combining regression estimators As in section, we have data (x i ; y i ); i = ; 2;...N, with x i = (x i...x ip ), and let z be the entire sample. We assume that each pair (x i ; y i ) is an independent realization of random variables (X; Y ) having distribution F. Let Cz(x 0 ) = (c z(x 0 );...c K z (x 0 )) T be a K-vector of estimators evaluated at X = x 0, based on data z. We seek coecients = ( ;... K ) T so that the linear combination estimator Cz(x) T = P k c k z(x) has low prediction error. Our criterion for selecting is ~ = argmin E 0F (Y 0? Cz(X 0 ) T ) 2 : () Here z is xed and E 0F denotes expectation under (X 0 ; Y 0 ) F. Of course in real problems we don't know E 0F and so we have to derive sample-based estimates. The obvious estimate of g(z; F; ) = E 0F (Y 0? Cz(X 0 ) T ) 2 is the plug-in or resubstitution estimate g(z; ^F ; ) = E 0 ^F (Y 0? Cz(X 0 ) T ) 2 = N whose minimizer is the least squares estimator (y i? Cz(x i ) T ) 2 ; (5) ^ LS = [ X i (Cz(x i )Cz(x i ) T ]? [ X i Cz(x i )y i ]: (6) What's wrong with ^ LS? The problem is that the data z = (x ; y ; )...(x n ; y n ) is used in two places: in constructing the Cz and in evaluating the error between y i and Cz(x i ). As a result:. P i C z(x i )Cz(x i ) T and P i C z(x i )y i are biased estimates of their population analogues 2. g(z; ^F ; ) is a biased estimator of g(z; F; ): 3

Although description () above may seem to be more direct, in section 3 we will nd that description (2) is more useful. The stacking method estimates the prediction error g(z; F; ) by N (y i? C z (?i)(x i ) T ) 2 : (7) The corresponding minimizer gives the stacking estimator of ~ : ^ St = [ X i (Cz (?i) (x i)cz (?i)(x i) T ]? X i Cz (?i)(x i)y i : (8) Stacking uses dierent data to construct C and to evaluate the error between y and C, and hence should produce a less biased estimator of g(z; F; ). Remark A. Note that there is no intercept in the linear combination in (). This makes sense when each estimator c k z(x) is an estimator of Y, so that E[c k z(x)] EY, and hence there are values of giving E[ P k c k z(x)] EY (e.g. take P k = ). In the more general case, each c k z(x) does not necessarily estimate Y : for example each c k z(x) might be an adaptively chosen basis function in a nonlinear regression model. Then we would want to include an intercept in the linear combination: this causes no diculty in the above framework, since we can just set c z(x). 3 Bootstrap estimates POne can apply the bootstrap P to this problem by bias-correcting the quantities i C z(x i )Cz(x i ) T and i C z(x i )y i appearing in the least squares estimator (6). However, it is more useful to approach the problem in terms of bias correction of the prediction error function g(z; ^F; ). We then minimize the biascorrected function ~g(z; ^F; ), to obtain an improved estimator ~. The estimator that we obtain from this procedure is in fact the least squares estimator that uses bias corrected versions of P i C z(x i )Cz(x i ) T and P i C z(x i )y i. The advantage of approaching the problem through the prediction error function is that regularization of the estimator can be incorporated in a straightforward manner (section 5). The method presented here is somewhat novel in that we bias-correct a function of, rather than a single estimator, as is usually the case. A similar proposal in a dierent setting can be found in McCullagh and Tibshirani (988). The bias of g(z; ^F ; ) is (F; ) = E F [g(z; F; )? g(z; ^F ; )]: (9) Given an estimate ^(F; ), our improved estimate is ^g(z; F; ) = g(z; ^F; ) + ^(F; ): (0)

Finally, an estimate of is obtained by minimization of expression (0). How can we estimate (F; )? Note that the stacking method implicitly uses the quantity N nx (y i? C z (?i)(x i ) T ) 2? N nx (y i? Cz(x i ) T ) 2 to estimate (F; ). Here we outline a bootstrap method that is similar to the estimation of the optimism of an error rate (Efron, 982; Efron and Tibshirani, 99 chapter 7). The bootstrap estimates (F; ) by ( ^F ; ) = E ^F [g(z ; ^F; )? g(z ; ^F ; )] () where ^F is the empirical distribution of a sample drawn with replacement from z. The advantage of the simple linear combination estimator is that the the resulting minimizer of (0) can be written down explicitly (c.f section 8). Let g (z; ^F) = N g 2 (z; ^F) = N Then the minimizer of (0) is Cz(x i )Cz(x i ) T Cz(x i )y i ( ^F ) = E ^F [g (z ; ^F)? g (z ; ^F )] 2 ( ^F ) = E ^F [g 2 (z ; ^F)? g2 (z ; ^F )] (2) ^ Bo = [g (z; ^F) + ( ^F )]? [g 2 (z; ^F) + 2 ( ^F )]: (3) P P N In simple terms, we bias-correct both C z(x i )Cz(x i ) T N =N and C z(x i )y i =N, and the resulting estimator is just the least squares estimator that uses the biascorrected versions. As with most applications of the bootstrap, we must use bootstrap sampling (Monte Carlo simulation) to approximate these quantities. Full details are given in the algorithm below. 5

Summary of bootstrap algorithm for combining estimators. Draw B bootstrap samples z b = (x b ; y b ), b = ; 2;... B, and from each sample derive the estimators C z b. 2. Evaluate C z b on both the original sample and on the bootstrap sample, and compute h ^ = i C B N z b (x i)c z b(x i) T? C N z b(x b i )C z b(x b i ) T i= i= h ^2 = C B N z b (x i)y i? N i= i= This gives corrected variance and covariances M CC = N M Cy = N C z b (x b i )y b i Cz(x i)cz(x i) T + ^ i= Cz(x i)y i + ^2: i= 3. Use M CC and M Cy to produce a (possibly regularized) regression of y on C. The regression coecients ^ are the (estimated) optimal combination weights. i : The regularized regression mentioned in step 3 is discussed is section 5. Remark B. A key advantage of the simple linear combination estimator is that the minimizers of the bias-corrected prediction error (0) can be written down explicitly. In section 8) we discuss more general tuning parameter selection problems where this will typically not be possible, so that the proposed procedure may not be computationally feasible. Remark C. Suppose that each c k z (x i) is a linear least squares t for a xed subset of p k variables. Then it is easy to show that the kth element of the bootstrap correction 2 ( ^F ) is?p k^ 2 () where ^ P 2 is the estimated variance of y i. Thus the bootstrap correction 2 ( ^F ) adjusts Cz(x i )y i =N downward to account for the number of regressors used in each c k z. The elements of ( ^F) are more complicated. Let the design matrix for c k z be Z,with a corresponding bootstrap value of Zk. Then the kkth element of ( ^F ) is ^ 2 fe ^F [(Z T k Z k)? ]Z T k Z k? pg 0; (5) 6

the inequality following P from Jensen's inequality. Thus ( ^F ) will tend to inate the diagonal of Cz(x i )Cz(x i ) T =N and hence shrink the least squares estimator ^ LS. The o-diagonal elements of ( ^F ) are more dicult to analyze: empirically, they seem to be negative when the c k zs are positively correlated. Linear estimators and generalized cross-validation Suppose each of the estimators c k z (x i), i = ; 2;...N, can be written as H k y for some xed matrix H k. For example H k y might be the least squares t for a xed subset of the variables X ; X 2...X p, or it might be a cubic smoothing spline t. We can obtain an analytic estimate of the combination weights, by approximating the cross-validation estimate. Let h k ii be the iith element of H k. A standard approximation used in generalized cross-validation gives c k z (x (?i) i ) = ck z(x i )? y i h k ii? h k ii ck z(x i )? y i tr(h k )=N? tr(h k )=N ~c k z(x i ): (6) Therefore a simple estimate of the combination weights can be obtained by least squares regression of y i on ~c k z (x i). Denote the resulting estimate by ^ GCV. When c k z(x i ) is an adaptively chosen linear t, for example a best subset regression, the above derivation does not apply. However, one might try ignoring the adaptivity in the estimators and use above analytic correction anyway. We explore this idea in example of section 6. 5 Regularization From the previous discussion we have four dierent estimators of the combination weights :. ^LS : the least-squares estimator dened in (6). 2. ^St : the stacked regression estimator dened in (8). 3. ^Bo : the bootstrap estimator dened in (3). ^GCV : the generalized cross-validation estimator dened below equation (6), available for linear estimators, c k z(x) = H k y. In our discussion we have derived each of these estimates as minimizers of some roughly unbiased estimate of prediction error ^g(z; ^F ; ). However in our simulation study of the next section (and also in Breiman's 993 study), 7

we nd that none of these estimators work well. All of these are unrestricted least squares estimators, and it turns out that some sort of regularization may needed to improve their performance. This is not surprising, since the same phenomenon occurs in multiple linear regression. In that setting, the average residual sum of squares is unbiased for the true prediction error, but its minimizer the least squares estimator does not necessarily possess the minimum prediction error. Often a regularized estimate, for example a ridge estimator or a subset regression, has lower prediction error than the least squares estimator. In terms of our development of section 2, the prediction error estimate ^g(z; F; ) is approximately unbiassed for g(z; F; ) for each xed value, but it is no longer unbiassed when a estimator ^ is substituted for. Roughly unbiassed estimators for g(z; F; ^) can be constructed by adding a regularization term to ^g(z; F; ) and this leads to regularized versions of the four estimators listed above. We investigate a number of forms of shrinkage in this paper. Most regularizations can be dened by the addition of a penalty function J() to the corrected prediction error function g(z; F; ): ~ = argmin [^g(z; F; ) + J()] (7) where 0 is a regularization parameter. Let M CC and M Cy be the bias corrected versions of P i C z(x i )Cz(x i ) T and P i C z(x i )y i respectively, obtained by either the bootstrap, cross-validation or generalized cross-validation as described earlier. We consider two choices for J(). Rather than shrink towards zero (as in ridge regression) it seems somewhat more natural to shrink ^ towards (=K; =K;...=K) T, since we might put prior weight =K on each model t c k z. To regularize in this way, we choose J() = jj? (=K)jj 2, leading to the estimate ~ = (M CC + I)? [M Cy + (=K)]: (8) We could choose by another layer of cross-validation or bootstrapping, but a xed choice is more attractive computationally. After some experimentation we found that a good choice the value of such that the Euclidean distance between ^ and (=K;...=K) T is reduced by a xed factor (75%) Another regularization forces P ^i = ; this can be achieved by choosing J() = T? leading to ~ = M? CC [M Cy? ] (9) where is chosen so that ~ T =. To generalize both of these, one might consider estimators of the form (M CC + I)? [M Cy + 2 ] and use a layer of cross-validation or bootstrapping to estimate the best values of and 2. This is very computationally intensive, and we did not try it in this study. 8

Still another form of regularization is a non-negativity constraint, suggested by Breiman (993). An algorithm for least squares regression under the constraint k 0; k = ; 2...K is given in Lawson and Hanson (97), and this can be used to constrain the weights in the stacking and GCV procedures. To apply this procedure in general, we minimize ^g(z; F; ) (equation 0) under the constraint k 0; k = ; 2...K leading to ~ = argmin ( T M CC? 2M Cy ) (20) This problem can be solved by a simple modication of the algorithm given in Lawson and Hansen (97). Finally, we consider a simple combination method that is somewhat dierent in spirit than other proposed methods, but which constrains ~ k 0 and ~ T = : We let ~ k be the relative performance of the kth model. For instance, one might use ~ k = L(y; ^ k ; c k z P ) j L(y; ^ j ; c k z) where L(y; ^ k ; c k z ) is the maximized likelihood for model k: For a normal model ~ k = P ^2?n=2 k j ^2?n=2 j where ^ k 2 is the mean residual error. The estimator can also be motivated as an \estimated posterior mean" where one assumes a uniform prior assigning mass =K to each model as we remark below. We replace ^ k 2, the resubstitution estimate of prediction error for model k, with a K-fold cross-validation estimate of prediction error. Remark D. The relative t estimator and the other combination estimators of the linear form KX k= ; k c k z(x) (2) can be related to the Bayesian formulation for the prediction problem. For instance, the predictive mean of an a new observation Y 0 with a predictor vector X 0 can be expressed as E(Y 0 jz; X 0 ) = Z E(Y 0 jx 0 ; ; M k ; )p(; ; M k jz)dddm k where M k represents the kth component model and = ( ; k ) represents the parameters corresponding to all component models. The predictive mean can be re-expressed as 9

E(Y 0 jz; X 0 ) = = Z Z " X k Z Z " X k E(Y 0 jx 0 ; ; M k ; )p(m k j; ; z) C(X 0 ; ; M k )p(m k j; ; z) # # p(; jz)dd p(; jz)dd where C(X 0 ; ; M k ) represents the expected output for model M k given (and ) at X 0. This formulation also motivates the non-negative weights (and weights that sum to one) for the component model outputs. Note, the choice of independent normal priors for i corresponds to the ridge type shrinkage and a prior consisting of point masses on the coordinate unit vectors corresponds to the relative t weights. 6 Examples 6. Introduction In the following examples we compare the combination methods described earlier. Throughout we use 0-fold cross-validation, and B = 0 bootstrap samples. Ten-fold cross-validation was found to be superior to leave-one out crossvalidation by Breiman (993), and it is computationally much less demanding. Estimated model error (f(x i )? ^f(x i )) 2 is reported; ^f(xi ) is the estimated regression function and f(x i ) is the true regression function evaluated the observed predictor values x i : Twenty-ve Monte Carlo simulations are used in each example, and the Monte-Carlo means and standard deviations of the mean are reported. 6.2 Combining linear regression models In this example, which is modied from Breiman (993), we investigate combinations of best subset regression models. The predictor variables X ; ; X 30 were independent standard normal random variables and the response was generated from the linear model Y = X + + m X 30 + where is a standard normal random variable. The m were calculated by m = m, where m are dened in clusters: m = (h? jm? 7j) 2 Ifjm? 7j < hg + 0

(h? jm? j) 2 Ifjm? j < hg + (h? jm? 2j) 2 Ifjm? 2j < hg: The constant was determined so that the signal to noise ratio was approximately equal to one. Each simulated data set consisted of 60 observations. Two values h = and h = were considered; the case h = corresponds to a model with three large non-zero coecients and h = corresponds to a model with many small coecients. We report the average model errors in Tables and 2. As expected, for the uncombined methods, ridge regression performs well when there are many small coecients and best subset regression is the clear choice when there were a few large coecients. Most of the combination methods without regularization do not yield smaller model errors than standard methods; the cross-validation combination method and generalized cross-validation method give substantially larger model errors than ordinary least squares, ridge regression or best subsets. Only, the bootstrap method for the three large coecient model yields smaller model errors than ordinary least squares and ridge regression. However, regularization substantially reduces the model error of the combination methods. The non-negative estimators seem to perform as well as, and sometimes far better than, the shrinkage estimators and the estimators constrained to sum to one. The bootstrap and the cross validation combination methods with non-negativity constraints yield smaller average model errors than ridge regression and ordinary least squares for both the many small coecient model and the three large coecient model and yield results very close to best subsets for the three coecient model. The relative t weights seem to perform well relative to the other combination methods for the model with three large coecients, but yield somewhat larger errors than some of the combination methods with regularization for the model with many small coecients. 6.3 Combining tree-based regression models Tree-based regression procedures typically grow a large model that overts the data and then prune the tree and select the model, among a nested sequence of models or sub-trees, that minimizes an estimate of prediction error. We consider the sequence of models derived by the cost-complexity pruning algorithm of the Classication and Regression Tree (CART) algorithm (Breiman, Friedman, Olshen and Stone, 98). The cost-complexity measure of tree performance is R (T ) = X h2 ~ T R(h) + [ # of terminal nodes in T ];

Table : Average model errors (and standard errors) for Example. Many weak coecients Method Regularization None Non-negativity Shrinkage Sum to one Least squares 29.3 (.2) - - - Ridge (by CV) 25.2 (.5) - - - Best Subset (by CV) 33.3 (2.2) - - - Relative Fit Weights (by CV) 27.6 (.5) - - - Combination by least squares 29.3 (.2) 29.3 (.2) 29.0 (.2) 29.3 (.2) Combination by CV 69.8 (5.9) 2.0 (.6) 9. (.2) 66.7 (5.7) Combination by GCV 60.0 (2.6) 22. (.0) 29. (.) 30.9 (.5) Combination by bootstrap 32.7 (5.6) 2.0 (3.0) 22.2 (3.) 23.7 (2.9) Table 2: Average model errors (and standard errors) for Example. Three large coecients Method Regularization None Non-negativity Shrinkage Sum to one Least squares 29.3 (.2) - - - Ridge (by CV) 2. (0.9) - - - Best Subset (by CV) 8.5 (.5) - - - Relative Fit Weights (by CV) 8.7 (.5) - - - Combination by least squares 29.3 (.2) 29.3 (.2) 28.9 (.2) 29.3 (.2) Combination by CV 5.3 (3.6) 9.5 (.) 28. (2.7) 36.7 (2.7) Combination by GCV 52.5 (8.2).0 (.0) 27.0 (.) 29.5 (.) Combination by bootstrap 6.5 (2.9) 0. (.0).3 (.) 2.6 (.2) 2

Table 3: Average model error (and standard errors) for regression tree combinations Method Regularization None Non-negativity shrinkage Unpruned tree 92.9 (2.5) - - Best tree by CV 68. (2.6) - - Combination by Relative Fit 66.2 (2.6) - - Recursive Shrinkage 57.8 (.5) - - Combination by bootstrap 88.5 (5.0) 60. (.8) 68.8 (2.5) where is a penalty per terminal node, R(h) is the resubstitution estimate of prediction error for node h and ~ T are the terminal nodes in the tree. For any value of the cost complexity pruning algorithm can eciently obtain the sub-tree of T that minimizes R (T 0 ) over all subtrees T 0 of T: We apply the combination methods to the sequence of optimally pruned sub-trees for 0 < : For this example, we assume predictors X ; ; X 0 are independent standard normal distributed, f = IfX > 0g + IfX 2 > 0g + IfX 2 > 0gfX > 0g + X 3 and add noise generated from the standard normal distribution. Data sets of size 250 observations were generated. The bootstrap combination method with ridge regression, and the tree selected by a 0-fold cross validation estimate of prediction error yielded similar model errors. The bootstrap combination method with the non-negativity constraint yielded slightly larger errors on these data sets than a recursive shrinkage method for tree models (Hastie and Pregibon, 99, Clark and Pregibon, 992). However, both of these methods, for this example, give substantially smaller errors than the standard method which selects the best tree by 0-fold cross validation. We consider the results from one typical simulated data set in more detail. The unpruned tree and the best tree selected by 0-fold cross validation are given in Figures and 2. The bootstrap combination method with non-negative coecients yielded an estimate of with three non-zero coecients ^ T = (0; :093; :396; 0; 0; 0; 0; 0; 0; 0; 0; 0; 0; 0; 0; 0; 0; 0; 0; :8; 0); corresponding to pruned sub-trees of size 2, 3 and 20 terminal nodes. Note, the resulting combination model can be represented by the unpruned tree with modied estimates for the terminal nodes. The three tree models corresponding the non-negative coecients are given in Figure 3. 3

x.2<.55 x.3<.0 x.<.005 x.<-.63 x.9<-.9 x.3<-. x.3<.3 x.9<.6 x.3<-.3.2 2..36 x.0<.62 x.5<.0x.<.8 x.7<.23 -.33. x.9<-.8 x.6<-.5.88 x.<.78 3.. 3. -.26.58.5 x.3<.59.2.9 x.3<-.62 2.0 x.9<.2 x.9<.7.38.2 2. 3.0.6.85 Figure : The unpruned tree for a typical realization x.2<0.55 x.3<.0 x.<0.005 0.6.70.0 3.00 Figure 2: Pruned sub-tree selected by 0-fold cross validation for a typical realization

Tree Tree 2 Tree 3 Figure 3: Sub-trees with positive weight for a typical realization 5

Interpretation of the non-negative coecient model is aided by using an analysis of variance decomposition to represent the eect branches in the tree. Let ^(T j ) be the tted values for the tree corresponding to the jth non-zero coecient. The tted values of combination model can be expressed as ^f = :973^(T ) + :880 [^(T 2 )? ^(T )] + :8 [^(T 3 )? ^(T 2 )] + 0 ^(T full )? ^(T 3 ) (22) where T full is the unpruned tree. Therefore, one can interpret the combination with non-negative constraints as an increasing shrinkage of the eects in the lower branches of the tree. This explains, in part, why the performance of this combination method and the recursive shrinkage method were similar on average. Note, for data sets which yield some shrinkage factors close to 0 in a decomposition analogous to (22), one could set those factors to zero, eectively pruning the tree, to obtain a more interpretable model. 7 Classication problems In a classication problem, the outcome is not continuous but rather falls into one of J outcome classes. We can view this as a regression problem with a multivariate J-valued response y having a one in the jth position if the observation falls in class j and zero otherwise. Most classication procedures provide estimated class probabilities functions ^p(x) = (^p (x);... ^p J (x)) T. Given K such functions ^p k (x) = (^p k(x);... ^pk J (x))t, k = ; 2;...K, it seems reasonable to apply the combination strategies either to the probabilities themselves or their log ratios. That is, we take c kj z (x), the jth component of c k z(x), to be either or c kj z (x) = ^pk j (x) (23) c kj z (x) = log ^pk j (x) ^p k J (x) (2) In our experiments, we found that the use of probabilities (23) gives better classication results and is more interpretable. Let ^P(x) = (^p (x); ^p 2 (x);... ^p J (x); ^p2 (x); ^p2 2 (x);...)t, a vector of length JK. Then the combination estimator has the form j (x) = T j ^P(x); j = ; 2;...J; (25) where j is a JK-vector of weights. We then apply the straightforward multivariate analogues of the procedures described earlier. The population regression problem () becomes a multi-response regression, producing vector-valued 6

x2-2 - 0 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 3 3 3 33 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3-2 0 2 6 8 x Figure : Typical realization for example 2. estimates ^ T j. The combination estimators are computed in an analogous fashion to the regression case. Finally, given the combination estimators ^ j (x) = ^ T j ^P(x), the estimated class probabilities are taken to be ^ j (x)= P j ^ j (x). A non-negativity constraint for the weights is especially attractive here, because it ensures that these estimated probabilities lie between 0 and. 7. Example 2.: combining linear discriminant analysis and nearest neighbours In this example the data are two dimensional, with classes. A typical data realization is shown in gure. Classes, 2 and 3 are standard independent normal, centered at (0,0), (,) and (6,0) respectively. Class was generated as standard independent normal with mean (6,0), conditional on :5 x 2 +x 2 2 8. There were 0 observations in each class. We consider combination of linear discriminant analysis (LDA) and -nearest neighbour method. The idea is that LDA should work best for classes and 2, but not for classes 3 and. Nearest neighbour should work moderately well for 7

Table : Average % test error (and standard deviations) for Example 3. Method Regularization None Non-negativity LDA 36.3 (.6) - -nearest neighbour 27.5 (.87) - Combination by least squares 26. (.9) 25.2 (.90) Combination by bootstrap 27.0 (.95) 25.7 (.95) Table 5: Average combination weights for Example 3. LDA -nearest neighbour Class 2 3 2 3 0.75 0.00 0.00 0.00 0.3 0.00 0.00 0.00 2 0.00 0.80 0.00 0.00 0.00 0.29 0.00 0.00 3 0.00 0.00 0.02 0.07 0.00 0.00 0.95 0.00 0.00 0.00 0.0 0.0 0.00 0.00 0.00.3 classes and 2, and much better than LDA for classes 3 and. By combining the two, we might be able to improve on both of them. To proceed, we require from each method, a set of estimated class probabilities for each observation. We used the estimated Gaussian probabilities for LDA; for -nearest neighbour P we estimated the class j probability for feature vector x by exp(?d 2 J (x; j))= j= exp(?d2 (x; j)), where d 2 (x; j) is the squared distance from x to the nearest neighbour in class j. The results of 0 simulations of the combined LDA/nearest neighbour procedure are shown in Table We see that simple least squares combination, or regularized combination by bootstrap, slightly improves upon the LDA and -nearest neighbour rules. More interesting are the estimated combination weights produced by the nonnegativity constraint. Table 5 shows the average weights from the combination by bootstrap (the combination by least squares weights are very similar). LDA gets higher weight for classes and 2, while -nearest neighbour is used almost exclusively for classes 3 and, where the structure is highly nonlinear. In general, this procedure might prove to be useful in determining which outcome classes are linearly separable, and which are not. 8

8 More general combination schemes More general combination schemes would allow the k s to vary with x: KX k (x)c k z (x): (26) A special case of this would allow k to vary only with c k z, that is KX k (c k z (x))ck z (x): (27) Both of these models fall into the category of the \varying coecient" model discussed in Hastie and Tibshirani (993). The diculty in general is how to estimate the functions k (); this might be easier in model (27) than in model (26) since the c k zs are all real-valued and are probably of lower dimension than x. One application of varying combination weights would be in scatterplot smoothing. Here each c k z(x) is a scatterplot smooth with a dierent smoothing parameter k. The simple combination P k c k z(x) gives another family of estimators that are typically outside of the original family of estimators indexed by. However, it would seem more promising to allow each k to be a function of x and hence allow local combination of the estimates. Requiring k (x) to be a smooth function of x would ensure that the combination estimator is also smooth. Potentially one could use the ideas here for more general parameter selection problems. Suppose we have a regression estimator denoted by z(x; ). The estimator is computed on the data set z, with an adjustable (tuning) parameter vector, and gives a prediction at x. We wish to estimate the value of giving the smallest prediction error g(z; F; ) = E 0F (Y 0? z(x 0 ; )) 2 : (28) Then we can apply the bootstrap technique of section 3 to estimate the bias in g(z; ^F; ) = P N i= (y i? z(x i ; ) 2. Using the notation of section 3, the biased corrected estimator has the form ^g(z; ^F; ) = N X i=(y i? z(x i ; ^)) 2 + ^( ^F; ) (29) where ^( ^F; ) = B h N P N i= (y i? z b(x i; )) 2? N P N i= (yb i? z b(xb i ; )) 2 i: We would then minimize ^g(z; ^F; ) over. This idea may only be useful if z(x; ) can be written as an explicit function of, so that ^g(z; ^F ; ) and its 9

derivatives can be easily computed. In this paper we have considered the linear estimator z(x; ) = P k c k z(x) for which the minimizer of ^g(z; ^F; ) can be explicitly derived by least squares. The variable kernel estimator of Lowe (993) is another example where this procedure could be applied: in fact, Lowe uses the cross-validation version of the above procedure to estimate the adjustable parameters. In many adaptive procedures, e.g. tree-based regression, the estimator z(x; ) is a complicated function of its tuning parameters, so that minimization of ^g(z; ^F ; ) would be quite dicult. 9 Discussion This investigation suggests that combining of estimators, used with some regularization, can be a useful tool both for improving prediction performance and for learning about the structure of the problem. We have derived and studied a number of procedures and found that cross-validation (stacking) and the bootstrap, used with a non-negativity constraint, seemed to work best. The bootstrap estimates seem to require far less regularization; the reason for this is not clear. An interesting question for further study is: how can we choose estimators for which combining will be eective? ACKNOWLEDGEMENTS We would like to thank Mike Escobar, Trevor Hastie and Georey Hinton for helpful discussions and suggestions. Both authors gratefully acknowledge the support of the Natural Science and Engineering Research Council of Canada. References Breiman, L. (993). Stacked regression. Technical report, Univ. of Cal, Berkeley. Breiman, L., Friedman, J., Olshen, R., and Stone, C. (98). Classication and Regression Trees. Wadsworth International Group. Clark L. and Pregibon D. (992). Statistical Models in S, chapter Tree-Based models. Wadsworth International Group. Efron, B. (983). Estimating the error rate of a prediction rule: some improvements on cross-validation. J. Amer. Statist. Assoc., 78:36{33. Efron, B. and Tibshirani, R. (993). An Introduction to the Bootstrap. Chapman and Hall, New York. Hastie, T.J. and Tibshirani, R.J. (993). Varying coecient models (with discussion). J. Royal Statist. Soc. B, 55:757{796. 20

Lawson, C.L. and Hanson, R.J. (97). Prentice-Hall: Englewood Clis, N.J. Solving Least Squares Problems. Lowe, David G. (993). Similarity metric learning for a variable kernel classier. Technical report, Dept. of Comp Sci, Univ. of British Columbia. McCullagh, P. and Tibshirani, R. (988). A simple adjustment for prole likelihoods. J. Royal. Statist. Soc. B., 52(2):325{3. Wolpert, D. (992). Stacked generalization. Neural Networks, 5:2{259. 2