Outlier detection and variable selection via difference based regression model and penalized regression

Size: px
Start display at page:

Download "Outlier detection and variable selection via difference based regression model and penalized regression"

Transcription

1 Journal of the Korean Data & Information Science Society 2018, 29(3), 한국데이터정보과학회지 Outlier detection and variable selection via difference based regression model and penalized regression InHae Choi 1 Chun Gun Park 2 Kyeong Eun Lee 3 13 Department of Statistics, Kyungpook National University 2 Department of Mathematics, Kyonggi University Received 3 May 2018, revised 22 May 2018, accepted 22 May 2018 Abstract This paper studies an efficient procedure for the outlier detection and variable selection problem in linear regression. The effect of outliers is added in linear regression as a mean shift parameter, nonzero or zero constant. To fit this mean shift model, most penalized regressions have used some adaptive penalties on the parameters to shrink most of the parameters to zero. Such penalized models do select the true variables well, but do not detect the outliers correctly. To overcome this problem, we first determine a group of possibly suspected outliers using difference-based regression model (DBRM) and add the group to the linear model as the parameters of the effect of each suspected outlier. Then, we perform outlier detection and variable selection ultaneously using Lasso regression or Elastic net regression for the linear regression with the effect term of each suspected outlier added. The proposed method is more efficient than the previous penalized regression. We compare the proposed procedure with other methods using a ulation study and apply this procedure to the real data. Keywords: Difference-based regression model, Elastic net, Lasso, outliers detection, variable selection. 1. Introduction Outliers are observations that significantly differ from the others and frequently occur in the collection of real data. They distort statistical inference; for instance, ordinary least squares estimator is very susceptible to outliers (Joo and Cho, 2016). To address this, many methods have studied on outlier detection or robust techniques. We consider the mean shift linear regression, y = β X 1 β 1 + X 2 β 2 + γ + ɛ, where X 1 is an n p 1 design matrix of relevant predictors, X 2 is a n p 2 design matrix of irrelevant predictors, β i is a parameter vector corresponding to X i, γ is the effect vector of outliers which consists of zero or nonzero This work was extracted from Master s thesis of InHae Choi at Kyungpook National University in December Master, Department of Statistics, Kyungpook National University, Daegu 41566, Korea. 2 Associate Professor, Department of Mathematics, Kyonggi University, Suwan 16227, Korea. 3 Corresponding author: Associate Professor, Department of Statistics, Kyungpook National University, Daegu 41566, Korea. artlee@knu.ac.kr

2 816 InHae Choi Chun Gun Park Kyeong Eun Lee constant and ɛ is an error vector (McCann and Welsch, 2007; She and Owen, 2011; Park and Kim, 2017). In this study, we focus on an efficient procedure for the outlier detection and variable selection in the above mean shift linear regression. Recently penalized regression methods such as Lasso (Tibshirani, 1996) and Elastic net regression (Zou and Hastie, 2005) are popular for variable selection. Also, a type of these penalized regressions can be used outlier detection. For example, She and Owen (2011) proposed a nonconvex penalty function on γ i s, the effect of the outliers. Such penalties do select the true variables well, but do not detect the outliers correctly. To overcome this problem, we first determine a group of possibly suspected outliers using difference-based regression model (DBRM) (Park and Kim, 2017; Park, 2018) and add the group to the linear model as the parameters of the effect of each suspected outlier. Then, we perform outlier detection and variable selection ultaneously using Lasso regression or Elastic net regression for the linear regression with the effect term of each suspected outlier added. The proposed method is more efficient than the previous penalized regression. In the literature, some methods such as Forward Search Algorithm (Hadi and Simono, 1993) directly detect outliers without estimating the mean function, and others such as least trimmed squares (LTS) (Rousseeuw and Leroy, 1987), thresholding based iteratively procedure (Θ-IPOD) (She and Owen, 2011) indirectly identify outliers using residuals from robust regression model. While the latter methods can detect outliers by estimating the mean trend function, our outlier detection method uses the difference-based regression model without estimating the mean trend function. The rest of this article is organized as follows. In Section 2, we introduce a differencebased regression model which is used for picking out a group of possibly suspected outliers and Lasso regression and Elastic net regression for selecting important variables. Also, we describe an efficient procedure or the outlier detection and variable selection via two penalized regressions and DBRM. In Section 3, we conduct a ulation study to compare our approach with other existing methods. In Section 4, we apply the proposed method to the consumption of petrol. Finally in Section 5, we provide some conclusions. 2. Methods In this section, we will briefly review DBRM (Park and Kim, 2017), Lasso regression (Tibshirani, 1996) and Elastic net regression (Zou and Hastie, 2005) and then propose our method for outlier detection and variable selection using DBRM and Lasso or Elastic net regression Difference based regression model Consider a multiple linear regression model without outlier y = 1 n β 0 + X 1 β 1 + X 2 β 2 + ɛ, where β 0 is an intercept, 1 n = [1, 1,..., 1] T is the n-dimensional unit vector, X a is the n p a matrix of rank p a, β a = [β a1,..., β apa ] T is the p a -vector of coefficient, a = 1, 2, E(ɛ) = 0 and V ar(ɛ) = σ 2 R with unknown correlation matrix R and unknown variance σ 2.

3 Outlier detection and variable selection via DBRM and penalized regression 817 For detecting outliers in linear regression model, the mean-shift model (Cook and Weisberg, 1982) has been used for a long time. So we consider the following mean shift model: w = y + γ + ɛ = 1 n β 0 + Xβ + γ + ɛ where X = [X 1, X 2 ], β = [β T 1, β T 2 ] T and p = p 1 + p 2. DBRM detects outliers using the following property - when the jth case is outlier, the absolute value of estimated intercept of the fitted model without the jth case is large. The DBRM can be written as: D (j) w = D (j) y + D (j) γ = 1 n 1 ( γ j ) + D (j) Xβ + A (j) γ + D (j) ɛ [ ] γj = [1 n 1 : D (j) X] + A β (j) γ + D (j) ɛ, j = 1,..., n (2.1) where ( γ j ) is the intercept for detecting outliers and D (j) and A (j) are the (n 1) n matrices as follows: [ ] [ ] I D (j) = j 1 1 j 1 0 (j 1),(n j) I and A 0 (n j),(j 1) 1 n j I (j) = j 1 0 j 1 0 (j 1),(n j). n j 0 (n j),(j 1) 0 n j I n j where I a is an a a identity matrix and O a,b is an a b null matrix. Since intercept ( γ j ) is highly influenced by outliers, we propose to determine outlier candidates using the intercepts based on the difference based regression model. Note that we are not interested in other parameters Lasso regression Tibshirani (1996) proposed Lasso Regression Model for regression shrinkage and selection. For any fixed non-negative λ, the Lasso estimate ˆβ L is defined by ˆβ L = arg min L(λ, β) β = arg min (y i β 0 β i=1 β j x ij ) 2 + λ β j. (2.2) Here, the larger λ, the closer β 1,..., β p are to zero. That is, the Lasso regression shrinks the coefficient estimates. But it is known to have there are some limitations. First, the Lasso method selects at most n variables before it saturates at high dimensional. Second, if there is a group of highly correlated variables, then the Lasso method tends to select one variable from a group and ignore the others (Zou and Hastie, 2005) Elastic net regression Zou and Hastie (2005) proposed Elastic net for regularization and variable selection to overcome these limitations of Lasso. For any fixed non-negative λ 1 and λ 2, the Elastic net

4 818 InHae Choi Chun Gun Park Kyeong Eun Lee estimate ˆβ E is defined by ˆβ E = arg min L(λ 1, λ 2, β) β = arg min (y i β 0 β i=1 β j x ij ) 2 + λ 1 βj 2 + λ 2 β j. (2.3) The Elastic net method allows sparsity and grouping effect together (Zou and Hastie, 2005) Algorithm We propose to use DBRM method to select outlier candidates and to perform outlier detection and variable selection ultaneously using Lasso regression or Elastic net method. This approach consists of the following steps: Step 1. Estimation of intercepts in DBRM. D (j) w = 1 n 1 ( γ j ) + D (j) Xβ + A (j) γ + D (j) ɛ Estimate intercepts, ˆγ j, j = 1,..., n and rewrite δ j = abs( ˆγ j ) Ascend them, δ (j), δ (1) δ (n). Step 2. Determination outlier candidates Set the percentage of outlier candidates Consider ones with large δ (j) value as outlier candidates Include outlier candidates indicator variables in the multiple linear model. Step 3. Outlier detection and variable selection using Lasso or Elastic net. w = Xβ + γ 1 z γ k z k + ɛ where z k is an indicator vector of outlier candidate Find Lasso estimates: (w i β 0 i=1 β j x ij 3.2. Find Elastic net estimates: k γ j z ij ) 2 + λ( β j + k γ j ). (w i β 0 β j x ij i=1 k k k γ j z ij ) 2 + λ 1 ( βj 2 + γj 2 ) + λ 2 ( β j + γ j ). j In step 3, we need to set the optimal λ for the Lasso regression and Elastic net. So, we use a method to obtain the optimal λ in a multiple regression model without outliers.

5 Outlier detection and variable selection via DBRM and penalized regression Simulation studies We conduct ulations to compare the performance of our method with the other existing methods. We compare the outlier detection with the following four methods and compare the variable selection with LAD Lasso : Least trimmed squares (LT S) (Rousseeuw and Leroy, 1987), LAD regression with the Lasso penalty (LAD Lasso ) (Wang et al., 2007), Hard thresholding (denoted by θ) based iteratively procedure for outlier detection (HIP OD) (She and Owen, 2011), Soft thresholding (denoted by θ) based iteratively procedure for outlier detection (SIP OD) (She and Owen, 2011) Simulation setting We consider three cases of samples sizes (n = 30, 100, 300) each with 10% outliers and p = n 10 covariates (p 1 = p 2 = p 2 ). Covariates are generated from multivariate normal distribution with mean 0 and covariance matrix Σ = {σ ij }, σ ij = ρ i j. We consider two cases of correlation of covariates (ρ = 0, 0.5) and two different errors distributions (N(0, 1), t df=2 ). So the total number of ulation cases is 12. Outlier locations and signs are randomly selected and outlier sizes are 8 or 10. We generate 100 data sets for each case. We set the number of outlier candidate group as 30% of the sample size Criteria Let n O be the number of true outliers, n D be the number of detected outliers, n CD be the number of correctly detected outliers, n ID be the number of incorrectly detected outliers, n IU be the number of incorrectly undetected outliers and n CU be the number of correctly undetected non-outliers. Detection Table 3.1 Outlier detection True Outlier Non-outlier Sum Outlier n CD n ID n D Non-outlier n IU n CU n UD Sum n O n NO n Two kinds of errors can occur in outlier detection: masking (a true outlier is not detected) and swamping (a non-outlier is detected as an outlier). We define relative frequencies of perfection, only masking, only swamping, masking and swamping and complete failure in order to show the strength of our proposed method as follows: - Relative frequency of perfect detection: P D = 1 I(n CD(s) = n O(s) ). n - Relative frequency of only-swamping with detection (overdetection): OS = 1 I(n IU(s) = 0, n ID(s) > 0, n CD(s) > 0). n

6 820 InHae Choi Chun Gun Park Kyeong Eun Lee - Relative frequency of only-masking with partial detection: OM = 1 I(n IU(s) > 0, n ID(s) = 0, n CD(s) > 0). n - Relative frequency of masking and swamping with partial detection: MS = 1 I(n IU(s) > 0, n ID(s) > 0, n CD(s) > 0). n - Relative frequency of complete failure relative frequency: CF = 1 I(n CD(s) = 0). n Finally, we use three criteria for variable selection. Let CS be the relative frequency of correct selection, CR be relative frequency of correct variable reduction and AN be the average number of selected variables: - CS : the relative frequency of correct selection: CS = 1 ( I {j : n ˆβ ) j,s 0} = {j : β j 0}. - CR : the relative frequency of correct variable reduction: CR = 1 ( I {j : n ˆβ ) j,s 0} {j : β j 0}. - AN : the average number of selected variables: 3.3. Simulation result AN = 1 n ( #{j : ˆβ ) j,s 0}. We show the results of the 12 ulation cases. Table 3.2 is the results of outliers detection. METHOD1 combines the Difference-Based Regression Model with the Lasso regression and METHOD2 combines Difference-Based Regression Model with the Elastic net regression. Regardless of sample size or other conditions, our method has comparable performance, mainly in detection power. And the larger the number of samples and the correlation between variables, the better the performance of detecting outlier. Although our performance is not superior to other methods, overall performance is comparable. Next, we show the result of the variable selection. Table 3.3 is the results of variable selection.

7 Outlier detection and variable selection via DBRM and penalized regression 821 Our method gets comparable results with other existing methods through ulation studies and it is superior in most cases. Although the number of variables selected is larger than the number of important variables, the rate of correct variable reduction is much better than the other methods. Table 3.2 Comparison of outlier detection ɛ ρ n Criteria LTS LAD Lasso HIPOD SIPOD METHOD1 METHOD2 PD OS n = 30 OM PD OS ρ = 0 n = 100 OM PD OS n = 300 OM N(0, 1) PD OS n = 30 OM PD OS ρ = 0.5 n = 100 OM PD OS n = 300 OM PD OS n = 30 OM MS CF PD OS t df=2 ρ = 0 n = 100 OM MS CF PD OS n = 300 OM MS CF

8 822 InHae Choi Chun Gun Park Kyeong Eun Lee Table 3.2 Continued ɛ ρ n Criteria LTS LAD Lasso HIPOD SIPOD METHOD1 METHOD2 PD OS n = 30 OM MS CF PD OS t df=2 ρ = 0.5 n = 100 OM MS CF PS OS n = 300 OM MS CF Table 3.3 Comparison of variable selection ɛ ρ n Criteria LAD Lasso METHOD1 METHOD2 CS n = 30 CR AN CS ρ = 0 n = 100 CR AN CS n = 300 CR N(0, 1) AN CS n = 30 CR AN CS ρ = 0.5 n = 100 CR AN CS n = 300 CR AN CS n = 30 CR AN CS t df=2 ρ = 0 n = 100 CR AN CS n = 300 CR AN

9 Outlier detection and variable selection via DBRM and penalized regression 823 Table 3.3 Continued ɛ ρ n Criteria LAD Lasso METHOD1 METHOD2 CS n = 30 CR AN CS t df=2 ρ = 0.5 n = 100 CR AN CS n = 300 CR AN Real data analysis We apply our procedure to the consumption of fuel. Because the consumption of fuel was measured in 48 states, there are 48 rows of data. The response variable (Y ) and four explanatory variables (X 1, X 2, X 3, and X 4 ) are as the followings: - Y = Consumption of fuel (millions of gallons). - X 1 = fuel tax (cents per gallon). - X 2 = Average income (dollars). - X 3 = Paved Highways (miles). - X 4 = Proportion of population with driver s licenses. Looking at Figure 4.1, we see that the 40th observation is an outlier. Table 4.1 is the result of outlier detection in fuel consumption data. Among the methods, HIPOD, SIPOD, METHOD1 and METHOD2 identify that the 40th observation is an outlier. But HIPOD, METHOD1 and METHOD2 tend to detect too many outliers. Table 4.1 The result of outlier detection in fuel consumption data Method Number of Outliers LTS 0 { } LAD Lasso 0 { } HIPOD 22 Outlier index set {5, 8, 9, 11, 15, 16, 18, 19, 20, 22, 23, 24, 31, 33, 34, 38, 39, 40, 42, 43, 44, 45} SIPOD 2 {18, 40} METHOD1 10 {5, 11, 18, 19, 33, 38, 40, 42, 44, 45} METHOD2 14 {5, 11, 15, 18, 19, 20, 33, 36, 38, 39, 40, 42, 44, 45} Table 4.2 is the results of variable selection in fuel consumption data. LAD-Lasso and METHOD1 consider X 3 as the unimportant variable and select 3 variables, METHOD2 does not find the unimportant variable and selects all 4 variables.

10 824 InHae Choi Chun Gun Park Kyeong Eun Lee Figure 4.1 Residual plot Table 4.2 The result of variable selection in fuel consumption data Method Number of selected variables Unselected variable LAD Lasso 3 X 3 METHOD1 3 X 3 METHOD Conclusion In this paper, using the properties of intercept estimator in a difference-based regression model, we propose a ultaneous procedure of an outlier detection and variable selection which is more efficient and pler than using each procedure separately. Our procedure does not require us to estimate mean function. Instead, we determine the candidate group of outliers using the intercept estimates in the difference-based regression model and add the candidate group to the model. Then, we use outlier detection and variable selection using Lasso regression and Elastic net regression. Our procedure gets results comparable with other existing methods through ulation studies, and it is superior in most cases. We apply our procedure to real data. For future study, we can extend to the nonparametric regression model for outliers detection. References Cook, R.D. and Weisberg, S. (1982). Residuals and influence in regression, Chapman & Hall, London, UK. Hadi, A. S. and Simonoff, J. S. (1993). Procedures for the identification of multiple outliers in linear models. Journal of the American Statistical Association, 88, Joo, Y. S. and Cho, G-Y. (2016). Outlier detection and treatment in industrial sampling survey. Journal of the Korean Data & Information Science Society, 27,

11 Outlier detection and variable selection via DBRM and penalized regression 825 McCann, L., and Welsch, R. E. (2007). Robust variable selection using least angle regression and elemental set sampling. Computational Statistics & Data Analysis, 52, Park, C. G. (2018). Distinction of an outlier(s) using difference based regression models. Journal of the Korean Data & Information Science Society, 29, Park, C. G. and Kim, I. (2017). Outlier detection using difference based regression Model. Manuscript. Rousseeuw, P. J., and Leroy, A. M. (1987). Robust regression and outlier detection, Wiley Series in Probability and Mathematical Statistics: Applied Probability and Statistics, John Wiley & Sons Inc, New York. She, Y. and Owen, A. B. (2011). Outlier detection using nonconvex penalized regression. Journal of the American Statistical Association, 106, Tibshirani, R. (1996). Regression shrinkage and selection via the Lasso. Journal of the Royal Statistical Society. Series B, 58, Wang, H., Li, G. and Jiang, G. (2007). Robust regression shrinkage and consistent variable selection through the lad-lasso. Journal of Business & Economic Statistics, 25, Zou, H and Hastie, T. (2005). Regularization and variable selection via the Elastic net. Journal of the Royal Statistical Society. Series B, 67,

PENALIZED PRINCIPAL COMPONENT REGRESSION. Ayanna Byrd. (Under the direction of Cheolwoo Park) Abstract

PENALIZED PRINCIPAL COMPONENT REGRESSION. Ayanna Byrd. (Under the direction of Cheolwoo Park) Abstract PENALIZED PRINCIPAL COMPONENT REGRESSION by Ayanna Byrd (Under the direction of Cheolwoo Park) Abstract When using linear regression problems, an unbiased estimate is produced by the Ordinary Least Squares.

More information

Effect of outliers on the variable selection by the regularized regression

Effect of outliers on the variable selection by the regularized regression Communications for Statistical Applications and Methods 2018, Vol. 25, No. 2, 235 243 https://doi.org/10.29220/csam.2018.25.2.235 Print ISSN 2287-7843 / Online ISSN 2383-4757 Effect of outliers on the

More information

Variable Selection in Restricted Linear Regression Models. Y. Tuaç 1 and O. Arslan 1

Variable Selection in Restricted Linear Regression Models. Y. Tuaç 1 and O. Arslan 1 Variable Selection in Restricted Linear Regression Models Y. Tuaç 1 and O. Arslan 1 Ankara University, Faculty of Science, Department of Statistics, 06100 Ankara/Turkey ytuac@ankara.edu.tr, oarslan@ankara.edu.tr

More information

High-dimensional Ordinary Least-squares Projection for Screening Variables

High-dimensional Ordinary Least-squares Projection for Screening Variables 1 / 38 High-dimensional Ordinary Least-squares Projection for Screening Variables Chenlei Leng Joint with Xiangyu Wang (Duke) Conference on Nonparametric Statistics for Big Data and Celebration to Honor

More information

Linear Model Selection and Regularization

Linear Model Selection and Regularization Linear Model Selection and Regularization Recall the linear model Y = β 0 + β 1 X 1 + + β p X p + ɛ. In the lectures that follow, we consider some approaches for extending the linear model framework. In

More information

TECHNICAL REPORT NO. 1091r. A Note on the Lasso and Related Procedures in Model Selection

TECHNICAL REPORT NO. 1091r. A Note on the Lasso and Related Procedures in Model Selection DEPARTMENT OF STATISTICS University of Wisconsin 1210 West Dayton St. Madison, WI 53706 TECHNICAL REPORT NO. 1091r April 2004, Revised December 2004 A Note on the Lasso and Related Procedures in Model

More information

Regularization and Variable Selection via the Elastic Net

Regularization and Variable Selection via the Elastic Net p. 1/1 Regularization and Variable Selection via the Elastic Net Hui Zou and Trevor Hastie Journal of Royal Statistical Society, B, 2005 Presenter: Minhua Chen, Nov. 07, 2008 p. 2/1 Agenda Introduction

More information

Robust Variable Selection Methods for Grouped Data. Kristin Lee Seamon Lilly

Robust Variable Selection Methods for Grouped Data. Kristin Lee Seamon Lilly Robust Variable Selection Methods for Grouped Data by Kristin Lee Seamon Lilly A dissertation submitted to the Graduate Faculty of Auburn University in partial fulfillment of the requirements for the Degree

More information

Regression Analysis for Data Containing Outliers and High Leverage Points

Regression Analysis for Data Containing Outliers and High Leverage Points Alabama Journal of Mathematics 39 (2015) ISSN 2373-0404 Regression Analysis for Data Containing Outliers and High Leverage Points Asim Kumer Dey Department of Mathematics Lamar University Md. Amir Hossain

More information

Sparse regression. Optimization-Based Data Analysis. Carlos Fernandez-Granda

Sparse regression. Optimization-Based Data Analysis.   Carlos Fernandez-Granda Sparse regression Optimization-Based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_spring16 Carlos Fernandez-Granda 3/28/2016 Regression Least-squares regression Example: Global warming Logistic

More information

Smoothly Clipped Absolute Deviation (SCAD) for Correlated Variables

Smoothly Clipped Absolute Deviation (SCAD) for Correlated Variables Smoothly Clipped Absolute Deviation (SCAD) for Correlated Variables LIB-MA, FSSM Cadi Ayyad University (Morocco) COMPSTAT 2010 Paris, August 22-27, 2010 Motivations Fan and Li (2001), Zou and Li (2008)

More information

ESL Chap3. Some extensions of lasso

ESL Chap3. Some extensions of lasso ESL Chap3 Some extensions of lasso 1 Outline Consistency of lasso for model selection Adaptive lasso Elastic net Group lasso 2 Consistency of lasso for model selection A number of authors have studied

More information

Machine Learning for OR & FE

Machine Learning for OR & FE Machine Learning for OR & FE Regression II: Regularization and Shrinkage Methods Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com

More information

Sparse Linear Models (10/7/13)

Sparse Linear Models (10/7/13) STA56: Probabilistic machine learning Sparse Linear Models (0/7/) Lecturer: Barbara Engelhardt Scribes: Jiaji Huang, Xin Jiang, Albert Oh Sparsity Sparsity has been a hot topic in statistics and machine

More information

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28 Sparsity Models Tong Zhang Rutgers University T. Zhang (Rutgers) Sparsity Models 1 / 28 Topics Standard sparse regression model algorithms: convex relaxation and greedy algorithm sparse recovery analysis:

More information

Consistent Group Identification and Variable Selection in Regression with Correlated Predictors

Consistent Group Identification and Variable Selection in Regression with Correlated Predictors Consistent Group Identification and Variable Selection in Regression with Correlated Predictors Dhruv B. Sharma, Howard D. Bondell and Hao Helen Zhang Abstract Statistical procedures for variable selection

More information

KANSAS STATE UNIVERSITY Manhattan, Kansas

KANSAS STATE UNIVERSITY Manhattan, Kansas ROBUST MIXTURE MODELING by CHUN YU M.S., Kansas State University, 2008 AN ABSTRACT OF A DISSERTATION submitted in partial fulfillment of the requirements for the degree Doctor of Philosophy Department

More information

The lasso. Patrick Breheny. February 15. The lasso Convex optimization Soft thresholding

The lasso. Patrick Breheny. February 15. The lasso Convex optimization Soft thresholding Patrick Breheny February 15 Patrick Breheny High-Dimensional Data Analysis (BIOS 7600) 1/24 Introduction Last week, we introduced penalized regression and discussed ridge regression, in which the penalty

More information

Outlier Detection via Feature Selection Algorithms in

Outlier Detection via Feature Selection Algorithms in Int. Statistical Inst.: Proc. 58th World Statistical Congress, 2011, Dublin (Session CPS032) p.4638 Outlier Detection via Feature Selection Algorithms in Covariance Estimation Menjoge, Rajiv S. M.I.T.,

More information

Different types of regression: Linear, Lasso, Ridge, Elastic net, Ro

Different types of regression: Linear, Lasso, Ridge, Elastic net, Ro Different types of regression: Linear, Lasso, Ridge, Elastic net, Robust and K-neighbors Faculty of Mathematics, Informatics and Mechanics, University of Warsaw 04.10.2009 Introduction We are given a linear

More information

Variable Selection for Highly Correlated Predictors

Variable Selection for Highly Correlated Predictors Variable Selection for Highly Correlated Predictors Fei Xue and Annie Qu Department of Statistics, University of Illinois at Urbana-Champaign WHOA-PSI, Aug, 2017 St. Louis, Missouri 1 / 30 Background Variable

More information

Linear Regression (9/11/13)

Linear Regression (9/11/13) STA561: Probabilistic machine learning Linear Regression (9/11/13) Lecturer: Barbara Engelhardt Scribes: Zachary Abzug, Mike Gloudemans, Zhuosheng Gu, Zhao Song 1 Why use linear regression? Figure 1: Scatter

More information

Sparse PCA with applications in finance

Sparse PCA with applications in finance Sparse PCA with applications in finance A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley Available online at www.princeton.edu/~aspremon 1 Introduction

More information

Iterative Selection Using Orthogonal Regression Techniques

Iterative Selection Using Orthogonal Regression Techniques Iterative Selection Using Orthogonal Regression Techniques Bradley Turnbull 1, Subhashis Ghosal 1 and Hao Helen Zhang 2 1 Department of Statistics, North Carolina State University, Raleigh, NC, USA 2 Department

More information

MSA220/MVE440 Statistical Learning for Big Data

MSA220/MVE440 Statistical Learning for Big Data MSA220/MVE440 Statistical Learning for Big Data Lecture 9-10 - High-dimensional regression Rebecka Jörnsten Mathematical Sciences University of Gothenburg and Chalmers University of Technology Recap from

More information

MS-C1620 Statistical inference

MS-C1620 Statistical inference MS-C1620 Statistical inference 10 Linear regression III Joni Virta Department of Mathematics and Systems Analysis School of Science Aalto University Academic year 2018 2019 Period III - IV 1 / 32 Contents

More information

Linear model selection and regularization

Linear model selection and regularization Linear model selection and regularization Problems with linear regression with least square 1. Prediction Accuracy: linear regression has low bias but suffer from high variance, especially when n p. It

More information

A Survey of L 1. Regression. Céline Cunen, 20/10/2014. Vidaurre, Bielza and Larranaga (2013)

A Survey of L 1. Regression. Céline Cunen, 20/10/2014. Vidaurre, Bielza and Larranaga (2013) A Survey of L 1 Regression Vidaurre, Bielza and Larranaga (2013) Céline Cunen, 20/10/2014 Outline of article 1.Introduction 2.The Lasso for Linear Regression a) Notation and Main Concepts b) Statistical

More information

Bayesian Grouped Horseshoe Regression with Application to Additive Models

Bayesian Grouped Horseshoe Regression with Application to Additive Models Bayesian Grouped Horseshoe Regression with Application to Additive Models Zemei Xu, Daniel F. Schmidt, Enes Makalic, Guoqi Qian, and John L. Hopper Centre for Epidemiology and Biostatistics, Melbourne

More information

Linear regression methods

Linear regression methods Linear regression methods Most of our intuition about statistical methods stem from linear regression. For observations i = 1,..., n, the model is Y i = p X ij β j + ε i, j=1 where Y i is the response

More information

A Short Introduction to the Lasso Methodology

A Short Introduction to the Lasso Methodology A Short Introduction to the Lasso Methodology Michael Gutmann sites.google.com/site/michaelgutmann University of Helsinki Aalto University Helsinki Institute for Information Technology March 9, 2016 Michael

More information

Generalized Elastic Net Regression

Generalized Elastic Net Regression Abstract Generalized Elastic Net Regression Geoffroy MOURET Jean-Jules BRAULT Vahid PARTOVINIA This work presents a variation of the elastic net penalization method. We propose applying a combined l 1

More information

Outlier Detection and Robust Estimation in Nonparametric Regression

Outlier Detection and Robust Estimation in Nonparametric Regression Outlier Detection and Robust Estimation in Nonparametric Regression Dehan Kong Howard Bondell Weining Shen University of Toronto University of Melbourne University of California, Irvine Abstract This paper

More information

Prediction & Feature Selection in GLM

Prediction & Feature Selection in GLM Tarigan Statistical Consulting & Coaching statistical-coaching.ch Doctoral Program in Computer Science of the Universities of Fribourg, Geneva, Lausanne, Neuchâtel, Bern and the EPFL Hands-on Data Analysis

More information

CHAPTER 5. Outlier Detection in Multivariate Data

CHAPTER 5. Outlier Detection in Multivariate Data CHAPTER 5 Outlier Detection in Multivariate Data 5.1 Introduction Multivariate outlier detection is the important task of statistical analysis of multivariate data. Many methods have been proposed for

More information

A Significance Test for the Lasso

A Significance Test for the Lasso A Significance Test for the Lasso Lockhart R, Taylor J, Tibshirani R, and Tibshirani R Ashley Petersen May 14, 2013 1 Last time Problem: Many clinical covariates which are important to a certain medical

More information

Machine Learning for Economists: Part 4 Shrinkage and Sparsity

Machine Learning for Economists: Part 4 Shrinkage and Sparsity Machine Learning for Economists: Part 4 Shrinkage and Sparsity Michal Andrle International Monetary Fund Washington, D.C., October, 2018 Disclaimer #1: The views expressed herein are those of the authors

More information

Robust Variable Selection Through MAVE

Robust Variable Selection Through MAVE Robust Variable Selection Through MAVE Weixin Yao and Qin Wang Abstract Dimension reduction and variable selection play important roles in high dimensional data analysis. Wang and Yin (2008) proposed sparse

More information

Selection of Smoothing Parameter for One-Step Sparse Estimates with L q Penalty

Selection of Smoothing Parameter for One-Step Sparse Estimates with L q Penalty Journal of Data Science 9(2011), 549-564 Selection of Smoothing Parameter for One-Step Sparse Estimates with L q Penalty Masaru Kanba and Kanta Naito Shimane University Abstract: This paper discusses the

More information

Chris Fraley and Daniel Percival. August 22, 2008, revised May 14, 2010

Chris Fraley and Daniel Percival. August 22, 2008, revised May 14, 2010 Model-Averaged l 1 Regularization using Markov Chain Monte Carlo Model Composition Technical Report No. 541 Department of Statistics, University of Washington Chris Fraley and Daniel Percival August 22,

More information

UNIVERSITY OF MASSACHUSETTS. Department of Mathematics and Statistics. Basic Exam - Applied Statistics. Tuesday, January 17, 2017

UNIVERSITY OF MASSACHUSETTS. Department of Mathematics and Statistics. Basic Exam - Applied Statistics. Tuesday, January 17, 2017 UNIVERSITY OF MASSACHUSETTS Department of Mathematics and Statistics Basic Exam - Applied Statistics Tuesday, January 17, 2017 Work all problems 60 points are needed to pass at the Masters Level and 75

More information

Biostatistics Advanced Methods in Biostatistics IV

Biostatistics Advanced Methods in Biostatistics IV Biostatistics 140.754 Advanced Methods in Biostatistics IV Jeffrey Leek Assistant Professor Department of Biostatistics jleek@jhsph.edu Lecture 12 1 / 36 Tip + Paper Tip: As a statistician the results

More information

Or How to select variables Using Bayesian LASSO

Or How to select variables Using Bayesian LASSO Or How to select variables Using Bayesian LASSO x 1 x 2 x 3 x 4 Or How to select variables Using Bayesian LASSO x 1 x 2 x 3 x 4 Or How to select variables Using Bayesian LASSO On Bayesian Variable Selection

More information

Lecture 14: Shrinkage

Lecture 14: Shrinkage Lecture 14: Shrinkage Reading: Section 6.2 STATS 202: Data mining and analysis October 27, 2017 1 / 19 Shrinkage methods The idea is to perform a linear regression, while regularizing or shrinking the

More information

Lecture 14: Variable Selection - Beyond LASSO

Lecture 14: Variable Selection - Beyond LASSO Fall, 2017 Extension of LASSO To achieve oracle properties, L q penalty with 0 < q < 1, SCAD penalty (Fan and Li 2001; Zhang et al. 2007). Adaptive LASSO (Zou 2006; Zhang and Lu 2007; Wang et al. 2007)

More information

An Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models

An Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models Proceedings 59th ISI World Statistics Congress, 25-30 August 2013, Hong Kong (Session CPS023) p.3938 An Algorithm for Bayesian Variable Selection in High-dimensional Generalized Linear Models Vitara Pungpapong

More information

arxiv: v3 [stat.ml] 14 Apr 2016

arxiv: v3 [stat.ml] 14 Apr 2016 arxiv:1307.0048v3 [stat.ml] 14 Apr 2016 Simple one-pass algorithm for penalized linear regression with cross-validation on MapReduce Kun Yang April 15, 2016 Abstract In this paper, we propose a one-pass

More information

Robust model selection criteria for robust S and LT S estimators

Robust model selection criteria for robust S and LT S estimators Hacettepe Journal of Mathematics and Statistics Volume 45 (1) (2016), 153 164 Robust model selection criteria for robust S and LT S estimators Meral Çetin Abstract Outliers and multi-collinearity often

More information

Shrinkage Methods: Ridge and Lasso

Shrinkage Methods: Ridge and Lasso Shrinkage Methods: Ridge and Lasso Jonathan Hersh 1 Chapman University, Argyros School of Business hersh@chapman.edu February 27, 2019 J.Hersh (Chapman) Ridge & Lasso February 27, 2019 1 / 43 1 Intro and

More information

Simultaneous regression shrinkage, variable selection, and supervised clustering of predictors with OSCAR

Simultaneous regression shrinkage, variable selection, and supervised clustering of predictors with OSCAR Simultaneous regression shrinkage, variable selection, and supervised clustering of predictors with OSCAR Howard D. Bondell and Brian J. Reich Department of Statistics, North Carolina State University,

More information

Forward Selection and Estimation in High Dimensional Single Index Models

Forward Selection and Estimation in High Dimensional Single Index Models Forward Selection and Estimation in High Dimensional Single Index Models Shikai Luo and Subhashis Ghosal North Carolina State University August 29, 2016 Abstract We propose a new variable selection and

More information

Package Grace. R topics documented: April 9, Type Package

Package Grace. R topics documented: April 9, Type Package Type Package Package Grace April 9, 2017 Title Graph-Constrained Estimation and Hypothesis Tests Version 0.5.3 Date 2017-4-8 Author Sen Zhao Maintainer Sen Zhao Description Use

More information

PENALIZING YOUR MODELS

PENALIZING YOUR MODELS PENALIZING YOUR MODELS AN OVERVIEW OF THE GENERALIZED REGRESSION PLATFORM Michael Crotty & Clay Barker Research Statisticians JMP Division, SAS Institute Copyr i g ht 2012, SAS Ins titut e Inc. All rights

More information

On High-Dimensional Cross-Validation

On High-Dimensional Cross-Validation On High-Dimensional Cross-Validation BY WEI-CHENG HSIAO Institute of Statistical Science, Academia Sinica, 128 Academia Road, Section 2, Nankang, Taipei 11529, Taiwan hsiaowc@stat.sinica.edu.tw 5 WEI-YING

More information

Single Index Quantile Regression for Heteroscedastic Data

Single Index Quantile Regression for Heteroscedastic Data Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University SMAC, November 6, 2015 E. Christou, M. G. Akritas (PSU) SIQR

More information

Analysis Methods for Supersaturated Design: Some Comparisons

Analysis Methods for Supersaturated Design: Some Comparisons Journal of Data Science 1(2003), 249-260 Analysis Methods for Supersaturated Design: Some Comparisons Runze Li 1 and Dennis K. J. Lin 2 The Pennsylvania State University Abstract: Supersaturated designs

More information

A Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression

A Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression A Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression Noah Simon Jerome Friedman Trevor Hastie November 5, 013 Abstract In this paper we purpose a blockwise descent

More information

ISyE 691 Data mining and analytics

ISyE 691 Data mining and analytics ISyE 691 Data mining and analytics Regression Instructor: Prof. Kaibo Liu Department of Industrial and Systems Engineering UW-Madison Email: kliu8@wisc.edu Office: Room 3017 (Mechanical Engineering Building)

More information

Data Mining Stat 588

Data Mining Stat 588 Data Mining Stat 588 Lecture 02: Linear Methods for Regression Department of Statistics & Biostatistics Rutgers University September 13 2011 Regression Problem Quantitative generic output variable Y. Generic

More information

Bayesian linear regression

Bayesian linear regression Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding

More information

Regression Shrinkage and Selection via the Lasso

Regression Shrinkage and Selection via the Lasso Regression Shrinkage and Selection via the Lasso ROBERT TIBSHIRANI, 1996 Presenter: Guiyun Feng April 27 () 1 / 20 Motivation Estimation in Linear Models: y = β T x + ɛ. data (x i, y i ), i = 1, 2,...,

More information

MSA220/MVE440 Statistical Learning for Big Data

MSA220/MVE440 Statistical Learning for Big Data MSA220/MVE440 Statistical Learning for Big Data Lecture 7/8 - High-dimensional modeling part 1 Rebecka Jörnsten Mathematical Sciences University of Gothenburg and Chalmers University of Technology Classification

More information

A Modern Look at Classical Multivariate Techniques

A Modern Look at Classical Multivariate Techniques A Modern Look at Classical Multivariate Techniques Yoonkyung Lee Department of Statistics The Ohio State University March 16-20, 2015 The 13th School of Probability and Statistics CIMAT, Guanajuato, Mexico

More information

CS 4491/CS 7990 SPECIAL TOPICS IN BIOINFORMATICS

CS 4491/CS 7990 SPECIAL TOPICS IN BIOINFORMATICS CS 4491/CS 7990 SPECIAL TOPICS IN BIOINFORMATICS * Some contents are adapted from Dr. Hung Huang and Dr. Chengkai Li at UT Arlington Mingon Kang, Ph.D. Computer Science, Kennesaw State University Problems

More information

A simulation study of model fitting to high dimensional data using penalized logistic regression

A simulation study of model fitting to high dimensional data using penalized logistic regression A simulation study of model fitting to high dimensional data using penalized logistic regression Ellinor Krona Kandidatuppsats i matematisk statistik Bachelor Thesis in Mathematical Statistics Kandidatuppsats

More information

Stability and the elastic net

Stability and the elastic net Stability and the elastic net Patrick Breheny March 28 Patrick Breheny High-Dimensional Data Analysis (BIOS 7600) 1/32 Introduction Elastic Net Our last several lectures have concentrated on methods for

More information

Lasso: Algorithms and Extensions

Lasso: Algorithms and Extensions ELE 538B: Sparsity, Structure and Inference Lasso: Algorithms and Extensions Yuxin Chen Princeton University, Spring 2017 Outline Proximal operators Proximal gradient methods for lasso and its extensions

More information

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Sparse Recovery using L1 minimization - algorithms Yuejie Chi Department of Electrical and Computer Engineering Spring

More information

EXTENDING PARTIAL LEAST SQUARES REGRESSION

EXTENDING PARTIAL LEAST SQUARES REGRESSION EXTENDING PARTIAL LEAST SQUARES REGRESSION ATHANASSIOS KONDYLIS UNIVERSITY OF NEUCHÂTEL 1 Outline Multivariate Calibration in Chemometrics PLS regression (PLSR) and the PLS1 algorithm PLS1 from a statistical

More information

The Adaptive Lasso and Its Oracle Properties Hui Zou (2006), JASA

The Adaptive Lasso and Its Oracle Properties Hui Zou (2006), JASA The Adaptive Lasso and Its Oracle Properties Hui Zou (2006), JASA Presented by Dongjun Chung March 12, 2010 Introduction Definition Oracle Properties Computations Relationship: Nonnegative Garrote Extensions:

More information

A New Bayesian Variable Selection Method: The Bayesian Lasso with Pseudo Variables

A New Bayesian Variable Selection Method: The Bayesian Lasso with Pseudo Variables A New Bayesian Variable Selection Method: The Bayesian Lasso with Pseudo Variables Qi Tang (Joint work with Kam-Wah Tsui and Sijian Wang) Department of Statistics University of Wisconsin-Madison Feb. 8,

More information

IDENTIFYING MULTIPLE OUTLIERS IN LINEAR REGRESSION : ROBUST FIT AND CLUSTERING APPROACH

IDENTIFYING MULTIPLE OUTLIERS IN LINEAR REGRESSION : ROBUST FIT AND CLUSTERING APPROACH SESSION X : THEORY OF DEFORMATION ANALYSIS II IDENTIFYING MULTIPLE OUTLIERS IN LINEAR REGRESSION : ROBUST FIT AND CLUSTERING APPROACH Robiah Adnan 2 Halim Setan 3 Mohd Nor Mohamad Faculty of Science, Universiti

More information

Statistics 203: Introduction to Regression and Analysis of Variance Penalized models

Statistics 203: Introduction to Regression and Analysis of Variance Penalized models Statistics 203: Introduction to Regression and Analysis of Variance Penalized models Jonathan Taylor - p. 1/15 Today s class Bias-Variance tradeoff. Penalized regression. Cross-validation. - p. 2/15 Bias-variance

More information

The OSCAR for Generalized Linear Models

The OSCAR for Generalized Linear Models Sebastian Petry & Gerhard Tutz The OSCAR for Generalized Linear Models Technical Report Number 112, 2011 Department of Statistics University of Munich http://www.stat.uni-muenchen.de The OSCAR for Generalized

More information

A direct formulation for sparse PCA using semidefinite programming

A direct formulation for sparse PCA using semidefinite programming A direct formulation for sparse PCA using semidefinite programming A. d Aspremont, L. El Ghaoui, M. Jordan, G. Lanckriet ORFE, Princeton University & EECS, U.C. Berkeley Available online at www.princeton.edu/~aspremon

More information

Robust Regression Diagnostics. Regression Analysis

Robust Regression Diagnostics. Regression Analysis Robust Regression Diagnostics 1.1 A Graduate Course Presented at the Faculty of Economics and Political Sciences, Cairo University Professor Ali S. Hadi The American University in Cairo and Cornell University

More information

Regularization: Ridge Regression and the LASSO

Regularization: Ridge Regression and the LASSO Agenda Wednesday, November 29, 2006 Agenda Agenda 1 The Bias-Variance Tradeoff 2 Ridge Regression Solution to the l 2 problem Data Augmentation Approach Bayesian Interpretation The SVD and Ridge Regression

More information

Variable Selection for Highly Correlated Predictors

Variable Selection for Highly Correlated Predictors Variable Selection for Highly Correlated Predictors Fei Xue and Annie Qu arxiv:1709.04840v1 [stat.me] 14 Sep 2017 Abstract Penalty-based variable selection methods are powerful in selecting relevant covariates

More information

A Bootstrap Lasso + Partial Ridge Method to Construct Confidence Intervals for Parameters in High-dimensional Sparse Linear Models

A Bootstrap Lasso + Partial Ridge Method to Construct Confidence Intervals for Parameters in High-dimensional Sparse Linear Models A Bootstrap Lasso + Partial Ridge Method to Construct Confidence Intervals for Parameters in High-dimensional Sparse Linear Models Jingyi Jessica Li Department of Statistics University of California, Los

More information

Robust variable selection through MAVE

Robust variable selection through MAVE This is the author s final, peer-reviewed manuscript as accepted for publication. The publisher-formatted version may be available through the publisher s web site or your institution s library. Robust

More information

Standardization and the Group Lasso Penalty

Standardization and the Group Lasso Penalty Standardization and the Group Lasso Penalty Noah Simon and Rob Tibshirani Corresponding author, email: nsimon@stanfordedu Sequoia Hall, Stanford University, CA 9435 March, Abstract We re-examine the original

More information

Package covtest. R topics documented:

Package covtest. R topics documented: Package covtest February 19, 2015 Title Computes covariance test for adaptive linear modelling Version 1.02 Depends lars,glmnet,glmpath (>= 0.97),MASS Author Richard Lockhart, Jon Taylor, Ryan Tibshirani,

More information

DIMENSION REDUCTION OF THE EXPLANATORY VARIABLES IN MULTIPLE LINEAR REGRESSION. P. Filzmoser and C. Croux

DIMENSION REDUCTION OF THE EXPLANATORY VARIABLES IN MULTIPLE LINEAR REGRESSION. P. Filzmoser and C. Croux Pliska Stud. Math. Bulgar. 003), 59 70 STUDIA MATHEMATICA BULGARICA DIMENSION REDUCTION OF THE EXPLANATORY VARIABLES IN MULTIPLE LINEAR REGRESSION P. Filzmoser and C. Croux Abstract. In classical multiple

More information

Gaussian Graphical Models and Graphical Lasso

Gaussian Graphical Models and Graphical Lasso ELE 538B: Sparsity, Structure and Inference Gaussian Graphical Models and Graphical Lasso Yuxin Chen Princeton University, Spring 2017 Multivariate Gaussians Consider a random vector x N (0, Σ) with pdf

More information

Business Statistics. Lecture 10: Correlation and Linear Regression

Business Statistics. Lecture 10: Correlation and Linear Regression Business Statistics Lecture 10: Correlation and Linear Regression Scatterplot A scatterplot shows the relationship between two quantitative variables measured on the same individuals. It displays the Form

More information

High-dimensional regression

High-dimensional regression High-dimensional regression Advanced Methods for Data Analysis 36-402/36-608) Spring 2014 1 Back to linear regression 1.1 Shortcomings Suppose that we are given outcome measurements y 1,... y n R, and

More information

Statistical Inference

Statistical Inference Statistical Inference Liu Yang Florida State University October 27, 2016 Liu Yang, Libo Wang (Florida State University) Statistical Inference October 27, 2016 1 / 27 Outline The Bayesian Lasso Trevor Park

More information

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract Journal of Data Science,17(1). P. 145-160,2019 DOI:10.6339/JDS.201901_17(1).0007 WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION Wei Xiong *, Maozai Tian 2 1 School of Statistics, University of

More information

Improved Ridge Estimator in Linear Regression with Multicollinearity, Heteroscedastic Errors and Outliers

Improved Ridge Estimator in Linear Regression with Multicollinearity, Heteroscedastic Errors and Outliers Journal of Modern Applied Statistical Methods Volume 15 Issue 2 Article 23 11-1-2016 Improved Ridge Estimator in Linear Regression with Multicollinearity, Heteroscedastic Errors and Outliers Ashok Vithoba

More information

Bivariate Relationships Between Variables

Bivariate Relationships Between Variables Bivariate Relationships Between Variables BUS 735: Business Decision Making and Research 1 Goals Specific goals: Detect relationships between variables. Be able to prescribe appropriate statistical methods

More information

Indian Statistical Institute

Indian Statistical Institute Indian Statistical Institute Introductory Computer programming Robust Regression methods with high breakdown point Author: Roll No: MD1701 February 24, 2018 Contents 1 Introduction 2 2 Criteria for evaluating

More information

Model Selection. Frank Wood. December 10, 2009

Model Selection. Frank Wood. December 10, 2009 Model Selection Frank Wood December 10, 2009 Standard Linear Regression Recipe Identify the explanatory variables Decide the functional forms in which the explanatory variables can enter the model Decide

More information

Institute of Statistics Mimeo Series No Simultaneous regression shrinkage, variable selection and clustering of predictors with OSCAR

Institute of Statistics Mimeo Series No Simultaneous regression shrinkage, variable selection and clustering of predictors with OSCAR DEPARTMENT OF STATISTICS North Carolina State University 2501 Founders Drive, Campus Box 8203 Raleigh, NC 27695-8203 Institute of Statistics Mimeo Series No. 2583 Simultaneous regression shrinkage, variable

More information

Variable Selection under Measurement Error: Comparing the Performance of Subset Selection and Shrinkage Methods

Variable Selection under Measurement Error: Comparing the Performance of Subset Selection and Shrinkage Methods Variable Selection under Measurement Error: Comparing the Performance of Subset Selection and Shrinkage Methods Ellen Sasahara Bachelor s Thesis Supervisor: Prof. Dr. Thomas Augustin Department of Statistics

More information

The Risk of James Stein and Lasso Shrinkage

The Risk of James Stein and Lasso Shrinkage Econometric Reviews ISSN: 0747-4938 (Print) 1532-4168 (Online) Journal homepage: http://tandfonline.com/loi/lecr20 The Risk of James Stein and Lasso Shrinkage Bruce E. Hansen To cite this article: Bruce

More information

A CONNECTION BETWEEN LOCAL AND DELETION INFLUENCE

A CONNECTION BETWEEN LOCAL AND DELETION INFLUENCE Sankhyā : The Indian Journal of Statistics 2000, Volume 62, Series A, Pt. 1, pp. 144 149 A CONNECTION BETWEEN LOCAL AND DELETION INFLUENCE By M. MERCEDES SUÁREZ RANCEL and MIGUEL A. GONZÁLEZ SIERRA University

More information

Pathwise coordinate optimization

Pathwise coordinate optimization Stanford University 1 Pathwise coordinate optimization Jerome Friedman, Trevor Hastie, Holger Hoefling, Robert Tibshirani Stanford University Acknowledgements: Thanks to Stephen Boyd, Michael Saunders,

More information

A General Framework for Variable Selection in Linear Mixed Models with Applications to Genetic Studies with Structured Populations

A General Framework for Variable Selection in Linear Mixed Models with Applications to Genetic Studies with Structured Populations A General Framework for Variable Selection in Linear Mixed Models with Applications to Genetic Studies with Structured Populations Joint work with Karim Oualkacha (UQÀM), Yi Yang (McGill), Celia Greenwood

More information

Chapter 3. Linear Models for Regression

Chapter 3. Linear Models for Regression Chapter 3. Linear Models for Regression Wei Pan Division of Biostatistics, School of Public Health, University of Minnesota, Minneapolis, MN 55455 Email: weip@biostat.umn.edu PubH 7475/8475 c Wei Pan Linear

More information

The MNet Estimator. Patrick Breheny. Department of Biostatistics Department of Statistics University of Kentucky. August 2, 2010

The MNet Estimator. Patrick Breheny. Department of Biostatistics Department of Statistics University of Kentucky. August 2, 2010 Department of Biostatistics Department of Statistics University of Kentucky August 2, 2010 Joint work with Jian Huang, Shuangge Ma, and Cun-Hui Zhang Penalized regression methods Penalized methods have

More information