COMBINING THE LIU-TYPE ESTIMATOR AND THE PRINCIPAL COMPONENT REGRESSION ESTIMATOR

Size: px
Start display at page:

Download "COMBINING THE LIU-TYPE ESTIMATOR AND THE PRINCIPAL COMPONENT REGRESSION ESTIMATOR"

Transcription

1 Noname manuscript No. (will be inserted by the editor) COMBINING THE LIU-TYPE ESTIMATOR AND THE PRINCIPAL COMPONENT REGRESSION ESTIMATOR Deniz Inan Received: date / Accepted: date Abstract In this study a new two-parameter estimator which includes the ordinary least squares (OLS), the principal components regression (PCR) and the Liu-type estimator is proposed. Conditions for the superiority of this new estimator over the PCR, r-k class estimator and Liu-type estimator are derived. Furthermore the performance of this estimator is compared with the other estimators in different conditions with simulation studies. Keywords Liu-type estimator; r-k class estimators; principal component regression estimator; ridge estimator; multicollinearity. Mathematics Subject Classification (0) 62J07 1 Introduction Multicollinearity can be a problem for many empirical researchers, and this problem will influence how you are able to interpret the coefficients in your model. It is well known that especially ordinary least squares (OLS) estimator can be effected seriously by the presence of multicollinearity so parameter estimates have large sampling variances. To deal with this problem different methods have been proposed such as ordinary ridge regression (ORR) estimator (Hoerl and Kennard, 1970), principal component regression (PCR) estimator (Draper and Smith, 1981), least absolute shrinkage and selection operator (LASSO) estimator (Tibshirani, 1996), least angle regression (LARS) estimator (Efron et al.,4) and Liu-type estimator (Liu, 3). Also combinations of these estimators have been proposed with the expectation that the combination of different estimators might inherit the advantages of both. Baye and D. Inan Marmara University Department of Statistics Istanbul Turkey Tel.: Fax: denizlukuslu@marmara.edu.tr

2 2 Deniz Inan Parker (1984) combined ORR estimator with PCR estimator and proposed r-k class estimator which provides an alternative method of dealing with multicollinearity. The r-k class estimator was compared with the ORR estimator and OLS estimator according to scalar mean square error (MSE) criteria in Nomura and Ohkuba (1985). Sarkar (1996) studied the properties of r-k class estimators in the sense of matrix mean square error. Differently, Kaciranlar and Sakallioglu (1) proposed r-d class estimators by combining the Liu and PCR estimators and compared this new estimator with the OLS, PCR and Liu estimators according to MSE criteria. Conditions for superiority of the r-d class estimators over these estimators and the r-k class estimators are derived in Ozkale and Kaciranlar (7). Aim of this paper is to introduce a new class of estimators which include OLS, PCR and Liu-type estimators that can be used as an alternative method for combating multicollinearity and compare this estimator with the OLS, ORR, Liu-type, PCR and r-k class estimators in the sense of MSE criteria. 2 A new class of estimators Consider the linear regression model Y = Xβ +ǫ (2.1) where X is a known n p matrix of standardized regressors with rank(x) = p, Y is a n 1 response vector, β is a vector of parameters and ǫ is a n 1 vector of errors with E(ǫ) = 0 and Cov(ǫ) = σ 2 I. Let us rewrite model (2.1) in canonical form Y = Zα+ǫ where Z = Xθ, α = θ β and θ is the p p orthogonal matrix whose columns constitute the eigenvectors of X X. Then, Z Z = θ X Xθ = Λ = diag(λ 1,,λ p ) where λ 1 λ 2 λ p > 0 are the ordered eigenvalues of X X. The columns of Z, which define the new set of orthogonal regressors Z = [Z 1,Z 2,,Z p ] are referred to as principal components and the OLS estimator of α is ˆα = (Z Z) 1 Z y = Λ 1 Z y. The PCR regression approach combats multicollinearity by using less than the full set of principal components in the model. Suppose that the last p r of the eigenvalues are approximately near zero. In PCR the principal components corresponding to near-zero eigenvalues are removed from the analysis and OLS is applied to the remaining components. That is, ˆα PC = Bˆα

3 Title Suppressed Due to Excessive Length 3 where B is a diagonal matrix whose diagonal elements are as b 1,1 = b 2,2 = = b r,r = 1 and b r+1,r+1 = b r+2,r+2 = = b p,p = 0 (Montgomery et. al., 1). Likewise in terms of original regressors, the PCR estimator can be given as ˆβ PC = θα PC ˆ (2.2). Let θ r be the remaining eigenvectors of X X after having deleted p r of columns of θ which correspond to the near-zero eigenvalues. Thus, θ rx Xθ r = Λ r = diag(λ 1,,λ r ) and PCR estimator given in (2.2), can also be given as in the form, ˆβ PC = θ r (θ rx Xθ r ) 1 θ rx y and r-k class estimator is ˆβ r (k) = θ r (θ rx Xθ r +ki) 1 θ rx y where r p and k 0. A new class of estimators which includes the OLS, PCR and Liu-Type estimators as a special cases is defined as follows. Definition 1. The r-(k,d) class estimator of β in (2.1) is ˆβ r (k,d) = θ r (θ rx Xθ r +ki) 1 (θ rx Xθ r di r )θ rˆβ PC, where k 0 and < d < are shrinkage and biasing parameters that must be chosen by the researcher and 0 < r p. Special cases of r-(k,d) class estimators can be given as: ˆβ p (0,0) = ˆβ LS = (X X) 1 X Y, the OLS estimator, ˆβ r (k, k) = ˆβ PC = θ r (θ rx Xθ r ) 1 θ rx y, the PCR estimator, ˆβ r (k,0) = ˆβ r (k) = θ r (θ rx Xθ r +ki) 1 θ rx y, the r-k class estimator, ˆβ p (k,d) = ˆβ(k,d) = (X X +ki) 1 (X X di)ˆβ, Liu-Type estimator. The MSEs ˆβ r (k,d), ˆβ LS, ˆβ PC, ˆβr (k), ˆβ(k,d) are given MSE( ˆβ r (k,d)) = σ 2 (λ i d) 2 λ i (λ i +k) 2 + ( d+k p λ i +k )2 αi 2 + αi 2 (2.3) respectively. MSE(ˆβ LS ) = σ 2 p MSE(ˆβ PC ) = σ 2 1 λ i 1 λ i + p i=r+1 MSE( ˆβ r (k)) = σ 2 (λ i ) (λ i +k) 2 + k ( λ i +k )2 αi 2 + α 2 i i=r+1 p i=r+1 p MSE(ˆβ(k,d)) = σ 2 (λ i d) 2 p λ i (λ i +k) 2 + ( d+k λ i +k )2 αi, 2 α 2 i

4 4 Deniz Inan Theorem 1. For any k > 0 and 0 < r p, if the biasing parameter chosen as d = ((σ 2 kαi)/(λ 2 i +k) 2 )/ ((σ 2 +λ i αi)/(λ 2 i (λ i +k) 2 )) then MSE( ˆβ r (k,d)) MSE(ˆβ PC ) and MSE( ˆβ r (k,d)) MSE( ˆβ r (k)). Proof. From (2.1.1) it is easy to see that, for fixed k, MSE( ˆβ r (k,d)) is a quadratic function of d. If the first derivative of this function is taken and solved for d, the optimum value of d which minimizes MSE( ˆβ r (k,d)) can be obtained as d opt = ((σ 2 kαi)/(λ 2 i +k) 2 )/ ((σ 2 +λ i αi)/(λ 2 i (λ i +k) 2 )) (2.4). In view of MSE( ˆβ r (k, k)) = MSE(ˆβ PC ), we have MSE( ˆβ r (k,d opt )) MSE(ˆβ PC ). And the equality occurs when d opt = k. Similarly in view of MSE( ˆβ r (k,0)) = MSE( ˆβ r (k)),wehavemse( ˆβ r (k,d opt )) MSE( ˆβ r (k)).equality occurs when d opt = 0. Theorem 2. For any k > 0, 0 < r p and i = 1,,p, when λ 1 < d < min[(σ 2 +α 2 i( 2k λ i ))λ i /(σ 2 +α 2 iλ i )] we have MSE( ˆβ r (k,d)) MSE(ˆβ(k,d)). Proof. For a given r the solution of MSE( ˆβ r (k,d)) MSE(ˆβ r 1 (k,d)) > 0 for d can be obtained as λ r < d < (σ 2 +α 2 r( 2k λ r ))λ r /(σ 2 +α 2 rλ r ). Generalizing this solution, if we choose d as, λ 1 < d < min[(σ 2 +α 2 i( 2k λ i ))λ i /(σ 2 +α 2 iλ i )] we get MSE( ˆβ r (k,d)) MSE(ˆβ r 1 (k,d)) > 0 for all r = 1,,p. As a result, MSE( ˆβ p (k,d)) > MSE(ˆβ p 1 (k,d)) > > MSE(ˆβ r (k,d)) MSE(ˆβ r (k,d)) < MSE(ˆβ(k,d)). This completes the proof of Theorem 2 (See Appendix).

5 Title Suppressed Due to Excessive Length 5 3 Simulation Results Monte Carlo simulations were used to illustrate the behaviour of the proposed estimator. Simulations were carried out under different degrees of multicollinearity. In first simulation study the number of regressors was chosen to be six and the parameters were set to be (22,1.4,1.2, 1.5, 0.2,0.7,1.7). The size of the sample n was set to be 30, 50 and 70. The regressors were generated from x i,1 = (1 ϕ 2 ) 1/2 w i,1 +ϕw i,5 +20 x i,2 = (1 ϕ 2 ) 1/2 w i,2 +ϕw i,5 +45 x i,3 = (1 ρ 2 ) 1/2 w i,3 +ρw i,6 24 x i,4 = (1 ρ 2 ) 1/2 w i,4 +ρw i,6 +7 x i,5 = (1 ρ 2 ) 1/2 w i,3 +ρw i,2 +19 x i,6 = w i,2 14 wherew i,j aregeneratedindependentlyfromn(0,2).ϕandρcorrespondtothe correlations between x i,1, x i,2 and x i,3, x i,4 are ϕ 2 and ρ 2 respectively (Gibbons, 1981). Three different sets of correlations were considered corresponding to ϕ = 0.95 and ρ = 0.95, ϕ = 0.98 and ρ = 0.95, ϕ = 0.95 and ρ = Simulations were repeated times. In each experiment inhomogeneous model was applied to the data, estimated error variances of each estimator and the following estimates were computed: 1.OLS : Ordinary least square estimator 2.ORR: Ordinary ridge regression estimator, k = ˆσ LS 2 /ˆβ LSˆβ LS (Schafer et 3.Liu: al., 1984) Liu-type estimator, k = ˆσ LS 2 /ˆβ LSˆβ LS, ˆd = p ((ˆσ2 kˆα i 2)/(λ i +k) 2 )/ p ((ˆσ2 +λ iˆα i 2)/(λ i(λ i +k) 2 )) (Liu, 3) and ˆβ = ˆβ LS 4.PCR: Principal component regression estimator 5.r-k: r-k class estimator, k = ˆσ LS 2 /ˆβ LSˆβ LS 6.r-(k,d): r-(k,d) class estimator, k = ˆσ LS 2 /ˆβ LSˆβ LS, ˆd = r ((ˆσ2 kˆα i 2)/ (λ i +k) 2 )/ r ((ˆσ2 +λ iˆα i 2)/(λ i(λ i +k) 2 )). Unfortunately,thed opt valuesgiveninthetheoremsdependonunknownmodel parameters and so, for practical purposes, these unknown parameters were decided to replace with their suitable estimates. If we substitute α and σ with their ORR estimates (ˆα R and ˆσ R ) in (2.2.1), we obtain ˆd opt = ((ˆσ R 2 kˆα R(i) 2 )/(λ i +k) 2 )/ ((ˆσ R 2 +λ iˆα R(i) 2 )/(λ i(λ i +k) 2 ))

6 6 Deniz Inan Table 1 EMSE values for six estimators (continuous regressors) ϕ ρ n OLS ORR Liu PCR r-k r-(k,d) Table 2 Estimated error variances for six estimators (continuous regressors) ϕ ρ n OLS ORR Liu PCR r-k r-(k,d) as an operational estimator for d opt. In simulation studies this estimator was used for r-(k,d) class estimator and considering the proposal of Liu (3), ˆd opt = p ((ˆσ R 2 kˆα R(i) 2 )/(λ i +k) 2 )/ p ((ˆσ R 2 +λ iˆα R(i) 2 )/(λ i(λ i +k) 2 )) was used for Liu-type estimator. The estimated MSEs (EMSEs) and means of the estimated error variances were computed for each of the above six estimators. The results are presented in Table 1 and Table 2. As expected, our proposed estimators have smaller EMSEs than those of the other estimators. Although the EMSEs of all estimators increased along with multicollinearity, the increase in EMSEs for proposed estimator was not as much as the others. It can be concluded that the effect of multicollinearity on the other estimators is more pronounced than proposed estimator. The EMSE reduction is much better when the multicollinearity is high. Also there is no significant difference among the estimators according to estimated error variances.

7 Title Suppressed Due to Excessive Length 7 Table 3 EMSE values for six estimator (both continuous and ordinal regressors) ϕ ρ n OLS ORR Liu PCR r-k r-(k,d) The case when the model has both continuous and ordinal regressors was also considered. For this case, six regressors were generated as in the first simulation study, the parameters were set to be (25,2.3,4.1, 1.5, 0.9,0.9,1.7) and the size of the sample n was set to be 30, 50 and 70. Two different sets of correlations were considered corresponding to ϕ = 0.95 and ρ = 0.95, ϕ = 0.95 and ρ = After six continuous regressors were generated, the first two of them were discretized in order to make them ordinal. In this way the impact of collinearity was slightly reduced. Simulations were repeated times. The results are presented in Table 3. As seen from the Table 3, in all cases r-(k,d) class estimator has smaller EMSE than the other estimators. 4 Numerical Example In this section, the acetylene data described and analysed in detail by Marquardt (1980) was employed. This data set is a typical set of response surface data for which a full quadratic model in three regressors is often considered to be an appropriate candidate model. The standardized full quadratic model for the data is y = β 0 +β 1 x 1 +β 2 x 2 +β 3 x 3 +β 1,2 x 1 x 2 +β 1,3 x 1 x 3 +β 2,3 x 2 x 3 +β 1,1 x 2 1+β 2,2 x 2 2+β 3,3 x 2 3+ǫ wherey=percentageofconversion,x 1 =(temperature )/80.623,x 2 =(H 2 (nheptane)-12.44)/5.662 and x 3 =(contact time )/ Each of the six estimators was applied to these data. k, r, ˆd opt and ˆd opt were determined respectively as , 6, and The parameter estimations are given in Table 4. Under this model, the condition number is κ = λ max λ min = = 42048

8 8 Deniz Inan Table 4 Parameter estimations for the acetylene data ˆβ 0 ˆβ1 ˆβ2 ˆβ3 ˆβ4 ˆβ5 ˆβ6 ˆβ7 ˆβ8 ˆβ9 OLS ORR Liu PCR r-k r-(k,d) Table 5 EMSEs, estimated parameter variances and biases for the acetylene data Estimated parameter Estimated parameter EMSE variance bias OLS ORR Liu PCR r-k r-(k,d) which indicates severe multicollinearity. In this case, the OLS estimator does not provide good estimations. Although the OLS estimator achieves an optimum fit to the estimation data, it does not provide a good prediction for the new data. Therefore, we can not trust the OLS estimator under this model. The bootstrap sampling method was used for these data to estimate the MSE values of relevant estimators and illustrate the distributional properties of the parameters. For this purpose, bootstrap samples were generated, and parameter estimates of six estimators were computed for each of these samples. The mean of the OLS estimates was considered as β, and EMSEs, estimated parameter variances and estimated biases were computed as given in Table 5. Because of the severe multicollinearity, small changes in the data cause large differences in the parameter estimates. To illustrate this situation, for each estimator, histograms of ˆβ 9 are given in Figure 1 as an example. As seen in Figure 1, the OLS, ORR and Liu-type estimates vary over a large interval. Compared to the OLS, the distributions for the ORR and Liutype estimations have stronger peaks and more rapid decays, but they still have heavy tails. Therefore, we can not say that these estimates are stable. The PCR, r-k and r-(k,d) class estimations are distributed within a narrower interval according to the OLS, ORR and Liu-type estimates. In addition, the distribution of r-(k,d) class estimations has a stronger peak and more rapid decay according to the distributions of the PCR and r-k class estimations. Considering all of these histograms, we can say that among the six estimators, the r-(k,d) class estimator gives the most stable estimates, indicating that this estimator provides a better prediction for the new data.

9 Title Suppressed Due to Excessive Length OLS ORR Liu PCR r k r k,d Fig. 1 Histograms of ˆβ 9 5 Summary In this paper conditions have been obtained for the superiority of the r-(k,d) class estimators over the PCR, r-k class and Liu-type estimators. It is seen that the conditions obtained in Theorem 1. and Theorem 2. depend on the unknown parameters. For practical purposes, thus, we have suggested to replace them with their ORR estimators. Simulation studies are constructed to decide whether the relevant conditions are satisfied. Simulation results obtained through the use of using ORR estimates are seen to be supporting our theoretical findings. Finally, considering these simulation studies it is concluded that especially for the severe multicollinearity case and small sample sizes r-(k,d) class estimator performs better than the other estimators in the sense of MSE. 6 Appendix MSE( ˆβ r (k,d)) MSE(ˆβ r 1 (k,d)) > 0 σ 2 ( r (λ i d) 2 /(λ i (λ i +k) 2 ) r 1 (λ i d) 2 /(λ i (λ i +k) 2 ))+ r ((d+k)/(λ i +k)) 2 α 2 i r 1 ((d+k)/(λ i +k)) 2 α 2 i + p i=r+1 α2 i p i=r α2 i > 0 σ 2 (λ r d) 2 /(λ r (λ r +k) 2 )+((d+k)/(λ r +k)) 2 α 2 r α 2 r > 0 λ r < d < (σ 2 +α 2 r( 2k λ r ))λ r /(σ 2 +α 2 rλ r ) Acknowledgements The author was supported by the Marmara University Scientific Research Project Unit (BAPKO, project number: FEN-D ).

10 10 Deniz Inan References 1. Baye M.R., Parker, D.F. (1984). Combining Ridge and Principal Component Regression: A Money Demand Illustration.Communications in Statistics-Theory and Methods 13, Draper, N.R., Smith, H.(1981). Applied Regression Analysis. John Willey & Sons. 3. Efron, B., Hastie, T., Johnstone, I., Tibshirani, R.(4). Least Angle Regression. The Annals of Statistics 32, Gibbons, D.G. (1981). A simulation study of some ridge estimators. Journal of the American Statistical Association, 76, Hoerl, A.E., Kennard, R.W.(1970). Ridge Regression: Biased Estimation for Nonorthogonal Problems. Technometrics 12, Kaciranlar S., Sakallioglu S. (1). Combining the Liu Estimator and the Principal Component Regression Estimator. Communications in Statistics-Theory and Methods 30, Liu K. (3).Using Liu Type Estimator to Combat Collinearity. Communications in Statistics-Theory and Methods 32, Marquardt, D.W.(1980) A Critique of Some Ridge Regression Methods. Journal of the American Statistical Association 75, Montgomery, D.C., Peck, A.E., Vining, G.G.(1) Introduction to the Linear Regression Analysis. John Willey & Sons Nomura M., Ohkubo T. (1985). A Note on Combining Ridge and Principal Component Regression. Communications in Statistics-Theory and Methods 14, Ozkale M.R., Kaciranlar S. (7). Superiorty of the r-d Class Estimator Over Some Estimators by the Mean Square Error Matrix Criterion. Statistics and Probability Letters 77, Sarkar,N.(1996). Mean Square Error Matrix Comparison of Some Estimators in Linear Regression with Multicollinearity. Statist.Probab.Lett. 30, Tibshirani, R.(1996). Regression Shrinkage and Selection via the Lasso. J.R. Statist. Soc.B 58,

Improved Liu Estimators for the Poisson Regression Model

Improved Liu Estimators for the Poisson Regression Model www.ccsenet.org/isp International Journal of Statistics and Probability Vol., No. ; May 202 Improved Liu Estimators for the Poisson Regression Model Kristofer Mansson B. M. Golam Kibria Corresponding author

More information

Evaluation of a New Estimator

Evaluation of a New Estimator Pertanika J. Sci. & Technol. 16 (2): 107-117 (2008) ISSN: 0128-7680 Universiti Putra Malaysia Press Evaluation of a New Estimator Ng Set Foong 1 *, Low Heng Chin 2 and Quah Soon Hoe 3 1 Department of Information

More information

Ridge Regression Revisited

Ridge Regression Revisited Ridge Regression Revisited Paul M.C. de Boer Christian M. Hafner Econometric Institute Report EI 2005-29 In general ridge (GR) regression p ridge parameters have to be determined, whereas simple ridge

More information

Multicollinearity and A Ridge Parameter Estimation Approach

Multicollinearity and A Ridge Parameter Estimation Approach Journal of Modern Applied Statistical Methods Volume 15 Issue Article 5 11-1-016 Multicollinearity and A Ridge Parameter Estimation Approach Ghadban Khalaf King Khalid University, albadran50@yahoo.com

More information

Comparison of Some Improved Estimators for Linear Regression Model under Different Conditions

Comparison of Some Improved Estimators for Linear Regression Model under Different Conditions Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 3-24-2015 Comparison of Some Improved Estimators for Linear Regression Model under

More information

Measuring Local Influential Observations in Modified Ridge Regression

Measuring Local Influential Observations in Modified Ridge Regression Journal of Data Science 9(2011), 359-372 Measuring Local Influential Observations in Modified Ridge Regression Aboobacker Jahufer 1 and Jianbao Chen 2 1 South Eastern University and 2 Xiamen University

More information

Robust model selection criteria for robust S and LT S estimators

Robust model selection criteria for robust S and LT S estimators Hacettepe Journal of Mathematics and Statistics Volume 45 (1) (2016), 153 164 Robust model selection criteria for robust S and LT S estimators Meral Çetin Abstract Outliers and multi-collinearity often

More information

EFFICIENCY of the PRINCIPAL COMPONENT LIU- TYPE ESTIMATOR in LOGISTIC REGRESSION

EFFICIENCY of the PRINCIPAL COMPONENT LIU- TYPE ESTIMATOR in LOGISTIC REGRESSION EFFICIENCY of the PRINCIPAL COMPONEN LIU- YPE ESIMAOR in LOGISIC REGRESSION Authors: Jibo Wu School of Mathematics and Finance, Chongqing University of Arts and Sciences, Chongqing, China, linfen52@126.com

More information

Confidence Intervals in Ridge Regression using Jackknife and Bootstrap Methods

Confidence Intervals in Ridge Regression using Jackknife and Bootstrap Methods Chapter 4 Confidence Intervals in Ridge Regression using Jackknife and Bootstrap Methods 4.1 Introduction It is now explicable that ridge regression estimator (here we take ordinary ridge estimator (ORE)

More information

Regularized Multiple Regression Methods to Deal with Severe Multicollinearity

Regularized Multiple Regression Methods to Deal with Severe Multicollinearity International Journal of Statistics and Applications 21, (): 17-172 DOI: 1.523/j.statistics.21.2 Regularized Multiple Regression Methods to Deal with Severe Multicollinearity N. Herawati *, K. Nisa, E.

More information

Relationship between ridge regression estimator and sample size when multicollinearity present among regressors

Relationship between ridge regression estimator and sample size when multicollinearity present among regressors Available online at www.worldscientificnews.com WSN 59 (016) 1-3 EISSN 39-19 elationship between ridge regression estimator and sample size when multicollinearity present among regressors ABSTACT M. C.

More information

Research Article On the Stochastic Restricted r-k Class Estimator and Stochastic Restricted r-d Class Estimator in Linear Regression Model

Research Article On the Stochastic Restricted r-k Class Estimator and Stochastic Restricted r-d Class Estimator in Linear Regression Model Applied Mathematics, Article ID 173836, 6 pages http://dx.doi.org/10.1155/2014/173836 Research Article On the Stochastic Restricted r-k Class Estimator and Stochastic Restricted r-d Class Estimator in

More information

Improved Ridge Estimator in Linear Regression with Multicollinearity, Heteroscedastic Errors and Outliers

Improved Ridge Estimator in Linear Regression with Multicollinearity, Heteroscedastic Errors and Outliers Journal of Modern Applied Statistical Methods Volume 15 Issue 2 Article 23 11-1-2016 Improved Ridge Estimator in Linear Regression with Multicollinearity, Heteroscedastic Errors and Outliers Ashok Vithoba

More information

Effects of Outliers and Multicollinearity on Some Estimators of Linear Regression Model

Effects of Outliers and Multicollinearity on Some Estimators of Linear Regression Model 204 Effects of Outliers and Multicollinearity on Some Estimators of Linear Regression Model S. A. Ibrahim 1 ; W. B. Yahya 2 1 Department of Physical Sciences, Al-Hikmah University, Ilorin, Nigeria. e-mail:

More information

ISyE 691 Data mining and analytics

ISyE 691 Data mining and analytics ISyE 691 Data mining and analytics Regression Instructor: Prof. Kaibo Liu Department of Industrial and Systems Engineering UW-Madison Email: kliu8@wisc.edu Office: Room 3017 (Mechanical Engineering Building)

More information

Regularization: Ridge Regression and the LASSO

Regularization: Ridge Regression and the LASSO Agenda Wednesday, November 29, 2006 Agenda Agenda 1 The Bias-Variance Tradeoff 2 Ridge Regression Solution to the l 2 problem Data Augmentation Approach Bayesian Interpretation The SVD and Ridge Regression

More information

Nonlinear Inequality Constrained Ridge Regression Estimator

Nonlinear Inequality Constrained Ridge Regression Estimator The International Conference on Trends and Perspectives in Linear Statistical Inference (LinStat2014) 24 28 August 2014 Linköping, Sweden Nonlinear Inequality Constrained Ridge Regression Estimator Dr.

More information

Efficient Choice of Biasing Constant. for Ridge Regression

Efficient Choice of Biasing Constant. for Ridge Regression Int. J. Contemp. Math. Sciences, Vol. 3, 008, no., 57-536 Efficient Choice of Biasing Constant for Ridge Regression Sona Mardikyan* and Eyüp Çetin Department of Management Information Systems, School of

More information

Ridge Regression and Ill-Conditioning

Ridge Regression and Ill-Conditioning Journal of Modern Applied Statistical Methods Volume 3 Issue Article 8-04 Ridge Regression and Ill-Conditioning Ghadban Khalaf King Khalid University, Saudi Arabia, albadran50@yahoo.com Mohamed Iguernane

More information

A Modern Look at Classical Multivariate Techniques

A Modern Look at Classical Multivariate Techniques A Modern Look at Classical Multivariate Techniques Yoonkyung Lee Department of Statistics The Ohio State University March 16-20, 2015 The 13th School of Probability and Statistics CIMAT, Guanajuato, Mexico

More information

Bayes Estimators & Ridge Regression

Bayes Estimators & Ridge Regression Readings Chapter 14 Christensen Merlise Clyde September 29, 2015 How Good are Estimators? Quadratic loss for estimating β using estimator a L(β, a) = (β a) T (β a) How Good are Estimators? Quadratic loss

More information

Research Article On the Weighted Mixed Almost Unbiased Ridge Estimator in Stochastic Restricted Linear Regression

Research Article On the Weighted Mixed Almost Unbiased Ridge Estimator in Stochastic Restricted Linear Regression Applied Mathematics Volume 2013, Article ID 902715, 10 pages http://dx.doi.org/10.1155/2013/902715 Research Article On the Weighted Mixed Almost Unbiased Ridge Estimator in Stochastic Restricted Linear

More information

Alternative Biased Estimator Based on Least. Trimmed Squares for Handling Collinear. Leverage Data Points

Alternative Biased Estimator Based on Least. Trimmed Squares for Handling Collinear. Leverage Data Points International Journal of Contemporary Mathematical Sciences Vol. 13, 018, no. 4, 177-189 HIKARI Ltd, www.m-hikari.com https://doi.org/10.1988/ijcms.018.8616 Alternative Biased Estimator Based on Least

More information

Regression Shrinkage and Selection via the Lasso

Regression Shrinkage and Selection via the Lasso Regression Shrinkage and Selection via the Lasso ROBERT TIBSHIRANI, 1996 Presenter: Guiyun Feng April 27 () 1 / 20 Motivation Estimation in Linear Models: y = β T x + ɛ. data (x i, y i ), i = 1, 2,...,

More information

Generalized Ridge Regression Estimator in Semiparametric Regression Models

Generalized Ridge Regression Estimator in Semiparametric Regression Models JIRSS (2015) Vol. 14, No. 1, pp 25-62 Generalized Ridge Regression Estimator in Semiparametric Regression Models M. Roozbeh 1, M. Arashi 2, B. M. Golam Kibria 3 1 Department of Statistics, Faculty of Mathematics,

More information

A Comparison between Biased and Unbiased Estimators in Ordinary Least Squares Regression

A Comparison between Biased and Unbiased Estimators in Ordinary Least Squares Regression Journal of Modern Alied Statistical Methods Volume Issue Article 7 --03 A Comarison between Biased and Unbiased Estimators in Ordinary Least Squares Regression Ghadban Khalaf King Khalid University, Saudi

More information

Ridge Estimator in Logistic Regression under Stochastic Linear Restrictions

Ridge Estimator in Logistic Regression under Stochastic Linear Restrictions British Journal of Mathematics & Computer Science 15(3): 1-14, 2016, Article no.bjmcs.24585 ISSN: 2231-0851 SCIENCEDOMAIN international www.sciencedomain.org Ridge Estimator in Logistic Regression under

More information

Ridge Estimation and its Modifications for Linear Regression with Deterministic or Stochastic Predictors

Ridge Estimation and its Modifications for Linear Regression with Deterministic or Stochastic Predictors Ridge Estimation and its Modifications for Linear Regression with Deterministic or Stochastic Predictors James Younker Thesis submitted to the Faculty of Graduate and Postdoctoral Studies in partial fulfillment

More information

Journal of Asian Scientific Research COMBINED PARAMETERS ESTIMATION METHODS OF LINEAR REGRESSION MODEL WITH MULTICOLLINEARITY AND AUTOCORRELATION

Journal of Asian Scientific Research COMBINED PARAMETERS ESTIMATION METHODS OF LINEAR REGRESSION MODEL WITH MULTICOLLINEARITY AND AUTOCORRELATION Journal of Asian Scientific Research ISSN(e): 3-1331/ISSN(p): 6-574 journal homepage: http://www.aessweb.com/journals/5003 COMBINED PARAMETERS ESTIMATION METHODS OF LINEAR REGRESSION MODEL WITH MULTICOLLINEARITY

More information

Generalized Maximum Entropy Estimators: Applications to the Portland Cement Dataset

Generalized Maximum Entropy Estimators: Applications to the Portland Cement Dataset The Open Statistics & Probability Journal 2011 3 13-20 13 Open Access Generalized Maximum Entropy Estimators: Applications to the Portland Cement Dataset Fikri Akdeniz a* Altan Çabuk b and Hüseyin Güler

More information

Large Sample Theory For OLS Variable Selection Estimators

Large Sample Theory For OLS Variable Selection Estimators Large Sample Theory For OLS Variable Selection Estimators Lasanthi C. R. Pelawa Watagoda and David J. Olive Southern Illinois University June 18, 2018 Abstract This paper gives large sample theory for

More information

SOME NEW PROPOSED RIDGE PARAMETERS FOR THE LOGISTIC REGRESSION MODEL

SOME NEW PROPOSED RIDGE PARAMETERS FOR THE LOGISTIC REGRESSION MODEL IMPACT: International Journal of Research in Applied, Natural and Social Sciences (IMPACT: IJRANSS) ISSN(E): 2321-8851; ISSN(P): 2347-4580 Vol. 3, Issue 1, Jan 2015, 67-82 Impact Journals SOME NEW PROPOSED

More information

Variable Selection in Restricted Linear Regression Models. Y. Tuaç 1 and O. Arslan 1

Variable Selection in Restricted Linear Regression Models. Y. Tuaç 1 and O. Arslan 1 Variable Selection in Restricted Linear Regression Models Y. Tuaç 1 and O. Arslan 1 Ankara University, Faculty of Science, Department of Statistics, 06100 Ankara/Turkey ytuac@ankara.edu.tr, oarslan@ankara.edu.tr

More information

Linear regression methods

Linear regression methods Linear regression methods Most of our intuition about statistical methods stem from linear regression. For observations i = 1,..., n, the model is Y i = p X ij β j + ε i, j=1 where Y i is the response

More information

ENHANCING THE EFFICIENCY OF THE RIDGE REGRESSION MODEL USING MONTE CARLO SIMULATIONS

ENHANCING THE EFFICIENCY OF THE RIDGE REGRESSION MODEL USING MONTE CARLO SIMULATIONS www.arpapress.com/volumes/vol27issue1/ijrras_27_1_02.pdf ENHANCING THE EFFICIENCY OF THE RIDGE REGRESSION MODEL USING MONTE CARLO SIMULATIONS Rania Ahmed Hamed Mohamed Department of Statistics, Mathematics

More information

APPLICATION OF RIDGE REGRESSION TO MULTICOLLINEAR DATA

APPLICATION OF RIDGE REGRESSION TO MULTICOLLINEAR DATA Journal of Research (Science), Bahauddin Zakariya University, Multan, Pakistan. Vol.15, No.1, June 2004, pp. 97-106 ISSN 1021-1012 APPLICATION OF RIDGE REGRESSION TO MULTICOLLINEAR DATA G. R. Pasha 1 and

More information

The prediction of house price

The prediction of house price 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050

More information

Dimension Reduction Methods

Dimension Reduction Methods Dimension Reduction Methods And Bayesian Machine Learning Marek Petrik 2/28 Previously in Machine Learning How to choose the right features if we have (too) many options Methods: 1. Subset selection 2.

More information

Linear Regression Models. Based on Chapter 3 of Hastie, Tibshirani and Friedman

Linear Regression Models. Based on Chapter 3 of Hastie, Tibshirani and Friedman Linear Regression Models Based on Chapter 3 of Hastie, ibshirani and Friedman Linear Regression Models Here the X s might be: p f ( X = " + " 0 j= 1 X j Raw predictor variables (continuous or coded-categorical

More information

Using Ridge Least Median Squares to Estimate the Parameter by Solving Multicollinearity and Outliers Problems

Using Ridge Least Median Squares to Estimate the Parameter by Solving Multicollinearity and Outliers Problems Modern Applied Science; Vol. 9, No. ; 05 ISSN 9-844 E-ISSN 9-85 Published by Canadian Center of Science and Education Using Ridge Least Median Squares to Estimate the Parameter by Solving Multicollinearity

More information

STK-IN4300 Statistical Learning Methods in Data Science

STK-IN4300 Statistical Learning Methods in Data Science Outline of the lecture STK-I4300 Statistical Learning Methods in Data Science Riccardo De Bin debin@math.uio.no Model Assessment and Selection Cross-Validation Bootstrap Methods Methods using Derived Input

More information

PENALIZED PRINCIPAL COMPONENT REGRESSION. Ayanna Byrd. (Under the direction of Cheolwoo Park) Abstract

PENALIZED PRINCIPAL COMPONENT REGRESSION. Ayanna Byrd. (Under the direction of Cheolwoo Park) Abstract PENALIZED PRINCIPAL COMPONENT REGRESSION by Ayanna Byrd (Under the direction of Cheolwoo Park) Abstract When using linear regression problems, an unbiased estimate is produced by the Ordinary Least Squares.

More information

Iterative Selection Using Orthogonal Regression Techniques

Iterative Selection Using Orthogonal Regression Techniques Iterative Selection Using Orthogonal Regression Techniques Bradley Turnbull 1, Subhashis Ghosal 1 and Hao Helen Zhang 2 1 Department of Statistics, North Carolina State University, Raleigh, NC, USA 2 Department

More information

Regularization and Variable Selection via the Elastic Net

Regularization and Variable Selection via the Elastic Net p. 1/1 Regularization and Variable Selection via the Elastic Net Hui Zou and Trevor Hastie Journal of Royal Statistical Society, B, 2005 Presenter: Minhua Chen, Nov. 07, 2008 p. 2/1 Agenda Introduction

More information

DIMENSION REDUCTION OF THE EXPLANATORY VARIABLES IN MULTIPLE LINEAR REGRESSION. P. Filzmoser and C. Croux

DIMENSION REDUCTION OF THE EXPLANATORY VARIABLES IN MULTIPLE LINEAR REGRESSION. P. Filzmoser and C. Croux Pliska Stud. Math. Bulgar. 003), 59 70 STUDIA MATHEMATICA BULGARICA DIMENSION REDUCTION OF THE EXPLANATORY VARIABLES IN MULTIPLE LINEAR REGRESSION P. Filzmoser and C. Croux Abstract. In classical multiple

More information

MEANINGFUL REGRESSION COEFFICIENTS BUILT BY DATA GRADIENTS

MEANINGFUL REGRESSION COEFFICIENTS BUILT BY DATA GRADIENTS Advances in Adaptive Data Analysis Vol. 2, No. 4 (2010) 451 462 c World Scientific Publishing Company DOI: 10.1142/S1793536910000574 MEANINGFUL REGRESSION COEFFICIENTS BUILT BY DATA GRADIENTS STAN LIPOVETSKY

More information

Technical Report No. 1/12, February 2012 JACKKNIFING THE RIDGE REGRESSION ESTIMATOR: A REVISIT Mansi Khurana, Yogendra P. Chaubey and Shalini Chandra

Technical Report No. 1/12, February 2012 JACKKNIFING THE RIDGE REGRESSION ESTIMATOR: A REVISIT Mansi Khurana, Yogendra P. Chaubey and Shalini Chandra Technical Report No. 1/12, February 2012 JACKKNIFING THE RIDGE REGRESSION ESTIMATOR: A REVISIT Mansi Khurana, Yogendra P. Chaubey and Shalini Chandra Jackknifing the Ridge Regression Estimator: A Revisit

More information

Research Article An Unbiased Two-Parameter Estimation with Prior Information in Linear Regression Model

Research Article An Unbiased Two-Parameter Estimation with Prior Information in Linear Regression Model e Scientific World Journal, Article ID 206943, 8 pages http://dx.doi.org/10.1155/2014/206943 Research Article An Unbiased Two-Parameter Estimation with Prior Information in Linear Regression Model Jibo

More information

Regression Analysis for Data Containing Outliers and High Leverage Points

Regression Analysis for Data Containing Outliers and High Leverage Points Alabama Journal of Mathematics 39 (2015) ISSN 2373-0404 Regression Analysis for Data Containing Outliers and High Leverage Points Asim Kumer Dey Department of Mathematics Lamar University Md. Amir Hossain

More information

Generalized Elastic Net Regression

Generalized Elastic Net Regression Abstract Generalized Elastic Net Regression Geoffroy MOURET Jean-Jules BRAULT Vahid PARTOVINIA This work presents a variation of the elastic net penalization method. We propose applying a combined l 1

More information

Biostatistics Advanced Methods in Biostatistics IV

Biostatistics Advanced Methods in Biostatistics IV Biostatistics 140.754 Advanced Methods in Biostatistics IV Jeffrey Leek Assistant Professor Department of Biostatistics jleek@jhsph.edu Lecture 12 1 / 36 Tip + Paper Tip: As a statistician the results

More information

Ill-conditioning and multicollinearity

Ill-conditioning and multicollinearity Linear Algebra and its Applications 2 (2000) 295 05 www.elsevier.com/locate/laa Ill-conditioning and multicollinearity Fikri Öztürk a,, Fikri Akdeniz b a Department of Statistics, Ankara University, Ankara,

More information

Data Analysis and Machine Learning Lecture 12: Multicollinearity, Bias-Variance Trade-off, Cross-validation and Shrinkage Methods.

Data Analysis and Machine Learning Lecture 12: Multicollinearity, Bias-Variance Trade-off, Cross-validation and Shrinkage Methods. TheThalesians Itiseasyforphilosopherstoberichiftheychoose Data Analysis and Machine Learning Lecture 12: Multicollinearity, Bias-Variance Trade-off, Cross-validation and Shrinkage Methods Ivan Zhdankin

More information

Comparisons of penalized least squares. methods by simulations

Comparisons of penalized least squares. methods by simulations Comparisons of penalized least squares arxiv:1405.1796v1 [stat.co] 8 May 2014 methods by simulations Ke ZHANG, Fan YIN University of Science and Technology of China, Hefei 230026, China Shifeng XIONG Academy

More information

Contents. 1 Review of Residuals. 2 Detecting Outliers. 3 Influential Observations. 4 Multicollinearity and its Effects

Contents. 1 Review of Residuals. 2 Detecting Outliers. 3 Influential Observations. 4 Multicollinearity and its Effects Contents 1 Review of Residuals 2 Detecting Outliers 3 Influential Observations 4 Multicollinearity and its Effects W. Zhou (Colorado State University) STAT 540 July 6th, 2015 1 / 32 Model Diagnostics:

More information

RXshrink_in_R. Generalized Ridge and Least Angle Regression estimates most likely to have minimum MSE risk

RXshrink_in_R. Generalized Ridge and Least Angle Regression estimates most likely to have minimum MSE risk RXshrink_in_R Generalized Ridge and Least Angle Regression estimates most likely to have minimum MSE risk vignette-like documentation for R-package RXshrink version 1.1 - November 2018 Personal computing

More information

Business Statistics. Tommaso Proietti. Model Evaluation and Selection. DEF - Università di Roma 'Tor Vergata'

Business Statistics. Tommaso Proietti. Model Evaluation and Selection. DEF - Università di Roma 'Tor Vergata' Business Statistics Tommaso Proietti DEF - Università di Roma 'Tor Vergata' Model Evaluation and Selection Predictive Ability of a Model: Denition and Estimation We aim at achieving a balance between parsimony

More information

A complex version of the LASSO algorithm and its application to beamforming

A complex version of the LASSO algorithm and its application to beamforming The 7 th International Telecommunications Symposium ITS 21) A complex version of the LASSO algorithm and its application to beamforming José F. de Andrade Jr. and Marcello L. R. de Campos Program of Electrical

More information

Dr. Maddah ENMG 617 EM Statistics 11/28/12. Multiple Regression (3) (Chapter 15, Hines)

Dr. Maddah ENMG 617 EM Statistics 11/28/12. Multiple Regression (3) (Chapter 15, Hines) Dr. Maddah ENMG 617 EM Statistics 11/28/12 Multiple Regression (3) (Chapter 15, Hines) Problems in multiple regression: Multicollinearity This arises when the independent variables x 1, x 2,, x k, are

More information

ADDRESSING MULTICOLLINEARITY IN REGRESSION MODELS: A RIDGE REGRESSION APPLICATION 1

ADDRESSING MULTICOLLINEARITY IN REGRESSION MODELS: A RIDGE REGRESSION APPLICATION 1 Vol. 6, No. 1, Summer 2017 2012 Published by JSES. ADDRESSING MULTICOLLINEARITY IN REGRESSION MODELS: A RIDGE REGRESSION APPLICATION 1 Ali BAGER a, Monica ROMAN b, Meshal ALGELIDH c, Bahr MOHAMMED d Abstract

More information

Ridge Regression. Chapter 335. Introduction. Multicollinearity. Effects of Multicollinearity. Sources of Multicollinearity

Ridge Regression. Chapter 335. Introduction. Multicollinearity. Effects of Multicollinearity. Sources of Multicollinearity Chapter 335 Introduction is a technique for analyzing multiple regression data that suffer from multicollinearity. When multicollinearity occurs, least squares estimates are unbiased, but their variances

More information

shrink: a vignette for the RXshrink R-package

shrink: a vignette for the RXshrink R-package shrink: a vignette for the RXshrink R-package Generalized Ridge and Least Angle Regression Version 1.0-7, December 2011 License: GPL (>= 2) Personal computing treatments for your data analysis infirmities

More information

SOME APPLICATIONS: NONLINEAR REGRESSIONS BASED ON KERNEL METHOD IN SOCIAL SCIENCES AND ENGINEERING

SOME APPLICATIONS: NONLINEAR REGRESSIONS BASED ON KERNEL METHOD IN SOCIAL SCIENCES AND ENGINEERING SOME APPLICATIONS: NONLINEAR REGRESSIONS BASED ON KERNEL METHOD IN SOCIAL SCIENCES AND ENGINEERING Antoni Wibowo Farewell Lecture PPI Ibaraki 27 June 2009 EDUCATION BACKGROUNDS Dr.Eng., Social Systems

More information

A Practical Guide for Creating Monte Carlo Simulation Studies Using R

A Practical Guide for Creating Monte Carlo Simulation Studies Using R International Journal of Mathematics and Computational Science Vol. 4, No. 1, 2018, pp. 18-33 http://www.aiscience.org/journal/ijmcs ISSN: 2381-7011 (Print); ISSN: 2381-702X (Online) A Practical Guide

More information

Ridge Regression Estimation Approach to Measurement Error Model

Ridge Regression Estimation Approach to Measurement Error Model Ridge Regression Estimation Approach to Measurement Error Model A.K.Md. Ehsanes Saleh Carleton University, Ottawa CANADA) E-mail: esaleh@math.carleton.ca Shalabh Department of Mathematics & Statistics

More information

Linear Regression Linear Regression with Shrinkage

Linear Regression Linear Regression with Shrinkage Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle

More information

ECON 3150/4150, Spring term Lecture 7

ECON 3150/4150, Spring term Lecture 7 ECON 3150/4150, Spring term 2014. Lecture 7 The multivariate regression model (I) Ragnar Nymoen University of Oslo 4 February 2014 1 / 23 References to Lecture 7 and 8 SW Ch. 6 BN Kap 7.1-7.8 2 / 23 Omitted

More information

Or How to select variables Using Bayesian LASSO

Or How to select variables Using Bayesian LASSO Or How to select variables Using Bayesian LASSO x 1 x 2 x 3 x 4 Or How to select variables Using Bayesian LASSO x 1 x 2 x 3 x 4 Or How to select variables Using Bayesian LASSO On Bayesian Variable Selection

More information

COMS 4771 Lecture Fixed-design linear regression 2. Ridge and principal components regression 3. Sparse regression and Lasso

COMS 4771 Lecture Fixed-design linear regression 2. Ridge and principal components regression 3. Sparse regression and Lasso COMS 477 Lecture 6. Fixed-design linear regression 2. Ridge and principal components regression 3. Sparse regression and Lasso / 2 Fixed-design linear regression Fixed-design linear regression A simplified

More information

Mitigating collinearity in linear regression models using ridge, surrogate and raised estimators

Mitigating collinearity in linear regression models using ridge, surrogate and raised estimators STATISTICS RESEARCH ARTICLE Mitigating collinearity in linear regression models using ridge, surrogate and raised estimators Diarmuid O Driscoll 1 and Donald E. Ramirez 2 * Received: 21 October 2015 Accepted:

More information

Preliminary test almost unbiased ridge estimator in a linear regression model with multivariate Student-t errors

Preliminary test almost unbiased ridge estimator in a linear regression model with multivariate Student-t errors ACTA ET COMMENTATIONES UNIVERSITATIS TARTUENSIS DE MATHEMATICA Volume 1, Number 1, 011 Available online at www.math.ut.ee/acta/ Preliminary test almost unbiased ridge estimator in a linear regression model

More information

MS-C1620 Statistical inference

MS-C1620 Statistical inference MS-C1620 Statistical inference 10 Linear regression III Joni Virta Department of Mathematics and Systems Analysis School of Science Aalto University Academic year 2018 2019 Period III - IV 1 / 32 Contents

More information

MSE Performance and Minimax Regret Significance Points for a HPT Estimator when each Individual Regression Coefficient is Estimated

MSE Performance and Minimax Regret Significance Points for a HPT Estimator when each Individual Regression Coefficient is Estimated This article was downloaded by: [Kobe University] On: 8 April 03, At: 8:50 Publisher: Taylor & Francis Informa Ltd Registered in England and Wales Registered Number: 07954 Registered office: Mortimer House,

More information

A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables

A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables Niharika Gauraha and Swapan Parui Indian Statistical Institute Abstract. We consider the problem of

More information

A New Asymmetric Interaction Ridge (AIR) Regression Method

A New Asymmetric Interaction Ridge (AIR) Regression Method A New Asymmetric Interaction Ridge (AIR) Regression Method by Kristofer Månsson, Ghazi Shukur, and Pär Sölander The Swedish Retail Institute, HUI Research, Stockholm, Sweden. Deartment of Economics and

More information

Inference After Variable Selection

Inference After Variable Selection Department of Mathematics, SIU Carbondale Inference After Variable Selection Lasanthi Pelawa Watagoda lasanthi@siu.edu June 12, 2017 Outline 1 Introduction 2 Inference For Ridge and Lasso 3 Variable Selection

More information

Bootstrapping Heteroskedasticity Consistent Covariance Matrix Estimator

Bootstrapping Heteroskedasticity Consistent Covariance Matrix Estimator Bootstrapping Heteroskedasticity Consistent Covariance Matrix Estimator by Emmanuel Flachaire Eurequa, University Paris I Panthéon-Sorbonne December 2001 Abstract Recent results of Cribari-Neto and Zarkos

More information

A Significance Test for the Lasso

A Significance Test for the Lasso A Significance Test for the Lasso Lockhart R, Taylor J, Tibshirani R, and Tibshirani R Ashley Petersen May 14, 2013 1 Last time Problem: Many clinical covariates which are important to a certain medical

More information

Machine Learning for OR & FE

Machine Learning for OR & FE Machine Learning for OR & FE Regression II: Regularization and Shrinkage Methods Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com

More information

An Modified PLSR Method in Prediction

An Modified PLSR Method in Prediction Journal of Data Science 4(2006), 257-274 An Modified PLSR Method in Prediction Bo Cheng 1 and Xizhi Wu 2 1 Altair Engineering Company and 2 Renmin University of China Abstract: Among many statistical methods

More information

Institute of Statistics Mimeo Series No Simultaneous regression shrinkage, variable selection and clustering of predictors with OSCAR

Institute of Statistics Mimeo Series No Simultaneous regression shrinkage, variable selection and clustering of predictors with OSCAR DEPARTMENT OF STATISTICS North Carolina State University 2501 Founders Drive, Campus Box 8203 Raleigh, NC 27695-8203 Institute of Statistics Mimeo Series No. 2583 Simultaneous regression shrinkage, variable

More information

Bayesian Grouped Horseshoe Regression with Application to Additive Models

Bayesian Grouped Horseshoe Regression with Application to Additive Models Bayesian Grouped Horseshoe Regression with Application to Additive Models Zemei Xu, Daniel F. Schmidt, Enes Makalic, Guoqi Qian, and John L. Hopper Centre for Epidemiology and Biostatistics, Melbourne

More information

Influencial Observations in Ridge Regression

Influencial Observations in Ridge Regression Influencial Observations in Ridge Regression by Robert L. Obenchain We augment traditional trace displays with dynamic graphics that can reveal dramatic changes in influence/outliers/leverages of observations

More information

Matematické Metody v Ekonometrii 7.

Matematické Metody v Ekonometrii 7. Matematické Metody v Ekonometrii 7. Multicollinearity Blanka Šedivá KMA zimní semestr 2016/2017 Blanka Šedivá (KMA) Matematické Metody v Ekonometrii 7. zimní semestr 2016/2017 1 / 15 One of the assumptions

More information

Selection Criteria Based on Monte Carlo Simulation and Cross Validation in Mixed Models

Selection Criteria Based on Monte Carlo Simulation and Cross Validation in Mixed Models Selection Criteria Based on Monte Carlo Simulation and Cross Validation in Mixed Models Junfeng Shang Bowling Green State University, USA Abstract In the mixed modeling framework, Monte Carlo simulation

More information

Lecture 1: OLS derivations and inference

Lecture 1: OLS derivations and inference Lecture 1: OLS derivations and inference Econometric Methods Warsaw School of Economics (1) OLS 1 / 43 Outline 1 Introduction Course information Econometrics: a reminder Preliminary data exploration 2

More information

Link to Paper. The latest iteration can be found at:

Link to Paper. The latest iteration can be found at: Link to Paper Introduction The latest iteration can be found at: http://learneconometrics.com/pdf/gc2017/collin_gretl_170523.pdf BKW dignostics in GRETL: Interpretation and Performance Oklahoma State University

More information

Linear Regression Linear Regression with Shrinkage

Linear Regression Linear Regression with Shrinkage Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle

More information

Machine Learning. Regression basics. Marc Toussaint University of Stuttgart Summer 2015

Machine Learning. Regression basics. Marc Toussaint University of Stuttgart Summer 2015 Machine Learning Regression basics Linear regression, non-linear features (polynomial, RBFs, piece-wise), regularization, cross validation, Ridge/Lasso, kernel trick Marc Toussaint University of Stuttgart

More information

PARTIAL RIDGE REGRESSION 1

PARTIAL RIDGE REGRESSION 1 1 2 This work was supported by NSF Grants GU-2059 and GU-19568 and by U.S. Air Force Grant No. AFOSR-68-1415. On leave from Punjab Agricultural University (India). Reproduction in whole or in part is permitted

More information

Effect of outliers on the variable selection by the regularized regression

Effect of outliers on the variable selection by the regularized regression Communications for Statistical Applications and Methods 2018, Vol. 25, No. 2, 235 243 https://doi.org/10.29220/csam.2018.25.2.235 Print ISSN 2287-7843 / Online ISSN 2383-4757 Effect of outliers on the

More information

Stat 5100 Handout #26: Variations on OLS Linear Regression (Ch. 11, 13)

Stat 5100 Handout #26: Variations on OLS Linear Regression (Ch. 11, 13) Stat 5100 Handout #26: Variations on OLS Linear Regression (Ch. 11, 13) 1. Weighted Least Squares (textbook 11.1) Recall regression model Y = β 0 + β 1 X 1 +... + β p 1 X p 1 + ε in matrix form: (Ch. 5,

More information

Panel Threshold Regression Models with Endogenous Threshold Variables

Panel Threshold Regression Models with Endogenous Threshold Variables Panel Threshold Regression Models with Endogenous Threshold Variables Chien-Ho Wang National Taipei University Eric S. Lin National Tsing Hua University This Version: June 29, 2010 Abstract This paper

More information

ABSTRACT. POST, JUSTIN BLAISE. Methods to Improve Prediction Accuracy under Structural Constraints. (Under the direction of Howard Bondell.

ABSTRACT. POST, JUSTIN BLAISE. Methods to Improve Prediction Accuracy under Structural Constraints. (Under the direction of Howard Bondell. ABSTRACT POST, JUSTIN BLAISE. Methods to Improve Prediction Accuracy under Structural Constraints. (Under the direction of Howard Bondell.) Statisticians are often faced with the difficult task of model

More information

The RXshrink Package

The RXshrink Package The RXshrink Package October 20, 2005 Title Maximum Likelihood Shrinkage via Ridge or Least Angle Regression Version 1.0-2 Date 2005-10-19 Author Maintainer Depends R (>= 1.8.0), lars Identify and display

More information

Spatial Lasso with Application to GIS Model Selection. F. Jay Breidt Colorado State University

Spatial Lasso with Application to GIS Model Selection. F. Jay Breidt Colorado State University Spatial Lasso with Application to GIS Model Selection F. Jay Breidt Colorado State University with Hsin-Cheng Huang, Nan-Jung Hsu, and Dave Theobald September 25 The work reported here was developed under

More information

On High-Dimensional Cross-Validation

On High-Dimensional Cross-Validation On High-Dimensional Cross-Validation BY WEI-CHENG HSIAO Institute of Statistical Science, Academia Sinica, 128 Academia Road, Section 2, Nankang, Taipei 11529, Taiwan hsiaowc@stat.sinica.edu.tw 5 WEI-YING

More information

PAijpam.eu M ESTIMATION, S ESTIMATION, AND MM ESTIMATION IN ROBUST REGRESSION

PAijpam.eu M ESTIMATION, S ESTIMATION, AND MM ESTIMATION IN ROBUST REGRESSION International Journal of Pure and Applied Mathematics Volume 91 No. 3 2014, 349-360 ISSN: 1311-8080 (printed version); ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu doi: http://dx.doi.org/10.12732/ijpam.v91i3.7

More information

Bootstrap & Confidence/Prediction intervals

Bootstrap & Confidence/Prediction intervals Bootstrap & Confidence/Prediction intervals Olivier Roustant Mines Saint-Étienne 2017/11 Olivier Roustant (EMSE) Bootstrap & Confidence/Prediction intervals 2017/11 1 / 9 Framework Consider a model with

More information

A Comparison Study of Principle Component Regression, Partial Least Squares Regression and Ridge Regression with Application to FTIR Data

A Comparison Study of Principle Component Regression, Partial Least Squares Regression and Ridge Regression with Application to FTIR Data A Comparison Study of Principle Component Regression, Partial Least Squares Regression and Ridge Regression with Application to FTIR Data Author: Ying Li Supervisor: Dietrich von Rosen Master thesis in

More information