COMBINING THE LIU-TYPE ESTIMATOR AND THE PRINCIPAL COMPONENT REGRESSION ESTIMATOR
|
|
- Augustine Willis McDaniel
- 5 years ago
- Views:
Transcription
1 Noname manuscript No. (will be inserted by the editor) COMBINING THE LIU-TYPE ESTIMATOR AND THE PRINCIPAL COMPONENT REGRESSION ESTIMATOR Deniz Inan Received: date / Accepted: date Abstract In this study a new two-parameter estimator which includes the ordinary least squares (OLS), the principal components regression (PCR) and the Liu-type estimator is proposed. Conditions for the superiority of this new estimator over the PCR, r-k class estimator and Liu-type estimator are derived. Furthermore the performance of this estimator is compared with the other estimators in different conditions with simulation studies. Keywords Liu-type estimator; r-k class estimators; principal component regression estimator; ridge estimator; multicollinearity. Mathematics Subject Classification (0) 62J07 1 Introduction Multicollinearity can be a problem for many empirical researchers, and this problem will influence how you are able to interpret the coefficients in your model. It is well known that especially ordinary least squares (OLS) estimator can be effected seriously by the presence of multicollinearity so parameter estimates have large sampling variances. To deal with this problem different methods have been proposed such as ordinary ridge regression (ORR) estimator (Hoerl and Kennard, 1970), principal component regression (PCR) estimator (Draper and Smith, 1981), least absolute shrinkage and selection operator (LASSO) estimator (Tibshirani, 1996), least angle regression (LARS) estimator (Efron et al.,4) and Liu-type estimator (Liu, 3). Also combinations of these estimators have been proposed with the expectation that the combination of different estimators might inherit the advantages of both. Baye and D. Inan Marmara University Department of Statistics Istanbul Turkey Tel.: Fax: denizlukuslu@marmara.edu.tr
2 2 Deniz Inan Parker (1984) combined ORR estimator with PCR estimator and proposed r-k class estimator which provides an alternative method of dealing with multicollinearity. The r-k class estimator was compared with the ORR estimator and OLS estimator according to scalar mean square error (MSE) criteria in Nomura and Ohkuba (1985). Sarkar (1996) studied the properties of r-k class estimators in the sense of matrix mean square error. Differently, Kaciranlar and Sakallioglu (1) proposed r-d class estimators by combining the Liu and PCR estimators and compared this new estimator with the OLS, PCR and Liu estimators according to MSE criteria. Conditions for superiority of the r-d class estimators over these estimators and the r-k class estimators are derived in Ozkale and Kaciranlar (7). Aim of this paper is to introduce a new class of estimators which include OLS, PCR and Liu-type estimators that can be used as an alternative method for combating multicollinearity and compare this estimator with the OLS, ORR, Liu-type, PCR and r-k class estimators in the sense of MSE criteria. 2 A new class of estimators Consider the linear regression model Y = Xβ +ǫ (2.1) where X is a known n p matrix of standardized regressors with rank(x) = p, Y is a n 1 response vector, β is a vector of parameters and ǫ is a n 1 vector of errors with E(ǫ) = 0 and Cov(ǫ) = σ 2 I. Let us rewrite model (2.1) in canonical form Y = Zα+ǫ where Z = Xθ, α = θ β and θ is the p p orthogonal matrix whose columns constitute the eigenvectors of X X. Then, Z Z = θ X Xθ = Λ = diag(λ 1,,λ p ) where λ 1 λ 2 λ p > 0 are the ordered eigenvalues of X X. The columns of Z, which define the new set of orthogonal regressors Z = [Z 1,Z 2,,Z p ] are referred to as principal components and the OLS estimator of α is ˆα = (Z Z) 1 Z y = Λ 1 Z y. The PCR regression approach combats multicollinearity by using less than the full set of principal components in the model. Suppose that the last p r of the eigenvalues are approximately near zero. In PCR the principal components corresponding to near-zero eigenvalues are removed from the analysis and OLS is applied to the remaining components. That is, ˆα PC = Bˆα
3 Title Suppressed Due to Excessive Length 3 where B is a diagonal matrix whose diagonal elements are as b 1,1 = b 2,2 = = b r,r = 1 and b r+1,r+1 = b r+2,r+2 = = b p,p = 0 (Montgomery et. al., 1). Likewise in terms of original regressors, the PCR estimator can be given as ˆβ PC = θα PC ˆ (2.2). Let θ r be the remaining eigenvectors of X X after having deleted p r of columns of θ which correspond to the near-zero eigenvalues. Thus, θ rx Xθ r = Λ r = diag(λ 1,,λ r ) and PCR estimator given in (2.2), can also be given as in the form, ˆβ PC = θ r (θ rx Xθ r ) 1 θ rx y and r-k class estimator is ˆβ r (k) = θ r (θ rx Xθ r +ki) 1 θ rx y where r p and k 0. A new class of estimators which includes the OLS, PCR and Liu-Type estimators as a special cases is defined as follows. Definition 1. The r-(k,d) class estimator of β in (2.1) is ˆβ r (k,d) = θ r (θ rx Xθ r +ki) 1 (θ rx Xθ r di r )θ rˆβ PC, where k 0 and < d < are shrinkage and biasing parameters that must be chosen by the researcher and 0 < r p. Special cases of r-(k,d) class estimators can be given as: ˆβ p (0,0) = ˆβ LS = (X X) 1 X Y, the OLS estimator, ˆβ r (k, k) = ˆβ PC = θ r (θ rx Xθ r ) 1 θ rx y, the PCR estimator, ˆβ r (k,0) = ˆβ r (k) = θ r (θ rx Xθ r +ki) 1 θ rx y, the r-k class estimator, ˆβ p (k,d) = ˆβ(k,d) = (X X +ki) 1 (X X di)ˆβ, Liu-Type estimator. The MSEs ˆβ r (k,d), ˆβ LS, ˆβ PC, ˆβr (k), ˆβ(k,d) are given MSE( ˆβ r (k,d)) = σ 2 (λ i d) 2 λ i (λ i +k) 2 + ( d+k p λ i +k )2 αi 2 + αi 2 (2.3) respectively. MSE(ˆβ LS ) = σ 2 p MSE(ˆβ PC ) = σ 2 1 λ i 1 λ i + p i=r+1 MSE( ˆβ r (k)) = σ 2 (λ i ) (λ i +k) 2 + k ( λ i +k )2 αi 2 + α 2 i i=r+1 p i=r+1 p MSE(ˆβ(k,d)) = σ 2 (λ i d) 2 p λ i (λ i +k) 2 + ( d+k λ i +k )2 αi, 2 α 2 i
4 4 Deniz Inan Theorem 1. For any k > 0 and 0 < r p, if the biasing parameter chosen as d = ((σ 2 kαi)/(λ 2 i +k) 2 )/ ((σ 2 +λ i αi)/(λ 2 i (λ i +k) 2 )) then MSE( ˆβ r (k,d)) MSE(ˆβ PC ) and MSE( ˆβ r (k,d)) MSE( ˆβ r (k)). Proof. From (2.1.1) it is easy to see that, for fixed k, MSE( ˆβ r (k,d)) is a quadratic function of d. If the first derivative of this function is taken and solved for d, the optimum value of d which minimizes MSE( ˆβ r (k,d)) can be obtained as d opt = ((σ 2 kαi)/(λ 2 i +k) 2 )/ ((σ 2 +λ i αi)/(λ 2 i (λ i +k) 2 )) (2.4). In view of MSE( ˆβ r (k, k)) = MSE(ˆβ PC ), we have MSE( ˆβ r (k,d opt )) MSE(ˆβ PC ). And the equality occurs when d opt = k. Similarly in view of MSE( ˆβ r (k,0)) = MSE( ˆβ r (k)),wehavemse( ˆβ r (k,d opt )) MSE( ˆβ r (k)).equality occurs when d opt = 0. Theorem 2. For any k > 0, 0 < r p and i = 1,,p, when λ 1 < d < min[(σ 2 +α 2 i( 2k λ i ))λ i /(σ 2 +α 2 iλ i )] we have MSE( ˆβ r (k,d)) MSE(ˆβ(k,d)). Proof. For a given r the solution of MSE( ˆβ r (k,d)) MSE(ˆβ r 1 (k,d)) > 0 for d can be obtained as λ r < d < (σ 2 +α 2 r( 2k λ r ))λ r /(σ 2 +α 2 rλ r ). Generalizing this solution, if we choose d as, λ 1 < d < min[(σ 2 +α 2 i( 2k λ i ))λ i /(σ 2 +α 2 iλ i )] we get MSE( ˆβ r (k,d)) MSE(ˆβ r 1 (k,d)) > 0 for all r = 1,,p. As a result, MSE( ˆβ p (k,d)) > MSE(ˆβ p 1 (k,d)) > > MSE(ˆβ r (k,d)) MSE(ˆβ r (k,d)) < MSE(ˆβ(k,d)). This completes the proof of Theorem 2 (See Appendix).
5 Title Suppressed Due to Excessive Length 5 3 Simulation Results Monte Carlo simulations were used to illustrate the behaviour of the proposed estimator. Simulations were carried out under different degrees of multicollinearity. In first simulation study the number of regressors was chosen to be six and the parameters were set to be (22,1.4,1.2, 1.5, 0.2,0.7,1.7). The size of the sample n was set to be 30, 50 and 70. The regressors were generated from x i,1 = (1 ϕ 2 ) 1/2 w i,1 +ϕw i,5 +20 x i,2 = (1 ϕ 2 ) 1/2 w i,2 +ϕw i,5 +45 x i,3 = (1 ρ 2 ) 1/2 w i,3 +ρw i,6 24 x i,4 = (1 ρ 2 ) 1/2 w i,4 +ρw i,6 +7 x i,5 = (1 ρ 2 ) 1/2 w i,3 +ρw i,2 +19 x i,6 = w i,2 14 wherew i,j aregeneratedindependentlyfromn(0,2).ϕandρcorrespondtothe correlations between x i,1, x i,2 and x i,3, x i,4 are ϕ 2 and ρ 2 respectively (Gibbons, 1981). Three different sets of correlations were considered corresponding to ϕ = 0.95 and ρ = 0.95, ϕ = 0.98 and ρ = 0.95, ϕ = 0.95 and ρ = Simulations were repeated times. In each experiment inhomogeneous model was applied to the data, estimated error variances of each estimator and the following estimates were computed: 1.OLS : Ordinary least square estimator 2.ORR: Ordinary ridge regression estimator, k = ˆσ LS 2 /ˆβ LSˆβ LS (Schafer et 3.Liu: al., 1984) Liu-type estimator, k = ˆσ LS 2 /ˆβ LSˆβ LS, ˆd = p ((ˆσ2 kˆα i 2)/(λ i +k) 2 )/ p ((ˆσ2 +λ iˆα i 2)/(λ i(λ i +k) 2 )) (Liu, 3) and ˆβ = ˆβ LS 4.PCR: Principal component regression estimator 5.r-k: r-k class estimator, k = ˆσ LS 2 /ˆβ LSˆβ LS 6.r-(k,d): r-(k,d) class estimator, k = ˆσ LS 2 /ˆβ LSˆβ LS, ˆd = r ((ˆσ2 kˆα i 2)/ (λ i +k) 2 )/ r ((ˆσ2 +λ iˆα i 2)/(λ i(λ i +k) 2 )). Unfortunately,thed opt valuesgiveninthetheoremsdependonunknownmodel parameters and so, for practical purposes, these unknown parameters were decided to replace with their suitable estimates. If we substitute α and σ with their ORR estimates (ˆα R and ˆσ R ) in (2.2.1), we obtain ˆd opt = ((ˆσ R 2 kˆα R(i) 2 )/(λ i +k) 2 )/ ((ˆσ R 2 +λ iˆα R(i) 2 )/(λ i(λ i +k) 2 ))
6 6 Deniz Inan Table 1 EMSE values for six estimators (continuous regressors) ϕ ρ n OLS ORR Liu PCR r-k r-(k,d) Table 2 Estimated error variances for six estimators (continuous regressors) ϕ ρ n OLS ORR Liu PCR r-k r-(k,d) as an operational estimator for d opt. In simulation studies this estimator was used for r-(k,d) class estimator and considering the proposal of Liu (3), ˆd opt = p ((ˆσ R 2 kˆα R(i) 2 )/(λ i +k) 2 )/ p ((ˆσ R 2 +λ iˆα R(i) 2 )/(λ i(λ i +k) 2 )) was used for Liu-type estimator. The estimated MSEs (EMSEs) and means of the estimated error variances were computed for each of the above six estimators. The results are presented in Table 1 and Table 2. As expected, our proposed estimators have smaller EMSEs than those of the other estimators. Although the EMSEs of all estimators increased along with multicollinearity, the increase in EMSEs for proposed estimator was not as much as the others. It can be concluded that the effect of multicollinearity on the other estimators is more pronounced than proposed estimator. The EMSE reduction is much better when the multicollinearity is high. Also there is no significant difference among the estimators according to estimated error variances.
7 Title Suppressed Due to Excessive Length 7 Table 3 EMSE values for six estimator (both continuous and ordinal regressors) ϕ ρ n OLS ORR Liu PCR r-k r-(k,d) The case when the model has both continuous and ordinal regressors was also considered. For this case, six regressors were generated as in the first simulation study, the parameters were set to be (25,2.3,4.1, 1.5, 0.9,0.9,1.7) and the size of the sample n was set to be 30, 50 and 70. Two different sets of correlations were considered corresponding to ϕ = 0.95 and ρ = 0.95, ϕ = 0.95 and ρ = After six continuous regressors were generated, the first two of them were discretized in order to make them ordinal. In this way the impact of collinearity was slightly reduced. Simulations were repeated times. The results are presented in Table 3. As seen from the Table 3, in all cases r-(k,d) class estimator has smaller EMSE than the other estimators. 4 Numerical Example In this section, the acetylene data described and analysed in detail by Marquardt (1980) was employed. This data set is a typical set of response surface data for which a full quadratic model in three regressors is often considered to be an appropriate candidate model. The standardized full quadratic model for the data is y = β 0 +β 1 x 1 +β 2 x 2 +β 3 x 3 +β 1,2 x 1 x 2 +β 1,3 x 1 x 3 +β 2,3 x 2 x 3 +β 1,1 x 2 1+β 2,2 x 2 2+β 3,3 x 2 3+ǫ wherey=percentageofconversion,x 1 =(temperature )/80.623,x 2 =(H 2 (nheptane)-12.44)/5.662 and x 3 =(contact time )/ Each of the six estimators was applied to these data. k, r, ˆd opt and ˆd opt were determined respectively as , 6, and The parameter estimations are given in Table 4. Under this model, the condition number is κ = λ max λ min = = 42048
8 8 Deniz Inan Table 4 Parameter estimations for the acetylene data ˆβ 0 ˆβ1 ˆβ2 ˆβ3 ˆβ4 ˆβ5 ˆβ6 ˆβ7 ˆβ8 ˆβ9 OLS ORR Liu PCR r-k r-(k,d) Table 5 EMSEs, estimated parameter variances and biases for the acetylene data Estimated parameter Estimated parameter EMSE variance bias OLS ORR Liu PCR r-k r-(k,d) which indicates severe multicollinearity. In this case, the OLS estimator does not provide good estimations. Although the OLS estimator achieves an optimum fit to the estimation data, it does not provide a good prediction for the new data. Therefore, we can not trust the OLS estimator under this model. The bootstrap sampling method was used for these data to estimate the MSE values of relevant estimators and illustrate the distributional properties of the parameters. For this purpose, bootstrap samples were generated, and parameter estimates of six estimators were computed for each of these samples. The mean of the OLS estimates was considered as β, and EMSEs, estimated parameter variances and estimated biases were computed as given in Table 5. Because of the severe multicollinearity, small changes in the data cause large differences in the parameter estimates. To illustrate this situation, for each estimator, histograms of ˆβ 9 are given in Figure 1 as an example. As seen in Figure 1, the OLS, ORR and Liu-type estimates vary over a large interval. Compared to the OLS, the distributions for the ORR and Liutype estimations have stronger peaks and more rapid decays, but they still have heavy tails. Therefore, we can not say that these estimates are stable. The PCR, r-k and r-(k,d) class estimations are distributed within a narrower interval according to the OLS, ORR and Liu-type estimates. In addition, the distribution of r-(k,d) class estimations has a stronger peak and more rapid decay according to the distributions of the PCR and r-k class estimations. Considering all of these histograms, we can say that among the six estimators, the r-(k,d) class estimator gives the most stable estimates, indicating that this estimator provides a better prediction for the new data.
9 Title Suppressed Due to Excessive Length OLS ORR Liu PCR r k r k,d Fig. 1 Histograms of ˆβ 9 5 Summary In this paper conditions have been obtained for the superiority of the r-(k,d) class estimators over the PCR, r-k class and Liu-type estimators. It is seen that the conditions obtained in Theorem 1. and Theorem 2. depend on the unknown parameters. For practical purposes, thus, we have suggested to replace them with their ORR estimators. Simulation studies are constructed to decide whether the relevant conditions are satisfied. Simulation results obtained through the use of using ORR estimates are seen to be supporting our theoretical findings. Finally, considering these simulation studies it is concluded that especially for the severe multicollinearity case and small sample sizes r-(k,d) class estimator performs better than the other estimators in the sense of MSE. 6 Appendix MSE( ˆβ r (k,d)) MSE(ˆβ r 1 (k,d)) > 0 σ 2 ( r (λ i d) 2 /(λ i (λ i +k) 2 ) r 1 (λ i d) 2 /(λ i (λ i +k) 2 ))+ r ((d+k)/(λ i +k)) 2 α 2 i r 1 ((d+k)/(λ i +k)) 2 α 2 i + p i=r+1 α2 i p i=r α2 i > 0 σ 2 (λ r d) 2 /(λ r (λ r +k) 2 )+((d+k)/(λ r +k)) 2 α 2 r α 2 r > 0 λ r < d < (σ 2 +α 2 r( 2k λ r ))λ r /(σ 2 +α 2 rλ r ) Acknowledgements The author was supported by the Marmara University Scientific Research Project Unit (BAPKO, project number: FEN-D ).
10 10 Deniz Inan References 1. Baye M.R., Parker, D.F. (1984). Combining Ridge and Principal Component Regression: A Money Demand Illustration.Communications in Statistics-Theory and Methods 13, Draper, N.R., Smith, H.(1981). Applied Regression Analysis. John Willey & Sons. 3. Efron, B., Hastie, T., Johnstone, I., Tibshirani, R.(4). Least Angle Regression. The Annals of Statistics 32, Gibbons, D.G. (1981). A simulation study of some ridge estimators. Journal of the American Statistical Association, 76, Hoerl, A.E., Kennard, R.W.(1970). Ridge Regression: Biased Estimation for Nonorthogonal Problems. Technometrics 12, Kaciranlar S., Sakallioglu S. (1). Combining the Liu Estimator and the Principal Component Regression Estimator. Communications in Statistics-Theory and Methods 30, Liu K. (3).Using Liu Type Estimator to Combat Collinearity. Communications in Statistics-Theory and Methods 32, Marquardt, D.W.(1980) A Critique of Some Ridge Regression Methods. Journal of the American Statistical Association 75, Montgomery, D.C., Peck, A.E., Vining, G.G.(1) Introduction to the Linear Regression Analysis. John Willey & Sons Nomura M., Ohkubo T. (1985). A Note on Combining Ridge and Principal Component Regression. Communications in Statistics-Theory and Methods 14, Ozkale M.R., Kaciranlar S. (7). Superiorty of the r-d Class Estimator Over Some Estimators by the Mean Square Error Matrix Criterion. Statistics and Probability Letters 77, Sarkar,N.(1996). Mean Square Error Matrix Comparison of Some Estimators in Linear Regression with Multicollinearity. Statist.Probab.Lett. 30, Tibshirani, R.(1996). Regression Shrinkage and Selection via the Lasso. J.R. Statist. Soc.B 58,
Improved Liu Estimators for the Poisson Regression Model
www.ccsenet.org/isp International Journal of Statistics and Probability Vol., No. ; May 202 Improved Liu Estimators for the Poisson Regression Model Kristofer Mansson B. M. Golam Kibria Corresponding author
More informationEvaluation of a New Estimator
Pertanika J. Sci. & Technol. 16 (2): 107-117 (2008) ISSN: 0128-7680 Universiti Putra Malaysia Press Evaluation of a New Estimator Ng Set Foong 1 *, Low Heng Chin 2 and Quah Soon Hoe 3 1 Department of Information
More informationRidge Regression Revisited
Ridge Regression Revisited Paul M.C. de Boer Christian M. Hafner Econometric Institute Report EI 2005-29 In general ridge (GR) regression p ridge parameters have to be determined, whereas simple ridge
More informationMulticollinearity and A Ridge Parameter Estimation Approach
Journal of Modern Applied Statistical Methods Volume 15 Issue Article 5 11-1-016 Multicollinearity and A Ridge Parameter Estimation Approach Ghadban Khalaf King Khalid University, albadran50@yahoo.com
More informationComparison of Some Improved Estimators for Linear Regression Model under Different Conditions
Florida International University FIU Digital Commons FIU Electronic Theses and Dissertations University Graduate School 3-24-2015 Comparison of Some Improved Estimators for Linear Regression Model under
More informationMeasuring Local Influential Observations in Modified Ridge Regression
Journal of Data Science 9(2011), 359-372 Measuring Local Influential Observations in Modified Ridge Regression Aboobacker Jahufer 1 and Jianbao Chen 2 1 South Eastern University and 2 Xiamen University
More informationRobust model selection criteria for robust S and LT S estimators
Hacettepe Journal of Mathematics and Statistics Volume 45 (1) (2016), 153 164 Robust model selection criteria for robust S and LT S estimators Meral Çetin Abstract Outliers and multi-collinearity often
More informationEFFICIENCY of the PRINCIPAL COMPONENT LIU- TYPE ESTIMATOR in LOGISTIC REGRESSION
EFFICIENCY of the PRINCIPAL COMPONEN LIU- YPE ESIMAOR in LOGISIC REGRESSION Authors: Jibo Wu School of Mathematics and Finance, Chongqing University of Arts and Sciences, Chongqing, China, linfen52@126.com
More informationConfidence Intervals in Ridge Regression using Jackknife and Bootstrap Methods
Chapter 4 Confidence Intervals in Ridge Regression using Jackknife and Bootstrap Methods 4.1 Introduction It is now explicable that ridge regression estimator (here we take ordinary ridge estimator (ORE)
More informationRegularized Multiple Regression Methods to Deal with Severe Multicollinearity
International Journal of Statistics and Applications 21, (): 17-172 DOI: 1.523/j.statistics.21.2 Regularized Multiple Regression Methods to Deal with Severe Multicollinearity N. Herawati *, K. Nisa, E.
More informationRelationship between ridge regression estimator and sample size when multicollinearity present among regressors
Available online at www.worldscientificnews.com WSN 59 (016) 1-3 EISSN 39-19 elationship between ridge regression estimator and sample size when multicollinearity present among regressors ABSTACT M. C.
More informationResearch Article On the Stochastic Restricted r-k Class Estimator and Stochastic Restricted r-d Class Estimator in Linear Regression Model
Applied Mathematics, Article ID 173836, 6 pages http://dx.doi.org/10.1155/2014/173836 Research Article On the Stochastic Restricted r-k Class Estimator and Stochastic Restricted r-d Class Estimator in
More informationImproved Ridge Estimator in Linear Regression with Multicollinearity, Heteroscedastic Errors and Outliers
Journal of Modern Applied Statistical Methods Volume 15 Issue 2 Article 23 11-1-2016 Improved Ridge Estimator in Linear Regression with Multicollinearity, Heteroscedastic Errors and Outliers Ashok Vithoba
More informationEffects of Outliers and Multicollinearity on Some Estimators of Linear Regression Model
204 Effects of Outliers and Multicollinearity on Some Estimators of Linear Regression Model S. A. Ibrahim 1 ; W. B. Yahya 2 1 Department of Physical Sciences, Al-Hikmah University, Ilorin, Nigeria. e-mail:
More informationISyE 691 Data mining and analytics
ISyE 691 Data mining and analytics Regression Instructor: Prof. Kaibo Liu Department of Industrial and Systems Engineering UW-Madison Email: kliu8@wisc.edu Office: Room 3017 (Mechanical Engineering Building)
More informationRegularization: Ridge Regression and the LASSO
Agenda Wednesday, November 29, 2006 Agenda Agenda 1 The Bias-Variance Tradeoff 2 Ridge Regression Solution to the l 2 problem Data Augmentation Approach Bayesian Interpretation The SVD and Ridge Regression
More informationNonlinear Inequality Constrained Ridge Regression Estimator
The International Conference on Trends and Perspectives in Linear Statistical Inference (LinStat2014) 24 28 August 2014 Linköping, Sweden Nonlinear Inequality Constrained Ridge Regression Estimator Dr.
More informationEfficient Choice of Biasing Constant. for Ridge Regression
Int. J. Contemp. Math. Sciences, Vol. 3, 008, no., 57-536 Efficient Choice of Biasing Constant for Ridge Regression Sona Mardikyan* and Eyüp Çetin Department of Management Information Systems, School of
More informationRidge Regression and Ill-Conditioning
Journal of Modern Applied Statistical Methods Volume 3 Issue Article 8-04 Ridge Regression and Ill-Conditioning Ghadban Khalaf King Khalid University, Saudi Arabia, albadran50@yahoo.com Mohamed Iguernane
More informationA Modern Look at Classical Multivariate Techniques
A Modern Look at Classical Multivariate Techniques Yoonkyung Lee Department of Statistics The Ohio State University March 16-20, 2015 The 13th School of Probability and Statistics CIMAT, Guanajuato, Mexico
More informationBayes Estimators & Ridge Regression
Readings Chapter 14 Christensen Merlise Clyde September 29, 2015 How Good are Estimators? Quadratic loss for estimating β using estimator a L(β, a) = (β a) T (β a) How Good are Estimators? Quadratic loss
More informationResearch Article On the Weighted Mixed Almost Unbiased Ridge Estimator in Stochastic Restricted Linear Regression
Applied Mathematics Volume 2013, Article ID 902715, 10 pages http://dx.doi.org/10.1155/2013/902715 Research Article On the Weighted Mixed Almost Unbiased Ridge Estimator in Stochastic Restricted Linear
More informationAlternative Biased Estimator Based on Least. Trimmed Squares for Handling Collinear. Leverage Data Points
International Journal of Contemporary Mathematical Sciences Vol. 13, 018, no. 4, 177-189 HIKARI Ltd, www.m-hikari.com https://doi.org/10.1988/ijcms.018.8616 Alternative Biased Estimator Based on Least
More informationRegression Shrinkage and Selection via the Lasso
Regression Shrinkage and Selection via the Lasso ROBERT TIBSHIRANI, 1996 Presenter: Guiyun Feng April 27 () 1 / 20 Motivation Estimation in Linear Models: y = β T x + ɛ. data (x i, y i ), i = 1, 2,...,
More informationGeneralized Ridge Regression Estimator in Semiparametric Regression Models
JIRSS (2015) Vol. 14, No. 1, pp 25-62 Generalized Ridge Regression Estimator in Semiparametric Regression Models M. Roozbeh 1, M. Arashi 2, B. M. Golam Kibria 3 1 Department of Statistics, Faculty of Mathematics,
More informationA Comparison between Biased and Unbiased Estimators in Ordinary Least Squares Regression
Journal of Modern Alied Statistical Methods Volume Issue Article 7 --03 A Comarison between Biased and Unbiased Estimators in Ordinary Least Squares Regression Ghadban Khalaf King Khalid University, Saudi
More informationRidge Estimator in Logistic Regression under Stochastic Linear Restrictions
British Journal of Mathematics & Computer Science 15(3): 1-14, 2016, Article no.bjmcs.24585 ISSN: 2231-0851 SCIENCEDOMAIN international www.sciencedomain.org Ridge Estimator in Logistic Regression under
More informationRidge Estimation and its Modifications for Linear Regression with Deterministic or Stochastic Predictors
Ridge Estimation and its Modifications for Linear Regression with Deterministic or Stochastic Predictors James Younker Thesis submitted to the Faculty of Graduate and Postdoctoral Studies in partial fulfillment
More informationJournal of Asian Scientific Research COMBINED PARAMETERS ESTIMATION METHODS OF LINEAR REGRESSION MODEL WITH MULTICOLLINEARITY AND AUTOCORRELATION
Journal of Asian Scientific Research ISSN(e): 3-1331/ISSN(p): 6-574 journal homepage: http://www.aessweb.com/journals/5003 COMBINED PARAMETERS ESTIMATION METHODS OF LINEAR REGRESSION MODEL WITH MULTICOLLINEARITY
More informationGeneralized Maximum Entropy Estimators: Applications to the Portland Cement Dataset
The Open Statistics & Probability Journal 2011 3 13-20 13 Open Access Generalized Maximum Entropy Estimators: Applications to the Portland Cement Dataset Fikri Akdeniz a* Altan Çabuk b and Hüseyin Güler
More informationLarge Sample Theory For OLS Variable Selection Estimators
Large Sample Theory For OLS Variable Selection Estimators Lasanthi C. R. Pelawa Watagoda and David J. Olive Southern Illinois University June 18, 2018 Abstract This paper gives large sample theory for
More informationSOME NEW PROPOSED RIDGE PARAMETERS FOR THE LOGISTIC REGRESSION MODEL
IMPACT: International Journal of Research in Applied, Natural and Social Sciences (IMPACT: IJRANSS) ISSN(E): 2321-8851; ISSN(P): 2347-4580 Vol. 3, Issue 1, Jan 2015, 67-82 Impact Journals SOME NEW PROPOSED
More informationVariable Selection in Restricted Linear Regression Models. Y. Tuaç 1 and O. Arslan 1
Variable Selection in Restricted Linear Regression Models Y. Tuaç 1 and O. Arslan 1 Ankara University, Faculty of Science, Department of Statistics, 06100 Ankara/Turkey ytuac@ankara.edu.tr, oarslan@ankara.edu.tr
More informationLinear regression methods
Linear regression methods Most of our intuition about statistical methods stem from linear regression. For observations i = 1,..., n, the model is Y i = p X ij β j + ε i, j=1 where Y i is the response
More informationENHANCING THE EFFICIENCY OF THE RIDGE REGRESSION MODEL USING MONTE CARLO SIMULATIONS
www.arpapress.com/volumes/vol27issue1/ijrras_27_1_02.pdf ENHANCING THE EFFICIENCY OF THE RIDGE REGRESSION MODEL USING MONTE CARLO SIMULATIONS Rania Ahmed Hamed Mohamed Department of Statistics, Mathematics
More informationAPPLICATION OF RIDGE REGRESSION TO MULTICOLLINEAR DATA
Journal of Research (Science), Bahauddin Zakariya University, Multan, Pakistan. Vol.15, No.1, June 2004, pp. 97-106 ISSN 1021-1012 APPLICATION OF RIDGE REGRESSION TO MULTICOLLINEAR DATA G. R. Pasha 1 and
More informationThe prediction of house price
000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050
More informationDimension Reduction Methods
Dimension Reduction Methods And Bayesian Machine Learning Marek Petrik 2/28 Previously in Machine Learning How to choose the right features if we have (too) many options Methods: 1. Subset selection 2.
More informationLinear Regression Models. Based on Chapter 3 of Hastie, Tibshirani and Friedman
Linear Regression Models Based on Chapter 3 of Hastie, ibshirani and Friedman Linear Regression Models Here the X s might be: p f ( X = " + " 0 j= 1 X j Raw predictor variables (continuous or coded-categorical
More informationUsing Ridge Least Median Squares to Estimate the Parameter by Solving Multicollinearity and Outliers Problems
Modern Applied Science; Vol. 9, No. ; 05 ISSN 9-844 E-ISSN 9-85 Published by Canadian Center of Science and Education Using Ridge Least Median Squares to Estimate the Parameter by Solving Multicollinearity
More informationSTK-IN4300 Statistical Learning Methods in Data Science
Outline of the lecture STK-I4300 Statistical Learning Methods in Data Science Riccardo De Bin debin@math.uio.no Model Assessment and Selection Cross-Validation Bootstrap Methods Methods using Derived Input
More informationPENALIZED PRINCIPAL COMPONENT REGRESSION. Ayanna Byrd. (Under the direction of Cheolwoo Park) Abstract
PENALIZED PRINCIPAL COMPONENT REGRESSION by Ayanna Byrd (Under the direction of Cheolwoo Park) Abstract When using linear regression problems, an unbiased estimate is produced by the Ordinary Least Squares.
More informationIterative Selection Using Orthogonal Regression Techniques
Iterative Selection Using Orthogonal Regression Techniques Bradley Turnbull 1, Subhashis Ghosal 1 and Hao Helen Zhang 2 1 Department of Statistics, North Carolina State University, Raleigh, NC, USA 2 Department
More informationRegularization and Variable Selection via the Elastic Net
p. 1/1 Regularization and Variable Selection via the Elastic Net Hui Zou and Trevor Hastie Journal of Royal Statistical Society, B, 2005 Presenter: Minhua Chen, Nov. 07, 2008 p. 2/1 Agenda Introduction
More informationDIMENSION REDUCTION OF THE EXPLANATORY VARIABLES IN MULTIPLE LINEAR REGRESSION. P. Filzmoser and C. Croux
Pliska Stud. Math. Bulgar. 003), 59 70 STUDIA MATHEMATICA BULGARICA DIMENSION REDUCTION OF THE EXPLANATORY VARIABLES IN MULTIPLE LINEAR REGRESSION P. Filzmoser and C. Croux Abstract. In classical multiple
More informationMEANINGFUL REGRESSION COEFFICIENTS BUILT BY DATA GRADIENTS
Advances in Adaptive Data Analysis Vol. 2, No. 4 (2010) 451 462 c World Scientific Publishing Company DOI: 10.1142/S1793536910000574 MEANINGFUL REGRESSION COEFFICIENTS BUILT BY DATA GRADIENTS STAN LIPOVETSKY
More informationTechnical Report No. 1/12, February 2012 JACKKNIFING THE RIDGE REGRESSION ESTIMATOR: A REVISIT Mansi Khurana, Yogendra P. Chaubey and Shalini Chandra
Technical Report No. 1/12, February 2012 JACKKNIFING THE RIDGE REGRESSION ESTIMATOR: A REVISIT Mansi Khurana, Yogendra P. Chaubey and Shalini Chandra Jackknifing the Ridge Regression Estimator: A Revisit
More informationResearch Article An Unbiased Two-Parameter Estimation with Prior Information in Linear Regression Model
e Scientific World Journal, Article ID 206943, 8 pages http://dx.doi.org/10.1155/2014/206943 Research Article An Unbiased Two-Parameter Estimation with Prior Information in Linear Regression Model Jibo
More informationRegression Analysis for Data Containing Outliers and High Leverage Points
Alabama Journal of Mathematics 39 (2015) ISSN 2373-0404 Regression Analysis for Data Containing Outliers and High Leverage Points Asim Kumer Dey Department of Mathematics Lamar University Md. Amir Hossain
More informationGeneralized Elastic Net Regression
Abstract Generalized Elastic Net Regression Geoffroy MOURET Jean-Jules BRAULT Vahid PARTOVINIA This work presents a variation of the elastic net penalization method. We propose applying a combined l 1
More informationBiostatistics Advanced Methods in Biostatistics IV
Biostatistics 140.754 Advanced Methods in Biostatistics IV Jeffrey Leek Assistant Professor Department of Biostatistics jleek@jhsph.edu Lecture 12 1 / 36 Tip + Paper Tip: As a statistician the results
More informationIll-conditioning and multicollinearity
Linear Algebra and its Applications 2 (2000) 295 05 www.elsevier.com/locate/laa Ill-conditioning and multicollinearity Fikri Öztürk a,, Fikri Akdeniz b a Department of Statistics, Ankara University, Ankara,
More informationData Analysis and Machine Learning Lecture 12: Multicollinearity, Bias-Variance Trade-off, Cross-validation and Shrinkage Methods.
TheThalesians Itiseasyforphilosopherstoberichiftheychoose Data Analysis and Machine Learning Lecture 12: Multicollinearity, Bias-Variance Trade-off, Cross-validation and Shrinkage Methods Ivan Zhdankin
More informationComparisons of penalized least squares. methods by simulations
Comparisons of penalized least squares arxiv:1405.1796v1 [stat.co] 8 May 2014 methods by simulations Ke ZHANG, Fan YIN University of Science and Technology of China, Hefei 230026, China Shifeng XIONG Academy
More informationContents. 1 Review of Residuals. 2 Detecting Outliers. 3 Influential Observations. 4 Multicollinearity and its Effects
Contents 1 Review of Residuals 2 Detecting Outliers 3 Influential Observations 4 Multicollinearity and its Effects W. Zhou (Colorado State University) STAT 540 July 6th, 2015 1 / 32 Model Diagnostics:
More informationRXshrink_in_R. Generalized Ridge and Least Angle Regression estimates most likely to have minimum MSE risk
RXshrink_in_R Generalized Ridge and Least Angle Regression estimates most likely to have minimum MSE risk vignette-like documentation for R-package RXshrink version 1.1 - November 2018 Personal computing
More informationBusiness Statistics. Tommaso Proietti. Model Evaluation and Selection. DEF - Università di Roma 'Tor Vergata'
Business Statistics Tommaso Proietti DEF - Università di Roma 'Tor Vergata' Model Evaluation and Selection Predictive Ability of a Model: Denition and Estimation We aim at achieving a balance between parsimony
More informationA complex version of the LASSO algorithm and its application to beamforming
The 7 th International Telecommunications Symposium ITS 21) A complex version of the LASSO algorithm and its application to beamforming José F. de Andrade Jr. and Marcello L. R. de Campos Program of Electrical
More informationDr. Maddah ENMG 617 EM Statistics 11/28/12. Multiple Regression (3) (Chapter 15, Hines)
Dr. Maddah ENMG 617 EM Statistics 11/28/12 Multiple Regression (3) (Chapter 15, Hines) Problems in multiple regression: Multicollinearity This arises when the independent variables x 1, x 2,, x k, are
More informationADDRESSING MULTICOLLINEARITY IN REGRESSION MODELS: A RIDGE REGRESSION APPLICATION 1
Vol. 6, No. 1, Summer 2017 2012 Published by JSES. ADDRESSING MULTICOLLINEARITY IN REGRESSION MODELS: A RIDGE REGRESSION APPLICATION 1 Ali BAGER a, Monica ROMAN b, Meshal ALGELIDH c, Bahr MOHAMMED d Abstract
More informationRidge Regression. Chapter 335. Introduction. Multicollinearity. Effects of Multicollinearity. Sources of Multicollinearity
Chapter 335 Introduction is a technique for analyzing multiple regression data that suffer from multicollinearity. When multicollinearity occurs, least squares estimates are unbiased, but their variances
More informationshrink: a vignette for the RXshrink R-package
shrink: a vignette for the RXshrink R-package Generalized Ridge and Least Angle Regression Version 1.0-7, December 2011 License: GPL (>= 2) Personal computing treatments for your data analysis infirmities
More informationSOME APPLICATIONS: NONLINEAR REGRESSIONS BASED ON KERNEL METHOD IN SOCIAL SCIENCES AND ENGINEERING
SOME APPLICATIONS: NONLINEAR REGRESSIONS BASED ON KERNEL METHOD IN SOCIAL SCIENCES AND ENGINEERING Antoni Wibowo Farewell Lecture PPI Ibaraki 27 June 2009 EDUCATION BACKGROUNDS Dr.Eng., Social Systems
More informationA Practical Guide for Creating Monte Carlo Simulation Studies Using R
International Journal of Mathematics and Computational Science Vol. 4, No. 1, 2018, pp. 18-33 http://www.aiscience.org/journal/ijmcs ISSN: 2381-7011 (Print); ISSN: 2381-702X (Online) A Practical Guide
More informationRidge Regression Estimation Approach to Measurement Error Model
Ridge Regression Estimation Approach to Measurement Error Model A.K.Md. Ehsanes Saleh Carleton University, Ottawa CANADA) E-mail: esaleh@math.carleton.ca Shalabh Department of Mathematics & Statistics
More informationLinear Regression Linear Regression with Shrinkage
Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle
More informationECON 3150/4150, Spring term Lecture 7
ECON 3150/4150, Spring term 2014. Lecture 7 The multivariate regression model (I) Ragnar Nymoen University of Oslo 4 February 2014 1 / 23 References to Lecture 7 and 8 SW Ch. 6 BN Kap 7.1-7.8 2 / 23 Omitted
More informationOr How to select variables Using Bayesian LASSO
Or How to select variables Using Bayesian LASSO x 1 x 2 x 3 x 4 Or How to select variables Using Bayesian LASSO x 1 x 2 x 3 x 4 Or How to select variables Using Bayesian LASSO On Bayesian Variable Selection
More informationCOMS 4771 Lecture Fixed-design linear regression 2. Ridge and principal components regression 3. Sparse regression and Lasso
COMS 477 Lecture 6. Fixed-design linear regression 2. Ridge and principal components regression 3. Sparse regression and Lasso / 2 Fixed-design linear regression Fixed-design linear regression A simplified
More informationMitigating collinearity in linear regression models using ridge, surrogate and raised estimators
STATISTICS RESEARCH ARTICLE Mitigating collinearity in linear regression models using ridge, surrogate and raised estimators Diarmuid O Driscoll 1 and Donald E. Ramirez 2 * Received: 21 October 2015 Accepted:
More informationPreliminary test almost unbiased ridge estimator in a linear regression model with multivariate Student-t errors
ACTA ET COMMENTATIONES UNIVERSITATIS TARTUENSIS DE MATHEMATICA Volume 1, Number 1, 011 Available online at www.math.ut.ee/acta/ Preliminary test almost unbiased ridge estimator in a linear regression model
More informationMS-C1620 Statistical inference
MS-C1620 Statistical inference 10 Linear regression III Joni Virta Department of Mathematics and Systems Analysis School of Science Aalto University Academic year 2018 2019 Period III - IV 1 / 32 Contents
More informationMSE Performance and Minimax Regret Significance Points for a HPT Estimator when each Individual Regression Coefficient is Estimated
This article was downloaded by: [Kobe University] On: 8 April 03, At: 8:50 Publisher: Taylor & Francis Informa Ltd Registered in England and Wales Registered Number: 07954 Registered office: Mortimer House,
More informationA New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables
A New Combined Approach for Inference in High-Dimensional Regression Models with Correlated Variables Niharika Gauraha and Swapan Parui Indian Statistical Institute Abstract. We consider the problem of
More informationA New Asymmetric Interaction Ridge (AIR) Regression Method
A New Asymmetric Interaction Ridge (AIR) Regression Method by Kristofer Månsson, Ghazi Shukur, and Pär Sölander The Swedish Retail Institute, HUI Research, Stockholm, Sweden. Deartment of Economics and
More informationInference After Variable Selection
Department of Mathematics, SIU Carbondale Inference After Variable Selection Lasanthi Pelawa Watagoda lasanthi@siu.edu June 12, 2017 Outline 1 Introduction 2 Inference For Ridge and Lasso 3 Variable Selection
More informationBootstrapping Heteroskedasticity Consistent Covariance Matrix Estimator
Bootstrapping Heteroskedasticity Consistent Covariance Matrix Estimator by Emmanuel Flachaire Eurequa, University Paris I Panthéon-Sorbonne December 2001 Abstract Recent results of Cribari-Neto and Zarkos
More informationA Significance Test for the Lasso
A Significance Test for the Lasso Lockhart R, Taylor J, Tibshirani R, and Tibshirani R Ashley Petersen May 14, 2013 1 Last time Problem: Many clinical covariates which are important to a certain medical
More informationMachine Learning for OR & FE
Machine Learning for OR & FE Regression II: Regularization and Shrinkage Methods Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com
More informationAn Modified PLSR Method in Prediction
Journal of Data Science 4(2006), 257-274 An Modified PLSR Method in Prediction Bo Cheng 1 and Xizhi Wu 2 1 Altair Engineering Company and 2 Renmin University of China Abstract: Among many statistical methods
More informationInstitute of Statistics Mimeo Series No Simultaneous regression shrinkage, variable selection and clustering of predictors with OSCAR
DEPARTMENT OF STATISTICS North Carolina State University 2501 Founders Drive, Campus Box 8203 Raleigh, NC 27695-8203 Institute of Statistics Mimeo Series No. 2583 Simultaneous regression shrinkage, variable
More informationBayesian Grouped Horseshoe Regression with Application to Additive Models
Bayesian Grouped Horseshoe Regression with Application to Additive Models Zemei Xu, Daniel F. Schmidt, Enes Makalic, Guoqi Qian, and John L. Hopper Centre for Epidemiology and Biostatistics, Melbourne
More informationInfluencial Observations in Ridge Regression
Influencial Observations in Ridge Regression by Robert L. Obenchain We augment traditional trace displays with dynamic graphics that can reveal dramatic changes in influence/outliers/leverages of observations
More informationMatematické Metody v Ekonometrii 7.
Matematické Metody v Ekonometrii 7. Multicollinearity Blanka Šedivá KMA zimní semestr 2016/2017 Blanka Šedivá (KMA) Matematické Metody v Ekonometrii 7. zimní semestr 2016/2017 1 / 15 One of the assumptions
More informationSelection Criteria Based on Monte Carlo Simulation and Cross Validation in Mixed Models
Selection Criteria Based on Monte Carlo Simulation and Cross Validation in Mixed Models Junfeng Shang Bowling Green State University, USA Abstract In the mixed modeling framework, Monte Carlo simulation
More informationLecture 1: OLS derivations and inference
Lecture 1: OLS derivations and inference Econometric Methods Warsaw School of Economics (1) OLS 1 / 43 Outline 1 Introduction Course information Econometrics: a reminder Preliminary data exploration 2
More informationLink to Paper. The latest iteration can be found at:
Link to Paper Introduction The latest iteration can be found at: http://learneconometrics.com/pdf/gc2017/collin_gretl_170523.pdf BKW dignostics in GRETL: Interpretation and Performance Oklahoma State University
More informationLinear Regression Linear Regression with Shrinkage
Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle
More informationMachine Learning. Regression basics. Marc Toussaint University of Stuttgart Summer 2015
Machine Learning Regression basics Linear regression, non-linear features (polynomial, RBFs, piece-wise), regularization, cross validation, Ridge/Lasso, kernel trick Marc Toussaint University of Stuttgart
More informationPARTIAL RIDGE REGRESSION 1
1 2 This work was supported by NSF Grants GU-2059 and GU-19568 and by U.S. Air Force Grant No. AFOSR-68-1415. On leave from Punjab Agricultural University (India). Reproduction in whole or in part is permitted
More informationEffect of outliers on the variable selection by the regularized regression
Communications for Statistical Applications and Methods 2018, Vol. 25, No. 2, 235 243 https://doi.org/10.29220/csam.2018.25.2.235 Print ISSN 2287-7843 / Online ISSN 2383-4757 Effect of outliers on the
More informationStat 5100 Handout #26: Variations on OLS Linear Regression (Ch. 11, 13)
Stat 5100 Handout #26: Variations on OLS Linear Regression (Ch. 11, 13) 1. Weighted Least Squares (textbook 11.1) Recall regression model Y = β 0 + β 1 X 1 +... + β p 1 X p 1 + ε in matrix form: (Ch. 5,
More informationPanel Threshold Regression Models with Endogenous Threshold Variables
Panel Threshold Regression Models with Endogenous Threshold Variables Chien-Ho Wang National Taipei University Eric S. Lin National Tsing Hua University This Version: June 29, 2010 Abstract This paper
More informationABSTRACT. POST, JUSTIN BLAISE. Methods to Improve Prediction Accuracy under Structural Constraints. (Under the direction of Howard Bondell.
ABSTRACT POST, JUSTIN BLAISE. Methods to Improve Prediction Accuracy under Structural Constraints. (Under the direction of Howard Bondell.) Statisticians are often faced with the difficult task of model
More informationThe RXshrink Package
The RXshrink Package October 20, 2005 Title Maximum Likelihood Shrinkage via Ridge or Least Angle Regression Version 1.0-2 Date 2005-10-19 Author Maintainer Depends R (>= 1.8.0), lars Identify and display
More informationSpatial Lasso with Application to GIS Model Selection. F. Jay Breidt Colorado State University
Spatial Lasso with Application to GIS Model Selection F. Jay Breidt Colorado State University with Hsin-Cheng Huang, Nan-Jung Hsu, and Dave Theobald September 25 The work reported here was developed under
More informationOn High-Dimensional Cross-Validation
On High-Dimensional Cross-Validation BY WEI-CHENG HSIAO Institute of Statistical Science, Academia Sinica, 128 Academia Road, Section 2, Nankang, Taipei 11529, Taiwan hsiaowc@stat.sinica.edu.tw 5 WEI-YING
More informationPAijpam.eu M ESTIMATION, S ESTIMATION, AND MM ESTIMATION IN ROBUST REGRESSION
International Journal of Pure and Applied Mathematics Volume 91 No. 3 2014, 349-360 ISSN: 1311-8080 (printed version); ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu doi: http://dx.doi.org/10.12732/ijpam.v91i3.7
More informationBootstrap & Confidence/Prediction intervals
Bootstrap & Confidence/Prediction intervals Olivier Roustant Mines Saint-Étienne 2017/11 Olivier Roustant (EMSE) Bootstrap & Confidence/Prediction intervals 2017/11 1 / 9 Framework Consider a model with
More informationA Comparison Study of Principle Component Regression, Partial Least Squares Regression and Ridge Regression with Application to FTIR Data
A Comparison Study of Principle Component Regression, Partial Least Squares Regression and Ridge Regression with Application to FTIR Data Author: Ying Li Supervisor: Dietrich von Rosen Master thesis in
More information