1. Introduction Over the last three decades a number of model selection criteria have been proposed, including AIC (Akaike, 1973), AICC (Hurvich & Tsa

Size: px
Start display at page:

Download "1. Introduction Over the last three decades a number of model selection criteria have been proposed, including AIC (Akaike, 1973), AICC (Hurvich & Tsa"

Transcription

1 On the Use of Marginal Likelihood in Model Selection Peide Shi Department of Probability and Statistics Peking University, Beijing P. R. China Chih-Ling Tsai Graduate School of Management University of California Davis, California U. S. A. SUMMARY Based on the marginal likelihood approach, we develop a model selection criterion, MIC, for regression models with the general variance structure. These include weighted regression models, regression models with ARMA errors, growth curve models, and spatial correlation models. We show that MIC is a consistent criterion. For regression models with either constant or non-constant variance, simulation studies indicate that MIC not only provides better model order choices than the asymptotic ecient criteria AIC, AICC, F P E and C p, but is also superior to the consistent criteria BIC and F IC in both small and large sample sizes. These results also hold for regression models with AR(1) errors, except that BIC performs slightly better than MIC in the large sample size case. Finally, Monte Carlo studies show that the eectiveness of model selection criteria decreases as the degree of heterogeneity (or the rst order autocorrelation) increases. Some key words: AIC; AICC; BIC; F IC; MIC; Marginal likelihood. 1

2 1. Introduction Over the last three decades a number of model selection criteria have been proposed, including AIC (Akaike, 1973), AICC (Hurvich & Tsai, 1989), BIC (Schwartz, 1978), F IC (Wei, 1992), F P E (Akaike, 1970), and C p (Mallows, 1973). All of these criteria are comprised of two basic components where the rst is a function of the scaled parameter estimator, which measures the goodness-of-t, and the second is a function of the number of unknown parameters, which penalizes overtting. Hence, when considering two candidate models with the same number of parameters (e.g., one including explanatory variables x 1 ; x 3 and x 5, and the other including x 2 ; x 4 and x 6 ), the model with the smaller value for the rst component will be selected. This implies that the more interesting parameter for model selection is the scaled parameter, and that the regression parameters can be viewed as nuisance parameters. In applications, one may use a linear model with a general variance structure, which needs to be selected based on data. For this kind of models, we are interested in the parameters included in the variance structure, and the regression parameters are considered nuisance parameters. A widely used procedure for inference about parameters of interest when nuisance parameters are also present is based on the prole log-likelihood function. However, this procedure may give inconsistent or inecient estimates. In addition, the prole loglikelihood itself is not a log-likelihood. Thus we must consider conditional log-likelihood (Cox & Hinkley, 1974), modied prole log-likelihood (Barndor-Nielsen, 1983 and 1985), and marginal log-likelihood (McCullagh & Nelder, 1989) as alternative approaches. For the sake of simplicity, in this paper we only consider the marginal log-likelihood to obtain a model selection criterion. The aim of this paper is to consider model selection for regression models with the general variance structure and obtain the modied information criterion, MIC. We will achieve this rst by obtaining an unbiased estimator of the expected Kullback-Leibler information of the marginal log-likelihood function of the tted model. We then derive MIC criterion in Section 2, and show that it is a consistent criterion in Section 3. In Section 4, we compare MIC with other model selection criteria via Monte Carlo studies. The results indicate that MIC is not only outperforms the ecient criteria AIC, AICC, F P E and C p, but is also superior (or comparable) to the consistent criteria BIC and F IC in both small and large sample sizes. Finally, we give some concluding remarks in 2

3 Section Derivation of the M I C Criterion 2.1. Marginal Log-likelihood Function Consider the candidate family of models Y = X + e; (2:1) where Y is an n1 vector, X is an nk matrix of known constants, is a k 1 vector of unknown regression coecients, e = (e 1 ; ; e n ) 0 has a multivariate normal distribution with mean E(e) = 0 and variance E(ee 0 ) = 2 W (), is an unknown scalar and is a q 1 unknown vector. In this model, ( 0 ; 2 ) 0 are the parameters of interest and the are nuisance parameters. Next, by adopting McCullagh and Nelder's (1989, Section 7.2) results, we obtain the marginal log-likelihood of ( 0 ; 2 ) 0 : L(; 2 ; Y ) =? 1 2 log j2 W j? 1 2 log jx0 W?1 X= 2 j? 1 2 Y 0 (I? H)Y = 2 ; (2:2) where Y = W?1=2 Y, X = W?1=2 X and H = X (X 0 X )?1 X 0. Suppose, however, that the response was generated by the true model Y = X "; where X 0 is an n k 0 matrix of known constants, 0 is a k 0 1 vector of unknown regression coecients, " = (" 1 ; ; " n ) 0 has a multivariate normal distribution with mean E(") = 0 and variance E("" 0 ) = 2 0W ( 0 ), 0 is an unknown scalar and 0 is a m 1 unknown vector. Then the marginal log-likelihood of ( 0 0; 2 0) 0 is L 0 ( 0 ; 0; 2 Y ) =? 1 2 log j2 0W 0 j? 1 2 log jx0 0W?1 0 X 0 =0j 2? 1 2 Y 0 0 (I? H 0 )Y 0 =0; 2 (2:3) where Y 0 = W?1=2 0 Y, X = W?1=2 0 X 0, H 0 = X0(X 0 0 X 0 )?1 X 0 0 and W 0 = W ( 0 ). Using equations (2.2) and (2.3), we can now obtain a model selection criterion as shown in the next subsection Marginal Information Criterion 3

4 A useful measure of the discrepancy between the true and candidate marginal loglikelihood functions is the Kullback-Leibler information (; ) =?2E 0 fl(; 2 ; Y )g = log j 2 W j + log jx 0 W?1 X= 2 j + trf 0 0X 0 0W 0?1=2 (I? H)W?1=2 X 0 0 g +trf(i? H)W?1=2 W 0 W 0?1=2 g 2 0= 2 ; (2:4) where E 0 denotes the expectation under the true model. We assume that the candidate family includes the true model, an assumption that is also used in the derivation of the Akaike information criterion, AIC (Linhart & Zucchini, 1986), and the corrected Akaike information criterion, AICC (Hurvich & Tsai, 1989). Under this assumption, the columns of X can be rearranged so that X 0 0 = X, where = ( 0 0; 0 0 ) 0 and 0 0 is a 1 (k? k 0 ) vector. Hence, the the third component of the right hand side of (2.4) is 0. Furthermore, replacing W?1=2 by its approximation W?1=2 0 in the second trace component of equation (2.4), (; ) can be approximated by (n? k) log( 2 ) + log jw j + log jx 0 W?1 Xj + (n? k) 0 = 2 : (2:5) A reasonable criterion for judging the quality of the candidate family in the light of the data is E 0 f(^; ^)g, where ^ 2 and ^ are the maximum likelihood estimates of 2 and in the candidate family, respectively. Note that when the candidate models include the true model, n^ 2 = 2 0 is approximately distributed as 2 n?k and E 0 (0=^ 2 2 ) n. This, n?k?2 together with equation (2.5), gives us (^; ^) (n? k)e 0 flog(^ 2 )g + E 0 flog j ^W j + log jx 0 ^W?1 Xjg + n(n? k) n? k? 2 ; where ^W = W (^). This leads to the corrected Akaike information criterion obtained from the marginal log-likelihood function, which is MIC = (n? k) log(^ 2 ) + log j ^W j + log jx 0 ^W?1 Xj + Three analytical examples are given below to illustrate the use of MIC. Example 2.1 (Regression model with constant variance) n(n? k) n? k? 2 : (2:6) Consider the case where W () = I. Here (2.1) represents a multiple regression model with constant variance, and the resulting selection criterion is MIC = (n? k) log(^ 2 ) + log jx 0 Xj + 4 n(n? k) n? k? 2 : (2:7)

5 It is interesting to note that the log jx 0 Xj term also appears in Wei's (1992) F IC criterion (see the expression of FIC in Section 4). Example 2.2 (Regression model with non-constant variance) Assume that W () = diag(w i ()), where W i () = W (z i ; ), z i is a known r 1 vector and is an unknown m 1 vector for i = 1; ; n. Under this assumption, (2.1) is a multiple regression model with non-constant variance, and the resulting marginal log-likelihood has the same form as equation (2.2). Hence, the selection criterion MIC is given by (2.6). For this weighted regression model, McCullagh & Tibshirani (1990) showed that its marginal log-likelihood is the rst order bias corrected prole log-likelihood. In addition, Verbyla (1993) used the marginal likelihood to obtain a test statistic to assess the homogeneity of the errors, and recently Lyon & Tsai (1996) compared various tests for homogeneity obtained when using the log-likelihood, marginal log-likelihood, conditional log-likelihood, and modied log-likelihood functions. Example 2.3 (Regression model with ARMA errors) Consider the regression model given by (2.1), where the random errors are generated by an ARMA(p; q) process dened as e t? 1 e t?1? ;? p e t?p = a t? ' 1 a t?1?? ' q a t?q ; where a t is a sequence of independent normal random variables having mean zero and variance 2. The log-likelihood function for e (ignoring the constant) is L(ej 2 ; ) = jw j?1=2 (2)?n=2 exp(? 1 2 e0 W?1 e= 2 ); (2:8) where = ( 1 ; ; p ; ' 1 ; ; ' q ) 0 and W is the variance matrix of e. Equation (2.8) can be found in Cooper & Thompson, 1977 and Harvey & Phillips, The resulting marginal log-likelihood has the same form as (2.2) (see Wilson, 1989), and hence the selection criterion MIC is given by (2.6). Note that MIC is used to select regression parameters by assuming that the the orders of p and q in the ARMA process are specied. This is dierent from the approach taken by Tsay (1984), who proposed a method to identify the order of p and q when the dimension of the regression parameters is known. Note that by using arguments similar to those in Examples 2 and 3, we also can obtain MIC for growth curve models and for regression models with spatial correlation errors 5

6 (Sen & Shrivastava, 1990, p. 138 and p. 143). The theoretical justication for applying MIC to variable selection under the above conditions is given in the next section. 3. Asymptotic Results for M IC Let A = f : is a nonempty subset of f1; ; kgg and A 0 = f 2 A : E(Y ) = X()()g be a subset of A, where each in A 0 is associated with a model that includes the true model, and X() and () contain the components X and, respectively. Finally, let ^ denote the model selected by MIC so that its value is the smallest of those for all possible candidate models, and let 0 be the model in A 0 with the smallest dimension. We will rst obtain asymptotic results for the multiple regression model with constant variance, and then extend the results to the general model setting given by (2.1). In order to state and prove our asymptotic results, we will need the following two denitions and assumptions. Denition 1. The selection procedure is said to be consistent if P f^ = 0 g?! 1 as n tends to innity, and the selection procedure is said to be strongly consistent if P flim n ^ = 0 g = 1: Denition 2. A random variable is sub-gaussian if there exists a constant ' > 0 such that for all real d Efexp(d)g exp('d 2 ): This denition is adopted from Zheng & Loh (1995). Assumption 1. X()(X() 0 X())?1 X 0 (). Assumption 2. lim inf n!1 n?1 inf 2A?A 0 kx 0 0? H()X 0 0 k 2 > 0, where H() = lim sup n!1 max 2A jx()0 X()=nj < 1 and 0 < lim inf n!1 min 2A jx()0 X()=nj: We can now present the asymptotic results for the selection criterion, MIC. 6

7 Theorem 3.1. If Assumptions 1 and 2 are satised and the " i are assumed to be iid sub-gaussian, then MIC is a consistent criterion. The proof for Theorem 3.1 is given in the Appendix. Note that if the sub-gaussian assumption is replaced by the moment condition, E(" 4h 1 ) < 1 for some positive integer h, Theorem 3.1 remains true. Theorem 3.2. Let MIC (s) () = (n? k()) log(^ 2 ()) + n log jx() 0 X()j + 2(k() + 2) n? k()? 2 ; where k() is the dimension of the submodel 2 A, j s = maxfj : j 1=2s < log(j)g, and n = 8 < : 1 when n j s n 1=2s = log(n) when n > j s : Under the same assumptions given for Theorem 3.1, MIC (s) is a strongly consistent criterion for all nite s > 1. The proof of Theorem 3.2 is not presented here since the techniques used are similar to those for the proof of Theorem 3.1. However, the details can be obtained from the authors by request. Note that under the assumptions of Theorem 3.1, MIC is near strong consistency, since, except for a constant term n+2, MIC and MIC (s) are identical when n < j s and j s are very large; for example, j 35 = 7: E and j 54 = 7: E Theorem 3.1 shows that MIC is a consistent criterion for multiple regression models with a constant variance error structure. In fact, this result can be extended to the general model setting given by (2.1), as we see below. Theorem 3.3 In model (2.1), assume that both ^? 0 and ^ 2 (; ^)?^ 2 (; 0 ) tend to zero in probability as n tends to innity for all 2 A. In addition, assume that the elements of W () are continuous functions of, W () is positive denite in the neighborhood of 0, and Assumptions 1 and 2 are satised when X() is replaced by W ()?1=2 X(). If the " i are iid sub-gaussian, then MIC is a consistent criterion. The proof of Theorem 3.3 is given in the Appendix. Its results also are still valid if the sub-gaussian assumption is replaced by the moment condition, E(" 4h 1 ) < 1 for some positive integer h. 7

8 4. Monte Carlo Results In this section we use simulation studies to compare the performance of MIC with AIC, AICC, BIC, F IC, CP and F P E, dened as: AIC = n log(^ 2 ) + 2k, AICC = n log(^ 2 ) + 2n(k + 1)=(n? k? 2), BIC = n log(^ 2 ) + k log(n), F IC = n^ 2 + ~ 2 log jx 0 Xj, C p = n^ 2 =~ 2? n + 2k, and F P E = n^ 2 (n + k)=(n? k), where ^ 2 and ~ 2 are the maximum likelihood estimates of under the tted model and the full model, respectively. We consider regression models with either constant variance, non-constant variance, or AR(1) errors. For each of ve sample sizes (n=15, 20, 40, 80, 160), 1000 realizations were generated from the true model (2.1), where 0 = (1; ; 1) 0 is an k 0 1 vector, 0 = 1, and x 0i N(0; I k0 ) for i = 1; ; n. There are nine candidate variables, stored in an n9 matrix of independent identically distributed normal random variables. The candidate models are linear, and are listed in columns in a sequential nested fashion. Hence the set of candidate models includes the true model. Example 4.1. (Regression model with constant variance) Here we assume that k 0 = 5 and that the " i are iid normal random variables with mean zero and variance 1. Table 1 presents the proportions of the true order selected by the various criteria. In the small sample sizes (n=15 and 20), it is obvious that AICC and MIC outperform other criteria, and F IC and C P perform the worst. As the sample size increases, the proportions of correct model order chosen by BIC, F IC, and MIC also increase. This is not surprising since all these criteria are consistent. MIC performs the best overall for both large and small sample sizes. Example 4.2. (Regression model with non-constant variance) Consider the weighted regression model y i = x 0 0i 0 + w i " i ; i = 1; ; n; where w i = exp( 0 z i ), the z i are iid standard normal, and the 0 are 0.2, 0.5 and 0.7. Table 2 shows that the performances of all six criteria are similar to those for Example 4.1; that is, MIC is superior to AICC in small sample sizes and is slightly better than F IC and BIC in large sample sizes, and is best overall. The criteria AIC, C P and F P E 8

9 all perform comparably poorly. It is also interesting to note that the proportion of correct model order selections for each criterion decreases as the 0 increases. This indicates that strong heterogeneity has a negative eect on model selection. Example 4.3. (Regression model with AR(1) errors) Consider the linear regression with AR(1) errors where the " i are random errors satisfying y i = x 0 0i 0 + " i ; i = 1; ; n; " 1 N(0; 1=(1? 2 0)); " i = 0 " i?1 + i ; i N(0; 1); for i = 2; 3; ; n; and 0 = 0:1; 0:5 and 0:9. Table 3 only presents simulation results for positive autocorrelation, since the corresponding negative autocorrelation gives similar results. It shows that MIC outperforms the other criteria except when n = 80 and 160, where BIC performs slightly better. Not surprisingly, large positive autocorrelation reduces the eectness of model selection criteria, especially when 0 is close to 1 and the sample size is small or moderate. We also obtain the same conclusion when the autocorrelation is negative. The above three examples give us more insight into why AICC and MIC perform quite dierently. For regression models with the constant variance (Example 4.1), AICC and MIC are unbiased estimators of the Kullback-Leibler informations computed from the log-likelihood function and the marginal log-likelihood function, respectively. Hence, both performance comparably in small samples. As the sample size increases, MIC, because it is a consistent criterion, outperforms AICC. For regression models with heteroscedastic error (Example 4.2) or autocorrelated error (Example 4.3), MIC is still an approximately unbiased estimator of the Kullback-Leibler information computed from the marginal loglikelihood function, but AICC is not the unbiased estimator of the Kullback-Leibler information computed from the log-likelihood function. Therefore, MIC outperforms AICC in both small and large samples. 5. Conclusions 9

10 We have applied the marginal log-likelihood approach to regression models with the general variance structure in order to obtain a good criterion, MIC. We show that MIC is a consistent criterion. Simulation studies indicate that MIC is not only superior to the ecient criteria AIC, AICC, C P and F IC, but also performs better than (or comparable to) the consistent criteria BIC and F IC. This means that the marginal log-likelihood approach results in better variable selections than the log-likelihood approach under these conditions, and therefore it is natural to use the marginal log-likelihood approach to revise some existing model selection criteria. MIC also can be extended to other model settings, such as regression models with ARCH errors (Gourieroux, 1997) and multivarite regression model with a general variance structure (Greene, 1993, Chapters 16 & 17, Diggle, Liang & Zeger, 1994, Chapter 4, and Bedrick & Tsai, 1994). Finally, in future the conditional loglikelihood and modied prole log-likelihood approaches can be applied to obtain model selection criteria, and then compared to MIC. ACKNOWLEDGMENTS The research of Peide Shi was supported by NSF of China. The research of Chih-Ling Tsai was supported by National Science Foundation grant DMS

11 REFERENCES Akaike, H. (1970). Statistical predictor identication. Ann. Inst. Statist. Math. 22, Akaike, H.(1973). Information theory and an extension of the maximum likelihood principle. In 2nd International Symposium on Information Theory, Ed. B.N. Petrov and F. Csaki, pp Budapest: Akademia Kiado. Barndor-Nielsen, O. E. (1983). On a formula for the distribution of the maximum likelihood estimator. Biometrika 70, Barndor-Nielsen, O. E. (1985). Properties of modied prole likelihood. In Contributions to Probability and Statistics in Honour of Gunnar Blom, Ed. J. Lanke and G. Lindgren, pp Lund. Bedrick, E. J. & Tsai, C. L. (1994). Model selection for multivariate regression in small samples. Biometrics 50, Chatterjee, S. & Hadi, A. S. (1988). Sensitivity Analysis in Linear Regression. New York: Wiley. Cooper, D. M. & Thompson, R. (1977). A note on the estimation of the parameters of the autoregressive-moving average process. Biometrika 64, Cox, D. R. & Hinkley, D. V. (1974). Theoretical Statistics. New York: Chapman and Hall. Diggle, P. J., Liang, K. Y. & Zeger, S. L. (1994). Analysis of Longitudinal Data. New York: Oxford. Gourieroux, C. (1997). ARCH Models and Financial Applications. New York: Springer. Greene, W. H. (1993). Econometric Analysis. New York: Macmillan. Harvey, A. C. & Phillips, G. D. A. (1979). Maximum likelihood estimation of regression models with autoregressive-moving average disturbances. Biometrika 66, Hurvich, C. M. & Tsai, C.-L. (1989). Regression and time series model selection in small samples. Biometrika 76, Linhart, H. & Zucchini, W. (1986). Model Selection. New York: Wiley. Lyon, J. & Tsai, C. L. (1996). A comparison of tests for heteroscedasticity. The Statistician 45, Mallows, C. L. (1973). Some comments on C p. Technometrics 15, McCullagh, P. & Nelder, J. A. (1989). Generalized Linear Models. New York: Chapman and Hall. 11

12 McCullagh, P. & Tibshirani, R. (1990). A simple method for the adjustment of prole likelihoods. J. R. Statist. Soc. B. 52, Rao, C. R. & Klee, J. (1988). Estimation of Variance Components and Applications. Amsterdam: North-Holland. Schwarz, G. (1978). Estimating the dimension of a model. Ann. Statist. 6, Sen, A. & Shrivastava, M. (1990). Regression Analysis: Theory, Methods, and Applications. New York: Springer-Verlag. Tsay, R. S. (1984). Regression models with time series errors. J. Amer. Statist. Assoc. 79, Verbyla, A. P. (1993). Modelling variance heterogeneity: residual maximum likelihood and diagnostics. J. R. Statist. Soc. B. 55, Wei, C. Z. (1992). On predictive least squares principles. Ann. Statist. 20, Wilson, G. T. (1989). On the use of marginal likelihood in time series model estimation. J. R. Statist. Soc. B. 51, Zheng, X. D. & Loh, W. Y. (1995). Consistency variable selection in linear models. J. Amer. Statist. Assoc. 90,

13 Table 1. Proportions of correct model order selection by AIC, AICC, BIC, F IC, C p, F P E, and MIC criteria in 1000 realizations for the regression model. Criteria n=15 n=20 n=40 n=80 n=160 AIC AICC BIC F IC C p F P E MIC

14 Table 2. Proportions of correct model order selection by AIC, AICC, BIC, F IC, C p, F P E, and MIC criteria in 1000 realizations for the weighted regression model. 0 Criteria n=15 n=20 n=40 n=80 n=160 AIC AICC BIC F IC C p F P E MIC AIC AICC BIC F IC C p F P E MIC AIC AICC BIC F IC C p F P E MIC

15 Table 3. Proportions of correct model order selection by AIC, AICC, BIC, F IC, C p, F P E, and MIC criteria in 1000 realizations for the regression model with AR(1) errors. 0 Criteria n=15 n=20 n=40 n=80 n=160 AIC AICC BIC F IC C p F P E MIC AIC AICC BIC F IC C p F P E MIC AIC AICC BIC F IC C p F P E MIC

Bias-corrected AIC for selecting variables in Poisson regression models

Bias-corrected AIC for selecting variables in Poisson regression models Bias-corrected AIC for selecting variables in Poisson regression models Ken-ichi Kamo (a), Hirokazu Yanagihara (b) and Kenichi Satoh (c) (a) Corresponding author: Department of Liberal Arts and Sciences,

More information

Discrepancy-Based Model Selection Criteria Using Cross Validation

Discrepancy-Based Model Selection Criteria Using Cross Validation 33 Discrepancy-Based Model Selection Criteria Using Cross Validation Joseph E. Cavanaugh, Simon L. Davies, and Andrew A. Neath Department of Biostatistics, The University of Iowa Pfizer Global Research

More information

Model Selection for Semiparametric Bayesian Models with Application to Overdispersion

Model Selection for Semiparametric Bayesian Models with Application to Overdispersion Proceedings 59th ISI World Statistics Congress, 25-30 August 2013, Hong Kong (Session CPS020) p.3863 Model Selection for Semiparametric Bayesian Models with Application to Overdispersion Jinfang Wang and

More information

An Unbiased C p Criterion for Multivariate Ridge Regression

An Unbiased C p Criterion for Multivariate Ridge Regression An Unbiased C p Criterion for Multivariate Ridge Regression (Last Modified: March 7, 2008) Hirokazu Yanagihara 1 and Kenichi Satoh 2 1 Department of Mathematics, Graduate School of Science, Hiroshima University

More information

Regression and Time Series Model Selection in Small Samples. Clifford M. Hurvich; Chih-Ling Tsai

Regression and Time Series Model Selection in Small Samples. Clifford M. Hurvich; Chih-Ling Tsai Regression and Time Series Model Selection in Small Samples Clifford M. Hurvich; Chih-Ling Tsai Biometrika, Vol. 76, No. 2. (Jun., 1989), pp. 297-307. Stable URL: http://links.jstor.org/sici?sici=0006-3444%28198906%2976%3a2%3c297%3aratsms%3e2.0.co%3b2-4

More information

Selection Criteria Based on Monte Carlo Simulation and Cross Validation in Mixed Models

Selection Criteria Based on Monte Carlo Simulation and Cross Validation in Mixed Models Selection Criteria Based on Monte Carlo Simulation and Cross Validation in Mixed Models Junfeng Shang Bowling Green State University, USA Abstract In the mixed modeling framework, Monte Carlo simulation

More information

On Properties of QIC in Generalized. Estimating Equations. Shinpei Imori

On Properties of QIC in Generalized. Estimating Equations. Shinpei Imori On Properties of QIC in Generalized Estimating Equations Shinpei Imori Graduate School of Engineering Science, Osaka University 1-3 Machikaneyama-cho, Toyonaka, Osaka 560-8531, Japan E-mail: imori.stat@gmail.com

More information

Order Selection for Vector Autoregressive Models

Order Selection for Vector Autoregressive Models IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 51, NO. 2, FEBRUARY 2003 427 Order Selection for Vector Autoregressive Models Stijn de Waele and Piet M. T. Broersen Abstract Order-selection criteria for vector

More information

Variable Selection in Restricted Linear Regression Models. Y. Tuaç 1 and O. Arslan 1

Variable Selection in Restricted Linear Regression Models. Y. Tuaç 1 and O. Arslan 1 Variable Selection in Restricted Linear Regression Models Y. Tuaç 1 and O. Arslan 1 Ankara University, Faculty of Science, Department of Statistics, 06100 Ankara/Turkey ytuac@ankara.edu.tr, oarslan@ankara.edu.tr

More information

A Variant of AIC Using Bayesian Marginal Likelihood

A Variant of AIC Using Bayesian Marginal Likelihood CIRJE-F-971 A Variant of AIC Using Bayesian Marginal Likelihood Yuki Kawakubo Graduate School of Economics, The University of Tokyo Tatsuya Kubokawa The University of Tokyo Muni S. Srivastava University

More information

Bias Correction of Cross-Validation Criterion Based on Kullback-Leibler Information under a General Condition

Bias Correction of Cross-Validation Criterion Based on Kullback-Leibler Information under a General Condition Bias Correction of Cross-Validation Criterion Based on Kullback-Leibler Information under a General Condition Hirokazu Yanagihara 1, Tetsuji Tonda 2 and Chieko Matsumoto 3 1 Department of Social Systems

More information

Bootstrap Simulation Procedure Applied to the Selection of the Multiple Linear Regressions

Bootstrap Simulation Procedure Applied to the Selection of the Multiple Linear Regressions JKAU: Sci., Vol. 21 No. 2, pp: 197-212 (2009 A.D. / 1430 A.H.); DOI: 10.4197 / Sci. 21-2.2 Bootstrap Simulation Procedure Applied to the Selection of the Multiple Linear Regressions Ali Hussein Al-Marshadi

More information

Analysis of the AIC Statistic for Optimal Detection of Small Changes in Dynamic Systems

Analysis of the AIC Statistic for Optimal Detection of Small Changes in Dynamic Systems Analysis of the AIC Statistic for Optimal Detection of Small Changes in Dynamic Systems Jeremy S. Conner and Dale E. Seborg Department of Chemical Engineering University of California, Santa Barbara, CA

More information

Model comparison and selection

Model comparison and selection BS2 Statistical Inference, Lectures 9 and 10, Hilary Term 2008 March 2, 2008 Hypothesis testing Consider two alternative models M 1 = {f (x; θ), θ Θ 1 } and M 2 = {f (x; θ), θ Θ 2 } for a sample (X = x)

More information

An Akaike Criterion based on Kullback Symmetric Divergence in the Presence of Incomplete-Data

An Akaike Criterion based on Kullback Symmetric Divergence in the Presence of Incomplete-Data An Akaike Criterion based on Kullback Symmetric Divergence Bezza Hafidi a and Abdallah Mkhadri a a University Cadi-Ayyad, Faculty of sciences Semlalia, Department of Mathematics, PB.2390 Marrakech, Moroco

More information

Model selection criteria Λ

Model selection criteria Λ Model selection criteria Λ Jean-Marie Dufour y Université de Montréal First version: March 1991 Revised: July 1998 This version: April 7, 2002 Compiled: April 7, 2002, 4:10pm Λ This work was supported

More information

Performance of Autoregressive Order Selection Criteria: A Simulation Study

Performance of Autoregressive Order Selection Criteria: A Simulation Study Pertanika J. Sci. & Technol. 6 (2): 7-76 (2008) ISSN: 028-7680 Universiti Putra Malaysia Press Performance of Autoregressive Order Selection Criteria: A Simulation Study Venus Khim-Sen Liew, Mahendran

More information

Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis

Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis Consistency of Test-based Criterion for Selection of Variables in High-dimensional Two Group-Discriminant Analysis Yasunori Fujikoshi and Tetsuro Sakurai Department of Mathematics, Graduate School of Science,

More information

Akaike Information Criterion

Akaike Information Criterion Akaike Information Criterion Shuhua Hu Center for Research in Scientific Computation North Carolina State University Raleigh, NC February 7, 2012-1- background Background Model statistical model: Y j =

More information

A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED

A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED by W. Robert Reed Department of Economics and Finance University of Canterbury, New Zealand Email: bob.reed@canterbury.ac.nz

More information

Consistency of test based method for selection of variables in high dimensional two group discriminant analysis

Consistency of test based method for selection of variables in high dimensional two group discriminant analysis https://doi.org/10.1007/s42081-019-00032-4 ORIGINAL PAPER Consistency of test based method for selection of variables in high dimensional two group discriminant analysis Yasunori Fujikoshi 1 Tetsuro Sakurai

More information

Outlier detection in ARIMA and seasonal ARIMA models by. Bayesian Information Type Criteria

Outlier detection in ARIMA and seasonal ARIMA models by. Bayesian Information Type Criteria Outlier detection in ARIMA and seasonal ARIMA models by Bayesian Information Type Criteria Pedro Galeano and Daniel Peña Departamento de Estadística Universidad Carlos III de Madrid 1 Introduction The

More information

KULLBACK-LEIBLER INFORMATION THEORY A BASIS FOR MODEL SELECTION AND INFERENCE

KULLBACK-LEIBLER INFORMATION THEORY A BASIS FOR MODEL SELECTION AND INFERENCE KULLBACK-LEIBLER INFORMATION THEORY A BASIS FOR MODEL SELECTION AND INFERENCE Kullback-Leibler Information or Distance f( x) gx ( ±) ) I( f, g) = ' f( x)log dx, If, ( g) is the "information" lost when

More information

Some connections between Bayesian and non-bayesian methods for regression model selection

Some connections between Bayesian and non-bayesian methods for regression model selection Statistics & Probability Letters 57 (00) 53 63 Some connections between Bayesian and non-bayesian methods for regression model selection Faming Liang Department of Statistics and Applied Probability, The

More information

Calculation of maximum entropy densities with application to income distribution

Calculation of maximum entropy densities with application to income distribution Journal of Econometrics 115 (2003) 347 354 www.elsevier.com/locate/econbase Calculation of maximum entropy densities with application to income distribution Ximing Wu Department of Agricultural and Resource

More information

Improved model selection criteria for SETAR time series models

Improved model selection criteria for SETAR time series models Journal of Statistical Planning and Inference 37 2007 2802 284 www.elsevier.com/locate/jspi Improved model selection criteria for SEAR time series models Pedro Galeano a,, Daniel Peña b a Departamento

More information

Generalized Linear Models

Generalized Linear Models Generalized Linear Models Lecture 3. Hypothesis testing. Goodness of Fit. Model diagnostics GLM (Spring, 2018) Lecture 3 1 / 34 Models Let M(X r ) be a model with design matrix X r (with r columns) r n

More information

Regression and Statistical Inference

Regression and Statistical Inference Regression and Statistical Inference Walid Mnif wmnif@uwo.ca Department of Applied Mathematics The University of Western Ontario, London, Canada 1 Elements of Probability 2 Elements of Probability CDF&PDF

More information

An Information Criteria for Order-restricted Inference

An Information Criteria for Order-restricted Inference An Information Criteria for Order-restricted Inference Nan Lin a, Tianqing Liu 2,b, and Baoxue Zhang,2,b a Department of Mathematics, Washington University in Saint Louis, Saint Louis, MO 633, U.S.A. b

More information

Variable Selection in Predictive Regressions

Variable Selection in Predictive Regressions Variable Selection in Predictive Regressions Alessandro Stringhi Advanced Financial Econometrics III Winter/Spring 2018 Overview This chapter considers linear models for explaining a scalar variable when

More information

Averaging Estimators for Regressions with a Possible Structural Break

Averaging Estimators for Regressions with a Possible Structural Break Averaging Estimators for Regressions with a Possible Structural Break Bruce E. Hansen University of Wisconsin y www.ssc.wisc.edu/~bhansen September 2007 Preliminary Abstract This paper investigates selection

More information

Detection of Outliers in Regression Analysis by Information Criteria

Detection of Outliers in Regression Analysis by Information Criteria Detection of Outliers in Regression Analysis by Information Criteria Seppo PynnÄonen, Department of Mathematics and Statistics, University of Vaasa, BOX 700, 65101 Vaasa, FINLAND, e-mail sjp@uwasa., home

More information

On Mixture Regression Shrinkage and Selection via the MR-LASSO

On Mixture Regression Shrinkage and Selection via the MR-LASSO On Mixture Regression Shrinage and Selection via the MR-LASSO Ronghua Luo, Hansheng Wang, and Chih-Ling Tsai Guanghua School of Management, Peing University & Graduate School of Management, University

More information

AR-order estimation by testing sets using the Modified Information Criterion

AR-order estimation by testing sets using the Modified Information Criterion AR-order estimation by testing sets using the Modified Information Criterion Rudy Moddemeijer 14th March 2006 Abstract The Modified Information Criterion (MIC) is an Akaike-like criterion which allows

More information

Asymptotic inference for a nonstationary double ar(1) model

Asymptotic inference for a nonstationary double ar(1) model Asymptotic inference for a nonstationary double ar() model By SHIQING LING and DONG LI Department of Mathematics, Hong Kong University of Science and Technology, Hong Kong maling@ust.hk malidong@ust.hk

More information

Diagnostic tools for random eects in the repeated measures growth curve model

Diagnostic tools for random eects in the repeated measures growth curve model Computational Statistics & Data Analysis 33 (2000) 79 100 www.elsevier.com/locate/csda Diagnostic tools for random eects in the repeated measures growth curve model P.J. Lindsey, J.K. Lindsey Biostatistics,

More information

Alfredo A. Romero * College of William and Mary

Alfredo A. Romero * College of William and Mary A Note on the Use of in Model Selection Alfredo A. Romero * College of William and Mary College of William and Mary Department of Economics Working Paper Number 6 October 007 * Alfredo A. Romero is a Visiting

More information

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures FE661 - Statistical Methods for Financial Engineering 9. Model Selection Jitkomut Songsiri statistical models overview of model selection information criteria goodness-of-fit measures 9-1 Statistical models

More information

Estimating the Marginal Odds Ratio in Observational Studies

Estimating the Marginal Odds Ratio in Observational Studies Estimating the Marginal Odds Ratio in Observational Studies Travis Loux Christiana Drake Department of Statistics University of California, Davis June 20, 2011 Outline The Counterfactual Model Odds Ratios

More information

Augustin: Some Basic Results on the Extension of Quasi-Likelihood Based Measurement Error Correction to Multivariate and Flexible Structural Models

Augustin: Some Basic Results on the Extension of Quasi-Likelihood Based Measurement Error Correction to Multivariate and Flexible Structural Models Augustin: Some Basic Results on the Extension of Quasi-Likelihood Based Measurement Error Correction to Multivariate and Flexible Structural Models Sonderforschungsbereich 386, Paper 196 (2000) Online

More information

1978 M.S. Mathematics, University of Illinois, Chicago Circle Distinguished Professor of Management, University of California, Davis

1978 M.S. Mathematics, University of Illinois, Chicago Circle Distinguished Professor of Management, University of California, Davis EDUCATION CHIH-LING TSAI Graduate School of Management One Shields Avenue University of California, Davis Davis, California 95616-8609 CLTSAI@UCDAVIS.EDU (530) 752-8565 1983 Ph.D. Statistics, University

More information

How the mean changes depends on the other variable. Plots can show what s happening...

How the mean changes depends on the other variable. Plots can show what s happening... Chapter 8 (continued) Section 8.2: Interaction models An interaction model includes one or several cross-product terms. Example: two predictors Y i = β 0 + β 1 x i1 + β 2 x i2 + β 12 x i1 x i2 + ɛ i. How

More information

Monte Carlo Methods for Statistical Inference: Variance Reduction Techniques

Monte Carlo Methods for Statistical Inference: Variance Reduction Techniques Monte Carlo Methods for Statistical Inference: Variance Reduction Techniques Hung Chen hchen@math.ntu.edu.tw Department of Mathematics National Taiwan University 3rd March 2004 Meet at NS 104 On Wednesday

More information

Comparison of New Approach Criteria for Estimating the Order of Autoregressive Process

Comparison of New Approach Criteria for Estimating the Order of Autoregressive Process IOSR Journal of Mathematics (IOSRJM) ISSN: 78-578 Volume 1, Issue 3 (July-Aug 1), PP 1- Comparison of New Approach Criteria for Estimating the Order of Autoregressive Process Salie Ayalew 1 M.Chitti Babu,

More information

Combining time series models for forecasting

Combining time series models for forecasting Statistics Preprints Statistics 10-001 Combining time series models for forecasting Hui Zou Stanford University Yuhong Yang Iowa State University Follow this and additional works at: http://lib.dr.iastate.edu/stat_las_preprints

More information

UNBIASED ESTIMATE FOR b-value OF MAGNITUDE FREQUENCY

UNBIASED ESTIMATE FOR b-value OF MAGNITUDE FREQUENCY J. Phys. Earth, 34, 187-194, 1986 UNBIASED ESTIMATE FOR b-value OF MAGNITUDE FREQUENCY Yosihiko OGATA* and Ken'ichiro YAMASHINA** * Institute of Statistical Mathematics, Tokyo, Japan ** Earthquake Research

More information

Regression I: Mean Squared Error and Measuring Quality of Fit

Regression I: Mean Squared Error and Measuring Quality of Fit Regression I: Mean Squared Error and Measuring Quality of Fit -Applied Multivariate Analysis- Lecturer: Darren Homrighausen, PhD 1 The Setup Suppose there is a scientific problem we are interested in solving

More information

Obtaining simultaneous equation models from a set of variables through genetic algorithms

Obtaining simultaneous equation models from a set of variables through genetic algorithms Obtaining simultaneous equation models from a set of variables through genetic algorithms José J. López Universidad Miguel Hernández (Elche, Spain) Domingo Giménez Universidad de Murcia (Murcia, Spain)

More information

Lecture 7: Model Building Bus 41910, Time Series Analysis, Mr. R. Tsay

Lecture 7: Model Building Bus 41910, Time Series Analysis, Mr. R. Tsay Lecture 7: Model Building Bus 41910, Time Series Analysis, Mr R Tsay An effective procedure for building empirical time series models is the Box-Jenkins approach, which consists of three stages: model

More information

Mathematical Institute, University of Utrecht. The problem of estimating the mean of an observed Gaussian innite-dimensional vector

Mathematical Institute, University of Utrecht. The problem of estimating the mean of an observed Gaussian innite-dimensional vector On Minimax Filtering over Ellipsoids Eduard N. Belitser and Boris Y. Levit Mathematical Institute, University of Utrecht Budapestlaan 6, 3584 CD Utrecht, The Netherlands The problem of estimating the mean

More information

CHOOSING AMONG GENERALIZED LINEAR MODELS APPLIED TO MEDICAL DATA

CHOOSING AMONG GENERALIZED LINEAR MODELS APPLIED TO MEDICAL DATA STATISTICS IN MEDICINE, VOL. 17, 59 68 (1998) CHOOSING AMONG GENERALIZED LINEAR MODELS APPLIED TO MEDICAL DATA J. K. LINDSEY AND B. JONES* Department of Medical Statistics, School of Computing Sciences,

More information

Projektpartner. Sonderforschungsbereich 386, Paper 163 (1999) Online unter:

Projektpartner. Sonderforschungsbereich 386, Paper 163 (1999) Online unter: Toutenburg, Shalabh: Estimation of Regression Coefficients Subject to Exact Linear Restrictions when some Observations are Missing and Balanced Loss Function is Used Sonderforschungsbereich 386, Paper

More information

On Autoregressive Order Selection Criteria

On Autoregressive Order Selection Criteria On Autoregressive Order Selection Criteria Venus Khim-Sen Liew Faculty of Economics and Management, Universiti Putra Malaysia, 43400 UPM, Serdang, Malaysia This version: 1 March 2004. Abstract This study

More information

2.2 Classical Regression in the Time Series Context

2.2 Classical Regression in the Time Series Context 48 2 Time Series Regression and Exploratory Data Analysis context, and therefore we include some material on transformations and other techniques useful in exploratory data analysis. 2.2 Classical Regression

More information

ISyE 691 Data mining and analytics

ISyE 691 Data mining and analytics ISyE 691 Data mining and analytics Regression Instructor: Prof. Kaibo Liu Department of Industrial and Systems Engineering UW-Madison Email: kliu8@wisc.edu Office: Room 3017 (Mechanical Engineering Building)

More information

Weighted tests of homogeneity for testing the number of components in a mixture

Weighted tests of homogeneity for testing the number of components in a mixture Computational Statistics & Data Analysis 41 (2003) 367 378 www.elsevier.com/locate/csda Weighted tests of homogeneity for testing the number of components in a mixture Edward Susko Department of Mathematics

More information

Bootstrap prediction and Bayesian prediction under misspecified models

Bootstrap prediction and Bayesian prediction under misspecified models Bernoulli 11(4), 2005, 747 758 Bootstrap prediction and Bayesian prediction under misspecified models TADAYOSHI FUSHIKI Institute of Statistical Mathematics, 4-6-7 Minami-Azabu, Minato-ku, Tokyo 106-8569,

More information

Unbiased prediction in linear regression models with equi-correlated responses

Unbiased prediction in linear regression models with equi-correlated responses ') -t CAA\..-ll' ~ j... "1-' V'~ /'. uuo. ;). I ''''- ~ ( \ '.. /' I ~, Unbiased prediction in linear regression models with equi-correlated responses Shalabh Received: May 13, 1996; revised version: December

More information

Additive Outlier Detection in Seasonal ARIMA Models by a Modified Bayesian Information Criterion

Additive Outlier Detection in Seasonal ARIMA Models by a Modified Bayesian Information Criterion 13 Additive Outlier Detection in Seasonal ARIMA Models by a Modified Bayesian Information Criterion Pedro Galeano and Daniel Peña CONTENTS 13.1 Introduction... 317 13.2 Formulation of the Outlier Detection

More information

Comparison with RSS-Based Model Selection Criteria for Selecting Growth Functions

Comparison with RSS-Based Model Selection Criteria for Selecting Growth Functions TR-No. 14-07, Hiroshima Statistical Research Group, 1 12 Comparison with RSS-Based Model Selection Criteria for Selecting Growth Functions Keisuke Fukui 2, Mariko Yamamura 1 and Hirokazu Yanagihara 2 1

More information

DYNAMIC ECONOMETRIC MODELS Vol. 9 Nicolaus Copernicus University Toruń Mariola Piłatowska Nicolaus Copernicus University in Toruń

DYNAMIC ECONOMETRIC MODELS Vol. 9 Nicolaus Copernicus University Toruń Mariola Piłatowska Nicolaus Copernicus University in Toruń DYNAMIC ECONOMETRIC MODELS Vol. 9 Nicolaus Copernicus University Toruń 2009 Mariola Piłatowska Nicolaus Copernicus University in Toruń Combined Forecasts Using the Akaike Weights A b s t r a c t. The focus

More information

Bootstrap Variants of the Akaike Information Criterion for Mixed Model Selection

Bootstrap Variants of the Akaike Information Criterion for Mixed Model Selection Bootstrap Variants of the Akaike Information Criterion for Mixed Model Selection Junfeng Shang, and Joseph E. Cavanaugh 2, Bowling Green State University, USA 2 The University of Iowa, USA Abstract: Two

More information

A NOTE ON THE SELECTION OF TIME SERIES MODELS. June 2001

A NOTE ON THE SELECTION OF TIME SERIES MODELS. June 2001 A NOTE ON THE SELECTION OF TIME SERIES MODELS Serena Ng Pierre Perron June 2001 Abstract We consider issues related to the order of an autoregression selected using information criteria. We study the sensitivity

More information

Goodness-of-Fit Tests for Time Series Models: A Score-Marked Empirical Process Approach

Goodness-of-Fit Tests for Time Series Models: A Score-Marked Empirical Process Approach Goodness-of-Fit Tests for Time Series Models: A Score-Marked Empirical Process Approach By Shiqing Ling Department of Mathematics Hong Kong University of Science and Technology Let {y t : t = 0, ±1, ±2,

More information

Statistical Practice. Selecting the Best Linear Mixed Model Under REML. Matthew J. GURKA

Statistical Practice. Selecting the Best Linear Mixed Model Under REML. Matthew J. GURKA Matthew J. GURKA Statistical Practice Selecting the Best Linear Mixed Model Under REML Restricted maximum likelihood (REML) estimation of the parameters of the mixed model has become commonplace, even

More information

Comparison with Residual-Sum-of-Squares-Based Model Selection Criteria for Selecting Growth Functions

Comparison with Residual-Sum-of-Squares-Based Model Selection Criteria for Selecting Growth Functions c 215 FORMATH Research Group FORMATH Vol. 14 (215): 27 39, DOI:1.15684/formath.14.4 Comparison with Residual-Sum-of-Squares-Based Model Selection Criteria for Selecting Growth Functions Keisuke Fukui 1,

More information

Combining time series models for forecasting

Combining time series models for forecasting International Journal of Forecasting 20 (2004) 69 84 www.elsevier.com/locate/ijforecast Combining time series models for forecasting Hui Zou a, Yuhong Yang b, * a Department of Statistics, Sequoia Hall,

More information

Regression, Ridge Regression, Lasso

Regression, Ridge Regression, Lasso Regression, Ridge Regression, Lasso Fabio G. Cozman - fgcozman@usp.br October 2, 2018 A general definition Regression studies the relationship between a response variable Y and covariates X 1,..., X n.

More information

Marcia Gumpertz and Sastry G. Pantula Department of Statistics North Carolina State University Raleigh, NC

Marcia Gumpertz and Sastry G. Pantula Department of Statistics North Carolina State University Raleigh, NC A Simple Approach to Inference in Random Coefficient Models March 8, 1988 Marcia Gumpertz and Sastry G. Pantula Department of Statistics North Carolina State University Raleigh, NC 27695-8203 Key Words

More information

Performance of lag length selection criteria in three different situations

Performance of lag length selection criteria in three different situations MPRA Munich Personal RePEc Archive Performance of lag length selection criteria in three different situations Zahid Asghar and Irum Abid Quaid-i-Azam University, Islamabad Aril 2007 Online at htts://mra.ub.uni-muenchen.de/40042/

More information

Nonlinear Autoregressive Processes with Optimal Properties

Nonlinear Autoregressive Processes with Optimal Properties Nonlinear Autoregressive Processes with Optimal Properties F. Blasques S.J. Koopman A. Lucas VU University Amsterdam, Tinbergen Institute, CREATES OxMetrics User Conference, September 2014 Cass Business

More information

Thomas J. Fisher. Research Statement. Preliminary Results

Thomas J. Fisher. Research Statement. Preliminary Results Thomas J. Fisher Research Statement Preliminary Results Many applications of modern statistics involve a large number of measurements and can be considered in a linear algebra framework. In many of these

More information

Re-sampling and exchangeable arrays University Ave. November Revised January Summary

Re-sampling and exchangeable arrays University Ave. November Revised January Summary Re-sampling and exchangeable arrays Peter McCullagh Department of Statistics University of Chicago 5734 University Ave Chicago Il 60637 November 1997 Revised January 1999 Summary The non-parametric, or

More information

FULL LIKELIHOOD INFERENCES IN THE COX MODEL

FULL LIKELIHOOD INFERENCES IN THE COX MODEL October 20, 2007 FULL LIKELIHOOD INFERENCES IN THE COX MODEL BY JIAN-JIAN REN 1 AND MAI ZHOU 2 University of Central Florida and University of Kentucky Abstract We use the empirical likelihood approach

More information

Geographically Weighted Regression as a Statistical Model

Geographically Weighted Regression as a Statistical Model Geographically Weighted Regression as a Statistical Model Chris Brunsdon Stewart Fotheringham Martin Charlton October 6, 2000 Spatial Analysis Research Group Department of Geography University of Newcastle-upon-Tyne

More information

Combining estimates in regression and. classication. Michael LeBlanc. and. Robert Tibshirani. and. Department of Statistics. cuniversity of Toronto

Combining estimates in regression and. classication. Michael LeBlanc. and. Robert Tibshirani. and. Department of Statistics. cuniversity of Toronto Combining estimates in regression and classication Michael LeBlanc and Robert Tibshirani Department of Preventive Medicine and Biostatistics and Department of Statistics University of Toronto December

More information

2 FRED J. HICKERNELL the sample mean of the y (i) : (2) ^ 1 N The mean square error of this estimate may be written as a sum of two parts, a bias term

2 FRED J. HICKERNELL the sample mean of the y (i) : (2) ^ 1 N The mean square error of this estimate may be written as a sum of two parts, a bias term GOODNESS OF FIT STATISTICS, DISCREPANCIES AND ROBUST DESIGNS FRED J. HICKERNELL Abstract. The Cramer{Von Mises goodness-of-t statistic, also known as the L 2 -star discrepancy, is the optimality criterion

More information

Non-maximum likelihood estimation and statistical inference for linear and nonlinear mixed models

Non-maximum likelihood estimation and statistical inference for linear and nonlinear mixed models Optimum Design for Mixed Effects Non-Linear and generalized Linear Models Cambridge, August 9-12, 2011 Non-maximum likelihood estimation and statistical inference for linear and nonlinear mixed models

More information

ADJUSTED POWER ESTIMATES IN. Ji Zhang. Biostatistics and Research Data Systems. Merck Research Laboratories. Rahway, NJ

ADJUSTED POWER ESTIMATES IN. Ji Zhang. Biostatistics and Research Data Systems. Merck Research Laboratories. Rahway, NJ ADJUSTED POWER ESTIMATES IN MONTE CARLO EXPERIMENTS Ji Zhang Biostatistics and Research Data Systems Merck Research Laboratories Rahway, NJ 07065-0914 and Dennis D. Boos Department of Statistics, North

More information

A Study on the Bias-Correction Effect of the AIC for Selecting Variables in Normal Multivariate Linear Regression Models under Model Misspecification

A Study on the Bias-Correction Effect of the AIC for Selecting Variables in Normal Multivariate Linear Regression Models under Model Misspecification TR-No. 13-08, Hiroshima Statistical Research Group, 1 23 A Study on the Bias-Correction Effect of the AIC for Selecting Variables in Normal Multivariate Linear Regression Models under Model Misspecification

More information

Variable Selection in Multivariate Linear Regression Models with Fewer Observations than the Dimension

Variable Selection in Multivariate Linear Regression Models with Fewer Observations than the Dimension Variable Selection in Multivariate Linear Regression Models with Fewer Observations than the Dimension (Last Modified: November 3 2008) Mariko Yamamura 1, Hirokazu Yanagihara 2 and Muni S. Srivastava 3

More information

arxiv: v2 [stat.co] 6 Oct 2014

arxiv: v2 [stat.co] 6 Oct 2014 Model selection criteria in beta regression with varying dispersion Fábio M. Bayer Francisco Cribari Neto Abstract arxiv:1405.3718v [stat.co] 6 Oct 014 We address the issue of model selection in beta regressions

More information

Journal of Statistical Research 2007, Vol. 41, No. 1, pp Bangladesh

Journal of Statistical Research 2007, Vol. 41, No. 1, pp Bangladesh Journal of Statistical Research 007, Vol. 4, No., pp. 5 Bangladesh ISSN 056-4 X ESTIMATION OF AUTOREGRESSIVE COEFFICIENT IN AN ARMA(, ) MODEL WITH VAGUE INFORMATION ON THE MA COMPONENT M. Ould Haye School

More information

Gaussian processes. Basic Properties VAG002-

Gaussian processes. Basic Properties VAG002- Gaussian processes The class of Gaussian processes is one of the most widely used families of stochastic processes for modeling dependent data observed over time, or space, or time and space. The popularity

More information

Shrinkage Estimation of the Slope Parameters of two Parallel Regression Lines Under Uncertain Prior Information

Shrinkage Estimation of the Slope Parameters of two Parallel Regression Lines Under Uncertain Prior Information Shrinkage Estimation of the Slope Parameters of two Parallel Regression Lines Under Uncertain Prior Information Shahjahan Khan Department of Mathematics & Computing University of Southern Queensland Toowoomba,

More information

Modfiied Conditional AIC in Linear Mixed Models

Modfiied Conditional AIC in Linear Mixed Models CIRJE-F-895 Modfiied Conditional AIC in Linear Mixed Models Yuki Kawakubo Graduate School of Economics, University of Tokyo Tatsuya Kubokawa University of Tokyo July 2013 CIRJE Discussion Papers can be

More information

Simple Regression Model Setup Estimation Inference Prediction. Model Diagnostic. Multiple Regression. Model Setup and Estimation.

Simple Regression Model Setup Estimation Inference Prediction. Model Diagnostic. Multiple Regression. Model Setup and Estimation. Statistical Computation Math 475 Jimin Ding Department of Mathematics Washington University in St. Louis www.math.wustl.edu/ jmding/math475/index.html October 10, 2013 Ridge Part IV October 10, 2013 1

More information

STAT Financial Time Series

STAT Financial Time Series STAT 6104 - Financial Time Series Chapter 4 - Estimation in the time Domain Chun Yip Yau (CUHK) STAT 6104:Financial Time Series 1 / 46 Agenda 1 Introduction 2 Moment Estimates 3 Autoregressive Models (AR

More information

11. Bootstrap Methods

11. Bootstrap Methods 11. Bootstrap Methods c A. Colin Cameron & Pravin K. Trivedi 2006 These transparencies were prepared in 20043. They can be used as an adjunct to Chapter 11 of our subsequent book Microeconometrics: Methods

More information

BAYESIAN ESTIMATION OF LINEAR STATISTICAL MODEL BIAS

BAYESIAN ESTIMATION OF LINEAR STATISTICAL MODEL BIAS BAYESIAN ESTIMATION OF LINEAR STATISTICAL MODEL BIAS Andrew A. Neath 1 and Joseph E. Cavanaugh 1 Department of Mathematics and Statistics, Southern Illinois University, Edwardsville, Illinois 606, USA

More information

Hi-Stat Discussion Paper. Research Unit for Statistical and Empirical Analysis in Social Sciences (Hi-Stat)

Hi-Stat Discussion Paper. Research Unit for Statistical and Empirical Analysis in Social Sciences (Hi-Stat) Global COE Hi-Stat Discussion Paper Series 006 Research Unit for Statistical and Empirical Analysis in Social Sciences (Hi-Stat) Model Selection Criteria for the Leads-and-Lags Cointegrating Regression

More information

The New Palgrave Dictionary of Economics Online

The New Palgrave Dictionary of Economics Online The New Palgrave Dictionary of Economics Online model selection Jean-Marie Dufour From The New Palgrave Dictionary of Economics, Second Edition, 2008 Edited by Steven N. Durlauf and Lawrence E. Blume Abstract

More information

INTERNALLY STUDENTIZED RESIDUALS TEST FOR MULTIPLE NON-NESTED REGRESSION MODELS

INTERNALLY STUDENTIZED RESIDUALS TEST FOR MULTIPLE NON-NESTED REGRESSION MODELS International Journal of Physics and Mathematical Sciences ISSN: 77-111 (Online) 014 Vol. 4 (1) January-March, pp.49-56/naidu et al. INTERNALLY STUENTIZE RESIUALS TEST FOR MULTIPLE NON-NESTE REGRESSION

More information

MODEL SELECTION BASED ON QUASI-LIKELIHOOD WITH APPLICATION TO OVERDISPERSED DATA

MODEL SELECTION BASED ON QUASI-LIKELIHOOD WITH APPLICATION TO OVERDISPERSED DATA J. Jpn. Soc. Comp. Statist., 26(2013), 53 69 DOI:10.5183/jjscs.1212002 204 MODEL SELECTION BASED ON QUASI-LIKELIHOOD WITH APPLICATION TO OVERDISPERSED DATA Yiping Tang ABSTRACT Overdispersion is a common

More information

Econometric Forecasting

Econometric Forecasting Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 1, 2014 Outline Introduction Model-free extrapolation Univariate time-series models Trend

More information

Model selection for good estimation or prediction over a user-specified covariate distribution

Model selection for good estimation or prediction over a user-specified covariate distribution Graduate Theses and Dissertations Iowa State University Capstones, Theses and Dissertations 2010 Model selection for good estimation or prediction over a user-specified covariate distribution Adam Lee

More information

DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS

DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS ISSN 1440-771X AUSTRALIA DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS Empirical Information Criteria for Time Series Forecasting Model Selection Md B Billah, R.J. Hyndman and A.B. Koehler Working

More information

2 T. Schneider and A. Neumaier m-dimensional random vectors with zero mean and covariance matrix C. The m- dimensional parameter vector w of intercept

2 T. Schneider and A. Neumaier m-dimensional random vectors with zero mean and covariance matrix C. The m- dimensional parameter vector w of intercept Algorithm: ARfit A Matlab Package for Estimation and Spectral Decomposition of Multivariate Autoregressive Models Tapio Schneider Princeton University and Arnold Neumaier Universitat Wien ARfit is a collection

More information

A characterization of consistency of model weights given partial information in normal linear models

A characterization of consistency of model weights given partial information in normal linear models Statistics & Probability Letters ( ) A characterization of consistency of model weights given partial information in normal linear models Hubert Wong a;, Bertrand Clare b;1 a Department of Health Care

More information

ON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES. 1. Introduction

ON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES. 1. Introduction Acta Math. Univ. Comenianae Vol. LXV, 1(1996), pp. 129 139 129 ON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES V. WITKOVSKÝ Abstract. Estimation of the autoregressive

More information