Doubly Constrained Factor Models with Applications. Summary

Size: px
Start display at page:

Download "Doubly Constrained Factor Models with Applications. Summary"

Transcription

1 Doubly Constrained Factor Models with Applications Henghsiu Tsai 1 Institute of Statistical Science, Academia Sinica, Taiwan, R.O.C. Ruey S. Tsay Booth School of Business, University of Chicago, Illinois, U.S.A. Edward M. H. Lin Institute of Statistical Science, Academia Sinica, Taiwan, R.O.C. and Ching-Wei Cheng Department of Statistics, Purdue University, Indiana, U.S.A. Summary This paper focuses on factor analysis of high-dimensional data. We propose statistical methods that enable an analyst to make use of prior knowledge or substantive information to sharpen the estimation of common factors. Specifically, we consider a doubly constrained factor model that enables analysts to specify both row and column constraints of the data matrix to improve the estimation of common factors. The row constraints may represent classifications of individual subjects whereas the column constraints may show the categories of variables. We derive both the maximum likelihood and least squares estimates of the proposed doubly constrained factor model and use simulation to study the performance of the analysis in finite samples. Akaike information criterion is used for model selection. Monthly U.S. housing starts of nine geographical divisions are used to demonstrate the application of the proposed model. Keywords: Akaike information criterion, Constrained factor model, Eigenvalues, Factor model, Housing starts, Principal component analysis, Seasonality. 1 Corresponding author: Henghsiu Tsai, Institute of Statistical Science, Academia Sinica, 128 Academia Road, Section 2, Nankang, Taipei 11529, Taiwan, R.O.C. Tel: , Fax: , htsai@stat.sinica.edu.tw. 1

2 1 Introduction Big data have become common in statistical applications. In many situations, it is natural to entertain the data as a 2-dimensional array with row representing subjects and column denoting variables. See, for instance, the large panel data in the econometric literature and the multivariate time series analysis in statistics. Consider specifically the United States (U.S.) housing markets. The U.S. Census Bureau publishes monthly housing starts of nine geographical divisions shown in Figure 1. We employ 10 years of the data from January 1997 to December Here the data matrix Z is a 120-by-9 matrix with each column representing a division and each row denoting a particular calendar month. Figure 2 shows the time plots of the logarithms of monthly housing starts of the nine divisions. From the plots, it is clear that U.S. housing starts have strong seasonality. Furthermore, the housing starts also exhibit some common characteristics. It is then natural to consider both the seasonality and geographical divisions in searching for the common factors driving the U.S. housing markets. In this particular example, the seasonality leads naturally to row constraints whereas the geographical considerations give rise to column constraints. The goal of this paper is to consider such constraints when we search for common factors in a big data set. Factor models are widely used in econometric and statistical applications, and constrained factor models have also been studied in the literature. See, for instance, Lam and Yao (2012) and the references therein. Tsai and Tsay (2010) proposed constrained and partially constrained factor models for high-dimensional time series analysis. They show that column constraints can be used effectively to obtain parsimonious factor models for high-dimensional series. Only column constraints are considered in that paper, however. On the other hand, as illustrated by the U.S. housing starts data, both row and column constraints might be informative in some applications. Therefore, we investigate doubly constrained factor models in this paper. The theoretical framework of the proposed model is the constrained principal component analysis of Takane and Hunter (2001), and our study focuses on estimation of the proposed model. Principal component analysis was proposed originally for independent data, but it has been widely used in time series analysis. See, for instance, Peña and Box (1987) and Tiao, Tsay, and Wang (1993). Consider a T by N data matrix Z, rows and columns of which often represent subjects and 2

3 variables, respectively. Let G be a T by m matrix of row constraints of rank m, and H be an N by s matrix of column constraints of rank s. Both G and H are known a priori based on some prior knowledge or substantive information of the problem at hand. For instance, Tsai and Tsay (2010) use H to obtain the level, slope, and curvature of interest rates. They also use H to denote the industrial classification of stock returns. Let ω 1 (s by r), ω 2 (N by p) and ω 3 (s by q) be the loading matrices of full rank, and E (T by N) a matrix of residuals, where p < N, max{r, q} s < N, and q min{r, p}. The postulated doubly constrained factor (DCF) model for Z = [Z i,j ] = [Z 1,, Z T ] is Z = F 1 ω 1H + GF 2 ω 2 + GF 3 ω 3H + E, (1) where A denotes the transpose matrix of A, F 1 = [F (1) 1,..., F (T ) 1 ] (T by r), F 2 = [F (1) 2,..., F (m) 2 ] (m by p), F 3 = [F (1) 3,..., F (m) 3 ] (m by q), and E = [e 1,..., e T ] (T by N) with E(e i ) = 0 and var(e i ) = Ψ. We refer to the model in Equation (1) as a DCF model of order (r, p, q) with r, p and q denoting the number of common factors in F 1, F 2 and F 3, respectively. For statistical factor models, one further assumes that Ψ is a diagonal matrix. In the econometric and finance literature, Ψ is not necessarily diagonal and the model becomes an approximate factor model. For the DCF model in Equation (1), F i are common factors. The first term of model pertains to what in Z can be explained by H but not by G, the second term to what can be explained by G but not by H, the third term to what can be explained by both G and H, and the last term to what can be explained by neither G nor H. Often the third term of model (1) denotes the interaction between the constraints G and H. Thus, F 1, F 2 and F 3 can be interpreted as column, row, and interaction factors, respectively. Similar to the conventional factor models, the scales and orderings of the latent common factors F i are not identifiable. The paper is organized as follows. In Section 2 we consider estimation of the proposed DCF model, including model selection and the common factors. We use simulation in Section 3 to investigate the efficacy of the estimation methods. Section 4 applies the proposed analysis to the monthly U.S. housing starts, and Section 5 concludes. 3

4 2 Estimation The proposed doubly constrained factor model in Equation (1) can be estimated by either the least squares (LS) method or the maximum likelihood method. The LS estimates are less efficient, but easier to obtain. We begin with the LS method. 2.1 Least Squares Estimation Least squares estimates (LSE) of the doubly constrained factor model in (1) can be obtained by several steps under certain assumptions. Consider first the simple factor model, Z = F ω +E, with the assumption that the factor matrix F satisfies F F = T I r, where I r denotes the r by r identity matrix with r being the rank of ω. In this case, LSEs of F and ω can be obtained by minimizing the objective function l(f, ω) = tr { (Z F ω )(Z F ω ) } = tr(zz ) + tr(f ω ωf ) 2 tr(zωf ). Therefore, it can easily be seen that F = [g 1,..., g r ] (2) where g i is an eigenvector corresponding to the i th largest eigenvalue of the matrix ZZ, and the estimator for ω is ω = 1 T Z F. (3) See, for instance, Proposition A.5 of Lütkepohl (2005). Consider the doubly constrained factor model in (1) with the following assumptions: F 1F 1 = T I r, F 2F 2 = mi p, and F 3F 3 = mi q, G F 1 ω 1 = O and F 2ω 2H = O. The proposed LS estimation procedure is as follows: (i) Pre- and post-multiplying (1) by (G G) 1 G and H(H H) 1, we have (G G) 1 G ZH(H H) 1 = F 3 ω 3 + (G G) 1 G EH(H H) 1, which becomes an unconstrained factor model, and therefore we can solve F 3 and ω 3 using (2) and (3) by substituting (G G) 1 G ZH(H H) 1 for Z. 4

5 (ii) Let Z = Z G F 3 ω 3H and the model becomes Z = F 1 ω 1H + GF 2 ω 2 + E. (4) (iii) Post-multiplying (4) by H(H H) 1, we have Z H(H H) 1 = F 1 ω 1 + E H(H H) 1. Solve F 1 and ω 1 using (2) and (3) by substituting Z H(H H) 1 for Z. (iv) Let Z = Z F 1 ω 1H and the model becomes Z = GF 2 ω 2 + E. (5) (v) Pre-multiplying (5) by (G G) 1 G, we have (G G) 1 G Z = F 2 ω 2 + (G G) 1 G E. Solve F 2 and ω 2 using (2) and (3) by substituting (G G) 1 G Z for Z. (vi) Finally, Ê = Z F 1 ω 1H G F 2 ω 2 G F 3 ω 3H and Ψ = Ê Ê/T. It is understood that Ψ = diag(ê Ê/T ) if Ψ is diagonal. 2.2 Maximum Likelihood Estimation For maximum likelihood estimation, we assume that, for 1 k T, var(e t ) = Ψ is a diagonal N by N matrix. We further assume that E(F (k) i ) = 0, and var(f (k) i ) = I, the identity matrix, for 1 i 3, and all k. We also assume cov(f (k) i, F (l) j ) = 0 for k l or i j, cov(e i, e j ) = 0 for all i j, cov(f (k) i, e j ) = 0 for all i, j, and k, and e k is a N-dimensional Gaussian random vector with mean zero and diagonal covariance matrix Ψ. For the purpose of identifiability, we adopt the approach of Anderson (2003) by imposing the restrictions that the matrices Γ 1, Γ 2, and Γ 3 are all diagonal, where Γ 1 = ω 1H Ψ 1 Hω 1, Γ 2 = ω 2Ψ 1 ω 2, Γ 3 = ω 3H Ψ 1 Hω 3. (6) We also assume that the diagonal elements of Γ 1, Γ 2 and Γ 3 are ordered and distinct (γ 1 11 > γ 1 22 > > γ1 rr, γ 2 11 > γ2 22 > > γ2 pp, and γ 3 11 > γ3 22 > > γ3 qq), and the first non-zero 5

6 element in each column of the matrix ω i, i = 1, 2, 3, is positive, so ω 1, ω 2 and ω 3 are uniquely determined. It can readily be checked that the covariance of vec(z ) is Σ = I T A+GG B, where A = Hω 1 ω 1 H + Ψ, and B = ω 2 ω 2 + Hω 3ω 3 H. For the definitions of the matrix operators vec( ) and, see, for example, Schott (1997). We divide the discussion of maximum likelihood estimation into subsections to better understand the flexibility of the proposed model. Also, the existence of row constraints requires an additional condition to simplify the estimation Case 1: ω 2 = ω 3 = 0 In this particular case, the proposed model becomes Z = F 1 ω 1H + E, which is the column constrained factor model of Tsai and Tsay (2010). An iterated procedure was proposed there to perform estimation Case 2: ω 1 = ω 3 = 0 When ω 1 = ω 3 = 0, the doubly constrained factor model becomes Z = GF 2 ω 2 + E. Here the model can be estimated by an iterated procedure similar to that of Tsai and Tsay (2010). Let Y = (G G) 1 G Z, and C Y = Y Y /m. The estimating procedure is as follows: 1. Compute initial estimates of the diagonal matrix Ψ = [ Ψ ij ]. Following Jöreskog (1975), we set Ψ ii = (1 r/(2n))/s ii, i = 1,..., N, where s ii is the ith diagonal element of S 1 and S = Z Z/(T 1). 2. Construct the symmetric matrix R B = Ψ 1/2 (C Y m Ψ/T ) Ψ 1/2 and perform a spectral decomposition on R B, say R B = L B W B L B, where W B = diag(ˆγ j ) and ˆγ 1 > ˆγ 2 > > ˆγ N are the ordered eigenvalues of R B. 3. Let Γ B = W B and Γ 2 = W 2, where W 2 is the left-upper r r submatrix of W B. Obtain ω 2 from Ψ 1/2 ω 2 = L 2, where L 2 consists of the first r columns of L B. The eigenvectors 6

7 are normalized such that ω 2 Ψ 1 ω 2 = Γ 2. More precisely, ω 2 is a normalized version of ω 2 = Ψ 1/2 L 2, where the normalization is to ensure that ω 2 Ψ 1 ω 2 = Γ 2, a diagonal matrix. 4. Substitute ω 2 obtained in Step 3 into the objective function m T ln Q + T m ln T Ψ + tr(c Q 1 Y ) + tr((c CY ) Ψ 1 ), (7) where Q = T ω 2 ω 2/m + Ψ, and minimize (7) with respect to Ψ 11,..., Ψ NN. A numerical search routine must be used. The resulting values Ψ 11,..., Ψ NN are employed at Steps 2 and 3 to create a new ω 2. Steps 2, 3 and 4 are repeated until convergence, i.e., until the differences between successive values of ω 2 ij in ω 2 = [ ω 2 ij ] and Ψ ii are negligible Case 3. The Full Model In this case, the log-likelihood function of vec(z ) is log f(vec(z )) = T N 2 log(2π) 1 2 log Σ 1 2 {vec(z )} Σ 1 vec(z ). To evaluate the prior log-likelihood function, we assume, for simplicity, that the row constraint G satisfies G G = T m I m. (8) This is not a strong condition and it can be met easily. For example, if (i) G = I m 1 T/m, where 1 T/m is the T/m-dimensional vector of 1, or if (ii) G = 1 T/m I m, then Assumption (8) holds. Note that (ii) is the case for the U.S. housing starts data. Lemma 1. If G G = T m I m, then (a) Σ = Q m A T m, where Q = A + T m B, (b) Σ 1 = I T A 1 + GG U, where U = m T (Q 1 A 1 ). (c) {vec(z )} Σ 1 vec(z ) = tr(za 1 Z ) + tr(zuz GG ). Proof: See the Appendix. 7

8 Let C = Z Z/T, Y = (G G) 1 G Z, and C Y = Y Y /m, then by Equation (8), Lemma 1 (a) and (c), the log likelihood function of ω 1, ω 2, ω 3 and Ψ based on Z is ln L(ω 1, ω 2, ω 3, Ψ) = T N 2 ln(2π) m 2 = T N 2 ln(2π) m 2 T 2 tr[(c C Y )A 1 ]. T m ln Q 2 T m ln Q 2 Thus the objective function can be written as ln A T 2 2m tr(c Y U) T 2 tr(ca 1 ) ln A T 2 tr(c Y Q 1 ) 2 ln L(θ) = T N ln(2π) + m ln Q + (T m) ln A + T tr(c Y Q 1 ) + T tr[(c C Y )A 1 ], (9) where A = Hω 1 ω 1 H + Ψ, B = ω 2 ω 2 + Hω 3ω 3 H, Q = A + T B/m, and we minimize (9) with respect of θ = (ω 1, ω 2, ω 3, Ψ) to obtain the maximum likelihood estimate ˆθ. Note that equation (7) is a special case of equation (9) Case 4. ω 3 = 0 In this case, there is no interaction between the row and column constraints, and the model becomes The associated objective function is Z = F 1 ω 1H + GF 2 ω 2 + E. 2 ln L(θ) = T N ln(2π) + m ln Q + (T m) ln A + T tr(c Y Q 1 ) + T tr((c C Y )A 1 ), (10) where θ = (ω 1, ω 2, Ψ), A = Hω 1 ω 1 H + Ψ, and Q = A + T ω 2 ω 2 /m. We minimize (10) to obtain the estimate ˆθ Initial Estimates for Cases 3 and 4 For Cases 1 and 2, the MLE are computed by iterated procedures. For Cases 3 and 4, no iterative procedure is available, and the MLE must be obtained by some numerical optimization method with certain initial estimates. We use the LS estimates of subsection 2.1 as the initial estimates. 8

9 2.3 Estimation of Latent Factors for Maximum Likelihood Approach Treating the ML estimates of ω i as given, we can estimate the latent factors F i by using the weighted least squares method. Specifically, given ω 1, ω 2, and ω 3, the weighted least squares estimates of F 1, F 2, and F 3 can be obtained by minimizing f(f 1, F 2, F 3 ) = tr(eψ 1 E ) = tr((z F 1 ω 1 H GF 2 ω 2 GF 3ω 3 H )Ψ 1 (Z F 1 ω 1 H GF 2 ω 2 GF 3ω 3 H ) ). Taking the partial derivative of f(f 1, F 2, F 3 ) with respect to F 1, and equating the result to zero, we obtain f F 1 = F 1 tr( 2F 1 ω 1H Ψ 1 (Z GF 2 ω 2 GF 3 ω 3H ) + F 1 ω 1H Ψ 1 Hω 1 F 1) = 2(Z GF 2 ω 2 GF 3 ω 3H )Ψ 1 Hω 1 + 2F 1 ω 1H Ψ 1 Hω 1 = 0. (11) The second equality follows from the fact that Equation (11) implies that X tr(ax) = A, X tr(xax B) = BXA + B XA. F 1 = (Z GF 2 ω 2 GF 3 ω 3H )Ψ 1 Hω 1 (ω 1H Ψ 1 Hω 1 ) 1. (12) Similarly, f F 2 = F 2 tr( 2GF 2 ω 2Ψ 1 (Z F 1 ω 1H GF 3 ω 3H ) + GF 2 ω 2Ψ 1 ω 2 F 2G ) = 2G (Z F 1 ω 1H GF 3 ω 3H )Ψ 1 ω 2 + 2G GF 2 ω 2Ψ 1 ω 2 = 0. Let G = (G G) 1 G, and Ḡ = G(G G) 1 G, then F 2 = G(Z F 1 ω 1H GF 3 ω 3H )Ψ 1 ω 2 (ω 2Ψ 1 ω 2 ) 1. (13) Thirdly, F F 3 = F 3 tr( GF 3 ω 3H Ψ 1 (Z F 1 ω 1H GF 2 ω 2) + GF 3 ω 3H Ψ 1 Hω 3 F 3G ) = 2G (Z F 1 ω 1H GF 2 ω 2)Ψ 1 Hω 3 + 2G GF 3 ω 3H Ψ 1 Hω 3 = 0. 9

10 Therefore, F 3 = G(Z F 1 ω 1H GF 2 ω 2)Ψ 1 Hω 3 (ω 3H Ψ 1 Hω 3 ) 1. (14) Using (6) and let Γ 12 = Γ 21 = ω 1 H Ψ 1 ω 2, Γ 13 = Γ 31 = ω 1 H Ψ 1 Hω 3, Γ 23 = Γ 32 = ω 2 Ψ 1 Hω 3, Γ 01 = ZΨ 1 Hω 1, Γ 02 = ZΨ 1 ω 2, and Γ 03 = ZΨ 1 Hω 3, then Equations (12), (13), and (14) becomes Multiplying both sides of (17) by Γ 3 to get Plugging (16) into (18) to get F 1 = (Γ 01 GF 2 Γ 21 GF 3 Γ 31 )Γ 1 1, (15) F 2 = ( GΓ 02 GF 1 Γ 12 F 3 Γ 32 )Γ 1 2, (16) F 3 = ( GΓ 03 GF 1 Γ 13 F 2 Γ 23 )Γ 1 3, (17) F 3 Γ 3 = GΓ 03 GF 1 Γ 13 F 2 Γ 23, (18) F 3 Γ 3 = GΓ 03 GF 1 Γ 13 ( GΓ 02 GF 1 Γ 12 F 3 Γ 32 )Γ 1 2 Γ 23 = G{F 1 (Γ 12 Γ 1 2 Γ 23 Γ 13 ) + Γ 03 Γ 02 Γ 1 2 Γ 23} + F 3 Γ 32 Γ 1 2 Γ 23. Subtracting both sides by F 3 Γ 32 Γ 1 2 Γ 23, and then post-multiplying by 32 = (Γ 3 Γ 32 Γ 1 2 Γ 23) 1 to get F 3 = G{F 1 (Γ 12 Γ 1 2 Γ 23 Γ 13 ) + Γ 03 Γ 02 Γ 1 2 Γ 23} 32. (19) Similarly, multiplying both sides of (16) by Γ 2 to get Plugging (17) into (20) to get F 2 Γ 2 = GΓ 02 GF 1 Γ 12 F 3 Γ 32. (20) F 2 Γ 2 = GΓ 02 GF 1 Γ 12 ( GΓ 03 GF 1 Γ 13 F 2 Γ 23 )Γ 1 3 Γ 32 = G{F 1 (Γ 13 Γ 1 3 Γ 32 Γ 12 ) + Γ 02 Γ 03 Γ 1 3 Γ 32} + F 2 Γ 23 Γ 1 3 Γ 32. Subtracting both sides by F 2 Γ 23 Γ 1 3 Γ 32, and then post-multiplying by 23 = (Γ 2 Γ 23 Γ 1 3 Γ 32) 1 to get F 2 = G{F 1 (Γ 13 Γ 1 3 Γ 32 Γ 12 ) + Γ 02 Γ 03 Γ 1 3 Γ 32} 23. (21) 10

11 Now, multiplying both sides of (15) by Γ 1, we have Plugging (19) and (21) into (22), we obtain F 1 Γ 1 = Γ 01 GF 2 Γ 21 GF 3 Γ 31. (22) F 1 Γ 1 = Γ 01 Ḡ{F 1 (Γ 13 Γ 1 3 Γ 32 Γ 12 ) + Γ 02 Γ 03 Γ 1 3 Γ 32} 23 Γ 21 Ḡ{F 1 (Γ 12 Γ 1 2 Γ 23 Γ 13 ) + Γ 03 Γ 02 Γ 1 2 Γ 23} 32 Γ 31. (23) Pre-multiplying both sides of (23) by G, and noting that G Ḡ = G, we have G F 1 Γ 1 = G Γ 01 G {F 1 (Γ 13 Γ 1 3 Γ 32 Γ 12 ) + Γ 02 Γ 03 Γ 1 3 Γ 32} 23 Γ 21 One solution to equation (24) is G {F 1 (Γ 12 Γ 1 2 Γ 23 Γ 13 ) + Γ 03 Γ 02 Γ 1 2 Γ 23} 32 Γ 31. (24) F 1 Γ 1 = Γ 01 {F 1 (Γ 13 Γ 1 3 Γ 32 Γ 12 ) + Γ 02 Γ 03 Γ 1 3 Γ 32} 23 Γ 21 From equation (25), we obtain {F 1 (Γ 12 Γ 1 2 Γ 23 Γ 13 ) + Γ 03 Γ 02 Γ 1 2 Γ 23} 32 Γ 31. (25) F 1 = { Γ 01 (Γ 02 Γ 03 Γ 1 3 Γ 32) 23 Γ 21 (Γ 03 Γ 02 Γ 1 2 Γ } 23) 32 Γ 31 { Γ1 + (Γ 13 Γ 1 3 Γ 32 Γ 12 ) 23 Γ 21 + (Γ 12 Γ 1 2 Γ } 1 23 Γ 13 ) 32 Γ 31. (26) Therefore, we use Equation (26) to compute F 1 first, then we use Equation (21) to compute F 2, and Equation (19) to compute F Model Selection In applications, the data generating process is unknown and one needs to select a proper constrained factor model based on the available data. In particular, the validity of row and/or column constraints must be verified. To this end, we consider the Akaike information criterion (AIC) (Akaike, 1974) for each of the fitted model, AIC = 2 ln L(ˆθ) + 2λ, where λ is the number of parameters of the model, and ˆθ is the MLE. Our simulation study and empirical example show that AIC works well in model selection. 11

12 Tsai and Tsay (2010) used hypothesis testing to check the validity of column constraints. The testing procedure becomes complicated for doubly constrained factor models because it would involve non-nested hypothesis testing. For instance, the model with only column constraints is not a sub-model of the one with only row constraints. 3 Simulation Study In this section, we report some finite-sample performance of the MLE and the AIC in Subsections 3.1 and 3.2, respectively. All computations in this section were performed using some Fortran code with IMSL subroutines. 3.1 Finite Sample Properties of the MLE We are interested in evaluating the performance of the numerical optimization in finding the MLE. The true model considered is the full model described in Subsection with N = 6, r = 2, p = 2, q = 1, s = 2, m = 12, G = 1 T/m I m, and H = I This is a DCF model of order (2,2,1). The values of ω i, i = 1, 2, 3, and Ψ used in the simulation are given in Tables 1 and 2, and the sample sizes employed are T = 480, 960 and 1,920. The estimation results of 10,000 repetitions are summarized also in Tables 1 and 2. From the tables, the maximum likelihood estimation works well in general and, as expected, the biases and the standard errors of the MLEs become smaller when the sample size increases. The estimates of ω 2 are relatively uncertain and their standard errors improve slowly with the increase in sample size. 12

13 Table 1: Simulation results of Ψ and ω 1 estimation via maximum likelihood method for the doubly constrained factor model, where the model is of order (2,2,1). Least squares estimates are used as initial values. (a) Ψ (True) Ψ ii Estimate when T=480 ˆΨ ii (St.E.) Estimate when T=960 ˆΨ ii (St.E.) Estimate when T=1920 ˆΨ ii (St.E.) (b) ω 1 (True) ω 1[, 1] ω 1[, 2] Estimate when T=480 ω 1[, 1] (St.E.) ω 1[, 2] (St.E.) Estimate when T=960 ω 1[, 1] (St.E.) ω 1[, 2] (St.E.) Estimate when T=1920 ω 1[, 1] (St.E.) ω 1[, 2] (St.E.) The numbers shown are the sample means and standard deviations of the estimates from 10,000 repetitions. the diagonal elements of Ψ, i = 1,..., 6. 13

14 Table 2: Simulation results of ω 2 and ω 3 estimation via the maximum likelihood method for the doubly constrained factor model, where the model is of order (2,2,1). Least squares estimates are used as initial values. (c) ω 2 (True) ω 2[, 1] ω 2[, 2] Estimate when T=480 ω 2[, 1] (St.E.) ω 2[, 2] (St.E.) Estimate when T=960 ω 2[, 1] (St.E.) ω 2[, 2] (St.E.) Estimate when T=1920 ω 2[, 1] (St.E.) ω 2[, 2] (St.E.) (d) ω 3 (True) ω 3[, 1] Estimate when T=480 ω 3[, 1] (St.E.) Estimate when T=960 ω 3[, 1] (St.E.) Estimate when T=1920 ω 3[, 1] (St.E.) The numbers shown are the sample means and standard deviations of the estimates from 10,000 repetitions. 14

15 3.2 Performance of AIC As mentioned in Subsection 2.4, the testing procedure of Tsai and Tsay (2010) becomes complicated for doubly constrained factor models because it would involve non-nested hypothesis testing. Therefore, AIC is used for checking the adequacy of the column and/or row constraints. In this subsection, we consider the finite sample performance of the AIC in selecting the data generating model among Cases 1-4 below. The data generating models considered are Case 1: ω 2 = ω 3 = 0, and ω 1 is the same as that of Subsection 3.1, Case 2: ω 1 = ω 3 = 0, and ω 2 is the same as the one used in Subsection 3.1, Case 3: the Full Model studied in Subsection 3.1, Case 4: ω 3 = 0, and ω 1 and ω 2 are the same as those studied in Subsection 3.1. The matrices H and G, the values of Ψ, N, r, p, q, m, and s are also the same as those of Subsection 3.1. For singly constrained factor models (Cases 1 and 2), we implement the estimation procedures as described in Subsections and 2.2.2, respectively. For doubly constrained factor models (Cases 3 and 4), we estimate all parameters via the objective functions (9) and (10), using the optimizing subroutine DNCONF from FORTRAN s IMSL library. The least squares estimates discussed in Subsection 2.1 are used as the initial values. The percentages of correct model specification based on 1,000 repetitions are reported in Table 3. The results show that the AIC works well in selecting a proper doubly constrained factor model. 4 Application To demonstrate the application of the proposed doubly constrained factor model, we consider the total housing starts of the United States, obtained from the U.S. Census Bureau website. The data period is from January 1997 to December 2006, so that we have 120 monthly data for the nine main divisions of the U.S. shown in Figure 1. The LOESS regression is applied 15

16 Table 3: The percentages of correct model specifications of doubly constrained factor model via the AIC criterion. The results are based on 1,000 repetitions. True model Fitted model Case 1 Case 2 Case 3 Case 4 T=480 Case % 0.0% 0.0% 0.0% Case 2 0.0% 98.4% 0.0% 0.0% Case 3 0.0% 0.0% 84.5% 17.6% Case 4 0.2% 1.6% 15.5% 82.4% T=960 Case % 0.0% 0.0% 0.0% Case 2 0.0% 98.8% 0.0% 0.0% Case 3 0.0% 0.0% 98.0% 17.7% Case 4 0.0% 1.2% 2.0% 82.3% T=1,920 Case % 0.0% 0.0% 0.0% Case 2 0.0% 99.1% 0.0% 0.0% Case 3 0.0% 0.1% 99.8% 19.2% Case 4 0.1% 0.8% 0.2% 80.8% to the log transformed data before fitting the doubly constrained factor model. This step is taken to remove the trend of the series. To specify the constraint matrix H, prior experience or geographical clustering may be helpful. In this paper, we apply the hierarchical clustering to the variables to specify H. It turns out that the result is consistent with the geographical clustering. Therefore, we employ three groups for the variables (divisions) and they are as follows: Group 1: New England, Middle Atlantic, East North Central, West North Central ; Group 2: South Atlantic, East South Central, West South Central ; Group 3: Mountain, Pacific. The H matrix simply consists of the indicator variables for the 3 groups. From Figure 1, Group 1 consists of the Northeast and Midwest of the U.S., Group 2 denotes the South, whereas Group 3 is the West. The time plots of Figure 2 show that the housing starts exhibit strong seasonality of period 12. Therefore, we let G = 1 10 I 12. Consequently, for this particular instance, we have 16

17 Figure 1: The census regions and divisions of the United States m = 12, T = 120, N = 9, and s = 3. We consider the DCF models of order (r, p, q) with 0 r, p, q 3, and q min{r, p}. Therefore, a total of 30 models were entertained. Table 4 shows the ranking of the entertained DCF models based on the AIC criterion, where the model of order (0,0,0) means an unrestricted model. Based on the AIC criterion, the doubly constrained factor model of order (2,2,1) is selected with the model of order (2,2,2) as a close second. Model checking shows that the residuals of the fitted DCF model of order (2,2,1) have some minor serial correlations, but those of the model of order (2,2,2) are close to being white noises. Therefore, we adopt the DCF model of order (2,2,2). Figure 3 shows the time plots of the residuals, Ê, of the entertained DCF(2,2,2) model. The left panel consists of the residuals of least square estimation whereas the right panel those of the maximum likelihood estimates. The two sets of residuals show similar pattern, but also contain certain differences. However, their sample autocorrelation functions confirm that the residuals have no significant serial dependence; see Figure 4. Table 5 gives the maximum likelihood estimates of the ω i for the selected DCF model of order (2,2,2). The corresponding 17

18 New England West Noth Central West South Central Middle Atlantic East North Central South Atlantic East South Central moutain Pacific year year year Figure 2: Time plots of monthly housing starts (in logarithms) of nine U.S. divisions: LSE of ω i are given in Table 6. These estimates are different from those of MLE of Table 5 because different normalizations are used. Figure 5 shows the time plots of the fitted common factors. The upper three panels show the common factors obtained by the least squares method whereas the lower three panels give the corresponding results for the maximum likelihood estimation. Care must be exercised in comparing the fitted common factors because their scales and orderings are not identifiable. For instance, consider the fitted common factors F 3. The orderings seem to be interchanged between the two estimation methods. Overall, the common factors F 1 of the maximum likelihood estimation appear to have some seasonality. We shall return to this point in our discussion later. 4.1 Discussion To gain insight into the decomposition of the housing starts implied by the fitted DCF model of order (2,2,2), we consider in details the results of maximum likelihood estimation. Figures 6 to 8 show the time plots of the decompositions of the housing starts series. The plots in Figure 6 consist of G F 2 ω 2 of Equation(1). Since the row constraints used are monthly indicator variables, these plots signify the deterministic seasonal pattern of each housing starts series that is orthogonal to the geographical divisions. From the plots, the deterministic 18

19 Table 4: The rankings of AIC for the proposed constrained factor models. Model AIC ranks Model AIC ranks (r,p,q) (r,p,q) (0,0,0) (3,3,0) (0,1,0) (1,1,1) (0,2,0) (2,1,1) (0,3,0) (3,1,1) (1,0,0) (1,2,1) (2,0,0) (2,3,1) (3,0,0) (2,3,2) (1,1,0) (1,3,1) (2,1,0) (3,2,1) (3,1,0) (3,2,2) (1,2,0) (2,2,1) (2,3,0) (2,2,2) (1,3,0) (3,3,1) (3,2,0) (3,3,2) (2,2,0) (3,3,3) Table 5: Maximum likelihood estimates of the doubly constrained factor model of order (2,2,2) for the U.S. housing starts data from 1997 to (a) MLE of ω 1 ω 1[, 1] ω 1[, 2] (b) MLE of ω 2 ω 2[, 1] ω 2[, 2] (c) MLE of ω 3 ω 3[, 1] ω 3[, 2] seasonality varies from series to series, but those of the East North Central and West North Central are similar. This seems reasonable as these two divisions are the Midwest and share close weather characteristics. New England and Middle Atlantic divisions have their own deterministic seasonal patterns. Finally, the Mountain and West South Central also share similar deterministic seasonal pattern. The plots in Figure 7 consist of F 1 ω 1H of Equation (1), which denotes housing variations due to the geographical locations, but is orthogonal to the deterministic seasonality. The 19

20 Table 6: Least squares estimates of the doubly constrained factor model of order (2,2,2) for the U.S. housing data from 1997 to LSE of ω 1 ω 1[, 1] ω 1[, 2] LSE of ω 2 ω 2[, 1] ω 2[, 2] LSE of ω 3 ω 3[, 1] ω 3[, 2] column constraints essentially pool information within each group to obtain the geographical housing variations. The series in Figure 7 also contain certain seasonality and we believe that they describe the stochastic seasonality of the three geographical groups. These stochastic seasonalities differ from group to group. Figure 8 shows the interactions G F 3 ω 3H between geographical grouping and deterministic seasonality of Equation (1). The plots show marked differencs between the three interactions. For this particular example, the proposed DCF model is capable of describing the seasonal and geographical patterns of U.S. housing starts. The example also demonstrates that the row and column constraints can be used to gain insight into the common structure of a multivariate time series. 5 Concluding Remarks In this paper, we considered both the least squares and maximum likelihood estimations of a doubly constrained factor model, and demonstrated the proposed methods by analyzing nine U.S. monthly housing starts series. The decomposition of the housing starts series shows that the proposed model is capable of describing the characteristics of the data. Much work of the constrained factor models, however, remains open. For instance, the maximum likelihood estimation is obtained under the normality assumption. In real applications, such an assumption might not be valid and the innovations of Equation (1) may contain conditional heteroscedasticity. In addition, we only consider deterministic constraints in the paper. It is of interest to investigate the proposed analysis when the constraints are stochastic. 20

21 Appendix: Proof of Lemma 1 To prove part (a), write B = ω B ω B, where ω B = [ω 2 Hω 3 ], then we have Σ = I T A + (G ω B )(G ω B) (by the definition of Σ and Theorem 7.7 of Schott, 1997) = I T A I m(p+q) + (G ω B)(I T A 1 )(G ω B ) (by Theorem of Harville, 1997, and Theorem 7.9 (a) of Schott, 1997) = A T I m(p+q) + T m I m ω BA 1 ω B (by Equation (8), Theorems 7.7 and 7.11 of Schott, 1997) = A T I mn + T m I m A 1/2 BA 1/2 (by Theorem 7.7 of Schott, 1997, and Theorem of Harville, 1997) = A T Im A 1/2 2 (I m A 1/2 ) (I m I N + Tm ) I m A 1/2 BA 1/2 (I m A 1/2 ) ( = A T m I m A + I m T ) m B (by Theorems 7.7 and 7.11 of Schott, 1997) = A T m I m Q (by the definition of Q and Theorem 7.6 (e) of Schott, 1997) = Q m A T m (by Theorem 7.11 of Schott, 1997). This proves part (a). Now, we prove part (b). We will prove part (b) by showing that Σ Σ 1 = Σ 1 Σ = INT. First note that, by the definitions of U and Q, we have QUA = B, and so QU = BA 1. Therefore, Σ Σ 1 = (I T A + GG B)(I T A 1 + GG U) = (I T I N ) + (GG AU) + (GG BA 1 ) + = I NT + (GG QU) + (GG BA 1 ) = I NT. ( ) T m GG BU Similarly, it can be shown that Σ 1 Σ = INT. This proves (b). Part (c) follows from part (b) and Theorem 7.17 of Schott (1997). 21

22 References [1] Akaike, H. (1974). A new look at the statistical model identification. IEEE Transactions on Automatic Control 19, [2] Anderson, T. W. (2003). An introduction to multivariate statistical analysis. N.J.: Wiley- Interscience. [3] Harville, D. A. (1997). Matrix Algebra From a Statistician s Perspective. New York: Springer. [4] Jöreskog, K.G. (1975),. Factor analysis by least squares and maximum likelihood, in Statistical Methods for Digital Computers, eds. K. Enslein, A. Ralston, and H. S. Wilf, New York: John Wiley. [5] Lam, C. and Yao, Q.W. (2012). Factor modeling for high-dimensional time series: inference for the number of factor. The Annals of Statistics 40, [6] Lütkepohl, H. (2005). New Introduction to Multiple Time Series Analysis, Berlin: Springer-Verlag. [7] Peña, D. and Box, G. E. P. (1987). Identifying a simplifying structure in time series. Journal of the American Statistical Association 82, [8] Schott, James R. (1997). Matrix Analysis for Statistics. New York : Wiley. [9] Takane, Y. and Hunter, M. A. (2001). Constrained principal component analysis: a comprehensive theory. Applicable Algebra in Engineering, Communication and Computing 12, [10] Tiao, G. C., Tsay, R. S., and Wang, T. C. (1993). Usefulness of linear transformations in multiple time series analysis. Empirical Econometrics, 18, [11] Tsai, H. and Tsay, R.S. (2010). Constrained factor models. Journal of the American Statistical Association 105,

23 ts(e) ts(e) Pacific Pacific Mountain Mountain West South Central West South Central East South Central East South Central South Atlantic West North Central South Atlantic West North Central East North Central Middle Atlantic New England East North Central Middle Atlantic New England Index Index (a) residuals for LSE (b) residuals for MLE Figure 3: Time series plots for (a) the least squares residuals and (b) the maximum likelihood residuals of the DCF model order (r,p,q) = (2,2,2). 23

24 E_r2p2q2, New England E_r2p2q2, New England E_r2p2q2, Middle Atlantic E_r2p2q2, East North Central E_r2p2q2, West North Central E_r2p2q2, South Atlantic E_r2p2q2, East South Central E_r2p2q2, West South Central E_r2p2q2, Mountain E_r2p2q2, Pacific E_r2p2q2, Middle Lag Atlantic E_r2p2q2, East LagNorth Central E_r2p2q2, West LagNorth Central E_r2p2q2, South Lag Atlantic E_r2p2q2, East LagSouth Central E_r2p2q2, West LagSouth Central E_r2p2q2, LagMountain E_r2p2q2, Lag Pacific (a) for LSE (b) for MLE Figure 4: for the residuals of DCF model with order (r,p,q) = (2,2,2). Results of the least squares estimation and the maximum likelihood estimation are shown. 24

25 F1[,1] F1[,2] F2[,1] F2[,2] F3[,1] F3[,2] (a) F i for LSE F1[,1] F1[,2] F2[,1] F2[,2] F3[,1] F3[,2] (b) F i for MLE Figure 5: Time series plots of common factors for a DCF model of order (r,p,q) = (2,2,2) via least squares estimation and maximum likelihood estimation. 25

26 ts(gterm) Pacific Mountain West South Central East South Central South Atlantic West North Central East North Central Middle Atlantic New England Index Figure 6: Time series plots for G F 2 ω 2 of a fitted DCF model of order (2,2,2). Maximum likelihood estimation is used. 26

27 ts(hterm) Pacific Mountain West South Central East South Central South Atlantic West North Central East North Central Middle Atlantic New England Index Figure 7: Time series plots for F 1 ω 1H of a fitted DCF model of order (2,2,2). Maximum likelihood estimation is used. 27

28 ts(ghterm) Pacific Mountain West South Central East South Central South Atlantic West North Central East North Central Middle Atlantic New England Index Figure 8: Time series plots for G F 3 ω 3H of a fitted DCF model of order (2,2,2). Maximum likelihood estimation is used. 28

DOUBLY CONSTRAINED FACTOR MODELS WITH APPLICATIONS

DOUBLY CONSTRAINED FACTOR MODELS WITH APPLICATIONS Statistica Sinica 26 (2016), 1453-1478 doi:http://dx.doi.org/10.5705/ss.2013.332t DOUBLY CONSTRAINED FACTOR MODELS WITH APPLICATIONS Henghsiu Tsai, Ruey S. Tsay, Edward M. H. Lin and Ching-Wei Cheng Academia

More information

FACTOR ANALYSIS AND MULTIDIMENSIONAL SCALING

FACTOR ANALYSIS AND MULTIDIMENSIONAL SCALING FACTOR ANALYSIS AND MULTIDIMENSIONAL SCALING Vishwanath Mantha Department for Electrical and Computer Engineering Mississippi State University, Mississippi State, MS 39762 mantha@isip.msstate.edu ABSTRACT

More information

Estimation and Testing for Common Cycles

Estimation and Testing for Common Cycles Estimation and esting for Common Cycles Anders Warne February 27, 2008 Abstract: his note discusses estimation and testing for the presence of common cycles in cointegrated vector autoregressions A simple

More information

4.1 Order Specification

4.1 Order Specification THE UNIVERSITY OF CHICAGO Booth School of Business Business 41914, Spring Quarter 2009, Mr Ruey S Tsay Lecture 7: Structural Specification of VARMA Models continued 41 Order Specification Turn to data

More information

Analysis of the AIC Statistic for Optimal Detection of Small Changes in Dynamic Systems

Analysis of the AIC Statistic for Optimal Detection of Small Changes in Dynamic Systems Analysis of the AIC Statistic for Optimal Detection of Small Changes in Dynamic Systems Jeremy S. Conner and Dale E. Seborg Department of Chemical Engineering University of California, Santa Barbara, CA

More information

Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8]

Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8] 1 Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8] Insights: Price movements in one market can spread easily and instantly to another market [economic globalization and internet

More information

9.1 Orthogonal factor model.

9.1 Orthogonal factor model. 36 Chapter 9 Factor Analysis Factor analysis may be viewed as a refinement of the principal component analysis The objective is, like the PC analysis, to describe the relevant variables in study in terms

More information

Cointegration Lecture I: Introduction

Cointegration Lecture I: Introduction 1 Cointegration Lecture I: Introduction Julia Giese Nuffield College julia.giese@economics.ox.ac.uk Hilary Term 2008 2 Outline Introduction Estimation of unrestricted VAR Non-stationarity Deterministic

More information

Robust Testing and Variable Selection for High-Dimensional Time Series

Robust Testing and Variable Selection for High-Dimensional Time Series Robust Testing and Variable Selection for High-Dimensional Time Series Ruey S. Tsay Booth School of Business, University of Chicago May, 2017 Ruey S. Tsay HTS 1 / 36 Outline 1 Focus on high-dimensional

More information

THE UNIVERSITY OF CHICAGO Graduate School of Business Business 41912, Spring Quarter 2008, Mr. Ruey S. Tsay. Solutions to Final Exam

THE UNIVERSITY OF CHICAGO Graduate School of Business Business 41912, Spring Quarter 2008, Mr. Ruey S. Tsay. Solutions to Final Exam THE UNIVERSITY OF CHICAGO Graduate School of Business Business 41912, Spring Quarter 2008, Mr. Ruey S. Tsay Solutions to Final Exam 1. (13 pts) Consider the monthly log returns, in percentages, of five

More information

Cointegrated VAR s. Eduardo Rossi University of Pavia. November Rossi Cointegrated VAR s Financial Econometrics / 56

Cointegrated VAR s. Eduardo Rossi University of Pavia. November Rossi Cointegrated VAR s Financial Econometrics / 56 Cointegrated VAR s Eduardo Rossi University of Pavia November 2013 Rossi Cointegrated VAR s Financial Econometrics - 2013 1 / 56 VAR y t = (y 1t,..., y nt ) is (n 1) vector. y t VAR(p): Φ(L)y t = ɛ t The

More information

Statistics 910, #5 1. Regression Methods

Statistics 910, #5 1. Regression Methods Statistics 910, #5 1 Overview Regression Methods 1. Idea: effects of dependence 2. Examples of estimation (in R) 3. Review of regression 4. Comparisons and relative efficiencies Idea Decomposition Well-known

More information

Heteroskedasticity; Step Changes; VARMA models; Likelihood ratio test statistic; Cusum statistic.

Heteroskedasticity; Step Changes; VARMA models; Likelihood ratio test statistic; Cusum statistic. 47 3!,57 Statistics and Econometrics Series 5 Febrary 24 Departamento de Estadística y Econometría Universidad Carlos III de Madrid Calle Madrid, 126 2893 Getafe (Spain) Fax (34) 91 624-98-49 VARIANCE

More information

Elements of Multivariate Time Series Analysis

Elements of Multivariate Time Series Analysis Gregory C. Reinsel Elements of Multivariate Time Series Analysis Second Edition With 14 Figures Springer Contents Preface to the Second Edition Preface to the First Edition vii ix 1. Vector Time Series

More information

THE UNIVERSITY OF CHICAGO Booth School of Business Business 41912, Spring Quarter 2016, Mr. Ruey S. Tsay

THE UNIVERSITY OF CHICAGO Booth School of Business Business 41912, Spring Quarter 2016, Mr. Ruey S. Tsay THE UNIVERSITY OF CHICAGO Booth School of Business Business 41912, Spring Quarter 2016, Mr. Ruey S. Tsay Lecture 5: Multivariate Multiple Linear Regression The model is Y n m = Z n (r+1) β (r+1) m + ɛ

More information

Generalized Autoregressive Score Models

Generalized Autoregressive Score Models Generalized Autoregressive Score Models by: Drew Creal, Siem Jan Koopman, André Lucas To capture the dynamic behavior of univariate and multivariate time series processes, we can allow parameters to be

More information

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions International Journal of Control Vol. 00, No. 00, January 2007, 1 10 Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions I-JENG WANG and JAMES C.

More information

Consistent Bivariate Distribution

Consistent Bivariate Distribution A Characterization of the Normal Conditional Distributions MATSUNO 79 Therefore, the function ( ) = G( : a/(1 b2)) = N(0, a/(1 b2)) is a solu- tion for the integral equation (10). The constant times of

More information

Analysis of Big Dependent Data in Economics and Finance

Analysis of Big Dependent Data in Economics and Finance Analysis of Big Dependent Data in Economics and Finance Ruey S. Tsay Booth Shool of Business, University of Chicago September 2016 Ruey S. Tsay Big Dependent Data 1 / 72 Outline 1 Big data? Machine learning?

More information

CS281 Section 4: Factor Analysis and PCA

CS281 Section 4: Factor Analysis and PCA CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we

More information

Online Appendix. j=1. φ T (ω j ) vec (EI T (ω j ) f θ0 (ω j )). vec (EI T (ω) f θ0 (ω)) = O T β+1/2) = o(1), M 1. M T (s) exp ( isω)

Online Appendix. j=1. φ T (ω j ) vec (EI T (ω j ) f θ0 (ω j )). vec (EI T (ω) f θ0 (ω)) = O T β+1/2) = o(1), M 1. M T (s) exp ( isω) Online Appendix Proof of Lemma A.. he proof uses similar arguments as in Dunsmuir 979), but allowing for weak identification and selecting a subset of frequencies using W ω). It consists of two steps.

More information

Estimating Variances and Covariances in a Non-stationary Multivariate Time Series Using the K-matrix

Estimating Variances and Covariances in a Non-stationary Multivariate Time Series Using the K-matrix Estimating Variances and Covariances in a Non-stationary Multivariate ime Series Using the K-matrix Stephen P Smith, January 019 Abstract. A second order time series model is described, and generalized

More information

FACTOR ANALYSIS AS MATRIX DECOMPOSITION 1. INTRODUCTION

FACTOR ANALYSIS AS MATRIX DECOMPOSITION 1. INTRODUCTION FACTOR ANALYSIS AS MATRIX DECOMPOSITION JAN DE LEEUW ABSTRACT. Meet the abstract. This is the abstract. 1. INTRODUCTION Suppose we have n measurements on each of taking m variables. Collect these measurements

More information

Inference using structural equations with latent variables

Inference using structural equations with latent variables This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this

More information

TAMS39 Lecture 10 Principal Component Analysis Factor Analysis

TAMS39 Lecture 10 Principal Component Analysis Factor Analysis TAMS39 Lecture 10 Principal Component Analysis Factor Analysis Martin Singull Department of Mathematics Mathematical Statistics Linköping University, Sweden Content - Lecture Principal component analysis

More information

VAR Models and Applications

VAR Models and Applications VAR Models and Applications Laurent Ferrara 1 1 University of Paris West M2 EIPMC Oct. 2016 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions

More information

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Article Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Fei Jin 1,2 and Lung-fei Lee 3, * 1 School of Economics, Shanghai University of Finance and Economics,

More information

arxiv: v2 [stat.me] 21 Jun 2017

arxiv: v2 [stat.me] 21 Jun 2017 Factor Models for Matrix-Valued High-Dimensional Time Series Dong Wang arxiv:1610.01889v2 [stat.me] 21 Jun 2017 Department of Operations Research and Financial Engineering, Princeton University, Princeton,

More information

Geographically weighted regression approach for origin-destination flows

Geographically weighted regression approach for origin-destination flows Geographically weighted regression approach for origin-destination flows Kazuki Tamesue 1 and Morito Tsutsumi 2 1 Graduate School of Information and Engineering, University of Tsukuba 1-1-1 Tennodai, Tsukuba,

More information

Appendix A : rational of the spatial Principal Component Analysis

Appendix A : rational of the spatial Principal Component Analysis Appendix A : rational of the spatial Principal Component Analysis In this appendix, the following notations are used : X is the n-by-p table of centred allelic frequencies, where rows are observations

More information

Vector autoregressions, VAR

Vector autoregressions, VAR 1 / 45 Vector autoregressions, VAR Chapter 2 Financial Econometrics Michael Hauser WS17/18 2 / 45 Content Cross-correlations VAR model in standard/reduced form Properties of VAR(1), VAR(p) Structural VAR,

More information

1 Data Arrays and Decompositions

1 Data Arrays and Decompositions 1 Data Arrays and Decompositions 1.1 Variance Matrices and Eigenstructure Consider a p p positive definite and symmetric matrix V - a model parameter or a sample variance matrix. The eigenstructure is

More information

7. Forecasting with ARIMA models

7. Forecasting with ARIMA models 7. Forecasting with ARIMA models 309 Outline: Introduction The prediction equation of an ARIMA model Interpreting the predictions Variance of the predictions Forecast updating Measuring predictability

More information

New Introduction to Multiple Time Series Analysis

New Introduction to Multiple Time Series Analysis Helmut Lütkepohl New Introduction to Multiple Time Series Analysis With 49 Figures and 36 Tables Springer Contents 1 Introduction 1 1.1 Objectives of Analyzing Multiple Time Series 1 1.2 Some Basics 2

More information

The exact bias of S 2 in linear panel regressions with spatial autocorrelation SFB 823. Discussion Paper. Christoph Hanck, Walter Krämer

The exact bias of S 2 in linear panel regressions with spatial autocorrelation SFB 823. Discussion Paper. Christoph Hanck, Walter Krämer SFB 83 The exact bias of S in linear panel regressions with spatial autocorrelation Discussion Paper Christoph Hanck, Walter Krämer Nr. 8/00 The exact bias of S in linear panel regressions with spatial

More information

EE731 Lecture Notes: Matrix Computations for Signal Processing

EE731 Lecture Notes: Matrix Computations for Signal Processing EE731 Lecture Notes: Matrix Computations for Signal Processing James P. Reilly c Department of Electrical and Computer Engineering McMaster University September 22, 2005 0 Preface This collection of ten

More information

Degenerate Expectation-Maximization Algorithm for Local Dimension Reduction

Degenerate Expectation-Maximization Algorithm for Local Dimension Reduction Degenerate Expectation-Maximization Algorithm for Local Dimension Reduction Xiaodong Lin 1 and Yu Zhu 2 1 Statistical and Applied Mathematical Science Institute, RTP, NC, 27709 USA University of Cincinnati,

More information

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley Review of Classical Least Squares James L. Powell Department of Economics University of California, Berkeley The Classical Linear Model The object of least squares regression methods is to model and estimate

More information

Vector Auto-Regressive Models

Vector Auto-Regressive Models Vector Auto-Regressive Models Laurent Ferrara 1 1 University of Paris Nanterre M2 Oct. 2018 Overview of the presentation 1. Vector Auto-Regressions Definition Estimation Testing 2. Impulse responses functions

More information

Intro VEC and BEKK Example Factor Models Cond Var and Cor Application Ref 4. MGARCH

Intro VEC and BEKK Example Factor Models Cond Var and Cor Application Ref 4. MGARCH ntro VEC and BEKK Example Factor Models Cond Var and Cor Application Ref 4. MGARCH JEM 140: Quantitative Multivariate Finance ES, Charles University, Prague Summer 2018 JEM 140 () 4. MGARCH Summer 2018

More information

Title. Description. var intro Introduction to vector autoregressive models

Title. Description. var intro Introduction to vector autoregressive models Title var intro Introduction to vector autoregressive models Description Stata has a suite of commands for fitting, forecasting, interpreting, and performing inference on vector autoregressive (VAR) models

More information

Estimation of Unique Variances Using G-inverse Matrix in Factor Analysis

Estimation of Unique Variances Using G-inverse Matrix in Factor Analysis International Mathematical Forum, 3, 2008, no. 14, 671-676 Estimation of Unique Variances Using G-inverse Matrix in Factor Analysis Seval Süzülmüş Osmaniye Korkut Ata University Vocational High School

More information

Quick Review on Linear Multiple Regression

Quick Review on Linear Multiple Regression Quick Review on Linear Multiple Regression Mei-Yuan Chen Department of Finance National Chung Hsing University March 6, 2007 Introduction for Conditional Mean Modeling Suppose random variables Y, X 1,

More information

Lecture 7: Model Building Bus 41910, Time Series Analysis, Mr. R. Tsay

Lecture 7: Model Building Bus 41910, Time Series Analysis, Mr. R. Tsay Lecture 7: Model Building Bus 41910, Time Series Analysis, Mr R Tsay An effective procedure for building empirical time series models is the Box-Jenkins approach, which consists of three stages: model

More information

Section 9.2: Matrices.. a m1 a m2 a mn

Section 9.2: Matrices.. a m1 a m2 a mn Section 9.2: Matrices Definition: A matrix is a rectangular array of numbers: a 11 a 12 a 1n a 21 a 22 a 2n A =...... a m1 a m2 a mn In general, a ij denotes the (i, j) entry of A. That is, the entry in

More information

Economics 620, Lecture 5: exp

Economics 620, Lecture 5: exp 1 Economics 620, Lecture 5: The K-Variable Linear Model II Third assumption (Normality): y; q(x; 2 I N ) 1 ) p(y) = (2 2 ) exp (N=2) 1 2 2(y X)0 (y X) where N is the sample size. The log likelihood function

More information

Cointegrated VARIMA models: specification and. simulation

Cointegrated VARIMA models: specification and. simulation Cointegrated VARIMA models: specification and simulation José L. Gallego and Carlos Díaz Universidad de Cantabria. Abstract In this note we show how specify cointegrated vector autoregressive-moving average

More information

Variational Principal Components

Variational Principal Components Variational Principal Components Christopher M. Bishop Microsoft Research 7 J. J. Thomson Avenue, Cambridge, CB3 0FB, U.K. cmbishop@microsoft.com http://research.microsoft.com/ cmbishop In Proceedings

More information

1 Phelix spot and futures returns: descriptive statistics

1 Phelix spot and futures returns: descriptive statistics MULTIVARIATE VOLATILITY MODELING OF ELECTRICITY FUTURES: ONLINE APPENDIX Luc Bauwens 1, Christian Hafner 2, and Diane Pierret 3 October 13, 2011 1 Phelix spot and futures returns: descriptive statistics

More information

Appendix A: Matrices

Appendix A: Matrices Appendix A: Matrices A matrix is a rectangular array of numbers Such arrays have rows and columns The numbers of rows and columns are referred to as the dimensions of a matrix A matrix with, say, 5 rows

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1 MA 575 Linear Models: Cedric E Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1 1 Within-group Correlation Let us recall the simple two-level hierarchical

More information

TRAFFIC FLOW MODELING AND FORECASTING THROUGH VECTOR AUTOREGRESSIVE AND DYNAMIC SPACE TIME MODELS

TRAFFIC FLOW MODELING AND FORECASTING THROUGH VECTOR AUTOREGRESSIVE AND DYNAMIC SPACE TIME MODELS TRAFFIC FLOW MODELING AND FORECASTING THROUGH VECTOR AUTOREGRESSIVE AND DYNAMIC SPACE TIME MODELS Kamarianakis Ioannis*, Prastacos Poulicos Foundation for Research and Technology, Institute of Applied

More information

ARIMA Models. Jamie Monogan. January 25, University of Georgia. Jamie Monogan (UGA) ARIMA Models January 25, / 38

ARIMA Models. Jamie Monogan. January 25, University of Georgia. Jamie Monogan (UGA) ARIMA Models January 25, / 38 ARIMA Models Jamie Monogan University of Georgia January 25, 2012 Jamie Monogan (UGA) ARIMA Models January 25, 2012 1 / 38 Objectives By the end of this meeting, participants should be able to: Describe

More information

Labor-Supply Shifts and Economic Fluctuations. Technical Appendix

Labor-Supply Shifts and Economic Fluctuations. Technical Appendix Labor-Supply Shifts and Economic Fluctuations Technical Appendix Yongsung Chang Department of Economics University of Pennsylvania Frank Schorfheide Department of Economics University of Pennsylvania January

More information

A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED

A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED by W. Robert Reed Department of Economics and Finance University of Canterbury, New Zealand Email: bob.reed@canterbury.ac.nz

More information

Testing an Autoregressive Structure in Binary Time Series Models

Testing an Autoregressive Structure in Binary Time Series Models ömmföäflsäafaäsflassflassflas ffffffffffffffffffffffffffffffffffff Discussion Papers Testing an Autoregressive Structure in Binary Time Series Models Henri Nyberg University of Helsinki and HECER Discussion

More information

Multivariate Statistics

Multivariate Statistics Multivariate Statistics Chapter 4: Factor analysis Pedro Galeano Departamento de Estadística Universidad Carlos III de Madrid pedro.galeano@uc3m.es Course 2017/2018 Master in Mathematical Engineering Pedro

More information

R = µ + Bf Arbitrage Pricing Model, APM

R = µ + Bf Arbitrage Pricing Model, APM 4.2 Arbitrage Pricing Model, APM Empirical evidence indicates that the CAPM beta does not completely explain the cross section of expected asset returns. This suggests that additional factors may be required.

More information

Lecture Stat Information Criterion

Lecture Stat Information Criterion Lecture Stat 461-561 Information Criterion Arnaud Doucet February 2008 Arnaud Doucet () February 2008 1 / 34 Review of Maximum Likelihood Approach We have data X i i.i.d. g (x). We model the distribution

More information

DATABASE AND METHODOLOGY

DATABASE AND METHODOLOGY CHAPTER 3 DATABASE AND METHODOLOGY In the present chapter, sources of database used and methodology applied for the empirical analysis has been presented The whole chapter has been divided into three sections

More information

IS THE NORTH ATLANTIC OSCILLATION A RANDOM WALK? A COMMENT WITH FURTHER RESULTS

IS THE NORTH ATLANTIC OSCILLATION A RANDOM WALK? A COMMENT WITH FURTHER RESULTS INTERNATIONAL JOURNAL OF CLIMATOLOGY Int. J. Climatol. 24: 377 383 (24) Published online 11 February 24 in Wiley InterScience (www.interscience.wiley.com). DOI: 1.12/joc.13 IS THE NORTH ATLANTIC OSCILLATION

More information

Inference in VARs with Conditional Heteroskedasticity of Unknown Form

Inference in VARs with Conditional Heteroskedasticity of Unknown Form Inference in VARs with Conditional Heteroskedasticity of Unknown Form Ralf Brüggemann a Carsten Jentsch b Carsten Trenkler c University of Konstanz University of Mannheim University of Mannheim IAB Nuremberg

More information

On Expected Gaussian Random Determinants

On Expected Gaussian Random Determinants On Expected Gaussian Random Determinants Moo K. Chung 1 Department of Statistics University of Wisconsin-Madison 1210 West Dayton St. Madison, WI 53706 Abstract The expectation of random determinants whose

More information

An Akaike Criterion based on Kullback Symmetric Divergence in the Presence of Incomplete-Data

An Akaike Criterion based on Kullback Symmetric Divergence in the Presence of Incomplete-Data An Akaike Criterion based on Kullback Symmetric Divergence Bezza Hafidi a and Abdallah Mkhadri a a University Cadi-Ayyad, Faculty of sciences Semlalia, Department of Mathematics, PB.2390 Marrakech, Moroco

More information

Thomas J. Fisher. Research Statement. Preliminary Results

Thomas J. Fisher. Research Statement. Preliminary Results Thomas J. Fisher Research Statement Preliminary Results Many applications of modern statistics involve a large number of measurements and can be considered in a linear algebra framework. In many of these

More information

A Non-Parametric Approach of Heteroskedasticity Robust Estimation of Vector-Autoregressive (VAR) Models

A Non-Parametric Approach of Heteroskedasticity Robust Estimation of Vector-Autoregressive (VAR) Models Journal of Finance and Investment Analysis, vol.1, no.1, 2012, 55-67 ISSN: 2241-0988 (print version), 2241-0996 (online) International Scientific Press, 2012 A Non-Parametric Approach of Heteroskedasticity

More information

AR, MA and ARMA models

AR, MA and ARMA models AR, MA and AR by Hedibert Lopes P Based on Tsay s Analysis of Financial Time Series (3rd edition) P 1 Stationarity 2 3 4 5 6 7 P 8 9 10 11 Outline P Linear Time Series Analysis and Its Applications For

More information

Alfredo A. Romero * College of William and Mary

Alfredo A. Romero * College of William and Mary A Note on the Use of in Model Selection Alfredo A. Romero * College of William and Mary College of William and Mary Department of Economics Working Paper Number 6 October 007 * Alfredo A. Romero is a Visiting

More information

Topic 1: Matrix diagonalization

Topic 1: Matrix diagonalization Topic : Matrix diagonalization Review of Matrices and Determinants Definition A matrix is a rectangular array of real numbers a a a m a A = a a m a n a n a nm The matrix is said to be of order n m if it

More information

Math Linear Algebra Final Exam Review Sheet

Math Linear Algebra Final Exam Review Sheet Math 15-1 Linear Algebra Final Exam Review Sheet Vector Operations Vector addition is a component-wise operation. Two vectors v and w may be added together as long as they contain the same number n of

More information

Time Series Modeling of Financial Data. Prof. Daniel P. Palomar

Time Series Modeling of Financial Data. Prof. Daniel P. Palomar Time Series Modeling of Financial Data Prof. Daniel P. Palomar The Hong Kong University of Science and Technology (HKUST) MAFS6010R- Portfolio Optimization with R MSc in Financial Mathematics Fall 2018-19,

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2 MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS SYSTEMS OF EQUATIONS AND MATRICES Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a

More information

Principal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis. Chris Funk. Lecture 17

Principal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis. Chris Funk. Lecture 17 Principal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis Chris Funk Lecture 17 Outline Filters and Rotations Generating co-varying random fields Translating co-varying fields into

More information

Moment Aberration Projection for Nonregular Fractional Factorial Designs

Moment Aberration Projection for Nonregular Fractional Factorial Designs Moment Aberration Projection for Nonregular Fractional Factorial Designs Hongquan Xu Department of Statistics University of California Los Angeles, CA 90095-1554 (hqxu@stat.ucla.edu) Lih-Yuan Deng Department

More information

Lecture 2: Univariate Time Series

Lecture 2: Univariate Time Series Lecture 2: Univariate Time Series Analysis: Conditional and Unconditional Densities, Stationarity, ARMA Processes Prof. Massimo Guidolin 20192 Financial Econometrics Spring/Winter 2017 Overview Motivation:

More information

Booth School of Business, University of Chicago Business 41914, Spring Quarter 2017, Mr. Ruey S. Tsay. Solutions to Midterm

Booth School of Business, University of Chicago Business 41914, Spring Quarter 2017, Mr. Ruey S. Tsay. Solutions to Midterm Booth School of Business, University of Chicago Business 41914, Spring Quarter 017, Mr Ruey S Tsay Solutions to Midterm Problem A: (51 points; 3 points per question) Answer briefly the following questions

More information

Principal Component Analysis (PCA) Our starting point consists of T observations from N variables, which will be arranged in an T N matrix R,

Principal Component Analysis (PCA) Our starting point consists of T observations from N variables, which will be arranged in an T N matrix R, Principal Component Analysis (PCA) PCA is a widely used statistical tool for dimension reduction. The objective of PCA is to find common factors, the so called principal components, in form of linear combinations

More information

Akaike criterion: Kullback-Leibler discrepancy

Akaike criterion: Kullback-Leibler discrepancy Model choice. Akaike s criterion Akaike criterion: Kullback-Leibler discrepancy Given a family of probability densities {f ( ; ψ), ψ Ψ}, Kullback-Leibler s index of f ( ; ψ) relative to f ( ; θ) is (ψ

More information

Kalman-Filter-Based Time-Varying Parameter Estimation via Retrospective Optimization of the Process Noise Covariance

Kalman-Filter-Based Time-Varying Parameter Estimation via Retrospective Optimization of the Process Noise Covariance 2016 American Control Conference (ACC) Boston Marriott Copley Place July 6-8, 2016. Boston, MA, USA Kalman-Filter-Based Time-Varying Parameter Estimation via Retrospective Optimization of the Process Noise

More information

Empirical Approach to Modelling and Forecasting Inflation in Ghana

Empirical Approach to Modelling and Forecasting Inflation in Ghana Current Research Journal of Economic Theory 4(3): 83-87, 2012 ISSN: 2042-485X Maxwell Scientific Organization, 2012 Submitted: April 13, 2012 Accepted: May 06, 2012 Published: June 30, 2012 Empirical Approach

More information

Minimax design criterion for fractional factorial designs

Minimax design criterion for fractional factorial designs Ann Inst Stat Math 205 67:673 685 DOI 0.007/s0463-04-0470-0 Minimax design criterion for fractional factorial designs Yue Yin Julie Zhou Received: 2 November 203 / Revised: 5 March 204 / Published online:

More information

FORECASTING. Methods and Applications. Third Edition. Spyros Makridakis. European Institute of Business Administration (INSEAD) Steven C Wheelwright

FORECASTING. Methods and Applications. Third Edition. Spyros Makridakis. European Institute of Business Administration (INSEAD) Steven C Wheelwright FORECASTING Methods and Applications Third Edition Spyros Makridakis European Institute of Business Administration (INSEAD) Steven C Wheelwright Harvard University, Graduate School of Business Administration

More information

Panel Threshold Regression Models with Endogenous Threshold Variables

Panel Threshold Regression Models with Endogenous Threshold Variables Panel Threshold Regression Models with Endogenous Threshold Variables Chien-Ho Wang National Taipei University Eric S. Lin National Tsing Hua University This Version: June 29, 2010 Abstract This paper

More information

Arma-Arch Modeling Of The Returns Of First Bank Of Nigeria

Arma-Arch Modeling Of The Returns Of First Bank Of Nigeria Arma-Arch Modeling Of The Returns Of First Bank Of Nigeria Emmanuel Alphonsus Akpan Imoh Udo Moffat Department of Mathematics and Statistics University of Uyo, Nigeria Ntiedo Bassey Ekpo Department of

More information

Regression. Oscar García

Regression. Oscar García Regression Oscar García Regression methods are fundamental in Forest Mensuration For a more concise and general presentation, we shall first review some matrix concepts 1 Matrices An order n m matrix is

More information

STAT 730 Chapter 9: Factor analysis

STAT 730 Chapter 9: Factor analysis STAT 730 Chapter 9: Factor analysis Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Data Analysis 1 / 15 Basic idea Factor analysis attempts to explain the

More information

FINANCIAL ECONOMETRICS AND EMPIRICAL FINANCE -MODULE2 Midterm Exam Solutions - March 2015

FINANCIAL ECONOMETRICS AND EMPIRICAL FINANCE -MODULE2 Midterm Exam Solutions - March 2015 FINANCIAL ECONOMETRICS AND EMPIRICAL FINANCE -MODULE2 Midterm Exam Solutions - March 205 Time Allowed: 60 minutes Family Name (Surname) First Name Student Number (Matr.) Please answer all questions by

More information

Factor Analysis Edpsy/Soc 584 & Psych 594

Factor Analysis Edpsy/Soc 584 & Psych 594 Factor Analysis Edpsy/Soc 584 & Psych 594 Carolyn J. Anderson University of Illinois, Urbana-Champaign April 29, 2009 1 / 52 Rotation Assessing Fit to Data (one common factor model) common factors Assessment

More information

Estimating AR/MA models

Estimating AR/MA models September 17, 2009 Goals The likelihood estimation of AR/MA models AR(1) MA(1) Inference Model specification for a given dataset Why MLE? Traditional linear statistics is one methodology of estimating

More information

Chapter 2:Determinants. Section 2.1: Determinants by cofactor expansion

Chapter 2:Determinants. Section 2.1: Determinants by cofactor expansion Chapter 2:Determinants Section 2.1: Determinants by cofactor expansion [ ] a b Recall: The 2 2 matrix is invertible if ad bc 0. The c d ([ ]) a b function f = ad bc is called the determinant and it associates

More information

Ross (1976) introduced the Arbitrage Pricing Theory (APT) as an alternative to the CAPM.

Ross (1976) introduced the Arbitrage Pricing Theory (APT) as an alternative to the CAPM. 4.2 Arbitrage Pricing Model, APM Empirical evidence indicates that the CAPM beta does not completely explain the cross section of expected asset returns. This suggests that additional factors may be required.

More information

On the equivalence of confidence interval estimation based on frequentist model averaging and least-squares of the full model in linear regression

On the equivalence of confidence interval estimation based on frequentist model averaging and least-squares of the full model in linear regression Working Paper 2016:1 Department of Statistics On the equivalence of confidence interval estimation based on frequentist model averaging and least-squares of the full model in linear regression Sebastian

More information

Switching Regime Estimation

Switching Regime Estimation Switching Regime Estimation Series de Tiempo BIrkbeck March 2013 Martin Sola (FE) Markov Switching models 01/13 1 / 52 The economy (the time series) often behaves very different in periods such as booms

More information

Multivariate Time Series: VAR(p) Processes and Models

Multivariate Time Series: VAR(p) Processes and Models Multivariate Time Series: VAR(p) Processes and Models A VAR(p) model, for p > 0 is X t = φ 0 + Φ 1 X t 1 + + Φ p X t p + A t, where X t, φ 0, and X t i are k-vectors, Φ 1,..., Φ p are k k matrices, with

More information

Elementary Linear Algebra

Elementary Linear Algebra Matrices J MUSCAT Elementary Linear Algebra Matrices Definition Dr J Muscat 2002 A matrix is a rectangular array of numbers, arranged in rows and columns a a 2 a 3 a n a 2 a 22 a 23 a 2n A = a m a mn We

More information

Multilevel Modeling Day 2 Intermediate and Advanced Issues: Multilevel Models as Mixed Models. Jian Wang September 18, 2012

Multilevel Modeling Day 2 Intermediate and Advanced Issues: Multilevel Models as Mixed Models. Jian Wang September 18, 2012 Multilevel Modeling Day 2 Intermediate and Advanced Issues: Multilevel Models as Mixed Models Jian Wang September 18, 2012 What are mixed models The simplest multilevel models are in fact mixed models:

More information

Illustration of the Varying Coefficient Model for Analyses the Tree Growth from the Age and Space Perspectives

Illustration of the Varying Coefficient Model for Analyses the Tree Growth from the Age and Space Perspectives TR-No. 14-06, Hiroshima Statistical Research Group, 1 11 Illustration of the Varying Coefficient Model for Analyses the Tree Growth from the Age and Space Perspectives Mariko Yamamura 1, Keisuke Fukui

More information

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA Principle Components Analysis: Uses one group of variables (we will call this X) In

More information

1. Introduction. Hang Qian 1 Iowa State University

1. Introduction. Hang Qian 1 Iowa State University Users Guide to the VARDAS Package Hang Qian 1 Iowa State University 1. Introduction The Vector Autoregression (VAR) model is widely used in macroeconomics. However, macroeconomic data are not always observed

More information

Chapter 12 REML and ML Estimation

Chapter 12 REML and ML Estimation Chapter 12 REML and ML Estimation C. R. Henderson 1984 - Guelph 1 Iterative MIVQUE The restricted maximum likelihood estimator (REML) of Patterson and Thompson (1971) can be obtained by iterating on MIVQUE,

More information