Corrigendum to Inference on impulse. response functions in structural VAR models. [J. Econometrics 177 (2013), 1-13]
|
|
- Mae Morrison
- 5 years ago
- Views:
Transcription
1 Corrigendum to Inference on impulse response functions in structural VAR models [J. Econometrics 177 (2013), 1-13] Atsushi Inoue a Lutz Kilian b a Department of Economics, Vanderbilt University, Nashville TN b University of Michigan, Department of Economics, Ann Arbor, MI JEL classification: C32; C52; E37 Keywords: Vector autoregression; Simultaneous inference; Impulse responses; Sign restrictions; Median; Mode; Credible set Corresponding author: Lutz Kilian, University of Michigan, Department of Economics, 309 Lorch Hall, 611 Tappan Street, Ann Arbor, MI Tel: Fax: lkilian@umich.edu. 1
2 Proposition 1 in Inoue and Kilian (2013) for the posterior density of the set of structural impulse responses in the fully sign-identified vector autoregressive (VAR) model is not correct as stated. The correct statement and proof is provided below. The correction does not affect the substance of the empirical findings based on this proposition. The statement and proof of Proposition 2 in the original article and the empirical results based on that proposition are not affected. Consider an n-dimensional VAR(p) model with an intercept. Ignoring the intercept for notational convenience, let B = [B 1 B p ] denote the slope parameters of the VAR model where B i is an n n matrix for i = 1, 2,..., n. Let Σ denote the error covariance matrix. Let A be the lower-triangular Cholesky decomposition of Σ such that AA = Σ, and let vech(a) denote the n(n + 1)/2 1 vector that consists of the on-diagonal elements and the below-diagonal elements of A. Let n n rotation matrix U be an element of O(n) (i.e., U is orthogonal and satisfies U U = I n ) and be Haar-distributed over O(n). By construction, we have U = 1 with probability with probability 1 2 Define Ũ = U if U = 1 W U if U = 1 2
3 and à = A if U = 1 AW if U = 1 where W is the n n diagonal matrix whose diagonal elements whose first diagonal element is -1 and other diagonal elements are 1. Then Ũ belongs to SO(n) (Ũ Ũ = 1 and Ũ = 1) and we have ÃŨ = AU. Note that U is Haardistributed when U = 1 where U belongs to O(n). Note also that W U is Haar-distributed when U = 1. Thus, Ũ is also Haar-distributed on SO(n) even unconditionally. Moreover, conditional on U = 1 or U = 1, Ũ has a one-to-one relationship with U and the Jacobian of the transformation from U to Ũ is always one. Because Ũ belongs to SO(n), S = I n 2(I n + Ũ) 1 (1) is an n n skew-symmetric matrix (León et al., 2006, p. 414). Let s denote the n(n 1)/2 1 vector that consists of the below-diagonal elements of S. the density of s is given by f(s) = ( ) Π n Γ(i/2) 2 (n 1)(n 2)/2 i=2 (2) π i/2 I n + S n 1 (see equation 4 of León et al. (2006), p. 415). 1 1 León et al. (2006, p. 416) define s as the vector that consists of the above-diagonal elements of S. Let s leon denote this vector. Because S is skew-symmetric, there is a n(n 1)/2 n(n 1)/2 permutation matrix P such that s leon = P s. Because the 3
4 Let Φ = [Φ 1 Φ 2 Φ p] where Φ i is the ith reduced-form vector moving average coefficient matrix. There is a one-to-one mapping between the first p + 1 structural impulse responses Θ = [A, Φ 1 A, Φ 2 A,, Φ p A], where A = ÃŨ, on the one hand, and the tuple formed by the reduced-form VAR parameters and s, (B, vech(a), s), on the other. The nonlinear function Θ = h(b, vech(ã), s) is known. Using the change-of-variables method, the posterior density f of Θ can be written as f( Θ) = = [vec(b) vech(ã) s ] vec( Θ) f(b, vech(ã), s) vec( Θ) 1 vech(σ) [vec(b) vech(ã) s ] f(b, Σ, s) vech(ã) vec( Θ) 1 vech(σ) [vec(b) vech(ã) s ] f(b Σ)f(Σ)f(s), (3) vech(ã) where B, Σ = ÃÃ, and Ũ are the unique values that satisfy the nonlinear function Θ = h(b, vech(ã), s). The conditioning on the data is omitted for notational simplicity. Let D n denote the n 2 n(n + 1)/2 duplication matrix of zeros and ones such that vec(m) = D n vech(m) for any n n symmetric matrix M (see Definition 4.1 of Magnus, 1988, p. 55). D + n denotes the Moore-Penrose inverse of D n, i.e., D + n = (D nd n ) 1 D n, so that we can write vech(m) = D + n vec(m) (see Theorem 4.1 of Magnus, 1988, p. 56). Similarly, D n is the n 2 n(n 1)/2 absolute value of the determinant of P is always 1, the density remains the same as in (2) even when s is defined to be the vector obtained from the below-diagonal elements of S. 4
5 matrix such that vec(s) = D n s (see Definition 6.1 in Magnus, 1988, p. 94) and D n + = ( D n D n ) 1 D n. Let E ij = e i e j denote a square matrix with the (i, j)th element equal to one and zeros elsewhere. Let L n denote the n(n + 1)/2 n 2 elimination matrix of zeros and ones such that vec(m) = L nvech(m) for any lower triangular matrix M (see Definition 5.1 of Magnus, 1988, p. 76). K n denotes the n 2 n 2 communication matrix such that vec(m ) = K n vec(m) for any n n matrix M (see Magnus and Neudecker, 1999, pp ). Proposition 1. The posterior density of Θ is f( Θ) = 2 n(n+1) 2 Ã (np 1) I n + Ũ (n 1) f(b Σ)f(Σ)f(s). (4) Proof of Proposition 1: First, we will show that f( Θ) = = vec( Θ) 1 vech(σ) [vec(b) vech(ã) s ] f(b Σ)f(Σ)f(s) ( vech(ã) ( Ũ I n )L n (I n Ã)J U Ã np ) 1 D n + [(Ã I n) + (I n Ã)K n]l n f(b Σ)f(Σ)f(s), (5) where J U = 2[(I n S ) 1 (I n S) 1 ] D n. Because Ũ = 2(I n S) 1 I n, dũ = 2(I n S) 1 (ds)(i n S) 1, (6) 5
6 and thus vec(dũ) = 2[(I n S) 1 (I n S) 1 ]vec(ds) = 2[(I n S) 1 (I n S) 1 ] D n ds = J U ds. (7) Because d(ãũ) = (dã)ũ + Ã(dŨ) = (dã)ũ + 2Ã(I n S) 1 (ds)(i n S) 1,(8) d(φãũ) = (dφ)ãũ + Φ(dÃ)Ũ + ΦÃ(dŨ) = (dφ)ãũ + Φ(dÃ)Ũ + 2ΦÃ(I n S) 1 (ds)(i n S) 1, (9) it follows that dvec(ãũ) = (Ũ I n )dvec(ã) + 2(I n Ã)((I n S ) 1 (I n S) 1 )dvec(s), = (Ũ I n )L ndvech(ã) + 2(I n Ã)((I n S ) 1 (I n S) 1 ) D n ds, = (Ũ I n )L ndvech(ã) + (I n Ã)J Uds, (10) dvec(φãũ) = (Ũ Ã I np )dvec(φ) + (Ũ Φ)dvec(Ã) +2(I n ΦÃ)((I n S ) 1 (I n S) 1 )dvec(s) = (Ũ Ã I np )dvec(φ) + (Ũ Φ)L ndvech(ã) + (I n ΦÃ)J Uds. (11) 6
7 It follows from expressions (10) and (11) that J 1 vec( Θ) [vech(ã) s vec(φ) ] = (Ũ I n )L n (I n Ã)J U O n 2 n 2 p (Ũ Φ)L n (I n ΦÃ)J U Ũ Ã I np (12). Since the upper-left submatrix, [(Ũ I n )L n (I n Ã)J U], is n 2 n 2, the upper-right submatrix is the n 2 n 2 p submatrix of zeros, and the lower-right submatrix, Ũ Ã I np, is n 2 p n 2 p, J 1 is block lower triangular. Thus its determinant is given by the product of the determinants of its blocks: J 1 = = (Ũ I n )L n (Ũ I n )L n (I n Ã)J U Ũ Ã I np (I n Ã)J U Ã np. (13) Because of the recursive relationship defined by equations (11), (14) and (17) in Inoue and Kilian (2013), the Jacobian matrix of Φ with respect to B is block-diagonal and each diagonal block has a unit determinant. Thus J 2 vec(φ) vec(b) = 1. (14) Since the Jacobian of vec(σ) with respect to vec(ã) is [(Ã I n) + (I n Ã)K n], (15) the determinant of the Jacobian of vech(σ) with respect to vech(ã) is given 7
8 by J 3 D n + [(Ã I n) + (I n Ã)K n]l n (16) Thus, expression (5) follows from expressions (13), (14) and (16). Next, we will show that the absolute values of two of the determinants in expression (5) can be written as: ( ) n(n 1) (Ũ I n )L n (I n Ã)J 1 2 U = I n + 2 Ũ n 1 Π n 1 i=1 ãn i ii,(17) D n + [(Ã I n) + (I n Ã)K n]l n = 2 n Π n i=1ã n i+1 ii, (18) where ã ii is the (i, i)th element of Ã. Express the left-hand side of equation (17) as [(Ũ I n )L n (I n Ã)J U] = (Ũ I n )Z, (19) where Z = [L n 2(F G) D n ], (20) F = Ũ(I n S ) 1 = 1 2 (I n + Ũ) = (I n S) 1, (21) G = Ã(I n S) 1 = ÃF. (22) where F and G are well-defined because I n +Ũ is nonsingular with probability 8
9 one. Then the determinant of (Ũ I n )Z equals that of Z. Because Z Z = I n(n+1) 2 2L n (F G) D n 2 D n(f G )L n 4 D n(f F G G) D n, (23) we have Z 2 = Z Z = 2 n(n 1) D n(f G )(I n 2 L nl n )(F G) D n (24) Because L nl n is a diagonal matrix with zeros and ones on the diagonal, so is I n 2 L nl n. Thus I n 2 L nl n = i>j E ii E jj, (25) where the summation is over i, j = 1, 2,...n such that i > j. There exists an n(n 1)/2 n matrix, say Q, that contains only zeros and ones and has full row rank such that QQ = I n(n 1), Q Q = I n 2 L nl n = (E ii E jj ). (26) 2 i>j Thus, using Q, we can write D n(f G )(I n 2 L nl n )(F G) D n = D n(f G )Q Q(F G) D n 9
10 = D n(f F )(I n à )Q Q(I n Ã)(F F ) D n = ( D + n (F F ) D n ) ( D n(i n à )Q )(Q(I n Ã) D n )( D + n (F F ) D n ), (27) where Theorem 6.11(i) of Magnus (1988, p. 100) is used in deriving the last equality. Let ũ ij denote the vector of the below-diagonal elements of E ij where i > j. Then Q = i<j ũ ji vec(e ij ) = i>j ũ ij vec(e ji ), (28) and D n = i>j vec(e ij E ji )ũ ij, (29) where the latter equality follows from Theorem 6.1 of Magnus (1988, p. 95). Hence Q(I n Ã) D n = ũ ij vec(e ji ) (I n Ã)vec(E st E ts )ũ st i>j s>t = ũ ij (e i e j)(i n Ã)(e t e s e s e t )ũ st i>j s>t = ũ ij (e ie t )(e jãe s)ũ st ũ ij (e ie s )(e jãe t)ũ st i>j s>t i>j s>t = ã js ũ ij ũ st ã jt ũ ij ũ it = ã jt ũ ij ũ it s>i>j i>j i>t i>j t = ã jj ũ ij ũ ij ã jt ũ ij ũ it. (30) i>j i>j>t 10
11 Element i + f(j) of ũ ij is unity, where f(j) = (j 1)n j(j + 1), (31) 2 and all other elements are zero. Note that f(j + 1) = f(j) j + n 1 such that f(j + 1) > f(j) if j < n 1 f(j + 1) = f(j) if j = n 1 f(j + 1) < f(j) if j = n. (32) This implies that the matrix ũ ij ũ it is lower triangular for i > j > t. Moreover, ũ ij ũ ij is diagonal for i > j. Hence the matrix Q(I n Ã) D n is lower triangular and thus Q(I n Ã) D n = ( 1) n(n 1) 2 ã jj ũ ij ũ ij = ( 1) n(n 1) 2 Π n 1 j=1 ãn j jj (33) i>j It follows from (27) and (33) that Z 2 = 2 n(n 1) D n(f G )(I n L nl n )(F G) D n = 2 n(n 1) D + n (F F ) D n 2 Q(I n Ã) D n 2 = 2 n(n 1) F 2(n 1) Q(I n Ã) D n 2 = 2 n(n 1) I n + Ũ 2(n 1) Q(I n Ã) D n 2 = 2 n(n 1) I n + Ũ 2(n 1) (Π n 1 j=1 ãn j jj ) 2, (34) 11
12 where Theorem 6.13(iii) of Magnus (1988, p. 100) is used in deriving the third equality. Thus equation (17) follows from equations (19) and (34). Turning to equation (18), it follows from Theorem 3.1 and Theorem 4.2(ii) in Magnus (1988) that D + n [(Ã I n) + (I n Ã)K n]l n = D + n [(Ã I n) + K n (Ã I n)]l n = D + n (I n + K n )(Ã I n)l n = 2D + n (Ã I n)l n = 2(D nd n ) 1 (L n (Ã I n )D n ). (35) Given Theorem 4.4(iii) and Theorem 5.12 in Magnus (1988), we have that D n + [(Ã I n) + (I n Ã)K n]l n = 2(D nd n ) 1 (L n (Ã I n )D n ) = 2 n(n+1)/2 2 n(n 1)/2 Ln (Ã I n )D n ) = 2 n Π n i=1ã n i+1 ii (36) which proves equation (18). Lastly, combining equations (5), (17) and (18), we obtain the desired result. Acknowledgments We thank Tom Doan, Jan Magnus and Jonas Arias for helpful discussions. 12
13 References 1. Inoue, A., Kilian, L., Inference on impulse response functions in structural VAR models. Journal of Econometrics 177, León, C.A., Massé, J.-C., Rivest, L.P., A statistical model for random rotations. Journal of Multivariate Analysis 976, Magnus, J.R., Linear structures. Oxford University Press, Oxford, UK. 4. Magnus, J.R., Neudecker, H., Matrix differential calculus with applications in statistics and econometrics. second edition. Wiley, Chichester, UK. 13
Corrigendum to Inference on impulse response functions in structural VAR models [J. Econometrics 177 (2013), 1-13]
Corrigendum to Inference on impulse response functions in structural VAR models [J. Econometrics 177 (2013), 1-13] Atsushi Inoue a Lutz Kilian b a Department of Economics, Vanderbilt University, Nashville
More informationLemma 8: Suppose the N by N matrix A has the following block upper triangular form:
17 4 Determinants and the Inverse of a Square Matrix In this section, we are going to use our knowledge of determinants and their properties to derive an explicit formula for the inverse of a square matrix
More informationWeb Appendix to Multivariate High-Frequency-Based Volatility (HEAVY) Models
Web Appendix to Multivariate High-Frequency-Based Volatility (HEAVY) Models Diaa Noureldin Department of Economics, University of Oxford, & Oxford-Man Institute, Eagle House, Walton Well Road, Oxford OX
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 3: Positive-Definite Systems; Cholesky Factorization Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 11 Symmetric
More informationNew Introduction to Multiple Time Series Analysis
Helmut Lütkepohl New Introduction to Multiple Time Series Analysis With 49 Figures and 36 Tables Springer Contents 1 Introduction 1 1.1 Objectives of Analyzing Multiple Time Series 1 1.2 Some Basics 2
More informationVector Autoregressive Model. Vector Autoregressions II. Estimation of Vector Autoregressions II. Estimation of Vector Autoregressions I.
Vector Autoregressive Model Vector Autoregressions II Empirical Macroeconomics - Lect 2 Dr. Ana Beatriz Galvao Queen Mary University of London January 2012 A VAR(p) model of the m 1 vector of time series
More informationIdentification of Economic Shocks by Inequality Constraints in Bayesian Structural Vector Autoregression
Identification of Economic Shocks by Inequality Constraints in Bayesian Structural Vector Autoregression Markku Lanne Department of Political and Economic Studies, University of Helsinki Jani Luoto Department
More informationOn Expected Gaussian Random Determinants
On Expected Gaussian Random Determinants Moo K. Chung 1 Department of Statistics University of Wisconsin-Madison 1210 West Dayton St. Madison, WI 53706 Abstract The expectation of random determinants whose
More informationMAT 2037 LINEAR ALGEBRA I web:
MAT 237 LINEAR ALGEBRA I 2625 Dokuz Eylül University, Faculty of Science, Department of Mathematics web: Instructor: Engin Mermut http://kisideuedutr/enginmermut/ HOMEWORK 2 MATRIX ALGEBRA Textbook: Linear
More informationTime-Varying Vector Autoregressive Models with Structural Dynamic Factors
Time-Varying Vector Autoregressive Models with Structural Dynamic Factors Paolo Gorgi, Siem Jan Koopman, Julia Schaumburg http://sjkoopman.net Vrije Universiteit Amsterdam School of Business and Economics
More informationPhys 201. Matrices and Determinants
Phys 201 Matrices and Determinants 1 1.1 Matrices 1.2 Operations of matrices 1.3 Types of matrices 1.4 Properties of matrices 1.5 Determinants 1.6 Inverse of a 3 3 matrix 2 1.1 Matrices A 2 3 7 =! " 1
More informationReview of Matrices and Block Structures
CHAPTER 2 Review of Matrices and Block Structures Numerical linear algebra lies at the heart of modern scientific computing and computational science. Today it is not uncommon to perform numerical computations
More informationLabor-Supply Shifts and Economic Fluctuations. Technical Appendix
Labor-Supply Shifts and Economic Fluctuations Technical Appendix Yongsung Chang Department of Economics University of Pennsylvania Frank Schorfheide Department of Economics University of Pennsylvania January
More informationGeneralized Impulse Response Analysis: General or Extreme?
MPRA Munich Personal RePEc Archive Generalized Impulse Response Analysis: General or Extreme? Kim Hyeongwoo Auburn University April 2009 Online at http://mpra.ub.uni-muenchen.de/17014/ MPRA Paper No. 17014,
More informationLinear Algebra and Matrix Inversion
Jim Lambers MAT 46/56 Spring Semester 29- Lecture 2 Notes These notes correspond to Section 63 in the text Linear Algebra and Matrix Inversion Vector Spaces and Linear Transformations Matrices are much
More informationTitle. Description. var intro Introduction to vector autoregressive models
Title var intro Introduction to vector autoregressive models Description Stata has a suite of commands for fitting, forecasting, interpreting, and performing inference on vector autoregressive (VAR) models
More informationSection 9.2: Matrices.. a m1 a m2 a mn
Section 9.2: Matrices Definition: A matrix is a rectangular array of numbers: a 11 a 12 a 1n a 21 a 22 a 2n A =...... a m1 a m2 a mn In general, a ij denotes the (i, j) entry of A. That is, the entry in
More informationInference when identifying assumptions are doubted. A. Theory B. Applications
Inference when identifying assumptions are doubted A. Theory B. Applications 1 A. Theory Structural model of interest: A y t B 1 y t1 B m y tm u t nn n1 u t i.i.d. N0, D D diagonal 2 Bayesian approach:
More informationMatrix Algebra, Class Notes (part 2) by Hrishikesh D. Vinod Copyright 1998 by Prof. H. D. Vinod, Fordham University, New York. All rights reserved.
Matrix Algebra, Class Notes (part 2) by Hrishikesh D. Vinod Copyright 1998 by Prof. H. D. Vinod, Fordham University, New York. All rights reserved. 1 Converting Matrices Into (Long) Vectors Convention:
More informationCointegrated VAR s. Eduardo Rossi University of Pavia. November Rossi Cointegrated VAR s Financial Econometrics / 56
Cointegrated VAR s Eduardo Rossi University of Pavia November 2013 Rossi Cointegrated VAR s Financial Econometrics - 2013 1 / 56 VAR y t = (y 1t,..., y nt ) is (n 1) vector. y t VAR(p): Φ(L)y t = ɛ t The
More informationAN EFFICIENT GLS ESTIMATOR OF TRIANGULAR MODELS WITH COVARIANCE RESTRICTIONS*
Journal of Econometrics 42 (1989) 267-273. North-Holland AN EFFICIENT GLS ESTIMATOR OF TRIANGULAR MODELS WITH COVARIANCE RESTRICTIONS* Manuel ARELLANO Institute of Economics and Statistics, Oxford OXI
More informationOn the Skeel condition number, growth factor and pivoting strategies for Gaussian elimination
On the Skeel condition number, growth factor and pivoting strategies for Gaussian elimination J.M. Peña 1 Introduction Gaussian elimination (GE) with a given pivoting strategy, for nonsingular matrices
More informationEstimation and Testing for Common Cycles
Estimation and esting for Common Cycles Anders Warne February 27, 2008 Abstract: his note discusses estimation and testing for the presence of common cycles in cointegrated vector autoregressions A simple
More informationMassachusetts Institute of Technology Department of Economics Time Series Lecture 6: Additional Results for VAR s
Massachusetts Institute of Technology Department of Economics Time Series 14.384 Guido Kuersteiner Lecture 6: Additional Results for VAR s 6.1. Confidence Intervals for Impulse Response Functions There
More informationMath 408 Advanced Linear Algebra
Math 408 Advanced Linear Algebra Chi-Kwong Li Chapter 4 Hermitian and symmetric matrices Basic properties Theorem Let A M n. The following are equivalent. Remark (a) A is Hermitian, i.e., A = A. (b) x
More informationLinear Models for Multivariate Repeated Measures Data
THE UNIVERSITY OF TEXAS AT SAN ANTONIO, COLLEGE OF BUSINESS Working Paper SERIES Date December 3, 200 WP # 007MSS-253-200 Linear Models for Multivariate Repeated Measures Data Anuradha Roy Management Science
More informationShrinkage in Set-Identified SVARs
Shrinkage in Set-Identified SVARs Alessio Volpicella (Queen Mary, University of London) 2018 IAAE Annual Conference, Université du Québec à Montréal (UQAM) and Université de Montréal (UdeM) 26-29 June
More informationMath113: Linear Algebra. Beifang Chen
Math3: Linear Algebra Beifang Chen Spring 26 Contents Systems of Linear Equations 3 Systems of Linear Equations 3 Linear Systems 3 2 Geometric Interpretation 3 3 Matrices of Linear Systems 4 4 Elementary
More information5.6. PSEUDOINVERSES 101. A H w.
5.6. PSEUDOINVERSES 0 Corollary 5.6.4. If A is a matrix such that A H A is invertible, then the least-squares solution to Av = w is v = A H A ) A H w. The matrix A H A ) A H is the left inverse of A and
More informationReduced rank regression in cointegrated models
Journal of Econometrics 06 (2002) 203 26 www.elsevier.com/locate/econbase Reduced rank regression in cointegrated models.w. Anderson Department of Statistics, Stanford University, Stanford, CA 94305-4065,
More informationTOPICS IN ADVANCED ECONOMETRICS: SVAR MODELS PART I LESSON 2. Luca Fanelli. University of Bologna.
TOPICS IN ADVANCED ECONOMETRICS: SVAR MODELS PART I LESSON 2 Luca Fanelli University of Bologna luca.fanelli@unibo.it http://www.rimini.unibo.it/fanelli The material in these slides is based on the following
More informationEquality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same.
Introduction Matrix Operations Matrix: An m n matrix A is an m-by-n array of scalars from a field (for example real numbers) of the form a a a n a a a n A a m a m a mn The order (or size) of A is m n (read
More informationVector autoregressions, VAR
1 / 45 Vector autoregressions, VAR Chapter 2 Financial Econometrics Michael Hauser WS17/18 2 / 45 Content Cross-correlations VAR model in standard/reduced form Properties of VAR(1), VAR(p) Structural VAR,
More informationJim Lambers MAT 610 Summer Session Lecture 1 Notes
Jim Lambers MAT 60 Summer Session 2009-0 Lecture Notes Introduction This course is about numerical linear algebra, which is the study of the approximate solution of fundamental problems from linear algebra
More information1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )
Direct Methods for Linear Systems Chapter Direct Methods for Solving Linear Systems Per-Olof Persson persson@berkeleyedu Department of Mathematics University of California, Berkeley Math 18A Numerical
More informationAn Extrapolated Gauss-Seidel Iteration
mathematics of computation, volume 27, number 124, October 1973 An Extrapolated Gauss-Seidel Iteration for Hessenberg Matrices By L. J. Lardy Abstract. We show that for certain systems of linear equations
More informationInference when identifying assumptions are doubted. A. Theory. Structural model of interest: B 1 y t1. u t. B m y tm. u t i.i.d.
Inference when identifying assumptions are doubted A. Theory B. Applications Structural model of interest: A y t B y t B m y tm nn n i.i.d. N, D D diagonal A. Theory Bayesian approach: Summarize whatever
More informationEstimation of a multivariate normal covariance matrix with staircase pattern data
AISM (2007) 59: 211 233 DOI 101007/s10463-006-0044-x Xiaoqian Sun Dongchu Sun Estimation of a multivariate normal covariance matrix with staircase pattern data Received: 20 January 2005 / Revised: 1 November
More informationFisher information for generalised linear mixed models
Journal of Multivariate Analysis 98 2007 1412 1416 www.elsevier.com/locate/jmva Fisher information for generalised linear mixed models M.P. Wand Department of Statistics, School of Mathematics and Statistics,
More informationLecture 2 INF-MAT : , LU, symmetric LU, Positve (semi)definite, Cholesky, Semi-Cholesky
Lecture 2 INF-MAT 4350 2009: 7.1-7.6, LU, symmetric LU, Positve (semi)definite, Cholesky, Semi-Cholesky Tom Lyche and Michael Floater Centre of Mathematics for Applications, Department of Informatics,
More informationDeterminants Chapter 3 of Lay
Determinants Chapter of Lay Dr. Doreen De Leon Math 152, Fall 201 1 Introduction to Determinants Section.1 of Lay Given a square matrix A = [a ij, the determinant of A is denoted by det A or a 11 a 1j
More informationA primer on Structural VARs
A primer on Structural VARs Claudia Foroni Norges Bank 10 November 2014 Structural VARs 1/ 26 Refresh: what is a VAR? VAR (p) : where y t K 1 y t = ν + B 1 y t 1 +... + B p y t p + u t, (1) = ( y 1t...
More informationMatrix Differential Calculus with Applications in Statistics and Econometrics
Matrix Differential Calculus with Applications in Statistics and Econometrics Revised Edition JAN. R. MAGNUS CentERjor Economic Research, Tilburg University and HEINZ NEUDECKER Cesaro, Schagen JOHN WILEY
More informationCME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6
CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6 GENE H GOLUB Issues with Floating-point Arithmetic We conclude our discussion of floating-point arithmetic by highlighting two issues that frequently
More informationSecrets of Matrix Factorization: Supplementary Document
Secrets of Matrix Factorization: Supplementary Document Je Hyeong Hong University of Cambridge jhh37@cam.ac.uk Andrew Fitzgibbon Microsoft, Cambridge, UK awf@microsoft.com Abstract This document consists
More informationIdentification in Structural VAR models with different volatility regimes
Identification in Structural VAR models with different volatility regimes Emanuele Bacchiocchi 28th September 2010 PRELIMINARY VERSION Abstract In this paper we study the identification conditions in structural
More informationSecond-order approximation of dynamic models without the use of tensors
Second-order approximation of dynamic models without the use of tensors Paul Klein a, a University of Western Ontario First draft: May 17, 2005 This version: January 24, 2006 Abstract Several approaches
More informationFoundations of Matrix Analysis
1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the
More informationMeasures and Jacobians of Singular Random Matrices. José A. Díaz-Garcia. Comunicación de CIMAT No. I-07-12/ (PE/CIMAT)
Measures and Jacobians of Singular Random Matrices José A. Díaz-Garcia Comunicación de CIMAT No. I-07-12/21.08.2007 (PE/CIMAT) Measures and Jacobians of singular random matrices José A. Díaz-García Universidad
More informationRegression systems for unbalanced panel data: a stepwise maximum likelihood procedure
Journal of Econometrics 1 (004) 81 91 www.elsevier.com/locate/econbase Regression systems for unbalanced panel data: a stepwise maximum likelihood procedure Erik Birn Department of Economics, University
More informationTopics. Vectors (column matrices): Vector addition and scalar multiplication The matrix of a linear function y Ax The elements of a matrix A : A ij
Topics Vectors (column matrices): Vector addition and scalar multiplication The matrix of a linear function y Ax The elements of a matrix A : A ij or a ij lives in row i and column j Definition of a matrix
More informationOrthogonal Transformations
Orthogonal Transformations Tom Lyche University of Oslo Norway Orthogonal Transformations p. 1/3 Applications of Qx with Q T Q = I 1. solving least squares problems (today) 2. solving linear equations
More informationLinear algebra for computational statistics
University of Seoul May 3, 2018 Vector and Matrix Notation Denote 2-dimensional data array (n p matrix) by X. Denote the element in the ith row and the jth column of X by x ij or (X) ij. Denote by X j
More informationNon-Gaussian Maximum Entropy Processes
Non-Gaussian Maximum Entropy Processes Georgi N. Boshnakov & Bisher Iqelan First version: 3 April 2007 Research Report No. 3, 2007, Probability and Statistics Group School of Mathematics, The University
More informationMatrix & Linear Algebra
Matrix & Linear Algebra Jamie Monogan University of Georgia For more information: http://monogan.myweb.uga.edu/teaching/mm/ Jamie Monogan (UGA) Matrix & Linear Algebra 1 / 84 Vectors Vectors Vector: A
More informationHands-on Matrix Algebra Using R
Preface vii 1. R Preliminaries 1 1.1 Matrix Defined, Deeper Understanding Using Software.. 1 1.2 Introduction, Why R?.................... 2 1.3 Obtaining R.......................... 4 1.4 Reference Manuals
More informationLinear Equations and Matrix
1/60 Chia-Ping Chen Professor Department of Computer Science and Engineering National Sun Yat-sen University Linear Algebra Gaussian Elimination 2/60 Alpha Go Linear algebra begins with a system of linear
More informationLinear Algebra and Matrices
Linear Algebra and Matrices 4 Overview In this chapter we studying true matrix operations, not element operations as was done in earlier chapters. Working with MAT- LAB functions should now be fairly routine.
More informationLAPACK-Style Codes for Pivoted Cholesky and QR Updating
LAPACK-Style Codes for Pivoted Cholesky and QR Updating Sven Hammarling 1, Nicholas J. Higham 2, and Craig Lucas 3 1 NAG Ltd.,Wilkinson House, Jordan Hill Road, Oxford, OX2 8DR, England, sven@nag.co.uk,
More informationMatrix Algebra. Matrix Algebra. Chapter 8 - S&B
Chapter 8 - S&B Algebraic operations Matrix: The size of a matrix is indicated by the number of its rows and the number of its columns. A matrix with k rows and n columns is called a k n matrix. The number
More informationGeneralized Impulse Response Analysis: General or Extreme?
Auburn University Department of Economics Working Paper Series Generalized Impulse Response Analysis: General or Extreme? Hyeongwoo Kim Auburn University AUWP 2012-04 This paper can be downloaded without
More informationMath 407: Linear Optimization
Math 407: Linear Optimization Lecture 16: The Linear Least Squares Problem II Math Dept, University of Washington February 28, 2018 Lecture 16: The Linear Least Squares Problem II (Math Dept, University
More informationMATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS SYSTEMS OF EQUATIONS AND MATRICES Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
More informationDiagonal and Monomial Solutions of the Matrix Equation AXB = C
Iranian Journal of Mathematical Sciences and Informatics Vol. 9, No. 1 (2014), pp 31-42 Diagonal and Monomial Solutions of the Matrix Equation AXB = C Massoud Aman Department of Mathematics, Faculty of
More informationSection 9.2: Matrices. Definition: A matrix A consists of a rectangular array of numbers, or elements, arranged in m rows and n columns.
Section 9.2: Matrices Definition: A matrix A consists of a rectangular array of numbers, or elements, arranged in m rows and n columns. That is, a 11 a 12 a 1n a 21 a 22 a 2n A =...... a m1 a m2 a mn A
More informationMultivariate Distributions
IEOR E4602: Quantitative Risk Management Spring 2016 c 2016 by Martin Haugh Multivariate Distributions We will study multivariate distributions in these notes, focusing 1 in particular on multivariate
More informationFundamentals of Engineering Analysis (650163)
Philadelphia University Faculty of Engineering Communications and Electronics Engineering Fundamentals of Engineering Analysis (6563) Part Dr. Omar R Daoud Matrices: Introduction DEFINITION A matrix is
More informationThe QR Factorization
The QR Factorization How to Make Matrices Nicer Radu Trîmbiţaş Babeş-Bolyai University March 11, 2009 Radu Trîmbiţaş ( Babeş-Bolyai University) The QR Factorization March 11, 2009 1 / 25 Projectors A projector
More informationComputational Linear Algebra
Computational Linear Algebra PD Dr. rer. nat. habil. Ralf Peter Mundani Computation in Engineering / BGU Scientific Computing in Computer Science / INF Winter Term 2017/18 Part 2: Direct Methods PD Dr.
More information1 Last time: least-squares problems
MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that
More informationChapter 4. Matrices and Matrix Rings
Chapter 4 Matrices and Matrix Rings We first consider matrices in full generality, i.e., over an arbitrary ring R. However, after the first few pages, it will be assumed that R is commutative. The topics,
More information. =. a i1 x 1 + a i2 x 2 + a in x n = b i. a 11 a 12 a 1n a 21 a 22 a 1n. i1 a i2 a in
Vectors and Matrices Continued Remember that our goal is to write a system of algebraic equations as a matrix equation. Suppose we have the n linear algebraic equations a x + a 2 x 2 + a n x n = b a 2
More informationLECTURE 2 LINEAR REGRESSION MODEL AND OLS
SEPTEMBER 29, 2014 LECTURE 2 LINEAR REGRESSION MODEL AND OLS Definitions A common question in econometrics is to study the effect of one group of variables X i, usually called the regressors, on another
More informationA Short Note on Resolving Singularity Problems in Covariance Matrices
International Journal of Statistics and Probability; Vol. 1, No. 2; 2012 ISSN 1927-7032 E-ISSN 1927-7040 Published by Canadian Center of Science and Education A Short Note on Resolving Singularity Problems
More informationA distance measure between cointegration spaces
Economics Letters 70 (001) 1 7 www.elsevier.com/ locate/ econbase A distance measure between cointegration spaces Rolf Larsson, Mattias Villani* Department of Statistics, Stockholm University S-106 1 Stockholm,
More informationSimultaneous Diagonalization of Positive Semi-definite Matrices
Simultaneous Diagonalization of Positive Semi-definite Matrices Jan de Leeuw Version 21, May 21, 2017 Abstract We give necessary and sufficient conditions for solvability of A j = XW j X, with the A j
More informationMathematics. EC / EE / IN / ME / CE. for
Mathematics for EC / EE / IN / ME / CE By www.thegateacademy.com Syllabus Syllabus for Mathematics Linear Algebra: Matrix Algebra, Systems of Linear Equations, Eigenvalues and Eigenvectors. Probability
More informationGraduate Mathematical Economics Lecture 1
Graduate Mathematical Economics Lecture 1 Yu Ren WISE, Xiamen University September 23, 2012 Outline 1 2 Course Outline ematical techniques used in graduate level economics courses Mathematics for Economists
More informationUndergraduate Mathematical Economics Lecture 1
Undergraduate Mathematical Economics Lecture 1 Yu Ren WISE, Xiamen University September 15, 2014 Outline 1 Courses Description and Requirement 2 Course Outline ematical techniques used in economics courses
More informationProblem 1. CS205 Homework #2 Solutions. Solution
CS205 Homework #2 s Problem 1 [Heath 3.29, page 152] Let v be a nonzero n-vector. The hyperplane normal to v is the (n-1)-dimensional subspace of all vectors z such that v T z = 0. A reflector is a linear
More informationData-Driven Inference on Sign Restrictions in Bayesian Structural Vector Autoregression. Markku Lanne and Jani Luoto. CREATES Research Paper
Data-Driven Inference on Sign Restrictions in Bayesian Structural Vector Autoregression Markku Lanne and Jani Luoto CREATES Research Paper 2016-4 Department of Economics and Business Economics Aarhus University
More informationMatrix Algebra for Engineers Jeffrey R. Chasnov
Matrix Algebra for Engineers Jeffrey R. Chasnov The Hong Kong University of Science and Technology The Hong Kong University of Science and Technology Department of Mathematics Clear Water Bay, Kowloon
More informationVanderbilt University Department of Economics Working Papers
Vanderbilt University Department of Economics Working Papers 9-0000 The uniform validity of impulse response inference in autoregressions Atsushi Inoue Vanderbilt University Lutz Kilian University of Michigan
More informationLAPACK-Style Codes for Pivoted Cholesky and QR Updating. Hammarling, Sven and Higham, Nicholas J. and Lucas, Craig. MIMS EPrint: 2006.
LAPACK-Style Codes for Pivoted Cholesky and QR Updating Hammarling, Sven and Higham, Nicholas J. and Lucas, Craig 2007 MIMS EPrint: 2006.385 Manchester Institute for Mathematical Sciences School of Mathematics
More informationReview of Vectors and Matrices
A P P E N D I X D Review of Vectors and Matrices D. VECTORS D.. Definition of a Vector Let p, p, Á, p n be any n real numbers and P an ordered set of these real numbers that is, P = p, p, Á, p n Then P
More informationforms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms
Christopher Engström November 14, 2014 Hermitian LU QR echelon Contents of todays lecture Some interesting / useful / important of matrices Hermitian LU QR echelon Rewriting a as a product of several matrices.
More informationEstimating and Identifying Vector Autoregressions Under Diagonality and Block Exogeneity Restrictions
Estimating and Identifying Vector Autoregressions Under Diagonality and Block Exogeneity Restrictions William D. Lastrapes Department of Economics Terry College of Business University of Georgia Athens,
More informationPenalized varimax. Abstract
Penalized varimax 1 Penalized varimax Nickolay T. Trendafilov and Doyo Gragn Department of Mathematics and Statistics, The Open University, Walton Hall, Milton Keynes MK7 6AA, UK Abstract A common weakness
More informationMatrix Algebra Review
APPENDIX A Matrix Algebra Review This appendix presents some of the basic definitions and properties of matrices. Many of the matrices in the appendix are named the same as the matrices that appear in
More informationEE731 Lecture Notes: Matrix Computations for Signal Processing
EE731 Lecture Notes: Matrix Computations for Signal Processing James P. Reilly c Department of Electrical and Computer Engineering McMaster University September 22, 2005 0 Preface This collection of ten
More informationSome results on the existence of t-all-or-nothing transforms over arbitrary alphabets
Some results on the existence of t-all-or-nothing transforms over arbitrary alphabets Navid Nasr Esfahani, Ian Goldberg and Douglas R. Stinson David R. Cheriton School of Computer Science University of
More informationChapter SSM: Linear Algebra Section Fails to be invertible; since det = 6 6 = Invertible; since det = = 2.
SSM: Linear Algebra Section 61 61 Chapter 6 1 2 1 Fails to be invertible; since det = 6 6 = 0 3 6 3 5 3 Invertible; since det = 33 35 = 2 7 11 5 Invertible; since det 2 5 7 0 11 7 = 2 11 5 + 0 + 0 0 0
More informationSpectral inequalities and equalities involving products of matrices
Spectral inequalities and equalities involving products of matrices Chi-Kwong Li 1 Department of Mathematics, College of William & Mary, Williamsburg, Virginia 23187 (ckli@math.wm.edu) Yiu-Tung Poon Department
More informationJoint Confidence Sets for Structural Impulse Responses
Joint Confidence Sets for Structural Impulse Responses Atsushi Inoue Lutz Kilian Vanderbilt University University of Michigan February 24, 2015 Abstract Many questions of economic interest in structural
More informationLinear Systems and Matrices
Department of Mathematics The Chinese University of Hong Kong 1 System of m linear equations in n unknowns (linear system) a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.......
More informationNew Information Response Functions
New Information Response Functions Caroline JARDET (1) Banque de France Alain MONFORT (2) CNAM, CREST and Banque de France Fulvio PEGORARO (3) Banque de France and CREST First version : February, 2009.
More information14.2 QR Factorization with Column Pivoting
page 531 Chapter 14 Special Topics Background Material Needed Vector and Matrix Norms (Section 25) Rounding Errors in Basic Floating Point Operations (Section 33 37) Forward Elimination and Back Substitution
More informationLectures on Linear Algebra for IT
Lectures on Linear Algebra for IT by Mgr. Tereza Kovářová, Ph.D. following content of lectures by Ing. Petr Beremlijski, Ph.D. Department of Applied Mathematics, VSB - TU Ostrava Czech Republic 11. Determinants
More informationELEMENTARY LINEAR ALGEBRA WITH APPLICATIONS. 1. Linear Equations and Matrices
ELEMENTARY LINEAR ALGEBRA WITH APPLICATIONS KOLMAN & HILL NOTES BY OTTO MUTZBAUER 11 Systems of Linear Equations 1 Linear Equations and Matrices Numbers in our context are either real numbers or complex
More informationThe Determinant: a Means to Calculate Volume
The Determinant: a Means to Calculate Volume Bo Peng August 16, 2007 Abstract This paper gives a definition of the determinant and lists many of its well-known properties Volumes of parallelepipeds are
More information