Estimation of state space models with skewed shocks
|
|
- Maude Lillian Flynn
- 5 years ago
- Views:
Transcription
1 Estimation of state space models with skewed shocks Grzegorz Koloch April 30, 2012 Abstract This paper provides analytical formulae for the likelihood function, predictive densities and the filtration densities for linear state space models in which shocks are allowed to follow a skewed distribution. In particular we work with the closed skew normal distribution introduced in Genton (2004), which nests a normal distribution as a special case. Closure of the distribution in the state space setting with respect to all necessary transformations is guaranteed by a simple regularization procedure which does not influence the likelihood function. Presented results allow for estimation, filtration and prediction of such models as vector autoregressions and first order perturbations of DSGE models with skewed shocks. This allows economists to assess in their models asymmetries of observed data, of shocks, impulse responses and forecasts confidence intervals. Keywords: Maximum Likelihood Estimation, State Space Models, Closed Skewed Normal Distribution, DSGE, VAR. JEL: C51, C13, E32 1 Introduction In this paper we provide a maximum likelihood estimator for a linear state space model which allows the shocks in the transition equation to follow a skewed distri- National Bank of Poland, Economic Institute, Bureau of Applied Research, Swietokrzyska 19/21, Warsaw, Poland. Warsaw School of Economics, Institute of Econometrics, Department of Decision Analysis and Support, Al. Niepodleglosci 164, Warsaw, Poland. 1
2 bution 1. Let us consider a following state space model: y t = F x t + Hζ t, x t = Ax t 1 + Bξ t, ζ t N(0, Ψ ζ ), ξ t p ξ (θ ξ ) x 0 N(µ x0, Ψ x0 ) for t T = {1, 2,..., T }, where x t R n and y t R m denote states and observables respectively, R n n A 0, B R n n ξ, F R m n, F R m n ζ, n ξ, n ζ > 0, moreover Ψ ζ R n ζ n ζ, Ψ x0 R n n, and Ψ ζ, Ψ ζ 0. Finally, p ξ (θ ξ ) denotes a probability distribution of martingale difference shocks, which depends on a vector of parameters θ ξ. The usual assumption is that p ξ is a multivariate normal distribution, independent across its dimensions. In such a case the Kalman filter constitutes an optimal filtration procedure. If normality assumption is relaxed, Kalman filter remains an optimal linear filter, see e.g. Simon (2006). I assume that shocks are independent, but, for some values of θ ξ, p ξ is a skewed distribution. In particular, I assume that shocks follow a closed skew normal distribution (csn henceforth), see Genton (2004). The csn distribution is chosen, since it is closed under most transformations required for optimal filtration and ML estimation in the state space setting 2. Let us note, however, that matrix A can be rank deficient 3, which will turn out to be an obstacle in ML estimation as singularity of A precludes propagation of the csn distribution through the state space setting. I provide a regularization procedure which deals with this problem. To allow for prediction, filtration and estimation I provide exact formulae for p(y t Y t 1 ), p(x t Y t 1 ), p(ξ t Y t 1 ), p(ζ t Y t 1 ), p(x t Y t ), p(ξ t Y t ), p(ζ t Y t ) and for the likelihood function p(θ Y t ), where θ = (θ F, θ H, θ ζ, θ A, θ F, θ ξ, θ x0 ). When one goes 1 Measurement shocks are assumed to be normally distributed but extension to skewed measurement errors is straightforward. We prefer to work with normal measurement errors because of the appealing interpretation based on the central limit theorem. 2 Details are provided in section 2 3 Which is a standard case when the state space form represents a first order perturbation of a DSGE model. 2
3 through the article, one should also find it easy to obtain formulae for any type of filtration, eg. for the smoother. Since the state space form represents some very popular macroeconomic tools like VARs and reduced form first order perturbations of DSGE model, provided results allow for extending the agenda of empirical macroeconomics by issues related to skewness in the data. In particular it allows estimating VARs and DSGE models with skewed shocks and getting asymmetric confidence intervals of impulse responses and forecasts, hence to assess asymmetries of eg. inflation risks and finally, without referring to nonlinearities. Remainder of the paper is arranged as follows. Section 2 describes the closed skewed normal distribution and discusses its basic properties, section?? derives the filter and section 3 derives the quasi maximum-likelihood estimation procedure. Section 4 provides exemplary application and the last section concludes. 2 The skewed normal distribution 2.1 Definition Let us denote a density function of a p-dimensional normal distribution with mean 4 µ and positive-definite covariance matrix Σ by φ p (z; µ, Σ). Let us also denote a cumulative distribution function of a q-dimensional normal distribution with mean µ and nonnegative-definite covariance matrix Σ by Φ q (z; µ, Σ). For q > 1 function Φ q does not have a closed form. By R p q, p, q 1, let us denote a space of linear operators from R p to R q. For every M R p q let M denote a determinant of M and let r(m) denote a rank of M. We will define the closed skewed normal, possibly singular, distribution by means of the moment generating function (mgf). Then, under nonsingularity conditions, probability density function (pdf) will be provided. Definition 2.1. (csn distribution mgf ) Let µ R p and ϑ R q, p, q 1. Let Σ R p p and R q q, Σ, 0, and let D R q p. We say that random variable z has a (p, q) dimensional closed skewed normal distribution with parameters 4 All vectors are column vectors throughout the paper. 3
4 µ, Σ, D, ϑ and if moment generating function of z, M z (t), is given by: which henceforth will be denoted by: M z (t) = Φ q(d Σt; ϑ, + DΣD T ) Φ q (0; ϑ, + DΣD T e tt µ+ 1 tt Σt 2 ) z csn p,q ( µ, Σ, D, ϑ, ) Note that matrices Σ and are allowed to be singular. If Σ is not positive definite, i.e. Σ = 0, resulting distribution is called singular. If Σ is positive definite, i.e. Σ > 0, distribution is called nonsingular. The csn distribution is closed in the sense, that it is closed under full rank linear transformations 5. Isomorphic linear transformations transform nonsingular csn variables into nonsigular ones and singular variables into singular ones. Full row, but column rank deficient linear transformations (dimension shrinkage) transform nonsingular csn variables into nonsigular ones and singular variables into singular or nonsingular ones. Full column, but row rank deficient linear transformations (dimension expansion) transform nonsingular csn variables into singular ones, whreas singular variables remain singular. Both singular and nonsingular variables can be transformed into a non-csn distributed variables under a rank deficient transformation. The skewed normal distribution when considered as consisting of both singular and nonsingular cns variables is therefore not closed under arbitrary linear transformations, which has negative consequences for maximum likelihood estimation of state space models with csn shocks when the transition matrix in state space equations is singular, which typically is the case in DSGE modeling. If Σ > 0, a csn random variable z has a probability density function accordin to: Definition 2.2. (csn distribution pdf ) If a random variable z follows a (p, q)- dimensional, p, q 1, closed skewed normal distribution with parameters µ, Σ, D, ϑ and, where µ R p, ϑ R q, Σ R p p, Σ > 0, R p p, 0 and D R q p, 5 Under full rank transformation we mean full row or full column rank transformation and this definition embraces the case when matrix of the transformation is square and has a full rank. In the latter case the transformation is called an isomorphism. When both the row and the column ranks are not full, transformation is called rank deficient. 4
5 than probability density function of z is given by: p(z) = φ p (z; µ, Σ) Φ q(d(z µ); ϑ, ) Φ q (0; ϑ, + DΣD T ) (1) Density function (1) defines a (p, q)-dimensional nonsingular closed skewed normal distribution in the sense that a random variable has (p, q)-dimensional nonsingular closed skewed normal distribution with parameters µ, Σ, D, ϑ and if and only if its density function for every z R p equals p(z) in (1). The probability density function (1) involves a probability distribution function of a q-dimensional normal distribution for, in principle, arbitrarily large q, which entails computational difficulties when working with a likelihood function based on p(z) in (1). Parameters µ, Σ and D have interpretation of location, scale and skewness parameters respectively. Parameters ϑ and are artificial, but inclusion of these additional dimensions allows for closure of the csn distribution under conditioning and marginalization respectively. The q-dimension is also artificial, but it allows for closure for sums and the joint distribution of independent (not necessarily iid) variables. When Σ, D and are scalars, they will be denoted respectively by σ, d and δ. 2.2 Basic properties Let us make use of the following: Remark 2.3. For p = q = 1, ϑ = 0 and = 1 the csn distribution reduces to the Azzalini skewed normal distribution, see Azzalini (1996), Azzalini (1999). Such a case will be denoted by: In the next sections we will find useful the following: z sn( µ, σ, d) (2) Corollary 2.4. Let z csn 1,1 ( µ, σ, d, ϑ, δ) for parameters as in definition (2.2), and assume that δ + d 2 σ 0, then: 2 d σ E(z) = µ + π δ + d2 σ var(z) = σ 2 d 2 σ 2 π δ + d 2 σ ) 3 ( E(z E(z)) 3 = (2 π 2 ) ( 2 π d σ (δ + σd 2 ) 1 2 (3) (4) ) 3 (5) 5
6 It follows that: Remark 2.5. Let z csn 1,1 ( µ, σ, d, ϑ, δ) for parameters as in definition (2.4), then d σ E(z) = 0 if and only if µ =. δ+d2 σ 2 π Which we use in sampling csn variables withe zero mean. We also need the following: Corollary 2.6. Let z csn p,q ( µ, Σ, D, ϑ, ), p, q 1, for parameters as in definition (2.1). Elements of z are independent if and only if matrices Σ and D are diagonal. Since Σ is p p and D is q p, corollary (2.6) implies that it is impossible to have q = 1 while keeping elements of z independent for p > 1, because it has to be the case that q = p > 1 in order for D to be diagonal. This is relevant for state space models with csn distributed iid disturbances eg. shocks in the transition equation, because the state variable, say ξ t, in every period consists of the csn distributed state from the previous period, say ξ t 1, plus the csn-distributed disturbance 6, say u t, and when we add two csn variables we have to add their q- dimensions, so thet the q-dimension of ξ t is the sum of q-dimensions of ξ t 1 and u t, hence, according to corollary (2.6), contribution of u t to q-dimension of ξ t in every period cannot be squeezed to eg. 1, but must equal size of u t, hence q-dimension of ξ t quickly increases which poses serious numerical difficulties. In further sections we will also need the following corollaries ( ): Corollary 2.7. Let z csn p,q ( µ, Σ, D, ϑ, ), p, q 1, for parameters as in definition (2.1). Let also x N(µ x, Σ x ), Σ x > 0, be independent of z, then: z + x csn p,q ( µ + µ x, Σ + Σ x, D Σ( Σ + Σ x ) 1, ϑ, + (D(I Σ( Σ + Σ x ) 1 )) ΣD T ) Corollary 2.8. Let z csn 1,q ( µ, σ, d, ϑ, δ), q 1 and for parameters as in definition (2.1), let also ρ 0 and b R, then: ρz + b csn 1,q (ρ µ + b, ρ 2 σ, 1 d, ϑ, δ) ρ 6 Both state from the previous period and the disturbance are transformed by the linear transformation in state space models, but let us ignore this fact for the present argumentation (or assume this transformations are identities). 6
7 Corollary 2.9. Let z csn p,q ( µ, Σ, D, ϑ, ), p, q 1, for parameters as in definition (2.1), let also A R p p, A > 1, and b R p, then: Az + b csn p,q (A µ + b, A ΣA T, D ΣA 1, ϑ, ) Corollary Let z i csn p,qi ( µ i, Σ i, D i, ϑ i, i ), p, q i 1, i = 1, 2,..., n, for parameters as in definition (2.1), then n i=1 z i csn p, P n i=1 q i ( µ, Σ, D, ϑ, ), where: µ = n µ i, Σ = i=1 n Σ i, D = (Σ 1 D1 T,..., Σ n Dn T ) T (Σ ) 1, i=1 n n ϑ = (ϑ T 1, ϑ T 2,..., ϑ T n ) T, = + D Σ D [ D i Σi ]( Σ ) 1 [ D i Σi ] 1 for = n i=1 i, D = n i=1 D i and Σ = Σi, where operator, arbitrary matrices A and B, is defined as: A B = A 0 0 B 2.3 The regularization procedure The following proposition is central for providing analytical maximum likelihood estimation. i=1 i=1 Proposition 2.1. Let η be distributed according to a csn p,q for some p, q 1 with parameters µ η, Σ η 0, D η, ϑ η and η > 0. Let z t = Gη, A R p,p. Then, z t has a csn (possibly singular) distribution if and only if r(g T ) = r([g T w i ]) for all i = 1, 2,..., q, where r(g) denotes rank of G and w i denotes the i-th row D η. Proposition (2.1) states, that for a csn variable η, variable z t = Gη has a csn distribution if and only if rows of D η are linear combinations of rows of G. In other words, all rows of D η must belong to the image of A T, i.e. to span(a T ). This condition is always satisfied (regardless of D η ) if G has a full column rank. However, for a rank deficient operator G this is a very demanding condition, since D η can be in principle arbitrary. Although proposition (2.1) constitutes a negative result for η as a p-dimensional variable, it has to be stressed that it is assured that elements of η are csn distributed. 7
8 The correspondence between proposition (2.1) and the state space formulation with which we work goes as follows: z t = x t G = [A B] η = (x t 1, ξ t ) T therefore for the csn distribution to propagate, we need G to have full column rank. When one works with medium- or large-size DSGE models, the reduced form representation matrix G can be rank deficient. Since the following argument applies to full rank matrices G, one needs to reformulate the model so that G has full column rank, but the value of the likelihood function is unaffected. This turns out to be simple. If G is rank deficient, than some of the states are linear combination of the remaining ones, which means, that they can be substitutes out from the state-space representation both in the transition and in the measurement equation. This obviously does not affect the value of the likelihood function and, moreover, this can be done automatically, since a priori, i.e. before the reduced form has been calculated, it is difficult to judge which variables will be redundant. Such a regularization allows for further steps to apply. 3 The filter In this section I provide formulae for predictive densities p(y t Y t 1 ), p(x t Y t 1 ), p(ξ t Y t 1 ), p(ζ t Y t 1 ) and for filtration densities p(x t Y t ), p(ξ t Y t ), p(ζ t Y t ). All the densities depend on appropriate elements of the parameter vector θ = (θ F, θ H, θ ζ, θ A, θ F, θ ξ, θ x0 ). First, however, I derive unconditional distribution for states and observables. 3.1 Unconditional distributions In DSGE modeling it usually is the case that n ξ < p, hence Bξ t does not have a distribution function since matrix BΣ ξ B T is singular. Assume that n ξ < p and r(b) = n ξ. In this case Bξ t follows: Bξ t CSN p,q (µ B, Σ B, D B, ϑ B, B ) 8
9 where: µ B = Bµ ξ Σ B = BΣ ξ B T D B = D ξ (B T B) 1 B T (6) ϑ B = ϑ ξ B = ξ which is a (p, n ξ )-dimensional singular closed skewed normal distribution. This is true for every t = 1, 2,..., T. Since normal distribution is a special case of CSN distribution, it can be written that: x 0 CSN p,1 (µ x0, Σ x0, D x0, ϑ x0, x0 ) for 7 : µ x0 = x Σ x0 = Ψ x D x0 = 0 (7) ϑ x0 = 0 x0 = 1 Generally, i.e. for t = 1, 2,..., T, if: x t 1 CSN p,qt 1 (µ xt 1, Σ xt 1, D xt 1, ϑ xt 1, xt 1 ) which is true for t = 1, i.e. for x 0, with q 0 = 1, than Ax t 1 follows: Ax t 1 CSN p,qt 1 (µ A,t 1, Σ A,t 1, D A,t 1, ϑ A,t 1, A,t 1 ) where: µ A,t 1 = Aµ x,t 1 Σ A,t 1 = AΣ x,t 1 A T D A,t 1 = D x,t 1 Σ x,t 1 A T Σ 1 A,t 1 (8) ϑ A,t 1 = ϑ x,t 1 A,t 1 = x,t 1 + D x,t 1 Σ x,t 1 D T x,t 1 D x,t 1 Σ x,t 1 A T Σ 1 A,t 1 AΣ x,t 1D T x,t 1 7 x0 > 0 can be arbitrary. 9
10 which is a (p, q t 1 )-dimensional closed skewed normal distribution, and this distribution is non-singular since it is assumed that r(a) = p. This is true for t = 0 and also, by induction, for every = 1, 2,..., T. Induction is true, since: x t = Ax t 1 + Bξ t where Ax t 1 and Bξ t are independent random variables, from which it follows, that x t is distributed according to: x t CSN p,qt 1 +q(µ xt, Σ xt, D xt, ϑ xt, xt ) this means: x t CSN p,qt (µ xt, Σ xt, D xt, ϑ xt, xt ) for q t = q t 1 + q, where: µ x,t = µ A,t + µ B Σ x,t = Σ A,t + Σ B D x,t = [ Σ A,t DA,t, T Σ B DB T ] T Σ 1 x,t ϑ A,t = [ ϑ T A,t, ϑ T ] T B,t (9) x,t = A,t B + (D A,t D B )(Σ A,t Σ B )(D A,t D B ) T + (D A,t Σ A,t D B Σ B )(Σ x,t ) 1 (Σ A,t DA,t T Σ B DB) T Since q 0 = 1 and q t = q t 1 + q, we have q t = tq + q 0 = tq + 1 and: x t CSN p,tq+1 (µ xt, Σ xt, D xt, ϑ xt, xt ) For aplications we will work with q = 1, which yields q t = t + 1. So far we have derived formulae for distribution of states x t for all t = 1, 2,..., T. Now we will pin down the distribution of observables y t for all t = 1, 2,..., T, which is given by measurement equation: y t = F x t + Hu t where: u t N(0, Ψ u ) for all t = 0, 1, 2,..., T. Since normal distribution is a special case of CSN distribution, it can be written that: u t CSN nu,1(µ u, Σ u, D u, ϑ u, u ) 10
11 for 8 : µ u = 0 Σ u = Ψ u D u = 0 (10) ϑ u = 0 u = 1 In DSGE modeling, since measurement errors are introduced in order to avoid stochastic singularity, it usually is the case that n = n u + n ξ, and since n ξ > 0, it is the case that n u < n, hence Hu t does not have a distribution function since matrix HΨ u H T is singular. Assume that n u < n and r(h) = n u. In this case Hu t follows: Hu t CSN n,1 (µ H, Σ H, D H, ϑ H, H ) where: µ H = Hµ u Σ H = HΨ u H T D H = D u (H T H) 1 H T (11) ϑ H = ϑ u H = u for all t = 1, 2,..., T, which is a (n, 1)-dimensional singular normal distribution. Now, form distribution of x t it follows that: F x t CSN n,tq+1 (µ F,t, Σ F,t, D F,t, ϑ F,t, F,t ) where: µ F,t = F µ x,t Σ F,t = F Σ x,t F T D F,t = D x,t Σ x,t F T Σ 1 F,t (12) ϑ F,t = ϑ x,t 8 u > 0 can be arbitrary. F,t = x,t + D x,t Σ x,t D T x,t D x,t Σ x,t F T Σ 1 F,t F Σ x,td T x,t 11
12 which is a (p, tq + 1)-dimensional closed skewed normal distribution, and this distribution is non-singular since it is assumed that r(f ) = n. From the measurement equation it follows that y t is distributed according to: y t CSN n,tq+1 (µ yt, Σ yt, D yt, ϑ yt, yt ) or: where: y t CSN n,tq+2 (µ yt, Σ yt, D yt, ϑ yt, yt ) µ y,t = µ F,t + µ H Σ y,t = Σ F,t + Σ H D y,t = [ Σ F,t DF,t, T Σ H DH T ] T Σ 1 y,t ϑ y,t = [ ϑ T F,t, ϑ T ] T H,t (13) y,t = F,t H + (D F,t D H )(Σ F,t Σ H )(D F,t D H ) T + (D F,t Σ F,t D H Σ H ) (Σ y,t ) 1 ( Σ F,t DF,t T Σ H DH T ) 3.2 Conditional distributions For t = 1, 2,..., T, let us define an information set: Y t = {y 1, y 2,..., y t } which consists of realized values of observables up to time t. We want to derive the a posteriori distribution (x t Y t ) and we will do so on the basis of the joint distribution (x t, y t Y t 1 ). Assume, that the a posteriori distribution of states x t 1, i.e. conditional distribution of states x t 1 with respect to the information set Y t 1, therefore after observing y t 1, is given by: (x t 1 Y t 1 ) = (µ t 1, Σ t 1, D t 1, ϑ t 1, t 1 ) for some parameters µ t 1, Σ t 1, D t 1, ϑ t 1 and t 1. If so, the a priori, i.e. prior to observing y t, conditional distribution of x t is given by: (x t Y t 1 ) = (Ax t 1 + Bξ t Y t 1 ) = (Ax t 1 Y t 1 ) + Bξ t = = (µ, Σ, D, ϑ, ) 12
13 for: µ = µ A + µ B Σ = Σ A + Σ B D = [ Σ A DA, T Σ B DB T ] T Σ 1 ϑ = [ ϑ T A, 0 T ] T = A B + (D A D H )(Σ A Σ D )(D A D H ) T + (D A Σ A D H Σ H )(Σ) 1 (D A Σ A D H Σ H ) T where: µ A = Aµ t 1 Σ A = AΣ t 1 A T D A = D t 1 Σ t 1 A T Σ 1 A ϑ A = ϑ t 1 = t 1 + D t 1 Σ t 1 D T t 1 D t 1 Σ t 1 A T Σ 1 A AΣ t 1D T t 1 Information contained in the joint distribution (x t, y t Y t 1 ) is equivalent to the information contained in the distribution (x t, e t Y t 1 ) where: e t = y t E(y t Y t 1 ) which is approximated by 9 : e t = y t F µ Distribution of e t conditional on x t andy t 1 is normal: (e t x t, Y t 1 ) = (F x t + Hu t F µ x t, Y t 1 ) = (F (x t µ) + (Hu t Y t 1 )) = (F (x t µ) + Hu t ) = N(F (x t µ), Σ H ) Therefore joint distribution (x t, e t Y t 1 ) = (x t, y t Y t 1 ) is given by: (x t, e t Y t 1 ) = ( µ, Σ, D, ϑ, ) 9 Can be approximated explicitly from state-space eq-s for observed x t 1. 13
14 which parameters are approximated by: µ = µ 0 Σ = Σ ΣF T F T Σ F ΣF T + Σ H D = D ϑ = ϑ 0 (14) = Therefore, the a posteriori distribution is: (x t Y t ) = (µ t, Σ t, D t, ϑ t, t ) for: 4 Likelihood function µ t = µ + ΣF T (F T ΣF T + Σ D ) 1 (Y t F µ) Σ t = Σ ΣF T (F T ΣF T + Σ D ) 1 F Σ T D t = D ϑ t = ϑ t = The likelihood function of the state space model which we work with is given by: T L = p(y 0 ) p(y t Y t 1 ) t=2 Because Hu t is independent of Y t 1, using the measurement equation we see that the conditional a priori distribution of y t Y t 1 is: (y t Y t 1 ) = (F x t + Hu t Y t 1 ) = (F x t Y t 1 + Hu t ) Since distribution of x t Y t 1 is: x t Y t 1 (µ, Σ, D, ϑ, ) 14
15 distribution of y t Y t 1 = F x t Y t 1 + Hu t is: y t Y t 1 CSN n, (µ y, Σ y, D y, ϑ y, y ) for: µ y = µ F + µ H Σ y = Σ F + Σ H D y = [ Σ F DF T, Σ H DH T ] T Σ 1 y ϑ y = [ ϑ T F, ϑ T ] T H (15) y = F H + (D F D H )(Σ F Σ H )(D F D H ) T + (D F Σ F D H Σ H ) (Σ y ) 1 ( Σ F DF T Σ H DH T ) where: µ F = F µ x Σ F = F Σ x F T D F = D x Σ x F T Σ 1 F (16) ϑ F = ϑ x F = x + D x Σ x D T x D x Σ x F T Σ 1 F F Σ xd T x therefore: p(y t Y t 1 ) = φ p (y t ; µ y, Σ y ) Φ q(d y (y t µ y ); ϑ y, y ) Φ q (0; ϑ y, y + D y Σ y D T y ) (17) Numerical calculation of the likelihood function can be carried-out by means of the following sequential procedure: 1. Assume parameters of p(y 1 ) and parameters of p(x 1 Y 1 ). 2. For t = 2, 3,..., T implement the following steps: (a) For t > 2 take parameters of p(x t 1 Y t 1 ), i.e. µ t 1, Σ t 1, D t 1, ϑ t 1 and t 1, from the previous iteration, and for t = 2 take them from the initial assumption about p(x 1 Y 1 ). (b) Calculate parameters of p(x t Y t 1 ), i.e. µ, Σ, D, ϑ and. (c) Calculate parameters of p(y t Y t 1 ), i.e. µ y, Σ y, D y, ϑ y and y. (d) Update value of the likelihood function. (e) Calculate parameters of p(y t Y t ), i.e. µ t, Σ t, D t, ϑ t and t. 15
16 Value of the approximate likelihood function L = p(y θ) is now calculated for given θ, which can be implemented in any numerical optimization routine. References [Azalini and Dalla Valle (1996)] Azzalini, A. and A. Dalla Valle.: The Multivariate Skew-Normal Distribution, Biometrika (1996), Vol. 83, No. 4, pp [Azzalini and Capitanio (1999)] Azzalini, A. and A. Dalla Valle.: Applications of the Multivariate Skew Normal Distribution, Journal of the Royal Statistical Society. Series B (Statistical Methodology) (1999), Vol. 61, No. 3, pp [Genton (2004)] Genton, M. G. (ed.): Skew-elliptical distributions and their applications. A journel beyond normality., A CRC Press Company (2004). [Simon (2006)] Simon, D.: Optimal State Estimation. Kalman, H and Nonlinear Approches, John Willey & Sons (2006). 16
Smets and Wouters model estimated with skewed shocks - empirical study of forecasting properties
COLLEGIUM OF ECONOMIC ANALYSIS WORKING PAPER SERIES Smets and Wouters model estimated with skewed shocks - empirical study of forecasting properties Grzegorz Koloch SGH KAE Working Papers Series Number:
More informationLabor-Supply Shifts and Economic Fluctuations. Technical Appendix
Labor-Supply Shifts and Economic Fluctuations Technical Appendix Yongsung Chang Department of Economics University of Pennsylvania Frank Schorfheide Department of Economics University of Pennsylvania January
More informationECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering
ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering Lecturer: Nikolay Atanasov: natanasov@ucsd.edu Teaching Assistants: Siwei Guo: s9guo@eng.ucsd.edu Anwesan Pal:
More informationThe Multivariate Normal Distribution. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 36
The Multivariate Normal Distribution Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 1 / 36 The Moment Generating Function (MGF) of a random vector X is given by M X (t) = E(e t X
More informationNotes on Random Vectors and Multivariate Normal
MATH 590 Spring 06 Notes on Random Vectors and Multivariate Normal Properties of Random Vectors If X,, X n are random variables, then X = X,, X n ) is a random vector, with the cumulative distribution
More informationState-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Fin. Econometrics / 53
State-space Model Eduardo Rossi University of Pavia November 2014 Rossi State-space Model Fin. Econometrics - 2014 1 / 53 Outline 1 Motivation 2 Introduction 3 The Kalman filter 4 Forecast errors 5 State
More informationDynamic Factor Models and Factor Augmented Vector Autoregressions. Lawrence J. Christiano
Dynamic Factor Models and Factor Augmented Vector Autoregressions Lawrence J Christiano Dynamic Factor Models and Factor Augmented Vector Autoregressions Problem: the time series dimension of data is relatively
More informationThe Multivariate Normal Distribution. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 36
The Multivariate Normal Distribution Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 1 / 36 The Moment Generating Function (MGF) of a random vector X is given by M X (t) = E(e t X
More informationSKEW-NORMALITY IN STOCHASTIC FRONTIER ANALYSIS
SKEW-NORMALITY IN STOCHASTIC FRONTIER ANALYSIS J. Armando Domínguez-Molina, Graciela González-Farías and Rogelio Ramos-Quiroga Comunicación Técnica No I-03-18/06-10-003 PE/CIMAT 1 Skew-Normality in Stochastic
More informationResearch Division Federal Reserve Bank of St. Louis Working Paper Series
Research Division Federal Reserve Bank of St Louis Working Paper Series Kalman Filtering with Truncated Normal State Variables for Bayesian Estimation of Macroeconomic Models Michael Dueker Working Paper
More informationMultivariate Distributions
IEOR E4602: Quantitative Risk Management Spring 2016 c 2016 by Martin Haugh Multivariate Distributions We will study multivariate distributions in these notes, focusing 1 in particular on multivariate
More informationGaussian Process Approximations of Stochastic Differential Equations
Gaussian Process Approximations of Stochastic Differential Equations Cédric Archambeau Dan Cawford Manfred Opper John Shawe-Taylor May, 2006 1 Introduction Some of the most complex models routinely run
More informationECO 513 Fall 2008 C.Sims KALMAN FILTER. s t = As t 1 + ε t Measurement equation : y t = Hs t + ν t. u t = r t. u 0 0 t 1 + y t = [ H I ] u t.
ECO 513 Fall 2008 C.Sims KALMAN FILTER Model in the form 1. THE KALMAN FILTER Plant equation : s t = As t 1 + ε t Measurement equation : y t = Hs t + ν t. Var(ε t ) = Ω, Var(ν t ) = Ξ. ε t ν t and (ε t,
More informationKalman Filter and Parameter Identification. Florian Herzog
Kalman Filter and Parameter Identification Florian Herzog 2013 Continuous-time Kalman Filter In this chapter, we shall use stochastic processes with independent increments w 1 (.) and w 2 (.) at the input
More informationMeasurement Errors and the Kalman Filter: A Unified Exposition
Luiss Lab of European Economics LLEE Working Document no. 45 Measurement Errors and the Kalman Filter: A Unified Exposition Salvatore Nisticò February 2007 Outputs from LLEE research in progress, as well
More informationAn EM algorithm for Gaussian Markov Random Fields
An EM algorithm for Gaussian Markov Random Fields Will Penny, Wellcome Department of Imaging Neuroscience, University College, London WC1N 3BG. wpenny@fil.ion.ucl.ac.uk October 28, 2002 Abstract Lavine
More informationMultivariate ARMA Processes
LECTURE 8 Multivariate ARMA Processes A vector y(t) of n elements is said to follow an n-variate ARMA process of orders p and q if it satisfies the equation (1) A 0 y(t) + A 1 y(t 1) + + A p y(t p) = M
More informationBIOS 2083 Linear Models Abdus S. Wahed. Chapter 2 84
Chapter 2 84 Chapter 3 Random Vectors and Multivariate Normal Distributions 3.1 Random vectors Definition 3.1.1. Random vector. Random vectors are vectors of random variables. For instance, X = X 1 X 2.
More informationSwitching Regime Estimation
Switching Regime Estimation Series de Tiempo BIrkbeck March 2013 Martin Sola (FE) Markov Switching models 01/13 1 / 52 The economy (the time series) often behaves very different in periods such as booms
More informationECONOMETRIC METHODS II: TIME SERIES LECTURE NOTES ON THE KALMAN FILTER. The Kalman Filter. We will be concerned with state space systems of the form
ECONOMETRIC METHODS II: TIME SERIES LECTURE NOTES ON THE KALMAN FILTER KRISTOFFER P. NIMARK The Kalman Filter We will be concerned with state space systems of the form X t = A t X t 1 + C t u t 0.1 Z t
More informationNew Information Response Functions
New Information Response Functions Caroline JARDET (1) Banque de France Alain MONFORT (2) CNAM, CREST and Banque de France Fulvio PEGORARO (3) Banque de France and CREST First version : February, 2009.
More informationBayesian Inference for DSGE Models. Lawrence J. Christiano
Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Bayesian inference Bayes rule. Monte Carlo integation.
More informationThe Instability of Correlations: Measurement and the Implications for Market Risk
The Instability of Correlations: Measurement and the Implications for Market Risk Prof. Massimo Guidolin 20254 Advanced Quantitative Methods for Asset Pricing and Structuring Winter/Spring 2018 Threshold
More informationOn detection of unit roots generalizing the classic Dickey-Fuller approach
On detection of unit roots generalizing the classic Dickey-Fuller approach A. Steland Ruhr-Universität Bochum Fakultät für Mathematik Building NA 3/71 D-4478 Bochum, Germany February 18, 25 1 Abstract
More informationEmpirical Bayes Estimation in Multiple Linear Regression with Multivariate Skew-Normal Distribution as Prior
Journal of Mathematical Extension Vol. 5, No. 2 (2, (2011, 37-50 Empirical Bayes Estimation in Multiple Linear Regression with Multivariate Skew-Normal Distribution as Prior M. Khounsiavash Islamic Azad
More informationTime-Varying Parameters
Kalman Filter and state-space models: time-varying parameter models; models with unobservable variables; basic tool: Kalman filter; implementation is task-specific. y t = x t β t + e t (1) β t = µ + Fβ
More informationIntroduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak
Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak 1 Introduction. Random variables During the course we are interested in reasoning about considered phenomenon. In other words,
More informationWhat do we know about EnKF?
What do we know about EnKF? David Kelly Kody Law Andrew Stuart Andrew Majda Xin Tong Courant Institute New York University New York, NY April 10, 2015 CAOS seminar, Courant. David Kelly (NYU) EnKF April
More informationLecture Note 1: Probability Theory and Statistics
Univ. of Michigan - NAME 568/EECS 568/ROB 530 Winter 2018 Lecture Note 1: Probability Theory and Statistics Lecturer: Maani Ghaffari Jadidi Date: April 6, 2018 For this and all future notes, if you would
More informationState-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Financial Econometrics / 49
State-space Model Eduardo Rossi University of Pavia November 2013 Rossi State-space Model Financial Econometrics - 2013 1 / 49 Outline 1 Introduction 2 The Kalman filter 3 Forecast errors 4 State smoothing
More information... Econometric Methods for the Analysis of Dynamic General Equilibrium Models
... Econometric Methods for the Analysis of Dynamic General Equilibrium Models 1 Overview Multiple Equation Methods State space-observer form Three Examples of Versatility of state space-observer form:
More informationBayesian Inference for DSGE Models. Lawrence J. Christiano
Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Bayesian inference Bayes rule. Monte Carlo integation.
More informationMachine Learning 4771
Machine Learning 4771 Instructor: ony Jebara Kalman Filtering Linear Dynamical Systems and Kalman Filtering Structure from Motion Linear Dynamical Systems Audio: x=pitch y=acoustic waveform Vision: x=object
More informationCombining Macroeconomic Models for Prediction
Combining Macroeconomic Models for Prediction John Geweke University of Technology Sydney 15th Australasian Macro Workshop April 8, 2010 Outline 1 Optimal prediction pools 2 Models and data 3 Optimal pools
More information(Extended) Kalman Filter
(Extended) Kalman Filter Brian Hunt 7 June 2013 Goals of Data Assimilation (DA) Estimate the state of a system based on both current and all past observations of the system, using a model for the system
More informationErgodicity in data assimilation methods
Ergodicity in data assimilation methods David Kelly Andy Majda Xin Tong Courant Institute New York University New York NY www.dtbkelly.com April 15, 2016 ETH Zurich David Kelly (CIMS) Data assimilation
More informationChapter 2: Fundamentals of Statistics Lecture 15: Models and statistics
Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:
More informationLecture 15: Multivariate normal distributions
Lecture 15: Multivariate normal distributions Normal distributions with singular covariance matrices Consider an n-dimensional X N(µ,Σ) with a positive definite Σ and a fixed k n matrix A that is not of
More informationGeneralized Autoregressive Score Models
Generalized Autoregressive Score Models by: Drew Creal, Siem Jan Koopman, André Lucas To capture the dynamic behavior of univariate and multivariate time series processes, we can allow parameters to be
More information1 Appendix A: Matrix Algebra
Appendix A: Matrix Algebra. Definitions Matrix A =[ ]=[A] Symmetric matrix: = for all and Diagonal matrix: 6=0if = but =0if 6= Scalar matrix: the diagonal matrix of = Identity matrix: the scalar matrix
More informationLecture 11. Multivariate Normal theory
10. Lecture 11. Multivariate Normal theory Lecture 11. Multivariate Normal theory 1 (1 1) 11. Multivariate Normal theory 11.1. Properties of means and covariances of vectors Properties of means and covariances
More informationNonlinear and/or Non-normal Filtering. Jesús Fernández-Villaverde University of Pennsylvania
Nonlinear and/or Non-normal Filtering Jesús Fernández-Villaverde University of Pennsylvania 1 Motivation Nonlinear and/or non-gaussian filtering, smoothing, and forecasting (NLGF) problems are pervasive
More informationEstimation of moment-based models with latent variables
Estimation of moment-based models with latent variables work in progress Ra aella Giacomini and Giuseppe Ragusa UCL/Cemmap and UCI/Luiss UPenn, 5/4/2010 Giacomini and Ragusa (UCL/Cemmap and UCI/Luiss)Moments
More informationAccounting for Missing Values in Score- Driven Time-Varying Parameter Models
TI 2016-067/IV Tinbergen Institute Discussion Paper Accounting for Missing Values in Score- Driven Time-Varying Parameter Models André Lucas Anne Opschoor Julia Schaumburg Faculty of Economics and Business
More informationChapter 6. Maximum Likelihood Analysis of Dynamic Stochastic General Equilibrium (DSGE) Models
Chapter 6. Maximum Likelihood Analysis of Dynamic Stochastic General Equilibrium (DSGE) Models Fall 22 Contents Introduction 2. An illustrative example........................... 2.2 Discussion...................................
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 12: Gaussian Belief Propagation, State Space Models and Kalman Filters Guest Kalman Filter Lecture by
More information5: MULTIVARATE STATIONARY PROCESSES
5: MULTIVARATE STATIONARY PROCESSES 1 1 Some Preliminary Definitions and Concepts Random Vector: A vector X = (X 1,..., X n ) whose components are scalarvalued random variables on the same probability
More informationThe Size and Power of Four Tests for Detecting Autoregressive Conditional Heteroskedasticity in the Presence of Serial Correlation
The Size and Power of Four s for Detecting Conditional Heteroskedasticity in the Presence of Serial Correlation A. Stan Hurn Department of Economics Unversity of Melbourne Australia and A. David McDonald
More informationRecursive Estimation
Recursive Estimation Raffaello D Andrea Spring 08 Problem Set 3: Extracting Estimates from Probability Distributions Last updated: April 9, 08 Notes: Notation: Unless otherwise noted, x, y, and z denote
More informationThe Multivariate Normal Distribution 1
The Multivariate Normal Distribution 1 STA 302 Fall 2017 1 See last slide for copyright information. 1 / 40 Overview 1 Moment-generating Functions 2 Definition 3 Properties 4 χ 2 and t distributions 2
More informationBayesian inference for multivariate skew-normal and skew-t distributions
Bayesian inference for multivariate skew-normal and skew-t distributions Brunero Liseo Sapienza Università di Roma Banff, May 2013 Outline Joint research with Antonio Parisi (Roma Tor Vergata) 1. Inferential
More informationMultivariate Gaussian Distribution. Auxiliary notes for Time Series Analysis SF2943. Spring 2013
Multivariate Gaussian Distribution Auxiliary notes for Time Series Analysis SF2943 Spring 203 Timo Koski Department of Mathematics KTH Royal Institute of Technology, Stockholm 2 Chapter Gaussian Vectors.
More informationFormulas for probability theory and linear models SF2941
Formulas for probability theory and linear models SF2941 These pages + Appendix 2 of Gut) are permitted as assistance at the exam. 11 maj 2008 Selected formulae of probability Bivariate probability Transforms
More informationST5215: Advanced Statistical Theory
Department of Statistics & Applied Probability Monday, September 26, 2011 Lecture 10: Exponential families and Sufficient statistics Exponential Families Exponential families are important parametric families
More informationBayesian Estimation of DSGE Models
Bayesian Estimation of DSGE Models Stéphane Adjemian Université du Maine, GAINS & CEPREMAP stephane.adjemian@univ-lemans.fr http://www.dynare.org/stepan June 28, 2011 June 28, 2011 Université du Maine,
More informationFundamentals of Data Assimila1on
014 GSI Community Tutorial NCAR Foothills Campus, Boulder, CO July 14-16, 014 Fundamentals of Data Assimila1on Milija Zupanski Cooperative Institute for Research in the Atmosphere Colorado State University
More informationOptimal Policy with Partial Information in a Forward-Looking Model: Certainty-Equivalence Redux
SWCE206.tex Optimal Policy with Partial Information in a Forward-Looking Model: Certainty-Equivalence Redux Lars E.O. Svensson and Michael Woodford First draft: January 1999 This version: June 2002 Abstract
More informationA new Hierarchical Bayes approach to ensemble-variational data assimilation
A new Hierarchical Bayes approach to ensemble-variational data assimilation Michael Tsyrulnikov and Alexander Rakitko HydroMetCenter of Russia College Park, 20 Oct 2014 Michael Tsyrulnikov and Alexander
More informationThe Metropolis-Hastings Algorithm. June 8, 2012
The Metropolis-Hastings Algorithm June 8, 22 The Plan. Understand what a simulated distribution is 2. Understand why the Metropolis-Hastings algorithm works 3. Learn how to apply the Metropolis-Hastings
More informationLinear models. Linear models are computationally convenient and remain widely used in. applied econometric research
Linear models Linear models are computationally convenient and remain widely used in applied econometric research Our main focus in these lectures will be on single equation linear models of the form y
More informationThe Multivariate Normal Distribution 1
The Multivariate Normal Distribution 1 STA 302 Fall 2014 1 See last slide for copyright information. 1 / 37 Overview 1 Moment-generating Functions 2 Definition 3 Properties 4 χ 2 and t distributions 2
More informationBayesian Inference for DSGE Models. Lawrence J. Christiano
Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Preliminaries. Probabilities. Maximum Likelihood. Bayesian
More informationData assimilation in high dimensions
Data assimilation in high dimensions David Kelly Kody Law Andy Majda Andrew Stuart Xin Tong Courant Institute New York University New York NY www.dtbkelly.com February 3, 2016 DPMMS, University of Cambridge
More informationData assimilation in high dimensions
Data assimilation in high dimensions David Kelly Courant Institute New York University New York NY www.dtbkelly.com February 12, 2015 Graduate seminar, CIMS David Kelly (CIMS) Data assimilation February
More informationx. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).
.8.6 µ =, σ = 1 µ = 1, σ = 1 / µ =, σ =.. 3 1 1 3 x Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ ). The Gaussian distribution Probably the most-important distribution in all of statistics
More informationTail dependence in bivariate skew-normal and skew-t distributions
Tail dependence in bivariate skew-normal and skew-t distributions Paola Bortot Department of Statistical Sciences - University of Bologna paola.bortot@unibo.it Abstract: Quantifying dependence between
More informationLasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices
Article Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Fei Jin 1,2 and Lung-fei Lee 3, * 1 School of Economics, Shanghai University of Finance and Economics,
More informationPhysics 403. Segev BenZvi. Propagation of Uncertainties. Department of Physics and Astronomy University of Rochester
Physics 403 Propagation of Uncertainties Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Maximum Likelihood and Minimum Least Squares Uncertainty Intervals
More informationUC Berkeley Department of Electrical Engineering and Computer Sciences. EECS 126: Probability and Random Processes
UC Berkeley Department of Electrical Engineering and Computer Sciences EECS 6: Probability and Random Processes Problem Set 3 Spring 9 Self-Graded Scores Due: February 8, 9 Submit your self-graded scores
More informationRevision of the ECMWF humidity analysis: Construction of a gaussian control variable
Revision of the ECMWF humidity analysis: Construction of a gaussian control variable Elías Valur Hólm ECMWF, Shinfield Park, Reading GR2 9AX, United Kingdom 1 Introduction In the ECMWF analysis specific
More informationEstimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator
Estimation Theory Estimation theory deals with finding numerical values of interesting parameters from given set of data. We start with formulating a family of models that could describe how the data were
More informationQuadratic forms in skew normal variates
J. Math. Anal. Appl. 73 (00) 558 564 www.academicpress.com Quadratic forms in skew normal variates Arjun K. Gupta a,,1 and Wen-Jang Huang b a Department of Mathematics and Statistics, Bowling Green State
More informationIntroduction to Computational Finance and Financial Econometrics Probability Review - Part 2
You can t see this text! Introduction to Computational Finance and Financial Econometrics Probability Review - Part 2 Eric Zivot Spring 2015 Eric Zivot (Copyright 2015) Probability Review - Part 2 1 /
More informationRandom Vectors and Multivariate Normal Distributions
Chapter 3 Random Vectors and Multivariate Normal Distributions 3.1 Random vectors Definition 3.1.1. Random vector. Random vectors are vectors of random 75 variables. For instance, X = X 1 X 2., where each
More informationDynamic Identification of DSGE Models
Dynamic Identification of DSGE Models Ivana Komunjer and Serena Ng UCSD and Columbia University All UC Conference 2009 Riverside 1 Why is the Problem Non-standard? 2 Setup Model Observables 3 Identification
More informationCS281A/Stat241A Lecture 17
CS281A/Stat241A Lecture 17 p. 1/4 CS281A/Stat241A Lecture 17 Factor Analysis and State Space Models Peter Bartlett CS281A/Stat241A Lecture 17 p. 2/4 Key ideas of this lecture Factor Analysis. Recall: Gaussian
More informationLecture 3 Stationary Processes and the Ergodic LLN (Reference Section 2.2, Hayashi)
Lecture 3 Stationary Processes and the Ergodic LLN (Reference Section 2.2, Hayashi) Our immediate goal is to formulate an LLN and a CLT which can be applied to establish sufficient conditions for the consistency
More informationMassachusetts Institute of Technology Department of Economics Time Series Lecture 6: Additional Results for VAR s
Massachusetts Institute of Technology Department of Economics Time Series 14.384 Guido Kuersteiner Lecture 6: Additional Results for VAR s 6.1. Confidence Intervals for Impulse Response Functions There
More informationProbability Lecture III (August, 2006)
robability Lecture III (August, 2006) 1 Some roperties of Random Vectors and Matrices We generalize univariate notions in this section. Definition 1 Let U = U ij k l, a matrix of random variables. Suppose
More informationDSGE Methods. Estimation of DSGE models: Maximum Likelihood & Bayesian. Willi Mutschler, M.Sc.
DSGE Methods Estimation of DSGE models: Maximum Likelihood & Bayesian Willi Mutschler, M.Sc. Institute of Econometrics and Economic Statistics University of Münster willi.mutschler@uni-muenster.de Summer
More informationModel estimation through matrix equations in financial econometrics
Model estimation through matrix equations in financial econometrics Federico Poloni 1 Joint work with Giacomo Sbrana 2 1 Technische Universität Berlin (A. Von Humboldt postdoctoral fellow) 2 Rouen Business
More informationHaruhiko Ogasawara. This article gives the first half of an expository supplement to Ogasawara (2015).
Economic Review (Otaru University of Commerce, Vol.66, No. & 3, 9-58. December, 5. Expository supplement I to the paper Asymptotic expansions for the estimators of Lagrange multipliers and associated parameters
More informationX t = a t + r t, (7.1)
Chapter 7 State Space Models 71 Introduction State Space models, developed over the past 10 20 years, are alternative models for time series They include both the ARIMA models of Chapters 3 6 and the Classical
More information14 - Gaussian Stochastic Processes
14-1 Gaussian Stochastic Processes S. Lall, Stanford 211.2.24.1 14 - Gaussian Stochastic Processes Linear systems driven by IID noise Evolution of mean and covariance Example: mass-spring system Steady-state
More informationParameter identification using the skewed Kalman Filter
Parameter identification using the sewed Kalman Filter Katrin Runtemund, Gerhard Müller, U München Motivation and the investigated identification problem Problem: damped SDOF system excited by a sewed
More information01 Probability Theory and Statistics Review
NAVARCH/EECS 568, ROB 530 - Winter 2018 01 Probability Theory and Statistics Review Maani Ghaffari January 08, 2018 Last Time: Bayes Filters Given: Stream of observations z 1:t and action data u 1:t Sensor/measurement
More informationIdentifiability problems in some non-gaussian spatial random fields
Chilean Journal of Statistics Vol. 3, No. 2, September 2012, 171 179 Probabilistic and Inferential Aspects of Skew-Symmetric Models Special Issue: IV International Workshop in honour of Adelchi Azzalini
More informationDecentralized Stochastic Control with Partial Sharing Information Structures: A Common Information Approach
Decentralized Stochastic Control with Partial Sharing Information Structures: A Common Information Approach 1 Ashutosh Nayyar, Aditya Mahajan and Demosthenis Teneketzis Abstract A general model of decentralized
More informationVolatility. Gerald P. Dwyer. February Clemson University
Volatility Gerald P. Dwyer Clemson University February 2016 Outline 1 Volatility Characteristics of Time Series Heteroskedasticity Simpler Estimation Strategies Exponentially Weighted Moving Average Use
More informationLeast Squares Estimation-Finite-Sample Properties
Least Squares Estimation-Finite-Sample Properties Ping Yu School of Economics and Finance The University of Hong Kong Ping Yu (HKU) Finite-Sample 1 / 29 Terminology and Assumptions 1 Terminology and Assumptions
More informationPreliminaries. Copyright c 2018 Dan Nettleton (Iowa State University) Statistics / 38
Preliminaries Copyright c 2018 Dan Nettleton (Iowa State University) Statistics 510 1 / 38 Notation for Scalars, Vectors, and Matrices Lowercase letters = scalars: x, c, σ. Boldface, lowercase letters
More informationOnline appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US
Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Gerdie Everaert 1, Lorenzo Pozzi 2, and Ruben Schoonackers 3 1 Ghent University & SHERPPA 2 Erasmus
More informationPolynomial Chaos and Karhunen-Loeve Expansion
Polynomial Chaos and Karhunen-Loeve Expansion 1) Random Variables Consider a system that is modeled by R = M(x, t, X) where X is a random variable. We are interested in determining the probability of the
More informationMonetary and Exchange Rate Policy Under Remittance Fluctuations. Technical Appendix and Additional Results
Monetary and Exchange Rate Policy Under Remittance Fluctuations Technical Appendix and Additional Results Federico Mandelman February In this appendix, I provide technical details on the Bayesian estimation.
More informationChapter 5. Chapter 5 sections
1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationEnKF and filter divergence
EnKF and filter divergence David Kelly Andrew Stuart Kody Law Courant Institute New York University New York, NY dtbkelly.com December 12, 2014 Applied and computational mathematics seminar, NIST. David
More informationDSGE Model Forecasting
University of Pennsylvania EABCN Training School May 1, 216 Introduction The use of DSGE models at central banks has triggered a strong interest in their forecast performance. The subsequent material draws
More informationPart IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015
Part IB Statistics Theorems with proof Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly)
More information1 First-order di erence equation
References Hamilton, J. D., 1994. Time Series Analysis. Princeton University Press. (Chapter 1,2) The task facing the modern time-series econometrician is to develop reasonably simple and intuitive models
More informationStochastic Analogues to Deterministic Optimizers
Stochastic Analogues to Deterministic Optimizers ISMP 2018 Bordeaux, France Vivak Patel Presented by: Mihai Anitescu July 6, 2018 1 Apology I apologize for not being here to give this talk myself. I injured
More information