Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US
|
|
- Eleanore Robbins
- 5 years ago
- Views:
Transcription
1 Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Gerdie Everaert 1, Lorenzo Pozzi 2, and Ruben Schoonackers 3 1 Ghent University & SHERPPA 2 Erasmus University Rotterdam & Tinbergen Institute 3 National Bank of Belgium & Ghent University August 30, 2016 Abstract This online appendix provides technical details on the the estimation methodology used in the main paper, including an offset mixture representation for the stochastic volatility components, a general outline of the Gibbs sampler and details on the exact implementation for each of the Gibbs blocks. A1 Offset mixture representation for the stochastic volatility components Using ν t N 0, e hνt and µt N 0, e hµt together with equation 18 in the main paper, we can write ν t = e hνt/2 µ 1t and µ t = e hµt/2 µ 2t with µ 1t and µ 2t i.i.d. error terms with unit variance. A key feature of these stochastic volatility components is that they are nonlinear but can be transformed into linear components by taking the logarithm of their squares 2 ln e hνt/2 µ 1t = hνt + ln µ 1t 2, 2 ln e hµt/2 µ 2t = hµt + ln µ 2t 2, A-1 where ln µ 1t 2 and ln µ 2t 2 are log-chi-square distributed with expected value and variance Following Kim et al. 1998, we approximate the linear models in A-1 by an offset mixture time series model as g jt = h jt + ɛ jt, A-2 Corresponding author at: Department of Economics, P.O. Box 1738, 3000 DR Rotterdam, the Netherlands, Tel: , pozzi@ese.eur.nl, Website: 1
2 e h for j = ν, µ, where g νt = ln νt/2 2 e µ 1t + c h and g µt = ln µt/2 2 µ 2t + c, with c =.001 being an offset constant, and the distribution of ɛ jt given by the following mixture of normals f ɛ jt = M q n f N ɛjt m n , s 2 n, A-3 n=1 with component probabilities q n, means m n and variances s 2 j. Equivalently, this mixture density can be written in terms of the component indicator variable κ jt as ɛ jt κ jt = n N m n , s 2 n, with P r κjt = n = q n. A-4 Following Kim et al. 1998, we use a mixture of M = 7 normal distributions to make the approximation to the log-chi-square distribution sufficiently good. Values for {q n, m n, s 2 n} are provided by Kim et al. in their Table 4. A2 General outline of the Gibbs sampler Taken together, equations 14 and 26 in the main paper constitute the observation equations of a State Space SS model, with the unobserved states β it and h jt evolving according to the state equations and respectively. In a standard linear Gaussian SS model, the Kalman filter can be used to filter the unobserved states from the data and to construct the likelihood function such that the unknown parameters can be estimated using Maximum Likelihood ML. However, the stochastic volatility components introduced in Section 3.1 and the stochastic model specification search outlined in Section 3.2 imply a non-regular estimation problem for which the standard approach via the Kalman filter and ML is not feasible. Instead, we use the Gibbs sampler which is a Markov Chain Monte Carlo MCMC method to simulate draws from the intractable joint and marginal posterior distributions of the unknown parameters and the unobserved states using only tractable conditional distributions. Intuitively, this amounts to reducing the complex non-linear model into a sequence of blocks for subsets of parameters/states that are tractable conditional on the other blocks in the sequence. For notational convenience, define the non-centered time-varying parameter vector βt = β0t, β1t, the non-centered stochastic volatilities vector h t = h νt, hµt, the mixture indicators vector κt = κ νt, κ µt, the unknown parameter vectors θ = θ 1,..., θ q, φ β = β 00, β 10, γ, ρ, σ η0, σ η1, φ h = h ν0, h µ0, σ υν, σ υµ, φ = δ, θ, φ β, φ h and the model indicators M θ = ι θ1,..., ι θq, M φβ = ι β00, ι β10, ι γ, ι ρ, ι β0t, ι β1t, M φh = ι hν, ι hµ, M = M θ, M φβ, M φh. Further let D t = ln C t, ln Y t, Z t be the data vector. Stacking observations over time, we denote D = {D t } T t=1 and similarly for β, h and κ. The posterior density of interest is then given by f φ, β, h, κ, M D. Building on Frühwirth-Schnatter and Wagner 2010 for the stochastic model specification part, on Chib and Greenberg 1994 for the moving average MA part and on Kim et al for the stochastic volatility part, our MCMC scheme is as follows: 1. Sample the binary indicators M θ, M φβ and the constant parameters δ, θ, φ β conditional on the time-varying parameters β and on the stochastic volatilities h. This is accomplished as: 2
3 a Sample the first step parameters δ from f δ φ β, h, D. b Sample the binary indicators M θ and the MA coefficients θ: i. Sample the binary indicators M θ from f M θ δ, φ β, φ h, β, h, D marginalizing over the parameters θ for which variable selection is carried out. ii. Sample the unrestricted parameters in θ from f θ δ, φ β, φ h, β, h, M θ, D while setting the restricted parameters in θ for which the corresponding binary indicator in M θ is 0 equal to 0. c Sample the binary indicators M φβ and the parameters φ β : i. Sample the binary indicators M φβ from f M φβ δ, θ, φ h, β, h, D marginalizing over the parameters φ β for which variable selection is carried out. ii. Sample the unrestricted parameters in φ β from f φ β δ, θ, φ h, β, h, M φβ, D while setting the restricted parameters in φ β for which the corresponding binary indicator in M φβ 0 equal to 0. is 2. Sample the unrestricted time-varying parameters in β from fβ δ, θ, φ β, h, M φβ, D. The restricted time-varying parameters in β sampled directly from their prior distribution in equation 23. for which the corresponding binary indicator is 0 are 3. Sample the mixture indicators κ, the binary indicators M φh, the constant parameters φ h and the stochastic volatilities components h conditional on the constant parameters δ, θ, φ β and on the time-varying parameters β. This is accomplished as: a Sample the mixture indicators κ from f κ φ, β, h, D. 1 b Sample the binary indicators M φh and the constant parameters φ h conditional on the constant parameters δ, θ, φ β, on the time-varying parameters β, on the stochastic volatility components h and on the mixture indicators κ: i. Sample the binary indicators M φh from f M φh δ, θ, φ β, β, h, κ, D marginalizing over the parameters φ h for which variable selection is carried out. ii. Sample the unrestricted parameters in φ h from f φ h δ, θ, φ β, β, h, κ, M φh, D while setting the restricted parameters in φ h for which the corresponding binary indicator in M φh is 0 equal to 0. c Sample the unrestricted stochastic volatilities in h from fh φ, β, κ, M φh, D. The restricted stochastic volatilities in h for which the corresponding binary indicator is 0 are sampled directly from their prior distribution in equation Note that the ordering of the Gibbs steps is in line with Del Negro and Primiceri 2015 who argue that the mixture indicators κ should be drawn after the components that do not condition on κ directly but before the components that do condition on κ. 3
4 4. Perform a random sign switch for σ ηi and {βit }T t=1 and for σ υj and {h jt }T t=1, e.g., σ η1 and {β1t} T t=1 are left unchanged with probability 0.5 while with the same probability they are replaced by σ η1 and { β1t} T t=1. Given an arbitrary set of starting values, sampling from these blocks is iterated J times and, after a sufficiently large number of burn-in draws B, the sequence of draws B + 1,..., J approximates a sample from the virtual posterior distribution f φ, β, h, κ, M D. Details on the exact implementation of each of the blocks can be found in Section A3. The results reported in the main paper are based on iterations, with the first draws discarded as a burn-in sequence. A3 Details on the blocks in the Gibbs sampling algorithm In this section we provide details on the exact implementation of the Gibbs sampling algorithm outlined in Section A2 to jointly sample the binary indicators M, the constant parameters φ, the time-varying parameters β, the stochastic volatilities h and the mixture indicators κ. For notational convenience, let us first define a general regression model y = x m b m + θ L e, e N 0, Σ, A-5 where y = y 1,..., y T is a T 1 vector stacking observations on the dependent variable y t, x = x 1,..., x T is a T k unrestricted predictor matrix with typical row x t, b is a k 1 unrestricted parameter vector, θ L is a lag polynomial of order q and Σ is a diagonal matrix with elements σet 2 on the diagonal that may vary over time to allow for heteroskedasticity of a known form. The restricted predictor matrix x m and restricted parameter vector b m exclude those elements in x and b for which the corresponding binary indicator in the m is zero, with m being a subset of the model M. The MAq errors in equation A-5 imply a model that is non-linear in the parameters. As suggested by Ullah et al and Chib and Greenberg 1994, conditional on θ a linear model can be obtained from a recursive transformation of the data. For t = 1,... T let ỹ t = y t θ i ỹ t i, with ỹ t = 0 for t 0, A-6 x t = x t and further for j = 1,..., q θ i x t i, with x t = 0 for t 0, A-7 ω jt = θ i ω j,t i + θ t+j 1, with ω jt = 0 for t 0, A-8 where θ s = 0 for s > q. Equation A-5 can then be transformed as ỹ = x m b m + ωλ + e = w m Φ m + e, A-9 4
5 where w = x, ω with ω = ω 1,..., ω T and ω t = ω 1t,..., ω qt, Φ m = b m, λ and λ = e0,..., e q+1 initial conditions that can be estimated as unknown parameters. Conditional on θ and σ 2 et, equation A-9 is a standard linear regression with observed variables ỹ and w m and heteroskedastic normal errors with known covariance matrix Σ. A naive implementation of the Gibbs sampler would be to first sample the binary indicators in m from f m Φ, Σ, ỹ, w and next Φ m from f Φ m m, Σ, ỹ, w. However, this approach does not result in an irreducible Markov chain as whenever an indicator in m equals zero, the corresponding coefficient in Φ is also zero which implies that the chain has absorbing states. Therefore, as in Frühwirth-Schnatter and Wagner 2010 we marginalize over the parameters Φ when sampling m and next draw Φ m conditional on the binary indicators in m. The posterior distribution f m Σ, ỹ, w can be obtained using Bayes Theorem as f m Σ, ỹ, w f ỹ m, Σ, w p m, A-10 with p m being the prior probability of m and f ỹ m, Σ, w the marginal likelihood of the regression model A-9 where the effect of the parameters Φ has been integrated out. Under the normal conjugate prior p Φ m = N Φ m 0, V m 0, A-11 the closed form solution of the marginal likelihood is given by f ỹ m, Σ, w Σ 0.5 VT m 0.5 ỹ Σ 1 ỹ + Φ m 0 exp V0 m 1 Φ m 0 Φm T VT m 1 Φ m T, A-12 V0 m where the posterior moments Φ m T and V m T can be calculated as Φ m T = VT m w m Σ 1 ỹ + V0 m 1 Φ m 0, A-13 VT m = w m Σ 1 w m + V0 m 1 1. A-14 Following George and McCulloch 1993 we use a single-move sampler in which the binary indicators ι k in m are sampled recursively from the Bernoulli distribution with probability p ι k = 1 ι k, Σ, ỹ, w = f ι k = 1 ι k, Σ, ỹ, w f ι k = 0 ι k, Σ, ỹ, w + f ι k = 1 ι k, Σ, ỹ, w, A-15 for k depending on the particular subset m of M. We further randomize over the sequence in which the binary indicators are drawn. Conditional on m, the posterior distribution of Φ m is given by p Φ m m, Σ, ỹ, w = N Φ m T, V m T, A-16 with posterior moments b m T and V m T given in equations A-13-A-14. 5
6 Block 1: Sampling M θ, M φβ and δ, θ, φ β Block 1a: Sampling the first step parameters δ Conditional on the stochastic volatility component h νt, equation 14 can be written in the general notation of equation A-5 as: y t = ln Y t, x m t = x t = Z t, b m = b = δ and θl = 1 such that θle t = ν t and the elements in Σ are given by σ 2 et = e hνt. In the notation of equation A-9 we further have ỹ t = y t, w m t = w t = x t and Φ m = Φ = δ. The parameters in δ can then be sampled from the posterior distribution in equation A-16 and used to calculate E t 1 ln Y t = Z t δ, ν t = ln Y t Z t δ and ν t = θ L ν te hµt/2 1 ρ2 e hνt/2, conditional on θ, ρ, h νt and h µt. Block 1b: Sampling the binary indicators M θ and the MA coefficients θ Conditional on the parameters δ, φ h and φ β, on the time-varying coefficients β t, on the stochastic volatility h µt and on the binary indicators M φβ, equation 16 can be written in the general notation of equation A-5 as: y t = ln C t, x t = 1, Z t δ, β 0t, β 1tZ t δ, ln C t 1, b = β 00, β 10, σ η0, σ η1, γ and θle t = θlε t, such that the elements in Σ are given by σ 2 et = σ 2 εt = e hµt / 1 ρ 2. The values of the binary indicators in the subset M φβ of M then imply the restricted x M φ β t and b M φ β. Under the normal conjugate prior p θ = N b θ 0, V0 θ, the exact conditional distribution of θ is p θ Φ M φ β, Σ, y, x T exp e t θ 2 2σet 2 t=1 exp 1 θ b θ 2 0 V θ 1 0 θ b θ 0, A-17 where e t θ = ỹ t θ w t m θ Φ M φ β is calculated from the transformed model in equation A-9 further conditioning on the initial conditions λ to obtain Φ M φ β = b M. φ β, λ Direct sampling of θ using equation A-17 is not possible, though, as e t θ is a non-linear function of θ. To solve this issue, Chib and Greenberg 1994 propose to linearize e t θ around θ using a first-order Taylor expansion e t θ e t θ Ψ t θ θ, A-18 where Ψ t = Ψ 1t,..., Ψ qt is a 1 q vector including the first-order derivatives of e t θ evaluated at θ obtained using the following recursion Ψ it = e t i θ θj Ψ i,t j, j=1 A-19 where Ψ it = 0 for t 0. An adequate approximation can be obtained by choosing θ to be the non-linear 6
7 least squares estimate of θ conditional on the other parameters in the model, which can be obtained as θ = argmin θ T e t θ 2, A-20 t=1 Conditioning on θ, equation A-18 can then be rewritten as an approximate linear regression model e t θ + Ψ t θ Ψ t θ + e t θ, A-21 with dependent variable e t θ +Ψ t θ and explanatory variables Ψ t. As such, the approximate expression in A-21 can be written in the general notation of equation A-5 by setting y t = e t θ +Ψ t θ, x t = Ψ t, b = θ and θl = 1 such that θle t = ε t and σ 2 et = σ 2 εt = e hµt / 1 ρ 2. The values of the binary indicators in the subset M θ of M imply the restricted x M θ t further have ỹ t = y t, w m t = x M θ t and Φ m = b M θ. and b M θ. In the notation of equation A-9 we The binary indicators in M θ can then be sampled from the posterior distribution in equation A-10 with marginal likelihood calculated from equation A-12. Next, we sample θ M θ using a Metropolis- Hastings MH step. Suppose θ M θ i is the current draw in the Markov chain. To obtain the next draw θ M θ i + 1, first draw a candidate θ M θ c using the following normal proposal distribution q θ M θ θ, Σ, y, x N Φ M θ T, V M θ T, A-22 with posterior moments Φ M θ T and V M θ T calculated using equations A-13-A-14. The MH step then implies a further randomization which amounts to accepting the candidate draw θ M θ c with probability { α θ M θ i, θ M θ p θ M θ c Φ M θ, Σ, y, x q θ M θ i θ, Σ, y, x } c = min p θ M θ i Φ M θ, Σ, y, x q θ M θ c θ, Σ, y, x, 1. A-23 If θ M θ c is accepted, θ M θ i + 1 is set equal to θ M θ c while if θ M θ c is rejected, θ M θ i + 1 is set equal to θ M θ i. Note that the unrestricted θ is restricted to obtain θ M θ by excluding θ l when ι θl = 0. In this case θ l is not sampled but set equal to zero. Block 1c: Sampling the binary indicators M φβ and the second step parameters φ β Conditional on β t, δ and ν t, equation 26 can be written in the general notation of equation A-5 setting y t = ln C t, x t = 1, Z t δ, β 0t, β 1tZ t δ, ln C t 1, ν t, b = β 00, β 01, σ η0, σ η1, γ, ρ and θle t = θlµ t, such that the elements in Σ are given by σ 2 et = e hµt. Further conditioning on the MA parameters θ, the unrestricted transformed variables ỹ t and w t in equation A-9 are obtained, with corresponding unrestricted extended parameter vector Φ = b, λ. The values of the binary indicators in M φβ imply the restricted w M φ β t and Φ M φ β. First, the binary indicators in M φβ are sampled from the posterior distribution in equation A-10 with marginal likelihood calculated from equation A-12. Second, conditional on M φβ, φ β is sampled, together with λ, by drawing Φ M φ β from the general expression in equation A-16. Note that the unrestricted Φ = β 00, β 10, σ η0, σ η1, γ, ρ, λ is restricted to obtain Φ M φ β by excluding those coefficients for which the then 7
8 corresponding binary indicator in M φβ is zero. These restricted coefficients are not sampled but set equal to zero. Block 2: Sampling β In this block we use the forward-filtering and backward-sampling approach of Carter and Kohn 1994 and De Jong and Shephard 1995 to sample the time-varying parameters β. Conditional on the coefficients φ β, on the stochastic volatility h µt, on the first block results Z t δ and ν t M φβ, equation 26 can be rewritten as and on the binary indicators y t = ι β0t σ η0 β 0t + ι β1t σ η1 β 1tx 1t + θ L µ t, A-24 with y t = ln C t β 00 β 10 Z t δ γ ln C t 1 ρνt and x 1t = Z t δ. Using the recursive transformation suggested by Ullah et al and Chib and Greenberg 1994, the model in equation A-24 can be transformed to ỹ t = ι β0t σ η0 β0t + ι β1t σ η1 β1t + ω t λ + µ t, A-25 where ỹ t and ω t = ω 1t,..., ω qt are calculated conditional on θ from equations A-6 and A-8 and similarly β 0t = β0t θ i β0,t i, with β0t = 0 for t 0, A-26 β 1t = β 1tx 1t θ i β1,t i, with β1t = 0 for t 0. A-27 Substituting equation 23 in A-26-A-27 yields β 0,t+1 = β0t θ i β0,t+1 i + η0t, β 1,t+1 = β 1tx 1,t+1 θ i β1,t+1 i + x 1,t+1 η1t, A-28 A-29 such that the state space representation of the model in equations A-25, 23 and A-28-A-29 is given by σ η {{}}{ }} ]{ ] α0t ỹ t ω t λ = 0 σ η σ η µ t, A-30 α0,t+1 α 1,t+1 ] }{{} α t+1 = T0t 0 ] α0t α 1t ] 0 T 1t }{{}}{{} T t α t + K0t 0 } 0 K 1t {{ } K t η 1t α t α 1t ] ] η 0t, A-31 }{{} η t 8
9 with α i,t+1 given by β i,t+1 β i,t+1 β it. β i,t q 2 }{{} α i,t x i,t+1 θ 1... θ q 1 θ q = β it β it. β i,t q 2 β i,t q }{{}}{{} T it α it + 1 x i,t+1 0. } 0 {{ } K it η it ], A-32 for i = 0, 1 and where x 0t = 1 t, µ t N 0, e hµt and η it N 0, 1. In line with equations 23 and A-26-A-27, each of the states in α t is initialized at zero. Equations A-30-A-31 constitute a standard linear Gaussian state space model, from which the unknown state variables α t can be filtered using the standard Kalman filter. Sampling α t from its conditional distribution can then be done using the multimove simulation smoother of Carter and Kohn 1994 and De Jong and Shephard Using β i0, σ ηi and βit, the centered time-varying coefficients β it in equation 20 can easily be reconstructed from equation 22. Note that in a restricted model with ι βit = 0, σ ηi is excluded from σ η and α it is dropped from the state vector α t. In this case, no forward-filtering and backward-sampling for β it distribution using equation 23. is needed as this can be sampled directly from its prior Block 3: Sampling κ, M φh, φ h and h Block 3a: Sampling the mixture indicators κ Conditional on h jt and on ɛ jt, calculated from equation A-2 as ɛ jt = g jt h jt, we first sample the mixture indicators κ jt from the conditional probability mass p κ jt = n h jt, ɛ jt q n f N ɛjt h jt + m n , s 2 n, A-33 for j = ν, µ. Block 3b: Sampling the binary indicators M φh and the constant parameters φ h Conditional on h jt, g jt and κ jt and using the non-centered parameterization in equation 27, equation A-2 can be rewritten in the general linear regression format of A-5 setting y t = g jt m κjt , x t = 1, h jk, b = h j0, σ υj and θle t = ɛ jt = ɛ jt m κjt for j = ν, µ. Given the mixture distribution of ɛ jt defined in equation A-4, the centered error term ɛ jt has a heteroskedastic variance s 2 κ jt such that Σ =diags 2 κ j1,..., s 2 κ jt. In the notation of equation A-9 we further have ỹ t = y t and the unrestricted w t = x t and Φ = b. The values of the binary indicators in M φh w M φ h t then imply the restricted and Φ M φ h. Hence, the binary indicators ι hj, for j = ν, µ, in M φh can be sampled from the posterior distribution in equation A-10 with marginal likelihood calculated from equation A-12. Next conditioning on M φh, the parameters h j0, σ υj, for j = ν, µ in φ h are sampled by drawing Φ M φ h using the general expression in equation A-16. Note that the unrestricted Φ = h j0, σ υj is restricted 9
10 to obtain Φ M φ h by excluding σ υj when ι hj = 0. In this case σ υj is not sampled but set equal to zero. Block 3c: Sampling the stochastic volatilities h Conditional on the transformed errors g jt defined under equation A-2, on the mixture indicators κ jt and on the parameters φ h, equations A-2 and can we written in the following conditional state space representation g jt m κjt ] h j0 = ] = h j,t+1 ] σ υj ] 1 h jt ] ] + ɛ jt, A-34 ] ] ] + 1, A-35 h jt υ jt for j = ν, µ, where ɛ jt = ɛ jt m κjt is ɛ jt centered around zero with var ɛ jt = s 2 κ jt and varυjt = 1. In line with equation 25, the random walk component h jt is initialized at zero. As equations A-34-A-35 constitute a standard linear Gaussian state space model, h jt can be filtered using the Kalman filter and sampled using the multimove simulation smoother of Carter and Kohn 1994 and De Jong and Shephard Note that in a restricted model with ι hj = 0, and hence σ υj = 0, no forward-filtering and backward-sampling for h jt is needed as this can be sampled directly from its prior distribution in equation 25. Using h j0, σ υj and h jt, the centered stochastic volatility component h jt in equation 21 can easily be reconstructed from equation
11 References Carter, C. and Kohn, R On gibbs sampling for state space models. Biometrika, 81: Chib, S. and Greenberg, E Bayes inference in regression models with ARMAp,q errors. Journal of Econometrics, 64: De Jong, P. and Shephard, N : The simulation smoother fro time series models. Biometrika, Del Negro, M. and Primiceri, G Time varying structural vector autoregressions and monetary policy: a corrigendum. The review of economic studies, 82: Frühwirth-Schnatter, S. and Wagner, H Stochastic model specification search for Gaussian and partial non-gaussian state space models. Journal of Econometrics, 1541: George, E. and McCulloch, R Variable selection via gibbs sampling. Journal of the American Statistical Association, 88: Kim, S., Shephard, N., and Chib, S Stochastic Volatility: Likelihood Inference and Comparison with ARCH Models. Review of Economic Studies, 653: Ullah, A., Vinod, H., and Singh, R Estimation of linear models with moving average disturbances. Journal of Quantitative Economics, 2:
Lecture 4: Dynamic models
linear s Lecture 4: s Hedibert Freitas Lopes The University of Chicago Booth School of Business 5807 South Woodlawn Avenue, Chicago, IL 60637 http://faculty.chicagobooth.edu/hedibert.lopes hlopes@chicagobooth.edu
More informationThe Effects of Monetary Policy on Stock Market Bubbles: Some Evidence
The Effects of Monetary Policy on Stock Market Bubbles: Some Evidence Jordi Gali Luca Gambetti ONLINE APPENDIX The appendix describes the estimation of the time-varying coefficients VAR model. The model
More informationResearch Division Federal Reserve Bank of St. Louis Working Paper Series
Research Division Federal Reserve Bank of St Louis Working Paper Series Kalman Filtering with Truncated Normal State Variables for Bayesian Estimation of Macroeconomic Models Michael Dueker Working Paper
More informationNon-Markovian Regime Switching with Endogenous States and Time-Varying State Strengths
Non-Markovian Regime Switching with Endogenous States and Time-Varying State Strengths January 2004 Siddhartha Chib Olin School of Business Washington University chib@olin.wustl.edu Michael Dueker Federal
More informationTimevarying VARs. Wouter J. Den Haan London School of Economics. c Wouter J. Den Haan
Timevarying VARs Wouter J. Den Haan London School of Economics c Wouter J. Den Haan Time-Varying VARs Gibbs-Sampler general idea probit regression application (Inverted Wishart distribution Drawing from
More informationDynamic Factor Models and Factor Augmented Vector Autoregressions. Lawrence J. Christiano
Dynamic Factor Models and Factor Augmented Vector Autoregressions Lawrence J Christiano Dynamic Factor Models and Factor Augmented Vector Autoregressions Problem: the time series dimension of data is relatively
More informationBayesian Model Comparison:
Bayesian Model Comparison: Modeling Petrobrás log-returns Hedibert Freitas Lopes February 2014 Log price: y t = log p t Time span: 12/29/2000-12/31/2013 (n = 3268 days) LOG PRICE 1 2 3 4 0 500 1000 1500
More informationThe Kalman filter, Nonlinear filtering, and Markov Chain Monte Carlo
NBER Summer Institute Minicourse What s New in Econometrics: Time Series Lecture 5 July 5, 2008 The Kalman filter, Nonlinear filtering, and Markov Chain Monte Carlo Lecture 5, July 2, 2008 Outline. Models
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More informationSession 5B: A worked example EGARCH model
Session 5B: A worked example EGARCH model John Geweke Bayesian Econometrics and its Applications August 7, worked example EGARCH model August 7, / 6 EGARCH Exponential generalized autoregressive conditional
More informationECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS
ECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS 1. THE CLASS OF MODELS y t {y s, s < t} p(y t θ t, {y s, s < t}) θ t = θ(s t ) P[S t = i S t 1 = j] = h ij. 2. WHAT S HANDY ABOUT IT Evaluating the
More informationThe Metropolis-Hastings Algorithm. June 8, 2012
The Metropolis-Hastings Algorithm June 8, 22 The Plan. Understand what a simulated distribution is 2. Understand why the Metropolis-Hastings algorithm works 3. Learn how to apply the Metropolis-Hastings
More informationBayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence
Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns
More informationSession 3A: Markov chain Monte Carlo (MCMC)
Session 3A: Markov chain Monte Carlo (MCMC) John Geweke Bayesian Econometrics and its Applications August 15, 2012 ohn Geweke Bayesian Econometrics and its Session Applications 3A: Markov () chain Monte
More informationKatsuhiro Sugita Faculty of Law and Letters, University of the Ryukyus. Abstract
Bayesian analysis of a vector autoregressive model with multiple structural breaks Katsuhiro Sugita Faculty of Law and Letters, University of the Ryukyus Abstract This paper develops a Bayesian approach
More informationBayesian Inference for DSGE Models. Lawrence J. Christiano
Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Bayesian inference Bayes rule. Monte Carlo integation.
More informationF denotes cumulative density. denotes probability density function; (.)
BAYESIAN ANALYSIS: FOREWORDS Notation. System means the real thing and a model is an assumed mathematical form for the system.. he probability model class M contains the set of the all admissible models
More informationMarkov Chain Monte Carlo Methods
Markov Chain Monte Carlo Methods John Geweke University of Iowa, USA 2005 Institute on Computational Economics University of Chicago - Argonne National Laboaratories July 22, 2005 The problem p (θ, ω I)
More informationCore Inflation and Trend Inflation. Appendix
Core Inflation and Trend Inflation Appendix June 2015 (Revised November 2015) James H. Stock Department of Economics, Harvard University and the National Bureau of Economic Research and Mark W. Watson
More informationNon-homogeneous Markov Mixture of Periodic Autoregressions for the Analysis of Air Pollution in the Lagoon of Venice
Non-homogeneous Markov Mixture of Periodic Autoregressions for the Analysis of Air Pollution in the Lagoon of Venice Roberta Paroli 1, Silvia Pistollato, Maria Rosa, and Luigi Spezia 3 1 Istituto di Statistica
More informationECO 513 Fall 2008 C.Sims KALMAN FILTER. s t = As t 1 + ε t Measurement equation : y t = Hs t + ν t. u t = r t. u 0 0 t 1 + y t = [ H I ] u t.
ECO 513 Fall 2008 C.Sims KALMAN FILTER Model in the form 1. THE KALMAN FILTER Plant equation : s t = As t 1 + ε t Measurement equation : y t = Hs t + ν t. Var(ε t ) = Ω, Var(ν t ) = Ξ. ε t ν t and (ε t,
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project
More informationWorking Paper Series
Discussion of Cogley and Sargent s Drifts and Volatilities: Monetary Policy and Outcomes in the Post WWII U.S. Marco Del Negro Working Paper 23-26 October 23 Working Paper Series Federal Reserve Bank of
More informationBayesian Inference for DSGE Models. Lawrence J. Christiano
Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Preliminaries. Probabilities. Maximum Likelihood. Bayesian
More informationPart I State space models
Part I State space models 1 Introduction to state space time series analysis James Durbin Department of Statistics, London School of Economics and Political Science Abstract The paper presents a broad
More informationBayesian Monte Carlo Filtering for Stochastic Volatility Models
Bayesian Monte Carlo Filtering for Stochastic Volatility Models Roberto Casarin CEREMADE University Paris IX (Dauphine) and Dept. of Economics University Ca Foscari, Venice Abstract Modelling of the financial
More informationApril 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning
for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions
More informationComputational statistics
Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated
More informationPoint, Interval, and Density Forecast Evaluation of Linear versus Nonlinear DSGE Models
Point, Interval, and Density Forecast Evaluation of Linear versus Nonlinear DSGE Models Francis X. Diebold Frank Schorfheide Minchul Shin University of Pennsylvania May 4, 2014 1 / 33 Motivation The use
More informationOutline. Clustering. Capturing Unobserved Heterogeneity in the Austrian Labor Market Using Finite Mixtures of Markov Chain Models
Capturing Unobserved Heterogeneity in the Austrian Labor Market Using Finite Mixtures of Markov Chain Models Collaboration with Rudolf Winter-Ebmer, Department of Economics, Johannes Kepler University
More informationStatistical Inference and Methods
Department of Mathematics Imperial College London d.stephens@imperial.ac.uk http://stats.ma.ic.ac.uk/ das01/ 31st January 2006 Part VI Session 6: Filtering and Time to Event Data Session 6: Filtering and
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 12: Gaussian Belief Propagation, State Space Models and Kalman Filters Guest Kalman Filter Lecture by
More informationNon-Markovian Regime Switching with Endogenous States and Time-Varying State Strengths
WORKING PAPER SERIES Non-Markovian Regime Switching with Endogenous States and Time-Varying State Strengths Siddhartha Chib and Michael Dueker Working Paper 2004-030A http://research.stlouisfed.org/wp/2004/2004-030.pdf
More informationGibbs Sampling in Linear Models #2
Gibbs Sampling in Linear Models #2 Econ 690 Purdue University Outline 1 Linear Regression Model with a Changepoint Example with Temperature Data 2 The Seemingly Unrelated Regressions Model 3 Gibbs sampling
More informationModeling conditional distributions with mixture models: Theory and Inference
Modeling conditional distributions with mixture models: Theory and Inference John Geweke University of Iowa, USA Journal of Applied Econometrics Invited Lecture Università di Venezia Italia June 2, 2005
More informationControlled sequential Monte Carlo
Controlled sequential Monte Carlo Jeremy Heng, Department of Statistics, Harvard University Joint work with Adrian Bishop (UTS, CSIRO), George Deligiannidis & Arnaud Doucet (Oxford) Bayesian Computation
More informationDSGE MODELS WITH STUDENT-t ERRORS
Econometric Reviews, 33(1 4):152 171, 2014 Copyright Taylor & Francis Group, LLC ISSN: 0747-4938 print/1532-4168 online DOI: 10.1080/07474938.2013.807152 DSGE MODELS WITH STUDENT-t ERRORS Siddhartha Chib
More informationInfinite-State Markov-switching for Dynamic. Volatility Models : Web Appendix
Infinite-State Markov-switching for Dynamic Volatility Models : Web Appendix Arnaud Dufays 1 Centre de Recherche en Economie et Statistique March 19, 2014 1 Comparison of the two MS-GARCH approximations
More informationLikelihood-free MCMC
Bayesian inference for stable distributions with applications in finance Department of Mathematics University of Leicester September 2, 2011 MSc project final presentation Outline 1 2 3 4 Classical Monte
More informationSpatio-temporal precipitation modeling based on time-varying regressions
Spatio-temporal precipitation modeling based on time-varying regressions Oleg Makhnin Department of Mathematics New Mexico Tech Socorro, NM 87801 January 19, 2007 1 Abstract: A time-varying regression
More informationDynamic Generalized Linear Models
Dynamic Generalized Linear Models Jesse Windle Oct. 24, 2012 Contents 1 Introduction 1 2 Binary Data (Static Case) 2 3 Data Augmentation (de-marginalization) by 4 examples 3 3.1 Example 1: CDF method.............................
More informationLabor-Supply Shifts and Economic Fluctuations. Technical Appendix
Labor-Supply Shifts and Economic Fluctuations Technical Appendix Yongsung Chang Department of Economics University of Pennsylvania Frank Schorfheide Department of Economics University of Pennsylvania January
More informationSwitching Regime Estimation
Switching Regime Estimation Series de Tiempo BIrkbeck March 2013 Martin Sola (FE) Markov Switching models 01/13 1 / 52 The economy (the time series) often behaves very different in periods such as booms
More informationBayesian Semiparametric GARCH Models
Bayesian Semiparametric GARCH Models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics xibin.zhang@monash.edu Quantitative Methods
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning
More informationBayesian Semiparametric GARCH Models
Bayesian Semiparametric GARCH Models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics xibin.zhang@monash.edu Quantitative Methods
More informationReducing Dimensions in a Large TVP-VAR
Reducing Dimensions in a Large TVP-VAR Eric Eisenstat University of Queensland Joshua C.C. Chan University of Technology Sydney Rodney W. Strachan University of Queensland March, 8 Corresponding author:
More informationSequential Monte Carlo Methods for Bayesian Computation
Sequential Monte Carlo Methods for Bayesian Computation A. Doucet Kyoto Sept. 2012 A. Doucet (MLSS Sept. 2012) Sept. 2012 1 / 136 Motivating Example 1: Generic Bayesian Model Let X be a vector parameter
More informationHastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model
UNIVERSITY OF TEXAS AT SAN ANTONIO Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model Liang Jing April 2010 1 1 ABSTRACT In this paper, common MCMC algorithms are introduced
More informationIntroduction to Machine Learning
Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 25: Markov Chain Monte Carlo (MCMC) Course Review and Advanced Topics Many figures courtesy Kevin
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationBayesian Modeling of Conditional Distributions
Bayesian Modeling of Conditional Distributions John Geweke University of Iowa Indiana University Department of Economics February 27, 2007 Outline Motivation Model description Methods of inference Earnings
More informationMarkov Chain Monte Carlo
Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).
More informationTAKEHOME FINAL EXAM e iω e 2iω e iω e 2iω
ECO 513 Spring 2015 TAKEHOME FINAL EXAM (1) Suppose the univariate stochastic process y is ARMA(2,2) of the following form: y t = 1.6974y t 1.9604y t 2 + ε t 1.6628ε t 1 +.9216ε t 2, (1) where ε is i.i.d.
More informationEstimating Macroeconomic Models: A Likelihood Approach
Estimating Macroeconomic Models: A Likelihood Approach Jesús Fernández-Villaverde University of Pennsylvania, NBER, and CEPR Juan Rubio-Ramírez Federal Reserve Bank of Atlanta Estimating Dynamic Macroeconomic
More informationInterweaving Strategy to Improve the Mixing Properties of MCMC Estimation for Time-Varying Parameter Models
Interweaving Strategy to Improve the Mixing Properties of MCMC Estimation for Time-Varying Parameter Models Stef De Visscher 1,2 and Gerdie Everaert 1 1 Ghent University, Faculty of Economics and Business
More informationLECTURE 15 Markov chain Monte Carlo
LECTURE 15 Markov chain Monte Carlo There are many settings when posterior computation is a challenge in that one does not have a closed form expression for the posterior distribution. Markov chain Monte
More informationBayesian Dynamic Linear Modelling for. Complex Computer Models
Bayesian Dynamic Linear Modelling for Complex Computer Models Fei Liu, Liang Zhang, Mike West Abstract Computer models may have functional outputs. With no loss of generality, we assume that a single computer
More informationAdvances and Applications in Perfect Sampling
and Applications in Perfect Sampling Ph.D. Dissertation Defense Ulrike Schneider advisor: Jem Corcoran May 8, 2003 Department of Applied Mathematics University of Colorado Outline Introduction (1) MCMC
More informationSimulation Smoothing for State-Space Models: A. Computational Efficiency Analysis
Simulation Smoothing for State-Space Models: A Computational Efficiency Analysis William J. McCausland Université de Montréal, CIREQ and CIRANO Shirley Miller Université de Montréal Denis Pelletier North
More information(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis
Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals
More informationMonte Carlo Methods. Leon Gu CSD, CMU
Monte Carlo Methods Leon Gu CSD, CMU Approximate Inference EM: y-observed variables; x-hidden variables; θ-parameters; E-step: q(x) = p(x y, θ t 1 ) M-step: θ t = arg max E q(x) [log p(y, x θ)] θ Monte
More informationDSGE Methods. Estimation of DSGE models: Maximum Likelihood & Bayesian. Willi Mutschler, M.Sc.
DSGE Methods Estimation of DSGE models: Maximum Likelihood & Bayesian Willi Mutschler, M.Sc. Institute of Econometrics and Economic Statistics University of Münster willi.mutschler@uni-muenster.de Summer
More informationPaul Karapanagiotidis ECO4060
Paul Karapanagiotidis ECO4060 The way forward 1) Motivate why Markov-Chain Monte Carlo (MCMC) is useful for econometric modeling 2) Introduce Markov-Chain Monte Carlo (MCMC) - Metropolis-Hastings (MH)
More informationNonlinear Inequality Constrained Ridge Regression Estimator
The International Conference on Trends and Perspectives in Linear Statistical Inference (LinStat2014) 24 28 August 2014 Linköping, Sweden Nonlinear Inequality Constrained Ridge Regression Estimator Dr.
More informationDynamic models. Dependent data The AR(p) model The MA(q) model Hidden Markov models. 6 Dynamic models
6 Dependent data The AR(p) model The MA(q) model Hidden Markov models Dependent data Dependent data Huge portion of real-life data involving dependent datapoints Example (Capture-recapture) capture histories
More informationEstimation and Inference on Dynamic Panel Data Models with Stochastic Volatility
Estimation and Inference on Dynamic Panel Data Models with Stochastic Volatility Wen Xu Department of Economics & Oxford-Man Institute University of Oxford (Preliminary, Comments Welcome) Theme y it =
More informationOnline Appendix to: Marijuana on Main Street? Estimating Demand in Markets with Limited Access
Online Appendix to: Marijuana on Main Street? Estating Demand in Markets with Lited Access By Liana Jacobi and Michelle Sovinsky This appendix provides details on the estation methodology for various speci
More informationGaussian Multiscale Spatio-temporal Models for Areal Data
Gaussian Multiscale Spatio-temporal Models for Areal Data (University of Missouri) Scott H. Holan (University of Missouri) Adelmo I. Bertolde (UFES) Outline Motivation Multiscale factorization The multiscale
More informationMarkov Chain Monte Carlo (MCMC)
Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can
More informationSUPPLEMENT TO MARKET ENTRY COSTS, PRODUCER HETEROGENEITY, AND EXPORT DYNAMICS (Econometrica, Vol. 75, No. 3, May 2007, )
Econometrica Supplementary Material SUPPLEMENT TO MARKET ENTRY COSTS, PRODUCER HETEROGENEITY, AND EXPORT DYNAMICS (Econometrica, Vol. 75, No. 3, May 2007, 653 710) BY SANGHAMITRA DAS, MARK ROBERTS, AND
More informationMonte Carlo in Bayesian Statistics
Monte Carlo in Bayesian Statistics Matthew Thomas SAMBa - University of Bath m.l.thomas@bath.ac.uk December 4, 2014 Matthew Thomas (SAMBa) Monte Carlo in Bayesian Statistics December 4, 2014 1 / 16 Overview
More informationEstimating marginal likelihoods from the posterior draws through a geometric identity
Estimating marginal likelihoods from the posterior draws through a geometric identity Johannes Reichl Energy Institute at the Johannes Kepler University Linz E-mail for correspondence: reichl@energieinstitut-linz.at
More informationMCMC Sampling for Bayesian Inference using L1-type Priors
MÜNSTER MCMC Sampling for Bayesian Inference using L1-type Priors (what I do whenever the ill-posedness of EEG/MEG is just not frustrating enough!) AG Imaging Seminar Felix Lucka 26.06.2012 , MÜNSTER Sampling
More informationStock index returns density prediction using GARCH models: Frequentist or Bayesian estimation?
MPRA Munich Personal RePEc Archive Stock index returns density prediction using GARCH models: Frequentist or Bayesian estimation? Ardia, David; Lennart, Hoogerheide and Nienke, Corré aeris CAPITAL AG,
More informationOnline Appendix for: A Bounded Model of Time Variation in Trend Inflation, NAIRU and the Phillips Curve
Online Appendix for: A Bounded Model of Time Variation in Trend Inflation, NAIRU and the Phillips Curve Joshua CC Chan Australian National University Gary Koop University of Strathclyde Simon M Potter
More informationComputer Intensive Methods in Mathematical Statistics
Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of
More informationKernel adaptive Sequential Monte Carlo
Kernel adaptive Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) December 7, 2015 1 / 36 Section 1 Outline
More informationEcon 623 Econometrics II Topic 2: Stationary Time Series
1 Introduction Econ 623 Econometrics II Topic 2: Stationary Time Series In the regression model we can model the error term as an autoregression AR(1) process. That is, we can use the past value of the
More informationBusiness Cycle Comovements in Industrial Subsectors
Business Cycle Comovements in Industrial Subsectors Michael T. Owyang 1 Daniel Soques 2 1 Federal Reserve Bank of St. Louis 2 University of North Carolina Wilmington The views expressed here are those
More informationNonparametric Bayesian Methods (Gaussian Processes)
[70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent
More informationGeneralized Autoregressive Score Models
Generalized Autoregressive Score Models by: Drew Creal, Siem Jan Koopman, André Lucas To capture the dynamic behavior of univariate and multivariate time series processes, we can allow parameters to be
More informationDynamic probabilities of restrictions in state space models: An application to the New Keynesian Phillips Curve
Dynamic probabilities of restrictions in state space models: An application to the New Keynesian Phillips Curve Gary Koop Department of Economics University of Strathclyde Scotland Gary.Koop@strath.ac.uk
More informationWorking Papers in Econometrics and Applied Statistics
T h e U n i v e r s i t y o f NEW ENGLAND Working Papers in Econometrics and Applied Statistics Finite Sample Inference in the SUR Model Duangkamon Chotikapanich and William E. Griffiths No. 03 - April
More informationState-Space Methods for Inferring Spike Trains from Calcium Imaging
State-Space Methods for Inferring Spike Trains from Calcium Imaging Joshua Vogelstein Johns Hopkins April 23, 2009 Joshua Vogelstein (Johns Hopkins) State-Space Calcium Imaging April 23, 2009 1 / 78 Outline
More informationJuly 31, 2009 / Ben Kedem Symposium
ing The s ing The Department of Statistics North Carolina State University July 31, 2009 / Ben Kedem Symposium Outline ing The s 1 2 s 3 4 5 Ben Kedem ing The s Ben has made many contributions to time
More informationGaussian kernel GARCH models
Gaussian kernel GARCH models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics 7 June 2013 Motivation A regression model is often
More informationBayesian Nonparametric Regression for Diabetes Deaths
Bayesian Nonparametric Regression for Diabetes Deaths Brian M. Hartman PhD Student, 2010 Texas A&M University College Station, TX, USA David B. Dahl Assistant Professor Texas A&M University College Station,
More informationThe Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision
The Particle Filter Non-parametric implementation of Bayes filter Represents the belief (posterior) random state samples. by a set of This representation is approximate. Can represent distributions that
More informationDefault Priors and Effcient Posterior Computation in Bayesian
Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature
More informationTime-Varying Parameters
Kalman Filter and state-space models: time-varying parameter models; models with unobservable variables; basic tool: Kalman filter; implementation is task-specific. y t = x t β t + e t (1) β t = µ + Fβ
More informationVolatility. Gerald P. Dwyer. February Clemson University
Volatility Gerald P. Dwyer Clemson University February 2016 Outline 1 Volatility Characteristics of Time Series Heteroskedasticity Simpler Estimation Strategies Exponentially Weighted Moving Average Use
More informationBayesian Stochastic Volatility (SV) Model with non-gaussian Errors
rrors Bayesian Stochastic Volatility (SV) Model with non-gaussian Errors Seokwoo Lee 1, Hedibert F. Lopes 2 1 Department of Statistics 2 Graduate School of Business University of Chicago May 19, 2008 rrors
More informationMarkov-Switching Models with Endogenous Explanatory Variables. Chang-Jin Kim 1
Markov-Switching Models with Endogenous Explanatory Variables by Chang-Jin Kim 1 Dept. of Economics, Korea University and Dept. of Economics, University of Washington First draft: August, 2002 This version:
More informationThis is the author s version of a work that was submitted/accepted for publication in the following source:
This is the author s version of a work that was submitted/accepted for publication in the following source: Strickland, Christopher M., Turner, Ian W., Denham, Robert, & Mengersen, Kerrie L. (2008) Efficient
More informationPattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods
Pattern Recognition and Machine Learning Chapter 11: Sampling Methods Elise Arnaud Jakob Verbeek May 22, 2008 Outline of the chapter 11.1 Basic Sampling Algorithms 11.2 Markov Chain Monte Carlo 11.3 Gibbs
More informationMCMC and Gibbs Sampling. Kayhan Batmanghelich
MCMC and Gibbs Sampling Kayhan Batmanghelich 1 Approaches to inference l Exact inference algorithms l l l The elimination algorithm Message-passing algorithm (sum-product, belief propagation) The junction
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing
More informationCS242: Probabilistic Graphical Models Lecture 7B: Markov Chain Monte Carlo & Gibbs Sampling
CS242: Probabilistic Graphical Models Lecture 7B: Markov Chain Monte Carlo & Gibbs Sampling Professor Erik Sudderth Brown University Computer Science October 27, 2016 Some figures and materials courtesy
More informationMarkov Chain Monte Carlo
Markov Chain Monte Carlo Michael Johannes Columbia University Nicholas Polson University of Chicago August 28, 2007 1 Introduction The Bayesian solution to any inference problem is a simple rule: compute
More information