Particle Learning and Smoothing
|
|
- Constance Ross
- 5 years ago
- Views:
Transcription
1 Particle Learning and Smoothing Hedibert Freitas Lopes The University of Chicago Booth School of Business September 18th 2009 Instituto Nacional de Pesquisas Espaciais São José dos Campos, Brazil Joint with Nick Polson, Carlos Carvalho and Mike Johannes 1
2 Outline of the talk Let the general dynamic model be Observation equation : p(y t+1 x t+1, θ) System equation : p(x t+1 x t, θ) We will talk about... MCMC in normal dynamic linear models. Particle filters: learning states x t+1. Particle filters: learning parameters θ. Particle Learning (PL) framework. More general dynamic models. Final remarks. 1
3 Normal dynamic linear model (NDLM) West and Harrison (1997): Hidden Markovian structure y t+1 x t+1 N(F t+1 x t+1, t+1) x t+1 x t N(G t+1 x t, t+1) Sequential learning p(x t y t ) = p(x t+1 y t ) = p(y t+1 x t ) = p(x t+1 y t+1 ) where y t = (y 1,..., y t ). 2
4 Example i. Local level model Let θ = (, ), x 0 N(m 0, C 0 ) and Kalman filter recursions y t+1 x t+1, θ N(x t+1, ) x t+1 x t, θ N(x t, ) Posterior at t: (x t y t ) N(m t, C t ) Prior at t + 1: (x t+1 y t ) N(m t, R t+1 ) Predictive at t + 1: (y t+1 y t ) N(m t, Q t+1 ) Posterior at t + 1: (x t+1 y t+1 ) N(m t+1, C t+1 ) where R t+1 = C t +, Q t+1 = R t+1 +, A t+1 = R t+1 /Q t+1, C t+1 = A t+1, and m t+1 = (1 A t+1 )m t + A t+1 y t+1. 3
5 Example i. Backward smoothing For t = n, x n y n N(m n n, C n n ), where m n n = m n C n n = C n For t < n, x t y n N(m n t, C n t ), where m n t = (1 B t )m t + B t m n t+1 C n t = (1 B t )C t + B 2 t C n t+1 and B t = C t C t + 4
6 Example i. Backward sampling For t = n, x n y n N(a n n, R n n ), where a n n = m n R n n = C n For t < n, x t x t+1, y n N(a n t, R n t ), where at n = (1 B t )m t + B t x t+1 Rt n = B t and B t = C t C t + This is basically the Forward filtering, backward sampling (FFBS) algorithm commonly used to sample from p(x n y n ) (Carter and Kohn, 1994 and Frühwirth-Schnatter, 1994). 5
7 Example i. n = 100, = 1.0, = 0.5 and x 0 = y(t) x(t)
8 Example i. p(x t y t, θ) and p(x t y n, θ) for t n. m 0 = 0.0 and C 0 = Forward filtering Backward smoothing
9 Non-Gaussian, nonlinear dynamic models The dynamic model is for t = 1,..., n and p(x 0 ). p(y t+1 x t+1 ) and p(x t+1 x t ) Prior and posterior at t + 1, i.e. p(x t+1 y t ) = p(x t+1 x t )p(x t y t )dx t p(x t+1 y t+1 ) p(y t+1 x t+1 )p(x t+1 y t ) are usually unavailable in closed form. Over the last 20 years: Carlin, Polson and Stoffer (1992) for more general DMs; Carter and Kohn (1994) and Frühwirth-Schnatter (1994) for conditionally Gaussian DLMs; Gamerman (1998) for generalized DLMs. 8
10 The Bayesian boostrap filter (BBF) Gordon, Salmond and Smith (1993) use a propagate-sample scheme to go from p(x t y t ) to p(x t+1 y t ) to p(x t 1 y t 1 ). 9
11 Resample or not resample? PARTICLE PATHS Time 10
12 Weights Time PARTICLE WEIGHTS 11
13 Fully adapted BBF (FABBF) Posterior at t: {x (i) t } N i=1 p(x t y t ). Propagate Draw { x (i) t+1 }N i=1 from p(x t+1 x (i) t, y t+1 ). Resample Draw {x (j) t+1 }N (i) j=1 from { x t+1 }N i=1 with weights w (i) t+1 p(y t+1 x (i) t ). Posterior at t + 1: {x (i) t+1 }N i=1 p(x t+1 y t+1 ). 12
14 The auxiliary particle filter (APF) Posterior at t: {x (i) t } N i=1 p(x t y t ). Resample Draw { x (j) t } N j=1 where g(x t ) = E(x t x t 1 ). (i) from {x t } N i=1 with weights w (i) t+1 p(y t+1 g(x (i) t )) Propagate Draw {x (i) t+1 }N i=1 from p(x t+1 x (i) t ) and resample (SIR) with weights t+1 p(y t+1 x (j) t+1 ). p(y t+1 g(x (kj ) t )) ω (j) Posterior at t + 1: {x (i) t+1 }N i=1 p(x t+1 y t+1 ). 13
15 How about p(θ y n )? Two-step strategy: On the first step, approximate p(θ y n ) by p N (θ y n ) = pn (y n θ)p(θ) p(y n ) p N (y n θ)p(θ) where p N (y n θ) is a SMC approximation to p(y n θ). Then, on the 2nd step, sample θ via a MCMC scheme or a SIR scheme 1. Problem 1: SMC looses its appealing sequential nature. Problem 2: Overall sampling scheme is sensitive to p N (y θ). 1 See Fernándes-Villaverde and Rubio-Ramírez (2007) Estimating Macroeconomic Models: A Likelihood Approach, DeJong, Dharmarajan, Liesenfeld, Moura and Richard (2009) Efficient Likelihood Evaluation of State-Space Representations for applications of this two-step strategy to DSGE and related models. 14
16 Example ii. Exact integrated likelihood p(y n, ) n = 100, x 0 = 0, = 1, = 0.5 and x 0 N(0.0, 100) grid: = (0.1,..., 2) and = (0.1,..., 3) t=10 t=20 t=30 t=40 t= t=60 t=70 t=80 t=90 t=
17 Example ii. Approximated p N (y n, ) Based on N = 1000 particles t=10 t=20 t=30 t=40 t= t=60 t=70 t=80 t=90 t=
18 Another (perhaps more natural) idea Sequentially learning x t and θ. Posterior at t : p(x t θ, y t )p(θ y t ) Prior ate t+1 : p(x t+1 θ, y t )p(θ y t ) Posterior at t+1 : p(x t+1 θ, y t+1 )p(θ y t+1 ) Advantages: Sequential updates of p(θ y t ), p(x t y t ) and p(θ, x t y t ) Sequential h-steps ahead forecast p(y t+h y t ) Sequential approximations for p(y t y t 1 ) Sequential Bayes factors t j=1 B 12t = p(y j y j 1, M 1 ) t j=1 p(y j y j 1, M 2 ) 17
19 Liu and West filter (LWF) {(x t, θ) (i) } N i=1 p(x t, θ y t ). Compute θ, V and m(θ (i) ) = aθ (i) + (1 a) θ i. Resampling Compute g(x (i) t ) = E(x t+1 x (i) t, m(θ (i) )) i. Compute w (i) t+1 = p(y t+1 g(x (j) t ), m(θ (j) )) i. Resample ( x t, θ) (i) from {(x t, θ, w t+1 ) (j) } N j=1. Sampling Sample θ(i) N(m( θ(i) ), (1 a 2 )V ) i Sample x (i) t+1 p(x t+1 x (i) t, θ(i) ) i. Compute ω (i) t+1 = p(y t+1 x (i) t+1, θ(i) )/p(yt+1 g( x (i) t ), m( θ (i) )) i. Resample (x t+1, θ) (i) from {( x t+1, θ, ω t+1 ) (j) } N j=1. {(x t+1, θ) (i) } N i=1 p(x t+1, θ y t+1 ). 18
20 Choosing a They approximate p(θ y t ) by a N-component mixture of normals p(θ y t ) = N i=1 ω (i) t f N (θ aθ (i) + (1 a) θ, (1 a 2 )V ) where θ and V approximate mean and variance of (θ y t ). A discount factor argument is used to set up the shrinkage quantity For example, δ = 0.50 leads to a = δ = 0.75 leads to a = δ = 0.95 leads to a = δ = 1.00 leads to a = a = 3δ 1. 2δ When a = 1.0, the particles will degenerate over. 19
21 Example iii. N = 2000, x 0 = 25, =.1, =.05 LW1:log, LW2:, LW3:LW1 + o.p. LW4:LW2 + o.p. 2 LW1 with delta=0.75 LW2 with delta=0.75 LW3 with delta=0.75 LW4 with delta= LW1 with delta=0.95 LW2 with delta=0.95 LW3 with delta=0.95 LW4 with delta= o.p.=optimal propagation 20
22 Particle Learning (PL) Advantages: Sequentially learning about (x t, θ); A real/practical alternative to MCMC methods; Sequential model assessment; Applicable in a wide range of dynamic and static models. Key issues in our approach: Reverse the Kalman filter logic: resample/propagate; Estimation of fixed, unknown parameters; Work with conditional sufficient statistics; Use SMC for smoothing. 21
23 Parameter learning 3 It is assumed that p ( θ x t, y t) = p (θ s t ), where s t is a recursively defined sufficient statistic (SS), s t+1 = S (s t, x t+1, y t+1 ). SS are just another state with a deterministic evolution; Filtering SS provides a mechanism for replenishing the parameters avoiding degeneracy; Reduction of the variance of re-sampling weights: Rao-Blackwellization. 3 Storvik (2002) and Fearnhead (2002) 22
24 Reverse the Kalman filter logic Traditional logic for filtering (Kalman, 1960): Predict/Update p ( x t+1 y t) = p (x t+1 x t ) p ( x t y t) dx t p ( x t+1 y t+1) p (y t+1 x t+1 ) p ( x t+1 y t). By Bayes rule, we can reverse this logic through: p ( x t y t+1) p (y t+1 x t ) p ( x t y t) p ( x t+1 y t+1) = p (x t+1 x t, y t+1 ) p ( x t y t+1) dx t We will first re-sample (smooth) and then propagate. Since information in y t+1 is used in both steps, the algorithm will be more efficient. 23
25 The general approach Assume at t, {(x t, s t ) (i)} N approximates i=1 pn (x t, s t y t ); Once y t+1 is observed, the re-sample/propagation rule is Resampling p ( x t, s t y t+1) p (y t+1 x t, s t ) p ( x t, s t y t) Propagation p ( x t+1 y t+1) = p (x t+1 x t, s t, y t+1 ) p ( x t, s t y t+1) dx t ds t 24
26 Perfect adaptation Target: p ( x t+1, x t y t+1) IS weights: w p (x t+1 x t, y t+1 ) p (y t+1 x t ) p (x t y t ) q 1 (x t y t+1 ) q 2 (x t+1 x t, y t+1 ) w p (x t+1 x t, y t+1 ) p (y t+1 x t ) p (x t y t ) p (x t y t+1 ) p (x t+1 x t, y t+1 ) p (x t y t ) = 1 This is the ideal scenario and should serve as a guiding principle in the construction of sequential Monte Carlo filters 4. Marginalization is key! 4 Perfect adaptation is also discussed in Pitt and Shephard (1999) and Johansen and Doucet (2008). 25
27 The general PL algorithm Posterior at t: Φ t { (x t, θ) (i)} N i=1 p(x t, θ y t ). Compute, for i = 1,..., N, w (i) t+1 p(y t+1 x (i) t, θ (i) ) Resample from Φ t with weights w t+1 : Φ t Propagate states Update sufficient statistics Sample parameters x (i) t+1 p(x t+1 x (i) t, θ (i), y t+1 ) s (i) t+1 = S(s(i) t, x (i) t+1, y t+1) { ( x t, θ) (i) } N i=1. θ (i) p(θ s (i) t+1 ) 26
28 Example ii. p N (θ y t ) via PL for x t and MCMC for (, ) Prior: U(0.1, 2) and U(0.1, 3). Based on N = 1000 particles t= t= t= t= t= t= t= t= t= t=100 27
29 Example ii. Trace plots for and t=10 t=50 t= Time Time Time t=10 t=50 t= Time Time Time 28
30 Example ii. p N (θ y t ) via PL for (x t,, ) Prior: U(0.1, 2) and U(0.1, 3). Based on N = 1000 particles t= t= t= t= t= t= t= t= t= t=100 29
31 Example iv. Comparing BBF, APF, FABBF and PL Simulation M = 20 data sets with n = 100 observations each, = and {5, 0.13, 0.013, , } from with x 0 = 0. y t+1 x t+1 N(x t+1, ) x t+1 x t N(x t, ) Particle filters: R = 20 replications of N = 1000 particles. Prior set up: x 0 N(0, 10). 30
32 Example iv. Log relative mean square error Let q t α be such that Pr(x t < q α t y t ) = α and α = (0.05, 0.25, 0.5, 0.75, 0.95). Then, the average MSE for filter f, t and quantile 100α-percentile is MSE α t,f = 1 MR M i=1 j=1 R (ˆq tij,f α qα it) 2 We compare filters f 1 and f 2 based on log relative MSE, i.e. LRMLE α t,f 1,f 2 = log MSE α t,f 1 MSE α t,f 2 31
33 Example iv. = 5.0 and τ/σ = 0.05 (APF,BBF), (FABBF,BBF) and (PL,BBF) Percentile = 5 Percentile = 25 Percentile = 50 Percentile = 75 Percentile = Time Time Time Time Time APF FAAPF PL APF FAAPF PL APF FAAPF PL APF FAAPF PL APF FAAPF PL 32
34 Example iv. = 0.13 and τ/σ = 0.32 Percentile = 5 Percentile = 25 Percentile = 50 Percentile = 75 Percentile = Time Time Time Time Time APF FAAPF PL APF FAAPF PL APF FAAPF PL APF FAAPF PL APF FAAPF PL 33
35 Example iv. = and τ/σ = 1.00 Percentile = 5 Percentile = 25 Percentile = 50 Percentile = 75 Percentile = Time Time Time Time Time APF FAAPF PL APF FAAPF PL APF FAAPF PL APF FAAPF PL APF FAAPF PL 34
36 Example iv. = and τ/σ = 3.16 Percentile = 5 Percentile = 25 Percentile = 50 Percentile = 75 Percentile = Time Time Time Time Time APF FAAPF PL APF FAAPF PL APF FAAPF PL APF FAAPF PL APF FAAPF PL 35
37 SMC smoothers SMC smoothers are alternatives to MCMC in state-space models. Godsill, Doucet and West (2004) Monte Carlo smoothing for non-linear series introduced an O(TN 2 ) algorithm that relies on Forward particles, and Backward re-weighting via evolution equation. See also Briers, Doucet and Maskell s (2009) Smoothing Algorithms for State-Space Models for other O(TN 2 ) smoothers. An O(TN) smoothing algorithm is introduced by Fearnhead, Wyncoll and Tawn s (2008) A sequential smoothing algorithm with linear computational cost. 36
38 Example v. PL and MCMC (filtering and smoothing) n = 100, = 1, = 0.5 and x 0 = 0 and x 0 N(0, 100). N = 1000 PL particles M = 1000 MCMC draws, after discarding the first M 0 = Smoothed distributions Kullback Leibler divergence Time TRUE MCMC PL PL MCMC 37
39 Example v. Computing (in seconds) N = M 0 = M = 500 n PL MCMC n = 100 and N = M 0 = M N PL MCMC MacBook 2.4GHz Intel Duo processor, 4GB 1067MHz memory. 38
40 Example vi. Sample-resample or PL? Three series of lengtht = 1000 were simulated from y t x t, N(x t, ) x t x t 1, N(x t 1, ) with x 0 = 0 and (, ) in {(0.1, 0.01), (0.01, 0.01), (0.01, 0.1)}. Throughout is kept fixed. The independent prior distributions for x 0 and are x 0 N(m 0, V 0 ) and IG(a, b), for a = 10, b = (a + 1) 0, m 0 = 0 and V 0 = 1, where 0 is the true value of for a given study. We also include BBF in the comparison, for completion. In all filters is sampled offline from p( S t ) where S t is the vector of conditional sufficient statistics. 39
41 Example vi. Mean absolute error The three filters are rerun R = 100 s, all with the same seed within run, for each one of the three simulated data sets. Five different number of particles N were considered: 250, 500, 1000, 2000 and Mean absolute errors (MAE) taken over the 100 replications are constructed by comparing percentiles of the true sequential distributions p(x t y t ) and p( y t ) to percentiles of the estimated sequential distributions p N (x t y t ) and p N ( y t ). For α = 0.1, 0.5, 0.9, true and estimated values of q x t,α and q t,α were computed, for Pr(x t < q x t,α y t ) = Pr( < q t,α y t ) = α. For a in {x, } and α in {0.01, 0.50, 0.99}, MAE a t,α = 1 R R qt,α a ˆq t,α,r a r=1 40
42 Example vi. M = 500 and learning. BBF,sample-resample,PL. (tau2,sig2,q)=(0.01,0.1,10%) (tau2,sig2,q)=(0.01,0.1,50%) (tau2,sig2,q)=(0.01,0.1,90%) MAE MAE MAE (tau2,sig2,q)=(0.01,0.01,10%) (tau2,sig2,q)=(0.01,0.01,50%) (tau2,sig2,q)=(0.01,0.01,90%) MAE MAE MAE (tau2,sig2,q)=(0.1,0.01,10%) (tau2,sig2,q)=(0.1,0.01,50%) (tau2,sig2,q)=(0.1,0.01,90%) MAE MAE MAE
43 Example vi. M = 5000 and learning. (tau2,sig2,q)=(0.01,0.1,10%) (tau2,sig2,q)=(0.01,0.1,50%) (tau2,sig2,q)=(0.01,0.1,90%) MAE MAE MAE (tau2,sig2,q)=(0.01,0.01,10%) (tau2,sig2,q)=(0.01,0.01,50%) (tau2,sig2,q)=(0.01,0.01,90%) MAE 2e-04 3e-04 4e-04 5e-04 6e-04 MAE 2e-04 3e-04 4e-04 5e-04 6e-04 MAE 2e-04 3e-04 4e-04 5e-04 6e (tau2,sig2,q)=(0.1,0.01,10%) (tau2,sig2,q)=(0.1,0.01,50%) (tau2,sig2,q)=(0.1,0.01,90%) MAE MAE MAE
44 Example vi. M = 500 and learning x t. (tau2,sig2,q)=(0.01,0.1,10%) (tau2,sig2,q)=(0.01,0.1,50%) (tau2,sig2,q)=(0.01,0.1,90%) MAE MAE MAE (tau2,sig2,q)=(0.01,0.01,10%) (tau2,sig2,q)=(0.01,0.01,50%) (tau2,sig2,q)=(0.01,0.01,90%) MAE MAE MAE (tau2,sig2,q)=(0.1,0.01,10%) (tau2,sig2,q)=(0.1,0.01,50%) (tau2,sig2,q)=(0.1,0.01,90%) MAE MAE MAE
45 Example vi. M = 5000 and learning x t. (tau2,sig2,q)=(0.01,0.1,10%) (tau2,sig2,q)=(0.01,0.1,50%) (tau2,sig2,q)=(0.01,0.1,90%) MAE MAE MAE (tau2,sig2,q)=(0.01,0.01,10%) (tau2,sig2,q)=(0.01,0.01,50%) (tau2,sig2,q)=(0.01,0.01,90%) MAE MAE MAE (tau2,sig2,q)=(0.1,0.01,10%) (tau2,sig2,q)=(0.1,0.01,50%) (tau2,sig2,q)=(0.1,0.01,90%) MAE MAE MAE
46 Example vii. Computing sequential Bayes factors A series y t is simulated from a AR(1) plus noise model: for t = 1,..., T. (y t+1 x t+1, θ) N(x t+1, ) (x t+1 x t, θ) N(βx t, ) We set T = 100, x 0 = 0, θ = (β,, ) = (0.9, 1.0, 0.5). and are kept known and the independent prior distributions for β and x 0 are both N(0, 1). 45
47 Example vii. Simulated data y(t) and x(t) Time 46
48 Example vii. PL pure filter versus PL We run two filters: PL pure filter - our particle learning algorithm for learning x t and keeping β fixed; PL - our particle learning algorithm for learning x t and β sequentially. The filters are based on N = 10, 000 particles. 47
49 Example vii. PL pure filter versus PL β was fixed at the true value. x t Time PL pure filter PL 48
50 Example vii. PL - learning β β Time 49
51 Example vii. PL - learning β Comparing p N (β y t ) with true p(β y t ). t=10 t=20 t=30 t=40 t=50 Density Density Density Density Density t=60 t=70 t=80 t=90 t=100 Density Density Density Density Density
52 Example vii. Sequential Bayes factor Bayes factor PL versus PL pure filter Time 51
53 Example vii. Posterior model probabilities: 4 models Time 52
54 Example vii. Posterior model probabilities: 31 models PL pure filter PL Posterior probability Density
55 PL in Conditional Dynamic Linear Models (CDLM) The model is y t+1 = F λt+1 x t+1 + ɛ t+1 where ɛ t+1 N ( ) 0, V λt+1 x t+1 = G λt+1 x t + ɛ x t+1 where ɛ x t+1 N ( ) 0, W λt+1 The error distribution p (ɛ t+1 ) = N ( 0, V λt+1 ) p (λt+1 ) dλ t+1 The augmented latent state is λ t+1 p(λ t+1 λ t ) PL extends Liu and Chen s (2000) Mixture of Kalman Filters. 54
56 Algorithm Step 1 (Re-sample): Generate an index k(i) Multi ( w (i)) where w (i) p (y t+1 (st x, θ) (i)) Step 2 (Propagate): States ) λ t+1 p (λ t+1 (λ t, θ) k(i), y t+1 x t+1 p(x t+1 (x t, θ) k(i), λ t+1, y t+1 ) Step 3 (Propagate): Sufficient Statistics s x t+1 = K (s x t, θ, λ t+1, y t+1 ) s t+1 = S (s t, x t+1, λ t+1, y t+1 ) 55
57 Example A. Dynamic factor with switching loadings 5 For t = 1,..., T, the model is defined as follows: Observation equation y t z t, θ N(γ t x t, I 2 ) State equations x t x t 1, θ N(x t 1, σx) 2 λ t λ t 1, θ Ber((1 p) 1 λ t 1 q λ t 1 ) where z t = (x t, λ t ). Factor loadings: γ t = (1, β λt ). Parameters: θ = (β 1, β 2,, σx, 2 p, q). 5 Lopes and Carvalho (2007) and Lopes, Salazar and Gamerman (2008) 56
58 Example A. Conditionally conjugate prior (β i ) N ( b i0, ) B i0 ( ν00 IG 2, d ) 00 2 ( σx 2 ν10 IG 2, d ) 10 2 p Beta(p 1, p 2 ) q Beta(q 1, q 2 ) x 0 N(m 0, C 0 ) for i = 1, 2, 57
59 Example A. Particle representation At t, particles {(x t, λ t, θ, s x t, s t ) (i)} N i=1 approximating p ( x t, λ t, θ, s x t, s t y t) where st x = S(st 1 x, θ) are state sufficient statistics s t = S(s t 1, x t, λ t ) are fixed parameter sufficient statistics 58
60 Example A. Re-sampling (x t, λ t, θ, s x t, s t ) Let us redefine β i = (1, β i ) whenever necessary. Draw an index k(i) Multi(ω (i) ) with weights ω (i) p(y t+1 (st x, λ t, θ) k(i) ) with 2 p(y t+1 m t, C t, λ t, θ) = f N (y t+1 ; β j m t, V j ) Pr (λ t+1 = j λ t, θ) j=1 where V j = (C t + x)β j β j + σ2 I 2, m t and C t are components of s x t and f N denotes the normal density function. 59
61 Example A. Propagating states Draw auxiliary state λ t+1 where λ (i) t+1 p(λ t+1 (s x t, λ t, θ) k(i), y t+1 ) Pr(λ t+1 = j s x t, λ t, θ, y t+1 ) f N (y t+1 ; β j m t, V j ) p (λ t+1 = j λ t, θ). Draw state x t+1 conditionally on λ t+1 x (i) t+1 p(x t+1 λ (i) t+1, (sx t, θ) k(i), y t+1 ) by a simply Kalman filter update. 60
62 Example A. Updating sufficient statistics for states, s x t+1 The Kalman filter recursion yield where m t+1 = m t + A t+1 (y t+1 β λt+1 m t ) C t+1 = C t + x A t+1 Q 1 t+1 A t+1 Q t+1 = (C t + x)γ t+1 γ t+1 + I 2 A t+1 = (C t + x)γ t+1q 1 t+1 61
63 Example A. Updating suff. statistics for parameters, s t+1 Recall that s t+1 = S(s t, x t+1, λ t+1 ). Then, (β i, s t+1 ) N ( b i,t+1, ) B i,t+1 for i = 1, 2, ( ( ν0t s t+1 ) IG 2, d ) 0,t+1 2 ( (σx s 2 ν1t t+1 ) IG 2, d ) 1,t+1 2 (p s t+1 ) Beta(p 1,t+1, p 2,t+1 ) (q s t+1 ) Beta(q 1,t+1, q 2,t+1 ) where I λt+1 =i = I i, I λt=i,λt+1 =j = I ij, ν it = ν i,t 1 + 1, B 1 i,t+1 = B 1 it + xt+1 2, B 1 i,t+1 b i,t+1 = B 1 it b it + x t+1 y t+1,2 I i, p i,t+1 = p it + I 1i (similarly for q i,t+1 ) for i = 1, 2, d 0,t+1 = d 0,t + (y t+1,1 x t+1 ) 2 + ] 2 j=1 [(y t+1,2 b j,t+1 x t+1 ) y t+1,2 + B 1 j,t+1 b j,t+1 I j, and d 1,t+1 = d 1,t + (x t+1 x t ) 2. 62
64 Example A. Filtering and smoothing for states 63
65 Example A. Sequential parameter learning 64
66 PL in (state) non-linear normal dynamic models The model now is y t+1 = F λt+1 x t+1 + ɛ t+1 where ɛ t+1 N ( 0, V λt+1 ) x t+1 = G λt+1 Z (x t ) + ω t+1 where ω t+1 N ( 0, W λt+1 ) where ɛ t+1 and λ t+1 are modeled as before. Algorithm: Step 1 (Re-sample): Generate an index k(i) Multi ( w (i)) where w (i) p (y t+1 (x t, θ) (i)) Step 2 (Propagate): ) λ t+1 p (λ t+1 (λ t, θ) k(i), y t+1 ) x t+1 p (x t+1 (x t,, θ) k(i), λ t+1, y t+1 s t+1 = S (s t, x t+1, λ t+1, y t+1 ) 65
67 Example B. Fat-tailed nonlinear model 6 Let y t+1 = x t+1 + σ λ t+1 ɛ t+1 ( ν where λ t+1 IG 2, ν ) 2 x t+1 = g(x t )β + σ x u t+1 where g(x t ) = x t 1 + x 2 t where ɛ t+1 and u t+1 are independent standard normals and ν is known. The observation error term is non-normal λ t+1 ɛ t+1 t ν. 6 From dejong et al (2007) 66
68 Example B. Sequential inference 67
69 Example C. Dynamic multinomial logit model Let us study the multinomial logit model 7 P (y t+1 = 1 β t+1 ) = eftβt 1 + e Ftβt and β t+1 = φβ t + σ x ɛ β t+1 where β 0 N(0, /(1 ρ 2 )). Scott s (2007) data augmentation structure leads to a mixture Kalman filter model y t+1 = I(z t 0) z t+1 = Z t β + ɛ t+1 where ɛ t+1 lne(1) Here ɛ t is an extreme value distribution of type 1 where E(1) is an exponential of mean one. The key is that it is easily to simulate p(z t β, y t ) using ( ln Ui z t+1 = ln 1 + e β i β ln V ) i e β i β I y t+1 =0 7 Carvalho, Lopes and Polson (2008) 68
70 Example C. 10-component mixture of normals Frunwirth-Schnatter and Schnatter (2007) uses a 10-component mixture of normals: p(ɛ t ) = e ɛt e ɛ t 10 w j N (µ j, sj 2 ) j=1 Hence conditional on an indicator λ t we can analyze y t = I(z t 0) and z t = µ λt + Z t β + s λt ɛ t where ɛ t N(0, 1) and Pr(λ t = j) = w j. Also, ) s β t+1 = K (st β, z t+1, λ t+1, θ, y t+1 p(y t+1 s β t, θ) = λ t+1 p(y t+1 s β t, λ t+1, θ) for re-sampling. Propagation now requires λ t+1 p λ t+1 (s β t, θ)k(i), y t+1 z t+1 p z t+1 (s β t, θ)k(i), λ t+1, y t+1 β t+1 p z t+1 (s β, θ) k(i), λ t+1, z t+1 69
71 Example C. Simulated exercise Y t Time β t Time 5 α 0.4 σ Time Time PL based on 30,000 particles. 70
72 Example D. Sequential Bayesian Lasso We develop a sequential version of Bayesian Lasso 8 for a simple problem of signal detection. The model takes the form (y t θ t ) N(θ t, 1) p(θ t τ) = (2τ) 1 exp ( θ t /τ) for t = 1,..., n and IG(a 0, b 0 ). Data augmentation: It is easy to see that p(θ t τ) = p(θ t τ, λ t )p(λ t )dλ t where λ t Exp(2) θ t τ, λ t N(0, λ t ) 8 Carlin and Polson (1991) and Hans (2008) 71
73 Example D. Data augmentation The natural set of latent variables is given by the augmentation variable λ n+1 and conditional sufficient statistics leading to Z n = (λ n+1, a n, b n ) The sequence of variables λ n+1 are i.i.d. and so can be propagated directly with p(λ n+1 ). The conditional sufficient statistics (a n+1, b n+1 ) are deterministically determined based on parameters (θ n+1, λ n+1 ) and previous values (a n, b n ). 72
74 Example D. PL algorithm 1. After n observations: { (Z n, τ) (i)} N i=1. 2. Draw λ (i) n+1 Exp(2). 3. Resample old particles with weights 4. Sample θ (i) C 1 n w (i) n+1 p(y n+1; 0, 1+(i) λ (i) n+1 ). n+1 N(m(i) n 1(i) 2(i) λ n+1. = 1 + τ n, C (i) ), where m n (i) 5. Suff. stats: a (i) n+1 = ã(i) n + 1/2, b (i) n+1 = b n (i) 6. Sample (offline) (i) IG(a n+1, b n+1 ). 7. Let Z (i) n+1 = (λ(i) n+1, a(i) n+1, b(i) n+1 ). = C n (i) y n+1 and + θ 2(i) (i) n+1 /(2 λ n+1 ). 8. After n + 1 observations: { (Z n+1, τ) (i)} N i=1. 73
75 Example D. Sequential Bayes factor As the Lasso is a model for sparsity we would expect the evidence for it to increase when we observe y t = 0. We can sequentially estimate p(y n+1 y n, lasso) via p(y n+1 y n, lasso) = 1 N N p(y n+1 (λ n, τ) (i) ) i=1 with predictive p(y n+1 λ n, τ) N(0, λ n + 1). This leads to a sequential Bayes factor BF n+1 = p(y n+1 lasso) p(y n+1 normal). 74
76 Example D. Simulated data Data based on θ = (0, 0, 0, 0, 1, 0, 0, 0, 1, 1, 0, 0, 0, 0, 1) and priors IG(2, 1) for the double exponential case and IG(2, 3) for the normal case, reflecting the ratio of variances between those two distributions. Bayes factor Observations 75
77 Final remarks PL is a general framework for sequential Bayesian inference in dynamic and static models. PL is able to deal with filtering and learning and reduce the accumulation of error. The loose definition of sufficient statistics and the flexibility to freely augment x t makes PL a competitive alternative to MCMC in highly structured models. A powerful by-product of PL (and SMC in general) over MCMC schemes, is its ability to sequentially produce model comparison, assessment indicators. 76
Particle Learning and Smoothing
Particle Learning and Smoothing Carlos Carvalho, Michael Johannes, Hedibert Lopes and Nicholas Polson This version: September 2009 First draft: December 2007 Abstract In this paper we develop particle
More informationParticle Learning for General Mixtures
Particle Learning for General Mixtures Hedibert Freitas Lopes 1 Booth School of Business University of Chicago Dipartimento di Scienze delle Decisioni Università Bocconi, Milano 1 Joint work with Nicholas
More informationLecture 4: Dynamic models
linear s Lecture 4: s Hedibert Freitas Lopes The University of Chicago Booth School of Business 5807 South Woodlawn Avenue, Chicago, IL 60637 http://faculty.chicagobooth.edu/hedibert.lopes hlopes@chicagobooth.edu
More informationParticle Learning and Smoothing
Statistical Science 2010, Vol. 25, No. 1, 88 106 DOI: 10.1214/10-STS325 Institute of Mathematical Statistics, 2010 Particle Learning and Smoothing Carlos M. Carvalho, Michael S. Johannes, Hedibert F. Lopes
More informationParticle Learning and Smoothing
Statistical Science 2010, Vol. 25, No. 1, 88 106 DOI: 10.1214/10-STS325 c Institute of Mathematical Statistics, 2010 Particle Learning and Smoothing Carlos M. Carvalho, Michael S. Johannes, Hedibert F.
More informationQuantile Filtering and Learning
Quantile Filtering and Learning Michael Johannes, Nicholas Polson and Seung M. Yae October 5, 2009 Abstract Quantile and least-absolute deviations (LAD) methods are popular robust statistical methods but
More informationBayesian Monte Carlo Filtering for Stochastic Volatility Models
Bayesian Monte Carlo Filtering for Stochastic Volatility Models Roberto Casarin CEREMADE University Paris IX (Dauphine) and Dept. of Economics University Ca Foscari, Venice Abstract Modelling of the financial
More informationSequential Monte Carlo Methods (for DSGE Models)
Sequential Monte Carlo Methods (for DSGE Models) Frank Schorfheide University of Pennsylvania, PIER, CEPR, and NBER October 23, 2017 Some References These lectures use material from our joint work: Tempered
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing
More informationParticle Learning for Sequential Bayesian Computation Rejoinder
BAYESIAN STATISTICS 9, J. M. Bernardo, M. J. Bayarri, J. O. Berger, A. P. Dawid, D. Heckerman, A. F. M. Smith and M. West (Eds.) c Oxford University Press, 20 Particle Learning for Sequential Bayesian
More informationBayesian Model Comparison:
Bayesian Model Comparison: Modeling Petrobrás log-returns Hedibert Freitas Lopes February 2014 Log price: y t = log p t Time span: 12/29/2000-12/31/2013 (n = 3268 days) LOG PRICE 1 2 3 4 0 500 1000 1500
More informationA Note on Auxiliary Particle Filters
A Note on Auxiliary Particle Filters Adam M. Johansen a,, Arnaud Doucet b a Department of Mathematics, University of Bristol, UK b Departments of Statistics & Computer Science, University of British Columbia,
More informationSequential Monte Carlo Methods for Bayesian Computation
Sequential Monte Carlo Methods for Bayesian Computation A. Doucet Kyoto Sept. 2012 A. Doucet (MLSS Sept. 2012) Sept. 2012 1 / 136 Motivating Example 1: Generic Bayesian Model Let X be a vector parameter
More informationParticle Filtering for Data-Driven Simulation and Optimization
Particle Filtering for Data-Driven Simulation and Optimization John R. Birge The University of Chicago Booth School of Business Includes joint work with Nicholas Polson. JRBirge INFORMS Phoenix, October
More informationMarkov Chain Monte Carlo Methods for Stochastic Optimization
Markov Chain Monte Carlo Methods for Stochastic Optimization John R. Birge The University of Chicago Booth School of Business Joint work with Nicholas Polson, Chicago Booth. JRBirge U of Toronto, MIE,
More informationParticle Filtering Approaches for Dynamic Stochastic Optimization
Particle Filtering Approaches for Dynamic Stochastic Optimization John R. Birge The University of Chicago Booth School of Business Joint work with Nicholas Polson, Chicago Booth. JRBirge I-Sim Workshop,
More informationMarkov Chain Monte Carlo Methods for Stochastic
Markov Chain Monte Carlo Methods for Stochastic Optimization i John R. Birge The University of Chicago Booth School of Business Joint work with Nicholas Polson, Chicago Booth. JRBirge U Florida, Nov 2013
More informationIntroduction to Particle Filters for Data Assimilation
Introduction to Particle Filters for Data Assimilation Mike Dowd Dept of Mathematics & Statistics (and Dept of Oceanography Dalhousie University, Halifax, Canada STATMOS Summer School in Data Assimila5on,
More informationAuxiliary Particle Methods
Auxiliary Particle Methods Perspectives & Applications Adam M. Johansen 1 adam.johansen@bristol.ac.uk Oxford University Man Institute 29th May 2008 1 Collaborators include: Arnaud Doucet, Nick Whiteley
More informationAn introduction to Sequential Monte Carlo
An introduction to Sequential Monte Carlo Thang Bui Jes Frellsen Department of Engineering University of Cambridge Research and Communication Club 6 February 2014 1 Sequential Monte Carlo (SMC) methods
More informationAN EFFICIENT TWO-STAGE SAMPLING METHOD IN PARTICLE FILTER. Qi Cheng and Pascal Bondon. CNRS UMR 8506, Université Paris XI, France.
AN EFFICIENT TWO-STAGE SAMPLING METHOD IN PARTICLE FILTER Qi Cheng and Pascal Bondon CNRS UMR 8506, Université Paris XI, France. August 27, 2011 Abstract We present a modified bootstrap filter to draw
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation
More informationBayesian Computations for DSGE Models
Bayesian Computations for DSGE Models Frank Schorfheide University of Pennsylvania, PIER, CEPR, and NBER October 23, 2017 This Lecture is Based on Bayesian Estimation of DSGE Models Edward P. Herbst &
More informationAn Brief Overview of Particle Filtering
1 An Brief Overview of Particle Filtering Adam M. Johansen a.m.johansen@warwick.ac.uk www2.warwick.ac.uk/fac/sci/statistics/staff/academic/johansen/talks/ May 11th, 2010 Warwick University Centre for Systems
More informationLecture 8: Bayesian Estimation of Parameters in State Space Models
in State Space Models March 30, 2016 Contents 1 Bayesian estimation of parameters in state space models 2 Computational methods for parameter estimation 3 Practical parameter estimation in state space
More informationBayesian Ingredients. Hedibert Freitas Lopes
Normal Prior s Ingredients Hedibert Freitas Lopes The University of Chicago Booth School of Business 5807 South Woodlawn Avenue, Chicago, IL 60637 http://faculty.chicagobooth.edu/hedibert.lopes hlopes@chicagobooth.edu
More informationMonte Carlo Approximation of Monte Carlo Filters
Monte Carlo Approximation of Monte Carlo Filters Adam M. Johansen et al. Collaborators Include: Arnaud Doucet, Axel Finke, Anthony Lee, Nick Whiteley 7th January 2014 Context & Outline Filtering in State-Space
More informationAdvanced Computational Methods in Statistics: Lecture 5 Sequential Monte Carlo/Particle Filtering
Advanced Computational Methods in Statistics: Lecture 5 Sequential Monte Carlo/Particle Filtering Axel Gandy Department of Mathematics Imperial College London http://www2.imperial.ac.uk/~agandy London
More informationThe Kalman Filter ImPr Talk
The Kalman Filter ImPr Talk Ged Ridgway Centre for Medical Image Computing November, 2006 Outline What is the Kalman Filter? State Space Models Kalman Filter Overview Bayesian Updating of Estimates Kalman
More informationNonlinear and/or Non-normal Filtering. Jesús Fernández-Villaverde University of Pennsylvania
Nonlinear and/or Non-normal Filtering Jesús Fernández-Villaverde University of Pennsylvania 1 Motivation Nonlinear and/or non-gaussian filtering, smoothing, and forecasting (NLGF) problems are pervasive
More informationThe Unscented Particle Filter
The Unscented Particle Filter Rudolph van der Merwe (OGI) Nando de Freitas (UC Bereley) Arnaud Doucet (Cambridge University) Eric Wan (OGI) Outline Optimal Estimation & Filtering Optimal Recursive Bayesian
More informationLecture 6: Bayesian Inference in SDE Models
Lecture 6: Bayesian Inference in SDE Models Bayesian Filtering and Smoothing Point of View Simo Särkkä Aalto University Simo Särkkä (Aalto) Lecture 6: Bayesian Inference in SDEs 1 / 45 Contents 1 SDEs
More informationControlled sequential Monte Carlo
Controlled sequential Monte Carlo Jeremy Heng, Department of Statistics, Harvard University Joint work with Adrian Bishop (UTS, CSIRO), George Deligiannidis & Arnaud Doucet (Oxford) Bayesian Computation
More informationOnline appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US
Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Gerdie Everaert 1, Lorenzo Pozzi 2, and Ruben Schoonackers 3 1 Ghent University & SHERPPA 2 Erasmus
More informationIntegrated Non-Factorized Variational Inference
Integrated Non-Factorized Variational Inference Shaobo Han, Xuejun Liao and Lawrence Carin Duke University February 27, 2014 S. Han et al. Integrated Non-Factorized Variational Inference February 27, 2014
More informationPATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA Contents in latter part Linear Dynamical Systems What is different from HMM? Kalman filter Its strength and limitation Particle Filter
More informationBayesian Dynamic Linear Modelling for. Complex Computer Models
Bayesian Dynamic Linear Modelling for Complex Computer Models Fei Liu, Liang Zhang, Mike West Abstract Computer models may have functional outputs. With no loss of generality, we assume that a single computer
More informationDynamic linear models (aka state-space models) 1
Dynamic linear models (aka state-space models) 1 Advanced Econometris: Time Series Hedibert Freitas Lopes INSPER 1 Part of this lecture is based on Gamerman and Lopes (2006) Markov Chain Monte Carlo: Stochastic
More informationL09. PARTICLE FILTERING. NA568 Mobile Robotics: Methods & Algorithms
L09. PARTICLE FILTERING NA568 Mobile Robotics: Methods & Algorithms Particle Filters Different approach to state estimation Instead of parametric description of state (and uncertainty), use a set of state
More informationState-Space Methods for Inferring Spike Trains from Calcium Imaging
State-Space Methods for Inferring Spike Trains from Calcium Imaging Joshua Vogelstein Johns Hopkins April 23, 2009 Joshua Vogelstein (Johns Hopkins) State-Space Calcium Imaging April 23, 2009 1 / 78 Outline
More informationComputer Intensive Methods in Mathematical Statistics
Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 7 Sequential Monte Carlo methods III 7 April 2017 Computer Intensive Methods (1) Plan of today s lecture
More informationSequential Monte Carlo and Particle Filtering. Frank Wood Gatsby, November 2007
Sequential Monte Carlo and Particle Filtering Frank Wood Gatsby, November 2007 Importance Sampling Recall: Let s say that we want to compute some expectation (integral) E p [f] = p(x)f(x)dx and we remember
More informationHuman Pose Tracking I: Basics. David Fleet University of Toronto
Human Pose Tracking I: Basics David Fleet University of Toronto CIFAR Summer School, 2009 Looking at People Challenges: Complex pose / motion People have many degrees of freedom, comprising an articulated
More informationSequential Bayesian Updating
BS2 Statistical Inference, Lectures 14 and 15, Hilary Term 2009 May 28, 2009 We consider data arriving sequentially X 1,..., X n,... and wish to update inference on an unknown parameter θ online. In a
More informationLinear Dynamical Systems
Linear Dynamical Systems Sargur N. srihari@cedar.buffalo.edu Machine Learning Course: http://www.cedar.buffalo.edu/~srihari/cse574/index.html Two Models Described by Same Graph Latent variables Observations
More informationSelf Adaptive Particle Filter
Self Adaptive Particle Filter Alvaro Soto Pontificia Universidad Catolica de Chile Department of Computer Science Vicuna Mackenna 4860 (143), Santiago 22, Chile asoto@ing.puc.cl Abstract The particle filter
More informationThe Metropolis-Hastings Algorithm. June 8, 2012
The Metropolis-Hastings Algorithm June 8, 22 The Plan. Understand what a simulated distribution is 2. Understand why the Metropolis-Hastings algorithm works 3. Learn how to apply the Metropolis-Hastings
More informationSensor Fusion: Particle Filter
Sensor Fusion: Particle Filter By: Gordana Stojceska stojcesk@in.tum.de Outline Motivation Applications Fundamentals Tracking People Advantages and disadvantages Summary June 05 JASS '05, St.Petersburg,
More informationLecture : Probabilistic Machine Learning
Lecture : Probabilistic Machine Learning Riashat Islam Reasoning and Learning Lab McGill University September 11, 2018 ML : Many Methods with Many Links Modelling Views of Machine Learning Machine Learning
More informationIntroduction to Bayesian Inference
University of Pennsylvania EABCN Training School May 10, 2016 Bayesian Inference Ingredients of Bayesian Analysis: Likelihood function p(y φ) Prior density p(φ) Marginal data density p(y ) = p(y φ)p(φ)dφ
More informationExpectation Propagation for Approximate Bayesian Inference
Expectation Propagation for Approximate Bayesian Inference José Miguel Hernández Lobato Universidad Autónoma de Madrid, Computer Science Department February 5, 2007 1/ 24 Bayesian Inference Inference Given
More informationRobotics. Mobile Robotics. Marc Toussaint U Stuttgart
Robotics Mobile Robotics State estimation, Bayes filter, odometry, particle filter, Kalman filter, SLAM, joint Bayes filter, EKF SLAM, particle SLAM, graph-based SLAM Marc Toussaint U Stuttgart DARPA Grand
More informationMotivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University
Econ 690 Purdue University In virtually all of the previous lectures, our models have made use of normality assumptions. From a computational point of view, the reason for this assumption is clear: combined
More informationLecture 7: Optimal Smoothing
Department of Biomedical Engineering and Computational Science Aalto University March 17, 2011 Contents 1 What is Optimal Smoothing? 2 Bayesian Optimal Smoothing Equations 3 Rauch-Tung-Striebel Smoother
More informationTime Series and Dynamic Models
Time Series and Dynamic Models Section 1 Intro to Bayesian Inference Carlos M. Carvalho The University of Texas at Austin 1 Outline 1 1. Foundations of Bayesian Statistics 2. Bayesian Estimation 3. The
More informationGaussian Multiscale Spatio-temporal Models for Areal Data
Gaussian Multiscale Spatio-temporal Models for Areal Data (University of Missouri) Scott H. Holan (University of Missouri) Adelmo I. Bertolde (UFES) Outline Motivation Multiscale factorization The multiscale
More informationTowards inference for skewed alpha stable Levy processes
Towards inference for skewed alpha stable Levy processes Simon Godsill and Tatjana Lemke Signal Processing and Communications Lab. University of Cambridge www-sigproc.eng.cam.ac.uk/~sjg Overview Motivation
More informationExercises Tutorial at ICASSP 2016 Learning Nonlinear Dynamical Models Using Particle Filters
Exercises Tutorial at ICASSP 216 Learning Nonlinear Dynamical Models Using Particle Filters Andreas Svensson, Johan Dahlin and Thomas B. Schön March 18, 216 Good luck! 1 [Bootstrap particle filter for
More informationExpectation Propagation Algorithm
Expectation Propagation Algorithm 1 Shuang Wang School of Electrical and Computer Engineering University of Oklahoma, Tulsa, OK, 74135 Email: {shuangwang}@ou.edu This note contains three parts. First,
More informationSequential Bayesian Inference for Dynamic State Space. Model Parameters
Sequential Bayesian Inference for Dynamic State Space Model Parameters Arnab Bhattacharya and Simon Wilson 1. INTRODUCTION Dynamic state-space models [24], consisting of a latent Markov process X 0,X 1,...
More informationPatterns of Scalable Bayesian Inference Background (Session 1)
Patterns of Scalable Bayesian Inference Background (Session 1) Jerónimo Arenas-García Universidad Carlos III de Madrid jeronimo.arenas@gmail.com June 14, 2017 1 / 15 Motivation. Bayesian Learning principles
More informationApproximate Bayesian Computation and Particle Filters
Approximate Bayesian Computation and Particle Filters Dennis Prangle Reading University 5th February 2014 Introduction Talk is mostly a literature review A few comments on my own ongoing research See Jasra
More informationAn introduction to particle filters
An introduction to particle filters Andreas Svensson Department of Information Technology Uppsala University June 10, 2014 June 10, 2014, 1 / 16 Andreas Svensson - An introduction to particle filters Outline
More informationGeneralized Autoregressive Score Smoothers
Generalized Autoregressive Score Smoothers Giuseppe Buccheri 1, Giacomo Bormetti 2, Fulvio Corsi 3,4, and Fabrizio Lillo 2 1 Scuola Normale Superiore, Italy 2 University of Bologna, Italy 3 University
More informationRAO-BLACKWELLIZED PARTICLE FILTER FOR MARKOV MODULATED NONLINEARDYNAMIC SYSTEMS
RAO-BLACKWELLIZED PARTICLE FILTER FOR MARKOV MODULATED NONLINEARDYNAMIC SYSTEMS Saiat Saha and Gustaf Hendeby Linöping University Post Print N.B.: When citing this wor, cite the original article. 2014
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department
More informationParticle filters, the optimal proposal and high-dimensional systems
Particle filters, the optimal proposal and high-dimensional systems Chris Snyder National Center for Atmospheric Research Boulder, Colorado 837, United States chriss@ucar.edu 1 Introduction Particle filters
More informationResearch Division Federal Reserve Bank of St. Louis Working Paper Series
Research Division Federal Reserve Bank of St Louis Working Paper Series Kalman Filtering with Truncated Normal State Variables for Bayesian Estimation of Macroeconomic Models Michael Dueker Working Paper
More informationPROBABILISTIC REASONING OVER TIME
PROBABILISTIC REASONING OVER TIME In which we try to interpret the present, understand the past, and perhaps predict the future, even when very little is crystal clear. Outline Time and uncertainty Inference:
More informationBayesian Inference. Chapter 4: Regression and Hierarchical Models
Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Advanced Statistics and Data Mining Summer School
More informationDynamic Generalized Linear Models
Dynamic Generalized Linear Models Jesse Windle Oct. 24, 2012 Contents 1 Introduction 1 2 Binary Data (Static Case) 2 3 Data Augmentation (de-marginalization) by 4 examples 3 3.1 Example 1: CDF method.............................
More informationBayesian Inference. Chapter 4: Regression and Hierarchical Models
Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Master in Business Administration and Quantitative
More informationFlexible Particle Markov chain Monte Carlo methods with an application to a factor stochastic volatility model
Flexible Particle Markov chain Monte Carlo methods with an application to a factor stochastic volatility model arxiv:1401.1667v4 [stat.co] 23 Jan 2018 Eduardo F. Mendes School of Applied Mathematics Fundação
More informationDefault Priors and Effcient Posterior Computation in Bayesian
Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature
More informationBayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference
1 The views expressed in this paper are those of the authors and do not necessarily reflect the views of the Federal Reserve Board of Governors or the Federal Reserve System. Bayesian Estimation of DSGE
More informationSpatial Dynamic Factor Analysis
Spatial Dynamic Factor Analysis Esther Salazar Federal University of Rio de Janeiro Department of Statistical Methods Sixth Workshop on BAYESIAN INFERENCE IN STOCHASTIC PROCESSES Bressanone/Brixen, Italy
More informationBayesian Estimation of DSGE Models: Lessons from Second-order Approximations
Bayesian Estimation of DSGE Models: Lessons from Second-order Approximations Sungbae An Singapore Management University Bank Indonesia/BIS Workshop: STRUCTURAL DYNAMIC MACROECONOMIC MODELS IN ASIA-PACIFIC
More informationParticle Filtering a brief introductory tutorial. Frank Wood Gatsby, August 2007
Particle Filtering a brief introductory tutorial Frank Wood Gatsby, August 2007 Problem: Target Tracking A ballistic projectile has been launched in our direction and may or may not land near enough to
More informationLearning of dynamical systems
Learning of dynamical systems Particle filters and Markov chain methods Thomas B. Schön and Fredrik Lindsten c Draft date August 23, 2017 2 Contents 1 Introduction 3 1.1 A few words for readers of the
More informationHmms with variable dimension structures and extensions
Hmm days/enst/january 21, 2002 1 Hmms with variable dimension structures and extensions Christian P. Robert Université Paris Dauphine www.ceremade.dauphine.fr/ xian Hmm days/enst/january 21, 2002 2 1 Estimating
More informationECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering
ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering Lecturer: Nikolay Atanasov: natanasov@ucsd.edu Teaching Assistants: Siwei Guo: s9guo@eng.ucsd.edu Anwesan Pal:
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters
More informationBayesian Stochastic Volatility (SV) Model with non-gaussian Errors
rrors Bayesian Stochastic Volatility (SV) Model with non-gaussian Errors Seokwoo Lee 1, Hedibert F. Lopes 2 1 Department of Statistics 2 Graduate School of Business University of Chicago May 19, 2008 rrors
More informationEVALUATING SYMMETRIC INFORMATION GAP BETWEEN DYNAMICAL SYSTEMS USING PARTICLE FILTER
EVALUATING SYMMETRIC INFORMATION GAP BETWEEN DYNAMICAL SYSTEMS USING PARTICLE FILTER Zhen Zhen 1, Jun Young Lee 2, and Abdus Saboor 3 1 Mingde College, Guizhou University, China zhenz2000@21cn.com 2 Department
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee and Andrew O. Finley 2 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department
More informationE cient Likelihood Evaluation of State-Space Representations
E cient Likelihood Evaluation of State-Space Representations David N. DeJong, Roman Liesenfeld, Guilherme V. Moura, Jean-Francois Richard, Hariharan Dharmarajan University of Pittsburgh Universität Kiel
More informationBayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence
Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns
More informationUnsupervised Learning
Unsupervised Learning Bayesian Model Comparison Zoubin Ghahramani zoubin@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit, and MSc in Intelligent Systems, Dept Computer Science University College
More informationSequential Monte Carlo Samplers for Applications in High Dimensions
Sequential Monte Carlo Samplers for Applications in High Dimensions Alexandros Beskos National University of Singapore KAUST, 26th February 2014 Joint work with: Dan Crisan, Ajay Jasra, Nik Kantas, Alex
More informationIntroduction to Machine Learning
Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 25: Markov Chain Monte Carlo (MCMC) Course Review and Advanced Topics Many figures courtesy Kevin
More informationThe Kalman filter, Nonlinear filtering, and Markov Chain Monte Carlo
NBER Summer Institute Minicourse What s New in Econometrics: Time Series Lecture 5 July 5, 2008 The Kalman filter, Nonlinear filtering, and Markov Chain Monte Carlo Lecture 5, July 2, 2008 Outline. Models
More informationLecture 13 : Variational Inference: Mean Field Approximation
10-708: Probabilistic Graphical Models 10-708, Spring 2017 Lecture 13 : Variational Inference: Mean Field Approximation Lecturer: Willie Neiswanger Scribes: Xupeng Tong, Minxing Liu 1 Problem Setup 1.1
More informationExpectation propagation for signal detection in flat-fading channels
Expectation propagation for signal detection in flat-fading channels Yuan Qi MIT Media Lab Cambridge, MA, 02139 USA yuanqi@media.mit.edu Thomas Minka CMU Statistics Department Pittsburgh, PA 15213 USA
More informationEstimation of moment-based models with latent variables
Estimation of moment-based models with latent variables work in progress Ra aella Giacomini and Giuseppe Ragusa UCL/Cemmap and UCI/Luiss UPenn, 5/4/2010 Giacomini and Ragusa (UCL/Cemmap and UCI/Luiss)Moments
More informationDSGE Models in a Liquidity Trap and Japan s Lost Decade
DSGE Models in a Liquidity Trap and Japan s Lost Decade Koiti Yano Economic and Social Research Institute ESRI International Conference 2009 June 29, 2009 1 / 27 Definition of a Liquidity Trap Terminology
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 12: Gaussian Belief Propagation, State Space Models and Kalman Filters Guest Kalman Filter Lecture by
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is
More informationThe Hierarchical Particle Filter
and Arnaud Doucet http://go.warwick.ac.uk/amjohansen/talks MCMSki V Lenzerheide 7th January 2016 Context & Outline Filtering in State-Space Models: SIR Particle Filters [GSS93] Block-Sampling Particle
More informationEstimating Macroeconomic Models: A Likelihood Approach
Estimating Macroeconomic Models: A Likelihood Approach Jesús Fernández-Villaverde University of Pennsylvania, NBER, and CEPR Juan Rubio-Ramírez Federal Reserve Bank of Atlanta Estimating Dynamic Macroeconomic
More informationAn Introduction to Particle Filtering
An Introduction to Particle Filtering Author: Lisa Turner Supervisor: Dr. Christopher Sherlock 10th May 2013 Abstract This report introduces the ideas behind particle filters, looking at the Kalman filter
More information