Particle Filters in High Dimensions
|
|
- Arleen Green
- 5 years ago
- Views:
Transcription
1 Particle Filters in High Dimensions Dan Crisan Imperial College London Bayesian Computation for High-Dimensional Statistical Models Opening Workshop August 2018 Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
2 . Talk Synopsis Stochastic Filtering Particle filters/sequential Monte Carlo methods Tempering/Model Reduction Case study Motivation: Particle Filters for Numerical Weather Prediction A stochastic transport model and the associated filtering problem Particle filter (initial condition, add-on procedures, parameter choice) Final remarks Research partially supported by EPSRC grant EP/N023781/1. Numerical work done by Wei Pan (Imperial College London). C.J. Cotter, DC, D.D. Holm, W. Pan, I. Shevchenko, Numerically Modelling Stochastic Lie Transport in Fluid Dynamics, C.J. Cotter, DC, D.D. Holm, W. Pan, I. Shevchenko, Sequential Monte Carlo for Stochastic Advection by Lie Transport (SALT): A case study for the damped and forced incompressible 2D stochastic Euler equation, in preparation. A. Beskos, DC, A. Jasra, Ajay; K. Kamatani, Y. Zhou, Y A stable particle filter for a class of high-dimensional state-space models. Adv. in Appl. Probab. 49 (2017). A. Beskos, DC, A. Jasra, On the stability of sequential Monte Carlo methods in high dimensions, Ann. Appl. Probab. 24 (2014). Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
3 . What is stochastic filtering? Stochastic Filtering: The process of using partial observations and a stochastic model to make inferences about an evolving dynamical system. X the signal process - hidden component Y the observation process - the data Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
4 . What is stochastic filtering? The filtering problem: Find the conditional distribution of the signal X t given Y t = σ(y s, s [0, t]), i.e., π t (A) = P(X t A Y t ), t 0, A B(R d ). Discrete framework: {X t, Y t } t 0 Markov process The signal process {X t } t 0 Markov chain, X 0 π 0 (dx 0 ) P (X t dx t X t 1 = x t 1 ) = K t (x t 1, dx t ) = f t (x t x t 1 )dt, Example : X t = b (X t 1 ) + σ (X t 1 ) B t, The observation process B t N (0, 1) i.i.d. P ( Y t dy t X [0,t] = x [0,t], Y [0,t 1] = y [0,t 1] ) = P (Yt dy t X t = x t ) = g t (y t x t )dy Example : Y t = h (X t ) + V t, where X [0,t] (X 0,..., X t ), x [0,t] (x 0,..., x t ). V t N (0, 1) i.i.d. Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
5 . What is stochastic filtering? Notation: posterior measure: the conditional distribution of the signal X t given Y t π t (A) = P(X t A Y t ), t 0, A B(R d ). predictive measure: the conditional distribution of the signal X t given Y t 1 p t (A) = P(X t A Y t 1 ), t 0, A B(R d ). If μ is a measure and k is a kernel, then μk (A) μ (dx) k (x, A). Bayes recursion. Prediction Updating p t = π t 1 K t π t = g t p t In other words, dπt dp t = C 1 t g t, where C t R d g t (y t, x t ) p t (dx t ). Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
6 . Particle filters 1. Class of approximations: Particle filters/sequential Monte Carlo Methods: SMC (a j (t), vj 1 (t),..., vj d (t)) N j=1 }{{}}{{} weight position π t πt N = N j=1 a j (t) δ vj (t) 2. The law of evolution of the approximation: π N t 1 SMC selection {}}{ pt N {}}{ πt N mutation 3. The measure of the approximating error: A numerical example sup E [ πt n (ϕ) π t (ϕ) ], ˆπ t ˆπ t n. ϕ B(R d ) Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
7 . Particle filters Remarks: Particle filters are recursive algorithms: The approximation for π t and Y t+1 are the only information used in order to obtain the approximation for π t+1. In other words, the information gained from Y 1,..., Y t is embedded in the current approximation. The standard particle filter involves sampling from the prior distribution of the signal and then using a weighted bootstrap technique (or equivalent) with weights defined by the likelihood of the most recent observation data. If d is small to moderate, then the standard particle filter can perform very well in the time parameter t. The function x k g(x k, y k ) can convey a lot of information about the signal, especially so in high dimensions. If this is the case, using the prior transition kernel f (x k 1, x k ) as proposal will be ineffective and standard particle filter will typically perform poorly often requiring that N = O(κ d ). Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
8 . Particle filters Wall clock time per time step (seconds) Algorithm PF STPF Dimension Figure: Computational cost per time step to achieve a predetermined RMSE versus model dimension, for the standard particle filter and the space-time PF. Add-on procedures are required to resolve high-dimensional filtering problems: Tempering * OT prior posterior Sequential DA in space Jittering * Model Reduction (High Low Res)* Nudging Hybrid models Hamiltonian Monte Carlo Informed priors Localization Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
9 . Tempering A tempering procedure Framework: For i = 1 to d {X t } t 0 Markov chain P (X t dx t X t 1 = x t 1 ) = f t (x t x t 1 )dx t, {X t, Y t } t 0 P (Y t dy t X t = x t ) = g t (y t x t )dy t reweight the particle using g 1 d t and (possibly) resample from it move particles using an MCMC that leaves g k d t f t π [0,t 1] invariant Beskos, DC, Jasra, On the stability of SMC methods in high dimensions, Kantas, Beskos, Jasra, Sequential Monte Carlo for inverse problems, Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
10 . Model Reduction Model Reduction (High Low Res)* Model reduction is a valuable methodology that can lead to substantial computational savings This is the case when the order reduction is performed through a coarsening of the grid used for the numerical algorithm that approximates the dynamical system. For the following example we perform and state space order reduction from to = If applied erroneously, order reduction can introduce large errors. Model reduction can be theoretically justified through the continuity of the conditional distribution of the signal on (π 0, g 1,..., g t, K 1,..., K t ). Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
11 . Model Reduction Recall that the recursion formula for the conditional distribution of the signal where dπt dp t p t = π t 1 K t π t = g t p t, = C 1 t g t, where C t R d g t (y t, x t ) p t (dx t ). Remark. The conditional distribution of the signal is a continuous function of (π 0, g 1,..., g t, K 1,..., K t ). In other words if lim ε 0 (πε 0, g1, ε..., gt ε, K1 ε,..., Kt ε ) = (π 0, g 1,..., g t, K 1,..., K t ) in a suitably chosen topology and define p ε t Δ = π ε t 1K ε t π ε t Δ = g ε t p ε t, (1) then lim ε 0 πt ε = π t and lim ε 0 pt ε = p t (again, in a suitably chosen topology). NB. Note that πt ε is no longer the solution of a filtering problem, but simply the solution of the iteration (1) Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
12 . Application to high-dimensional problems A Case Study Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
13 . Motivation State estimation in Numerical Weather Prediction Data Assimilation for NWP set of methodologies that combines past knowledge of a system in the form of a numerical model with new information about that system in the form of observations of that system. designed to improve forecasting, reduce model uncertainties and adjust model parameters. termen used mainly in the computational geoscience community major component of Numerical Weather Prediction Variational DA: combines the model and the data through the optimisation of a given criterion (minimisation of a so-called cost-function). Ensemble based DA: uses a set of model trajectories/possible scenarios that are intermittently updated according to data and are used to infer the past, current or future position of a system. Hurricane Irma forecast: a. ECMWF, b. USA Global Forecast Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
14 . Motivation Numerical weather prediction requires massive computing capabilities 14,000 trillion calculations per second 2 petabytes of memory 460,000 compute cores. 24 petabytes of storage for saving data Dimension of the state space O(10 9 ) Dimension of the observation space O(10 7 ) Our computing capability The Cray XC40 supercomputer (source: Met Office website) 2x Intel Xeon 2.10GHz 32 logical cores 2x Nvidia Quadro M GB memory Dimension of the state space Dimension of the observation space The Beast (source: Wei s office) Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
15 . A stochastic transport model Example: fluid transport dynamics Consider a two dimensional incompressible fluid flow u defined on 2D-torus Ω = [0, L x ] [0, L y ] modelled by the two-dimensional Euler equations with forcing and dampening. Let q = ẑ curl u denote the vorticity of u, where ẑ denotes the z-axis. For a scalar field g : Ω R, we write g = ( y g, x g) T. Let ψ : Ω [0, ) R denote the stream function. t q + (u ) q = Q rq u = ψ Δψ = q. Q is the forcing term given by Q = 0.1 sin (8πx) r is a positive constant - the large scale dissipation time scale. we consider a slip flow boundary condition ψ Ω = 0. evolution of Lagrangian fluid parcels dx t dt = u(x t, t). Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
16 . A stochastic transport model The filtering problem Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
17 . A stochastic transport model Signal: A discrete space/time approximation of the PDE/SPDE on the domain [0, 1] 2 PDE System SPDE System t ω + u ω = Q rω dq + ū qdt + i ξ i q dw i t = (Q rq) dt u = ψ ū = ψ Δψ = ω Δ ψ = q with Q = 0.1 sin (8πx), r = Boundary Condition ψ Ω = 0. The space/time approximation is generated using a mixed continuous and discontinuous Galerkin finite element scheme + an optimal third order strong stability preserving Runge-Kutta, [Bernsen et al 2006, Gottlieb 2005]. PDE SPDE Grid Resolution 512x512 64x64 Numerical Δt Spin-up 40 ett ett: eddy turnover time L/u L 2.5 time units. Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
18 Initial configuration for the vorticity. A stochastic transport model ω spin = sin(8πx) sin(8πy) cos(6πx) cos(6πy) cos(10πx) cos(4πy) sin(2πy) sin(2πx) from which we spin up the system until an energy equilibrium state seems to have been reached. This equilibrium state, denoted by ω initial, is then chosen as the initial condition. (2) Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
19 . A stochastic transport model Plot of the numerical PDE solution at the initial time t initial and its coarse-grained version done via spatial averaging and projection of the fine grid stream-function to the coarse grid. Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
20 . A stochastic transport model Plot of the numerical PDE solution at the final time t = t initial large eddy turnover times (ett). The coarse-graining is done via spatial averaging and projection of the fine grid streamfunction to the coarse grid. Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
21 . A stochastic transport model Observations: u is observed on a subgrid of the signal grid (9 9 points) { u SPDE t (x) + αz x, z x N (0, 1) Experiment 1 Y t (x) = u PDE t (x) + αz x, z x N (0, 1) Experiment 2 α is calibrated to the standard deviation of the true solution over a coarse grid cell. Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
22 . SIR fails Histogram of weights The Standard Particle Filter fails! Figure: example: loglikelihoods histogram, period 1 ett, 100 particles Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
23 . SIR fails A (non-standard) particle filter Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
24 . Initial condition Initial Condition A good choice of the initial condition is esential for the successful implementation of the filter. In practice it is a reflection of the level of uncertainty of the estimate of initial position of the dynamical system. We use the initial condition is to obtain an ensemble which contain particles that are reasonably close to the truth. Choice for the running example deformation - physically consistent with the system, casimirs preserved. We take a nominal value ω t0 and deform it using the following modified Euler equation: tω + β i u(τ i ) ω = 0 (3) where β i N (0, ɛ), i = 1,..., N p are centered Gaussian weights with an apriori variance parameter ɛ, and τ i U (t initial, t 0 ), i = 1,..., N p are uniform random numbers. Thus each u (τ i ) corresponds to a PDE solution in the time period [t initial, t 0 ). Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
25 . Initial condition Alternative choices q + ζ where ζ is gaussian random field, doable but not physical, only works for q because it s the least smooth of the three fields of interest. The other fields are spatially smooth. also this breaks the SPDE well-posedness theorem (function space regularity). Figure (ux,uy) directly perturb ψ, by ψ + ˉψ where ˉψ = (I κδ) 1 ζ invert elliptic operator with boundary condition ˉψ = 0. Figure (ux, uy) Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
26 . Add-on techniques Model Reduction (High Low Res)* For the current numerical example we perform a state space order reduction from to = How do we reduce the model: We use a Stochastic PDE defined on a coarser grid: t q + (u ) q + (ξ k ) q dbt k = Q rq k=1 u = ψ Δψ = q. ξ k are divergence free given vector fields ξ k are computed from the true solution by using an empirical orthogonal functions (EOFs) procedure are scalar independent Brownian motions B k t dx t = u(x t, t) dt + i ξ i (x t ) dw i (t). Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
27 . Methodology to calibrate the additional stochasticity The reason for this stochastic parametrization is grounded in solid physical considerations, see D.D. Holm, Variational principles for stochastic fluids, Proc.Roy. Soc. A, dx t = u f t (x t )dt dx t = ut c (x t )dt + ξ i (x t ) dw i (t) For each m = 0, 1,..., M 1 i 1 Solve dxij f (t)/dt = uf t (x ij f f (t)) with initial condition xij (mδt ) = x ij. 2 Compute ut c by low-pass filtering ut f along the trajectory. 3 Compute xij c c (t) by solving dxij (t)/dt = uc t (x ij f (t)) with the same initial condition. 4 Compute the difference Δxij m = xij f c ((m + 1)ΔT ) xij ((m + 1)ΔT ), which measures the error between the fine and coarse trajectory. Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
28 . Methodology to calibrate the additional stochasticity Having obtained Δxij m, we would like to extract the basis for the noise. This amounts to a Gaussian model of the form Δxij m = Δx ij + δt N ξij k ΔW m, k where ΔWm k are i.i.d. Normal random variables with mean zero, variance 1. We estimate ξ by minimising 2 Δxij m N E Δx ij ξij k ΔW m k, δt ijm where the choice of N can be informed by using Empirical Orthogonal Function (EOFs). k=1 k=1 Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
29 . Methodology to calibrate the additional stochasticity Number of sources of noise (EoFs - empirical orthogonal functions) decide on a case by case basis too many will slow down the algorithm On the left: Number of EOFs 90% variance vs 50% (no change). On the right: Normalised spectrum of the Δx covariance operator, showing number of eofs required to capture 50%, 70% and 90% of total variance Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
30 . Methodology to calibrate the additional stochasticity Model reduction UQ pictures sytems 512x512 vs 128x128 vs 64x64 (a) ux (b) uy Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
31 . Methodology to calibrate the additional stochasticity (a) psi (b) q Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
32 . Methodology to calibrate the additional stochasticity The aim of the calibration is to capture the statistical properties of the fast fluctuations, rather than the trajectory of the flow. Validation of the stochastic parameterisation in terms of uncertainty quantification for the SPDE. Performance of DA algorithm relies on the correct modelling of the unresolved scales. Evaluating the uncertainty arising from the choice of EOFs is be part of the particle filter implementation. Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
33 . The tempering procedure The tempering procedure Initialisation t=0: For i = 1,..., n, sample q i 0 from π 0. Iteration (t i 1, t i ]: Given the ensemble {X n (t i 1 )} n=1,...,n, 1 Evolve X n (t i 1 ) using the SPDE to obtain X n (t i ). 2 Given X := {X n (t i )} n=1,...,n, define normalised tempered weights ˉλ n,i (φ, X) := exp ( φλ n,i) m exp ( φλ m,i) where the dependence on X means the Λ n,i are computed using X. Define effective sample size ESS i (φ, X) := ˉλi (φ, X) 1. l 2 Set φ = While ESS i (φ, X) < N threshold do: (a) Find 1 φ < φ < 1 such that ESS i (φ (1 φ), X) N threshold. Resample according to ˉλ n,i (φ (1 φ), X) and apply MCMC if required (i.e. when there are duplicated particles), to obtain a new set of particles X (φ ). Set φ = 1 φ and X = X (φ ). (b) If ESS i N threshold then STOP and go to the next filtering step with {( Xn (t i ), ˉλ n,i )} n=1,...,n. Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
34 . The tempering procedure MCMC Mutation (Jittering) Algorithm Given the ensemble {X n,k (t i )} n=1,...,n corresponding to the k th tempering step with temperatureφ k, and proposal step size ρ [0, 1], repeat the following steps. Propose ( X n (t i ) = G X n (t i 1 ), ρw (t i 1 : t i ; ω) + ) 1 ρ 2 Z (t i 1 : t i ; ω) where X n (t i ) = G (X n (t i 1 ), W (t i 1 : t i ; ω)), and W Z. Accept X n (t i ) with probability ( ˉλ φ k, X ) n (t i ) 1 ˉλ (φ k, X n (t i )) where λ (φ, x) = exp ( φλ(x)) is the unnormalised weight function. Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
35 . Implementation parameter Number of Particles decide on a case by case basis too few will not give a reasonable solution too many will slow down the algorithm Picture Number of particles 225 (good) vs 500 (no change), 225 (good) vs 25 (less good) 25 seems ok but we want as many as computationally feasible to tune the algorithm (a) psi 225 vs 500 (b) psi 225 vs 25 Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
36 . Resampling Intervals Resampling Intervals small resampling intervals lead to an unreasonable increase in the computational effort large resampling intervals make the algorithm fail the ESS can be used as criterion for choosing the resampling time adapted resampling times can be used ESS evolution in time/observation noise. DA Solution for DA periods: 1 ETT and 0.2 ETT (a) ux (b) uy (c) psi (d) q Figure: DA: obs spde, period 0.2 ett Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
37 . Results (a) ux (b) uy (c) psi (d) q Figure: DA: obs pde, period 0.2 ett (a) ux (b) uy (c) psi (d) q Figure: DA: obs pde, period 1 ett Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
38 . Results Number of tempering steps/average MCMC steps Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
39 . Final Remarks Final remarks Particle filters/sequential Monte Carlo methods are theoretically justified algorithms for approximating the state of dynamical systems partially (and noisily) observed. Standard particle filters do NOT work in high dimensions. Properly calibrated and modified, particle filters can be used to solve high dimensional problems (see also the work of Peter Jan van Leeuwen, Roland Potthast, Hans Kunsch). Important parameters: initial condition, number of particle, number of observations, correction times, observation error, etc. One can use methodology to assess the reliability of ensemble forecasting system. Add-on techniques: Tempering * OT prior posterior Sequential DA in space * Jittering * Model Reduction (High Low Res)* Nudging Hybrid models Hamiltonian Monte Carlo Informed priors Localization Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39
Particle Filters in High Dimensions
Particle Filters in High Dimensions Dan Crisan Imperial College London Workshop - Simulation and probability: recent trends The Henri Lebesgue Center for Mathematics 5-8 June 2018 Rennes Dan Crisan (Imperial
More informationSequential Monte Carlo Samplers for Applications in High Dimensions
Sequential Monte Carlo Samplers for Applications in High Dimensions Alexandros Beskos National University of Singapore KAUST, 26th February 2014 Joint work with: Dan Crisan, Ajay Jasra, Nik Kantas, Alex
More informationSequential Monte Carlo Methods in High Dimensions
Sequential Monte Carlo Methods in High Dimensions Alexandros Beskos Statistical Science, UCL Oxford, 24th September 2012 Joint work with: Dan Crisan, Ajay Jasra, Nik Kantas, Andrew Stuart Imperial College,
More informationSequential Monte Carlo Methods for Bayesian Computation
Sequential Monte Carlo Methods for Bayesian Computation A. Doucet Kyoto Sept. 2012 A. Doucet (MLSS Sept. 2012) Sept. 2012 1 / 136 Motivating Example 1: Generic Bayesian Model Let X be a vector parameter
More informationErgodicity in data assimilation methods
Ergodicity in data assimilation methods David Kelly Andy Majda Xin Tong Courant Institute New York University New York NY www.dtbkelly.com April 15, 2016 ETH Zurich David Kelly (CIMS) Data assimilation
More informationApproximate Bayesian Computation and Particle Filters
Approximate Bayesian Computation and Particle Filters Dennis Prangle Reading University 5th February 2014 Introduction Talk is mostly a literature review A few comments on my own ongoing research See Jasra
More informationData assimilation as an optimal control problem and applications to UQ
Data assimilation as an optimal control problem and applications to UQ Walter Acevedo, Angwenyi David, Jana de Wiljes & Sebastian Reich Universität Potsdam/ University of Reading IPAM, November 13th 2017
More informationMultilevel Sequential 2 Monte Carlo for Bayesian Inverse Problems
Jonas Latz 1 Multilevel Sequential 2 Monte Carlo for Bayesian Inverse Problems Jonas Latz Technische Universität München Fakultät für Mathematik Lehrstuhl für Numerische Mathematik jonas.latz@tum.de November
More informationData assimilation Schrödinger s perspective
Data assimilation Schrödinger s perspective Sebastian Reich (www.sfb1294.de) Universität Potsdam/ University of Reading IMS NUS, August 3, 218 Universität Potsdam/ University of Reading 1 Core components
More informationParticle Filters: Convergence Results and High Dimensions
Particle Filters: Convergence Results and High Dimensions Mark Coates mark.coates@mcgill.ca McGill University Department of Electrical and Computer Engineering Montreal, Quebec, Canada Bellairs 2012 Outline
More informationEnKF-based particle filters
EnKF-based particle filters Jana de Wiljes, Sebastian Reich, Wilhelm Stannat, Walter Acevedo June 20, 2017 Filtering Problem Signal dx t = f (X t )dt + 2CdW t Observations dy t = h(x t )dt + R 1/2 dv t.
More informationLagrangian Data Assimilation and Its Application to Geophysical Fluid Flows
Lagrangian Data Assimilation and Its Application to Geophysical Fluid Flows Laura Slivinski June, 3 Laura Slivinski (Brown University) Lagrangian Data Assimilation June, 3 / 3 Data Assimilation Setup:
More informationRegularization by noise in infinite dimensions
Regularization by noise in infinite dimensions Franco Flandoli, University of Pisa King s College 2017 Franco Flandoli, University of Pisa () Regularization by noise King s College 2017 1 / 33 Plan of
More informationA Note on the Particle Filter with Posterior Gaussian Resampling
Tellus (6), 8A, 46 46 Copyright C Blackwell Munksgaard, 6 Printed in Singapore. All rights reserved TELLUS A Note on the Particle Filter with Posterior Gaussian Resampling By X. XIONG 1,I.M.NAVON 1,2 and
More informationBayesian parameter estimation in predictive engineering
Bayesian parameter estimation in predictive engineering Damon McDougall Institute for Computational Engineering and Sciences, UT Austin 14th August 2014 1/27 Motivation Understand physical phenomena Observations
More informationBayesian Calibration of Simulators with Structured Discretization Uncertainty
Bayesian Calibration of Simulators with Structured Discretization Uncertainty Oksana A. Chkrebtii Department of Statistics, The Ohio State University Joint work with Matthew T. Pratola (Statistics, The
More informationMethods of Data Assimilation and Comparisons for Lagrangian Data
Methods of Data Assimilation and Comparisons for Lagrangian Data Chris Jones, Warwick and UNC-CH Kayo Ide, UCLA Andrew Stuart, Jochen Voss, Warwick Guillaume Vernieres, UNC-CH Amarjit Budiraja, UNC-CH
More informationNonparametric Drift Estimation for Stochastic Differential Equations
Nonparametric Drift Estimation for Stochastic Differential Equations Gareth Roberts 1 Department of Statistics University of Warwick Brazilian Bayesian meeting, March 2010 Joint work with O. Papaspiliopoulos,
More informationModel Uncertainty Quantification for Data Assimilation in partially observed Lorenz 96
Model Uncertainty Quantification for Data Assimilation in partially observed Lorenz 96 Sahani Pathiraja, Peter Jan Van Leeuwen Institut für Mathematik Universität Potsdam With thanks: Sebastian Reich,
More informationData assimilation in high dimensions
Data assimilation in high dimensions David Kelly Courant Institute New York University New York NY www.dtbkelly.com February 12, 2015 Graduate seminar, CIMS David Kelly (CIMS) Data assimilation February
More informationStochastic Collocation Methods for Polynomial Chaos: Analysis and Applications
Stochastic Collocation Methods for Polynomial Chaos: Analysis and Applications Dongbin Xiu Department of Mathematics, Purdue University Support: AFOSR FA955-8-1-353 (Computational Math) SF CAREER DMS-64535
More informationData assimilation in high dimensions
Data assimilation in high dimensions David Kelly Kody Law Andy Majda Andrew Stuart Xin Tong Courant Institute New York University New York NY www.dtbkelly.com February 3, 2016 DPMMS, University of Cambridge
More informationBayesian Inverse problem, Data assimilation and Localization
Bayesian Inverse problem, Data assimilation and Localization Xin T Tong National University of Singapore ICIP, Singapore 2018 X.Tong Localization 1 / 37 Content What is Bayesian inverse problem? What is
More informationGaussian Process Approximations of Stochastic Differential Equations
Gaussian Process Approximations of Stochastic Differential Equations Cédric Archambeau Centre for Computational Statistics and Machine Learning University College London c.archambeau@cs.ucl.ac.uk CSML
More informationExercises Tutorial at ICASSP 2016 Learning Nonlinear Dynamical Models Using Particle Filters
Exercises Tutorial at ICASSP 216 Learning Nonlinear Dynamical Models Using Particle Filters Andreas Svensson, Johan Dahlin and Thomas B. Schön March 18, 216 Good luck! 1 [Bootstrap particle filter for
More informationMCMC Sampling for Bayesian Inference using L1-type Priors
MÜNSTER MCMC Sampling for Bayesian Inference using L1-type Priors (what I do whenever the ill-posedness of EEG/MEG is just not frustrating enough!) AG Imaging Seminar Felix Lucka 26.06.2012 , MÜNSTER Sampling
More informationLagrangian Data Assimilation and Manifold Detection for a Point-Vortex Model. David Darmon, AMSC Kayo Ide, AOSC, IPST, CSCAMM, ESSIC
Lagrangian Data Assimilation and Manifold Detection for a Point-Vortex Model David Darmon, AMSC Kayo Ide, AOSC, IPST, CSCAMM, ESSIC Background Data Assimilation Iterative process Forecast Analysis Background
More informationAdaptive Population Monte Carlo
Adaptive Population Monte Carlo Olivier Cappé Centre Nat. de la Recherche Scientifique & Télécom Paris 46 rue Barrault, 75634 Paris cedex 13, France http://www.tsi.enst.fr/~cappe/ Recent Advances in Monte
More informationState-Space Methods for Inferring Spike Trains from Calcium Imaging
State-Space Methods for Inferring Spike Trains from Calcium Imaging Joshua Vogelstein Johns Hopkins April 23, 2009 Joshua Vogelstein (Johns Hopkins) State-Space Calcium Imaging April 23, 2009 1 / 78 Outline
More informationLatent state estimation using control theory
Latent state estimation using control theory Bert Kappen SNN Donders Institute, Radboud University, Nijmegen Gatsby Unit, UCL London August 3, 7 with Hans Christian Ruiz Bert Kappen Smoothing problem Given
More informationLatest thoughts on stochastic kinetic energy backscatter - good and bad
Latest thoughts on stochastic kinetic energy backscatter - good and bad by Glenn Shutts DARC Reading University May 15 2013 Acknowledgments ECMWF for supporting this work Martin Leutbecher Martin Steinheimer
More informationSequential Monte Carlo Methods for High-Dimensional Inverse Problems: A case study for the Navier-Stokes equations
SIAM/ASA J. UNCERTAINTY QUANTIFICATION Vol. xx, pp. x c xxxx Society for Industrial and Applied Mathematics x x Sequential Monte Carlo Methods for High-Dimensional Inverse Problems: A case study for the
More informationAdvanced uncertainty evaluation of climate models by Monte Carlo methods
Advanced uncertainty evaluation of climate models by Monte Carlo methods Marko Laine marko.laine@fmi.fi Pirkka Ollinaho, Janne Hakkarainen, Johanna Tamminen, Heikki Järvinen (FMI) Antti Solonen, Heikki
More informationA new Hierarchical Bayes approach to ensemble-variational data assimilation
A new Hierarchical Bayes approach to ensemble-variational data assimilation Michael Tsyrulnikov and Alexander Rakitko HydroMetCenter of Russia College Park, 20 Oct 2014 Michael Tsyrulnikov and Alexander
More information4. DATA ASSIMILATION FUNDAMENTALS
4. DATA ASSIMILATION FUNDAMENTALS... [the atmosphere] "is a chaotic system in which errors introduced into the system can grow with time... As a consequence, data assimilation is a struggle between chaotic
More informationTSRT14: Sensor Fusion Lecture 8
TSRT14: Sensor Fusion Lecture 8 Particle filter theory Marginalized particle filter Gustaf Hendeby gustaf.hendeby@liu.se TSRT14 Lecture 8 Gustaf Hendeby Spring 2018 1 / 25 Le 8: particle filter theory,
More informationShort tutorial on data assimilation
Mitglied der Helmholtz-Gemeinschaft Short tutorial on data assimilation 23 June 2015 Wolfgang Kurtz & Harrie-Jan Hendricks Franssen Institute of Bio- and Geosciences IBG-3 (Agrosphere), Forschungszentrum
More informationAuxiliary Particle Methods
Auxiliary Particle Methods Perspectives & Applications Adam M. Johansen 1 adam.johansen@bristol.ac.uk Oxford University Man Institute 29th May 2008 1 Collaborators include: Arnaud Doucet, Nick Whiteley
More informationGaussian Process Approximations of Stochastic Differential Equations
Gaussian Process Approximations of Stochastic Differential Equations Cédric Archambeau Dan Cawford Manfred Opper John Shawe-Taylor May, 2006 1 Introduction Some of the most complex models routinely run
More informationSequential Monte Carlo Methods for Bayesian Model Selection in Positron Emission Tomography
Methods for Bayesian Model Selection in Positron Emission Tomography Yan Zhou John A.D. Aston and Adam M. Johansen 6th January 2014 Y. Zhou J. A. D. Aston and A. M. Johansen Outline Positron emission tomography
More informationA new iterated filtering algorithm
A new iterated filtering algorithm Edward Ionides University of Michigan, Ann Arbor ionides@umich.edu Statistics and Nonlinear Dynamics in Biology and Medicine Thursday July 31, 2014 Overview 1 Introduction
More informationConvergence of the Ensemble Kalman Filter in Hilbert Space
Convergence of the Ensemble Kalman Filter in Hilbert Space Jan Mandel Center for Computational Mathematics Department of Mathematical and Statistical Sciences University of Colorado Denver Parts based
More informationIntroduction to Bayesian methods in inverse problems
Introduction to Bayesian methods in inverse problems Ville Kolehmainen 1 1 Department of Applied Physics, University of Eastern Finland, Kuopio, Finland March 4 2013 Manchester, UK. Contents Introduction
More informationPATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA Contents in latter part Linear Dynamical Systems What is different from HMM? Kalman filter Its strength and limitation Particle Filter
More informationIntroduction. log p θ (y k y 1:k 1 ), k=1
ESAIM: PROCEEDINGS, September 2007, Vol.19, 115-120 Christophe Andrieu & Dan Crisan, Editors DOI: 10.1051/proc:071915 PARTICLE FILTER-BASED APPROXIMATE MAXIMUM LIKELIHOOD INFERENCE ASYMPTOTICS IN STATE-SPACE
More informationAdvanced Computational Methods in Statistics: Lecture 5 Sequential Monte Carlo/Particle Filtering
Advanced Computational Methods in Statistics: Lecture 5 Sequential Monte Carlo/Particle Filtering Axel Gandy Department of Mathematics Imperial College London http://www2.imperial.ac.uk/~agandy London
More informationSmoothers: Types and Benchmarks
Smoothers: Types and Benchmarks Patrick N. Raanes Oxford University, NERSC 8th International EnKF Workshop May 27, 2013 Chris Farmer, Irene Moroz Laurent Bertino NERSC Geir Evensen Abstract Talk builds
More informationLecture 6: Multiple Model Filtering, Particle Filtering and Other Approximations
Lecture 6: Multiple Model Filtering, Particle Filtering and Other Approximations Department of Biomedical Engineering and Computational Science Aalto University April 28, 2010 Contents 1 Multiple Model
More informationImprecise Filtering for Spacecraft Navigation
Imprecise Filtering for Spacecraft Navigation Tathagata Basu Cristian Greco Thomas Krak Durham University Strathclyde University Ghent University Filtering for Spacecraft Navigation The General Problem
More informationSequential Monte Carlo Methods (for DSGE Models)
Sequential Monte Carlo Methods (for DSGE Models) Frank Schorfheide University of Pennsylvania, PIER, CEPR, and NBER October 23, 2017 Some References These lectures use material from our joint work: Tempered
More informationDivide-and-Conquer Sequential Monte Carlo
Divide-and-Conquer Joint work with: John Aston, Alexandre Bouchard-Côté, Brent Kirkpatrick, Fredrik Lindsten, Christian Næsseth, Thomas Schön University of Warwick a.m.johansen@warwick.ac.uk http://go.warwick.ac.uk/amjohansen/talks/
More informationA State Space Model for Wind Forecast Correction
A State Space Model for Wind Forecast Correction Valrie Monbe, Pierre Ailliot 2, and Anne Cuzol 1 1 Lab-STICC, Université Européenne de Bretagne, France (e-mail: valerie.monbet@univ-ubs.fr, anne.cuzol@univ-ubs.fr)
More informationWhat do we know about EnKF?
What do we know about EnKF? David Kelly Kody Law Andrew Stuart Andrew Majda Xin Tong Courant Institute New York University New York, NY April 10, 2015 CAOS seminar, Courant. David Kelly (NYU) EnKF April
More informationDynamic System Identification using HDMR-Bayesian Technique
Dynamic System Identification using HDMR-Bayesian Technique *Shereena O A 1) and Dr. B N Rao 2) 1), 2) Department of Civil Engineering, IIT Madras, Chennai 600036, Tamil Nadu, India 1) ce14d020@smail.iitm.ac.in
More informationLearning Static Parameters in Stochastic Processes
Learning Static Parameters in Stochastic Processes Bharath Ramsundar December 14, 2012 1 Introduction Consider a Markovian stochastic process X T evolving (perhaps nonlinearly) over time variable T. We
More informationLecture Particle Filters
FMS161/MASM18 Financial Statistics November 29, 2010 Monte Carlo filters The filter recursions could only be solved for HMMs and for linear, Gaussian models. Idea: Approximate any model with a HMM. Replace
More informationIntroduction to multiscale modeling and simulation. Explicit methods for ODEs : forward Euler. y n+1 = y n + tf(y n ) dy dt = f(y), y(0) = y 0
Introduction to multiscale modeling and simulation Lecture 5 Numerical methods for ODEs, SDEs and PDEs The need for multiscale methods Two generic frameworks for multiscale computation Explicit methods
More informationSampling the posterior: An approach to non-gaussian data assimilation
Physica D 230 (2007) 50 64 www.elsevier.com/locate/physd Sampling the posterior: An approach to non-gaussian data assimilation A. Apte a, M. Hairer b, A.M. Stuart b,, J. Voss b a Department of Mathematics,
More informationSuperparameterization and Dynamic Stochastic Superresolution (DSS) for Filtering Sparse Geophysical Flows
Superparameterization and Dynamic Stochastic Superresolution (DSS) for Filtering Sparse Geophysical Flows June 2013 Outline 1 Filtering Filtering: obtaining the best statistical estimation of a nature
More informationDevelopment of Stochastic Artificial Neural Networks for Hydrological Prediction
Development of Stochastic Artificial Neural Networks for Hydrological Prediction G. B. Kingston, M. F. Lambert and H. R. Maier Centre for Applied Modelling in Water Engineering, School of Civil and Environmental
More informationTutorial on ABC Algorithms
Tutorial on ABC Algorithms Dr Chris Drovandi Queensland University of Technology, Australia c.drovandi@qut.edu.au July 3, 2014 Notation Model parameter θ with prior π(θ) Likelihood is f(ý θ) with observed
More informationBayesian Inference: Principles and Practice 3. Sparse Bayesian Models and the Relevance Vector Machine
Bayesian Inference: Principles and Practice 3. Sparse Bayesian Models and the Relevance Vector Machine Mike Tipping Gaussian prior Marginal prior: single α Independent α Cambridge, UK Lecture 3: Overview
More informationGaussian processes for inference in stochastic differential equations
Gaussian processes for inference in stochastic differential equations Manfred Opper, AI group, TU Berlin November 6, 2017 Manfred Opper, AI group, TU Berlin (TU Berlin) inference in SDE November 6, 2017
More informationECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering
ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering Lecturer: Nikolay Atanasov: natanasov@ucsd.edu Teaching Assistants: Siwei Guo: s9guo@eng.ucsd.edu Anwesan Pal:
More informationA nonparametric ensemble transform method for Bayesian inference
A nonparametric ensemble transform method for Bayesian inference Article Published Version Reich, S. (2013) A nonparametric ensemble transform method for Bayesian inference. SIAM Journal on Scientific
More informationParametric Problems, Stochastics, and Identification
Parametric Problems, Stochastics, and Identification Hermann G. Matthies a B. Rosić ab, O. Pajonk ac, A. Litvinenko a a, b University of Kragujevac c SPT Group, Hamburg wire@tu-bs.de http://www.wire.tu-bs.de
More informationStrong Lens Modeling (II): Statistical Methods
Strong Lens Modeling (II): Statistical Methods Chuck Keeton Rutgers, the State University of New Jersey Probability theory multiple random variables, a and b joint distribution p(a, b) conditional distribution
More informationFundamentals of Data Assimilation
National Center for Atmospheric Research, Boulder, CO USA GSI Data Assimilation Tutorial - June 28-30, 2010 Acknowledgments and References WRFDA Overview (WRF Tutorial Lectures, H. Huang and D. Barker)
More informationData assimilation with and without a model
Data assimilation with and without a model Tim Sauer George Mason University Parameter estimation and UQ U. Pittsburgh Mar. 5, 2017 Partially supported by NSF Most of this work is due to: Tyrus Berry,
More informationApril 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning
for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions
More informationState Space Models for Wind Forecast Correction
for Wind Forecast Correction Valérie 1 Pierre Ailliot 2 Anne Cuzol 1 1 Université de Bretagne Sud 2 Université de Brest MAS - 2008/28/08 Outline 1 2 Linear Model : an adaptive bias correction Non Linear
More informationL09. PARTICLE FILTERING. NA568 Mobile Robotics: Methods & Algorithms
L09. PARTICLE FILTERING NA568 Mobile Robotics: Methods & Algorithms Particle Filters Different approach to state estimation Instead of parametric description of state (and uncertainty), use a set of state
More informationLecture 6: Bayesian Inference in SDE Models
Lecture 6: Bayesian Inference in SDE Models Bayesian Filtering and Smoothing Point of View Simo Särkkä Aalto University Simo Särkkä (Aalto) Lecture 6: Bayesian Inference in SDEs 1 / 45 Contents 1 SDEs
More informationEnKF and filter divergence
EnKF and filter divergence David Kelly Andrew Stuart Kody Law Courant Institute New York University New York, NY dtbkelly.com December 12, 2014 Applied and computational mathematics seminar, NIST. David
More informationDATA ASSIMILATION FOR FLOOD FORECASTING
DATA ASSIMILATION FOR FLOOD FORECASTING Arnold Heemin Delft University of Technology 09/16/14 1 Data assimilation is the incorporation of measurement into a numerical model to improve the model results
More informationDimension-Independent likelihood-informed (DILI) MCMC
Dimension-Independent likelihood-informed (DILI) MCMC Tiangang Cui, Kody Law 2, Youssef Marzouk Massachusetts Institute of Technology 2 Oak Ridge National Laboratory 2 August 25 TC, KL, YM DILI MCMC USC
More informationKernel adaptive Sequential Monte Carlo
Kernel adaptive Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) December 7, 2015 1 / 36 Section 1 Outline
More informationMonte Carlo Approximation of Monte Carlo Filters
Monte Carlo Approximation of Monte Carlo Filters Adam M. Johansen et al. Collaborators Include: Arnaud Doucet, Axel Finke, Anthony Lee, Nick Whiteley 7th January 2014 Context & Outline Filtering in State-Space
More informationA Comparison of Particle Filters for Personal Positioning
VI Hotine-Marussi Symposium of Theoretical and Computational Geodesy May 9-June 6. A Comparison of Particle Filters for Personal Positioning D. Petrovich and R. Piché Institute of Mathematics Tampere University
More informationTowards a Bayesian model for Cyber Security
Towards a Bayesian model for Cyber Security Mark Briers (mbriers@turing.ac.uk) Joint work with Henry Clausen and Prof. Niall Adams (Imperial College London) 27 September 2017 The Alan Turing Institute
More informationPar$cle Filters Part I: Theory. Peter Jan van Leeuwen Data- Assimila$on Research Centre DARC University of Reading
Par$cle Filters Part I: Theory Peter Jan van Leeuwen Data- Assimila$on Research Centre DARC University of Reading Reading July 2013 Why Data Assimila$on Predic$on Model improvement: - Parameter es$ma$on
More informationNegative Association, Ordering and Convergence of Resampling Methods
Negative Association, Ordering and Convergence of Resampling Methods Nicolas Chopin ENSAE, Paristech (Joint work with Mathieu Gerber and Nick Whiteley, University of Bristol) Resampling schemes: Informal
More informationStochastic Spectral Approaches to Bayesian Inference
Stochastic Spectral Approaches to Bayesian Inference Prof. Nathan L. Gibson Department of Mathematics Applied Mathematics and Computation Seminar March 4, 2011 Prof. Gibson (OSU) Spectral Approaches to
More informationSeminar: Data Assimilation
Seminar: Data Assimilation Jonas Latz, Elisabeth Ullmann Chair of Numerical Mathematics (M2) Technical University of Munich Jonas Latz, Elisabeth Ullmann (TUM) Data Assimilation 1 / 28 Prerequisites Bachelor:
More informationAn Efficient Ensemble Data Assimilation Approach To Deal With Range Limited Observation
An Efficient Ensemble Data Assimilation Approach To Deal With Range Limited Observation A. Shah 1,2, M. E. Gharamti 1, L. Bertino 1 1 Nansen Environmental and Remote Sensing Center 2 University of Bergen
More informationV (r,t) = i ˆ u( x, y,z,t) + ˆ j v( x, y,z,t) + k ˆ w( x, y, z,t)
IV. DIFFERENTIAL RELATIONS FOR A FLUID PARTICLE This chapter presents the development and application of the basic differential equations of fluid motion. Simplifications in the general equations and common
More informationSampling from complex probability distributions
Sampling from complex probability distributions Louis J. M. Aslett (louis.aslett@durham.ac.uk) Department of Mathematical Sciences Durham University UTOPIAE Training School II 4 July 2017 1/37 Motivation
More informationBrownian Motion. 1 Definition Brownian Motion Wiener measure... 3
Brownian Motion Contents 1 Definition 2 1.1 Brownian Motion................................. 2 1.2 Wiener measure.................................. 3 2 Construction 4 2.1 Gaussian process.................................
More informationBayesian Methods and Uncertainty Quantification for Nonlinear Inverse Problems
Bayesian Methods and Uncertainty Quantification for Nonlinear Inverse Problems John Bardsley, University of Montana Collaborators: H. Haario, J. Kaipio, M. Laine, Y. Marzouk, A. Seppänen, A. Solonen, Z.
More informationFundamentals of Data Assimila1on
014 GSI Community Tutorial NCAR Foothills Campus, Boulder, CO July 14-16, 014 Fundamentals of Data Assimila1on Milija Zupanski Cooperative Institute for Research in the Atmosphere Colorado State University
More informationComputer Intensive Methods in Mathematical Statistics
Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of
More informationBayesian Dynamic Linear Modelling for. Complex Computer Models
Bayesian Dynamic Linear Modelling for Complex Computer Models Fei Liu, Liang Zhang, Mike West Abstract Computer models may have functional outputs. With no loss of generality, we assume that a single computer
More informationData assimilation with and without a model
Data assimilation with and without a model Tyrus Berry George Mason University NJIT Feb. 28, 2017 Postdoc supported by NSF This work is in collaboration with: Tim Sauer, GMU Franz Hamilton, Postdoc, NCSU
More informationRobotics. Lecture 4: Probabilistic Robotics. See course website for up to date information.
Robotics Lecture 4: Probabilistic Robotics See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College London Review: Sensors
More informationSystematic strategies for real time filtering of turbulent signals in complex systems
Systematic strategies for real time filtering of turbulent signals in complex systems Statistical inversion theory for Gaussian random variables The Kalman Filter for Vector Systems: Reduced Filters and
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is
More informationKernel Sequential Monte Carlo
Kernel Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) * equal contribution April 25, 2016 1 / 37 Section
More informationStatistical Inference and Methods
Department of Mathematics Imperial College London d.stephens@imperial.ac.uk http://stats.ma.ic.ac.uk/ das01/ 31st January 2006 Part VI Session 6: Filtering and Time to Event Data Session 6: Filtering and
More informationAn introduction to Sequential Monte Carlo
An introduction to Sequential Monte Carlo Thang Bui Jes Frellsen Department of Engineering University of Cambridge Research and Communication Club 6 February 2014 1 Sequential Monte Carlo (SMC) methods
More informationInfinite-State Markov-switching for Dynamic. Volatility Models : Web Appendix
Infinite-State Markov-switching for Dynamic Volatility Models : Web Appendix Arnaud Dufays 1 Centre de Recherche en Economie et Statistique March 19, 2014 1 Comparison of the two MS-GARCH approximations
More information