Particle Filters in High Dimensions

Size: px
Start display at page:

Download "Particle Filters in High Dimensions"

Transcription

1 Particle Filters in High Dimensions Dan Crisan Imperial College London Bayesian Computation for High-Dimensional Statistical Models Opening Workshop August 2018 Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

2 . Talk Synopsis Stochastic Filtering Particle filters/sequential Monte Carlo methods Tempering/Model Reduction Case study Motivation: Particle Filters for Numerical Weather Prediction A stochastic transport model and the associated filtering problem Particle filter (initial condition, add-on procedures, parameter choice) Final remarks Research partially supported by EPSRC grant EP/N023781/1. Numerical work done by Wei Pan (Imperial College London). C.J. Cotter, DC, D.D. Holm, W. Pan, I. Shevchenko, Numerically Modelling Stochastic Lie Transport in Fluid Dynamics, C.J. Cotter, DC, D.D. Holm, W. Pan, I. Shevchenko, Sequential Monte Carlo for Stochastic Advection by Lie Transport (SALT): A case study for the damped and forced incompressible 2D stochastic Euler equation, in preparation. A. Beskos, DC, A. Jasra, Ajay; K. Kamatani, Y. Zhou, Y A stable particle filter for a class of high-dimensional state-space models. Adv. in Appl. Probab. 49 (2017). A. Beskos, DC, A. Jasra, On the stability of sequential Monte Carlo methods in high dimensions, Ann. Appl. Probab. 24 (2014). Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

3 . What is stochastic filtering? Stochastic Filtering: The process of using partial observations and a stochastic model to make inferences about an evolving dynamical system. X the signal process - hidden component Y the observation process - the data Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

4 . What is stochastic filtering? The filtering problem: Find the conditional distribution of the signal X t given Y t = σ(y s, s [0, t]), i.e., π t (A) = P(X t A Y t ), t 0, A B(R d ). Discrete framework: {X t, Y t } t 0 Markov process The signal process {X t } t 0 Markov chain, X 0 π 0 (dx 0 ) P (X t dx t X t 1 = x t 1 ) = K t (x t 1, dx t ) = f t (x t x t 1 )dt, Example : X t = b (X t 1 ) + σ (X t 1 ) B t, The observation process B t N (0, 1) i.i.d. P ( Y t dy t X [0,t] = x [0,t], Y [0,t 1] = y [0,t 1] ) = P (Yt dy t X t = x t ) = g t (y t x t )dy Example : Y t = h (X t ) + V t, where X [0,t] (X 0,..., X t ), x [0,t] (x 0,..., x t ). V t N (0, 1) i.i.d. Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

5 . What is stochastic filtering? Notation: posterior measure: the conditional distribution of the signal X t given Y t π t (A) = P(X t A Y t ), t 0, A B(R d ). predictive measure: the conditional distribution of the signal X t given Y t 1 p t (A) = P(X t A Y t 1 ), t 0, A B(R d ). If μ is a measure and k is a kernel, then μk (A) μ (dx) k (x, A). Bayes recursion. Prediction Updating p t = π t 1 K t π t = g t p t In other words, dπt dp t = C 1 t g t, where C t R d g t (y t, x t ) p t (dx t ). Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

6 . Particle filters 1. Class of approximations: Particle filters/sequential Monte Carlo Methods: SMC (a j (t), vj 1 (t),..., vj d (t)) N j=1 }{{}}{{} weight position π t πt N = N j=1 a j (t) δ vj (t) 2. The law of evolution of the approximation: π N t 1 SMC selection {}}{ pt N {}}{ πt N mutation 3. The measure of the approximating error: A numerical example sup E [ πt n (ϕ) π t (ϕ) ], ˆπ t ˆπ t n. ϕ B(R d ) Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

7 . Particle filters Remarks: Particle filters are recursive algorithms: The approximation for π t and Y t+1 are the only information used in order to obtain the approximation for π t+1. In other words, the information gained from Y 1,..., Y t is embedded in the current approximation. The standard particle filter involves sampling from the prior distribution of the signal and then using a weighted bootstrap technique (or equivalent) with weights defined by the likelihood of the most recent observation data. If d is small to moderate, then the standard particle filter can perform very well in the time parameter t. The function x k g(x k, y k ) can convey a lot of information about the signal, especially so in high dimensions. If this is the case, using the prior transition kernel f (x k 1, x k ) as proposal will be ineffective and standard particle filter will typically perform poorly often requiring that N = O(κ d ). Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

8 . Particle filters Wall clock time per time step (seconds) Algorithm PF STPF Dimension Figure: Computational cost per time step to achieve a predetermined RMSE versus model dimension, for the standard particle filter and the space-time PF. Add-on procedures are required to resolve high-dimensional filtering problems: Tempering * OT prior posterior Sequential DA in space Jittering * Model Reduction (High Low Res)* Nudging Hybrid models Hamiltonian Monte Carlo Informed priors Localization Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

9 . Tempering A tempering procedure Framework: For i = 1 to d {X t } t 0 Markov chain P (X t dx t X t 1 = x t 1 ) = f t (x t x t 1 )dx t, {X t, Y t } t 0 P (Y t dy t X t = x t ) = g t (y t x t )dy t reweight the particle using g 1 d t and (possibly) resample from it move particles using an MCMC that leaves g k d t f t π [0,t 1] invariant Beskos, DC, Jasra, On the stability of SMC methods in high dimensions, Kantas, Beskos, Jasra, Sequential Monte Carlo for inverse problems, Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

10 . Model Reduction Model Reduction (High Low Res)* Model reduction is a valuable methodology that can lead to substantial computational savings This is the case when the order reduction is performed through a coarsening of the grid used for the numerical algorithm that approximates the dynamical system. For the following example we perform and state space order reduction from to = If applied erroneously, order reduction can introduce large errors. Model reduction can be theoretically justified through the continuity of the conditional distribution of the signal on (π 0, g 1,..., g t, K 1,..., K t ). Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

11 . Model Reduction Recall that the recursion formula for the conditional distribution of the signal where dπt dp t p t = π t 1 K t π t = g t p t, = C 1 t g t, where C t R d g t (y t, x t ) p t (dx t ). Remark. The conditional distribution of the signal is a continuous function of (π 0, g 1,..., g t, K 1,..., K t ). In other words if lim ε 0 (πε 0, g1, ε..., gt ε, K1 ε,..., Kt ε ) = (π 0, g 1,..., g t, K 1,..., K t ) in a suitably chosen topology and define p ε t Δ = π ε t 1K ε t π ε t Δ = g ε t p ε t, (1) then lim ε 0 πt ε = π t and lim ε 0 pt ε = p t (again, in a suitably chosen topology). NB. Note that πt ε is no longer the solution of a filtering problem, but simply the solution of the iteration (1) Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

12 . Application to high-dimensional problems A Case Study Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

13 . Motivation State estimation in Numerical Weather Prediction Data Assimilation for NWP set of methodologies that combines past knowledge of a system in the form of a numerical model with new information about that system in the form of observations of that system. designed to improve forecasting, reduce model uncertainties and adjust model parameters. termen used mainly in the computational geoscience community major component of Numerical Weather Prediction Variational DA: combines the model and the data through the optimisation of a given criterion (minimisation of a so-called cost-function). Ensemble based DA: uses a set of model trajectories/possible scenarios that are intermittently updated according to data and are used to infer the past, current or future position of a system. Hurricane Irma forecast: a. ECMWF, b. USA Global Forecast Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

14 . Motivation Numerical weather prediction requires massive computing capabilities 14,000 trillion calculations per second 2 petabytes of memory 460,000 compute cores. 24 petabytes of storage for saving data Dimension of the state space O(10 9 ) Dimension of the observation space O(10 7 ) Our computing capability The Cray XC40 supercomputer (source: Met Office website) 2x Intel Xeon 2.10GHz 32 logical cores 2x Nvidia Quadro M GB memory Dimension of the state space Dimension of the observation space The Beast (source: Wei s office) Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

15 . A stochastic transport model Example: fluid transport dynamics Consider a two dimensional incompressible fluid flow u defined on 2D-torus Ω = [0, L x ] [0, L y ] modelled by the two-dimensional Euler equations with forcing and dampening. Let q = ẑ curl u denote the vorticity of u, where ẑ denotes the z-axis. For a scalar field g : Ω R, we write g = ( y g, x g) T. Let ψ : Ω [0, ) R denote the stream function. t q + (u ) q = Q rq u = ψ Δψ = q. Q is the forcing term given by Q = 0.1 sin (8πx) r is a positive constant - the large scale dissipation time scale. we consider a slip flow boundary condition ψ Ω = 0. evolution of Lagrangian fluid parcels dx t dt = u(x t, t). Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

16 . A stochastic transport model The filtering problem Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

17 . A stochastic transport model Signal: A discrete space/time approximation of the PDE/SPDE on the domain [0, 1] 2 PDE System SPDE System t ω + u ω = Q rω dq + ū qdt + i ξ i q dw i t = (Q rq) dt u = ψ ū = ψ Δψ = ω Δ ψ = q with Q = 0.1 sin (8πx), r = Boundary Condition ψ Ω = 0. The space/time approximation is generated using a mixed continuous and discontinuous Galerkin finite element scheme + an optimal third order strong stability preserving Runge-Kutta, [Bernsen et al 2006, Gottlieb 2005]. PDE SPDE Grid Resolution 512x512 64x64 Numerical Δt Spin-up 40 ett ett: eddy turnover time L/u L 2.5 time units. Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

18 Initial configuration for the vorticity. A stochastic transport model ω spin = sin(8πx) sin(8πy) cos(6πx) cos(6πy) cos(10πx) cos(4πy) sin(2πy) sin(2πx) from which we spin up the system until an energy equilibrium state seems to have been reached. This equilibrium state, denoted by ω initial, is then chosen as the initial condition. (2) Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

19 . A stochastic transport model Plot of the numerical PDE solution at the initial time t initial and its coarse-grained version done via spatial averaging and projection of the fine grid stream-function to the coarse grid. Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

20 . A stochastic transport model Plot of the numerical PDE solution at the final time t = t initial large eddy turnover times (ett). The coarse-graining is done via spatial averaging and projection of the fine grid streamfunction to the coarse grid. Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

21 . A stochastic transport model Observations: u is observed on a subgrid of the signal grid (9 9 points) { u SPDE t (x) + αz x, z x N (0, 1) Experiment 1 Y t (x) = u PDE t (x) + αz x, z x N (0, 1) Experiment 2 α is calibrated to the standard deviation of the true solution over a coarse grid cell. Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

22 . SIR fails Histogram of weights The Standard Particle Filter fails! Figure: example: loglikelihoods histogram, period 1 ett, 100 particles Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

23 . SIR fails A (non-standard) particle filter Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

24 . Initial condition Initial Condition A good choice of the initial condition is esential for the successful implementation of the filter. In practice it is a reflection of the level of uncertainty of the estimate of initial position of the dynamical system. We use the initial condition is to obtain an ensemble which contain particles that are reasonably close to the truth. Choice for the running example deformation - physically consistent with the system, casimirs preserved. We take a nominal value ω t0 and deform it using the following modified Euler equation: tω + β i u(τ i ) ω = 0 (3) where β i N (0, ɛ), i = 1,..., N p are centered Gaussian weights with an apriori variance parameter ɛ, and τ i U (t initial, t 0 ), i = 1,..., N p are uniform random numbers. Thus each u (τ i ) corresponds to a PDE solution in the time period [t initial, t 0 ). Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

25 . Initial condition Alternative choices q + ζ where ζ is gaussian random field, doable but not physical, only works for q because it s the least smooth of the three fields of interest. The other fields are spatially smooth. also this breaks the SPDE well-posedness theorem (function space regularity). Figure (ux,uy) directly perturb ψ, by ψ + ˉψ where ˉψ = (I κδ) 1 ζ invert elliptic operator with boundary condition ˉψ = 0. Figure (ux, uy) Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

26 . Add-on techniques Model Reduction (High Low Res)* For the current numerical example we perform a state space order reduction from to = How do we reduce the model: We use a Stochastic PDE defined on a coarser grid: t q + (u ) q + (ξ k ) q dbt k = Q rq k=1 u = ψ Δψ = q. ξ k are divergence free given vector fields ξ k are computed from the true solution by using an empirical orthogonal functions (EOFs) procedure are scalar independent Brownian motions B k t dx t = u(x t, t) dt + i ξ i (x t ) dw i (t). Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

27 . Methodology to calibrate the additional stochasticity The reason for this stochastic parametrization is grounded in solid physical considerations, see D.D. Holm, Variational principles for stochastic fluids, Proc.Roy. Soc. A, dx t = u f t (x t )dt dx t = ut c (x t )dt + ξ i (x t ) dw i (t) For each m = 0, 1,..., M 1 i 1 Solve dxij f (t)/dt = uf t (x ij f f (t)) with initial condition xij (mδt ) = x ij. 2 Compute ut c by low-pass filtering ut f along the trajectory. 3 Compute xij c c (t) by solving dxij (t)/dt = uc t (x ij f (t)) with the same initial condition. 4 Compute the difference Δxij m = xij f c ((m + 1)ΔT ) xij ((m + 1)ΔT ), which measures the error between the fine and coarse trajectory. Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

28 . Methodology to calibrate the additional stochasticity Having obtained Δxij m, we would like to extract the basis for the noise. This amounts to a Gaussian model of the form Δxij m = Δx ij + δt N ξij k ΔW m, k where ΔWm k are i.i.d. Normal random variables with mean zero, variance 1. We estimate ξ by minimising 2 Δxij m N E Δx ij ξij k ΔW m k, δt ijm where the choice of N can be informed by using Empirical Orthogonal Function (EOFs). k=1 k=1 Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

29 . Methodology to calibrate the additional stochasticity Number of sources of noise (EoFs - empirical orthogonal functions) decide on a case by case basis too many will slow down the algorithm On the left: Number of EOFs 90% variance vs 50% (no change). On the right: Normalised spectrum of the Δx covariance operator, showing number of eofs required to capture 50%, 70% and 90% of total variance Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

30 . Methodology to calibrate the additional stochasticity Model reduction UQ pictures sytems 512x512 vs 128x128 vs 64x64 (a) ux (b) uy Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

31 . Methodology to calibrate the additional stochasticity (a) psi (b) q Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

32 . Methodology to calibrate the additional stochasticity The aim of the calibration is to capture the statistical properties of the fast fluctuations, rather than the trajectory of the flow. Validation of the stochastic parameterisation in terms of uncertainty quantification for the SPDE. Performance of DA algorithm relies on the correct modelling of the unresolved scales. Evaluating the uncertainty arising from the choice of EOFs is be part of the particle filter implementation. Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

33 . The tempering procedure The tempering procedure Initialisation t=0: For i = 1,..., n, sample q i 0 from π 0. Iteration (t i 1, t i ]: Given the ensemble {X n (t i 1 )} n=1,...,n, 1 Evolve X n (t i 1 ) using the SPDE to obtain X n (t i ). 2 Given X := {X n (t i )} n=1,...,n, define normalised tempered weights ˉλ n,i (φ, X) := exp ( φλ n,i) m exp ( φλ m,i) where the dependence on X means the Λ n,i are computed using X. Define effective sample size ESS i (φ, X) := ˉλi (φ, X) 1. l 2 Set φ = While ESS i (φ, X) < N threshold do: (a) Find 1 φ < φ < 1 such that ESS i (φ (1 φ), X) N threshold. Resample according to ˉλ n,i (φ (1 φ), X) and apply MCMC if required (i.e. when there are duplicated particles), to obtain a new set of particles X (φ ). Set φ = 1 φ and X = X (φ ). (b) If ESS i N threshold then STOP and go to the next filtering step with {( Xn (t i ), ˉλ n,i )} n=1,...,n. Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

34 . The tempering procedure MCMC Mutation (Jittering) Algorithm Given the ensemble {X n,k (t i )} n=1,...,n corresponding to the k th tempering step with temperatureφ k, and proposal step size ρ [0, 1], repeat the following steps. Propose ( X n (t i ) = G X n (t i 1 ), ρw (t i 1 : t i ; ω) + ) 1 ρ 2 Z (t i 1 : t i ; ω) where X n (t i ) = G (X n (t i 1 ), W (t i 1 : t i ; ω)), and W Z. Accept X n (t i ) with probability ( ˉλ φ k, X ) n (t i ) 1 ˉλ (φ k, X n (t i )) where λ (φ, x) = exp ( φλ(x)) is the unnormalised weight function. Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

35 . Implementation parameter Number of Particles decide on a case by case basis too few will not give a reasonable solution too many will slow down the algorithm Picture Number of particles 225 (good) vs 500 (no change), 225 (good) vs 25 (less good) 25 seems ok but we want as many as computationally feasible to tune the algorithm (a) psi 225 vs 500 (b) psi 225 vs 25 Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

36 . Resampling Intervals Resampling Intervals small resampling intervals lead to an unreasonable increase in the computational effort large resampling intervals make the algorithm fail the ESS can be used as criterion for choosing the resampling time adapted resampling times can be used ESS evolution in time/observation noise. DA Solution for DA periods: 1 ETT and 0.2 ETT (a) ux (b) uy (c) psi (d) q Figure: DA: obs spde, period 0.2 ett Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

37 . Results (a) ux (b) uy (c) psi (d) q Figure: DA: obs pde, period 0.2 ett (a) ux (b) uy (c) psi (d) q Figure: DA: obs pde, period 1 ett Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

38 . Results Number of tempering steps/average MCMC steps Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

39 . Final Remarks Final remarks Particle filters/sequential Monte Carlo methods are theoretically justified algorithms for approximating the state of dynamical systems partially (and noisily) observed. Standard particle filters do NOT work in high dimensions. Properly calibrated and modified, particle filters can be used to solve high dimensional problems (see also the work of Peter Jan van Leeuwen, Roland Potthast, Hans Kunsch). Important parameters: initial condition, number of particle, number of observations, correction times, observation error, etc. One can use methodology to assess the reliability of ensemble forecasting system. Add-on techniques: Tempering * OT prior posterior Sequential DA in space * Jittering * Model Reduction (High Low Res)* Nudging Hybrid models Hamiltonian Monte Carlo Informed priors Localization Dan Crisan (Imperial College London) Particle Filters in High Dimensions 27 August / 39

Particle Filters in High Dimensions

Particle Filters in High Dimensions Particle Filters in High Dimensions Dan Crisan Imperial College London Workshop - Simulation and probability: recent trends The Henri Lebesgue Center for Mathematics 5-8 June 2018 Rennes Dan Crisan (Imperial

More information

Sequential Monte Carlo Samplers for Applications in High Dimensions

Sequential Monte Carlo Samplers for Applications in High Dimensions Sequential Monte Carlo Samplers for Applications in High Dimensions Alexandros Beskos National University of Singapore KAUST, 26th February 2014 Joint work with: Dan Crisan, Ajay Jasra, Nik Kantas, Alex

More information

Sequential Monte Carlo Methods in High Dimensions

Sequential Monte Carlo Methods in High Dimensions Sequential Monte Carlo Methods in High Dimensions Alexandros Beskos Statistical Science, UCL Oxford, 24th September 2012 Joint work with: Dan Crisan, Ajay Jasra, Nik Kantas, Andrew Stuart Imperial College,

More information

Sequential Monte Carlo Methods for Bayesian Computation

Sequential Monte Carlo Methods for Bayesian Computation Sequential Monte Carlo Methods for Bayesian Computation A. Doucet Kyoto Sept. 2012 A. Doucet (MLSS Sept. 2012) Sept. 2012 1 / 136 Motivating Example 1: Generic Bayesian Model Let X be a vector parameter

More information

Ergodicity in data assimilation methods

Ergodicity in data assimilation methods Ergodicity in data assimilation methods David Kelly Andy Majda Xin Tong Courant Institute New York University New York NY www.dtbkelly.com April 15, 2016 ETH Zurich David Kelly (CIMS) Data assimilation

More information

Approximate Bayesian Computation and Particle Filters

Approximate Bayesian Computation and Particle Filters Approximate Bayesian Computation and Particle Filters Dennis Prangle Reading University 5th February 2014 Introduction Talk is mostly a literature review A few comments on my own ongoing research See Jasra

More information

Data assimilation as an optimal control problem and applications to UQ

Data assimilation as an optimal control problem and applications to UQ Data assimilation as an optimal control problem and applications to UQ Walter Acevedo, Angwenyi David, Jana de Wiljes & Sebastian Reich Universität Potsdam/ University of Reading IPAM, November 13th 2017

More information

Multilevel Sequential 2 Monte Carlo for Bayesian Inverse Problems

Multilevel Sequential 2 Monte Carlo for Bayesian Inverse Problems Jonas Latz 1 Multilevel Sequential 2 Monte Carlo for Bayesian Inverse Problems Jonas Latz Technische Universität München Fakultät für Mathematik Lehrstuhl für Numerische Mathematik jonas.latz@tum.de November

More information

Data assimilation Schrödinger s perspective

Data assimilation Schrödinger s perspective Data assimilation Schrödinger s perspective Sebastian Reich (www.sfb1294.de) Universität Potsdam/ University of Reading IMS NUS, August 3, 218 Universität Potsdam/ University of Reading 1 Core components

More information

Particle Filters: Convergence Results and High Dimensions

Particle Filters: Convergence Results and High Dimensions Particle Filters: Convergence Results and High Dimensions Mark Coates mark.coates@mcgill.ca McGill University Department of Electrical and Computer Engineering Montreal, Quebec, Canada Bellairs 2012 Outline

More information

EnKF-based particle filters

EnKF-based particle filters EnKF-based particle filters Jana de Wiljes, Sebastian Reich, Wilhelm Stannat, Walter Acevedo June 20, 2017 Filtering Problem Signal dx t = f (X t )dt + 2CdW t Observations dy t = h(x t )dt + R 1/2 dv t.

More information

Lagrangian Data Assimilation and Its Application to Geophysical Fluid Flows

Lagrangian Data Assimilation and Its Application to Geophysical Fluid Flows Lagrangian Data Assimilation and Its Application to Geophysical Fluid Flows Laura Slivinski June, 3 Laura Slivinski (Brown University) Lagrangian Data Assimilation June, 3 / 3 Data Assimilation Setup:

More information

Regularization by noise in infinite dimensions

Regularization by noise in infinite dimensions Regularization by noise in infinite dimensions Franco Flandoli, University of Pisa King s College 2017 Franco Flandoli, University of Pisa () Regularization by noise King s College 2017 1 / 33 Plan of

More information

A Note on the Particle Filter with Posterior Gaussian Resampling

A Note on the Particle Filter with Posterior Gaussian Resampling Tellus (6), 8A, 46 46 Copyright C Blackwell Munksgaard, 6 Printed in Singapore. All rights reserved TELLUS A Note on the Particle Filter with Posterior Gaussian Resampling By X. XIONG 1,I.M.NAVON 1,2 and

More information

Bayesian parameter estimation in predictive engineering

Bayesian parameter estimation in predictive engineering Bayesian parameter estimation in predictive engineering Damon McDougall Institute for Computational Engineering and Sciences, UT Austin 14th August 2014 1/27 Motivation Understand physical phenomena Observations

More information

Bayesian Calibration of Simulators with Structured Discretization Uncertainty

Bayesian Calibration of Simulators with Structured Discretization Uncertainty Bayesian Calibration of Simulators with Structured Discretization Uncertainty Oksana A. Chkrebtii Department of Statistics, The Ohio State University Joint work with Matthew T. Pratola (Statistics, The

More information

Methods of Data Assimilation and Comparisons for Lagrangian Data

Methods of Data Assimilation and Comparisons for Lagrangian Data Methods of Data Assimilation and Comparisons for Lagrangian Data Chris Jones, Warwick and UNC-CH Kayo Ide, UCLA Andrew Stuart, Jochen Voss, Warwick Guillaume Vernieres, UNC-CH Amarjit Budiraja, UNC-CH

More information

Nonparametric Drift Estimation for Stochastic Differential Equations

Nonparametric Drift Estimation for Stochastic Differential Equations Nonparametric Drift Estimation for Stochastic Differential Equations Gareth Roberts 1 Department of Statistics University of Warwick Brazilian Bayesian meeting, March 2010 Joint work with O. Papaspiliopoulos,

More information

Model Uncertainty Quantification for Data Assimilation in partially observed Lorenz 96

Model Uncertainty Quantification for Data Assimilation in partially observed Lorenz 96 Model Uncertainty Quantification for Data Assimilation in partially observed Lorenz 96 Sahani Pathiraja, Peter Jan Van Leeuwen Institut für Mathematik Universität Potsdam With thanks: Sebastian Reich,

More information

Data assimilation in high dimensions

Data assimilation in high dimensions Data assimilation in high dimensions David Kelly Courant Institute New York University New York NY www.dtbkelly.com February 12, 2015 Graduate seminar, CIMS David Kelly (CIMS) Data assimilation February

More information

Stochastic Collocation Methods for Polynomial Chaos: Analysis and Applications

Stochastic Collocation Methods for Polynomial Chaos: Analysis and Applications Stochastic Collocation Methods for Polynomial Chaos: Analysis and Applications Dongbin Xiu Department of Mathematics, Purdue University Support: AFOSR FA955-8-1-353 (Computational Math) SF CAREER DMS-64535

More information

Data assimilation in high dimensions

Data assimilation in high dimensions Data assimilation in high dimensions David Kelly Kody Law Andy Majda Andrew Stuart Xin Tong Courant Institute New York University New York NY www.dtbkelly.com February 3, 2016 DPMMS, University of Cambridge

More information

Bayesian Inverse problem, Data assimilation and Localization

Bayesian Inverse problem, Data assimilation and Localization Bayesian Inverse problem, Data assimilation and Localization Xin T Tong National University of Singapore ICIP, Singapore 2018 X.Tong Localization 1 / 37 Content What is Bayesian inverse problem? What is

More information

Gaussian Process Approximations of Stochastic Differential Equations

Gaussian Process Approximations of Stochastic Differential Equations Gaussian Process Approximations of Stochastic Differential Equations Cédric Archambeau Centre for Computational Statistics and Machine Learning University College London c.archambeau@cs.ucl.ac.uk CSML

More information

Exercises Tutorial at ICASSP 2016 Learning Nonlinear Dynamical Models Using Particle Filters

Exercises Tutorial at ICASSP 2016 Learning Nonlinear Dynamical Models Using Particle Filters Exercises Tutorial at ICASSP 216 Learning Nonlinear Dynamical Models Using Particle Filters Andreas Svensson, Johan Dahlin and Thomas B. Schön March 18, 216 Good luck! 1 [Bootstrap particle filter for

More information

MCMC Sampling for Bayesian Inference using L1-type Priors

MCMC Sampling for Bayesian Inference using L1-type Priors MÜNSTER MCMC Sampling for Bayesian Inference using L1-type Priors (what I do whenever the ill-posedness of EEG/MEG is just not frustrating enough!) AG Imaging Seminar Felix Lucka 26.06.2012 , MÜNSTER Sampling

More information

Lagrangian Data Assimilation and Manifold Detection for a Point-Vortex Model. David Darmon, AMSC Kayo Ide, AOSC, IPST, CSCAMM, ESSIC

Lagrangian Data Assimilation and Manifold Detection for a Point-Vortex Model. David Darmon, AMSC Kayo Ide, AOSC, IPST, CSCAMM, ESSIC Lagrangian Data Assimilation and Manifold Detection for a Point-Vortex Model David Darmon, AMSC Kayo Ide, AOSC, IPST, CSCAMM, ESSIC Background Data Assimilation Iterative process Forecast Analysis Background

More information

Adaptive Population Monte Carlo

Adaptive Population Monte Carlo Adaptive Population Monte Carlo Olivier Cappé Centre Nat. de la Recherche Scientifique & Télécom Paris 46 rue Barrault, 75634 Paris cedex 13, France http://www.tsi.enst.fr/~cappe/ Recent Advances in Monte

More information

State-Space Methods for Inferring Spike Trains from Calcium Imaging

State-Space Methods for Inferring Spike Trains from Calcium Imaging State-Space Methods for Inferring Spike Trains from Calcium Imaging Joshua Vogelstein Johns Hopkins April 23, 2009 Joshua Vogelstein (Johns Hopkins) State-Space Calcium Imaging April 23, 2009 1 / 78 Outline

More information

Latent state estimation using control theory

Latent state estimation using control theory Latent state estimation using control theory Bert Kappen SNN Donders Institute, Radboud University, Nijmegen Gatsby Unit, UCL London August 3, 7 with Hans Christian Ruiz Bert Kappen Smoothing problem Given

More information

Latest thoughts on stochastic kinetic energy backscatter - good and bad

Latest thoughts on stochastic kinetic energy backscatter - good and bad Latest thoughts on stochastic kinetic energy backscatter - good and bad by Glenn Shutts DARC Reading University May 15 2013 Acknowledgments ECMWF for supporting this work Martin Leutbecher Martin Steinheimer

More information

Sequential Monte Carlo Methods for High-Dimensional Inverse Problems: A case study for the Navier-Stokes equations

Sequential Monte Carlo Methods for High-Dimensional Inverse Problems: A case study for the Navier-Stokes equations SIAM/ASA J. UNCERTAINTY QUANTIFICATION Vol. xx, pp. x c xxxx Society for Industrial and Applied Mathematics x x Sequential Monte Carlo Methods for High-Dimensional Inverse Problems: A case study for the

More information

Advanced uncertainty evaluation of climate models by Monte Carlo methods

Advanced uncertainty evaluation of climate models by Monte Carlo methods Advanced uncertainty evaluation of climate models by Monte Carlo methods Marko Laine marko.laine@fmi.fi Pirkka Ollinaho, Janne Hakkarainen, Johanna Tamminen, Heikki Järvinen (FMI) Antti Solonen, Heikki

More information

A new Hierarchical Bayes approach to ensemble-variational data assimilation

A new Hierarchical Bayes approach to ensemble-variational data assimilation A new Hierarchical Bayes approach to ensemble-variational data assimilation Michael Tsyrulnikov and Alexander Rakitko HydroMetCenter of Russia College Park, 20 Oct 2014 Michael Tsyrulnikov and Alexander

More information

4. DATA ASSIMILATION FUNDAMENTALS

4. DATA ASSIMILATION FUNDAMENTALS 4. DATA ASSIMILATION FUNDAMENTALS... [the atmosphere] "is a chaotic system in which errors introduced into the system can grow with time... As a consequence, data assimilation is a struggle between chaotic

More information

TSRT14: Sensor Fusion Lecture 8

TSRT14: Sensor Fusion Lecture 8 TSRT14: Sensor Fusion Lecture 8 Particle filter theory Marginalized particle filter Gustaf Hendeby gustaf.hendeby@liu.se TSRT14 Lecture 8 Gustaf Hendeby Spring 2018 1 / 25 Le 8: particle filter theory,

More information

Short tutorial on data assimilation

Short tutorial on data assimilation Mitglied der Helmholtz-Gemeinschaft Short tutorial on data assimilation 23 June 2015 Wolfgang Kurtz & Harrie-Jan Hendricks Franssen Institute of Bio- and Geosciences IBG-3 (Agrosphere), Forschungszentrum

More information

Auxiliary Particle Methods

Auxiliary Particle Methods Auxiliary Particle Methods Perspectives & Applications Adam M. Johansen 1 adam.johansen@bristol.ac.uk Oxford University Man Institute 29th May 2008 1 Collaborators include: Arnaud Doucet, Nick Whiteley

More information

Gaussian Process Approximations of Stochastic Differential Equations

Gaussian Process Approximations of Stochastic Differential Equations Gaussian Process Approximations of Stochastic Differential Equations Cédric Archambeau Dan Cawford Manfred Opper John Shawe-Taylor May, 2006 1 Introduction Some of the most complex models routinely run

More information

Sequential Monte Carlo Methods for Bayesian Model Selection in Positron Emission Tomography

Sequential Monte Carlo Methods for Bayesian Model Selection in Positron Emission Tomography Methods for Bayesian Model Selection in Positron Emission Tomography Yan Zhou John A.D. Aston and Adam M. Johansen 6th January 2014 Y. Zhou J. A. D. Aston and A. M. Johansen Outline Positron emission tomography

More information

A new iterated filtering algorithm

A new iterated filtering algorithm A new iterated filtering algorithm Edward Ionides University of Michigan, Ann Arbor ionides@umich.edu Statistics and Nonlinear Dynamics in Biology and Medicine Thursday July 31, 2014 Overview 1 Introduction

More information

Convergence of the Ensemble Kalman Filter in Hilbert Space

Convergence of the Ensemble Kalman Filter in Hilbert Space Convergence of the Ensemble Kalman Filter in Hilbert Space Jan Mandel Center for Computational Mathematics Department of Mathematical and Statistical Sciences University of Colorado Denver Parts based

More information

Introduction to Bayesian methods in inverse problems

Introduction to Bayesian methods in inverse problems Introduction to Bayesian methods in inverse problems Ville Kolehmainen 1 1 Department of Applied Physics, University of Eastern Finland, Kuopio, Finland March 4 2013 Manchester, UK. Contents Introduction

More information

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA Contents in latter part Linear Dynamical Systems What is different from HMM? Kalman filter Its strength and limitation Particle Filter

More information

Introduction. log p θ (y k y 1:k 1 ), k=1

Introduction. log p θ (y k y 1:k 1 ), k=1 ESAIM: PROCEEDINGS, September 2007, Vol.19, 115-120 Christophe Andrieu & Dan Crisan, Editors DOI: 10.1051/proc:071915 PARTICLE FILTER-BASED APPROXIMATE MAXIMUM LIKELIHOOD INFERENCE ASYMPTOTICS IN STATE-SPACE

More information

Advanced Computational Methods in Statistics: Lecture 5 Sequential Monte Carlo/Particle Filtering

Advanced Computational Methods in Statistics: Lecture 5 Sequential Monte Carlo/Particle Filtering Advanced Computational Methods in Statistics: Lecture 5 Sequential Monte Carlo/Particle Filtering Axel Gandy Department of Mathematics Imperial College London http://www2.imperial.ac.uk/~agandy London

More information

Smoothers: Types and Benchmarks

Smoothers: Types and Benchmarks Smoothers: Types and Benchmarks Patrick N. Raanes Oxford University, NERSC 8th International EnKF Workshop May 27, 2013 Chris Farmer, Irene Moroz Laurent Bertino NERSC Geir Evensen Abstract Talk builds

More information

Lecture 6: Multiple Model Filtering, Particle Filtering and Other Approximations

Lecture 6: Multiple Model Filtering, Particle Filtering and Other Approximations Lecture 6: Multiple Model Filtering, Particle Filtering and Other Approximations Department of Biomedical Engineering and Computational Science Aalto University April 28, 2010 Contents 1 Multiple Model

More information

Imprecise Filtering for Spacecraft Navigation

Imprecise Filtering for Spacecraft Navigation Imprecise Filtering for Spacecraft Navigation Tathagata Basu Cristian Greco Thomas Krak Durham University Strathclyde University Ghent University Filtering for Spacecraft Navigation The General Problem

More information

Sequential Monte Carlo Methods (for DSGE Models)

Sequential Monte Carlo Methods (for DSGE Models) Sequential Monte Carlo Methods (for DSGE Models) Frank Schorfheide University of Pennsylvania, PIER, CEPR, and NBER October 23, 2017 Some References These lectures use material from our joint work: Tempered

More information

Divide-and-Conquer Sequential Monte Carlo

Divide-and-Conquer Sequential Monte Carlo Divide-and-Conquer Joint work with: John Aston, Alexandre Bouchard-Côté, Brent Kirkpatrick, Fredrik Lindsten, Christian Næsseth, Thomas Schön University of Warwick a.m.johansen@warwick.ac.uk http://go.warwick.ac.uk/amjohansen/talks/

More information

A State Space Model for Wind Forecast Correction

A State Space Model for Wind Forecast Correction A State Space Model for Wind Forecast Correction Valrie Monbe, Pierre Ailliot 2, and Anne Cuzol 1 1 Lab-STICC, Université Européenne de Bretagne, France (e-mail: valerie.monbet@univ-ubs.fr, anne.cuzol@univ-ubs.fr)

More information

What do we know about EnKF?

What do we know about EnKF? What do we know about EnKF? David Kelly Kody Law Andrew Stuart Andrew Majda Xin Tong Courant Institute New York University New York, NY April 10, 2015 CAOS seminar, Courant. David Kelly (NYU) EnKF April

More information

Dynamic System Identification using HDMR-Bayesian Technique

Dynamic System Identification using HDMR-Bayesian Technique Dynamic System Identification using HDMR-Bayesian Technique *Shereena O A 1) and Dr. B N Rao 2) 1), 2) Department of Civil Engineering, IIT Madras, Chennai 600036, Tamil Nadu, India 1) ce14d020@smail.iitm.ac.in

More information

Learning Static Parameters in Stochastic Processes

Learning Static Parameters in Stochastic Processes Learning Static Parameters in Stochastic Processes Bharath Ramsundar December 14, 2012 1 Introduction Consider a Markovian stochastic process X T evolving (perhaps nonlinearly) over time variable T. We

More information

Lecture Particle Filters

Lecture Particle Filters FMS161/MASM18 Financial Statistics November 29, 2010 Monte Carlo filters The filter recursions could only be solved for HMMs and for linear, Gaussian models. Idea: Approximate any model with a HMM. Replace

More information

Introduction to multiscale modeling and simulation. Explicit methods for ODEs : forward Euler. y n+1 = y n + tf(y n ) dy dt = f(y), y(0) = y 0

Introduction to multiscale modeling and simulation. Explicit methods for ODEs : forward Euler. y n+1 = y n + tf(y n ) dy dt = f(y), y(0) = y 0 Introduction to multiscale modeling and simulation Lecture 5 Numerical methods for ODEs, SDEs and PDEs The need for multiscale methods Two generic frameworks for multiscale computation Explicit methods

More information

Sampling the posterior: An approach to non-gaussian data assimilation

Sampling the posterior: An approach to non-gaussian data assimilation Physica D 230 (2007) 50 64 www.elsevier.com/locate/physd Sampling the posterior: An approach to non-gaussian data assimilation A. Apte a, M. Hairer b, A.M. Stuart b,, J. Voss b a Department of Mathematics,

More information

Superparameterization and Dynamic Stochastic Superresolution (DSS) for Filtering Sparse Geophysical Flows

Superparameterization and Dynamic Stochastic Superresolution (DSS) for Filtering Sparse Geophysical Flows Superparameterization and Dynamic Stochastic Superresolution (DSS) for Filtering Sparse Geophysical Flows June 2013 Outline 1 Filtering Filtering: obtaining the best statistical estimation of a nature

More information

Development of Stochastic Artificial Neural Networks for Hydrological Prediction

Development of Stochastic Artificial Neural Networks for Hydrological Prediction Development of Stochastic Artificial Neural Networks for Hydrological Prediction G. B. Kingston, M. F. Lambert and H. R. Maier Centre for Applied Modelling in Water Engineering, School of Civil and Environmental

More information

Tutorial on ABC Algorithms

Tutorial on ABC Algorithms Tutorial on ABC Algorithms Dr Chris Drovandi Queensland University of Technology, Australia c.drovandi@qut.edu.au July 3, 2014 Notation Model parameter θ with prior π(θ) Likelihood is f(ý θ) with observed

More information

Bayesian Inference: Principles and Practice 3. Sparse Bayesian Models and the Relevance Vector Machine

Bayesian Inference: Principles and Practice 3. Sparse Bayesian Models and the Relevance Vector Machine Bayesian Inference: Principles and Practice 3. Sparse Bayesian Models and the Relevance Vector Machine Mike Tipping Gaussian prior Marginal prior: single α Independent α Cambridge, UK Lecture 3: Overview

More information

Gaussian processes for inference in stochastic differential equations

Gaussian processes for inference in stochastic differential equations Gaussian processes for inference in stochastic differential equations Manfred Opper, AI group, TU Berlin November 6, 2017 Manfred Opper, AI group, TU Berlin (TU Berlin) inference in SDE November 6, 2017

More information

ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering

ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering Lecturer: Nikolay Atanasov: natanasov@ucsd.edu Teaching Assistants: Siwei Guo: s9guo@eng.ucsd.edu Anwesan Pal:

More information

A nonparametric ensemble transform method for Bayesian inference

A nonparametric ensemble transform method for Bayesian inference A nonparametric ensemble transform method for Bayesian inference Article Published Version Reich, S. (2013) A nonparametric ensemble transform method for Bayesian inference. SIAM Journal on Scientific

More information

Parametric Problems, Stochastics, and Identification

Parametric Problems, Stochastics, and Identification Parametric Problems, Stochastics, and Identification Hermann G. Matthies a B. Rosić ab, O. Pajonk ac, A. Litvinenko a a, b University of Kragujevac c SPT Group, Hamburg wire@tu-bs.de http://www.wire.tu-bs.de

More information

Strong Lens Modeling (II): Statistical Methods

Strong Lens Modeling (II): Statistical Methods Strong Lens Modeling (II): Statistical Methods Chuck Keeton Rutgers, the State University of New Jersey Probability theory multiple random variables, a and b joint distribution p(a, b) conditional distribution

More information

Fundamentals of Data Assimilation

Fundamentals of Data Assimilation National Center for Atmospheric Research, Boulder, CO USA GSI Data Assimilation Tutorial - June 28-30, 2010 Acknowledgments and References WRFDA Overview (WRF Tutorial Lectures, H. Huang and D. Barker)

More information

Data assimilation with and without a model

Data assimilation with and without a model Data assimilation with and without a model Tim Sauer George Mason University Parameter estimation and UQ U. Pittsburgh Mar. 5, 2017 Partially supported by NSF Most of this work is due to: Tyrus Berry,

More information

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions

More information

State Space Models for Wind Forecast Correction

State Space Models for Wind Forecast Correction for Wind Forecast Correction Valérie 1 Pierre Ailliot 2 Anne Cuzol 1 1 Université de Bretagne Sud 2 Université de Brest MAS - 2008/28/08 Outline 1 2 Linear Model : an adaptive bias correction Non Linear

More information

L09. PARTICLE FILTERING. NA568 Mobile Robotics: Methods & Algorithms

L09. PARTICLE FILTERING. NA568 Mobile Robotics: Methods & Algorithms L09. PARTICLE FILTERING NA568 Mobile Robotics: Methods & Algorithms Particle Filters Different approach to state estimation Instead of parametric description of state (and uncertainty), use a set of state

More information

Lecture 6: Bayesian Inference in SDE Models

Lecture 6: Bayesian Inference in SDE Models Lecture 6: Bayesian Inference in SDE Models Bayesian Filtering and Smoothing Point of View Simo Särkkä Aalto University Simo Särkkä (Aalto) Lecture 6: Bayesian Inference in SDEs 1 / 45 Contents 1 SDEs

More information

EnKF and filter divergence

EnKF and filter divergence EnKF and filter divergence David Kelly Andrew Stuart Kody Law Courant Institute New York University New York, NY dtbkelly.com December 12, 2014 Applied and computational mathematics seminar, NIST. David

More information

DATA ASSIMILATION FOR FLOOD FORECASTING

DATA ASSIMILATION FOR FLOOD FORECASTING DATA ASSIMILATION FOR FLOOD FORECASTING Arnold Heemin Delft University of Technology 09/16/14 1 Data assimilation is the incorporation of measurement into a numerical model to improve the model results

More information

Dimension-Independent likelihood-informed (DILI) MCMC

Dimension-Independent likelihood-informed (DILI) MCMC Dimension-Independent likelihood-informed (DILI) MCMC Tiangang Cui, Kody Law 2, Youssef Marzouk Massachusetts Institute of Technology 2 Oak Ridge National Laboratory 2 August 25 TC, KL, YM DILI MCMC USC

More information

Kernel adaptive Sequential Monte Carlo

Kernel adaptive Sequential Monte Carlo Kernel adaptive Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) December 7, 2015 1 / 36 Section 1 Outline

More information

Monte Carlo Approximation of Monte Carlo Filters

Monte Carlo Approximation of Monte Carlo Filters Monte Carlo Approximation of Monte Carlo Filters Adam M. Johansen et al. Collaborators Include: Arnaud Doucet, Axel Finke, Anthony Lee, Nick Whiteley 7th January 2014 Context & Outline Filtering in State-Space

More information

A Comparison of Particle Filters for Personal Positioning

A Comparison of Particle Filters for Personal Positioning VI Hotine-Marussi Symposium of Theoretical and Computational Geodesy May 9-June 6. A Comparison of Particle Filters for Personal Positioning D. Petrovich and R. Piché Institute of Mathematics Tampere University

More information

Towards a Bayesian model for Cyber Security

Towards a Bayesian model for Cyber Security Towards a Bayesian model for Cyber Security Mark Briers (mbriers@turing.ac.uk) Joint work with Henry Clausen and Prof. Niall Adams (Imperial College London) 27 September 2017 The Alan Turing Institute

More information

Par$cle Filters Part I: Theory. Peter Jan van Leeuwen Data- Assimila$on Research Centre DARC University of Reading

Par$cle Filters Part I: Theory. Peter Jan van Leeuwen Data- Assimila$on Research Centre DARC University of Reading Par$cle Filters Part I: Theory Peter Jan van Leeuwen Data- Assimila$on Research Centre DARC University of Reading Reading July 2013 Why Data Assimila$on Predic$on Model improvement: - Parameter es$ma$on

More information

Negative Association, Ordering and Convergence of Resampling Methods

Negative Association, Ordering and Convergence of Resampling Methods Negative Association, Ordering and Convergence of Resampling Methods Nicolas Chopin ENSAE, Paristech (Joint work with Mathieu Gerber and Nick Whiteley, University of Bristol) Resampling schemes: Informal

More information

Stochastic Spectral Approaches to Bayesian Inference

Stochastic Spectral Approaches to Bayesian Inference Stochastic Spectral Approaches to Bayesian Inference Prof. Nathan L. Gibson Department of Mathematics Applied Mathematics and Computation Seminar March 4, 2011 Prof. Gibson (OSU) Spectral Approaches to

More information

Seminar: Data Assimilation

Seminar: Data Assimilation Seminar: Data Assimilation Jonas Latz, Elisabeth Ullmann Chair of Numerical Mathematics (M2) Technical University of Munich Jonas Latz, Elisabeth Ullmann (TUM) Data Assimilation 1 / 28 Prerequisites Bachelor:

More information

An Efficient Ensemble Data Assimilation Approach To Deal With Range Limited Observation

An Efficient Ensemble Data Assimilation Approach To Deal With Range Limited Observation An Efficient Ensemble Data Assimilation Approach To Deal With Range Limited Observation A. Shah 1,2, M. E. Gharamti 1, L. Bertino 1 1 Nansen Environmental and Remote Sensing Center 2 University of Bergen

More information

V (r,t) = i ˆ u( x, y,z,t) + ˆ j v( x, y,z,t) + k ˆ w( x, y, z,t)

V (r,t) = i ˆ u( x, y,z,t) + ˆ j v( x, y,z,t) + k ˆ w( x, y, z,t) IV. DIFFERENTIAL RELATIONS FOR A FLUID PARTICLE This chapter presents the development and application of the basic differential equations of fluid motion. Simplifications in the general equations and common

More information

Sampling from complex probability distributions

Sampling from complex probability distributions Sampling from complex probability distributions Louis J. M. Aslett (louis.aslett@durham.ac.uk) Department of Mathematical Sciences Durham University UTOPIAE Training School II 4 July 2017 1/37 Motivation

More information

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3 Brownian Motion Contents 1 Definition 2 1.1 Brownian Motion................................. 2 1.2 Wiener measure.................................. 3 2 Construction 4 2.1 Gaussian process.................................

More information

Bayesian Methods and Uncertainty Quantification for Nonlinear Inverse Problems

Bayesian Methods and Uncertainty Quantification for Nonlinear Inverse Problems Bayesian Methods and Uncertainty Quantification for Nonlinear Inverse Problems John Bardsley, University of Montana Collaborators: H. Haario, J. Kaipio, M. Laine, Y. Marzouk, A. Seppänen, A. Solonen, Z.

More information

Fundamentals of Data Assimila1on

Fundamentals of Data Assimila1on 014 GSI Community Tutorial NCAR Foothills Campus, Boulder, CO July 14-16, 014 Fundamentals of Data Assimila1on Milija Zupanski Cooperative Institute for Research in the Atmosphere Colorado State University

More information

Computer Intensive Methods in Mathematical Statistics

Computer Intensive Methods in Mathematical Statistics Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of

More information

Bayesian Dynamic Linear Modelling for. Complex Computer Models

Bayesian Dynamic Linear Modelling for. Complex Computer Models Bayesian Dynamic Linear Modelling for Complex Computer Models Fei Liu, Liang Zhang, Mike West Abstract Computer models may have functional outputs. With no loss of generality, we assume that a single computer

More information

Data assimilation with and without a model

Data assimilation with and without a model Data assimilation with and without a model Tyrus Berry George Mason University NJIT Feb. 28, 2017 Postdoc supported by NSF This work is in collaboration with: Tim Sauer, GMU Franz Hamilton, Postdoc, NCSU

More information

Robotics. Lecture 4: Probabilistic Robotics. See course website for up to date information.

Robotics. Lecture 4: Probabilistic Robotics. See course website   for up to date information. Robotics Lecture 4: Probabilistic Robotics See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College London Review: Sensors

More information

Systematic strategies for real time filtering of turbulent signals in complex systems

Systematic strategies for real time filtering of turbulent signals in complex systems Systematic strategies for real time filtering of turbulent signals in complex systems Statistical inversion theory for Gaussian random variables The Kalman Filter for Vector Systems: Reduced Filters and

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is

More information

Kernel Sequential Monte Carlo

Kernel Sequential Monte Carlo Kernel Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) * equal contribution April 25, 2016 1 / 37 Section

More information

Statistical Inference and Methods

Statistical Inference and Methods Department of Mathematics Imperial College London d.stephens@imperial.ac.uk http://stats.ma.ic.ac.uk/ das01/ 31st January 2006 Part VI Session 6: Filtering and Time to Event Data Session 6: Filtering and

More information

An introduction to Sequential Monte Carlo

An introduction to Sequential Monte Carlo An introduction to Sequential Monte Carlo Thang Bui Jes Frellsen Department of Engineering University of Cambridge Research and Communication Club 6 February 2014 1 Sequential Monte Carlo (SMC) methods

More information

Infinite-State Markov-switching for Dynamic. Volatility Models : Web Appendix

Infinite-State Markov-switching for Dynamic. Volatility Models : Web Appendix Infinite-State Markov-switching for Dynamic Volatility Models : Web Appendix Arnaud Dufays 1 Centre de Recherche en Economie et Statistique March 19, 2014 1 Comparison of the two MS-GARCH approximations

More information