Optimal filtering and the dual process
|
|
- Augustine Holt
- 5 years ago
- Views:
Transcription
1 Optimal filtering and the dual process Omiros Papaspiliopoulos ICREA & Universitat Pompeu Fabra Joint work with Matteo Ruggiero Collegio Carlo Alberto & University of Turin
2 The filtering problem X t0 X t1 X t2 Y 0 Y 1 Y 2 Hidden Markov model represented as a graphical model Signal: latent Markov chain X tn X, transition kernel P t(x, dx ) and initial distribution ν Emission densities: Data Y n Y with conditional density f x(y) (also consider non-dominated filtering models) Goal: evaluate the filtering distributions ν n(dx) := L(X tn Y 0,..., Y n) Subcase: if X t0 = X t1 =, it s a classical Bayesian inference problem
3 Filtering recursions Mathematically, it s the solutions to the recursion ν 0 = φ Y0 (ν), ν n = φ Yn (ψ tn t n 1 (ν n)), n > 0, where for probability measure ξ: Update: φ y(ξ)(dx) = fx(y)ξ(dx), p ξ (y) = p ξ (y) conjugate pair (f x, F) : ξ F = φ y(ξ) F f x(y)ξ(dx) X Prediction: ψ t(ξ)(dx ) = ξ(dx)p t(x, dx ) X Note that update and prediction operators satisfy ( n ) n ( w i p ξi (y) φ y w i ξ i = i=1 i=1 j w jp ξj (y) φy(ξ n ) n i), ψ t w i ξ i = w i ψ t(ξ i ) i=1 i=1
4 Finite-dimensional filters A finite-dimensional filter is a solution to the previous recursion s.t. there exists a finite-dimensional family F f of probability measures and ν F f φ y(ν), ψ t(ν) F f so evolution of the mixtures can be described by successive transformations of a finite-dimensional parameter : Nonparametric : any model where X is finite set (Baum filter) Parametric: linear Gaussian system (Kalman filter)
5 Motivating model: Cox-type process Dynamic version of the standard conjugate Bayesian model for count data: Y n X tn Poisson(X tn ) X tn Markov chain with gamma marginal For example X tn = U 2 n U n = a U n 1 + β η n with stationary N ( β 2 ) 0, 1 a 2 when a < 1
6 More generally, we can relate the signal to the Feller/square root/cir process/continuous-state branching process with immigration, e.g Kawazu & Watanabe (Theory Probab. Appl., 1971) dx t = (δσ 2 2γX t)dt + 2σ X tdb t, δ, γ, σ > 0 This process for δ 2 it has a stationary and limiting distribution, which is gamma(δ/2, γ/σ 2 ) For the economy of this talk, we focus on δ 2 (otherwise the boundary behaviour needs to be explicitly specified).
7 Computable filters We will classify a filter as computable when ν n is characterised in terms of a finite number of parameters that can be computed at a cost that grows polynomially with n. Special case are finite-dimensional filters for which the number of parameters does not grow with n, hence the cost of computing them grows linearly with n. We will identify filters that evolve in mixtures of finite-dimensional distributions indexed by (m, θ). We will present the framework and then provide examples of processes/signals that fit in it.
8 Conditions A1 (Reversibility): π(dx)p t(x, dx ) = π(dx )P t(x, dx). Notation: M = Z K + = { m = (m 1,..., m K ) : m j Z +, j = 1,..., K }. 0; e j ; m = i m i; product order on M; m i for i m
9 A 2D example for M (2,2) (2,1) (1,2) (2,0) (1,1) (0,2) (1,0) (0,1) (0,0)
10 A2 (Conjugacy): For Θ R l, l Z +, let h : X M Θ R + be such that sup x h(x, m, θ) < for all m M, θ Θ, and h(x, 0, θ) = 1 for some θ Θ. F = {h(x, m, θ)π(dx), m M, θ Θ} assumed to be a family of probability measures s.t. there exist functions t : Y M M and T : Y Θ Θ with m t(y, m) increasing and such that φ y(h(x, m, θ)π(dx)) = h(x, t(y, m), T (y, θ))π(dx). Note: p h(x,m,θ)π(dx) (y) =: c(m, θ, y) = which does not depend on x. fx(y)h(x, m, θ) h(x, t(y, m), T (y, θ))
11 A3 (Duality): Let r : Θ Θ, λ : Z + R + be increasing, ρ : Θ R + be continuous, and consider a Markov process (M t, Θ t) with state-space M Θ such that: dθ t/dt = r(θ t), Θ 0 = θ 0, when at (M t, Θ t) = (m, θ), the process jumps down to state (m e j, θ) with instantaneous rate λ( m )ρ(θ)m j. and it is dual to X t wrt functions h, i.e., x X, m M, θ Θ, t 0: E x [h(x t, m, θ)] = E (m,θ) [h(x, M t, Θ t)] When K = 0 or l = 0 the dual is just Θ t or M t
12 Remarks Absorption/Ergodicity M t can only jump to smaller states and 0 is absorbing Duality functions Radon-Nikodym derivatives between measures that are conjugate to the emission density Transition probabilities p m,n(t; θ) = P[M t = n M 0 = m, Θ 0 = θ], n, m M, n m. analytic expressions exist (Proposition 2.1)
13 Duality and optimal filtering: propagation Proposition Under A1-A2-A3 ψ t(h(x, m, θ)π(dx)) = p m,m i (t; θ)h(x, m i, Θ t)π(dx), 0 i m Proof. ψ t(h(x, m, θ)π(dx)) = h(x, m, θ)π(dx)p t(x, dx ) = h(x, m, θ)π(dx )P t(x, dx) X X = π(dx )E x [h(x t, m, θ)] = π(dx )E (m,θ) [h(x, M t, Θ t)] = p m,n(t; θ)h(x, n, Θ t)π(dx ) n m
14 Duality and optimal filtering: update Proposition For { F f = wmh(x, m, θ)π(dx) : Λ M, Λ <, m Λ wm 0, } m Λ wm = 1. and under A1-A2-A3, Ff is closed under prediction and update: ( ) φ y wmh(x, m, θ)π(dx) = ŵnh(x, n, T (y, θ))π(dx) m Λ n t(y,λ) with t(y, Λ) := {n : n = t(y, m), m Λ} ŵ n w mc(m, θ, y) for n = t(y, m), n t(y,λ) ŵn = 1, and ( ) ψ t w mh(x, m, θ)π(dx) = m Λ n G(Λ) ( m Λ,m n ) w mp m,n(t; θ) h(x, n, θ t)π(dx).
15 Generators (Care is needed with domains but will skip for economy of time) The generator of a Markov process X t, with semigroup operator P t, is a linear operator A with domain denoted D(A), linked to the semigroup operator via the Kolmogorov backward equation Ptf(x) = (APtf)(x), t f D(A), where on the left hand side P th(x) is differentiated in t for given x, whereas on the right hand side, A acts on P th(x) as a function of x for given t
16 Local duality (Care is needed with domains but will skip for economy of time) If X t solves an SDE on R d dx t = b(x t)dt + σ(x t)db t, then its generator is (Af)(x) = d i=1 b i (x) f(x) x i d i,j=1 a i,j (x) 2 f(x) x i x j, f D(A), for a i,j (x) := (σ(x)σ(x) T ) i,j If (M t, Θ t) is Markov process as in A3 its generator is K l g(m, θ) (Ag)(m, θ) = λ( m )ρ(θ) m i [g(m e i, θ) g(m, θ)]+ r i (θ), g D(A), θ i=1 i=1
17 A4 (Local duality): The function h(x, m, θ) defined in A2 is such that h(x, m, θ), as a function of x belongs to D(A) for all (m, θ) M Θ, as a function of (m, θ) belongs to D(A) for all x X, and x X, m M, θ Θ (Ah(, m, θ))(x) = (Ah(x,, ))(m, θ) (1) This implies the duality in A3. (Proof: resolvents/laplace transforms)
18 I will focus on the properties of the signal and illustrate the dual. Clearly, for computable filtering we need the emission density to be conjugate to the family F. I will present results for: Linear diffusion signals - Gaussian - only deterministic dual CIR signals - gamma - both deterministic and death process dual Due to time constraints I will skip results for: Multi-dimensional Wright-Fisher signals - Dirichlet - only death process dual Fleming Viot - Dirichlet process prior - only death process dual Dawson-Watanabe - gamma process - both deterministic and death process dual
19 Linear diffusion signals SDE dx t = σ2 α (Xt γ)dt + 2σdB t, Invariant measure π(dx) Normal(dx; γ, α). Generator A = (σ 2 γ/α σ 2 x/α) d d2 + σ2 dx dx 2 Duality functions h(x, µ, τ) = ( α ) 1/2 (x µ)2 exp { + τ 2τ } (x γ)2 Normal(dx; µ, τ) = 2α π(dx)
20 Linear diffusion signals Kalman filter - in 1 line A small calculation yields: Ah(, µ, τ)(x) = σ2 (γ µ) α µ h(x, µ, τ) + 2σ2 (1 τ/α) h(x, µ, τ). τ Hence the dual is purely deterministic and described in terms of the ODEs: dµ t/dt = σ2 α (γ µt)dt, dτt/dt = 2σ2 (1 τ t/α)dt.
21 CIR signals Some care is needed with domains SDE, for σ > 0, γ > 0, δ 2 dx t = (δσ 2 2γX t)dt + 2σ X tdb t Invariant measure π(dx) gamma(dx; δ/2, γ/σ 2 ). Generator Duality functions h(x, m, θ) = = A = (δσ 2 2γx) d dx + 2σ2 x d2 dx 2, Γ(δ/2) ( γ ) δ/2 θ δ/2+m Γ(δ/2 + m) σ 2 x m exp{ (θ γ/σ 2 )x} gamma(dx; δ/2 + m, θ) π(dx)
22 CIR signals A direct calculation yields: Ah(, m, θ)(x) = 2mσ 2 θ h(x, m 1, θ) + σ 2 (δ + 2m)(θ γ/σ 2 )h(x, m + 1, θ) σ 2 [2mθ + (δ + 2m)(θ γ/σ 2 )]h(x, m, θ) = 2mσ 2 θ (h(x, m 1, θ) h(x, m, θ)) + σ 2 (θ γ/σ 2 )(2θx δ 2m)h(x, m, θ) = 2mσ 2 θ (h(x, m 1, θ) h(x, m, θ)) + 2σ 2 θ(γ/σ 2 θ) h(x, m, θ) θ Therefore, we consider a two-component process (M t, Θ t) with generator A and λ(m) = 2mσ 2, r(θ) = 2σ 2 θ(γ/σ 2 θ), ρ(θ) = θ. and the above calculation gives Ah(, m, θ)(x) = Ah(x,, )(m, θ).
23 CIR signals Under this setting Θ t = γ θe 2γt σ 2 θe 2γt + γ/σ 2 θ which implies that the transition probabilities for the death process simplify to binomial probabilities p m,m i (t; θ) = Bin ( γ m i ; m, σ 2 (θe2γt + γ/σ 2 θ) 1) which in turns immediately yields the result ψ t(gamma(m + δ/2, θ)) = m ( Bin k ; m, k=0 γ σ 2 (θe2γt + γ/σ 2 θ) 1) gamma ( k + δ/2, γ σ 2 θe 2γt ) θe 2γt + γ/σ 2. θ
24 Discussion/Extensions Complexity Pruning Smoothing/Likelihood/Simulation Approximation Other models
25 Complexity L(D n Y 0,..., Y n 1 ) has support on Λ n {θ n} with Λ n+1 = G(t(Y n, Λ n)) and weights that can be computed recursively. The component probabilities at time n can be computed at a cost that is at most O( Λ n 2 ), but Λ n increases with n. Proposition Under the assumption that t(y, m) = m + N(y), where N : Y M, we have that ( ) n ( ) K Λ n = G m 0 + N(Y i ) 1 + dn K i=1 where d n = m 0 + n i=1 N(Y i). When the observations follow a stationary process, d n will be of order n, thus overall complexity of filtering O(n 2K ), where the constant depends on K but not n
Optimal filtering and the dual process
Optimal filtering and the dual process Omiros Papaspiliopoulos ICREA & Universitat Pompeu Fabra Joint work with Matteo Ruggiero Collegio Carlo Alberto & University of Turin Outline The filtering problem
More informationLecture 6: Bayesian Inference in SDE Models
Lecture 6: Bayesian Inference in SDE Models Bayesian Filtering and Smoothing Point of View Simo Särkkä Aalto University Simo Särkkä (Aalto) Lecture 6: Bayesian Inference in SDEs 1 / 45 Contents 1 SDEs
More informationConcentration inequalities for Feynman-Kac particle models. P. Del Moral. INRIA Bordeaux & IMB & CMAP X. Journées MAS 2012, SMAI Clermond-Ferrand
Concentration inequalities for Feynman-Kac particle models P. Del Moral INRIA Bordeaux & IMB & CMAP X Journées MAS 2012, SMAI Clermond-Ferrand Some hyper-refs Feynman-Kac formulae, Genealogical & Interacting
More informationParticle Filtering a brief introductory tutorial. Frank Wood Gatsby, August 2007
Particle Filtering a brief introductory tutorial Frank Wood Gatsby, August 2007 Problem: Target Tracking A ballistic projectile has been launched in our direction and may or may not land near enough to
More informationUnsupervised Learning
Unsupervised Learning Bayesian Model Comparison Zoubin Ghahramani zoubin@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit, and MSc in Intelligent Systems, Dept Computer Science University College
More informationBranching Processes II: Convergence of critical branching to Feller s CSB
Chapter 4 Branching Processes II: Convergence of critical branching to Feller s CSB Figure 4.1: Feller 4.1 Birth and Death Processes 4.1.1 Linear birth and death processes Branching processes can be studied
More informationSequential Monte Carlo and Particle Filtering. Frank Wood Gatsby, November 2007
Sequential Monte Carlo and Particle Filtering Frank Wood Gatsby, November 2007 Importance Sampling Recall: Let s say that we want to compute some expectation (integral) E p [f] = p(x)f(x)dx and we remember
More informationLinear Dynamical Systems (Kalman filter)
Linear Dynamical Systems (Kalman filter) (a) Overview of HMMs (b) From HMMs to Linear Dynamical Systems (LDS) 1 Markov Chains with Discrete Random Variables x 1 x 2 x 3 x T Let s assume we have discrete
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing
More informationNonparametric Drift Estimation for Stochastic Differential Equations
Nonparametric Drift Estimation for Stochastic Differential Equations Gareth Roberts 1 Department of Statistics University of Warwick Brazilian Bayesian meeting, March 2010 Joint work with O. Papaspiliopoulos,
More informationDynamic models. Dependent data The AR(p) model The MA(q) model Hidden Markov models. 6 Dynamic models
6 Dependent data The AR(p) model The MA(q) model Hidden Markov models Dependent data Dependent data Huge portion of real-life data involving dependent datapoints Example (Capture-recapture) capture histories
More informationLecture 4: Hidden Markov Models: An Introduction to Dynamic Decision Making. November 11, 2010
Hidden Lecture 4: Hidden : An Introduction to Dynamic Decision Making November 11, 2010 Special Meeting 1/26 Markov Model Hidden When a dynamical system is probabilistic it may be determined by the transition
More informationLinear Dynamical Systems
Linear Dynamical Systems Sargur N. srihari@cedar.buffalo.edu Machine Learning Course: http://www.cedar.buffalo.edu/~srihari/cse574/index.html Two Models Described by Same Graph Latent variables Observations
More informationStat 516, Homework 1
Stat 516, Homework 1 Due date: October 7 1. Consider an urn with n distinct balls numbered 1,..., n. We sample balls from the urn with replacement. Let N be the number of draws until we encounter a ball
More informationLAN property for ergodic jump-diffusion processes with discrete observations
LAN property for ergodic jump-diffusion processes with discrete observations Eulalia Nualart (Universitat Pompeu Fabra, Barcelona) joint work with Arturo Kohatsu-Higa (Ritsumeikan University, Japan) &
More informationMean-field dual of cooperative reproduction
The mean-field dual of systems with cooperative reproduction joint with Tibor Mach (Prague) A. Sturm (Göttingen) Friday, July 6th, 2018 Poisson construction of Markov processes Let (X t ) t 0 be a continuous-time
More informationImprecise Filtering for Spacecraft Navigation
Imprecise Filtering for Spacecraft Navigation Tathagata Basu Cristian Greco Thomas Krak Durham University Strathclyde University Ghent University Filtering for Spacecraft Navigation The General Problem
More informationMean field simulation for Monte Carlo integration. Part II : Feynman-Kac models. P. Del Moral
Mean field simulation for Monte Carlo integration Part II : Feynman-Kac models P. Del Moral INRIA Bordeaux & Inst. Maths. Bordeaux & CMAP Polytechnique Lectures, INLN CNRS & Nice Sophia Antipolis Univ.
More informationSequence labeling. Taking collective a set of interrelated instances x 1,, x T and jointly labeling them
HMM, MEMM and CRF 40-957 Special opics in Artificial Intelligence: Probabilistic Graphical Models Sharif University of echnology Soleymani Spring 2014 Sequence labeling aking collective a set of interrelated
More informationGeneral Bayesian Inference I
General Bayesian Inference I Outline: Basic concepts, One-parameter models, Noninformative priors. Reading: Chapters 10 and 11 in Kay-I. (Occasional) Simplified Notation. When there is no potential for
More informationBessel-like SPDEs. Lorenzo Zambotti, Sorbonne Université (joint work with Henri Elad-Altman) 15th May 2018, Luminy
Bessel-like SPDEs, Sorbonne Université (joint work with Henri Elad-Altman) Squared Bessel processes Let δ, y, and (B t ) t a BM. By Yamada-Watanabe s Theorem, there exists a unique (strong) solution (Y
More informationStochastic Differential Equations
Chapter 5 Stochastic Differential Equations We would like to introduce stochastic ODE s without going first through the machinery of stochastic integrals. 5.1 Itô Integrals and Itô Differential Equations
More informationThe Λ-Fleming-Viot process and a connection with Wright-Fisher diffusion. Bob Griffiths University of Oxford
The Λ-Fleming-Viot process and a connection with Wright-Fisher diffusion Bob Griffiths University of Oxford A d-dimensional Λ-Fleming-Viot process {X(t)} t 0 representing frequencies of d types of individuals
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning
More informationAdvanced Monte Carlo integration methods. P. Del Moral (INRIA team ALEA) INRIA & Bordeaux Mathematical Institute & X CMAP
Advanced Monte Carlo integration methods P. Del Moral (INRIA team ALEA) INRIA & Bordeaux Mathematical Institute & X CMAP MCQMC 2012, Sydney, Sunday Tutorial 12-th 2012 Some hyper-refs Feynman-Kac formulae,
More informationBirth and Death Processes. Birth and Death Processes. Linear Growth with Immigration. Limiting Behaviour for Birth and Death Processes
DTU Informatics 247 Stochastic Processes 6, October 27 Today: Limiting behaviour of birth and death processes Birth and death processes with absorbing states Finite state continuous time Markov chains
More informationMarkov Chain Monte Carlo Methods for Stochastic
Markov Chain Monte Carlo Methods for Stochastic Optimization i John R. Birge The University of Chicago Booth School of Business Joint work with Nicholas Polson, Chicago Booth. JRBirge U Florida, Nov 2013
More informationNash Type Inequalities for Fractional Powers of Non-Negative Self-adjoint Operators. ( Wroclaw 2006) P.Maheux (Orléans. France)
Nash Type Inequalities for Fractional Powers of Non-Negative Self-adjoint Operators ( Wroclaw 006) P.Maheux (Orléans. France) joint work with A.Bendikov. European Network (HARP) (to appear in T.A.M.S)
More informationConsistency of the maximum likelihood estimator for general hidden Markov models
Consistency of the maximum likelihood estimator for general hidden Markov models Jimmy Olsson Centre for Mathematical Sciences Lund University Nordstat 2012 Umeå, Sweden Collaborators Hidden Markov models
More informationProbabilistic Graphical Models
School of Computer Science Probabilistic Graphical Models Variational Inference II: Mean Field Method and Variational Principle Junming Yin Lecture 15, March 7, 2012 X 1 X 1 X 1 X 1 X 2 X 3 X 2 X 2 X 3
More informationPatterns of Scalable Bayesian Inference Background (Session 1)
Patterns of Scalable Bayesian Inference Background (Session 1) Jerónimo Arenas-García Universidad Carlos III de Madrid jeronimo.arenas@gmail.com June 14, 2017 1 / 15 Motivation. Bayesian Learning principles
More informationCS839: Probabilistic Graphical Models. Lecture 7: Learning Fully Observed BNs. Theo Rekatsinas
CS839: Probabilistic Graphical Models Lecture 7: Learning Fully Observed BNs Theo Rekatsinas 1 Exponential family: a basic building block For a numeric random variable X p(x ) =h(x)exp T T (x) A( ) = 1
More informationNonparametric Bayesian Methods - Lecture I
Nonparametric Bayesian Methods - Lecture I Harry van Zanten Korteweg-de Vries Institute for Mathematics CRiSM Masterclass, April 4-6, 2016 Overview of the lectures I Intro to nonparametric Bayesian statistics
More informationVariational Scoring of Graphical Model Structures
Variational Scoring of Graphical Model Structures Matthew J. Beal Work with Zoubin Ghahramani & Carl Rasmussen, Toronto. 15th September 2003 Overview Bayesian model selection Approximations using Variational
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation
More informationOn Reparametrization and the Gibbs Sampler
On Reparametrization and the Gibbs Sampler Jorge Carlos Román Department of Mathematics Vanderbilt University James P. Hobert Department of Statistics University of Florida March 2014 Brett Presnell Department
More informationManifold Monte Carlo Methods
Manifold Monte Carlo Methods Mark Girolami Department of Statistical Science University College London Joint work with Ben Calderhead Research Section Ordinary Meeting The Royal Statistical Society October
More informationPattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions
Pattern Recognition and Machine Learning Chapter 2: Probability Distributions Cécile Amblard Alex Kläser Jakob Verbeek October 11, 27 Probability Distributions: General Density Estimation: given a finite
More informationHmms with variable dimension structures and extensions
Hmm days/enst/january 21, 2002 1 Hmms with variable dimension structures and extensions Christian P. Robert Université Paris Dauphine www.ceremade.dauphine.fr/ xian Hmm days/enst/january 21, 2002 2 1 Estimating
More informationLAN property for sde s with additive fractional noise and continuous time observation
LAN property for sde s with additive fractional noise and continuous time observation Eulalia Nualart (Universitat Pompeu Fabra, Barcelona) joint work with Samy Tindel (Purdue University) Vlad s 6th birthday,
More informationLatent voter model on random regular graphs
Latent voter model on random regular graphs Shirshendu Chatterjee Cornell University (visiting Duke U.) Work in progress with Rick Durrett April 25, 2011 Outline Definition of voter model and duality with
More informationPattern Recognition and Machine Learning
Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability
More informationBayesian Inference and the Symbolic Dynamics of Deterministic Chaos. Christopher C. Strelioff 1,2 Dr. James P. Crutchfield 2
How Random Bayesian Inference and the Symbolic Dynamics of Deterministic Chaos Christopher C. Strelioff 1,2 Dr. James P. Crutchfield 2 1 Center for Complex Systems Research and Department of Physics University
More informationHidden Markov Models. Aarti Singh Slides courtesy: Eric Xing. Machine Learning / Nov 8, 2010
Hidden Markov Models Aarti Singh Slides courtesy: Eric Xing Machine Learning 10-701/15-781 Nov 8, 2010 i.i.d to sequential data So far we assumed independent, identically distributed data Sequential data
More informationErnesto Mordecki 1. Lecture III. PASI - Guanajuato - June 2010
Optimal stopping for Hunt and Lévy processes Ernesto Mordecki 1 Lecture III. PASI - Guanajuato - June 2010 1Joint work with Paavo Salminen (Åbo, Finland) 1 Plan of the talk 1. Motivation: from Finance
More informationMachine Learning 4771
Machine Learning 4771 Instructor: ony Jebara Kalman Filtering Linear Dynamical Systems and Kalman Filtering Structure from Motion Linear Dynamical Systems Audio: x=pitch y=acoustic waveform Vision: x=object
More informationBayesian model selection in graphs by using BDgraph package
Bayesian model selection in graphs by using BDgraph package A. Mohammadi and E. Wit March 26, 2013 MOTIVATION Flow cytometry data with 11 proteins from Sachs et al. (2005) RESULT FOR CELL SIGNALING DATA
More informationUndirected Graphical Models
Undirected Graphical Models 1 Conditional Independence Graphs Let G = (V, E) be an undirected graph with vertex set V and edge set E, and let A, B, and C be subsets of vertices. We say that C separates
More informationLecture 4: Introduction to stochastic processes and stochastic calculus
Lecture 4: Introduction to stochastic processes and stochastic calculus Cédric Archambeau Centre for Computational Statistics and Machine Learning Department of Computer Science University College London
More informationBayesian Regression Linear and Logistic Regression
When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we
More informationSequential Monte Carlo Methods for Bayesian Computation
Sequential Monte Carlo Methods for Bayesian Computation A. Doucet Kyoto Sept. 2012 A. Doucet (MLSS Sept. 2012) Sept. 2012 1 / 136 Motivating Example 1: Generic Bayesian Model Let X be a vector parameter
More informationContraction properties of Feynman-Kac semigroups
Journées de Statistique Marne La Vallée, January 2005 Contraction properties of Feynman-Kac semigroups Pierre DEL MORAL, Laurent MICLO Lab. J. Dieudonné, Nice Univ., LATP Univ. Provence, Marseille 1 Notations
More informationState Space and Hidden Markov Models
State Space and Hidden Markov Models Kunsch H.R. State Space and Hidden Markov Models. ETH- Zurich Zurich; Aliaksandr Hubin Oslo 2014 Contents 1. Introduction 2. Markov Chains 3. Hidden Markov and State
More informationAppendices: Stochastic Backpropagation and Approximate Inference in Deep Generative Models
Appendices: Stochastic Backpropagation and Approximate Inference in Deep Generative Models Danilo Jimenez Rezende Shakir Mohamed Daan Wierstra Google DeepMind, London, United Kingdom DANILOR@GOOGLE.COM
More informationVariational Learning : From exponential families to multilinear systems
Variational Learning : From exponential families to multilinear systems Ananth Ranganathan th February 005 Abstract This note aims to give a general overview of variational inference on graphical models.
More informationIntroduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak
Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak 1 Introduction. Random variables During the course we are interested in reasoning about considered phenomenon. In other words,
More informationProbabilistic Graphical Models for Image Analysis - Lecture 4
Probabilistic Graphical Models for Image Analysis - Lecture 4 Stefan Bauer 12 October 2018 Max Planck ETH Center for Learning Systems Overview 1. Repetition 2. α-divergence 3. Variational Inference 4.
More informationOn the posterior structure of NRMI
On the posterior structure of NRMI Igor Prünster University of Turin, Collegio Carlo Alberto and ICER Joint work with L.F. James and A. Lijoi Isaac Newton Institute, BNR Programme, 8th August 2007 Outline
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is
More informationp 1 ( Y p dp) 1/p ( X p dp) 1 1 p
Doob s inequality Let X(t) be a right continuous submartingale with respect to F(t), t 1 P(sup s t X(s) λ) 1 λ {sup s t X(s) λ} X + (t)dp 2 For 1 < p
More informationExact Simulation of Diffusions and Jump Diffusions
Exact Simulation of Diffusions and Jump Diffusions A work by: Prof. Gareth O. Roberts Dr. Alexandros Beskos Dr. Omiros Papaspiliopoulos Dr. Bruno Casella 28 th May, 2008 Content 1 Exact Algorithm Construction
More informationHidden Markov Models. By Parisa Abedi. Slides courtesy: Eric Xing
Hidden Markov Models By Parisa Abedi Slides courtesy: Eric Xing i.i.d to sequential data So far we assumed independent, identically distributed data Sequential (non i.i.d.) data Time-series data E.g. Speech
More informationControlled sequential Monte Carlo
Controlled sequential Monte Carlo Jeremy Heng, Department of Statistics, Harvard University Joint work with Adrian Bishop (UTS, CSIRO), George Deligiannidis & Arnaud Doucet (Oxford) Bayesian Computation
More informationQuantitative Biology II Lecture 4: Variational Methods
10 th March 2015 Quantitative Biology II Lecture 4: Variational Methods Gurinder Singh Mickey Atwal Center for Quantitative Biology Cold Spring Harbor Laboratory Image credit: Mike West Summary Approximate
More informationInverse Langevin approach to time-series data analysis
Inverse Langevin approach to time-series data analysis Fábio Macêdo Mendes Anníbal Dias Figueiredo Neto Universidade de Brasília Saratoga Springs, MaxEnt 2007 Outline 1 Motivation 2 Outline 1 Motivation
More informationBayesian Methods with Monte Carlo Markov Chains II
Bayesian Methods with Monte Carlo Markov Chains II Henry Horng-Shing Lu Institute of Statistics National Chiao Tung University hslu@stat.nctu.edu.tw http://tigpbp.iis.sinica.edu.tw/courses.htm 1 Part 3
More informationAsymptotics for posterior hazards
Asymptotics for posterior hazards Pierpaolo De Blasi University of Turin 10th August 2007, BNR Workshop, Isaac Newton Intitute, Cambridge, UK Joint work with Giovanni Peccati (Université Paris VI) and
More informationLikelihood NIPS July 30, Gaussian Process Regression with Student-t. Likelihood. Jarno Vanhatalo, Pasi Jylanki and Aki Vehtari NIPS-2009
with with July 30, 2010 with 1 2 3 Representation Representation for Distribution Inference for the Augmented Model 4 Approximate Laplacian Approximation Introduction to Laplacian Approximation Laplacian
More informationBayesian Machine Learning - Lecture 7
Bayesian Machine Learning - Lecture 7 Guido Sanguinetti Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh gsanguin@inf.ed.ac.uk March 4, 2015 Today s lecture 1
More informationCS Lecture 19. Exponential Families & Expectation Propagation
CS 6347 Lecture 19 Exponential Families & Expectation Propagation Discrete State Spaces We have been focusing on the case of MRFs over discrete state spaces Probability distributions over discrete spaces
More informationMachine Learning. Probability Basics. Marc Toussaint University of Stuttgart Summer 2014
Machine Learning Probability Basics Basic definitions: Random variables, joint, conditional, marginal distribution, Bayes theorem & examples; Probability distributions: Binomial, Beta, Multinomial, Dirichlet,
More informationp L yi z n m x N n xi
y i z n x n N x i Overview Directed and undirected graphs Conditional independence Exact inference Latent variables and EM Variational inference Books statistical perspective Graphical Models, S. Lauritzen
More information(Exact and efficient) Simulation of Conditioned Markov Processes
(Exact and efficient) Simulation of Conditioned Markov Processes Omiros Papaspiliopoulos Universitat Pompeu Fabra, Barcelona http://www.econ.upf.edu/ omiros GrStoc α, ΛEYKA A 2009 Acknowledgement: K.Kalogeropoulos
More informationAdvanced Machine Learning & Perception
Advanced Machine Learning & Perception Instructor: Tony Jebara Topic 6 Standard Kernels Unusual Input Spaces for Kernels String Kernels Probabilistic Kernels Fisher Kernels Probability Product Kernels
More informationAn Overview of Methods for Applying Semi-Markov Processes in Biostatistics.
An Overview of Methods for Applying Semi-Markov Processes in Biostatistics. Charles J. Mode Department of Mathematics and Computer Science Drexel University Philadelphia, PA 19104 Overview of Topics. I.
More informationWEYL S LEMMA, ONE OF MANY. Daniel W. Stroock
WEYL S LEMMA, ONE OF MANY Daniel W Stroock Abstract This note is a brief, and somewhat biased, account of the evolution of what people working in PDE s call Weyl s Lemma about the regularity of solutions
More informationFundamentals of Statistics
Chapter 2 Fundamentals of Statistics This chapter discusses some fundamental concepts of mathematical statistics. These concepts are essential for the material in later chapters. 2.1 Populations, Samples,
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Lecture 12 Dynamical Models CS/CNS/EE 155 Andreas Krause Homework 3 out tonight Start early!! Announcements Project milestones due today Please email to TAs 2 Parameter learning
More informationBasic math for biology
Basic math for biology Lei Li Florida State University, Feb 6, 2002 The EM algorithm: setup Parametric models: {P θ }. Data: full data (Y, X); partial data Y. Missing data: X. Likelihood and maximum likelihood
More informationChapter 8.8.1: A factorization theorem
LECTURE 14 Chapter 8.8.1: A factorization theorem The characterization of a sufficient statistic in terms of the conditional distribution of the data given the statistic can be difficult to work with.
More informationBayesian Nonparametrics: some contributions to construction and properties of prior distributions
Bayesian Nonparametrics: some contributions to construction and properties of prior distributions Annalisa Cerquetti Collegio Nuovo, University of Pavia, Italy Interview Day, CETL Lectureship in Statistics,
More informationMarkov Chain Monte Carlo Methods for Stochastic Optimization
Markov Chain Monte Carlo Methods for Stochastic Optimization John R. Birge The University of Chicago Booth School of Business Joint work with Nicholas Polson, Chicago Booth. JRBirge U of Toronto, MIE,
More informationCS281A/Stat241A Lecture 22
CS281A/Stat241A Lecture 22 p. 1/4 CS281A/Stat241A Lecture 22 Monte Carlo Methods Peter Bartlett CS281A/Stat241A Lecture 22 p. 2/4 Key ideas of this lecture Sampling in Bayesian methods: Predictive distribution
More informationComputer Intensive Methods in Mathematical Statistics
Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of
More informationStochastic Demography, Coalescents, and Effective Population Size
Demography Stochastic Demography, Coalescents, and Effective Population Size Steve Krone University of Idaho Department of Mathematics & IBEST Demographic effects bottlenecks, expansion, fluctuating population
More informationAuxiliary Particle Methods
Auxiliary Particle Methods Perspectives & Applications Adam M. Johansen 1 adam.johansen@bristol.ac.uk Oxford University Man Institute 29th May 2008 1 Collaborators include: Arnaud Doucet, Nick Whiteley
More informationParticle Filtering Approaches for Dynamic Stochastic Optimization
Particle Filtering Approaches for Dynamic Stochastic Optimization John R. Birge The University of Chicago Booth School of Business Joint work with Nicholas Polson, Chicago Booth. JRBirge I-Sim Workshop,
More informationChapter 2: Fundamentals of Statistics Lecture 15: Models and statistics
Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Hidden Markov Models Barnabás Póczos & Aarti Singh Slides courtesy: Eric Xing i.i.d to sequential data So far we assumed independent, identically distributed
More informationModern Methods of Statistical Learning sf2935 Auxiliary material: Exponential Family of Distributions Timo Koski. Second Quarter 2016
Auxiliary material: Exponential Family of Distributions Timo Koski Second Quarter 2016 Exponential Families The family of distributions with densities (w.r.t. to a σ-finite measure µ) on X defined by R(θ)
More informationSlice Sampling Mixture Models
Slice Sampling Mixture Models Maria Kalli, Jim E. Griffin & Stephen G. Walker Centre for Health Services Studies, University of Kent Institute of Mathematics, Statistics & Actuarial Science, University
More informationRandom Walks Conditioned to Stay Positive
1 Random Walks Conditioned to Stay Positive Bob Keener Let S n be a random walk formed by summing i.i.d. integer valued random variables X i, i 1: S n = X 1 + + X n. If the drift EX i is negative, then
More informationOptimal transportation and optimal control in a finite horizon framework
Optimal transportation and optimal control in a finite horizon framework Guillaume Carlier and Aimé Lachapelle Université Paris-Dauphine, CEREMADE July 2008 1 MOTIVATIONS - A commitment problem (1) Micro
More informationGaussian processes for inference in stochastic differential equations
Gaussian processes for inference in stochastic differential equations Manfred Opper, AI group, TU Berlin November 6, 2017 Manfred Opper, AI group, TU Berlin (TU Berlin) inference in SDE November 6, 2017
More informationChapter 4 Dynamic Bayesian Networks Fall Jin Gu, Michael Zhang
Chapter 4 Dynamic Bayesian Networks 2016 Fall Jin Gu, Michael Zhang Reviews: BN Representation Basic steps for BN representations Define variables Define the preliminary relations between variables Check
More informationDependent hierarchical processes for multi armed bandits
Dependent hierarchical processes for multi armed bandits Federico Camerlenghi University of Bologna, BIDSA & Collegio Carlo Alberto First Italian meeting on Probability and Mathematical Statistics, Torino
More informationHybrid Dirichlet processes for functional data
Hybrid Dirichlet processes for functional data Sonia Petrone Università Bocconi, Milano Joint work with Michele Guindani - U.T. MD Anderson Cancer Center, Houston and Alan Gelfand - Duke University, USA
More informationAsymptotics for posterior hazards
Asymptotics for posterior hazards Igor Prünster University of Turin, Collegio Carlo Alberto and ICER Joint work with P. Di Biasi and G. Peccati Workshop on Limit Theorems and Applications Paris, 16th January
More informationIntroduction to asymptotic techniques for stochastic systems with multiple time-scales
Introduction to asymptotic techniques for stochastic systems with multiple time-scales Eric Vanden-Eijnden Courant Institute Motivating examples Consider the ODE {Ẋ = Y 3 + sin(πt) + cos( 2πt) X() = x
More information