Lecture 5: Importance sampling and Hamilton-Jacobi equations

Size: px
Start display at page:

Download "Lecture 5: Importance sampling and Hamilton-Jacobi equations"

Transcription

1 Lecture 5: Importance sampling and Hamilton-Jacobi equations Henrik Hult Department of Mathematics KTH Royal Institute of Technology Sweden Summer School on Monte Carlo Methods and Rare Events Brown University, June 13-17, 2016

2 Outline 1 Large deviations and Hamilton-Jacobi equations 2 Exponential decay of the second moment 3 Construction of subsolutions

3 The subsolution approach to efficient importance sampling Quantify performance as the exponential rate of decay of the second moment The latter is given as the initial value of a solution to a Hamilton-Jacobi equation Construction of efficient importance sampling algorithms is essentially equivalent to the construction of classical subsolutions of the corresponding HJ-equation

4 Large deviations and Hamilton-Jacobi equations Consider a process {X n } with continuous trajectories, satisfying a large deviations principle (LDP). The large deviations principle says roughly that, given T > 0, an absolutely continuous φ : [0, T ] R d and a small δ > 0, { P sup t [0,T ] } X n (t) φ(t) > δ e ni T (φ), where I T is the rate function and takes the form: I T (φ) = T 0 L(φ(t), φ(t))dt, and L is the local rate function. For an expectation E[exp{ nf (X n (T ))}] we have, similarly, that { } E[exp{ nf (X n (T ))}] exp n inf {I T (φ) + F(φ(T ))}. φ

5 Large deviations and Hamilton-Jacobi equations Consider a process {X n } with continuous trajectories, satisfying a large deviations principle (LDP). The large deviations principle says roughly that, given T > 0, an absolutely continuous φ : [0, T ] R d and a small δ > 0, { P sup t [0,T ] } X n (t) φ(t) > δ e ni T (φ), where I T is the rate function and takes the form: I T (φ) = T 0 L(φ(t), φ(t))dt, and L is the local rate function. For an expectation E[exp{ nf (X n (T ))}] we have, similarly, that { } E[exp{ nf (X n (T ))}] exp n inf {I T (φ) + F(φ(T ))}. φ

6 Recall the Markov random walk model Let {v i (x), x R d, i 0} be independent and identically distributed random vector fields with distribution P{v i (x) } = θ( x), where θ is a regular conditional probability distribution. Let Xi+1 n = X i n + 1 n v i(xi n ), X0 n = x 0. Denote the log moment generating function of θ( x) by H(x, α) = log E[exp{ α, v 1 (x) }] and suppose H(x, α) < for all x and α in R d. The Fenchel-Legendre transform (convex conjugate) of H(x, ), denoted by L(x, β) = sup α R d [ α, β H(x, α)].

7 The backward equation Let A n denote the backward evolution operator associated with X n, that is, A n f (i, x) = E i,x [f (i + 1, Xi+1 n ) f (i, x)] [ = f (i + 1, x + 1 ] n z) f (i, x) θ(dz x). The (Kolmogorov) backward equation implies that V n (i, x) = E i,x [exp{ nf(x n n )}] satisfies A n V n (i, x) = 0, V n (n, x) = exp{ nf (x)}, where V n (0, x 0 ) = E[exp{ nf (X n n )}] is the quantity we are interested in computing.

8 The Hamilton-Jacobi equation Redefine V n by V n ( i n, x) := 1 n log E i,x[exp{ nf(x n n )}]. Plugging in this transformation in the backward equation leads to the equation for V n : [ 0 = exp{ n[v n ( i n + 1 n, x + 1 n z) V n ( i ] n, x)]} 1 θ(dz x).

9 The Hamilton-Jacobi equation Redefine V n by V n ( i n, x) := 1 n log E i,x[exp{ nf(x n n )}]. Plugging in this transformation in the backward equation leads to the equation for V n : [ 0 = exp{ n[v n ( i n + 1 n, x + 1 n z) V n ( i ] n, x)]} 1 θ(dz x).

10 The Hamilton-Jacobi equation Assuming V n V, with V smooth and i/n t, we may approximate V n ( i n + 1 n, x + 1 n z) V n ( i n, x) V ( i n + 1 n, x + 1 n z) V ( i n, x) 1 [ ] V t (t, x) + DV (t, x), z, n... which leads to 1 = exp{ V t (t, x) DV (t, x), z }θ(dz x) = e V t (t,x)+h(x, DV (t,x))... and taking logarithm on both sides yields V t (t, x) H(x, DV (t, x)) = 0, with terminal condition V (T, x) = F(x)

11 The Hamilton-Jacobi equation Assuming V n V, with V smooth and i/n t, we may approximate V n ( i n + 1 n, x + 1 n z) V n ( i n, x) V ( i n + 1 n, x + 1 n z) V ( i n, x) 1 [ ] V t (t, x) + DV (t, x), z, n... which leads to 1 = exp{ V t (t, x) DV (t, x), z }θ(dz x) = e V t (t,x)+h(x, DV (t,x))... and taking logarithm on both sides yields V t (t, x) H(x, DV (t, x)) = 0, with terminal condition V (T, x) = F(x)

12 The Hamilton-Jacobi equation Assuming V n V, with V smooth and i/n t, we may approximate V n ( i n + 1 n, x + 1 n z) V n ( i n, x) V ( i n + 1 n, x + 1 n z) V ( i n, x) 1 [ ] V t (t, x) + DV (t, x), z, n... which leads to 1 = exp{ V t (t, x) DV (t, x), z }θ(dz x) = e V t (t,x)+h(x, DV (t,x))... and taking logarithm on both sides yields V t (t, x) H(x, DV (t, x)) = 0, with terminal condition V (T, x) = F(x)

13 The Hamilton-Jacobi equation We have formally derived the Hamilton-Jacobi equation for the exponential rate of decay V (t, x): V t (t, x) H(x, DV (t, x)) = 0, V (T, x) = F(x). Using the theory of viscosity solutions and variational representation one can show that the unique viscosity solution can be given the variational representation { T } V (t, x) = inf L(φ(s), φ(s))ds + F(φ(T )), t where the infimum is taken over all absolutely continuous φ with φ(t) = x. In particular we can identify the exponential rate as the initial value: V (0, x 0 ) = inf{i T (φ) + F(φ(T ))}.

14 Exponential decay of the second moment Recall from Lecture 2 that the importance sampling estimator for computing E[exp{ nf (X n n )}] is given by dp dp α e nf( X n n ) = exp { n 1 i=0 ᾱ n i, Z i + H( X n i, ᾱ n i ) } exp{ nf ( X n n )}. The second moment of the estimator, starting from x at time j is given by: [( Ej,x ᾱ exp { n 1 i=j ᾱi n, Z i + H( X } i n, ᾱi n ) exp{ nf( X ) 2 ] n n )} [ ] = E j,x exp{sᾱn Sᾱj 2nF(Xn n )}.

15 Exponential decay of the second moment Recall from Lecture 2 that the importance sampling estimator for computing E[exp{ nf (X n n )}] is given by dp dp α e nf( X n n ) = exp { n 1 i=0 ᾱ n i, Z i + H( X n i, ᾱ n i ) } exp{ nf ( X n n )}. The second moment of the estimator, starting from x at time j is given by: [( Ej,x ᾱ exp { n 1 i=j ᾱi n, Z i + H( X } i n, ᾱi n ) exp{ nf( X ) 2 ] n n )} [ ] = E j,x exp{sᾱn Sᾱj 2nF(Xn n )}.

16 Exponential decay of the second moment Recall from Lecture 2 that the importance sampling estimator for computing E[exp{ nf (X n n )}] is given by dp dp α e nf( X n n ) = exp { n 1 i=0 ᾱ n i, Z i + H( X n i, ᾱ n i ) } exp{ nf ( X n n )}. The second moment of the estimator, starting from x at time j is given by: [( Ej,x ᾱ exp { n 1 i=j ᾱi n, Z i + H( X } i n, ᾱi n ) exp{ nf( X ) 2 ] n n )} [ ] = E j,x exp{sᾱn Sᾱj 2nF(Xn n )}.

17 Exponential decay of the second moment We expect that the second moment decays exponentially fast in n and as above we abuse notation (from Lecture 2) and redefine W n by W n ( j [ ] n, x) := 1 n log E j,x e Sᾱn Sᾱj 2nF(Xn n ). Using the backward equation for the second moment from Lecture 2 gives [ ] 0 = e n[w n ( j n + 1 n,x+ 1 n z) W n ( j n,x)] ᾱ j (x),z +H(x,ᾱ j (x)) 1 θ(dz x). Assuming W n W, with W smooth, j/n t, and ᾱj n (x) ᾱ(t, x) we may approximate W n ( j n + 1 n, x + 1 n z) W n ( j n, x) W ( j n + 1 n, x + 1 n z) W ( j n, x) 1 [ ] W t (t, x) + DW (t, x), z, n

18 Exponential decay of the second moment... which leads to 1 = e W t (t,x) DW (t,x)+ᾱ(t,x),z +H(x,ᾱ(t,x)) θ(dz x) = e W t (t,x)+h(x, DW (t,x) ᾱ(t,x))+h(x,ᾱ(t,x))... and taking logarithm on both sides yields W t (t, x) H(x, DW (t, x) ᾱ(t, x)) H(x, ᾱ(t, x)) = 0, with terminal condition W (T, x) = 2F(x).

19 Exponential decay of the second moment We have formally derived the Hamilton-Jacobi equation for the exponential rate of decay of the second moment of an importance sampling algorithm based on the change of measure ᾱ: W t (t, x) H(x, DW (t, x) ᾱ(t, x)) H(x, ᾱ(t, x)) = 0, W (T, x) = 2F(x). We have established a Hamilton-Jacobi equation for V (t, x) such that E[exp{ nf(x n (T ))}] e nv (0,x 0). Moreover, given a change of measure ᾱj n (x) ᾱ(t, x), as n and j/n t the second moment of the importance sampling estimator is approximately equal to exp{ nw (0, x 0 )}.

20 The role of subsolutions Let V is a classical subsolution: continuously differentiable s.t. V t (t, x) H(x, D V (t, x)) 0, V (T, x) F(x). Consider the importance sampling algorithm designed by taking ᾱ(t, x) = D V (t, x). Then W (t, x) = V (t, x) + V (t, x) is a viscosity subsolution to HJ-eqn for W.

21 The role of subsolutions Indeed, since V is a (viscosity) solution and V is a subsolution W t (t, x) H(x, D W (t, x) ᾱ(t, x)) H(x, ᾱ(t, x)) = V t (t, x) H(x, DV (t, x)) + V t (t, x) H(x, D V (t, x)) 0,... and for the terminal condition we have W (T, x) = V (T, x) + V (T, x) 2F(x). Hence, V + V is a viscosity subsolution and V (t, x) + V (t, x) W (t, x) for all t T and all x. In particular, at the starting point (0, x 0 ) we have V (0, x 0 ) + V (0, x 0 ) W (0, x 0 ).

22 The role of subsolutions The subsolution property leads to an asymptotic upper bound on the second moment exp{ nw (0, x 0 )} exp{ nv (0, x 0 )} exp{ n V (0, x 0 )}. We also have, from Jensen s inequality that the second moment is larger than the square of the first moment, exp{ nw (0, x 0 )} exp{ n2v (0, x 0 )}, which leads to W (0, x 0 ) 2V (0, x 0 ). Consequently, if we can find a classical subsolution V with V (0, x 0 ) = V (0, x 0 ), then the importance sampling algorithm based on taking ᾱ n j (x) = D V (j/n, x) will be asymptotically optimal.

23 The Gaussian random walk Construction of subsolutions Consider the mean of iid N(0, 1)-random variables and the probability P{X n n (,, a] [b, )}. Let For V i, i = 1, 2, we have V 1 (t, x) = a(b x) (1 t)h(a), V 2 (t, x) = b(a x) (1 t)h(a). V i t (t, x) H( DV i (t, x)) = H(a) H(a) = 0,

24 The Gaussian random walk Construction of subsolutions We propose to take V (t, x) = V 1 (t, x) V 2 (t, x). For the terminal condition we have V (1, x) = ab + ( ax) ( bx) 0, x / (a, b). Thus V is a subsolution (in fact a viscosity subsolution). A mollification argument leads us to take V as V (t, x) = δ log(e 1 δ V 1 (t,x) + e 1 δ V 2 (t,x) ), for some small δ > 0.

25 The epidemic model Construction of subsolutions Consider the probability that the infection, starting from x 0 > 1 ρ 1 reaches a high level x 1 > x 0 before returning to 1 ρ 1. This is a stationary problem and the design of an importance sampling algorithm is related to finding a classical subsolution to the stationary Hamilton-Jacobi equation H(x, DV (x)) = 0, V (x 1 ) = 0, V (1 ρ 1 ) =, where H(x, α) = λ 1 (x)(e α 1) + λ 1 (x)(e α 1).

26 The epidemic model Construction of subsolutions In this case we can actually work out the quasi-potential V (x). Indeed, first consider α(x) as a solution to H(x, α) = 0. We note that the solution is [ λ 1 (x) + λ 1 (x) α(x) = log ± 2λ 1 (x) (λ 1 (x) + λ 1 (x) 2λ 1 (x) To obtain V (x) we simply need to intergrate: V (x) = x1 x [ λ 1 (z) + λ 1 (z) log + 2λ 1 (z) ( λ 1 (z) + λ 1 (z) 2λ 1 (z) ) 2 λ 1 (x) ]. λ 1 (x) ) 2 λ 1 (z) ] dz. λ 1 (z)

Weak convergence and large deviation theory

Weak convergence and large deviation theory First Prev Next Go To Go Back Full Screen Close Quit 1 Weak convergence and large deviation theory Large deviation principle Convergence in distribution The Bryc-Varadhan theorem Tightness and Prohorov

More information

A Note On Large Deviation Theory and Beyond

A Note On Large Deviation Theory and Beyond A Note On Large Deviation Theory and Beyond Jin Feng In this set of notes, we will develop and explain a whole mathematical theory which can be highly summarized through one simple observation ( ) lim

More information

MIN-MAX REPRESENTATIONS OF VISCOSITY SOLUTIONS OF HAMILTON-JACOBI EQUATIONS AND APPLICATIONS IN RARE-EVENT SIMULATION

MIN-MAX REPRESENTATIONS OF VISCOSITY SOLUTIONS OF HAMILTON-JACOBI EQUATIONS AND APPLICATIONS IN RARE-EVENT SIMULATION MIN-MAX REPRESENTATIONS OF VISCOSITY SOLUTIONS OF HAMILTON-JACOBI EQUATIONS AND APPLICATIONS IN RARE-EVENT SIMULATION BOUALEM DJEHICHE, HENRIK HULT, AND PIERRE NYQUIST Abstract. In this paper a duality

More information

Importance Sampling for Jackson Networks

Importance Sampling for Jackson Networks Importance Sampling for Jackson Networks Paul Dupuis and Hui Wang Lefschetz Center for Dynamical Systems Brown University Providence, R.I. 02912, U.S.A. 12 March, 2008 Abstract Rare event simulation in

More information

Asymptotics and Simulation of Heavy-Tailed Processes

Asymptotics and Simulation of Heavy-Tailed Processes Asymptotics and Simulation of Heavy-Tailed Processes Department of Mathematics Stockholm, Sweden Workshop on Heavy-tailed Distributions and Extreme Value Theory ISI Kolkata January 14-17, 2013 Outline

More information

The Moment Method; Convex Duality; and Large/Medium/Small Deviations

The Moment Method; Convex Duality; and Large/Medium/Small Deviations Stat 928: Statistical Learning Theory Lecture: 5 The Moment Method; Convex Duality; and Large/Medium/Small Deviations Instructor: Sham Kakade The Exponential Inequality and Convex Duality The exponential

More information

March 1, Florida State University. Concentration Inequalities: Martingale. Approach and Entropy Method. Lizhe Sun and Boning Yang.

March 1, Florida State University. Concentration Inequalities: Martingale. Approach and Entropy Method. Lizhe Sun and Boning Yang. Florida State University March 1, 2018 Framework 1. (Lizhe) Basic inequalities Chernoff bounding Review for STA 6448 2. (Lizhe) Discrete-time martingales inequalities via martingale approach 3. (Boning)

More information

Proceedings of the 2014 Winter Simulation Conference A. Tolk, S. D. Diallo, I. O. Ryzhov, L. Yilmaz, S. Buckley, and J. A. Miller, eds.

Proceedings of the 2014 Winter Simulation Conference A. Tolk, S. D. Diallo, I. O. Ryzhov, L. Yilmaz, S. Buckley, and J. A. Miller, eds. Proceedings of the 014 Winter Simulation Conference A. Tolk, S. D. Diallo, I. O. Ryzhov, L. Yilmaz, S. Buckley, and J. A. Miller, eds. Rare event simulation in the neighborhood of a rest point Paul Dupuis

More information

Outline. Computer Science 331. Cost of Binary Search Tree Operations. Bounds on Height: Worst- and Average-Case

Outline. Computer Science 331. Cost of Binary Search Tree Operations. Bounds on Height: Worst- and Average-Case Outline Computer Science Average Case Analysis: Binary Search Trees Mike Jacobson Department of Computer Science University of Calgary Lecture #7 Motivation and Objective Definition 4 Mike Jacobson (University

More information

Gärtner-Ellis Theorem and applications.

Gärtner-Ellis Theorem and applications. Gärtner-Ellis Theorem and applications. Elena Kosygina July 25, 208 In this lecture we turn to the non-i.i.d. case and discuss Gärtner-Ellis theorem. As an application, we study Curie-Weiss model with

More information

Weak Convergence of Numerical Methods for Dynamical Systems and Optimal Control, and a relation with Large Deviations for Stochastic Equations

Weak Convergence of Numerical Methods for Dynamical Systems and Optimal Control, and a relation with Large Deviations for Stochastic Equations Weak Convergence of Numerical Methods for Dynamical Systems and, and a relation with Large Deviations for Stochastic Equations Mattias Sandberg KTH CSC 2010-10-21 Outline The error representation for weak

More information

Statistics 300B Winter 2018 Final Exam Due 24 Hours after receiving it

Statistics 300B Winter 2018 Final Exam Due 24 Hours after receiving it Statistics 300B Winter 08 Final Exam Due 4 Hours after receiving it Directions: This test is open book and open internet, but must be done without consulting other students. Any consultation of other students

More information

Exam February h

Exam February h Master 2 Mathématiques et Applications PUF Ho Chi Minh Ville 2009/10 Viscosity solutions, HJ Equations and Control O.Ley (INSA de Rennes) Exam February 2010 3h Written-by-hands documents are allowed. Printed

More information

21.1 Lower bounds on minimax risk for functional estimation

21.1 Lower bounds on minimax risk for functional estimation ECE598: Information-theoretic methods in high-dimensional statistics Spring 016 Lecture 1: Functional estimation & testing Lecturer: Yihong Wu Scribe: Ashok Vardhan, Apr 14, 016 In this chapter, we will

More information

Markov processes and queueing networks

Markov processes and queueing networks Inria September 22, 2015 Outline Poisson processes Markov jump processes Some queueing networks The Poisson distribution (Siméon-Denis Poisson, 1781-1840) { } e λ λ n n! As prevalent as Gaussian distribution

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing

More information

Computer Intensive Methods in Mathematical Statistics

Computer Intensive Methods in Mathematical Statistics Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of

More information

Chapter 8: Differential entropy. University of Illinois at Chicago ECE 534, Natasha Devroye

Chapter 8: Differential entropy. University of Illinois at Chicago ECE 534, Natasha Devroye Chapter 8: Differential entropy Chapter 8 outline Motivation Definitions Relation to discrete entropy Joint and conditional differential entropy Relative entropy and mutual information Properties AEP for

More information

Exercises in Extreme value theory

Exercises in Extreme value theory Exercises in Extreme value theory 2016 spring semester 1. Show that L(t) = logt is a slowly varying function but t ǫ is not if ǫ 0. 2. If the random variable X has distribution F with finite variance,

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation

More information

Large Deviations from the Hydrodynamic Limit for a System with Nearest Neighbor Interactions

Large Deviations from the Hydrodynamic Limit for a System with Nearest Neighbor Interactions Large Deviations from the Hydrodynamic Limit for a System with Nearest Neighbor Interactions Amarjit Budhiraja Department of Statistics & Operations Research University of North Carolina at Chapel Hill

More information

Ordinary Differential Equations II

Ordinary Differential Equations II Ordinary Differential Equations II February 9 217 Linearization of an autonomous system We consider the system (1) x = f(x) near a fixed point x. As usual f C 1. Without loss of generality we assume x

More information

Introduction to Self-normalized Limit Theory

Introduction to Self-normalized Limit Theory Introduction to Self-normalized Limit Theory Qi-Man Shao The Chinese University of Hong Kong E-mail: qmshao@cuhk.edu.hk Outline What is the self-normalization? Why? Classical limit theorems Self-normalized

More information

Convex Functions. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Convex Functions. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST) Convex Functions Daniel P. Palomar Hong Kong University of Science and Technology (HKUST) ELEC5470 - Convex Optimization Fall 2017-18, HKUST, Hong Kong Outline of Lecture Definition convex function Examples

More information

Constrained Optimization and Lagrangian Duality

Constrained Optimization and Lagrangian Duality CIS 520: Machine Learning Oct 02, 2017 Constrained Optimization and Lagrangian Duality Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may

More information

On Lyapunov Inequalities and Subsolutions for Efficient Importance Sampling

On Lyapunov Inequalities and Subsolutions for Efficient Importance Sampling On Lyapunov Inequalities and Subsolutions for Efficient Importance Sampling By Jose Blanchet, Kevin Leder and Peter Glynn Columbia University, Columbia University and Stanford University July 28, 2009

More information

Consistency of the maximum likelihood estimator for general hidden Markov models

Consistency of the maximum likelihood estimator for general hidden Markov models Consistency of the maximum likelihood estimator for general hidden Markov models Jimmy Olsson Centre for Mathematical Sciences Lund University Nordstat 2012 Umeå, Sweden Collaborators Hidden Markov models

More information

Annealed Brownian motion in a heavy tailed Poissonian potential

Annealed Brownian motion in a heavy tailed Poissonian potential Annealed Brownian motion in a heavy tailed Poissonian potential Ryoki Fukushima Tokyo Institute of Technology Workshop on Random Polymer Models and Related Problems, National University of Singapore, May

More information

Large deviation theory and applications

Large deviation theory and applications Large deviation theory and applications Peter Mörters November 0, 2008 Abstract Large deviation theory deals with the decay of the probability of increasingly unlikely events. It is one of the key techniques

More information

LECTURE 2. Convexity and related notions. Last time: mutual information: definitions and properties. Lecture outline

LECTURE 2. Convexity and related notions. Last time: mutual information: definitions and properties. Lecture outline LECTURE 2 Convexity and related notions Last time: Goals and mechanics of the class notation entropy: definitions and properties mutual information: definitions and properties Lecture outline Convexity

More information

LECTURE 3. Last time:

LECTURE 3. Last time: LECTURE 3 Last time: Mutual Information. Convexity and concavity Jensen s inequality Information Inequality Data processing theorem Fano s Inequality Lecture outline Stochastic processes, Entropy rate

More information

Concentration inequalities and the entropy method

Concentration inequalities and the entropy method Concentration inequalities and the entropy method Gábor Lugosi ICREA and Pompeu Fabra University Barcelona what is concentration? We are interested in bounding random fluctuations of functions of many

More information

Winter 2019 Math 106 Topics in Applied Mathematics. Lecture 8: Importance Sampling

Winter 2019 Math 106 Topics in Applied Mathematics. Lecture 8: Importance Sampling Winter 2019 Math 106 Topics in Applied Mathematics Data-driven Uncertainty Quantification Yoonsang Lee (yoonsang.lee@dartmouth.edu) Lecture 8: Importance Sampling 8.1 Importance Sampling Importance sampling

More information

CS281A/Stat241A Lecture 22

CS281A/Stat241A Lecture 22 CS281A/Stat241A Lecture 22 p. 1/4 CS281A/Stat241A Lecture 22 Monte Carlo Methods Peter Bartlett CS281A/Stat241A Lecture 22 p. 2/4 Key ideas of this lecture Sampling in Bayesian methods: Predictive distribution

More information

Mañé s Conjecture from the control viewpoint

Mañé s Conjecture from the control viewpoint Mañé s Conjecture from the control viewpoint Université de Nice - Sophia Antipolis Setting Let M be a smooth compact manifold of dimension n 2 be fixed. Let H : T M R be a Hamiltonian of class C k, with

More information

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version Convex Optimization Theory Chapter 5 Exercises and Solutions: Extended Version Dimitri P. Bertsekas Massachusetts Institute of Technology Athena Scientific, Belmont, Massachusetts http://www.athenasc.com

More information

in Bounded Domains Ariane Trescases CMLA, ENS Cachan

in Bounded Domains Ariane Trescases CMLA, ENS Cachan CMLA, ENS Cachan Joint work with Yan GUO, Chanwoo KIM and Daniela TONON International Conference on Nonlinear Analysis: Boundary Phenomena for Evolutionnary PDE Academia Sinica December 21, 214 Outline

More information

Monte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan

Monte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan Monte-Carlo MMD-MA, Université Paris-Dauphine Xiaolu Tan tan@ceremade.dauphine.fr Septembre 2015 Contents 1 Introduction 1 1.1 The principle.................................. 1 1.2 The error analysis

More information

Duality and dynamics in Hamilton-Jacobi theory for fully convex problems of control

Duality and dynamics in Hamilton-Jacobi theory for fully convex problems of control Duality and dynamics in Hamilton-Jacobi theory for fully convex problems of control RTyrrell Rockafellar and Peter R Wolenski Abstract This paper describes some recent results in Hamilton- Jacobi theory

More information

Explicit Bounds for the Distribution Function of the Sum of Dependent Normally Distributed Random Variables

Explicit Bounds for the Distribution Function of the Sum of Dependent Normally Distributed Random Variables Explicit Bounds for the Distribution Function of the Sum of Dependent Normally Distributed Random Variables Walter Schneider July 26, 20 Abstract In this paper an analytic expression is given for the bounds

More information

Introduction to Rare Event Simulation

Introduction to Rare Event Simulation Introduction to Rare Event Simulation Brown University: Summer School on Rare Event Simulation Jose Blanchet Columbia University. Department of Statistics, Department of IEOR. Blanchet (Columbia) 1 / 31

More information

l approche des grandes déviations

l approche des grandes déviations Épuisement de ressources partagées en environnement Markovien : l approche des grandes déviations Joint work with François Delarue (Paris Diderot) et René Schott (Nancy 1) Journées de Probabilité à Lille

More information

Importance Sampling for Rare Events

Importance Sampling for Rare Events Importance Sampling for Rare Events Soren Asmussen Department of Mathematical Sciences Aarhus University Ny Munkegade DK-8000 Aarhus C, Denmark asmus@imf.au.dk Paul Dupuis Division of Applied Mathematics

More information

18.175: Lecture 15 Characteristic functions and central limit theorem

18.175: Lecture 15 Characteristic functions and central limit theorem 18.175: Lecture 15 Characteristic functions and central limit theorem Scott Sheffield MIT Outline Characteristic functions Outline Characteristic functions Characteristic functions Let X be a random variable.

More information

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales.

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. Lecture 2 1 Martingales We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. 1.1 Doob s inequality We have the following maximal

More information

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016 Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016 1 Entropy Since this course is about entropy maximization,

More information

Lecture Quantitative Finance Spring Term 2015

Lecture Quantitative Finance Spring Term 2015 on bivariate Lecture Quantitative Finance Spring Term 2015 Prof. Dr. Erich Walter Farkas Lecture 07: April 2, 2015 1 / 54 Outline on bivariate 1 2 bivariate 3 Distribution 4 5 6 7 8 Comments and conclusions

More information

(Classical) Information Theory II: Source coding

(Classical) Information Theory II: Source coding (Classical) Information Theory II: Source coding Sibasish Ghosh The Institute of Mathematical Sciences CIT Campus, Taramani, Chennai 600 113, India. p. 1 Abstract The information content of a random variable

More information

6.1 Variational representation of f-divergences

6.1 Variational representation of f-divergences ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 Lecture 6: Variational representation, HCR and CR lower bounds Lecturer: Yihong Wu Scribe: Georgios Rovatsos, Feb 11, 2016

More information

Mean field games and related models

Mean field games and related models Mean field games and related models Fabio Camilli SBAI-Dipartimento di Scienze di Base e Applicate per l Ingegneria Facoltà di Ingegneria Civile ed Industriale Email: Camilli@dmmm.uniroma1.it Web page:

More information

Quantitative Biology II Lecture 4: Variational Methods

Quantitative Biology II Lecture 4: Variational Methods 10 th March 2015 Quantitative Biology II Lecture 4: Variational Methods Gurinder Singh Mickey Atwal Center for Quantitative Biology Cold Spring Harbor Laboratory Image credit: Mike West Summary Approximate

More information

18.175: Lecture 8 Weak laws and moment-generating/characteristic functions

18.175: Lecture 8 Weak laws and moment-generating/characteristic functions 18.175: Lecture 8 Weak laws and moment-generating/characteristic functions Scott Sheffield MIT 18.175 Lecture 8 1 Outline Moment generating functions Weak law of large numbers: Markov/Chebyshev approach

More information

Lecture 6: Bayesian Inference in SDE Models

Lecture 6: Bayesian Inference in SDE Models Lecture 6: Bayesian Inference in SDE Models Bayesian Filtering and Smoothing Point of View Simo Särkkä Aalto University Simo Särkkä (Aalto) Lecture 6: Bayesian Inference in SDEs 1 / 45 Contents 1 SDEs

More information

Intrinsic Noise in Nonlinear Gene Regulation Inference

Intrinsic Noise in Nonlinear Gene Regulation Inference Intrinsic Noise in Nonlinear Gene Regulation Inference Chao Du Department of Statistics, University of Virginia Joint Work with Wing H. Wong, Department of Statistics, Stanford University Transcription

More information

Control Theory: From Classical to Quantum Optimal, Stochastic, and Robust Control

Control Theory: From Classical to Quantum Optimal, Stochastic, and Robust Control Control Theory: From Classical to Quantum Optimal, Stochastic, and Robust Control Notes for Quantum Control Summer School, Caltech, August 2005 M.R. James Department of Engineering Australian National

More information

Lecture 1: Background on Convex Analysis

Lecture 1: Background on Convex Analysis Lecture 1: Background on Convex Analysis John Duchi PCMI 2016 Outline I Convex sets 1.1 Definitions and examples 2.2 Basic properties 3.3 Projections onto convex sets 4.4 Separating and supporting hyperplanes

More information

Spectral Theory of Orthogonal Polynomials

Spectral Theory of Orthogonal Polynomials Spectral Theory of Orthogonal Barry Simon IBM Professor of Mathematics and Theoretical Physics California Institute of Technology Pasadena, CA, U.S.A. Lecture 2: Szegö Theorem for OPUC for S Spectral Theory

More information

General Theory of Large Deviations

General Theory of Large Deviations Chapter 30 General Theory of Large Deviations A family of random variables follows the large deviations principle if the probability of the variables falling into bad sets, representing large deviations

More information

Large-deviation theory and coverage in mobile phone networks

Large-deviation theory and coverage in mobile phone networks Weierstrass Institute for Applied Analysis and Stochastics Large-deviation theory and coverage in mobile phone networks Paul Keeler, Weierstrass Institute for Applied Analysis and Stochastics, Berlin joint

More information

Large deviations for random projections of l p balls

Large deviations for random projections of l p balls 1/32 Large deviations for random projections of l p balls Nina Gantert CRM, september 5, 2016 Goal: Understanding random projections of high-dimensional convex sets. 2/32 2/32 Outline Goal: Understanding

More information

LANGEVIN EQUATION AND THERMODYNAMICS

LANGEVIN EQUATION AND THERMODYNAMICS LANGEVIN EQUATION AND THERMODYNAMICS RELATING STOCHASTIC DYNAMICS WITH THERMODYNAMIC LAWS November 10, 2017 1 / 20 MOTIVATION There are at least three levels of description of classical dynamics: thermodynamic,

More information

Lecture 22: Final Review

Lecture 22: Final Review Lecture 22: Final Review Nuts and bolts Fundamental questions and limits Tools Practical algorithms Future topics Dr Yao Xie, ECE587, Information Theory, Duke University Basics Dr Yao Xie, ECE587, Information

More information

Exercises with solutions (Set D)

Exercises with solutions (Set D) Exercises with solutions Set D. A fair die is rolled at the same time as a fair coin is tossed. Let A be the number on the upper surface of the die and let B describe the outcome of the coin toss, where

More information

Closed-Loop Impulse Control of Oscillating Systems

Closed-Loop Impulse Control of Oscillating Systems Closed-Loop Impulse Control of Oscillating Systems A. N. Daryin and A. B. Kurzhanski Moscow State (Lomonosov) University Faculty of Computational Mathematics and Cybernetics Periodic Control Systems, 2007

More information

Statistical physics models belonging to the generalised exponential family

Statistical physics models belonging to the generalised exponential family Statistical physics models belonging to the generalised exponential family Jan Naudts Universiteit Antwerpen 1. The generalized exponential family 6. The porous media equation 2. Theorem 7. The microcanonical

More information

3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions

3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions 3. Convex functions Convex Optimization Boyd & Vandenberghe basic properties and examples operations that preserve convexity the conjugate function quasiconvex functions log-concave and log-convex functions

More information

Average-cost temporal difference learning and adaptive control variates

Average-cost temporal difference learning and adaptive control variates Average-cost temporal difference learning and adaptive control variates Sean Meyn Department of ECE and the Coordinated Science Laboratory Joint work with S. Mannor, McGill V. Tadic, Sheffield S. Henderson,

More information

Concentration inequalities and tail bounds

Concentration inequalities and tail bounds Concentration inequalities and tail bounds John Duchi Outline I Basics and motivation 1 Law of large numbers 2 Markov inequality 3 Cherno bounds II Sub-Gaussian random variables 1 Definitions 2 Examples

More information

An Introduction to Malliavin calculus and its applications

An Introduction to Malliavin calculus and its applications An Introduction to Malliavin calculus and its applications Lecture 3: Clark-Ocone formula David Nualart Department of Mathematics Kansas University University of Wyoming Summer School 214 David Nualart

More information

Nonparametric estimation under Shape Restrictions

Nonparametric estimation under Shape Restrictions Nonparametric estimation under Shape Restrictions Jon A. Wellner University of Washington, Seattle Statistical Seminar, Frejus, France August 30 - September 3, 2010 Outline: Five Lectures on Shape Restrictions

More information

16. Working with the Langevin and Fokker-Planck equations

16. Working with the Langevin and Fokker-Planck equations 16. Working with the Langevin and Fokker-Planck equations In the preceding Lecture, we have shown that given a Langevin equation (LE), it is possible to write down an equivalent Fokker-Planck equation

More information

10-704: Information Processing and Learning Spring Lecture 8: Feb 5

10-704: Information Processing and Learning Spring Lecture 8: Feb 5 10-704: Information Processing and Learning Spring 2015 Lecture 8: Feb 5 Lecturer: Aarti Singh Scribe: Siheng Chen Disclaimer: These notes have not been subjected to the usual scrutiny reserved for formal

More information

SOLVABLE VARIATIONAL PROBLEMS IN N STATISTICAL MECHANICS

SOLVABLE VARIATIONAL PROBLEMS IN N STATISTICAL MECHANICS SOLVABLE VARIATIONAL PROBLEMS IN NON EQUILIBRIUM STATISTICAL MECHANICS University of L Aquila October 2013 Tullio Levi Civita Lecture 2013 Coauthors Lorenzo Bertini Alberto De Sole Alessandra Faggionato

More information

Upper Bound for Intermediate Singular Values of Random Sub-Gaussian Matrices 1

Upper Bound for Intermediate Singular Values of Random Sub-Gaussian Matrices 1 Upper Bound for Intermediate Singular Values of Random Sub-Gaussian Matrices 1 Feng Wei 2 University of Michigan July 29, 2016 1 This presentation is based a project under the supervision of M. Rudelson.

More information

Concentration Inequalities

Concentration Inequalities Chapter Concentration Inequalities I. Moment generating functions, the Chernoff method, and sub-gaussian and sub-exponential random variables a. Goal for this section: given a random variable X, how does

More information

Gradient Gibbs measures with disorder

Gradient Gibbs measures with disorder Gradient Gibbs measures with disorder Codina Cotar University College London April 16, 2015, Providence Partly based on joint works with Christof Külske Outline 1 The model 2 Questions 3 Known results

More information

Model Counting for Logical Theories

Model Counting for Logical Theories Model Counting for Logical Theories Wednesday Dmitry Chistikov Rayna Dimitrova Department of Computer Science University of Oxford, UK Max Planck Institute for Software Systems (MPI-SWS) Kaiserslautern

More information

x log x, which is strictly convex, and use Jensen s Inequality:

x log x, which is strictly convex, and use Jensen s Inequality: 2. Information measures: mutual information 2.1 Divergence: main inequality Theorem 2.1 (Information Inequality). D(P Q) 0 ; D(P Q) = 0 iff P = Q Proof. Let ϕ(x) x log x, which is strictly convex, and

More information

1 Stat 605. Homework I. Due Feb. 1, 2011

1 Stat 605. Homework I. Due Feb. 1, 2011 The first part is homework which you need to turn in. The second part is exercises that will not be graded, but you need to turn it in together with the take-home final exam. 1 Stat 605. Homework I. Due

More information

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R. Ergodic Theorems Samy Tindel Purdue University Probability Theory 2 - MA 539 Taken from Probability: Theory and examples by R. Durrett Samy T. Ergodic theorems Probability Theory 1 / 92 Outline 1 Definitions

More information

6 Markov Chain Monte Carlo (MCMC)

6 Markov Chain Monte Carlo (MCMC) 6 Markov Chain Monte Carlo (MCMC) The underlying idea in MCMC is to replace the iid samples of basic MC methods, with dependent samples from an ergodic Markov chain, whose limiting (stationary) distribution

More information

A large deviation principle for a RWRC in a box

A large deviation principle for a RWRC in a box A large deviation principle for a RWRC in a box 7th Cornell Probability Summer School Michele Salvi TU Berlin July 12, 2011 Michele Salvi (TU Berlin) An LDP for a RWRC in a nite box July 12, 2011 1 / 15

More information

EEL 5544 Noise in Linear Systems Lecture 30. X (s) = E [ e sx] f X (x)e sx dx. Moments can be found from the Laplace transform as

EEL 5544 Noise in Linear Systems Lecture 30. X (s) = E [ e sx] f X (x)e sx dx. Moments can be found from the Laplace transform as L30-1 EEL 5544 Noise in Linear Systems Lecture 30 OTHER TRANSFORMS For a continuous, nonnegative RV X, the Laplace transform of X is X (s) = E [ e sx] = 0 f X (x)e sx dx. For a nonnegative RV, the Laplace

More information

Convex Optimization & Lagrange Duality

Convex Optimization & Lagrange Duality Convex Optimization & Lagrange Duality Chee Wei Tan CS 8292 : Advanced Topics in Convex Optimization and its Applications Fall 2010 Outline Convex optimization Optimality condition Lagrange duality KKT

More information

4TE3/6TE3. Algorithms for. Continuous Optimization

4TE3/6TE3. Algorithms for. Continuous Optimization 4TE3/6TE3 Algorithms for Continuous Optimization (Algorithms for Constrained Nonlinear Optimization Problems) Tamás TERLAKY Computing and Software McMaster University Hamilton, November 2005 terlaky@mcmaster.ca

More information

A new Hellinger-Kantorovich distance between positive measures and optimal Entropy-Transport problems

A new Hellinger-Kantorovich distance between positive measures and optimal Entropy-Transport problems A new Hellinger-Kantorovich distance between positive measures and optimal Entropy-Transport problems Giuseppe Savaré http://www.imati.cnr.it/ savare Dipartimento di Matematica, Università di Pavia Nonlocal

More information

LARGE DEVIATIONS FOR STOCHASTIC PROCESSES

LARGE DEVIATIONS FOR STOCHASTIC PROCESSES LARGE DEVIATIONS FOR STOCHASTIC PROCESSES By Stefan Adams Abstract: The notes are devoted to results on large deviations for sequences of Markov processes following closely the book by Feng and Kurtz ([FK06]).

More information

Lecture 3 January 28

Lecture 3 January 28 EECS 28B / STAT 24B: Advanced Topics in Statistical LearningSpring 2009 Lecture 3 January 28 Lecturer: Pradeep Ravikumar Scribe: Timothy J. Wheeler Note: These lecture notes are still rough, and have only

More information

1 Solution to Problem 2.1

1 Solution to Problem 2.1 Solution to Problem 2. I incorrectly worked this exercise instead of 2.2, so I decided to include the solution anyway. a) We have X Y /3, which is a - function. It maps the interval, ) where X lives) onto

More information

Logarithmic Functions

Logarithmic Functions Name Student ID Number Group Name Group Members Logarithmic Functions 1. Solve the equations below. xx = xx = 5. Were you able solve both equations above? If so, was one of the equations easier to solve

More information

Lecture 8: Bayesian Estimation of Parameters in State Space Models

Lecture 8: Bayesian Estimation of Parameters in State Space Models in State Space Models March 30, 2016 Contents 1 Bayesian estimation of parameters in state space models 2 Computational methods for parameter estimation 3 Practical parameter estimation in state space

More information

8. Conjugate functions

8. Conjugate functions L. Vandenberghe EE236C (Spring 2013-14) 8. Conjugate functions closed functions conjugate function 8-1 Closed set a set C is closed if it contains its boundary: x k C, x k x = x C operations that preserve

More information

Lecture 15 Random variables

Lecture 15 Random variables Lecture 15 Random variables Weinan E 1,2 and Tiejun Li 2 1 Department of Mathematics, Princeton University, weinan@princeton.edu 2 School of Mathematical Sciences, Peking University, tieli@pku.edu.cn No.1

More information

Optimization and Optimal Control in Banach Spaces

Optimization and Optimal Control in Banach Spaces Optimization and Optimal Control in Banach Spaces Bernhard Schmitzer October 19, 2017 1 Convex non-smooth optimization with proximal operators Remark 1.1 (Motivation). Convex optimization: easier to solve,

More information

LECTURE NOTES 57. Lecture 9

LECTURE NOTES 57. Lecture 9 LECTURE NOTES 57 Lecture 9 17. Hypothesis testing A special type of decision problem is hypothesis testing. We partition the parameter space into H [ A with H \ A = ;. Wewrite H 2 H A 2 A. A decision problem

More information

On the interior of the simplex, we have the Hessian of d(x), Hd(x) is diagonal with ith. µd(w) + w T c. minimize. subject to w T 1 = 1,

On the interior of the simplex, we have the Hessian of d(x), Hd(x) is diagonal with ith. µd(w) + w T c. minimize. subject to w T 1 = 1, Math 30 Winter 05 Solution to Homework 3. Recognizing the convexity of g(x) := x log x, from Jensen s inequality we get d(x) n x + + x n n log x + + x n n where the equality is attained only at x = (/n,...,

More information

MARKOV CHAIN MONTE CARLO

MARKOV CHAIN MONTE CARLO MARKOV CHAIN MONTE CARLO RYAN WANG Abstract. This paper gives a brief introduction to Markov Chain Monte Carlo methods, which offer a general framework for calculating difficult integrals. We start with

More information

Max stable Processes & Random Fields: Representations, Models, and Prediction

Max stable Processes & Random Fields: Representations, Models, and Prediction Max stable Processes & Random Fields: Representations, Models, and Prediction Stilian Stoev University of Michigan, Ann Arbor March 2, 2011 Based on joint works with Yizao Wang and Murad S. Taqqu. 1 Preliminaries

More information

Math 141: Lecture 11

Math 141: Lecture 11 Math 141: Lecture 11 The Fundamental Theorem of Calculus and integration methods Bob Hough October 12, 2016 Bob Hough Math 141: Lecture 11 October 12, 2016 1 / 36 First Fundamental Theorem of Calculus

More information

Econ Lecture 14. Outline

Econ Lecture 14. Outline Econ 204 2010 Lecture 14 Outline 1. Differential Equations and Solutions 2. Existence and Uniqueness of Solutions 3. Autonomous Differential Equations 4. Complex Exponentials 5. Linear Differential Equations

More information