The Mabinogion Sheep Problem
|
|
- Melvin Sutton
- 6 years ago
- Views:
Transcription
1 The Mabinogion Sheep Problem Kun Dong Cornell University April 22, 2015 K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
2 Introduction (Williams 1991) we are given a herd of black and white sheep at time t = 0. At each time t = 1, 2,, a sheep is randomly selected, and a sheep of the opposite color (if any remain) becomes the same color as the chosen sheep. Just after t = 0 and each transition, we can remove any number of white sheep (if any remain) from the herd. The goal is to maximize the expected final number black sheep. K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
3 A naive lower bound Definition. A Policy is a function a : (Z ) 3 Z for which a(t, #white, #black) is the number of white sheep removed at time t. Suppose we start with w 0 white sheep and b 0 black sheep. One of the policy is that we immediately remove all white sheep at t = 0. Hence there is an easy lower bound for the optimal policy, which is the starting number of black sheep. K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
4 Decision Process Let (W t, B t ) be the number of white and black sheep at time t, with the natural filtration F t = σ((w s, B s ) s t ). ( ) ( ) P (W t+1, B t+1 ) = (w, b) F t = p (W t, B t ), a(t), (w, b) K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
5 Decision Process Let (W t, B t ) be the number of white and black sheep at time t, with the natural filtration F t = σ((w s, B s ) s t ). ( ) ( ) P (W t+1, B t+1 ) = (w, b) F t = p (W t, B t ), a(t), (w, b) What if a 0? Total number of sheep stays the same b 0 + w 0. p(b, b + 1) = b b 0 + w 0, b p(b, b 1) = 1 b 0 + w 0 for 0 < b < b 0 + w 0. K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
6 Transition probability For the Markov Chain B t with initially k sheep of each color, the transition matrix is 1 0 2k k 2k p = 2k i... i 2k 2k k 1 2k 2k 0 1 B t is not irreducible. {0, 2k} are recurrent but all other states are transient. Any π with π(0) + π(2k) = 1 is stationary. K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
7 Absorbing Markov Chain We can divide the communicating classes into transient classes and ergodic classes. K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
8 Absorbing Markov Chain We can divide the communicating classes into transient classes and ergodic classes. We call a chain an Absorbing Markov Chain when all ergodic classes are singleton sets. K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
9 Absorbing Markov Chain We can divide the communicating classes into transient classes and ergodic classes. We call a chain an Absorbing Markov Chain when all ergodic classes are singleton sets. K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
10 AMC cont. Theorem For any finite Markov chain, the probability that the process is in ergodic classes tends to 1 as n. Proof. For any transient state, there is a positive probability to leave the transient class, so there is also a positive probability to reach an ergodic class. Because we have finite states, there exists N and p such that the probability of entering an ergodic state within N steps is at least p. This means the probability of not entering an ergodic state after kn steps is less than (1 p) k 0 as k. K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
11 AMC cont. Theorem For any finite Markov chain, the probability that the process is in ergodic classes tends to 1 as n. Proof. For any transient state, there is a positive probability to leave the transient class, so there is also a positive probability to reach an ergodic class. Because we have finite states, there exists N and p such that the probability of entering an ergodic state within N steps is at least p. This means the probability of not entering an ergodic state after kn steps is less than (1 p) k 0 as k. Lemma P(B t {0, 2k}) 1 K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
12 Fundamental Matrix We can put the transition matrix P into the form 1 0 [ ] 0 1 S 0 P = = R Q 2k 1 1 2k 0 2k 2k 1 2k 2k i 2k... i 2k 1 2k 0 S - transition within ergodic classes T Q - transition within transient classes T R - transition from transient classes to ergodic classes The fundamental matrix is N = (I Q) 1 = I + Q + Q 2 + = k=0 Q k K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
13 N = (I + Q) 1 = k=0 Qk Remark 1. Let n j be the number of visits to a transient state j. E i [n j ] = p k ij = N ij Remark 2. In similar ways we can also get Var i (n j ), E i [ j T n j], Var i [ j T n j] from the fundamental matrix N. K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
14 N = (I + Q) 1 = k=0 Qk Remark 1. Let n j be the number of visits to a transient state j. E i [n j ] = p k ij = N ij Remark 2. In similar ways we can also get Var i (n j ), E i [ j T n j], Var i [ j T n j] from the fundamental matrix N. Theorem. Let b ij be the probability that the chain starts in transient state i and ends in absorbing state j, then {b ij } = B = NR. Proof. By Strong Markov property, b ij = p ij + k T p ik b kj = B = R + QB = (I Q)B = R K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
15 The sheep chain ( B 0 = 2k i=1 Figure: E[B ] with respect to b 0 1 (2k i)! E b0 [B ] = (2k)b b0,2k ( ) 2k 1 )/( 2k i 1 i=1 1 (2k i)! ( ) 2k 1 ) i 1 K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
16 Policy A Do Nothing, Policy A Reduce W t to B t 1, W t < B t W t B t Definition. The value function V (w, b) is the expected final number of black sheep under policy A if we start with w white sheep and b black sheep. K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
17 Policy A Do Nothing, Policy A Reduce W t to B t 1, W t < B t W t B t Definition. The value function V (w, b) is the expected final number of black sheep under policy A if we start with w white sheep and b black sheep. As a result, V has the following properties, (a1) V (0, b) = b (a2) V (w, b) = V (w 1, b) = V (b 1, b) whenever w b > 0. (a3) V (w, b) = w w+b V (w + 1, b 1) + b w+b V (w 1, b + 1) whenever b > w > 0. K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
18 Policy A Do Nothing, Policy A Reduce W t to B t 1, W t < B t W t B t Definition. The value function V (w, b) is the expected final number of black sheep under policy A if we start with w white sheep and b black sheep. As a result, V has the following properties, (a1) V (0, b) = b (a2) V (w, b) = V (w 1, b) = V (b 1, b) whenever w b > 0. (a3) V (w, b) = w w+b V (w + 1, b 1) + b b > w > 0. Claim. V (W n, B n ) is a martingale w.r.t {F n }. w+b V (w 1, b + 1) whenever V (W n+1, B n+1 ) = Wn W n+b n V (W n +1, B n 1)+ Bn W n+b n V (W n 1, B n +1) = V (W n, B n ) K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
19 Supermartingale Lemma 1. V (w, b) V (w 1, b) whenever w > 0 Lemma 2. V (w, b) w w+b V (w + 1, b 1) + b w+b V (w 1, b + 1) whenever w, b > 0. K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
20 Supermartingale Lemma 1. V (w, b) V (w 1, b) whenever w > 0 Lemma 2. V (w, b) w w+b V (w + 1, b 1) + b w+b V (w 1, b + 1) whenever w, b > 0. Suppose these two Lemmas hold, we have the theorem Theorem. For any policy, V (W n, B n ) is a supermartingale. Proof. Indeed, any action we take does not increase the value of V. Hence E[V (W n+1, B n+1 ) F n ] E[V (W n, B n ) F n ] = V (W n, B n ) K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
21 Martingale Convergence Because V (W n, B n ) is a supermartingale it converges almost surely. Hence the chain must end up in an absorbing state for which all sheep are of one color. In this case, for deterministic W 0 and B 0, E[B ] = E[V (W, B )] V (W 0, B 0 ) Therefore, for any initial number of black and white sheep, the expected final number of black sheep under any policy is no more than the expected final number of black sheep under Policy A. We conclude that Policy A is optimal K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
22 Long-term Limit Theorem V (k, k) (2k + π 4 πk) 0 Remark. If we start with 1000 black sheep and 1000 white sheep, we expect to finish with 1945 black sheep. Proof. p k = 2 2k( ) 2k k (πk) 1/2 by Stirling s Formula. Let α k = v k (2k + π 4 p k). For ρ k = 2p k 1+p k, v k+1 = v k + ρ k (2k + 1 v k ). α k+1 = (1 ρ k )α k ρ k c k for c k = (p k p k+1 )(1 + p k ) 2p 2 k p k+1 Because c k 0 and (1 ρ k ) 0, α k 0 π 4 K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
23 Continuous Case Let B t, W t be the number of black and white sheep, but for t R. A t is also a continuous time, non-decreasing process (Heuristically da t = a t dt, a t 0). X t = B t + W t, Y t = B t W t. T = inf{t 0 : Y t X t }. dx t = da t dy t = Y t X t dt + da t + 2dβ t V (x, y) = max A E (x,y)[x T 1 {YT X T }] K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
24 Application Analogous idea in Portfolio Selection with Transaction Cost (Davis and Norman, 1990). An invester decide between a bank account paying a fixed interest rate, and a stock with log-normal diffusion price. He consumes from the bank, and tries to maximize the expectation of consumption E 0 e δt u(c(t))dt. Let s 0 (t) be the holding in bank and s 1 (t) be the holding in stock. ds 0 (t) = (rs 0 (t) c(t))dt ds 1 (t) = αs 1 (t) + σs 1 (t)db In this case, we want s 0(t) s 1 (t) = π and c(t) s 0 (t)+s 1 (t) = C. K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
25 References Terence Chan, Some diffusion models for the mabinogion sheep problem of williams, Advances in applied probability (1996), Mark HA Davis and Andrew R Norman, Portfolio selection with transaction costs, Mathematics of Operations Research 15 (1990), no. 4, John G Kemeny and James Laurie Snell, Finite markov chains: with a new appendix generalization of a fundamental matrix, Springer, David Williams, Probability with martingales, Cambridge university press, K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
26 Thank you! K. Dong (Cornell University) The Mabinogion Sheep Problem April 22, / 18
Markov Chains Absorption Hamid R. Rabiee
Markov Chains Absorption Hamid R. Rabiee Absorbing Markov Chain An absorbing state is one in which the probability that the process remains in that state once it enters the state is (i.e., p ii = ). A
More informationMarkov Chains Absorption (cont d) Hamid R. Rabiee
Markov Chains Absorption (cont d) Hamid R. Rabiee 1 Absorbing Markov Chain An absorbing state is one in which the probability that the process remains in that state once it enters the state is 1 (i.e.,
More informationChapter 5. Continuous-Time Markov Chains. Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan
Chapter 5. Continuous-Time Markov Chains Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan Continuous-Time Markov Chains Consider a continuous-time stochastic process
More informationP i [B k ] = lim. n=1 p(n) ii <. n=1. V i :=
2.7. Recurrence and transience Consider a Markov chain {X n : n N 0 } on state space E with transition matrix P. Definition 2.7.1. A state i E is called recurrent if P i [X n = i for infinitely many n]
More informationMarkov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains
Markov Chains A random process X is a family {X t : t T } of random variables indexed by some set T. When T = {0, 1, 2,... } one speaks about a discrete-time process, for T = R or T = [0, ) one has a continuous-time
More informationThe Transition Probability Function P ij (t)
The Transition Probability Function P ij (t) Consider a continuous time Markov chain {X(t), t 0}. We are interested in the probability that in t time units the process will be in state j, given that it
More informationMATH 56A: STOCHASTIC PROCESSES CHAPTER 1
MATH 56A: STOCHASTIC PROCESSES CHAPTER. Finite Markov chains For the sake of completeness of these notes I decided to write a summary of the basic concepts of finite Markov chains. The topics in this chapter
More informationHJB equations. Seminar in Stochastic Modelling in Economics and Finance January 10, 2011
Department of Probability and Mathematical Statistics Faculty of Mathematics and Physics, Charles University in Prague petrasek@karlin.mff.cuni.cz Seminar in Stochastic Modelling in Economics and Finance
More informationMarkov Chains (Part 3)
Markov Chains (Part 3) State Classification Markov Chains - State Classification Accessibility State j is accessible from state i if p ij (n) > for some n>=, meaning that starting at state i, there is
More informationIrreducibility. Irreducible. every state can be reached from every other state For any i,j, exist an m 0, such that. Absorbing state: p jj =1
Irreducibility Irreducible every state can be reached from every other state For any i,j, exist an m 0, such that i,j are communicate, if the above condition is valid Irreducible: all states are communicate
More informationThe concentration of a drug in blood. Exponential decay. Different realizations. Exponential decay with noise. dc(t) dt.
The concentration of a drug in blood Exponential decay C12 concentration 2 4 6 8 1 C12 concentration 2 4 6 8 1 dc(t) dt = µc(t) C(t) = C()e µt 2 4 6 8 1 12 time in minutes 2 4 6 8 1 12 time in minutes
More informationPROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS
PROBABILITY: LIMIT THEOREMS II, SPRING 15. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please
More informationRECURSION EQUATION FOR
Math 46 Lecture 8 Infinite Horizon discounted reward problem From the last lecture: The value function of policy u for the infinite horizon problem with discount factor a and initial state i is W i, u
More informationSMSTC (2007/08) Probability.
SMSTC (27/8) Probability www.smstc.ac.uk Contents 12 Markov chains in continuous time 12 1 12.1 Markov property and the Kolmogorov equations.................... 12 2 12.1.1 Finite state space.................................
More informationMinimization of ruin probabilities by investment under transaction costs
Minimization of ruin probabilities by investment under transaction costs Stefan Thonhauser DSA, HEC, Université de Lausanne 13 th Scientific Day, Bonn, 3.4.214 Outline Introduction Risk models and controls
More informationLet (Ω, F) be a measureable space. A filtration in discrete time is a sequence of. F s F t
2.2 Filtrations Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of σ algebras {F t } such that F t F and F t F t+1 for all t = 0, 1,.... In continuous time, the second condition
More informationCS145: Probability & Computing Lecture 18: Discrete Markov Chains, Equilibrium Distributions
CS145: Probability & Computing Lecture 18: Discrete Markov Chains, Equilibrium Distributions Instructor: Erik Sudderth Brown University Computer Science April 14, 215 Review: Discrete Markov Chains Some
More informationLatent voter model on random regular graphs
Latent voter model on random regular graphs Shirshendu Chatterjee Cornell University (visiting Duke U.) Work in progress with Rick Durrett April 25, 2011 Outline Definition of voter model and duality with
More informationSocial network analysis: social learning
Social network analysis: social learning Donglei Du (ddu@unb.edu) Faculty of Business Administration, University of New Brunswick, NB Canada Fredericton E3B 9Y2 October 20, 2016 Donglei Du (UNB) AlgoTrading
More informationMarkov Model. Model representing the different resident states of a system, and the transitions between the different states
Markov Model Model representing the different resident states of a system, and the transitions between the different states (applicable to repairable, as well as non-repairable systems) System behavior
More informationSection Notes 9. Midterm 2 Review. Applied Math / Engineering Sciences 121. Week of December 3, 2018
Section Notes 9 Midterm 2 Review Applied Math / Engineering Sciences 121 Week of December 3, 2018 The following list of topics is an overview of the material that was covered in the lectures and sections
More informationErgodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.
Ergodic Theorems Samy Tindel Purdue University Probability Theory 2 - MA 539 Taken from Probability: Theory and examples by R. Durrett Samy T. Ergodic theorems Probability Theory 1 / 92 Outline 1 Definitions
More informationMarkov Chains. Arnoldo Frigessi Bernd Heidergott November 4, 2015
Markov Chains Arnoldo Frigessi Bernd Heidergott November 4, 2015 1 Introduction Markov chains are stochastic models which play an important role in many applications in areas as diverse as biology, finance,
More informationA MODEL FOR THE LONG-TERM OPTIMAL CAPACITY LEVEL OF AN INVESTMENT PROJECT
A MODEL FOR HE LONG-ERM OPIMAL CAPACIY LEVEL OF AN INVESMEN PROJEC ARNE LØKKA AND MIHAIL ZERVOS Abstract. We consider an investment project that produces a single commodity. he project s operation yields
More informationLecture 10. Theorem 1.1 [Ergodicity and extremality] A probability measure µ on (Ω, F) is ergodic for T if and only if it is an extremal point in M.
Lecture 10 1 Ergodic decomposition of invariant measures Let T : (Ω, F) (Ω, F) be measurable, and let M denote the space of T -invariant probability measures on (Ω, F). Then M is a convex set, although
More informationISM206 Lecture, May 12, 2005 Markov Chain
ISM206 Lecture, May 12, 2005 Markov Chain Instructor: Kevin Ross Scribe: Pritam Roy May 26, 2005 1 Outline of topics for the 10 AM lecture The topics are: Discrete Time Markov Chain Examples Chapman-Kolmogorov
More informationP(X 0 = j 0,... X nk = j k )
Introduction to Probability Example Sheet 3 - Michaelmas 2006 Michael Tehranchi Problem. Let (X n ) n 0 be a homogeneous Markov chain on S with transition matrix P. Given a k N, let Z n = X kn. Prove that
More informationMarkov Chains (Part 4)
Markov Chains (Part 4) Steady State Probabilities and First Passage Times Markov Chains - 1 Steady-State Probabilities Remember, for the inventory example we had (8) P &.286 =.286.286 %.286 For an irreducible
More informationMarkov Processes Hamid R. Rabiee
Markov Processes Hamid R. Rabiee Overview Markov Property Markov Chains Definition Stationary Property Paths in Markov Chains Classification of States Steady States in MCs. 2 Markov Property A discrete
More informationOptimism in the Face of Uncertainty Should be Refutable
Optimism in the Face of Uncertainty Should be Refutable Ronald ORTNER Montanuniversität Leoben Department Mathematik und Informationstechnolgie Franz-Josef-Strasse 18, 8700 Leoben, Austria, Phone number:
More informationLecture 4: Introduction to stochastic processes and stochastic calculus
Lecture 4: Introduction to stochastic processes and stochastic calculus Cédric Archambeau Centre for Computational Statistics and Machine Learning Department of Computer Science University College London
More informationStatistics 150: Spring 2007
Statistics 150: Spring 2007 April 23, 2008 0-1 1 Limiting Probabilities If the discrete-time Markov chain with transition probabilities p ij is irreducible and positive recurrent; then the limiting probabilities
More informationNon-homogeneous random walks on a semi-infinite strip
Non-homogeneous random walks on a semi-infinite strip Chak Hei Lo Joint work with Andrew R. Wade World Congress in Probability and Statistics 11th July, 2016 Outline Motivation: Lamperti s problem Our
More informationSTOCHASTIC PROCESSES Basic notions
J. Virtamo 38.3143 Queueing Theory / Stochastic processes 1 STOCHASTIC PROCESSES Basic notions Often the systems we consider evolve in time and we are interested in their dynamic behaviour, usually involving
More informationClassification of Countable State Markov Chains
Classification of Countable State Markov Chains Friday, March 21, 2014 2:01 PM How can we determine whether a communication class in a countable state Markov chain is: transient null recurrent positive
More informationExample: physical systems. If the state space. Example: speech recognition. Context can be. Example: epidemics. Suppose each infected
4. Markov Chains A discrete time process {X n,n = 0,1,2,...} with discrete state space X n {0,1,2,...} is a Markov chain if it has the Markov property: P[X n+1 =j X n =i,x n 1 =i n 1,...,X 0 =i 0 ] = P[X
More informationLecture 20 : Markov Chains
CSCI 3560 Probability and Computing Instructor: Bogdan Chlebus Lecture 0 : Markov Chains We consider stochastic processes. A process represents a system that evolves through incremental changes called
More informationProbability, Random Processes and Inference
INSTITUTO POLITÉCNICO NACIONAL CENTRO DE INVESTIGACION EN COMPUTACION Laboratorio de Ciberseguridad Probability, Random Processes and Inference Dr. Ponciano Jorge Escamilla Ambrosio pescamilla@cic.ipn.mx
More informationPROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS
PROBABILITY: LIMIT THEOREMS II, SPRING 218. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please
More informationExistence, Uniqueness and Stability of Invariant Distributions in Continuous-Time Stochastic Models
Existence, Uniqueness and Stability of Invariant Distributions in Continuous-Time Stochastic Models Christian Bayer and Klaus Wälde Weierstrass Institute for Applied Analysis and Stochastics and University
More informationIntroduction General Framework Toy models Discrete Markov model Data Analysis Conclusion. The Micro-Price. Sasha Stoikov. Cornell University
The Micro-Price Sasha Stoikov Cornell University Jim Gatheral @ NYU High frequency traders (HFT) HFTs are good: Optimal order splitting Pairs trading / statistical arbitrage Market making / liquidity provision
More informationMarkov Processes Cont d. Kolmogorov Differential Equations
Markov Processes Cont d Kolmogorov Differential Equations The Kolmogorov Differential Equations characterize the transition functions {P ij (t)} of a Markov process. The time-dependent behavior of the
More informationLectures on Markov Chains
Lectures on Markov Chains David M. McClendon Department of Mathematics Ferris State University 2016 edition 1 Contents Contents 2 1 Markov chains 4 1.1 The definition of a Markov chain.....................
More informationIntroduction to Algorithmic Trading Strategies Lecture 3
Introduction to Algorithmic Trading Strategies Lecture 3 Trend Following Haksun Li haksun.li@numericalmethod.com www.numericalmethod.com References Introduction to Stochastic Calculus with Applications.
More informationLecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past.
1 Markov chain: definition Lecture 5 Definition 1.1 Markov chain] A sequence of random variables (X n ) n 0 taking values in a measurable state space (S, S) is called a (discrete time) Markov chain, if
More informationLecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321
Lecture 11: Introduction to Markov Chains Copyright G. Caire (Sample Lectures) 321 Discrete-time random processes A sequence of RVs indexed by a variable n 2 {0, 1, 2,...} forms a discretetime random process
More informationProperties of an infinite dimensional EDS system : the Muller s ratchet
Properties of an infinite dimensional EDS system : the Muller s ratchet LATP June 5, 2011 A ratchet source : wikipedia Plan 1 Introduction : The model of Haigh 2 3 Hypothesis (Biological) : The population
More informationOptimal Control of Partially Observable Piecewise Deterministic Markov Processes
Optimal Control of Partially Observable Piecewise Deterministic Markov Processes Nicole Bäuerle based on a joint work with D. Lange Wien, April 2018 Outline Motivation PDMP with Partial Observation Controlled
More informationUNIVERSITY OF LONDON IMPERIAL COLLEGE LONDON
UNIVERSITY OF LONDON IMPERIAL COLLEGE LONDON BSc and MSci EXAMINATIONS (MATHEMATICS) MAY JUNE 23 This paper is also taken for the relevant examination for the Associateship. M3S4/M4S4 (SOLUTIONS) APPLIED
More informationStochastic process. X, a series of random variables indexed by t
Stochastic process X, a series of random variables indexed by t X={X(t), t 0} is a continuous time stochastic process X={X(t), t=0,1, } is a discrete time stochastic process X(t) is the state at time t,
More informationChapter 16 focused on decision making in the face of uncertainty about one future
9 C H A P T E R Markov Chains Chapter 6 focused on decision making in the face of uncertainty about one future event (learning the true state of nature). However, some decisions need to take into account
More informationMonopoly An Analysis using Markov Chains. Benjamin Bernard
Monopoly An Analysis using Markov Chains Benjamin Bernard Columbia University, New York PRIME event at Pace University, New York April 19, 2017 Introduction Applications of Markov chains Applications of
More informationNotes on Heterogeneity, Aggregation, and Market Wage Functions: An Empirical Model of Self-Selection in the Labor Market
Notes on Heterogeneity, Aggregation, and Market Wage Functions: An Empirical Model of Self-Selection in the Labor Market Heckman and Sedlacek, JPE 1985, 93(6), 1077-1125 James Heckman University of Chicago
More informationA Method for Evaluating the Distribution of the Total Cost of a Random Process over its Lifetime
A Method for Evaluating the Distribution of the Total Cost of a Random Process over its Lifetime P.K. Pollett a and V.T. Stefanov b a Department of Mathematics, The University of Queensland Queensland
More informationP (A G) dp G P (A G)
First homework assignment. Due at 12:15 on 22 September 2016. Homework 1. We roll two dices. X is the result of one of them and Z the sum of the results. Find E [X Z. Homework 2. Let X be a r.v.. Assume
More informationSummary of Results on Markov Chains. Abstract
Summary of Results on Markov Chains Enrico Scalas 1, 1 Laboratory on Complex Systems. Dipartimento di Scienze e Tecnologie Avanzate, Università del Piemonte Orientale Amedeo Avogadro, Via Bellini 25 G,
More informationSTAT STOCHASTIC PROCESSES. Contents
STAT 3911 - STOCHASTIC PROCESSES ANDREW TULLOCH Contents 1. Stochastic Processes 2 2. Classification of states 2 3. Limit theorems for Markov chains 4 4. First step analysis 5 5. Branching processes 5
More informationShadow prices and well-posedness in the problem of optimal investment and consumption with transaction costs
Shadow prices and well-posedness in the problem of optimal investment and consumption with transaction costs Mihai Sîrbu, The University of Texas at Austin based on joint work with Jin Hyuk Choi and Gordan
More informationCourse 16:198:520: Introduction To Artificial Intelligence Lecture 13. Decision Making. Abdeslam Boularias. Wednesday, December 7, 2016
Course 16:198:520: Introduction To Artificial Intelligence Lecture 13 Decision Making Abdeslam Boularias Wednesday, December 7, 2016 1 / 45 Overview We consider probabilistic temporal models where the
More informationONLINE APPENDIX TO: NONPARAMETRIC IDENTIFICATION OF THE MIXED HAZARD MODEL USING MARTINGALE-BASED MOMENTS
ONLINE APPENDIX TO: NONPARAMETRIC IDENTIFICATION OF THE MIXED HAZARD MODEL USING MARTINGALE-BASED MOMENTS JOHANNES RUF AND JAMES LEWIS WOLTER Appendix B. The Proofs of Theorem. and Proposition.3 The proof
More informationFunctional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals
Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals Noèlia Viles Cuadros BCAM- Basque Center of Applied Mathematics with Prof. Enrico
More informationReflected Brownian Motion
Chapter 6 Reflected Brownian Motion Often we encounter Diffusions in regions with boundary. If the process can reach the boundary from the interior in finite time with positive probability we need to decide
More informationFast-slow systems with chaotic noise
Fast-slow systems with chaotic noise David Kelly Ian Melbourne Courant Institute New York University New York NY www.dtbkelly.com May 12, 215 Averaging and homogenization workshop, Luminy. Fast-slow systems
More informationMarkov Chain BSDEs and risk averse networks
Markov Chain BSDEs and risk averse networks Samuel N. Cohen Mathematical Institute, University of Oxford (Based on joint work with Ying Hu, Robert Elliott, Lukas Szpruch) 2nd Young Researchers in BSDEs
More informationStochastic Differential Equations.
Chapter 3 Stochastic Differential Equations. 3.1 Existence and Uniqueness. One of the ways of constructing a Diffusion process is to solve the stochastic differential equation dx(t) = σ(t, x(t)) dβ(t)
More informationStochastic process for macro
Stochastic process for macro Tianxiao Zheng SAIF 1. Stochastic process The state of a system {X t } evolves probabilistically in time. The joint probability distribution is given by Pr(X t1, t 1 ; X t2,
More informationMarkov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued
Markov Chains X(t) is a Markov Process if, for arbitrary times t 1 < t 2
More informationQuantitative Model Checking (QMC) - SS12
Quantitative Model Checking (QMC) - SS12 Lecture 06 David Spieler Saarland University, Germany June 4, 2012 1 / 34 Deciding Bisimulations 2 / 34 Partition Refinement Algorithm Notation: A partition P over
More informationStochastic Shortest Path Problems
Chapter 8 Stochastic Shortest Path Problems 1 In this chapter, we study a stochastic version of the shortest path problem of chapter 2, where only probabilities of transitions along different arcs can
More informationINTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING
INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING ERIC SHANG Abstract. This paper provides an introduction to Markov chains and their basic classifications and interesting properties. After establishing
More informationWEAK VERSIONS OF STOCHASTIC ADAMS-BASHFORTH AND SEMI-IMPLICIT LEAPFROG SCHEMES FOR SDES. 1. Introduction
WEAK VERSIONS OF STOCHASTIC ADAMS-BASHFORTH AND SEMI-IMPLICIT LEAPFROG SCHEMES FOR SDES BRIAN D. EWALD 1 Abstract. We consider the weak analogues of certain strong stochastic numerical schemes considered
More informationReinforcement Learning
Reinforcement Learning March May, 2013 Schedule Update Introduction 03/13/2015 (10:15-12:15) Sala conferenze MDPs 03/18/2015 (10:15-12:15) Sala conferenze Solving MDPs 03/20/2015 (10:15-12:15) Aula Alpha
More informationChapter 2 SOME ANALYTICAL TOOLS USED IN THE THESIS
Chapter 2 SOME ANALYTICAL TOOLS USED IN THE THESIS 63 2.1 Introduction In this chapter we describe the analytical tools used in this thesis. They are Markov Decision Processes(MDP), Markov Renewal process
More informationLinear Dependence of Stationary Distributions in Ergodic Markov Decision Processes
Linear ependence of Stationary istributions in Ergodic Markov ecision Processes Ronald Ortner epartment Mathematik und Informationstechnologie, Montanuniversität Leoben Abstract In ergodic MPs we consider
More information1/2 1/2 1/4 1/4 8 1/2 1/2 1/2 1/2 8 1/2 6 P =
/ 7 8 / / / /4 4 5 / /4 / 8 / 6 P = 0 0 0 0 0 0 0 0 0 0 4 0 0 0 0 0 0 0 0 4 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 Andrei Andreevich Markov (856 9) In Example. 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 P (n) = 0
More informationLecture 12. F o s, (1.1) F t := s>t
Lecture 12 1 Brownian motion: the Markov property Let C := C(0, ), R) be the space of continuous functions mapping from 0, ) to R, in which a Brownian motion (B t ) t 0 almost surely takes its value. Let
More informationSimple Consumption / Savings Problems (based on Ljungqvist & Sargent, Ch 16, 17) Jonathan Heathcote. updated, March The household s problem X
Simple Consumption / Savings Problems (based on Ljungqvist & Sargent, Ch 16, 17) subject to for all t Jonathan Heathcote updated, March 2006 1. The household s problem max E β t u (c t ) t=0 c t + a t+1
More informationThe Markov Decision Process (MDP) model
Decision Making in Robots and Autonomous Agents The Markov Decision Process (MDP) model Subramanian Ramamoorthy School of Informatics 25 January, 2013 In the MAB Model We were in a single casino and the
More informationStationary Probabilities of Markov Chains with Upper Hessenberg Transition Matrices
Stationary Probabilities of Marov Chains with Upper Hessenberg Transition Matrices Y. Quennel ZHAO Department of Mathematics and Statistics University of Winnipeg Winnipeg, Manitoba CANADA R3B 2E9 Susan
More informationp 1 ( Y p dp) 1/p ( X p dp) 1 1 p
Doob s inequality Let X(t) be a right continuous submartingale with respect to F(t), t 1 P(sup s t X(s) λ) 1 λ {sup s t X(s) λ} X + (t)dp 2 For 1 < p
More informationStochastic Processes. Theory for Applications. Robert G. Gallager CAMBRIDGE UNIVERSITY PRESS
Stochastic Processes Theory for Applications Robert G. Gallager CAMBRIDGE UNIVERSITY PRESS Contents Preface page xv Swgg&sfzoMj ybr zmjfr%cforj owf fmdy xix Acknowledgements xxi 1 Introduction and review
More informationUniformly Uniformly-ergodic Markov chains and BSDEs
Uniformly Uniformly-ergodic Markov chains and BSDEs Samuel N. Cohen Mathematical Institute, University of Oxford (Based on joint work with Ying Hu, Robert Elliott, Lukas Szpruch) Centre Henri Lebesgue,
More informationAssortment Optimization under the Multinomial Logit Model with Nested Consideration Sets
Assortment Optimization under the Multinomial Logit Model with Nested Consideration Sets Jacob Feldman School of Operations Research and Information Engineering, Cornell University, Ithaca, New York 14853,
More informationOn Polynomial Cases of the Unichain Classification Problem for Markov Decision Processes
On Polynomial Cases of the Unichain Classification Problem for Markov Decision Processes Eugene A. Feinberg Department of Applied Mathematics and Statistics State University of New York at Stony Brook
More informationA comparison of numerical methods for the. Solution of continuous-time DSGE models. Juan Carlos Parra Alvarez
A comparison of numerical methods for the solution of continuous-time DSGE models Juan Carlos Parra Alvarez Department of Economics and Business, and CREATES Aarhus University, Denmark November 14, 2012
More informationMarkov processes Course note 2. Martingale problems, recurrence properties of discrete time chains.
Institute for Applied Mathematics WS17/18 Massimiliano Gubinelli Markov processes Course note 2. Martingale problems, recurrence properties of discrete time chains. [version 1, 2017.11.1] We introduce
More informationA Barrier Version of the Russian Option
A Barrier Version of the Russian Option L. A. Shepp, A. N. Shiryaev, A. Sulem Rutgers University; shepp@stat.rutgers.edu Steklov Mathematical Institute; shiryaev@mi.ras.ru INRIA- Rocquencourt; agnes.sulem@inria.fr
More informationOptimal exit strategies for investment projects. 7th AMaMeF and Swissquote Conference
Optimal exit strategies for investment projects Simone Scotti Université Paris Diderot Laboratoire de Probabilité et Modèles Aléatories Joint work with : Etienne Chevalier, Université d Evry Vathana Ly
More information8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains
8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains 8.1 Review 8.2 Statistical Equilibrium 8.3 Two-State Markov Chain 8.4 Existence of P ( ) 8.5 Classification of States
More informationIntroduction Optimality and Asset Pricing
Introduction Optimality and Asset Pricing Andrea Buraschi Imperial College Business School October 2010 The Euler Equation Take an economy where price is given with respect to the numéraire, which is our
More informationMathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )
Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio (2014-2015) Etienne Tanré - Olivier Faugeras INRIA - Team Tosca November 26th, 2014 E. Tanré (INRIA - Team Tosca) Mathematical
More informationMarch 16, Abstract. We study the problem of portfolio optimization under the \drawdown constraint" that the
ON PORTFOLIO OPTIMIZATION UNDER \DRAWDOWN" CONSTRAINTS JAKSA CVITANIC IOANNIS KARATZAS y March 6, 994 Abstract We study the problem of portfolio optimization under the \drawdown constraint" that the wealth
More informationPortfolio Optimization with unobservable Markov-modulated drift process
Portfolio Optimization with unobservable Markov-modulated drift process Ulrich Rieder Department of Optimization and Operations Research University of Ulm, Germany D-89069 Ulm, Germany e-mail: rieder@mathematik.uni-ulm.de
More informationOperations Research Letters. Instability of FIFO in a simple queueing system with arbitrarily low loads
Operations Research Letters 37 (2009) 312 316 Contents lists available at ScienceDirect Operations Research Letters journal homepage: www.elsevier.com/locate/orl Instability of FIFO in a simple queueing
More informationOptimal Stopping Problems and American Options
Optimal Stopping Problems and American Options Nadia Uys A dissertation submitted to the Faculty of Science, University of the Witwatersrand, in fulfilment of the requirements for the degree of Master
More informationOptimal portfolio strategies under partial information with expert opinions
1 / 35 Optimal portfolio strategies under partial information with expert opinions Ralf Wunderlich Brandenburg University of Technology Cottbus, Germany Joint work with Rüdiger Frey Research Seminar WU
More informationMATH3200, Lecture 31: Applications of Eigenvectors. Markov Chains and Chemical Reaction Systems
Lecture 31: Some Applications of Eigenvectors: Markov Chains and Chemical Reaction Systems Winfried Just Department of Mathematics, Ohio University April 9 11, 2018 Review: Eigenvectors and left eigenvectors
More informationA Detailed Look at a Discrete Randomw Walk with Spatially Dependent Moments and Its Continuum Limit
A Detailed Look at a Discrete Randomw Walk with Spatially Dependent Moments and Its Continuum Limit David Vener Department of Mathematics, MIT May 5, 3 Introduction In 8.366, we discussed the relationship
More informationHamilton-Jacobi-Bellman Equation of an Optimal Consumption Problem
Hamilton-Jacobi-Bellman Equation of an Optimal Consumption Problem Shuenn-Jyi Sheu Institute of Mathematics, Academia Sinica WSAF, CityU HK June 29-July 3, 2009 1. Introduction X c,π t is the wealth with
More informationLecture 9 Classification of States
Lecture 9: Classification of States of 27 Course: M32K Intro to Stochastic Processes Term: Fall 204 Instructor: Gordan Zitkovic Lecture 9 Classification of States There will be a lot of definitions and
More information