1 Gambler s Ruin Problem
|
|
- Barnaby Nelson
- 6 years ago
- Views:
Transcription
1 Coyright c 2017 by Karl Sigman 1 Gambler s Ruin Problem Let N 2 be an integer and let 1 i N 1. Consider a gambler who starts with an initial fortune of $i and then on each successive gamble either wins $1 or loses $1 indeendent of the ast with robabilities and q = 1 resectively. Let X n denote the total fortune after the n th gamble. The gambler s objective is to kee successively gambling so as to reach a total fortune of $N, without first getting ruined (running out of money). If the gambler succeeds, then the gambler is said to win the game. In any case, the gambler stos laying after winning or getting ruined, whichever haens first. In essence, we can view each successive gamble as an indeendent coin fli which lands heads (H) with robability and lands tails (T) with robability q = 1. Each time the coin is flied, the gambler wins 1 dollar if it lands H and loses 1 dollar if it lands T. As soon as total earnings hits N or 0, the gambler stos fliing; the game ends. While the game roceeds before stoing, {X n : n 0} forms a simle random walk X n = i n, n 1, X 0 = i, where { n } forms an indeendent and identically distributed (i.i.d.) sequence of r.v.s. distributed as P ( = 1) =, P ( = 1) = q = 1, and reresents the earnings on the successive gambles. E( ) = q = 2 1. If > 1/2, equivalently > q, then E( ) > 0; while if < 1/2, equivalently < q, then E( ) < 0. Finally, if = 1/2, equivalently = q, then E( ) = 0. Thus, when > 1/2, the gambling is (on average) in favor of the gambler, when < 1/2, the gambling is (on average) in favor of the casino. Only when = 1/2 is the game fair. Recursively, this can be re-written as X n+1 = X n + n+1, n 0. Since the game stos when either X n = 0 or X n = N, let τ i = min{n 0 : X n {0, N} X 0 = i}, denote the time at which the game stos when X 0 = i; that is the time until either 0 or N are hit (for the first time). If X τi = N, then the gambler wins, if X τi = 0, then the gambler is ruined. Let P i (N) = P (X τi = N) denote the robability that the gambler wins when X 0 = i. P i (N) denotes the robability that the gambler, starting initially with $i, reaches a total fortune of N before ruin; 1 P i (N) is thus the corresonding robably of ruin Clearly P 0 (N) = 0 and P N (N) = 1 by definition, and we next roceed to comute P i (N), 1 i N 1. Proosition 1.1 (Gambler s Ruin Problem Solution) 1 ( q )i P i (N) = 1 ( q, if q; )N (1) i N, if = q =
2 Proof : For our derivation, we let P i = P i (N), that is, we suress the deendence on N for ease of notation. The key idea is to condition on the outcome of the first gamble, 1 = 1 or 1 = 1, yielding P i = P i+1 + qp i 1. (2) The derivation of this recursion is as follows: If 1 = 1, then the gambler s total fortune increases to X 1 = i + 1 and what haens after this is indeendent of the ast; the gambler will now win with robability P i+1. Similarly, if 1 = 1, then the gambler s fortune decreases to X 1 = i 1 and so the gambler will now win with robability P i 1. The robabilities corresonding to the two outcomes are and q yielding (2). Since + q = 1, (2) can be re-written as P i + qp i = P i+1 + qp i 1, yielding P i+1 P i = q (P i P i 1 ). In articular, P 2 P 1 = (q/)(p 1 P 0 ) = (q/)p 1 (since P 0 = 0), so that P 3 P 2 = (q/)(p 2 P 1 ) = (q/) 2 P 1, and more generally Thus P i+1 P i = ( q )i P 1, 0 < i < N. P i+1 P 1 = = i (P k+1 P k ) k=1 i k=1 ( q )k P 1, yielding i P i+1 = P 1 + P 1 ( q i )k = P 1 ( q )k = k=1 k=0 1 ( q P )i ( q ), if q; P 1 (i + 1), if = q = 0.5. (3) (Here we are using the geometric series equation i n=0 a n = 1 ai+1 1 a, for any number a 1 and any integer i 1, and of course i n=0 1 n = 1 + i.) Choosing i = N 1 and using the fact that P N = 1 yields 1 ( q 1 = P N = P )N 1 1 ( q, if q; ) P 1 N, if = q = 0.5, from which we conclude that P 1 = 1 q 1 ( q )N, if q; 1 N, if = q = 0.5, 2
3 thus obtaining from (3) (after algebra) the solution 1 ( q )i P i = 1 ( q, if q; )N (4) i N, if = q = 0.5. Remark 1.1 {X n : n 0} yields a Markov chain (MC) on the state sace S = {0, 1,..., N}. def The transition robabilities, P i,j = P (X n+1 = j X n = i), are given by P i,i+1 =, P i,i 1 = q, 0 < i < N, and both 0 and N are absorbing states, P 00 = P NN = 1. 1 For examle, when N = 4 the transition matrix P = (P i,j ) is given by q P = 0 q q Becoming infinitely rich or getting ruined In the formula (1), it is of interest to see what haens as N ; denote this by P i ( ) = lim N P i (N). This limiting quantity denotes the robability that the gambler, if allowed to lay forever unless ruined, will in fact never get ruined and instead will obtain an infinitely large fortune. Formally, this is true because of the basic result in robability that if {A n : n 1} is either an increasing or decreasing sequence of events, then lim n P (A n ) = P (lim n A n ). Increasing means that A n A n+1, n 1, in which case lim A n def = n A n, n=1 while decreasing means that A n A n+1, n 1 in which case lim A n def = A n, n In our alication, A N = {Gambler reaches N before 0 given that X 0 = i} is decreasing in N, and P i (N) = P (A N ), and n=1 lim A N = {Gambler reaches before 0 given that X 0 = i}. N Proosition 1.2 Define P i ( ) = lim N P i (N). If > 0.5, then If 0.50, then P i ( ) = 1 ( q )i > 0. (5) P i ( ) = 0. (6) Thus, unless the gambles are strictly better than fair ( > 0.5), ruin is certain. 1 There are three communication classes: C 1 = {0}, C 2 = {1,..., N 1}, C 3 = {N}. C 1 and C 3 are recurrent whereas C 2 is transient. 3
4 Proof : If > 0.5, then q < 1; hence in the denominator of (1), ( q )N 0 yielding the result. If < 0.50, then q > 1; hence in the denominator of (1), ( q )N yielding the result. Finally, if = 0.5, then i (N) = i/n 0. Examles 1. John starts with $2, and = 0.6: What is the robability that John obtains a fortune of N = 4 without going broke? i = 2, N = 4 and q = 1 = 0.4, so q/ = 2/3, and we want P 2 (4) = 1 (2/3)2 1 (2/3) 4 = What is the robability that John will become infinitely rich? P 2 ( ) = 1 (2/3) 2 = 5/9 = If John instead started with i = $1, what is the robability that he would go broke? The robability he becomes infinitely rich is P 1 ( ) = 1 (q/) = 1/3, so the robability of ruin is 1 P 1 ( ) = 2/ Alications Risk insurance business Consider an insurance comany that earns $1 er day (from interest), but on each day, indeendent of the ast, might suffer a claim against it for the amount $2 with robability q = 1. Whenever such a claim is suffered, $2 is removed from its reserve of money. Thus on the n th day, the net income for that day is exactly n as in the gamblers ruin roblem: 1 with robability, 1 with robability q. If the insurance comany starts off initially with a reserve of $i 1, then what is the robability it will eventually get ruined (run out of money)? The answer is given by (5) and (6): If > 0.5 then the robability is given by ( q )i > 0, whereas if 0.5 ruin will always ocurr. This makes intuitive sense because if > 0.5, then the average net income er day is E( ) = q > 0, whereas if 0.5, then the average net income er day is E( ) = q 0. So the comany can not exect to stay in business unless earning (on average) more than is taken away by claims. In later lectures, we will learn how to solve for the ruin robabilities for the much more general model in which the claims arrive according to a Poisson rocess, money comes in at a constant rate c and the claim sizes have a general distribution G: The classic Cramér-Lundberg Model. If the insurance business starts out with a reserve of x > 0, what is the robability it will get ruined? The answer turns out to be deely related to a queueing model in stationarity. 4
5 1.3 Random walk hitting robabilities Let R n = n, n 1, R 0 = 0 denote a simle random walk initially at the origin, the { n } are iid with P ( = 1) =, P ( = 1) = q = 1. Note that E( ) = q = 2 1 and σ 2 = V ar( ) = E( 2 ) E 2 ( ) = 1 (2 1) 2 = 4(1 ). Let a > 0 and b > 0 be integers and define (a) = P ({R n } hits level a before hitting level b). In words this is the robability that the random walk goes u by the amount a before going down by the amount b. By letting i = b, and N = a + b, we can equivalently imagine a gambler who starts with i = b and wishes to reach N = a+b before going broke (e.g. go u by the amount a before going down by the amount b ). So we can comute (a) by casting the roblem into the framework of the gambler s ruin roblem: (a) = P i (N) where N = a + b, i = b. Thus Examles 1 ( q )b (a) = 1 ( q, if q; )a+b (7) b a+b, if = q = Ellen bought a share of stock for $10, and it is believed that the stock rice moves (day by day) as a simle random walk with = What is the robability that Ellen s stock reaches the high value of $15 before the low value of $5? We want the robability that the stock goes u by 5 before going down by 5. This is equivalent to starting the random walk at 0 with a = 5 and b = 5, and comuting (a). (a) = 1 ( q )b 1 (0.82)5 1 ( q = = 0.73 )a+b 1 (0.82) What is the robability that Ellen will become infinitely rich (without ever hitting 0 first)? Here we equivalently want to know the robability that a gambler starting with i = 10 becomes infinitely rich before going broke. Just like Examle 2 on Page 3: 1 (q/) i = 1 (0.82) = Maximums and minimums of the simle random walk Formula (7) can immediately be used for comuting the robability that the simle random walk {R n }, starting initially at R 0 = 0, will ever hit level a, for any given ositive integer a 1: Kee a fixed while taking the limit as b in (7). The result deends on wether < 0.50 or A little thought reveals that we can equivalently state this roblem as comuting the tail P (M a), a 0, where M def = max{r n : n 0} is the all-time maximum of the 5
6 random walk; a non-negative random variable, because the corresonding two events are equal; {M a} = {R n = a, for some n 1}: On the one hand, if R n at some time n ever hits as high as a, it might hit even higher at a later time, hence M a. On the other hand, if M a, then R n must have hit a for some n. The increasing/decreasing sequence of events framework/results that we used in justifying Proosition 1.2 is used here as well: the events A b = {R n hits a before b } are increasing in b, and P (A b ) = (a) and thus lim b P (A b ) = P (lim b A b ) = {R n = a, for some n 1} = {M a}. Proosition 1.3 Let M def = max{r n : n 0} for the simle random walk starting initially at the origin (R 0 = 0). 1. When < 0.50, P (M a) = (/q) a, a 0; M has a geometric distribution with success robability 1 (/q): P (M = k) = (/q) k (1 (/q)), k 0. In this case, the random walk drifts down to, w1, but before doing so reaches the finite maximum M. 2. If 0.50, then P (M a) = 1, a 0: P (M = ) = 1; the random walk will, with robability 1, reach any ositive integer a no matter how large. Proof : Taking the limit in (7) as b yields the result by considering the two cases < 0.5 or 0.5: If < 0.5, then (q/) > 1 and so both (q/) b and (q/) a+b tend to as b. But before taking the limit, multily both numerator and denominator by (q/) b = (/q) b, yielding (a) = (/q)b 1 (/q) b (q/) a. Since (/q) b 0 as b, the result follows. If > 0.5, then (q/) < 1 and so both (q/) b and (q/) a+b tend to 0 as b yielding the limit in (7) as 1. If = 0.5, then (a) = b/(b + a) 1 as b. If < 0.5, then E( ) < 0, and if > 0.5, then E( ) > 0; so Proosition 1.3 is consistent with the fact that any random walk with E( ) < 0 (called the negative drift case) satisfies lim n R n =, w1, and any random walk with E( ) > 0 ( called the ositive drift case) satisfies lim n R n = +, w1. 2 But furthermore we learn that when < 0.5, although w1 the chain drifts off to, it first reaches a finite maximum M before doing so, and this rv M has a geometric distribution. Finally Proosition 1.3 also offers us a roof that when = 0.5, the symmetric case, the random walk will w1 hit any ositive value, P (M a) = 1. By symmetry, we also obtain analogous results for the minimum, : 2 R From the strong law of large numbers, lim n n n + if E( ) > 0. = E( ), w1, so R n ne( ) if E( ) < 0 and 6
7 Corollary 1.1 Let m def = min{r n : n 0} for the simle random walk starting initially at the origin (R 0 = 0). 1. If > 0.5, then P (m b) = (q/) b, b 0. In this case, the random walk drifts u to +, but before doing so dros down to a finite minimum m 0. Taking absolute values of m makes it non-negative and so we can exress this result as P ( m b) = (q/) b, b 0; m has a geometric distribution with success robability 1 (q/): P ( m = k) = (q/) k (1 (q/)), k If 0.50, then P (m b) = 1, b 0: P (m = ) = 1; the random walk will, with robability 1, reach any negative integer a no matter how small. Note that when < 0.5, P (M = 0) = 1 (/q) > 0. This is because it is ossible that the random walk will never enter the ositive axis before drifting off to ; with ositive robability it will remain forever 0: P (R n 0, n 0) = P (M = 0) = 1 (/q). Similarly, if > 0.5, then P (m = 0) = 1 (q/) > 0; with ositive robability the random walk will never enter the negative axis before drifting off to +. Recurrence of the simle symmetric random walk Combining the results for both M and m in the revious section when = 0.5, we have Proosition 1.4 The simle symmetric ( = 0.50) random walk, starting at the origin, will w1 eventually hit any integer a, ositive or negative. In fact it will hit any given integer a infinitely often, always returning yet again after leaving; it is a recurrent Markov chain. Proof : The first statement follows directly from Proosition 1.3 and Corollary 1.1; P (M = ) = 1 = P (m = ) = 1. For the second statement we argue as follows: Using the first statement, we know that the simle symmetric random walk starting at 0 will hit 1 eventually. But when it does, we can use the same result to conclude that it will go back and hit 0 eventually after that, because that is stochastically equivalent to starting at R 0 = 0 and waiting for the chain to hit 1, which also will haen eventually. But then yet again it must hit 1 again and so on (back and forth infinitely often), all by the same logic. We conclude that the chain will, over and over again, return to state 0 w1; it will do so infinitely often; 0 is a recurrent state for the simle symmetric random walk. Thus (since the chain is irreducible) all states are recurrent. Let τ = min{n 1 : R n = 0 R 0 = 0}, the so-called return time to state 0. We just argued that τ is a roer random variable, that is, P (τ < ) = 1. This means that if the chain starts in state 0, then, if we wait long enough, we will (w1) see it return to state 0. What we will rove later is that E(τ) = ; meaning that on average our wait is infinite. This imlies that the simle symmetric random walk forms a null recurrent Markov chain. 7
1 Gambler s Ruin Problem
1 Gambler s Ruin Problem Consider a gambler who starts with an initial fortune of $1 and then on each successive gamble either wins $1 or loses $1 independent of the past with probabilities p and q = 1
More informationSolutions to In Class Problems Week 15, Wed.
Massachusetts Institute of Technology 6.04J/18.06J, Fall 05: Mathematics for Comuter Science December 14 Prof. Albert R. Meyer and Prof. Ronitt Rubinfeld revised December 14, 005, 1404 minutes Solutions
More information8 STOCHASTIC PROCESSES
8 STOCHASTIC PROCESSES The word stochastic is derived from the Greek στoχαστικoς, meaning to aim at a target. Stochastic rocesses involve state which changes in a random way. A Markov rocess is a articular
More informationECE 6960: Adv. Random Processes & Applications Lecture Notes, Fall 2010
ECE 6960: Adv. Random Processes & Alications Lecture Notes, Fall 2010 Lecture 16 Today: (1) Markov Processes, (2) Markov Chains, (3) State Classification Intro Please turn in H 6 today. Read Chater 11,
More informationHomework Solution 4 for APPM4/5560 Markov Processes
Homework Solution 4 for APPM4/556 Markov Processes 9.Reflecting random walk on the line. Consider the oints,,, 4 to be marked on a straight line. Let X n be a Markov chain that moves to the right with
More informationLecture 6. 2 Recurrence/transience, harmonic functions and martingales
Lecture 6 Classification of states We have shown that all states of an irreducible countable state Markov chain must of the same tye. This gives rise to the following classification. Definition. [Classification
More information6 Stationary Distributions
6 Stationary Distributions 6. Definition and Examles Definition 6.. Let {X n } be a Markov chain on S with transition robability matrix P. A distribution π on S is called stationary (or invariant) if π
More informationMarkov Chains (Part 3)
Markov Chains (Part 3) State Classification Markov Chains - State Classification Accessibility State j is accessible from state i if p ij (n) > for some n>=, meaning that starting at state i, there is
More informationSolution: (Course X071570: Stochastic Processes)
Solution I (Course X071570: Stochastic Processes) October 24, 2013 Exercise 1.1: Find all functions f from the integers to the real numbers satisfying f(n) = 1 2 f(n + 1) + 1 f(n 1) 1. 2 A secial solution
More informationChapter 11 Advanced Topic Stochastic Processes
Chapter 11 Advanced Topic Stochastic Processes CHAPTER OUTLINE Section 1 Simple Random Walk Section 2 Markov Chains Section 3 Markov Chain Monte Carlo Section 4 Martingales Section 5 Brownian Motion Section
More informationAnalysis of some entrance probabilities for killed birth-death processes
Analysis of some entrance robabilities for killed birth-death rocesses Master s Thesis O.J.G. van der Velde Suervisor: Dr. F.M. Sieksma July 5, 207 Mathematical Institute, Leiden University Contents Introduction
More informationStatistics 433 Practice Final Exam: Cover Sheet and Marking Sheet
Statistics 433 Practice Final Exam: Cover Sheet and Marking Sheet YOUR NAME INSTRUCTIONS: No notes, no calculators, and no communications devices are permitted. Please keep all materials away from your
More informationReadings: Finish Section 5.2
LECTURE 19 Readings: Finish Section 5.2 Lecture outline Markov Processes I Checkout counter example. Markov process: definition. -step transition probabilities. Classification of states. Example: Checkout
More informationMarkov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains
Markov Chains A random process X is a family {X t : t T } of random variables indexed by some set T. When T = {0, 1, 2,... } one speaks about a discrete-time process, for T = R or T = [0, ) one has a continuous-time
More informationIEOR 6711, HMWK 5, Professor Sigman
IEOR 6711, HMWK 5, Professor Sigman 1. Semi-Markov processes: Consider an irreducible positive recurrent discrete-time Markov chain {X n } with transition matrix P (P i,j ), i, j S, and finite state space.
More informationLecture 4 - Random walk, ruin problems and random processes
Lecture 4 - Random walk, ruin problems and random processes Jan Bouda FI MU April 19, 2009 Jan Bouda (FI MU) Lecture 4 - Random walk, ruin problems and random processesapril 19, 2009 1 / 30 Part I Random
More informationMATH 2710: NOTES FOR ANALYSIS
MATH 270: NOTES FOR ANALYSIS The main ideas we will learn from analysis center around the idea of a limit. Limits occurs in several settings. We will start with finite limits of sequences, then cover infinite
More informationLectures on Markov Chains
Lectures on Markov Chains David M. McClendon Department of Mathematics Ferris State University 2016 edition 1 Contents Contents 2 1 Markov chains 4 1.1 The definition of a Markov chain.....................
More information33 The Gambler's Ruin 1
33 The Gambler's Ruin 1 Figure 1 The Gambler's Ruin There are many variations of the gambler's ruin problem, but a principal one goes like this. We have i dollars out of a total of n. We flip a coin which
More informationAnswers to selected exercises
Answers to selected exercises A First Course in Stochastic Models, Henk C. Tijms 1.1 ( ) 1.2 (a) Let waiting time if passengers already arrived,. Then,, (b) { (c) Long-run fraction for is (d) Let waiting
More informationElementary Analysis in Q p
Elementary Analysis in Q Hannah Hutter, May Szedlák, Phili Wirth November 17, 2011 This reort follows very closely the book of Svetlana Katok 1. 1 Sequences and Series In this section we will see some
More informationChapter 16 focused on decision making in the face of uncertainty about one future
9 C H A P T E R Markov Chains Chapter 6 focused on decision making in the face of uncertainty about one future event (learning the true state of nature). However, some decisions need to take into account
More informationBrownian Motion and Random Prime Factorization
Brownian Motion and Random Prime Factorization Kendrick Tang June 4, 202 Contents Introduction 2 2 Brownian Motion 2 2. Develoing Brownian Motion.................... 2 2.. Measure Saces and Borel Sigma-Algebras.........
More informationLecture 9 Classification of States
Lecture 9: Classification of States of 27 Course: M32K Intro to Stochastic Processes Term: Fall 204 Instructor: Gordan Zitkovic Lecture 9 Classification of States There will be a lot of definitions and
More informationOn a Markov Game with Incomplete Information
On a Markov Game with Incomlete Information Johannes Hörner, Dinah Rosenberg y, Eilon Solan z and Nicolas Vieille x{ January 24, 26 Abstract We consider an examle of a Markov game with lack of information
More informationCOMMUNICATION BETWEEN SHAREHOLDERS 1
COMMUNICATION BTWN SHARHOLDRS 1 A B. O A : A D Lemma B.1. U to µ Z r 2 σ2 Z + σ2 X 2r ω 2 an additive constant that does not deend on a or θ, the agents ayoffs can be written as: 2r rθa ω2 + θ µ Y rcov
More informationSTOCHASTIC MODELS LECTURE 1 MARKOV CHAINS. Nan Chen MSc Program in Financial Engineering The Chinese University of Hong Kong (ShenZhen) Sept.
STOCHASTIC MODELS LECTURE 1 MARKOV CHAINS Nan Chen MSc Program in Financial Engineering The Chinese University of Hong Kong (ShenZhen) Sept. 6, 2016 Outline 1. Introduction 2. Chapman-Kolmogrov Equations
More informationEcon 101A Midterm 2 Th 8 April 2009.
Econ A Midterm Th 8 Aril 9. You have aroximately hour and minutes to answer the questions in the midterm. I will collect the exams at. shar. Show your work, and good luck! Problem. Production (38 oints).
More informationRANDOM WALKS IN ONE DIMENSION
RANDOM WALKS IN ONE DIMENSION STEVEN P. LALLEY 1. THE GAMBLER S RUIN PROBLEM 1.1. Statement of the problem. I have A dollars; my colleague Xinyi has B dollars. A cup of coffee at the Sacred Grounds in
More informationModule 8. Lecture 3: Markov chain
Lecture 3: Markov chain A Markov chain is a stochastic rocess having the roerty that the value of the rocess X t at time t, deends only on its value at time t-1, X t-1 and not on the sequence X t-2, X
More informationLecture: Condorcet s Theorem
Social Networs and Social Choice Lecture Date: August 3, 00 Lecture: Condorcet s Theorem Lecturer: Elchanan Mossel Scribes: J. Neeman, N. Truong, and S. Troxler Condorcet s theorem, the most basic jury
More informationLecture 20 : Markov Chains
CSCI 3560 Probability and Computing Instructor: Bogdan Chlebus Lecture 0 : Markov Chains We consider stochastic processes. A process represents a system that evolves through incremental changes called
More information1 Random Variables and Probability Distributions
1 Random Variables and Probability Distributions 1.1 Random Variables 1.1.1 Discrete random variables A random variable X is called discrete if the number of values that X takes is finite or countably
More informationThe Longest Run of Heads
The Longest Run of Heads Review by Amarioarei Alexandru This aer is a review of older and recent results concerning the distribution of the longest head run in a coin tossing sequence, roblem that arise
More informationISE/OR 760 Applied Stochastic Modeling
ISE/OR 760 Applied Stochastic Modeling Topic 2: Discrete Time Markov Chain Yunan Liu Department of Industrial and Systems Engineering NC State University Yunan Liu (NC State University) ISE/OR 760 1 /
More informationTopic: Lower Bounds on Randomized Algorithms Date: September 22, 2004 Scribe: Srinath Sridhar
15-859(M): Randomized Algorithms Lecturer: Anuam Guta Toic: Lower Bounds on Randomized Algorithms Date: Setember 22, 2004 Scribe: Srinath Sridhar 4.1 Introduction In this lecture, we will first consider
More informationMarkov Chains on Countable State Space
Markov Chains on Countable State Space 1 Markov Chains Introduction 1. Consider a discrete time Markov chain {X i, i = 1, 2,...} that takes values on a countable (finite or infinite) set S = {x 1, x 2,...},
More information1 Probability Spaces and Random Variables
1 Probability Saces and Random Variables 1.1 Probability saces Ω: samle sace consisting of elementary events (or samle oints). F : the set of events P: robability 1.2 Kolmogorov s axioms Definition 1.2.1
More informationB8.1 Martingales Through Measure Theory. Concept of independence
B8.1 Martingales Through Measure Theory Concet of indeendence Motivated by the notion of indeendent events in relims robability, we have generalized the concet of indeendence to families of σ-algebras.
More information2. Transience and Recurrence
Virtual Laboratories > 15. Markov Chains > 1 2 3 4 5 6 7 8 9 10 11 12 2. Transience and Recurrence The study of Markov chains, particularly the limiting behavior, depends critically on the random times
More informationMarkov Chains. Chapter 16. Markov Chains - 1
Markov Chains Chapter 16 Markov Chains - 1 Why Study Markov Chains? Decision Analysis focuses on decision making in the face of uncertainty about one future event. However, many decisions need to consider
More informationLecture 6 Random walks - advanced methods
Lecture 6: Random wals - advanced methods 1 of 11 Course: M362K Intro to Stochastic Processes Term: Fall 2014 Instructor: Gordan Zitovic Lecture 6 Random wals - advanced methods STOPPING TIMES Our last
More informationEE/Stats 376A: Information theory Winter Lecture 5 Jan 24. Lecturer: David Tse Scribe: Michael X, Nima H, Geng Z, Anton J, Vivek B.
EE/Stats 376A: Information theory Winter 207 Lecture 5 Jan 24 Lecturer: David Tse Scribe: Michael X, Nima H, Geng Z, Anton J, Vivek B. 5. Outline Markov chains and stationary distributions Prefix codes
More informationISM206 Lecture, May 12, 2005 Markov Chain
ISM206 Lecture, May 12, 2005 Markov Chain Instructor: Kevin Ross Scribe: Pritam Roy May 26, 2005 1 Outline of topics for the 10 AM lecture The topics are: Discrete Time Markov Chain Examples Chapman-Kolmogorov
More information1/2 1/2 1/4 1/4 8 1/2 1/2 1/2 1/2 8 1/2 6 P =
/ 7 8 / / / /4 4 5 / /4 / 8 / 6 P = 0 0 0 0 0 0 0 0 0 0 4 0 0 0 0 0 0 0 0 4 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 Andrei Andreevich Markov (856 9) In Example. 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 P (n) = 0
More informationLecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes
Lecture Notes 7 Random Processes Definition IID Processes Bernoulli Process Binomial Counting Process Interarrival Time Process Markov Processes Markov Chains Classification of States Steady State Probabilities
More informationOnline Appendix to Accompany AComparisonof Traditional and Open-Access Appointment Scheduling Policies
Online Aendix to Accomany AComarisonof Traditional and Oen-Access Aointment Scheduling Policies Lawrence W. Robinson Johnson Graduate School of Management Cornell University Ithaca, NY 14853-6201 lwr2@cornell.edu
More informationIntroduction to Stochastic Processes
18.445 Introduction to Stochastic Processes Lecture 1: Introduction to finite Markov chains Hao Wu MIT 04 February 2015 Hao Wu (MIT) 18.445 04 February 2015 1 / 15 Course description About this course
More informationAnalysis of M/M/n/K Queue with Multiple Priorities
Analysis of M/M/n/K Queue with Multile Priorities Coyright, Sanjay K. Bose For a P-riority system, class P of highest riority Indeendent, Poisson arrival rocesses for each class with i as average arrival
More information8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains
8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains 8.1 Review 8.2 Statistical Equilibrium 8.3 Two-State Markov Chain 8.4 Existence of P ( ) 8.5 Classification of States
More informationMarkov Processes Hamid R. Rabiee
Markov Processes Hamid R. Rabiee Overview Markov Property Markov Chains Definition Stationary Property Paths in Markov Chains Classification of States Steady States in MCs. 2 Markov Property A discrete
More information15-451/651: Design & Analysis of Algorithms October 23, 2018 Lecture #17: Prediction from Expert Advice last changed: October 25, 2018
5-45/65: Design & Analysis of Algorithms October 23, 208 Lecture #7: Prediction from Exert Advice last changed: October 25, 208 Prediction with Exert Advice Today we ll study the roblem of making redictions
More informationMATH HOMEWORK PROBLEMS D. MCCLENDON
MATH 46- HOMEWORK PROBLEMS D. MCCLENDON. Consider a Markov chain with state space S = {0, }, where p = P (0, ) and q = P (, 0); compute the following in terms of p and q: (a) P (X 2 = 0 X = ) (b) P (X
More informationAM 221: Advanced Optimization Spring Prof. Yaron Singer Lecture 6 February 12th, 2014
AM 221: Advanced Otimization Sring 2014 Prof. Yaron Singer Lecture 6 February 12th, 2014 1 Overview In our revious lecture we exlored the concet of duality which is the cornerstone of Otimization Theory.
More informationExtension of Minimax to Infinite Matrices
Extension of Minimax to Infinite Matrices Chris Calabro June 21, 2004 Abstract Von Neumann s minimax theorem is tyically alied to a finite ayoff matrix A R m n. Here we show that (i) if m, n are both inite,
More informationRandom variables. Lecture 5 - Discrete Distributions. Discrete Probability distributions. Example - Discrete probability model
Random Variables Random variables Lecture 5 - Discrete Distributions Sta02 / BME02 Colin Rundel Setember 8, 204 A random variable is a numeric uantity whose value deends on the outcome of a random event
More informationPROFIT MAXIMIZATION. π = p y Σ n i=1 w i x i (2)
PROFIT MAXIMIZATION DEFINITION OF A NEOCLASSICAL FIRM A neoclassical firm is an organization that controls the transformation of inuts (resources it owns or urchases into oututs or roducts (valued roducts
More informationINTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING
INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING ERIC SHANG Abstract. This paper provides an introduction to Markov chains and their basic classifications and interesting properties. After establishing
More informationRANDOM WALKS AND PERCOLATION: AN ANALYSIS OF CURRENT RESEARCH ON MODELING NATURAL PROCESSES
RANDOM WALKS AND PERCOLATION: AN ANALYSIS OF CURRENT RESEARCH ON MODELING NATURAL PROCESSES AARON ZWIEBACH Abstract. In this aer we will analyze research that has been recently done in the field of discrete
More informationSets of Real Numbers
Chater 4 Sets of Real Numbers 4. The Integers Z and their Proerties In our revious discussions about sets and functions the set of integers Z served as a key examle. Its ubiquitousness comes from the fact
More information1 Random Walks and Electrical Networks
CME 305: Discrete Mathematics and Algorithms Random Walks and Electrical Networks Random walks are widely used tools in algorithm design and probabilistic analysis and they have numerous applications.
More information1 IEOR 6712: Notes on Brownian Motion I
Copyright c 005 by Karl Sigman IEOR 67: Notes on Brownian Motion I We present an introduction to Brownian motion, an important continuous-time stochastic process that serves as a continuous-time analog
More informationDRAFT - do not circulate
An Introduction to Proofs about Concurrent Programs K. V. S. Prasad (for the course TDA383/DIT390) Deartment of Comuter Science Chalmers University Setember 26, 2016 Rough sketch of notes released since
More informationMarkov Chains Handout for Stat 110
Markov Chains Handout for Stat 0 Prof. Joe Blitzstein (Harvard Statistics Department) Introduction Markov chains were first introduced in 906 by Andrey Markov, with the goal of showing that the Law of
More informationOptimism, Delay and (In)Efficiency in a Stochastic Model of Bargaining
Otimism, Delay and In)Efficiency in a Stochastic Model of Bargaining Juan Ortner Boston University Setember 10, 2012 Abstract I study a bilateral bargaining game in which the size of the surlus follows
More informationImproved Capacity Bounds for the Binary Energy Harvesting Channel
Imroved Caacity Bounds for the Binary Energy Harvesting Channel Kaya Tutuncuoglu 1, Omur Ozel 2, Aylin Yener 1, and Sennur Ulukus 2 1 Deartment of Electrical Engineering, The Pennsylvania State University,
More informationRandom walk on a polygon
IMS Lecture Notes Monograph Series Recent Developments in Nonparametric Inference and Probability Vol. 50 (2006) 4 c Institute of Mathematical Statistics, 2006 DOI: 0.24/0749270600000058 Random walk on
More informationarxiv:cond-mat/ v2 25 Sep 2002
Energy fluctuations at the multicritical oint in two-dimensional sin glasses arxiv:cond-mat/0207694 v2 25 Se 2002 1. Introduction Hidetoshi Nishimori, Cyril Falvo and Yukiyasu Ozeki Deartment of Physics,
More information1 Some basic renewal theory: The Renewal Reward Theorem
Copyright c 27 by Karl Sigman Some basic renewal theory: The Renewal Reward Theorem Here, we will present some basic results in renewal theory such as the elementary renewal theorem, and then the very
More informationChapter 8 Markov Chains and Some Applications ( 馬哥夫鏈 )
Chater 8 arkov Chains and oe Alications ( 馬哥夫鏈 Consider a sequence of rando variables,,, and suose that the set of ossible values of these rando variables is {,,,, }, which is called the state sace. It
More informationMarkov Chains Introduction
Markov Chains 4 4.1. Introduction In this chapter, we consider a stochastic process {X n,n= 0, 1, 2,...} that takes on a finite or countable number of possible values. Unless otherwise mentioned, this
More informationOutline. Markov Chains and Markov Models. Outline. Markov Chains. Markov Chains Definitions Huizhen Yu
and Markov Models Huizhen Yu janey.yu@cs.helsinki.fi Det. Comuter Science, Univ. of Helsinki Some Proerties of Probabilistic Models, Sring, 200 Huizhen Yu (U.H.) and Markov Models Jan. 2 / 32 Huizhen Yu
More informationProbability, Random Processes and Inference
INSTITUTO POLITÉCNICO NACIONAL CENTRO DE INVESTIGACION EN COMPUTACION Laboratorio de Ciberseguridad Probability, Random Processes and Inference Dr. Ponciano Jorge Escamilla Ambrosio pescamilla@cic.ipn.mx
More informationDepartment of Mathematics
Deartment of Mathematics Ma 3/03 KC Border Introduction to Probability and Statistics Winter 209 Sulement : Series fun, or some sums Comuting the mean and variance of discrete distributions often involves
More informationAt the boundary states, we take the same rules except we forbid leaving the state space, so,.
Birth-death chains Monday, October 19, 2015 2:22 PM Example: Birth-Death Chain State space From any state we allow the following transitions: with probability (birth) with probability (death) with probability
More informationLecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321
Lecture 11: Introduction to Markov Chains Copyright G. Caire (Sample Lectures) 321 Discrete-time random processes A sequence of RVs indexed by a variable n 2 {0, 1, 2,...} forms a discretetime random process
More informationStatistics 253/317 Introduction to Probability Models. Winter Midterm Exam Monday, Feb 10, 2014
Statistics 253/317 Introduction to Probability Models Winter 2014 - Midterm Exam Monday, Feb 10, 2014 Student Name (print): (a) Do not sit directly next to another student. (b) This is a closed-book, closed-note
More informationTopic 7: Using identity types
Toic 7: Using identity tyes June 10, 2014 Now we would like to learn how to use identity tyes and how to do some actual mathematics with them. By now we have essentially introduced all inference rules
More informationEconomics 101. Lecture 7 - Monopoly and Oligopoly
Economics 0 Lecture 7 - Monooly and Oligooly Production Equilibrium After having exlored Walrasian equilibria with roduction in the Robinson Crusoe economy, we will now ste in to a more general setting.
More informationNANYANG TECHNOLOGICAL UNIVERSITY SEMESTER I EXAMINATION MH4702/MAS446/MTH437 Probabilistic Methods in OR
NANYANG TECHNOLOGICAL UNIVERSITY SEMESTER I EXAMINATION 2013-201 MH702/MAS6/MTH37 Probabilistic Methods in OR December 2013 TIME ALLOWED: 2 HOURS INSTRUCTIONS TO CANDIDATES 1. This examination paper contains
More information1 IEOR 4701: Continuous-Time Markov Chains
Copyright c 2006 by Karl Sigman 1 IEOR 4701: Continuous-Time Markov Chains A Markov chain in discrete time, {X n : n 0}, remains in any state for exactly one unit of time before making a transition (change
More informationLecture 2 : CS6205 Advanced Modeling and Simulation
Lecture 2 : CS6205 Advanced Modeling and Simulation Lee Hwee Kuan 21 Aug. 2013 For the purpose of learning stochastic simulations for the first time. We shall only consider probabilities on finite discrete
More informationNamed Entity Recognition using Maximum Entropy Model SEEM5680
Named Entity Recognition using Maximum Entroy Model SEEM5680 Named Entity Recognition System Named Entity Recognition (NER): Identifying certain hrases/word sequences in a free text. Generally it involves
More informationk- price auctions and Combination-auctions
k- rice auctions and Combination-auctions Martin Mihelich Yan Shu Walnut Algorithms March 6, 219 arxiv:181.3494v3 [q-fin.mf] 5 Mar 219 Abstract We rovide for the first time an exact analytical solution
More informationEinführung in Stochastische Prozesse und Zeitreihenanalyse Vorlesung, 2013S, 2.0h March 2015 Hubalek/Scherrer
Name: Mat.Nr.: Studium: Bitte keinen Rotstift verwenden! 15.593 Einführung in Stochastische Prozesse und Zeitreihenanalyse Vorlesung, 213S, 2.h March 215 Hubalek/Scherrer (Dauer 9 Minutes, Permissible
More information4. Score normalization technical details We now discuss the technical details of the score normalization method.
SMT SCORING SYSTEM This document describes the scoring system for the Stanford Math Tournament We begin by giving an overview of the changes to scoring and a non-technical descrition of the scoring rules
More informationDISCRIMINANTS IN TOWERS
DISCRIMINANTS IN TOWERS JOSEPH RABINOFF Let A be a Dedekind domain with fraction field F, let K/F be a finite searable extension field, and let B be the integral closure of A in K. In this note, we will
More informationChapter 1: PROBABILITY BASICS
Charles Boncelet, obability, Statistics, and Random Signals," Oxford University ess, 0. ISBN: 978-0-9-0005-0 Chater : PROBABILITY BASICS Sections. What Is obability?. Exeriments, Outcomes, and Events.
More informationMath Homework 5 Solutions
Math 45 - Homework 5 Solutions. Exercise.3., textbook. The stochastic matrix for the gambler problem has the following form, where the states are ordered as (,, 4, 6, 8, ): P = The corresponding diagram
More informationBOUNDS FOR THE COUPLING TIME IN QUEUEING NETWORKS PERFECT SIMULATION
BOUNDS FOR THE COUPLING TIME IN QUEUEING NETWORKS PERFECT SIMULATION JANTIEN G. DOPPER, BRUNO GAUJAL AND JEAN-MARC VINCENT Abstract. In this aer, the duration of erfect simulations for Markovian finite
More informationStatics and dynamics: some elementary concepts
1 Statics and dynamics: some elementary concets Dynamics is the study of the movement through time of variables such as heartbeat, temerature, secies oulation, voltage, roduction, emloyment, rices and
More informationTHE 2D CASE OF THE BOURGAIN-DEMETER-GUTH ARGUMENT
THE 2D CASE OF THE BOURGAIN-DEMETER-GUTH ARGUMENT ZANE LI Let e(z) := e 2πiz and for g : [0, ] C and J [0, ], define the extension oerator E J g(x) := g(t)e(tx + t 2 x 2 ) dt. J For a ositive weight ν
More informationMarkov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued
Markov Chains X(t) is a Markov Process if, for arbitrary times t 1 < t 2
More information4 Branching Processes
4 Branching Processes Organise by generations: Discrete time. If P(no offspring) 0 there is a probability that the process will die out. Let X = number of offspring of an individual p(x) = P(X = x) = offspring
More informationMarkov Chains (Part 4)
Markov Chains (Part 4) Steady State Probabilities and First Passage Times Markov Chains - 1 Steady-State Probabilities Remember, for the inventory example we had (8) P &.286 =.286.286 %.286 For an irreducible
More informationMartingale Theory for Finance
Martingale Theory for Finance Tusheng Zhang October 27, 2015 1 Introduction 2 Probability spaces and σ-fields 3 Integration with respect to a probability measure. 4 Conditional expectation. 5 Martingales.
More informationLecture 14: Introduction to Decision Making
Lecture 14: Introduction to Decision Making Preferences Utility functions Maximizing exected utility Value of information Actions and consequences So far, we have focused on ways of modeling a stochastic,
More informationPractice Final Solutions
Practice Final Solutions 1. True or false: (a) If a is a sum of three squares, and b is a sum of three squares, then so is ab. False: Consider a 14, b 2. (b) No number of the form 4 m (8n + 7) can be written
More informationThe Boundary Problem: Markov Chain Solution
MATH 529 The Boundary Problem: Markov Chain Solution Consider a random walk X that starts at positive height j, and on each independent step, moves upward a units with probability p, moves downward b units
More informationMATH 829: Introduction to Data Mining and Analysis Consistency of Linear Regression
1/9 MATH 829: Introduction to Data Mining and Analysis Consistency of Linear Regression Dominique Guillot Deartments of Mathematical Sciences University of Delaware February 15, 2016 Distribution of regression
More information