STOCHASTIC MODELS LECTURE 1 MARKOV CHAINS. Nan Chen MSc Program in Financial Engineering The Chinese University of Hong Kong (ShenZhen) Sept.
|
|
- Virgil Ambrose Poole
- 5 years ago
- Views:
Transcription
1 STOCHASTIC MODELS LECTURE 1 MARKOV CHAINS Nan Chen MSc Program in Financial Engineering The Chinese University of Hong Kong (ShenZhen) Sept. 6, 2016
2 Outline 1. Introduction 2. Chapman-Kolmogrov Equations 3. The Gambler s ruin Problem
3 1.1 INTRODUCTION
4 What is a Stochastic Process? A stochastic process is a collection of random variables that are indexed by time. Usually, we denote it by (X 1, X 2,!, X n,!) or (X t,t 0). Examples: Daily average temperature on CUHK-SZ campus Real-time stock price of Google
5 Motivation of Markov Chains Stochastic processes are widely used to characterize the temporal relationship between random variables. The simplest model should be that independent of each other. X n are But, the above model may not be able to provide a reasonable approximation to financial markets.
6 What is a Markov Chain? Let (X n, n = 0,1, 2,!) be a stochastic process that takes on a finite/countable number of possible states. We call it by a Markov chain, if the conditional distribution of X n+1 depends on the past observations (X 1, X 2,!, X n ) only through. Namely, X n P(X n+1 = j X 1 = i 1, X 2 = i 2,!, X n = i n ) = P(X n+1 = j X n = i n ) for all n.
7 The Markovian Property It can be shown that the definition of Markov chains is equivalent to stating that P(X n+1 = j, X 1 = i 1, X 2 = i 2,!, X n 1 = i n 1 X n = i n ) = P(X n+1 = j X n = i n )P(X 1 = i 1, X 2 = i 2,!, X n 1 = i n 1 X n = i n ) In words, given the current state of the process, its future and historical movements are independent.
8 Financial Rationale: Efficient Market Hypothesis The Markovian property turns out to be highly relevant to financial modeling in light of one of the most profound theory in the history of modern finance --- efficient market hypothesis. It states Market information, such as the information reflected in the past record or the information published in financial press, must be absorbed and reflected quickly in the stock price.
9 More about EMH: a Thought Experiment Let us start with the following thought experiment: Assume that Prof. Chen had invented an formula which we could use to predict the movements of Google stock price very accurately. What would happen if this formula was disclosed to the public?
10 More about EMH: a Thought Experiment Suppose that it predicted that Google s stock price would rise dramatically in three days to US$700 from US$650. The prediction would induce a great wave of immediate buy orders. Huge demands on Google s stocks would push its price to jump to $700 immediately. The formula fails! A true story of Edward Thorp and the Black- Scholes formula
11 Implication of Efficient Market Hypothesis One implication of EMH is that given the current stock price, knowing its history will help very little in predicting its future. Therefore, we should use Markov processes to model the dynamic of financial variables.
12 Transition Matrix In this lecture, we only consider timehomogenous Markov chains; that is, the transition probabilities P(X n+1 = j X n = i n ) are independent of time n. Denote p ij := P(X n+1 = j X n = i). We then can use the following matrix to characterize the process.! # # P := # # " p 11 p 12! p 1n p 21 p 22! p 2n!!!! p n1 p n2! p nn $ & & & & %
13 Transition Matrix The transition matrix of a Markov chain must be a stochastic matrix: p ij 0. n p ij =1. j=1
14 Example I: Forecasting the Weather Suppose that the chance of rain tomorrow in Shenzhen depends on previous weather conditions only through whether or not it is raining today. Assume that if it rains today, then it will rain tomorrow with probability 70%; and if it does not rain today, then it will rain tomorrow with prob. 50%. How can we use a Markov chain to model it?
15 Example II: 1-dimensional Random Walk A Markov chain whose state space is given by the integers 0,±1,±2,... is said to be a random walk if, for some number 0 < p <1, P i,i+1 = p =1 P i,i 1, i = 0,±1,±2,... We say the random walk is symmetric if p =1/ 2 ; asymmetric if p 1/ 2.
16 1.2 CHAPMAN- KOLMOGOROV EQUATIONS
17 The Chapman-Kolmogorov Equations The CK equations provide a method for computing the n -step transition probabilities of a Markov chain. P n+m ij = P( X n+m = j X 0 = i) = P n m ik P kj or in a matrix form, P n+m = P n P m. k
18 Example III: Rain Probability Reconsider the situation in Example I. Given that it is raining today, what is the probability that it will rain four days from today?
19 Example IV: Urn and Balls An urn always contains 2 balls. Ball colors are red and blue. At each stage a ball is randomly chosen and then replaced by a new ball, which with prob. 80% is the same color, and with prob. 20% is the opposite color. If initially both balls are red, find the probability that the fifth ball selected is red.
20 1.3 THE GAMBLER S RUIN PROBLEM
21 The Gambler s Ruin Problem Consider a gambler who at each play of the game has probability p of winning one dollar and probability q =1 p of losing one dollar. Assuming that successive plays of the game are independent, what is the probability that, starting with i dollars, the gambler fortune will win N dollars before he ruins (i.e., his fortune reaches 0)?
22 Markov Description of the Model If we let X n denote the player s fortune at time n, then the process {X n, n =1, 2,...} is a Markov chain with transition probabilities P 00 = P NN =1, P i,i+1 = p =1 P i,i 1 i =1, 2,..., N 1. The Markov chain has three classes: {0},{1, 2,..., N 1},{N}
23 Solution Let P i be the probability that, starting with dollars, the gambler fortune will eventually reach. By conditioning on the outcome of the initial play and N P i = pp i+1 + qp i 1 P 0 = 0, P N =1., i =1, 2,..., N 1. i
24 Solution (Continued) Hence, we obtain from the preceding slide that P 2 P 1 = q p (P 1 P 0 ) = q p P 1; P 3 P 2 = q p (P " 2 P 1 ) = q % $ ' # p & 2 P 1 ; P N P N 1 = q p (P " N 1 P N 2 ) = q % $ ' # p & N 1 P 1.
25 Solution (Continued) Adding all the equalities up, we obtain " $ P 1 = # $ % 1/ N, 1 (q / p) 1 (q / p) N, p =1/ 2; p 1/ 2. P i = " $ # $ % i / N, 1 (q / p) i 1 (q / p) N, p =1/ 2; p 1/ 2.
26 Solution (Continued) Note that, as N +, P i ) + * +, + 0, # 1 q & % ( $ p ' i, p 1/ 2; p >1/ 2. Thus, if p >1/ 2, there is a positive probability that the gambler s fortune will increase indefinitely; while if p 1/ 2, the gambler will, with probability 1, go ruin against an infinitely rich adversary (say, a casino).
27 Homework Assignments Read Ross Chapter 4.1, 4.2, and Answer Questions: Exercises 5, 6 (Page 261, Ross) Exercises 13, 14 (Page 262, Ross) Exercises 56, 57 (Page 270, Ross) Due on Sept. 13, Wed.
Markov Chains. Chapter 16. Markov Chains - 1
Markov Chains Chapter 16 Markov Chains - 1 Why Study Markov Chains? Decision Analysis focuses on decision making in the face of uncertainty about one future event. However, many decisions need to consider
More information2 DISCRETE-TIME MARKOV CHAINS
1 2 DISCRETE-TIME MARKOV CHAINS 21 FUNDAMENTAL DEFINITIONS AND PROPERTIES From now on we will consider processes with a countable or finite state space S {0, 1, 2, } Definition 1 A discrete-time discrete-state
More informationProbability, Random Processes and Inference
INSTITUTO POLITÉCNICO NACIONAL CENTRO DE INVESTIGACION EN COMPUTACION Laboratorio de Ciberseguridad Probability, Random Processes and Inference Dr. Ponciano Jorge Escamilla Ambrosio pescamilla@cic.ipn.mx
More informationMarkov Chains (Part 3)
Markov Chains (Part 3) State Classification Markov Chains - State Classification Accessibility State j is accessible from state i if p ij (n) > for some n>=, meaning that starting at state i, there is
More informationISM206 Lecture, May 12, 2005 Markov Chain
ISM206 Lecture, May 12, 2005 Markov Chain Instructor: Kevin Ross Scribe: Pritam Roy May 26, 2005 1 Outline of topics for the 10 AM lecture The topics are: Discrete Time Markov Chain Examples Chapman-Kolmogorov
More informationChapter 16 focused on decision making in the face of uncertainty about one future
9 C H A P T E R Markov Chains Chapter 6 focused on decision making in the face of uncertainty about one future event (learning the true state of nature). However, some decisions need to take into account
More informationISE/OR 760 Applied Stochastic Modeling
ISE/OR 760 Applied Stochastic Modeling Topic 2: Discrete Time Markov Chain Yunan Liu Department of Industrial and Systems Engineering NC State University Yunan Liu (NC State University) ISE/OR 760 1 /
More informationMarkov Processes Hamid R. Rabiee
Markov Processes Hamid R. Rabiee Overview Markov Property Markov Chains Definition Stationary Property Paths in Markov Chains Classification of States Steady States in MCs. 2 Markov Property A discrete
More informationStochastic processes. MAS275 Probability Modelling. Introduction and Markov chains. Continuous time. Markov property
Chapter 1: and Markov chains Stochastic processes We study stochastic processes, which are families of random variables describing the evolution of a quantity with time. In some situations, we can treat
More information1 Gambler s Ruin Problem
1 Gambler s Ruin Problem Consider a gambler who starts with an initial fortune of $1 and then on each successive gamble either wins $1 or loses $1 independent of the past with probabilities p and q = 1
More information1 Gambler s Ruin Problem
Coyright c 2017 by Karl Sigman 1 Gambler s Ruin Problem Let N 2 be an integer and let 1 i N 1. Consider a gambler who starts with an initial fortune of $i and then on each successive gamble either wins
More informationMarkov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains
Markov Chains A random process X is a family {X t : t T } of random variables indexed by some set T. When T = {0, 1, 2,... } one speaks about a discrete-time process, for T = R or T = [0, ) one has a continuous-time
More informationLecture 4 - Random walk, ruin problems and random processes
Lecture 4 - Random walk, ruin problems and random processes Jan Bouda FI MU April 19, 2009 Jan Bouda (FI MU) Lecture 4 - Random walk, ruin problems and random processesapril 19, 2009 1 / 30 Part I Random
More informationGambler s Ruin with Catastrophes and Windfalls
Journal of Grace Scientific Publishing Statistical Theory and Practice Volume 2, No2, June 2008 Gambler s Ruin with Catastrophes and Windfalls B unter, Department of Mathematics, University of California,
More informationMATH HOMEWORK PROBLEMS D. MCCLENDON
MATH 46- HOMEWORK PROBLEMS D. MCCLENDON. Consider a Markov chain with state space S = {0, }, where p = P (0, ) and q = P (, 0); compute the following in terms of p and q: (a) P (X 2 = 0 X = ) (b) P (X
More informationLesson Plan. AM 121: Introduction to Optimization Models and Methods. Lecture 17: Markov Chains. Yiling Chen SEAS. Stochastic process Markov Chains
AM : Introduction to Optimization Models and Methods Lecture 7: Markov Chains Yiling Chen SEAS Lesson Plan Stochastic process Markov Chains n-step probabilities Communicating states, irreducibility Recurrent
More informationChapter 11 Advanced Topic Stochastic Processes
Chapter 11 Advanced Topic Stochastic Processes CHAPTER OUTLINE Section 1 Simple Random Walk Section 2 Markov Chains Section 3 Markov Chain Monte Carlo Section 4 Martingales Section 5 Brownian Motion Section
More informationUncertainty Runs Rampant in the Universe C. Ebeling circa Markov Chains. A Stochastic Process. Into each life a little uncertainty must fall.
Uncertainty Runs Rampant in the Universe C. Ebeling circa 2000 Markov Chains A Stochastic Process Into each life a little uncertainty must fall. Our Hero - Andrei Andreyevich Markov Born: 14 June 1856
More informationLecture 4 An Introduction to Stochastic Processes
Lecture 4 An Introduction to Stochastic Processes Prof. Massimo Guidolin Prep Course in Quantitative Methods for Finance August-September 2017 Plan of the lecture Motivation and definitions Filtrations
More informationRandom Walk and Other Lattice Models
Random Walk and Other Lattice Models Christian Beneš Brooklyn College Math Club Brooklyn College Math Club 04-23-2013 (Brooklyn College Math Club) 04-23-2013 1 / 28 Outline 1 Lattices 2 Random Walk 3 Percolation
More informationLecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes
Lecture Notes 7 Random Processes Definition IID Processes Bernoulli Process Binomial Counting Process Interarrival Time Process Markov Processes Markov Chains Classification of States Steady State Probabilities
More informationThe Boundary Problem: Markov Chain Solution
MATH 529 The Boundary Problem: Markov Chain Solution Consider a random walk X that starts at positive height j, and on each independent step, moves upward a units with probability p, moves downward b units
More informationReadings: Finish Section 5.2
LECTURE 19 Readings: Finish Section 5.2 Lecture outline Markov Processes I Checkout counter example. Markov process: definition. -step transition probabilities. Classification of states. Example: Checkout
More information33 The Gambler's Ruin 1
33 The Gambler's Ruin 1 Figure 1 The Gambler's Ruin There are many variations of the gambler's ruin problem, but a principal one goes like this. We have i dollars out of a total of n. We flip a coin which
More informationMathematical Games and Random Walks
Mathematical Games and Random Walks Alexander Engau, Ph.D. Department of Mathematical and Statistical Sciences College of Liberal Arts and Sciences / Intl. College at Beijing University of Colorado Denver
More information18.175: Lecture 30 Markov chains
18.175: Lecture 30 Markov chains Scott Sheffield MIT Outline Review what you know about finite state Markov chains Finite state ergodicity and stationarity More general setup Outline Review what you know
More informationModule 6:Random walks and related areas Lecture 24:Random woalk and other areas. The Lecture Contains: Random Walk.
The Lecture Contains: Random Walk Ehrenfest Model file:///e /courses/introduction_stochastic_process_application/lecture24/24_1.htm[9/30/2013 1:03:45 PM] Random Walk As already discussed there is another
More informationLecture 20 : Markov Chains
CSCI 3560 Probability and Computing Instructor: Bogdan Chlebus Lecture 0 : Markov Chains We consider stochastic processes. A process represents a system that evolves through incremental changes called
More informationChapter 35 out of 37 from Discrete Mathematics for Neophytes: Number Theory, Probability, Algorithms, and Other Stuff by J. M. Cargal.
35 Mixed Chains In this chapter we learn how to analyze Markov chains that consists of transient and absorbing states. Later we will see that this analysis extends easily to chains with (nonabsorbing)
More informationIntroduction to Stochastic Processes
18.445 Introduction to Stochastic Processes Lecture 1: Introduction to finite Markov chains Hao Wu MIT 04 February 2015 Hao Wu (MIT) 18.445 04 February 2015 1 / 15 Course description About this course
More informationWe Live in Exciting Times. CSCI-567: Machine Learning (Spring 2019) Outline. Outline. ACM (an international computing research society) has named
We Live in Exciting Times ACM (an international computing research society) has named CSCI-567: Machine Learning (Spring 2019) Prof. Victor Adamchik U of Southern California Apr. 2, 2019 Yoshua Bengio,
More informationMarkov Chains Handout for Stat 110
Markov Chains Handout for Stat 0 Prof. Joe Blitzstein (Harvard Statistics Department) Introduction Markov chains were first introduced in 906 by Andrey Markov, with the goal of showing that the Law of
More informationGeneral Boundary Problem: System of Equations Solution
MATH 54 General Boundary Problem: System of Equations Solution Consider again the scenario of a random walk X that starts at positive height j, and on each independent step, moves upward a units with probability
More information1/2 1/2 1/4 1/4 8 1/2 1/2 1/2 1/2 8 1/2 6 P =
/ 7 8 / / / /4 4 5 / /4 / 8 / 6 P = 0 0 0 0 0 0 0 0 0 0 4 0 0 0 0 0 0 0 0 4 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 Andrei Andreevich Markov (856 9) In Example. 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 P (n) = 0
More informationMarkov Chains on Countable State Space
Markov Chains on Countable State Space 1 Markov Chains Introduction 1. Consider a discrete time Markov chain {X i, i = 1, 2,...} that takes values on a countable (finite or infinite) set S = {x 1, x 2,...},
More informationMarkov Chains Introduction
Markov Chains 4 4.1. Introduction In this chapter, we consider a stochastic process {X n,n= 0, 1, 2,...} that takes on a finite or countable number of possible values. Unless otherwise mentioned, this
More informationLecture 9 Classification of States
Lecture 9: Classification of States of 27 Course: M32K Intro to Stochastic Processes Term: Fall 204 Instructor: Gordan Zitkovic Lecture 9 Classification of States There will be a lot of definitions and
More informationDiscrete time Markov chains. Discrete Time Markov Chains, Definition and classification. Probability axioms and first results
Discrete time Markov chains Discrete Time Markov Chains, Definition and classification 1 1 Applied Mathematics and Computer Science 02407 Stochastic Processes 1, September 5 2017 Today: Short recap of
More informationChapter 10. Finite-State Markov Chains. Introductory Example: Googling Markov Chains
Chapter 0 Finite-State Markov Chains Introductory Example: Googling Markov Chains Google means many things: it is an Internet search engine, the company that produces the search engine, and a verb meaning
More informationLTCC. Exercises. (1) Two possible weather conditions on any day: {rainy, sunny} (2) Tomorrow s weather depends only on today s weather
1. Markov chain LTCC. Exercises Let X 0, X 1, X 2,... be a Markov chain with state space {1, 2, 3, 4} and transition matrix 1/2 1/2 0 0 P = 0 1/2 1/3 1/6. 0 0 0 1 (a) What happens if the chain starts in
More informationLectures on Probability and Statistical Models
Lectures on Probability and Statistical Models Phil Pollett Professor of Mathematics The University of Queensland c These materials can be used for any educational purpose provided they are are not altered
More informationCS 188: Artificial Intelligence Spring 2009
CS 188: Artificial Intelligence Spring 2009 Lecture 21: Hidden Markov Models 4/7/2009 John DeNero UC Berkeley Slides adapted from Dan Klein Announcements Written 3 deadline extended! Posted last Friday
More informationMarkov chain Monte Carlo
Markov chain Monte Carlo Feng Li feng.li@cufe.edu.cn School of Statistics and Mathematics Central University of Finance and Economics Revised on April 24, 2017 Today we are going to learn... 1 Markov Chains
More informationCOMS 4721: Machine Learning for Data Science Lecture 20, 4/11/2017
COMS 4721: Machine Learning for Data Science Lecture 20, 4/11/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University SEQUENTIAL DATA So far, when thinking
More informationMarkov Chains. Chapter Introduction. Specifying a Markov Chain
Chapter 11 Markov Chains 11.1 Introduction Most of our study of probability has dealt with independent trials processes. These processes are the basis of classical probability theory and much of statistics.
More informationDirect Proof Universal Statements
Direct Proof Universal Statements Lecture 13 Section 4.1 Robb T. Koether Hampden-Sydney College Wed, Feb 6, 2013 Robb T. Koether (Hampden-Sydney College) Direct Proof Universal Statements Wed, Feb 6, 2013
More informationThe probability of going from one state to another state on the next trial depends only on the present experiment and not on past history.
c Dr Oksana Shatalov, Fall 2010 1 9.1: Markov Chains DEFINITION 1. Markov process, or Markov Chain, is an experiment consisting of a finite number of stages in which the outcomes and associated probabilities
More informationRandom walk on a polygon
IMS Lecture Notes Monograph Series Recent Developments in Nonparametric Inference and Probability Vol. 50 (2006) 4 c Institute of Mathematical Statistics, 2006 DOI: 0.24/0749270600000058 Random walk on
More informationTreball final de grau GRAU DE MATEMÀTIQUES Facultat de Matemàtiques Universitat de Barcelona MARKOV CHAINS
Treball final de grau GRAU DE MATEMÀTIQUES Facultat de Matemàtiques Universitat de Barcelona MARKOV CHAINS Autor: Anna Areny Satorra Director: Dr. David Márquez Carreras Realitzat a: Departament de probabilitat,
More information8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains
8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains 8.1 Review 8.2 Statistical Equilibrium 8.3 Two-State Markov Chain 8.4 Existence of P ( ) 8.5 Classification of States
More informationLecture: Conditional Expectation
Lecture: Conditional Expectation Lutz Kruschwitz & Andreas Löffler Discounted Cash Flow, Section 1.2, Outline 1.2 Conditional expectation 1.2.1 Uncertainty and information 1.2.2 Rules 1.2.3 Application
More informationMathematical Statistics
Mathematical Statistics Kenichi ISHIKAWA http://ishiken.free.fr/lecture.html Oct. 18 Combination and Probability Oct. 25 Random variables and probability distributions Nov. 1 Representative probability
More informationIntroduction to Game Theory: Simple Decisions Models
Introduction to Game Theory: Simple Decisions Models John C.S. Lui Department of Computer Science & Engineering The Chinese University of Hong Kong John C.S. Lui (CUHK) Advanced Topics in Network Analysis
More informationISyE 6650 Probabilistic Models Fall 2007
ISyE 6650 Probabilistic Models Fall 2007 Homework 4 Solution 1. (Ross 4.3) In this case, the state of the system is determined by the weather conditions in the last three days. Letting D indicate a dry
More informationThe Markov Decision Process (MDP) model
Decision Making in Robots and Autonomous Agents The Markov Decision Process (MDP) model Subramanian Ramamoorthy School of Informatics 25 January, 2013 In the MAB Model We were in a single casino and the
More informationFINITE MARKOV CHAINS
Treball final de grau GRAU DE MATEMÀTIQUES Facultat de Matemàtiques Universitat de Barcelona FINITE MARKOV CHAINS Lidia Pinilla Peralta Director: Realitzat a: David Márquez-Carreras Departament de Probabilitat,
More informationProbability Models. 4. What is the definition of the expectation of a discrete random variable?
1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions
More informationReinforcement Learning
Reinforcement Learning March May, 2013 Schedule Update Introduction 03/13/2015 (10:15-12:15) Sala conferenze MDPs 03/18/2015 (10:15-12:15) Sala conferenze Solving MDPs 03/20/2015 (10:15-12:15) Aula Alpha
More informationT 1. The value function v(x) is the expected net gain when using the optimal stopping time starting at state x:
108 OPTIMAL STOPPING TIME 4.4. Cost functions. The cost function g(x) gives the price you must pay to continue from state x. If T is your stopping time then X T is your stopping state and f(x T ) is your
More informationHidden Markov Models (HMM) and Support Vector Machine (SVM)
Hidden Markov Models (HMM) and Support Vector Machine (SVM) Professor Joongheon Kim School of Computer Science and Engineering, Chung-Ang University, Seoul, Republic of Korea 1 Hidden Markov Models (HMM)
More informationLecture: Conditional Expectation
Discounted Cash Flow, Section 1.2 Outline 1.2 Conditional Expectation Uncertainty 1 Uncertainty is a distinguished feature of valuation usually modelled as different future states of nature ω with corresponding
More information1 Random Walks and Electrical Networks
CME 305: Discrete Mathematics and Algorithms Random Walks and Electrical Networks Random walks are widely used tools in algorithm design and probabilistic analysis and they have numerous applications.
More informationDoing Physics with Random Numbers
Doing Physics with Random Numbers Andrew J. Schultz Department of Chemical and Biological Engineering University at Buffalo The State University of New York Concepts Random numbers can be used to measure
More information9) Time series econometrics
30C00200 Econometrics 9) Time series econometrics Timo Kuosmanen Professor Management Science http://nomepre.net/index.php/timokuosmanen 1 Macroeconomic data: GDP Inflation rate Examples of time series
More informationLecture 5: Introduction to Markov Chains
Lecture 5: Introduction to Markov Chains Winfried Just Department of Mathematics, Ohio University January 24 26, 2018 weather.com light The weather is a stochastic process. For now we can assume that this
More informationMarkov Model. Model representing the different resident states of a system, and the transitions between the different states
Markov Model Model representing the different resident states of a system, and the transitions between the different states (applicable to repairable, as well as non-repairable systems) System behavior
More information12 Markov chains The Markov property
12 Markov chains Summary. The chapter begins with an introduction to discrete-time Markov chains, and to the use of matrix products and linear algebra in their study. The concepts of recurrence and transience
More informationNote that in the example in Lecture 1, the state Home is recurrent (and even absorbing), but all other states are transient. f ii (n) f ii = n=1 < +
Random Walks: WEEK 2 Recurrence and transience Consider the event {X n = i for some n > 0} by which we mean {X = i}or{x 2 = i,x i}or{x 3 = i,x 2 i,x i},. Definition.. A state i S is recurrent if P(X n
More informationDynamic Risk Measures and Nonlinear Expectations with Markov Chain noise
Dynamic Risk Measures and Nonlinear Expectations with Markov Chain noise Robert J. Elliott 1 Samuel N. Cohen 2 1 Department of Commerce, University of South Australia 2 Mathematical Insitute, University
More informationM3/4/5 S4 Applied Probability
M3/4/5 S4 Applied Probability Autumn 215 Badr Missaoui Room 545 Huxley Building Imperial College London E-Mail: badr.missaoui8@imperial.ac.uk Department of Mathematics Imperial College London 18 Queens
More informationLecture 6 Sept. 14, 2015
PHYS 7895: Quantum Information Theory Fall 205 Prof. Mark M. Wilde Lecture 6 Sept., 205 Scribe: Mark M. Wilde This document is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 3.0
More information18.600: Lecture 32 Markov Chains
18.600: Lecture 32 Markov Chains Scott Sheffield MIT Outline Markov chains Examples Ergodicity and stationarity Outline Markov chains Examples Ergodicity and stationarity Markov chains Consider a sequence
More information1 Presessional Probability
1 Presessional Probability Probability theory is essential for the development of mathematical models in finance, because of the randomness nature of price fluctuations in the markets. This presessional
More informationSpecial Mathematics. Tutorial 13. Markov chains
Tutorial 13 Markov chains "The future starts today, not tomorrow." Pope John Paul II A sequence of trials of an experiment is a nite Markov chain if: the outcome of each experiment is one of a nite set
More informationContinuous Time Markov Chains
Continuous Time Markov Chains Stochastic Processes - Lecture Notes Fatih Cavdur to accompany Introduction to Probability Models by Sheldon M. Ross Fall 2015 Outline Introduction Continuous-Time Markov
More informationINTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING
INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING ERIC SHANG Abstract. This paper provides an introduction to Markov chains and their basic classifications and interesting properties. After establishing
More informationMath-Stat-491-Fall2014-Notes-V
Math-Stat-491-Fall2014-Notes-V Hariharan Narayanan November 18, 2014 Martingales 1 Introduction Martingales were originally introduced into probability theory as a model for fair betting games. Essentially
More informationUndergraduate Research and Creative Practice
Grand Valley State University ScholarWorks@GVSU Honors Projects Undergraduate Research and Creative Practice 2014 The Application of Stochastic Processes in the Sciences: Using Markov Chains to Estimate
More informationExpectations, Markov chains, and the Metropolis algorithm
Expectations, Markov chains, and the Metropolis algorithm Peter Hoff Departments of Statistics and Biostatistics and the Center for Statistics and the Social Sciences University of Washington 7-27-05 1
More informationMarkov Chains (Part 4)
Markov Chains (Part 4) Steady State Probabilities and First Passage Times Markov Chains - 1 Steady-State Probabilities Remember, for the inventory example we had (8) P &.286 =.286.286 %.286 For an irreducible
More informationDiscrete Markov Chain. Theory and use
Discrete Markov Chain. Theory and use Andres Vallone PhD Student andres.vallone@predoc.uam.es 2016 Contents 1 Introduction 2 Concept and definition Examples Transitions Matrix Chains Classification 3 Empirical
More informationSome Definition and Example of Markov Chain
Some Definition and Example of Markov Chain Bowen Dai The Ohio State University April 5 th 2016 Introduction Definition and Notation Simple example of Markov Chain Aim Have some taste of Markov Chain and
More informationMarkov Chains. As part of Interdisciplinary Mathematical Modeling, By Warren Weckesser Copyright c 2006.
Markov Chains As part of Interdisciplinary Mathematical Modeling, By Warren Weckesser Copyright c 2006 1 Introduction A (finite) Markov chain is a process with a finite number of states (or outcomes, or
More informationMath Homework 5 Solutions
Math 45 - Homework 5 Solutions. Exercise.3., textbook. The stochastic matrix for the gambler problem has the following form, where the states are ordered as (,, 4, 6, 8, ): P = The corresponding diagram
More informationu( x)= Pr X() t hits C before 0 X( 0)= x ( ) 2 AMS 216 Stochastic Differential Equations Lecture #2
AMS 6 Stochastic Differential Equations Lecture # Gambler s Ruin (continued) Question #: How long can you play? Question #: What is the chance that you break the bank? Note that unlike in the case of deterministic
More informationMARKOV CHAINS. Discussion attached. Source:
LAVC - Math 385 Spring 2015 Shin/Dobson/Caleodis MARKOV CHAINS 1) A species is endangered. The survival/reproductive rates of females are dependent upon age. Is it possible to find an age distribution
More informationAPPM 4/5560 Markov Processes Fall 2019, Some Review Problems for Exam One
APPM /556 Markov Processes Fall 9, Some Review Problems for Exam One (Note: You will not have to invert any matrices on the exam and you will not be expected to solve systems of or more unknowns. There
More informationStochastic process. X, a series of random variables indexed by t
Stochastic process X, a series of random variables indexed by t X={X(t), t 0} is a continuous time stochastic process X={X(t), t=0,1, } is a discrete time stochastic process X(t) is the state at time t,
More informationMATH 564/STAT 555 Applied Stochastic Processes Homework 2, September 18, 2015 Due September 30, 2015
ID NAME SCORE MATH 56/STAT 555 Applied Stochastic Processes Homework 2, September 8, 205 Due September 30, 205 The generating function of a sequence a n n 0 is defined as As : a ns n for all s 0 for which
More informationHomework 2. Spring 2019 (Due Thursday February 7)
ECE 302: Probabilistic Methods in Electrical and Computer Engineering Spring 2019 Instructor: Prof. A. R. Reibman Homework 2 Spring 2019 (Due Thursday February 7) Homework is due on Thursday February 7
More informationA FUZZY TIME SERIES-MARKOV CHAIN MODEL WITH AN APPLICATION TO FORECAST THE EXCHANGE RATE BETWEEN THE TAIWAN AND US DOLLAR.
International Journal of Innovative Computing, Information and Control ICIC International c 2012 ISSN 1349-4198 Volume 8, Number 7(B), July 2012 pp. 4931 4942 A FUZZY TIME SERIES-MARKOV CHAIN MODEL WITH
More informationTMA4265 Stochastic processes ST2101 Stochastic simulation and modelling
Norwegian University of Science and Technology Department of Mathematical Sciences Page of 7 English Contact during examination: Øyvind Bakke Telephone: 73 9 8 26, 99 4 673 TMA426 Stochastic processes
More informationSTOCHASTIC PROCESSES Basic notions
J. Virtamo 38.3143 Queueing Theory / Stochastic processes 1 STOCHASTIC PROCESSES Basic notions Often the systems we consider evolve in time and we are interested in their dynamic behaviour, usually involving
More informationErrata List for the Second Printing
Errata List for the Second Printing The Second Revised Edition has had two printings, and it also exists on the web. The second printing can be distinguished from the first printing by the fact that on
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning Mark Schmidt University of British Columbia Winter 2019 Last Time: Monte Carlo Methods If we want to approximate expectations of random functions, E[g(x)] = g(x)p(x) or E[g(x)]
More informationAlgorithmic Game Theory and Applications. Lecture 4: 2-player zero-sum games, and the Minimax Theorem
Algorithmic Game Theory and Applications Lecture 4: 2-player zero-sum games, and the Minimax Theorem Kousha Etessami 2-person zero-sum games A finite 2-person zero-sum (2p-zs) strategic game Γ, is a strategic
More informationMarkov Chains. Sarah Filippi Department of Statistics TA: Luke Kelly
Markov Chains Sarah Filippi Department of Statistics http://www.stats.ox.ac.uk/~filippi TA: Luke Kelly With grateful acknowledgements to Prof. Yee Whye Teh's slides from 2013 14. Schedule 09:30-10:30 Lecture:
More informationDefinition and Examples of DTMCs
Definition and Examples of DTMCs Natarajan Gautam Department of Industrial and Systems Engineering Texas A&M University 235A Zachry, College Station, TX 77843-3131 Email: gautam@tamuedu Phone: 979-845-5458
More informationMarkov Chains, Stochastic Processes, and Matrix Decompositions
Markov Chains, Stochastic Processes, and Matrix Decompositions 5 May 2014 Outline 1 Markov Chains Outline 1 Markov Chains 2 Introduction Perron-Frobenius Matrix Decompositions and Markov Chains Spectral
More informationSlides 5: Reasoning with Random Numbers
Slides 5: Reasoning with Random Numbers Realistic Simulating What is convincing? Fast and efficient? Solving Problems Simulation Approach Using Thought Experiment or Theory Formal Language 1 Brain Teasers
More information