1 if the i-th toss is a Head (with probability p) ξ i = 0 if the i-th toss is a Tail (with probability 1 p)
|
|
- Elisabeth Gallagher
- 5 years ago
- Views:
Transcription
1 6 Chapter 3. Markov Chain: Introduction Whatever happened in the past, be it glory or misery, be Markov! 3.1. Examples Example 3.1. (Coin Tossing.) Let ξ 0 0 and, for i 1, 1 if the i-th toss is a Head (with probability p) ξ i 0 if the i-th toss is a Tail (with probability 1 p) Set X n n i0 ξ i, n 0. X n is the random number which counts the number of Heads up to the n-th toss. Then, P(X n+1 j X 0 0, X 1 i 1,..., X n 1 i n 1, X n i) P(ξn+1 1 X 0 0, X 1 i 1,..., X n 1 i n 1, X n i) if j i + 1 P(ξ n+1 0 X 0 0, X 1 i 1,..., X n 1 i n 1, X n i) if j i P(ξn+1 1) if j i + 1 P(ξ n+1 0) if j i p if j i p if j i P(ξn+1 1 X n i) if j i + 1 P(ξ n+1 0 X n i) if j i P(X n+1 j X n i) It implies, once the present X n is fixed, the past history X 0,..., X n 1, shall not affect the future distribution of X n+1. Here, we are using time n as present, time before n is past, and time beyond n is future. The above statement is the same as saying that the future X k : k n + 1} and the past X k : k n 1} are conditionally independent given the present X n (taking any fixed value.) Remark. Not only n + 1 but, all future distribution X n+1, X n+2,... shall depend on the past and now only through now. Example 3.2. (Mickey in Maze) Mickey mouse travels in a maze with nine 3 3 cells. The cells are numbered as 0, 1,..., 8 from left to right and top down. Each step Mickey travels from where it is to one of the surrounding connected cells with equal chance
2 7 Let X n denote the cell number of Mickey at step n. (X 0 4). Then, P(X n+1 j X 0 0, X 1 i 1,..., X n 1 i n 1, X n i) P(X n+1 j X n i). Suppose currently Mickey is in cell 5, for example, the future movement or path of Mickey is irrelevant with the past movement or path of Mickey. In other words, how Mickey has got to cell 5 in the past has nothing to do with how Mickey would move around in the future. The process X n : n 0, 1, 2,...} is a Markov chain. Example for fun ( First Blood. ) John Rambo only obeys the order from Colonel Samuel Trautman, who supposedly only obeys the order from the Pentagon. Then Pentagon Trautman Rambo forms a MC Definitions/Descriptions 1. Stochastic Process: a family of random variables X t } index by t. 2. State space: the set of values of the stochastic processes, which in general does not have to be real numbers. 3. Markov Process: A stochastic process X t } indexed by time t such that, at each time t, the future of the process X s : s > t} is conditionally independent of the past of the process X s : s < t} given the present of the process X t taking any fixed value. Another interpretation is: at each time t, the future of the process X s : s > t} depends on the the past of the process X s : s < t} only through the present X t. Caution: X s : s > t is in general not independent of X s : s < t. 4. Markov chain (M C): Markov process with discrete state space. Discrete state space is usually denoted by numbers 0, 1, 2, Discrete/continuous time Markov chain: Discrete time M C: the time domain is 0, 1, 2,...}. Continuous time MC: the time domain is [0, ). Examples 3.1 and 3.2 are discrete time Markov chains. Poisson process is continuous time M C. Brownian motion is a continuous time Markov process (but not MC). For a discrete time MC: X n : n 0, 1,...}: the defining equation is, for any n 0, P(X n+1 j X 0 0, X 1 i 1,..., X n 1 i n 1, X n i) P(X n+1 j X n i) Transition probabilities. One-step transition probability: P(X n+1 j X n i). If one-step transition probability is irrelevant with n, i.e., it s the same for all n, we call the MC X t } a MC with stationary transition probabilities. Throughout the course, we only consider MC with stationary transition probabilities. Let P P(X n+1 j X n i). The one-step transition probability matrix, or transition matrix in brief, is P P 00 P 01 P P 10 P 11 P P 20 P 21 P
3 8 For the Mickey in Maze (Example 3.2) example: P Denote the k-step transition probability as P (k) probability matrix as ( 1/4 1/4 1/4 1/4 P (k) P(X n+k j X n i) and the k-step transition P (k) For notational convenience, we always let P (1) P, P (0) 1 if i j 0 if i j. Then, Theorem 3.1. For all 0 m n and n 0, Proof. Write P (n) Observe that P (n) P (n) P (n) P n. ) P (m) il P (n m) lj P(X n j X 0 i) P(X n j, X 0 i)/p(x 0 i) P(X n j, X m l, X 0 i)/p(x 0 i) P(X n j X m l, X 0 i)p(x m l, X 0 i)/p(x 0 i) P(X n j X m l, X 0 i)p(x m l X 0 i) P(X n j X m l)p(x m l, X 0 i)/p(x 0 i) P (n m) lj P (m) il. is the (i + 1, j + 1)-th entry of the matrix P (n), P (m) il, l 0, 1, 2,... are the, l 0, 1, 2,... are the (j +1)-th column of the matrix (i+1)-th row of the matrix P (m), and P (n m) lj P (n m). Hence, P (n) P (m) P (n m)
4 9 for all 0 m n. Since P (1) P, we have P (2) PP P 2, P (3) PP (2) P 3,... P (n) P n by induction. Example 3.3 (Michey in Maze, Example 3.2 continued) Compute P (3) 48 Solution. Write P (3) P (3) k0 P 4 k P (2) k 8 k1,3,5,7 P 4 k P (2) k 8 1/4 k1,3,5,7 P (2) k 8 0; 1/3 1/2 1/3 + 1/3 1/4 1/3 + 1/3 1/4 1/3 0. and P (3) More Examples. Example 3.4. (An Inventory Model) Let X n be the number of TV sets at a store in the end of day n with X 0 2. Let ξ n be the sales of the number of TVs on day n. Assume ξ 1, ξ 2,... are iid (independent, identically distributed) such that 0.5 i 0 P(ξ n i) 0.4 i i 2 At the end of any day n, if X n 0 or 1, two TVs will be sent to the store overnight. Moreover, in the case of X n 1, another TV will be sent directly to the customer s house. If X n 1 or 2, nothing happens. With this inventory policy, Xn ξ X n+1 n+1 if X n 1, 2 2 ξ n+1 if X n 1, 0. Then, X 0, X 1, X 2,..., is a MC with state space 1, 0, 1, 2} and with one-step transition probability P Example 3.5. (The Ehrenfest Model) There are 2N particles in a jar separated by a membrane into two chambers A and B. Let Y n be the number of particles in A after n crossings. Each crossing is a particle from A to B or from B to A. Assume when any once crossing happens, it happens to any one of the 2N particles with each equal chance 1/2N. A B
5 10 Then Y n, n 0 is a MC with state space 0, 1, 2,..., 2N} and with one step transition probability: i/(2n) if j i 1 P P(Y n+1 j Y n i) 1 i/(2n) if j i else; for 0 i j 2N.
6 11 More DIY Exercises: Exercise 3.1 Bunny rabbit has three dens A, B and C. It likes A better than B and C. If it s in B or C on any night, it will always take chance 0.9 to go to A and chance 0.1 to go to the other den for the following night. Once it reaches A, it will stay there for two nights and the third night will be in B or C with equal chance 1/2. Let X n be the den Bunny stays for night n. What is the state space of the X n }? Is X n } a MC? Den C Den A Den B Exercise 3.2 For three events A, B and C, the following three statements are equivalent: (i) P(A B C) P(A C)P(B C); (ii). P(A C B) P(A C); and (iii) P(B A C) P(B C). (Assume all quantities here are well defined.) Notice that statement (i) says that A and B are conditionally independent given C. Exercise 3.3 Suppose X n, n 0, 1, 2,... is a discrete time MC. Let 0 n 0 < n 1 < n 2 <... be a subsequence of the nonnegative integers and Y k X nk. Is Y k : k 0, 1, 2,..., } a MC? Exercise 3.4 Suppose X n : n 2, 1, 0, 1, 2, } is a discrete time MC with the time being from all integers from to. Let Y n X n for all integers n. Is Y n : n 2, 1, 0, 1, 2, } a MC?
Math 180B Problem Set 3
Math 180B Problem Set 3 Problem 1. (Exercise 3.1.2) Solution. By the definition of conditional probabilities we have Pr{X 2 = 1, X 3 = 1 X 1 = 0} = Pr{X 3 = 1 X 2 = 1, X 1 = 0} Pr{X 2 = 1 X 1 = 0} = P
More information3.5. First step analysis
12 3.5. First step analysis A journey of a thousand miles begins with the first step. Lao Tse Example 3.6. (Coin Tossing) Repeatedly toss a fair coin a number of times. What s the expected number of tosses
More informationMarkov Chains. Chapter 16. Markov Chains - 1
Markov Chains Chapter 16 Markov Chains - 1 Why Study Markov Chains? Decision Analysis focuses on decision making in the face of uncertainty about one future event. However, many decisions need to consider
More informationMarkov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains
Markov Chains A random process X is a family {X t : t T } of random variables indexed by some set T. When T = {0, 1, 2,... } one speaks about a discrete-time process, for T = R or T = [0, ) one has a continuous-time
More informationIEOR 6711: Stochastic Models I. Solutions to Homework Assignment 9
IEOR 67: Stochastic Models I Solutions to Homework Assignment 9 Problem 4. Let D n be the random demand of time period n. Clearly D n is i.i.d. and independent of all X k for k < n. Then we can represent
More informationChapter 11 Advanced Topic Stochastic Processes
Chapter 11 Advanced Topic Stochastic Processes CHAPTER OUTLINE Section 1 Simple Random Walk Section 2 Markov Chains Section 3 Markov Chain Monte Carlo Section 4 Martingales Section 5 Brownian Motion Section
More informationMath 180B Homework 4 Solutions
Math 80B Homework 4 Solutions Note: We will make repeated use of the following result. Lemma. Let (X n ) be a time-homogeneous Markov chain with countable state space S, let A S, and let T = inf { n 0
More informationIEOR 6711: Professor Whitt. Introduction to Markov Chains
IEOR 6711: Professor Whitt Introduction to Markov Chains 1. Markov Mouse: The Closed Maze We start by considering how to model a mouse moving around in a maze. The maze is a closed space containing nine
More informationP i [B k ] = lim. n=1 p(n) ii <. n=1. V i :=
2.7. Recurrence and transience Consider a Markov chain {X n : n N 0 } on state space E with transition matrix P. Definition 2.7.1. A state i E is called recurrent if P i [X n = i for infinitely many n]
More informationLecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321
Lecture 11: Introduction to Markov Chains Copyright G. Caire (Sample Lectures) 321 Discrete-time random processes A sequence of RVs indexed by a variable n 2 {0, 1, 2,...} forms a discretetime random process
More informationLecture 1: Brief Review on Stochastic Processes
Lecture 1: Brief Review on Stochastic Processes A stochastic process is a collection of random variables {X t (s) : t T, s S}, where T is some index set and S is the common sample space of the random variables.
More informationStochastic processes. MAS275 Probability Modelling. Introduction and Markov chains. Continuous time. Markov property
Chapter 1: and Markov chains Stochastic processes We study stochastic processes, which are families of random variables describing the evolution of a quantity with time. In some situations, we can treat
More information18.175: Lecture 30 Markov chains
18.175: Lecture 30 Markov chains Scott Sheffield MIT Outline Review what you know about finite state Markov chains Finite state ergodicity and stationarity More general setup Outline Review what you know
More informationStatistics 253/317 Introduction to Probability Models. Winter Midterm Exam Monday, Feb 10, 2014
Statistics 253/317 Introduction to Probability Models Winter 2014 - Midterm Exam Monday, Feb 10, 2014 Student Name (print): (a) Do not sit directly next to another student. (b) This is a closed-book, closed-note
More informationStochastic Processes
qmc082.tex. Version of 30 September 2010. Lecture Notes on Quantum Mechanics No. 8 R. B. Griffiths References: Stochastic Processes CQT = R. B. Griffiths, Consistent Quantum Theory (Cambridge, 2002) DeGroot
More informationMARKOV PROCESSES. Valerio Di Valerio
MARKOV PROCESSES Valerio Di Valerio Stochastic Process Definition: a stochastic process is a collection of random variables {X(t)} indexed by time t T Each X(t) X is a random variable that satisfy some
More informationTHE QUEEN S UNIVERSITY OF BELFAST
THE QUEEN S UNIVERSITY OF BELFAST 0SOR20 Level 2 Examination Statistics and Operational Research 20 Probability and Distribution Theory Wednesday 4 August 2002 2.30 pm 5.30 pm Examiners { Professor R M
More informationChapter 16 focused on decision making in the face of uncertainty about one future
9 C H A P T E R Markov Chains Chapter 6 focused on decision making in the face of uncertainty about one future event (learning the true state of nature). However, some decisions need to take into account
More informationWhat is Probability? Probability. Sample Spaces and Events. Simple Event
What is Probability? Probability Peter Lo Probability is the numerical measure of likelihood that the event will occur. Simple Event Joint Event Compound Event Lies between 0 & 1 Sum of events is 1 1.5
More informationSample Spaces, Random Variables
Sample Spaces, Random Variables Moulinath Banerjee University of Michigan August 3, 22 Probabilities In talking about probabilities, the fundamental object is Ω, the sample space. (elements) in Ω are denoted
More informationProblem Points S C O R E Total: 120
PSTAT 160 A Final Exam Solution December 10, 2015 Name Student ID # Problem Points S C O R E 1 10 2 10 3 10 4 10 5 10 6 10 7 10 8 10 9 10 10 10 11 10 12 10 Total: 120 1. (10 points) Take a Markov chain
More informationProblem Points S C O R E Total: 120
PSTAT 160 A Final Exam December 10, 2015 Name Student ID # Problem Points S C O R E 1 10 2 10 3 10 4 10 5 10 6 10 7 10 8 10 9 10 10 10 11 10 12 10 Total: 120 1. (10 points) Take a Markov chain with the
More informationhttp://www.math.uah.edu/stat/markov/.xhtml 1 of 9 7/16/2009 7:20 AM Virtual Laboratories > 16. Markov Chains > 1 2 3 4 5 6 7 8 9 10 11 12 1. A Markov process is a random process in which the future is
More informationTCOM 501: Networking Theory & Fundamentals. Lecture 6 February 19, 2003 Prof. Yannis A. Korilis
TCOM 50: Networking Theory & Fundamentals Lecture 6 February 9, 003 Prof. Yannis A. Korilis 6- Topics Time-Reversal of Markov Chains Reversibility Truncating a Reversible Markov Chain Burke s Theorem Queues
More informationThe Boundary Problem: Markov Chain Solution
MATH 529 The Boundary Problem: Markov Chain Solution Consider a random walk X that starts at positive height j, and on each independent step, moves upward a units with probability p, moves downward b units
More informationMarkov Chains. As part of Interdisciplinary Mathematical Modeling, By Warren Weckesser Copyright c 2006.
Markov Chains As part of Interdisciplinary Mathematical Modeling, By Warren Weckesser Copyright c 2006 1 Introduction A (finite) Markov chain is a process with a finite number of states (or outcomes, or
More informationMATH 56A: STOCHASTIC PROCESSES CHAPTER 1
MATH 56A: STOCHASTIC PROCESSES CHAPTER. Finite Markov chains For the sake of completeness of these notes I decided to write a summary of the basic concepts of finite Markov chains. The topics in this chapter
More informationLecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes
Lecture Notes 7 Random Processes Definition IID Processes Bernoulli Process Binomial Counting Process Interarrival Time Process Markov Processes Markov Chains Classification of States Steady State Probabilities
More informationMarkov Processes Hamid R. Rabiee
Markov Processes Hamid R. Rabiee Overview Markov Property Markov Chains Definition Stationary Property Paths in Markov Chains Classification of States Steady States in MCs. 2 Markov Property A discrete
More informationInterlude: Practice Final
8 POISSON PROCESS 08 Interlude: Practice Final This practice exam covers the material from the chapters 9 through 8. Give yourself 0 minutes to solve the six problems, which you may assume have equal point
More informationOverview of Probability. Mark Schmidt September 12, 2017
Overview of Probability Mark Schmidt September 12, 2017 Dungeons & Dragons scenario: You roll dice 1: Practical Application Roll or you sneak past monster. Otherwise, you are eaten. If you survive, you
More information1. what conditional independencies are implied by the graph. 2. whether these independecies correspond to the probability distribution
NETWORK ANALYSIS Lourens Waldorp PROBABILITY AND GRAPHS The objective is to obtain a correspondence between the intuitive pictures (graphs) of variables of interest and the probability distributions of
More informationDiscrete Random Variables (1) Solutions
STAT/MATH 394 A - PROBABILITY I UW Autumn Quarter 06 Néhémy Lim Discrete Random Variables ( Solutions Problem. The probability mass function p X of some discrete real-valued random variable X is given
More informationProbability Theory and Applications
Probability Theory and Applications Videos of the topics covered in this manual are available at the following links: Lesson 4 Probability I http://faculty.citadel.edu/silver/ba205/online course/lesson
More informationCS 798: Homework Assignment 3 (Queueing Theory)
1.0 Little s law Assigned: October 6, 009 Patients arriving to the emergency room at the Grand River Hospital have a mean waiting time of three hours. It has been found that, averaged over the period of
More informationISE/OR 760 Applied Stochastic Modeling
ISE/OR 760 Applied Stochastic Modeling Topic 2: Discrete Time Markov Chain Yunan Liu Department of Industrial and Systems Engineering NC State University Yunan Liu (NC State University) ISE/OR 760 1 /
More information(b) What is the variance of the time until the second customer arrives, starting empty, assuming that we measure time in minutes?
IEOR 3106: Introduction to Operations Research: Stochastic Models Fall 2006, Professor Whitt SOLUTIONS to Final Exam Chapters 4-7 and 10 in Ross, Tuesday, December 19, 4:10pm-7:00pm Open Book: but only
More informationChapter 10 Markov Chains and Transition Matrices
Finite Mathematics (Mat 119) Lecture week 3 Dr. Firozzaman Department of Mathematics and Statistics Arizona State University Chapter 10 Markov Chains and Transition Matrices A Markov Chain is a sequence
More informationBefore you begin read these instructions carefully.
MATHEMATICAL TRIPOS Part IA Friday, 3 June, 2011 1:30 pm to 4:30 pm PAPER 2 Before you begin read these instructions carefully. The examination paper is divided into two sections. Each question in Section
More informationA REVIEW AND APPLICATION OF HIDDEN MARKOV MODELS AND DOUBLE CHAIN MARKOV MODELS
A REVIEW AND APPLICATION OF HIDDEN MARKOV MODELS AND DOUBLE CHAIN MARKOV MODELS Michael Ryan Hoff A Dissertation submitted to the Faculty of Science, University of the Witwatersrand, Johannesburg, in fulfilment
More informationMathematical Methods for Computer Science
Mathematical Methods for Computer Science Computer Science Tripos, Part IB Michaelmas Term 2016/17 R.J. Gibbens Problem sheets for Probability methods William Gates Building 15 JJ Thomson Avenue Cambridge
More informationMarkov Chains and Hidden Markov Models
Chapter 1 Markov Chains and Hidden Markov Models In this chapter, we will introduce the concept of Markov chains, and show how Markov chains can be used to model signals using structures such as hidden
More informationBudapest University of Tecnology and Economics. AndrásVetier Q U E U I N G. January 25, Supported by. Pro Renovanda Cultura Hunariae Alapítvány
Budapest University of Tecnology and Economics AndrásVetier Q U E U I N G January 25, 2000 Supported by Pro Renovanda Cultura Hunariae Alapítvány Klebelsberg Kunó Emlékére Szakalapitvány 2000 Table of
More informationProbability and Statistics
Probability and Statistics 1 Contents some stochastic processes Stationary Stochastic Processes 2 4. Some Stochastic Processes 4.1 Bernoulli process 4.2 Binomial process 4.3 Sine wave process 4.4 Random-telegraph
More informationLecture 6: Entropy Rate
Lecture 6: Entropy Rate Entropy rate H(X) Random walk on graph Dr. Yao Xie, ECE587, Information Theory, Duke University Coin tossing versus poker Toss a fair coin and see and sequence Head, Tail, Tail,
More informationDept. of Linguistics, Indiana University Fall 2015
L645 Dept. of Linguistics, Indiana University Fall 2015 1 / 34 To start out the course, we need to know something about statistics and This is only an introduction; for a fuller understanding, you would
More information1 Probability Spaces and Random Variables
1 Probability Saces and Random Variables 1.1 Probability saces Ω: samle sace consisting of elementary events (or samle oints). F : the set of events P: robability 1.2 Kolmogorov s axioms Definition 1.2.1
More informationIEOR 3106: Introduction to Operations Research: Stochastic Models. Professor Whitt. SOLUTIONS to Homework Assignment 1
IEOR 3106: Introduction to Operations Research: Stochastic Models Professor Whitt SOLUTIONS to Homework Assignment 1 Probability Review: Read Chapters 1 and 2 in the textbook, Introduction to Probability
More informationThe probability of going from one state to another state on the next trial depends only on the present experiment and not on past history.
c Dr Oksana Shatalov, Fall 2010 1 9.1: Markov Chains DEFINITION 1. Markov process, or Markov Chain, is an experiment consisting of a finite number of stages in which the outcomes and associated probabilities
More informationConditional Probability
Conditional Probability Conditional Probability The Law of Total Probability Let A 1, A 2,..., A k be mutually exclusive and exhaustive events. Then for any other event B, P(B) = P(B A 1 ) P(A 1 ) + P(B
More information5. Conditional Distributions
1 of 12 7/16/2009 5:36 AM Virtual Laboratories > 3. Distributions > 1 2 3 4 5 6 7 8 5. Conditional Distributions Basic Theory As usual, we start with a random experiment with probability measure P on an
More informationGeneral Boundary Problem: System of Equations Solution
MATH 54 General Boundary Problem: System of Equations Solution Consider again the scenario of a random walk X that starts at positive height j, and on each independent step, moves upward a units with probability
More informationStatistics for Managers Using Microsoft Excel/SPSS Chapter 4 Basic Probability And Discrete Probability Distributions
Statistics for Managers Using Microsoft Excel/SPSS Chapter 4 Basic Probability And Discrete Probability Distributions 1999 Prentice-Hall, Inc. Chap. 4-1 Chapter Topics Basic Probability Concepts: Sample
More information9. DISCRETE PROBABILITY DISTRIBUTIONS
9. DISCRETE PROBABILITY DISTRIBUTIONS Random Variable: A quantity that takes on different values depending on chance. Eg: Next quarter s sales of Coca Cola. The proportion of Super Bowl viewers surveyed
More informationBe sure this exam has 9 pages including the cover. The University of British Columbia
Be sure this exam has 9 pages including the cover The University of British Columbia Sessional Exams 2011 Term 2 Mathematics 303 Introduction to Stochastic Processes Dr. D. Brydges Last Name: First Name:
More informationChapter 29 out of 37 from Discrete Mathematics for Neophytes: Number Theory, Probability, Algorithms, and Other Stuff by J. M.
29 Markov Chains Definition of a Markov Chain Markov chains are one of the most fun tools of probability; they give a lot of power for very little effort. We will restrict ourselves to finite Markov chains.
More informationUncertainty Runs Rampant in the Universe C. Ebeling circa Markov Chains. A Stochastic Process. Into each life a little uncertainty must fall.
Uncertainty Runs Rampant in the Universe C. Ebeling circa 2000 Markov Chains A Stochastic Process Into each life a little uncertainty must fall. Our Hero - Andrei Andreyevich Markov Born: 14 June 1856
More informationPhysics 403 Probability Distributions II: More Properties of PDFs and PMFs
Physics 403 Probability Distributions II: More Properties of PDFs and PMFs Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Last Time: Common Probability Distributions
More informationIntroduction to Probability 2017/18 Supplementary Problems
Introduction to Probability 2017/18 Supplementary Problems Problem 1: Let A and B denote two events with P(A B) 0. Show that P(A) 0 and P(B) 0. A A B implies P(A) P(A B) 0, hence P(A) 0. Similarly B A
More informationCMI/BVR Probability notes 5. Conditional distributions and Conditional expectations:
CMI/BVR Probability notes 5 Conditional distributions and Conditional expectations: You now have all the basic material: probability and conditional probability; random variables; expectation and variance
More informationLECTURE 1. 1 Introduction. 1.1 Sample spaces and events
LECTURE 1 1 Introduction The first part of our adventure is a highly selective review of probability theory, focusing especially on things that are most useful in statistics. 1.1 Sample spaces and events
More informationHomework #2 Solutions Due: September 5, for all n N n 3 = n2 (n + 1) 2 4
Do the following exercises from the text: Chapter (Section 3):, 1, 17(a)-(b), 3 Prove that 1 3 + 3 + + n 3 n (n + 1) for all n N Proof The proof is by induction on n For n N, let S(n) be the statement
More informationPermutation. Permutation. Permutation. Permutation. Permutation
Conditional Probability Consider the possible arrangements of the letters A, B, and C. The possible arrangements are: ABC, ACB, BAC, BCA, CAB, CBA. If the order of the arrangement is important then we
More informationStochastic Processes MIT, fall 2011 Day by day lecture outline and weekly homeworks. A) Lecture Outline Suggested reading
Stochastic Processes 18445 MIT, fall 2011 Day by day lecture outline and weekly homeworks A) Lecture Outline Suggested reading Part 1: Random walk on Z Lecture 1: thursday, september 8, 2011 Presentation
More informationLaw of large numbers for Markov chains
Chapter 14 Law of large numbers for Markov chains In this chapter we consider the equilibrium state of a Markov chain in the long run limit of many steps. This limit of observing the dynamic chain over
More information18.600: Lecture 32 Markov Chains
18.600: Lecture 32 Markov Chains Scott Sheffield MIT Outline Markov chains Examples Ergodicity and stationarity Outline Markov chains Examples Ergodicity and stationarity Markov chains Consider a sequence
More informationLecture 20 : Markov Chains
CSCI 3560 Probability and Computing Instructor: Bogdan Chlebus Lecture 0 : Markov Chains We consider stochastic processes. A process represents a system that evolves through incremental changes called
More informationProbability Distributions for Discrete RV
An example: Assume we toss a coin 3 times and record the outcomes. Let X i be a random variable defined by { 1, if the i th outcome is Head; X i = 0, if the i th outcome is Tail; Let X be the random variable
More informationRapid Introduction to Machine Learning/ Deep Learning
Rapid Introduction to Machine Learning/ Deep Learning Hyeong In Choi Seoul National University 1/32 Lecture 5a Bayesian network April 14, 2016 2/32 Table of contents 1 1. Objectives of Lecture 5a 2 2.Bayesian
More informationMA 1125 Lecture 33 - The Sign Test. Monday, December 4, Objectives: Introduce an example of a non-parametric test.
MA 1125 Lecture 33 - The Sign Test Monday, December 4, 2017 Objectives: Introduce an example of a non-parametric test. For the last topic of the semester we ll look at an example of a non-parametric test.
More informationCMPSCI 240: Reasoning about Uncertainty
CMPSCI 240: Reasoning about Uncertainty Lecture 17: Representing Joint PMFs and Bayesian Networks Andrew McGregor University of Massachusetts Last Compiled: April 7, 2017 Warm Up: Joint distributions Recall
More informationGenerating Functions
8.30 lecture notes March, 0 Generating Functions Lecturer: Michel Goemans We are going to discuss enumeration problems, and how to solve them using a powerful tool: generating functions. What is an enumeration
More informationMath 341: Probability Eighth Lecture (10/6/09)
Math 341: Probability Eighth Lecture (10/6/09) Steven J Miller Williams College Steven.J.Miller@williams.edu http://www.williams.edu/go/math/sjmiller/ public html/341/ Bronfman Science Center Williams
More informationStability and Rare Events in Stochastic Models Sergey Foss Heriot-Watt University, Edinburgh and Institute of Mathematics, Novosibirsk
Stability and Rare Events in Stochastic Models Sergey Foss Heriot-Watt University, Edinburgh and Institute of Mathematics, Novosibirsk ANSAPW University of Queensland 8-11 July, 2013 1 Outline (I) Fluid
More informationMAT 135B Midterm 1 Solutions
MAT 35B Midterm Solutions Last Name (PRINT): First Name (PRINT): Student ID #: Section: Instructions:. Do not open your test until you are told to begin. 2. Use a pen to print your name in the spaces above.
More informationP(X 0 = j 0,... X nk = j k )
Introduction to Probability Example Sheet 3 - Michaelmas 2006 Michael Tehranchi Problem. Let (X n ) n 0 be a homogeneous Markov chain on S with transition matrix P. Given a k N, let Z n = X kn. Prove that
More informationClassical Information Theory Notes from the lectures by prof Suhov Trieste - june 2006
Classical Information Theory Notes from the lectures by prof Suhov Trieste - june 2006 Fabio Grazioso... July 3, 2006 1 2 Contents 1 Lecture 1, Entropy 4 1.1 Random variable...............................
More informationProbability: Part 1 Naima Hammoud
Probability: Part 1 Naima ammoud Feb 7, 2017 Motivation ossing a coin Rolling a die Outcomes: eads or ails Outcomes: 1, 2, 3, 4, 5 or 6 Defining Probability If I toss a coin, there is a 50% chance I will
More informationChapter 1 Axioms of Probability. Wen-Guey Tzeng Computer Science Department National Chiao University
Chapter 1 Axioms of Probability Wen-Guey Tzeng Computer Science Department National Chiao University What is probability? A branch of mathematics that deals with calculating the likelihood of a given event
More informationLet us think of the situation as having a 50 sided fair die; any one number is equally likely to appear.
Probability_Homework Answers. Let the sample space consist of the integers through. {, 2, 3,, }. Consider the following events from that Sample Space. Event A: {a number is a multiple of 5 5, 0, 5,, }
More informationArtificial Intelligence Uncertainty
Artificial Intelligence Uncertainty Ch. 13 Uncertainty Let action A t = leave for airport t minutes before flight Will A t get me there on time? A 25, A 60, A 3600 Uncertainty: partial observability (road
More informationBasic Probability space, sample space concepts and order of a Stochastic Process
The Lecture Contains: Basic Introduction Basic Probability space, sample space concepts and order of a Stochastic Process Examples Definition of Stochastic Process Marginal Distributions Moments Gaussian
More informationProbability and Stochastic Processes Homework Chapter 12 Solutions
Probability and Stochastic Processes Homework Chapter 1 Solutions Problem Solutions : Yates and Goodman, 1.1.1 1.1.4 1.3. 1.4.3 1.5.3 1.5.6 1.6.1 1.9.1 1.9.4 1.10.1 1.10.6 1.11.1 1.11.3 1.11.5 and 1.11.9
More informationn(1 p i ) n 1 p i = 1 3 i=1 E(X i p = p i )P(p = p i ) = 1 3 p i = n 3 (p 1 + p 2 + p 3 ). p i i=1 P(X i = 1 p = p i )P(p = p i ) = p1+p2+p3
Introduction to Probability Due:August 8th, 211 Solutions of Final Exam Solve all the problems 1. (15 points) You have three coins, showing Head with probabilities p 1, p 2 and p 3. You perform two different
More informationCSE 312 Final Review: Section AA
CSE 312 TAs December 8, 2011 General Information General Information Comprehensive Midterm General Information Comprehensive Midterm Heavily weighted toward material after the midterm Pre-Midterm Material
More informationPart IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationA COMPOUND POISSON APPROXIMATION INEQUALITY
J. Appl. Prob. 43, 282 288 (2006) Printed in Israel Applied Probability Trust 2006 A COMPOUND POISSON APPROXIMATION INEQUALITY EROL A. PEKÖZ, Boston University Abstract We give conditions under which the
More informationContents. CONTENTS Page 1
CONTENTS Page 1 Contents 1 Markov chains 5 1.1 Specifying and simulating a Markov chain................... 5 1.2 The Markov property.............................. 8 1.3 It s all just matrix theory...........................
More informationLecture 25 Goals: Chapter 18 Understand the molecular basis for pressure and the idealgas
Lecture 5 Goals: Chapter 18 Understand the molecular basis for pressure and the idealgas law. redict the molar specific heats of gases and solids. Understand how heat is transferred via molecular collisions
More information18.175: Lecture 8 Weak laws and moment-generating/characteristic functions
18.175: Lecture 8 Weak laws and moment-generating/characteristic functions Scott Sheffield MIT 18.175 Lecture 8 1 Outline Moment generating functions Weak law of large numbers: Markov/Chebyshev approach
More informationProbability Theory and Simulation Methods
Feb 28th, 2018 Lecture 10: Random variables Countdown to midterm (March 21st): 28 days Week 1 Chapter 1: Axioms of probability Week 2 Chapter 3: Conditional probability and independence Week 4 Chapters
More informationJUSTIN HARTMANN. F n Σ.
BROWNIAN MOTION JUSTIN HARTMANN Abstract. This paper begins to explore a rigorous introduction to probability theory using ideas from algebra, measure theory, and other areas. We start with a basic explanation
More informationApplied Statistics I
Applied Statistics I Liang Zhang Department of Mathematics, University of Utah June 17, 2008 Liang Zhang (UofU) Applied Statistics I June 17, 2008 1 / 22 Random Variables Definition A dicrete random variable
More informationProbability and Discrete Distributions
AMS 7L LAB #3 Fall, 2007 Objectives: Probability and Discrete Distributions 1. To explore relative frequency and the Law of Large Numbers 2. To practice the basic rules of probability 3. To work with the
More informationContinuous-Valued Probability Review
CS 6323 Continuous-Valued Probability Review Prof. Gregory Provan Department of Computer Science University College Cork 2 Overview Review of discrete distributions Continuous distributions 3 Discrete
More informationLecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past.
1 Markov chain: definition Lecture 5 Definition 1.1 Markov chain] A sequence of random variables (X n ) n 0 taking values in a measurable state space (S, S) is called a (discrete time) Markov chain, if
More informationDiscrete Mathematics and Probability Theory Summer 2017 Hongling Lu, Vrettos Moulos, and Allen Tang HW 6
CS 70 Discrete Mathematics and Probability Theory Summer 017 Hongling Lu, Vrettos Moulos, and Allen Tang HW 6 Sundry Before you start your homework, write down your team. Who else did you work with on
More informationAn Introduction to Stochastic Modeling
F An Introduction to Stochastic Modeling Fourth Edition Mark A. Pinsky Department of Mathematics Northwestern University Evanston, Illinois Samuel Karlin Department of Mathematics Stanford University Stanford,
More informationQuantitative Methods for Decision Making
January 14, 2012 Lecture 3 Probability Theory Definition Mutually exclusive events: Two events A and B are mutually exclusive if A B = φ Definition Special Addition Rule: Let A and B be two mutually exclusive
More informationGreat Theoretical Ideas in Computer Science
15-251 Great Theoretical Ideas in Computer Science Probability Theory: Counting in Terms of Proportions Lecture 10 (September 27, 2007) Some Puzzles Teams A and B are equally good In any one game, each
More information