Lecture 4. David Aldous. 2 September David Aldous Lecture 4

Similar documents
Lecture 3. David Aldous. 31 August David Aldous Lecture 3

Discrete random structures whose limits are described by a PDE: 3 open problems

IEOR 6711: Stochastic Models I. Solutions to Homework Assignment 9

Math-Stat-491-Fall2014-Notes-V

Inference for Stochastic Processes

IB Math Standard Level Year 1: Final Exam Review Alei - Desert Academy

Course: ESO-209 Home Work: 1 Instructor: Debasis Kundu

Class 26: review for final exam 18.05, Spring 2014

1 Gambler s Ruin Problem

1 Presessional Probability

CSE 312 Final Review: Section AA

Deep Learning for Computer Vision

Math , Fall 2012: HW 5 Solutions

6.041/6.431 Spring 2009 Quiz 1 Wednesday, March 11, 7:30-9:30 PM. SOLUTIONS

IEOR 3106: Introduction to Operations Research: Stochastic Models. Professor Whitt. SOLUTIONS to Homework Assignment 2

Lecture 1: Review on Probability and Statistics

n(1 p i ) n 1 p i = 1 3 i=1 E(X i p = p i )P(p = p i ) = 1 3 p i = n 3 (p 1 + p 2 + p 3 ). p i i=1 P(X i = 1 p = p i )P(p = p i ) = p1+p2+p3

UC Berkeley Department of Electrical Engineering and Computer Science. EE 126: Probablity and Random Processes. Solutions 5 Spring 2006

P i [B k ] = lim. n=1 p(n) ii <. n=1. V i :=

Discrete Random Variables (1) Solutions

Exercises in Probability Theory Paul Jung MA 485/585-1C Fall 2015 based on material of Nikolai Chernov

Solutionbank S1 Edexcel AS and A Level Modular Mathematics

T 1. The value function v(x) is the expected net gain when using the optimal stopping time starting at state x:

Martingale Theory for Finance

What is Probability? Probability. Sample Spaces and Events. Simple Event

MATH 3C: MIDTERM 1 REVIEW. 1. Counting

MAT 271E Probability and Statistics

. Find E(V ) and var(v ).

Math 493 Final Exam December 01

COMS 4721: Machine Learning for Data Science Lecture 20, 4/11/2017

Ismor Fischer, 2/27/2018 Solutions / 3.5-1

MATH 564/STAT 555 Applied Stochastic Processes Homework 2, September 18, 2015 Due September 30, 2015

Problem Points S C O R E Total: 120

5. Conditional Distributions

General Boundary Problem: System of Equations Solution

Interlude: Practice Final

Lecture notes for probability. Math 124

Probability and Distributions

Topic 5: Probability. 5.4 Combined Events and Conditional Probability Paper 1

18.440: Lecture 28 Lectures Review

1 Exercises for lecture 1

The genome encodes biology as patterns or motifs. We search the genome for biologically important patterns.

Midterm Exam 1 (Solutions)

Review of Probability. CS1538: Introduction to Simulations

Midterm Exam 2 (Solutions)

UCSD CSE 21, Spring 2014 [Section B00] Mathematics for Algorithm and System Analysis

Lecture 21. David Aldous. 16 October David Aldous Lecture 21

MATH HOMEWORK PROBLEMS D. MCCLENDON

EECS 126 Probability and Random Processes University of California, Berkeley: Spring 2017 Kannan Ramchandran March 21, 2017.

Problem 1: (Chernoff Bounds via Negative Dependence - from MU Ex 5.15)

2. Variance and Covariance: We will now derive some classic properties of variance and covariance. Assume real-valued random variables X and Y.

Math 510 midterm 3 answers

Section 2.5 Linear Inequalities

Probability Models. 4. What is the definition of the expectation of a discrete random variable?

Admin and Lecture 1: Recap of Measure Theory

Example continued. Math 425 Intro to Probability Lecture 37. Example continued. Example

Chapter 7 Probability Basics

CS 125 Section #12 (More) Probability and Randomized Algorithms 11/24/14. For random numbers X which only take on nonnegative integer values, E(X) =

Discrete Event Systems Solution to Exercise Sheet 6

18.440: Lecture 28 Lectures Review

Math 381 Discrete Mathematical Modeling

On random walks. i=1 U i, where x Z is the starting point of

18.440: Lecture 19 Normal random variables

Notes on Discrete Probability

The Geometric Distribution

Markov Chains. Chapter 16. Markov Chains - 1

Selected Exercises on Expectations and Some Probability Inequalities

Question Paper Code : AEC11T03

CHAPTER 6 SOME CONTINUOUS PROBABILITY DISTRIBUTIONS. 6.2 Normal Distribution. 6.1 Continuous Uniform Distribution

Lecture 4 - Random walk, ruin problems and random processes

STOR Lecture 4. Axioms of Probability - II

MATH 19B FINAL EXAM PROBABILITY REVIEW PROBLEMS SPRING, 2010

Homework 6: Solutions Sid Banerjee Problem 1: (The Flajolet-Martin Counter) ORIE 4520: Stochastics at Scale Fall 2015

Midterm Exam 1 Solution

Statistics Examples. Cathal Ormond

STAT J535: Introduction

(b). What is an expression for the exact value of P(X = 4)? 2. (a). Suppose that the moment generating function for X is M (t) = 2et +1 3

Random Variables and Their Distributions

The story of the film so far... Mathematics for Informatics 4a. Continuous-time Markov processes. Counting processes

Discrete Mathematics and Probability Theory Fall 2017 Ramchandran and Rao Final Solutions

Math 122L. Additional Homework Problems. Prepared by Sarah Schott

MATH Notebook 5 Fall 2018/2019

1.1 Review of Probability Theory

Markov Chains (Part 3)

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

Chapter 2 Random Variables

Notes for Math 324, Part 17

6.041/6.431 Fall 2010 Quiz 2 Solutions

MAS223 Statistical Inference and Modelling Exercises

STAT Chapter 5 Continuous Distributions

4 Branching Processes

Edexcel past paper questions

4th IIA-Penn State Astrostatistics School July, 2013 Vainu Bappu Observatory, Kavalur

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

MATH 56A SPRING 2008 STOCHASTIC PROCESSES

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

ECE 302: Probabilistic Methods in Engineering

Week 2: Review of probability and statistics

Mutually Exclusive Events

2. Suppose (X, Y ) is a pair of random variables uniformly distributed over the triangle with vertices (0, 0), (2, 0), (2, 1).

Transcription:

Lecture 4 David Aldous 2 September 2015

The specific examples I m discussing are not so important; the point of these first lectures is to illustrate a few of the 100 ideas from STAT134. Ideas used in Lecture 3. Distributions of max(x 1, X 2 ) and min(x 1, X 2 ) Properties of Exponential distribution. Basics of conditional probability and conditional expectation. Global and local interpretations of density function. Calculating E(X A) from conditional distributions. Conditional expectation as a random variable.

Conditional expectation as a random variable. Given r.v. s (W, Y ) consider E(W Y = y). This is a number depending on y in other words it s a function of y. Giving this function a name h we have ( ) E(W Y = y) = h(y) for all possible values y of Y. We now make a notational convention, to rewrite the assertion (*) as ( ) E(W Y ) = h(y ). The right side is a r.v., so we must regard E(W Y ) as a r.v. [PK] page 60 lists properties of (**), but rather hard to understand at first sight. One important property is that the law of total probability becomes E[ E(W Y ) ] = EW. I will give three examples to illustrate this notation.

Example. First consider S n = n ξ i for i.i.d. (ξ i ) with Eξ i = µ ξ. i=1 Then take N a {1, 2, 3,...}-valued r.v. with EN = µ N, independent of (ξ i ). What is ES N?

S n = n ξ i for i.i.d. (ξ i ) with Eξ i = µ ξ. i=1 N a {1, 2, 3,...}-valued r.v. with EN = µ N, independent of (ξ i ). - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - ES n = nµ ξ. E(S N N = n) = nµ ξ. E(S N N) = N µ ξ. ES N = E[E(S N N)] = E[N µ ξ ] = µ N µ ξ.

A geometric probability example. Saying that a random point in the plane is uniform on a set A is saying that the joint density function of its coordinates (X, Y ) is f (x, y) = 1 area(a), (x, y) A. Let s do some calculations in the case where A is the top half of the unit disc. In particular, we will calculate EY by first considering E(Y X ).

Conditional distribution of Y given X = x is uniform on [0, 1 x 2 ]. So E(Y X = x) = 1 2 1 x 2 ; that is E(Y X ) = 1 2 1 X 2. We can now calculate EY = E[E(Y X )] = 1 2 E 1 X 2 = 1 2 1 1 1 x 2 f X (x)dx 1 = π 1 (1 x 2 )dx = 4 3π. 1

Example: Likelihood of your team winning, at halftime. Most team sports are decided by point difference points scored by home team, minus points scored by visiting team. Write X 1 = point difference in first half X 2 = point difference in second half so home team wins if X 1 + X 2 > 0 visiting team wins if X 1 + X 2 < 0 tie if X 1 + X 2 = 0. A reasonable probability model assumes (i) X 1 and X 2 are i.i.d. random variables. If the teams are equally talented, then assume (ii) X 1 has symmetric distribution, that is the same distribution as X 1. Finally, let me simplify the math by making an unrealistic assumption (iii) the distribution of X 1 is continuous.

So ties cannot happen, and by (ii) P(Home team wins) = 1/2. This is the probability before the game starts. At half time we know the value of X 1, so there is a conditional probability P(Home team wins X 1 ). This is different in different matches, so we can ask What is the distribution of P(Home team wins X 1 )? First I show some data from baseball.

In this match the initial price (perceived probability of home team winning) was close to 50%, and halfway through the game it was about 64%. I have these halfway through the game numbers for 30 matches where the initial price was close to 50%

proportion of data probability 0 50 price 100 Figure. Empirical distribution function for the baseball data, compared with the uniform distribution.

Write F (x) for the distribution function of each X i. P(Home team wins X 1 = x) = P(X 1 + X 2 > 0 X 1 = x) and this says = P(X 2 > x) = P( X 2 < x) = F (x) by symmetry assumption P(Home team wins X 1 ) = F (X 1 ). But we know (STAT134) that for X 1 with continuous distribution, F (X 1 ) always has Uniform(0, 1) distribution. The baseball data fits this theory prediction quite well. In a low-scoring sport like soccer the continuous distribution of X i assumption is not realistic but we could modify the analysis.

Stochastic Processes For our purposes a stochastic process is just a sequence (X 0, X 1, X 2,...) of random variables. The range space of the X s might be the integers Z or the reals R or some more general space S. We envisage observing some physical process changing with time in some random way: X t is the observed value of the process at time t = 0, 1, 2,.... So X 0 is the initial value. Perhaps the simplest example is simple symmetric random walk X t = where (ξ i ) are i.i.d. with P(ξ i = 1) = P(ξ i = 1) = 1 2. Here X 0 = 0. There are many ways to study this process, but I want to illustrate the technique conditioning on the first step which will be a fundamental technique for studying Markov chains. t i=1 ξ i

Interpret this process as gambling at fair odds bet 1 unit on an event with probability 1/2, either win or lose the 1 unit. So X t is your fortune after t bets. Suppose start with fortune x continue until your fortune reach some target amount K or 0. There is some probability p(x) that you succeed in reaching K; this probability depends on x and on K, but let us take K as fixed. So we know p(k) = 1; p(0) = 0. What can we say about p(x) for 1 x K 1?

From the law of total probability we have That is, P(A) = P(B)P(A B) + P(B c )P(A B c ) p(x) = P( win first bet) P( reach K win first bet) + P( lose first bet) P( reach K lose first bet) ( ) p(x) = 1 2 p(x + 1) + 1 2p(x 1), 1 x K 1. This is the simplest case of a linear difference equation and the general solution is p(x) = a + bx. Because we know the boundary conditions p(0) = 0, p(k) = 1 we can solve to find a, b and find p(x) = x/k, 0 x K.

We can use the same conditioning on the first step method to study s(x) = E(number of steps until reach K or 0) starting from x. Here the equation is s(x) = 1 + 1 2 s(x + 1) + 1 2s(x 1), 1 x K 1 s(0) = 0, s(k) = 0. The general form of solution (see textbook for details) is s(x) = cx(k x); plugging into the equation gives c = 1, so s(x) = x(k x), 0 x K.

Conceptual point. The notion of independence is used in two conceptually different ways. We often use independence as an assumption in a model throwing dice, for instance. Given a well-defined math model, events or random variables X, Y either are independent, or are not independent, as a mathematical conclusion. For instance, for a uniform random pick of a playing card from a standard deck, the events card is a King and card is a Spade are independent. The same point will arise with the Markov property.