Probability Review. Chao Lan

Similar documents
Probability Review. Yutian Li. January 18, Stanford University. Yutian Li (Stanford University) Probability Review January 18, / 27

Dept. of Linguistics, Indiana University Fall 2015

Recitation 2: Probability

Joint Distribution of Two or More Random Variables

Appendix A : Introduction to Probability and stochastic processes

Lecture 2: Repetition of probability theory and statistics

More on Distribution Function

Algorithms for Uncertainty Quantification

Random Variables. Saravanan Vijayakumaran Department of Electrical Engineering Indian Institute of Technology Bombay

Review of Probability. CS1538: Introduction to Simulations

Probability. Paul Schrimpf. January 23, UBC Economics 326. Probability. Paul Schrimpf. Definitions. Properties. Random variables.

Random variables (discrete)

Chapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables

Probability. Paul Schrimpf. January 23, Definitions 2. 2 Properties 3

Bayesian statistics, simulation and software

E X A M. Probability Theory and Stochastic Processes Date: December 13, 2016 Duration: 4 hours. Number of pages incl.

STAT 430/510 Probability Lecture 7: Random Variable and Expectation

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

Fundamental Tools - Probability Theory II

Deep Learning for Computer Vision

Homework 4 Solution, due July 23

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

CME 106: Review Probability theory

Probability Theory for Machine Learning. Chris Cremer September 2015

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Probability Review Lecturer: Ji Liu Thank Jerry Zhu for sharing his slides

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016

More than one variable

Review: mostly probability and some statistics

Fundamentals. CS 281A: Statistical Learning Theory. Yangqing Jia. August, Based on tutorial slides by Lester Mackey and Ariel Kleiner

01 Probability Theory and Statistics Review

Basic Probability and Statistics

6.041/6.431 Fall 2010 Quiz 2 Solutions

Lecture 1: Probability Fundamentals

1 Random Variable: Topics

SDS 321: Introduction to Probability and Statistics

Lecture 16. Lectures 1-15 Review

Lecture 3. Discrete Random Variables

LECTURE 1. 1 Introduction. 1.1 Sample spaces and events

Quick Tour of Basic Probability Theory and Linear Algebra

Review of Probabilities and Basic Statistics

3 Multiple Discrete Random Variables

DISCRETE RANDOM VARIABLES: PMF s & CDF s [DEVORE 3.2]

Bivariate distributions

Introduction to Machine Learning

Lecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019

a zoo of (discrete) random variables

Single Maths B: Introduction to Probability

Chapter 2. Probability

Lecture 11. Probability Theory: an Overveiw

Chapter 2. Continuous random variables

MAS113 Introduction to Probability and Statistics. Proofs of theorems

CMPSCI 240: Reasoning Under Uncertainty

1 Presessional Probability

Probability. Computer Science Tripos, Part IA. R.J. Gibbens. Computer Laboratory University of Cambridge. Easter Term 2008/9

Lecture 10: Bayes' Theorem, Expected Value and Variance Lecturer: Lale Özkahya

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Data Modeling & Analysis Techniques. Probability & Statistics. Manfred Huber

Brief Review of Probability

M378K In-Class Assignment #1

Introduction to Stochastic Processes

Review of probability

Introduction to Statistical Inference Self-study

The Binomial distribution. Probability theory 2. Example. The Binomial distribution

Review of Probability Mark Craven and David Page Computer Sciences 760.

for valid PSD. PART B (Answer all five units, 5 X 10 = 50 Marks) UNIT I

Math Review Sheet, Fall 2008

Lecture 10. Variance and standard deviation

Chapter 4. Chapter 4 sections

1.1 Review of Probability Theory

MAS113 Introduction to Probability and Statistics. Proofs of theorems

Probability Theory Review

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN

Basics on Probability. Jingrui He 09/11/2007

Review of Probability Theory

A Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.

EE514A Information Theory I Fall 2013

STA 111: Probability & Statistical Inference

Lecture 2. October 21, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

Basic Probability. Introduction

ECE531: Principles of Detection and Estimation Course Introduction

Random Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline.

Exercises with solutions (Set D)

B4 Estimation and Inference

A random variable is a quantity whose value is determined by the outcome of an experiment.

1 Probability theory. 2 Random variables and probability theory.

L2: Review of probability and statistics

Lecture 2: Review of Probability

Lecture 3: Basic Statistical Tools. Bruce Walsh lecture notes Tucson Winter Institute 7-9 Jan 2013

Discrete Probability Refresher

Expectation of Random Variables

1 INFO Sep 05

Stochastic Models of Manufacturing Systems

CHAPTER 4 MATHEMATICAL EXPECTATION. 4.1 Mean of a Random Variable

STAT 430/510: Lecture 10

SDS 321: Introduction to Probability and Statistics

1: PROBABILITY REVIEW

Outline PMF, CDF and PDF Mean, Variance and Percentiles Some Common Distributions. Week 5 Random Variables and Their Distributions

Transcription:

Probability Review Chao Lan

Let s start with a single random variable

Random Experiment A random experiment has three elements 1. sample space Ω: set of all possible outcomes e.g.,ω={1,2,3,4,5,6} 2. event space E: collection of subsets of Ω e.g., E={{1},...,{1,2},...,{3,4,6},...,{1,2,3,4,5,6}}=2 Ω 3. probability measure P: a special function P : E R e.g., P({1})=1/6, P({3,4,6})=3/6

Axioms of Probability Probability measure P needs to satisfy 1. P(e) 0, for any event e E e.g.,p({3,4,6})=3/6 0 2. P(Ω)=1 e.g., P({1,2,3,4,5,6})=1 3. if e1,e2 E are disjoint, then P(e1 e2)=p(e1)+p(e2) e.g., P({1} {2,4,6})=P({1})+P({2,4,6})

Exercise Tomorrow s Temperature Tomorrow s Weather Sample Space Ω =? Sample Space Ω =? Event Space E =? Event Space E =? Probability Measure P? Probability Measure P?

Exercise A random experiment is flipping a coin twice (getting H or T in each trial) 1. sample space Ω =? 2. event space E =? 3. probability measure P? A random experiment can be a single trial or a sequence of trails.

Random Variable A random variable X is a function X : Ω R. Q: are the following random variables? 1. X({1}) = 1, X({3}) = 3, 2. X({1}) = 19, X({3}) = 45.7, 3. X({1}) = 1, X({3}) = 1,

Random Variable A random variable X is a function X : E R. Q: are the following random variables? 1. X({1}) = 1, X({3}) = 3, 2. X({1}) = 19, X({3}) = 45.7, 3. X({1}) = 1, X({3}) = 1, Q: why bother?

Random Variable Random variable makes presentation easier! Example 1 If X({1})=19 and X({3})=45.7, instead of writing P({1}), P({3}), we can write P(X=19), P(X=45.7). Example 2 If X counts odds over three trials, e.g., X({1, 2, 6})=1, X({1,2,3})=2, instead of writing P(1 odd) or P(2 odds), we can write P(X=1) or P(X=2).

Discrete Variable vs Continuous Variable X is discrete if it maps E to a discrete domain (our focus) - e.g., X({3})=1, X({1,2,3})=2 - specify P(X=1), P(X=2), etc X is continuous if it maps E to a continuous domain - e.g., X(angle) = 19.32 - specify P(0 X 30), etc For continuous random variable X, P(X=a)=0.

Properties of Probability 1. if e1 e2, then P(e1) P(e2) e.g., since {2} {2,4,6}, we have P({2}) P({2,4,6}) 2. P(e1 e2) min{p(e1), P(e2)} e.g., P({2,3} {2,4,6}) min{p({2,3}),p({2,4,6})} 3. P(e1 e2) P(e1) + P(e2) [union bound] e.g., P({2,3} {2,4,6}) P({2,3}) + P({2,4,6})

Properties of Probability 4. P(Ω \ e) = 1 - P(e) e.g., P({1,2,3,4,5,6}\{1,2,3}) = 1 - P({1,2,3}) 5. if e1,e2 are disjoint, e1 e2=ω, then P(e1)+P(e2)=1 [law of total probability] e.g., P({1,2,3})+P({4,5,6}) = 1

What is the distribution of a random variable?

Cumulative Distribution Function (CDF) CDF of X is a function CDFX: R [0,1] specifying a special probability CDFX(a) = P(X a) Example If X counts odds, P(X 3) is the chance of getting three trials with no more than 3 odds, i.e., P(X 3) = P({e:X(e) 3}) If X is temperature, P(X 27) is the chance of getting tomorrow s temperature of no higher than 27F i.e., P(X 27) = P({e:X(e) 27})

Properties of CDF CDF of X is a function CDFX: R [0,1] specifying a special probability CDFX(a) = P(X a) 1. 0 CDF(a) 1 2. lima - CDF(a) = 0 3. lima + CDF(a) = 1 4. if a b, then CDF(a) CDF(b)

Probability Density Function (PDF) PDF of a continuous random variable X is the derivative of its CDF p(x) = CDF(x) / x

Example: Gaussian Distribution If X follows a Gaussian distribution (or, normal distribution), then

Different PDFs

Properties of PDF PDF of a continuous random variable X is the derivative of its CDF p(x) = CDF(x) / x 1. p(x) 0 2. x (-,+ ) p(x) dx = 1 3. x S p(x) dx = P(X S)

Probability Mass Function (PMF) PMF of a discrete random variable X is a table of probabilities for all events.

Different PMFs

Properties of PMF PMF of a discrete random variable X is a table of probabilities for all events. 1. 1 p(x) 0 2. Σx (-,+ ) p(x) = 1 3. Σx S p(x) = P(X S)

Based on distribution, we can define expectation and variance of a variable...

Expectation Expectation of a random variable is its weighted average value. E[X] = Σx D x*p(x) X p(x) E[X] 1 0.5 = Σx D x*p(x) 2 0.25 = 1*0.5 + 2*0.25 + 3*0.25 3 0.25 = 1.75 or E[X] = x D x*p(x) dx

Expectation of Standard Normal Distribution Expectation of a normally distributed random variable ~N(0,σ2) is zero. One can similarly prove expectation of variable from N(m,σ2) is m.

Properties of Expectation Expectation of a random variable is its weighted average value. E[X] = Σx D x*p(x) 1. E[a] = a, for any constant a 2. E[a*X] = a*e[x] 3. E[X+Y] = E[X] + E[Y] 4. E[1X=k] = P(X=k) or E[X] = x D x*p(x) dx

Variance Variance of a random variable is its average deviation from its expectation. Var[X] = E [ (X-E[X])2 ] X p(x) E[X] Var[X] 1 0.5 = Σx D x*p(x) = Σx D (X-E[X])2*p(x) 2 0.25 = 1*0.5 + 2*0.25 + 3*0.25 = (1-1.75)2*0.5 + (2-1.75)2*0.25 + (3-1.75)2*0.25 3 0.25 = 1.75 = 0.6875 Note expectation E[X] is a constant.

Variance of Normally Distributed Variable

Variance of Other Distribution Poisson Distribution Rayleigh Distribution

Properties of Variance Variance of a random variable is its average deviation from its expectation. Var[X] = E [ (X-E[X])2 ] 1. Var[X] = E[X2] - E[X]2 2. Var[a] = 0 for constant a 3. Var[aX] = a2 Var[X]

Common Distributions

Now we look at two (discrete) random variables... The following slides are largely built on the lecture slides of probability by Klein and Abbeel.

Recap: Distribution of Single Variable Q: what is the probability that tomorrow is hot and sun, i.e. P(T=hot, W=sun)?

Joint Distribution of Two Variables Joint distribution is a table of probabilities of all joint events.

Marginal Distribution Sometimes we are interested in the distribution of one variable. We can obtain it from the joint distribution by marginalizing it. marginal distribution

Exercise

Conditional Distribution Sometimes we are interested in the distribution of one variable X while fixing the other Y. This is conditional distribution and defined as P(X Y) = P(X,Y) / P(Y).

Conditional Distribution Sometimes we are interested in the distribution of one variable X while fixing the other Y. This is conditional distribution and defined as P(X Y) = P(X,Y) / P(Y).

Exercise

Remark Variable X can have different distributions when conditioned on different Y.

Product Rule Infer joint probability from marginal probability and conditional probability.

Exercise

Chain Rule Split a joint probability of multiple variables into conditional probabilities of each.

Chain Rule Split a joint probability of multiple variables into conditional probabilities of each. Q: how to derive it?

Bayes s Rule Bayes s rule connects two conditional probabilities. - P(A) is prior of A, P(A B) is posterior of A - p(a=a B=b) = p(b=b A=a)*p(A=a) / p(b=b) [discrete] - p(a b) = p(b a)*p(a) / p(b) [continuous] Q: how to derive Bayes s Rule? Why Bayes s rule?

Example We randomly pick up a coin and flip it to get head (H) or tail (T). We know a priori that - 40% coins are fair (F), and 60% are unfair (U) - 80% chance we get a head (from an arbitrary coin) If we get head, what is the chance the flipped coin is fair? P(F H)?

Exercise

Independence Two variables X and Y are independent if P(X, Y) = P(X) P(Y) - or, equivalently P(X Y) = P(X) - p(x=x,y=y) = p(x=x)*p(y=y) [discrete] - p(x,y) = p(x)*p(y) [continuous]

Expectation Expectation of two random variables X,Y is E[g(X,Y)] = Σ(x,y) D g(x,y)*p(x,y) or E[g(X,Y)] = (x,y) D g(x,y)*p(x) dxdy

Covariance Covariance of two variables X, Y is Cov(X,Y) = E[(X-E[X])(Y - E[Y])] - if X=Y, then Cov(X,Y) = Var(X) = Var(Y) - Cov(X,Y) = E[XY] - E[X] E[Y] - Var[X+Y] = Var[X] + Var[Y] + Cov[X,Y] - if X, Y are independent, then Cov[X,Y] = 0 (uncorrelated)

Property of Covariance