Probability measures A probability measure, P, is a real valued function from the collection of possible events so that the following

Similar documents
Lecture 2. October 21, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

Outline. 1. Define likelihood 2. Interpretations of likelihoods 3. Likelihood plots 4. Maximum likelihood 5. Likelihood ratio benchmarks

Simple Linear Regression. John McGready Johns Hopkins University

Sampling Variability and Confidence Intervals. John McGready Johns Hopkins University

Lecture 21. December 19, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

1 Hypothesis testing for a single mean

Lecture 27. December 13, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

Stat 704 Data Analysis I Probability Review

EXAM. Exam #1. Math 3342 Summer II, July 21, 2000 ANSWERS

Section B. The Theoretical Sampling Distribution of the Sample Mean and Its Estimate Based on a Single Sample

Statistics and Econometrics I

Lecture 1: Probability Fundamentals

Probability Pr(A) 0, for any event A. 2. Pr(S) = 1, for the sample space S. 3. If A and B are mutually exclusive, Pr(A or B) = Pr(A) + Pr(B).

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016

Recitation 2: Probability

Probability Theory Review

DISCRETE RANDOM VARIABLES: PMF s & CDF s [DEVORE 3.2]

A brief review of basics of probabilities

Lecture 3: Random variables, distributions, and transformations

Random Variables. Definition: A random variable (r.v.) X on the probability space (Ω, F, P) is a mapping

Probability theory for Networks (Part 1) CS 249B: Science of Networks Week 02: Monday, 02/04/08 Daniel Bilar Wellesley College Spring 2008

LECTURE 1. 1 Introduction. 1.1 Sample spaces and events

Chapter 3. Chapter 3 sections

Independence Solutions STAT-UB.0103 Statistics for Business Control and Regression Models

Econ 325: Introduction to Empirical Economics

Review of probabilities

Properties of Probability

STAT 516: Basic Probability and its Applications

Review of Probability. CS1538: Introduction to Simulations

Lecture 3. Discrete Random Variables

Chapter 1. Sets and probability. 1.3 Probability space

Deep Learning for Computer Vision

Random Variables Example:

Dept. of Linguistics, Indiana University Fall 2015

CMPSCI 240: Reasoning about Uncertainty

Example. Test for a proportion

Lecture 2: Repetition of probability theory and statistics

M378K In-Class Assignment #1

MATH 3510: PROBABILITY AND STATS July 1, 2011 FINAL EXAM

Continuous Distributions

Example A. Define X = number of heads in ten tosses of a coin. What are the values that X may assume?

Dynamic Programming Lecture #4

Chapter 2.5 Random Variables and Probability The Modern View (cont.)

Math 151. Rumbos Fall Solutions to Review Problems for Exam 2. Pr(X = 1) = ) = Pr(X = 2) = Pr(X = 3) = p X. (k) =

MAS108 Probability I

Lecture 1. ABC of Probability

Relationship between probability set function and random variable - 2 -

Recap of Basic Probability Theory

EE514A Information Theory I Fall 2013

Probability and Probability Distributions. Dr. Mohammed Alahmed

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

Random variables. DS GA 1002 Probability and Statistics for Data Science.

Recap of Basic Probability Theory

Lecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019

Lecture 2: Review of Basic Probability Theory

Venn Diagrams; Probability Laws. Notes. Set Operations and Relations. Venn Diagram 2.1. Venn Diagrams; Probability Laws. Notes

What does independence look like?

Random Variables. Saravanan Vijayakumaran Department of Electrical Engineering Indian Institute of Technology Bombay

Conditional Probability

3. Review of Probability and Statistics

At least 8/9 of th measurements lie in the interval

Probability Review. Chao Lan

Introduction to Probability. Ariel Yadin. Lecture 1. We begin with an example [this is known as Bertrand s paradox]. *** Nov.

Copyright c 2006 Jason Underdown Some rights reserved. choose notation. n distinct items divided into r distinct groups.

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Algorithms for Uncertainty Quantification

HW MATH425/525 Lecture Notes 1

Toxicokinetics: Absorption, Distribution, and Excretion. Michael A. Trush, PhD Johns Hopkins University

Chapter 3 : Conditional Probability and Independence

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

Lecture 23. November 15, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

Why study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables

Probability Theory and Random Variables

Chapter 2. Continuous random variables

Calculus with Algebra and Trigonometry II Lecture 21 Probability applications

Chapter 2. Probability. Math 371. University of Hawai i at Mānoa. Summer 2011

Chapter 2. Probability

MATH MW Elementary Probability Course Notes Part I: Models and Counting

Announcements. Topics: To Do:

A Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.

Review of Basic Probability Theory

Statistical Inference

Single Maths B: Introduction to Probability

Conditional Probability

STAT 430/510 Probability Lecture 7: Random Variable and Expectation

Statistics 100A Homework 5 Solutions

Probability Theory for Machine Learning. Chris Cremer September 2015

STAT Chapter 3: Probability

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

L2: Review of probability and statistics

MATH Solutions to Probability Exercises

Lecture notes for probability. Math 124

Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology

The enumeration of all possible outcomes of an experiment is called the sample space, denoted S. E.g.: S={head, tail}

1 Probability and Random Variables

Probability theory basics

2. Variance and Covariance: We will now derive some classic properties of variance and covariance. Assume real-valued random variables X and Y.

Conditional Probability (cont'd)

Example. What is the sample space for flipping a fair coin? Rolling a 6-sided die? Find the event E where E = {x x has exactly one head}

Transcription:

This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this site. Copyright 2006, The Johns Hopkins University and Brian Caffo. All rights reserved. Use of these materials permitted only in accordance with license rights granted. Materials provided AS IS ; no representations or warranties provided. User assumes all responsibility for use, and all liability related thereto, and must independently review all materials for accuracy and efficacy. May contain materials owned by others. User is responsible for obtaining permissions for use from third parties as needed.

Lecture 2 1. Define probability calculus 2. Basic axioms of probability 3. Define random variables 4. Define density and mass functions 5. Define cumulative distribution functions and survivor functions 6. Define quantiles, percentiles, medians

Probability measures A probability measure, P, is a real valued function from the collection of possible events so that the following hold 1. For an event E Ω, 0 P(E) 1 2. P(Ω) = 1 3. If E 1 and E 2 are mutully exclusive events P(E 1 E 2 ) = P(E 1 ) + P(E 2 ).

Additivity Part 3 of the definition implies finite additivity P( n i=1 A i) = n P(A i ) i=1 where the {A i } are mutually exclusive. This is usually extended to countable additivity P( i=1 A i) = P(A i ) i=1

Note P is defined on F a collection of subsets of Ω Example Ω = {1, 2, 3} then F = {, {1}, {2}, {3}, {1, 2}, {1, 3}, {2, 3}, {1, 2, 3}}. When Ω is a continuous set, the definition gets much trickier. In this case we assume that F is sufficiently rich so that any set that we re interested in will be in it.

Consequences You should be able to prove all of the following: P( ) = 0 P(E) = 1 P(E c ) P(A B) = P(A) + P(B) P(A B) if A B then P(A) P(B) P (A B) = 1 P(A c B c ) P(A B c ) = P(A) P(A B) P( n i=1 E i) n i=1 P(E i ) P( n i=1 E i) max i P(E i )

Example Proof that P(E) = 1 P(E c ) 1 = P(Ω) = P(E E c ) = P(E) + P(E c )

Example Proof that P( n i=1 E i) n i=1 P(E i ). P(E 1 E 2 ) = P(E 1 ) + P(E 2 ) P(E 1 E 2 ) P(E 1 ) + P(E 2 ). Assume the statement is true for n 1 and consider n. P( n i=1 E i) P(E n ) + P( i=1 n 1 E i) P(E n ) + = n P(E i ). i=1 n 1 i=1 P(E i )

Example The National Sleep Foundation (www.sleepfoundation.org) reports that around 3% of the American population has sleep apnea. They also report that around 10% of the North American and European population has restless leg syndrome. Similarly, they report that 58% of adults in the US experience insomnia. Does this imply that 71% of people will have at least one sleep problems of these sorts?

Random variables A random variable is a numerical outcome of an experiment. The random variables that we study will come in two varieties, discrete or continous. Discrete random variable are random variables that take on only a countable number of possiblities. P(X = k) Continous random variable can take any value on the real line or some subset of the real line. P(X A)

Examples of random variables The (0 1) outcome of the flip of a coin The outcome from the roll of a die The BMI of a subject four years after a basline measurment The hypertension status of a subject randomly drawn from a population

PMF A probability mass function evaluated at a value corresponds to the probability that a random variable takes that value. To be a valid pmf a function, p, must satisfy 1. p(x) 0 for all x 2. x p(x) = 1 The sum is taken over all of the possible values for x.

Example Let X be the result of a coin flip where X = 0 represents tails and X = 1 represents heads. p(x) = (1/2) x (1/2) 1 x for x = 0, 1 Suppose that we do not know whether or not the coin is fair. Let θ be the probability of a head p(x) = θ x (1 θ) 1 x for x = 0, 1

PDF A probability density function (pdf), is a function associated with a continuous random variable Areas under pdfs corresond to probabilities for that random variable To be a valid pdf, a function f must satisfy 1. f(x) 0 for all x 2. f(x)dx = 1

Example Assume that the time in years from diagnosis until death of persons with a specific kind of cancer follows a density like { e x/5 f(x) = 5 for x > 0 0 otherwise Shorthand f(x) = e x/5 /5 for x > 0. Is this a valid density? 1. e raised to any power is always positive 2. 0 f(x)dx = 0 e x/5 /5dx = e x/5 0 = 1

What s the probability that a randomly selected person from this distribution survives more than 6 years? P(X 6) = 6 e t/5 5 dt = e t/5 6 = e 6/5.301. Approximation in R pexp(6, 1/5, lower.tail = FALSE) density 0.00 0.05 0.10 0.15 0.20 0 5 10 15 20 Survival time in years

CDF and survival function The cumulative distribution function (CDF) of a random variable X is defined as the function F(x) = P(X x) This definition applies regardless of whether X is discrete or continuous. The survival function of a random variable X is defined as Notice that S(x) = 1 F(X) S(x) = P(X > x) For continuous random variables, the PDF is the derivative of the CDF

Example What are the survival function and CDF from the exponential density considered before? S(x) = hence we know that x e t/5 5 dt = e t/5 F(x) = 1 S(x) = 1 e x/5 Notice that we can recover the PDF by x = e x/5 f(x) = F (x) = d dx (1 e x/5 ) = e x/5 /5

Quantiles The α th quantile of a distribution with distribution function F is the point x α so that F(x α ) = α A percentile is simply a quantile with α expressed as a percent The median is the 50 th percentile

Example What is the 25 th percentile of the exponential survival distribution considered before? We want to solve (for x).25 = F(x) = 1 e x/5 resulting in the solution x = log(.75)5 1.44 Therefore, 25% of the subjects from this population live less than 1.44 years R can approximate exponential quantiles for you qexp(.25, 1/5)