Lecture 2. October 21, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

Similar documents
Probability measures A probability measure, P, is a real valued function from the collection of possible events so that the following

Stat 704 Data Analysis I Probability Review

EXAM. Exam #1. Math 3342 Summer II, July 21, 2000 ANSWERS

Deep Learning for Computer Vision

LECTURE 1. 1 Introduction. 1.1 Sample spaces and events

Probability Theory Review

DISCRETE RANDOM VARIABLES: PMF s & CDF s [DEVORE 3.2]

A brief review of basics of probabilities

Lecture 3: Random variables, distributions, and transformations

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016

Lecture 2: Repetition of probability theory and statistics

Statistics and Econometrics I

Recitation 2: Probability

Chapter 3. Chapter 3 sections

Lecture 1: Probability Fundamentals

Review of probabilities

STAT 516: Basic Probability and its Applications

EE514A Information Theory I Fall 2013

CMPSCI 240: Reasoning about Uncertainty

M378K In-Class Assignment #1

Random Variables. Definition: A random variable (r.v.) X on the probability space (Ω, F, P) is a mapping

Dynamic Programming Lecture #4

Chapter 2.5 Random Variables and Probability The Modern View (cont.)

Continuous Distributions

Algorithms for Uncertainty Quantification

Econ 325: Introduction to Empirical Economics

Lecture 1. ABC of Probability

Lecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019

Dept. of Linguistics, Indiana University Fall 2015

Random Variables. Saravanan Vijayakumaran Department of Electrical Engineering Indian Institute of Technology Bombay

Review of Probability. CS1538: Introduction to Simulations

Chapter 1. Sets and probability. 1.3 Probability space

Venn Diagrams; Probability Laws. Notes. Set Operations and Relations. Venn Diagram 2.1. Venn Diagrams; Probability Laws. Notes

Lecture 8. October 22, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

Conditional Probability

Introduction to Probability. Ariel Yadin. Lecture 1. We begin with an example [this is known as Bertrand s paradox]. *** Nov.

Probability Review. Chao Lan

Lecture 3. Discrete Random Variables

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Random Variables Example:

HW MATH425/525 Lecture Notes 1

Probability and Probability Distributions. Dr. Mohammed Alahmed

Example A. Define X = number of heads in ten tosses of a coin. What are the values that X may assume?

Chapter 3 : Conditional Probability and Independence

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Math 151. Rumbos Fall Solutions to Review Problems for Exam 2. Pr(X = 1) = ) = Pr(X = 2) = Pr(X = 3) = p X. (k) =

A Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.

Why study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables

Chapter 2. Probability

Probability Review. Yutian Li. January 18, Stanford University. Yutian Li (Stanford University) Probability Review January 18, / 27

Chapter 2. Probability. Math 371. University of Hawai i at Mānoa. Summer 2011

3. Review of Probability and Statistics

Properties of Probability

Calculus with Algebra and Trigonometry II Lecture 21 Probability applications

CSE 312: Foundations of Computing II Random Variables, Linearity of Expectation 4 Solutions

MATH MW Elementary Probability Course Notes Part I: Models and Counting

Recap of Basic Probability Theory

Mathematical Foundations of Computer Science Lecture Outline October 18, 2018

Random variables. DS GA 1002 Probability and Statistics for Data Science.

Statistical Inference

Continuous random variables

What does independence look like?

Continuous Random Variables

Probability Theory for Machine Learning. Chris Cremer September 2015

L2: Review of probability and statistics

Independence Solutions STAT-UB.0103 Statistics for Business Control and Regression Models

Lecture notes for probability. Math 124

Probability theory for Networks (Part 1) CS 249B: Science of Networks Week 02: Monday, 02/04/08 Daniel Bilar Wellesley College Spring 2008

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

The enumeration of all possible outcomes of an experiment is called the sample space, denoted S. E.g.: S={head, tail}

Relationship between probability set function and random variable - 2 -

Chapter 2. Continuous random variables

2. Variance and Covariance: We will now derive some classic properties of variance and covariance. Assume real-valued random variables X and Y.

Conditional Probability (cont'd)

At least 8/9 of th measurements lie in the interval

Recap of Basic Probability Theory

Probability Theory and Random Variables

Lecture 2: Review of Basic Probability Theory

Probability theory basics

4.2 Probability Models

Probability: Axioms, Properties, Interpretations

I - Probability. What is Probability? the chance of an event occuring. 1classical probability. 2empirical probability. 3subjective probability

Design of Engineering Experiments

Sample Spaces, Random Variables

Origins of Probability Theory

An introduction to biostatistics: part 1

SDS 321: Introduction to Probability and Statistics

Chapter 2 Random Variables

Brief Review of Probability

Single Maths B: Introduction to Probability

Probability and Statisitcs

Topic 5 Basics of Probability

Announcements. Topics: To Do:

STAT 430/510 Probability

CMPSCI 240: Reasoning Under Uncertainty

Chapter 1 Probability Theory

MAT 271E Probability and Statistics

Brief Review of Probability

CMPSCI 240: Reasoning Under Uncertainty

Transcription:

Lecture 2 Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University October 21, 2007

1 2 3 4 5 6

Define probability calculus Basic axioms of probability Define random Define density and mass functions Define cumulative distribution survivor functions Define, percentiles, medians

measures A probability measure, P, is a real valued function from the collection of possible events so that the following hold 1. For an event E Ω, 0 P(E) 1 2. P(Ω) = 1 3. If E 1 and E 2 are mutually exclusive events P(E 1 E 2 ) = P(E 1 ) + P(E 2 ).

Part 3 of the definition implies finite additivity P( n i=1a i ) = n P(A i ) i=1 Additivity where the {A i } are mutually exclusive. This is usually extended to countable additivity P( i=1a i ) = P(A i ) i=1

Note P is defined on F a collection of subsets of Ω Example Ω = {1, 2, 3} then F = {, {1}, {2}, {3}, {1, 2}, {1, 3}, {2, 3}, {1, 2, 3}}. When Ω is a continuous set, the definition gets much trickier. In this case we assume that F is sufficiently rich so that any set that we re interested in will be in it.

Consequences You should be able to prove all of the following: P( ) = 0 P(E) = 1 P(E c ) P(A B) = P(A) + P(B) P(A B) if A B then P(A) P(B) P (A B) = 1 P(A c B c ) P(A B c ) = P(A) P(A B) P( n i=1 E i) n i=1 P(E i) P( n i=1 E i) max i P(E i )

Example Proof that P(E) = 1 P(E c ) 1 = P(Ω) = P(E E c ) = P(E) + P(E c )

Proof that P( n i=1 E i) n i=1 P(E i) Example P(E 1 E 2 ) = P(E 1 ) + P(E 2 ) P(E 1 E 2 ) P(E 1 ) + P(E 2 ) Assume the statement is true for n 1 and consider n P( n i=1e i ) P(E n ) + P( n 1 i=1 E i) n 1 P(E n ) + P(E i ) = n P(E i ) i=1 i=1

Example The National Sleep Foundation (www.sleepfoundation.org) reports that around 3% of the American population has sleep apnea. They also report that around 10% of the North American and European population has restless leg syndrome. Similarly, they report that 58% of adults in the US experience insomnia. Does this imply that 71% of people will have at least one sleep problems of these sorts?

Example continued Answer: No, the events are not mutually exclusive. To elaborate let: A 1 = {Person has sleep apnea} A 2 = {Person has RLS} A 3 = {Person has insomnia} Then (work out the details for yourself) P(A 1 A 2 A 3 ) = P(A 1 ) + P(A 2 ) + P(A 3 ) P(A 1 A 2 ) P(A 1 A 3 ) P(A 2 A 3 ) + P(A 1 A 2 A 3 ) =.71 + Other stuff where the Other stuff has to be less than 0

Example : LA Times from Rice page 26

A random variable is a numerical outcome of an experiment. The random that we study will come in two varieties, discrete or continuous. Discrete random variable are random that take on only a countable number of possibilities. P(X = k) Continuous random variable can take any value on the real line or some subset of the real line. P(X A)

Examples of random The (0 1) outcome of the flip of a coin The outcome from the roll of a die The BMI of a subject four years after a baseline measurement The hypertension status of a subject randomly drawn from a population

PMF A probability mass function evaluated at a value corresponds to the probability that a random variable takes that value. To be a valid pmf a function, p, must satisfy 1 p(x) 0 for all x 2 x p(x) = 1 The sum is taken over all of the possible values for x.

Example Let X be the result of a coin flip where X = 0 represents tails and X = 1 represents heads. p(x) = (1/2) x (1/2) 1 x for x = 0, 1 Suppose that we do not know whether or not the coin is fair; Let θ be the probability of a head p(x) = θ x (1 θ) 1 x for x = 0, 1

PDF A probability density function (pdf), is a function associated with a continuous random variable Areas under pdfs correspond to probabilities for that random variable To be a valid pdf, a function f must satisfy 1 f (x) 0 for all x 2 f (x)dx = 1

Example Assume that the time in years from diagnosis until death of persons with a specific kind of cancer follows a density like { e x/5 f (x) = 5 for x > 0 0 otherwise More compactly written: f (x) = 1 5 e x/5 for x > 0. Is this a valid density? 1 e raised to any power is always positive 2 0 f (x)dx = 0 e x/5 /5dx = e x/5 0 = 1

Example continued What s the probability that a randomly selected person from this distribution survives more than 6 years? P(X 6) = e t/5 6 5 dt = e t/5 6 = e 6/5.301. Approximation in R pexp(6, 1/5, lower.tail = FALSE)

density 0.00 0.05 0.10 0.15 0.20 Example continued 0 5 10 15 20 Survival time in years

CDF and survival function The cumulative distribution function (CDF) of a random variable X is defined as the function F (x) = P(X x) This definition applies regardless of whether X is discrete or continuous. The survival function of a random variable X is defined as S(x) = P(X > x) Notice that S(x) = 1 F (x) For continuous random, the PDF is the derivative of the CDF

Example What are the survival function and CDF from the exponential density considered before? S(x) = hence we know that e t/5 x 5 dt = e t/5 x = e x/5 F (x) = 1 S(x) = 1 e x/5 Notice that we can recover the PDF by f (x) = F (x) = d dx (1 e x/5 ) = e x/5 /5

Quantiles The α th quantile of a distribution with distribution function F is the point x α so that F (x α ) = α A percentile is simply a quantile with α expressed as a percent The median is the 50 th percentile

Example What is the 25 th percentile of the exponential survival distribution considered before? We want to solve (for x).25 = F (x) = 1 e x/5 resulting in the solution x = log(.75) 5 1.44 Therefore, 25% of the subjects from this population live less than 1.44 years R can approximate exponential for you qexp(.25, 1/5)