Continuous Probability Distributions

Similar documents
Continuous Probability Distributions

Probability and Probability Distributions. Dr. Mohammed Alahmed

Dr. Junchao Xia Center of Biophysics and Computational Biology. Fall /13/2016 1/33

Chapter 7: Theoretical Probability Distributions Variable - Measured/Categorized characteristic

EXAM. Exam #1. Math 3342 Summer II, July 21, 2000 ANSWERS

STAT Chapter 5 Continuous Distributions

Covariance. Lecture 20: Covariance / Correlation & General Bivariate Normal. Covariance, cont. Properties of Covariance

The normal distribution

A Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.

Probability Distributions Columns (a) through (d)

ECON Fundamentals of Probability

This does not cover everything on the final. Look at the posted practice problems for other topics.

Sections 3.4 and 3.5

Sections 3.4 and 3.5

Discrete Random Variables

MATH 3510: PROBABILITY AND STATS June 15, 2011 MIDTERM EXAM

Discrete Probability Distributions

Part 3: Parametric Models

Continuous Distributions

Lecture 2: Review of Probability

Chapter 5 continued. Chapter 5 sections

3 Modeling Process Quality

Point and Interval Estimation II Bios 662

Joint Probability Distributions and Random Samples (Devore Chapter Five)

Class 8 Review Problems solutions, 18.05, Spring 2014

1 Exercises for lecture 1

Random Variables and Expectations

Chapter 5. Chapter 5 sections

Twelfth Problem Assignment

Part 3: Parametric Models

6 The normal distribution, the central limit theorem and random samples

ECON 5350 Class Notes Review of Probability and Distribution Theory

MAS113 Introduction to Probability and Statistics. Proofs of theorems

1 Probability Distributions

Chapter 5 Formulas Distribution Formula Characteristics n. π is the probability Function. x trial and n is the. where x = 0, 1, 2, number of trials

Expectation, Variance and Standard Deviation for Continuous Random Variables Class 6, Jeremy Orloff and Jonathan Bloom

Bandits, Experts, and Games

STAT/MA 416 Answers Homework 6 November 15, 2007 Solutions by Mark Daniel Ward PROBLEMS

Guidelines for Solving Probability Problems

Chapter 2. Continuous random variables

Introduction to Computational Finance and Financial Econometrics Probability Review - Part 2

Statistical Methods in Particle Physics

SDS 321: Introduction to Probability and Statistics

LECTURE 1. Introduction to Econometrics

Chapter 4: Continuous Random Variables and Probability Distributions

Discrete Probability distribution Discrete Probability distribution

a table or a graph or an equation.

Probability and Distributions

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

2. In a clinical trial of certain new treatment, we may be interested in the proportion of patients cured.

PROBABILITY DISTRIBUTION

7. Be able to prove Rules in Section 7.3, using only the Kolmogorov axioms.

Special distributions

Random variables, Expectation, Mean and Variance. Slides are adapted from STAT414 course at PennState

Lecture 2: CENTRAL LIMIT THEOREM

UQ, Semester 1, 2017, Companion to STAT2201/CIVL2530 Exam Formulae and Tables

Econ 371 Problem Set #1 Answer Sheet

Continuous Random Variables. What continuous random variables are and how to use them. I can give a definition of a continuous random variable.

Practice Questions for Final

Counting principles, including permutations and combinations.

Problem 1. Problem 2. Problem 3. Problem 4

ST 371 (IX): Theories of Sampling Distributions

Random Variable. Pr(X = a) = Pr(s)

success and failure independent from one trial to the next?

DEPARTMENT OF MATHEMATICS AND STATISTICS

Chapter 4 - Lecture 3 The Normal Distribution

Applied Statistics I

18 Bivariate normal distribution I

P (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n

Probability, Random Processes and Inference

MATH c UNIVERSITY OF LEEDS Examination for the Module MATH2715 (January 2015) STATISTICAL METHODS. Time allowed: 2 hours

Chapter 3 Probability Distributions and Statistics Section 3.1 Random Variables and Histograms

Review of probability and statistics 1 / 31

Chapter 15 Sampling Distribution Models

Random Variables and Their Distributions

15 Discrete Distributions

STAT 4385 Topic 01: Introduction & Review

Math 151. Rumbos Fall Solutions to Review Problems for Final Exam

Statistics 351 Probability I Fall 2006 (200630) Final Exam Solutions. θ α β Γ(α)Γ(β) (uv)α 1 (v uv) β 1 exp v }

1 Review of Probability

Math438 Actuarial Probability

Statistics STAT:5100 (22S:193), Fall Sample Final Exam B

EXAMINATIONS OF THE HONG KONG STATISTICAL SOCIETY GRADUATE DIPLOMA, Statistical Theory and Methods I. Time Allowed: Three Hours

Topic 4: Continuous random variables

Topic 4: Continuous random variables

Discrete Distributions

Special Discrete RV s. Then X = the number of successes is a binomial RV. X ~ Bin(n,p).

Homework 10 (due December 2, 2009)

Lecture 2: Repetition of probability theory and statistics

The mean, variance and covariance. (Chs 3.4.1, 3.4.2)

Probability Distributions for Continuous Variables. Probability Distributions for Continuous Variables

STA 2201/442 Assignment 2

Probability Density Functions

FINAL EXAM: Monday 8-10am

1.1 Review of Probability Theory

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Homework 9 (due November 24, 2009)

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

5.6 The Normal Distributions

Statistics, Data Analysis, and Simulation SS 2015

Transcription:

1

Chapter 5 Continuous Probability Distributions 5.1 Probability density function Example 5.1.1. Revisit Example 3.1.1. 11 12 13 14 15 16 21 22 23 24 25 26 S = 31 32 33 34 35 36 41 42 43 44 45 46 (5.1.1) 51 52 53 54 55 56 61 62 63 64 65 66 Pr(11) = Pr(12) =...= Pr(66) = 1/36. Chapter 5 2

If X=sum of the two numbers. Then X can take values 2, 3,..., 12 with probabilities 1/36, 2/36, 3/36, 4/36, 5/36, 6/36, 5/36, 4/36, 3/36, 2/36, and 1/36 respectively. Now suppose I asked our own Peter to pick a point on a 6 8 rectangle. What is the probability that he picks the point (2,1)? What is the probability that he picks the point (0,0)? What is the probability that he picks the point (2,1)? = 1 =0. What is the probability that he picks the point (0,0)? = 1 =0. Let me ask this now: What is the probability that Peter picks a point in the first half of the rectangle? Chapter 5 3

What is the probability that he picks a point in the middle third of the rectangle? What is the probability that Peter picks a point in the first half of the rectangle? Pr(a point is picked in the first half of the rectangle) = 1/2. What is the probability that he picks a point in the middle third of the rectangle? Pr(a point is picked in the middle third of the rectangle) = 1/3. Therefore, even though the probability of each sample point is zero, we can define probabilities of specific events. What assumption did we make in calculating the above probabilities? What if that assumption is violated? Chapter 5 4

Example 5.1.2. Menstrual cycle example(revisited) Histogram of Menstrual Cycle Relative Frequency 40 30 20 10 0 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 Time (days) Figure 5.1: Distribution of time intervals between successive menstrual periods (days) of college women (Table 2.3; Rosner; Page 13). Mean=28.5; Median=28; Mode=28. If we randomly pick a college woman, what is the probability that her menstrual cycle will be longer than 28 days? If we randomly pick a college woman, what is the probability Chapter 5 5

that her menstrual cycle will be between 28 and 32 days? Can we approximate these probabilities by using the smooth curve? Can we find a function (formula) that represents the smooth curve? Probability density function Probability density function (pdf) of a random variable X is a function (curve) such that the area under the density function between any two values a and b is equal to the probability that the random variable X falls between a and b. Probability density function satisfies: (i)it is non-negative, and (ii) total area under the curve is 1. Chapter 5 6

Cumulative distribution function The cumulative distribution function (cdf) of a random variable X evaluated at x is given by F (x) =Pr(X x). Cumulative distribution function F (x) satisfies: (i) F (x) is non-decreasing in x, and (ii) F ( ) =0, (iii) F ( ) =1, (iv) Pr(a <X b) =F (b) F (a). Mean and variance of a continuous random variable μ = E(X), or σ 2 = Var(X) =E(X 2 ) μ 2, Chapter 5 7

5.2 Normal Distribution Normal distribution is the most widely used continuous distribution. It is symmetric, bell-shaped. Between one standard deviations of mean lies about 68% of the area under the curve. Between two standard deviations of mean lies about 95% of the area under the curve. Between three standard deviations of mean lies about 99.7% of the area under the curve. Mean = Median = Mode. Has the density function f(x) = 1 σ 2π exp [ 1 (x μ)2 2σ2 ], <x<. (5.2.1) Only two parameters μ and σ 2 characterizes the whole distribution. Chapter 5 8

A normal distribution with parameters μ and σ 2 is denoted by N(μ, σ 2 ). If X N(μ, σ 2 ), then Pr(X μ) =Pr(X μ) = 1 2. The center (mean) of the distribution is μ and the variance of the distribution is σ 2. i.e., E(X) =μ, Var(X) =σ 2.. Standard normal distribution Standard normal distribution is a special case of normal distribution with mean μ = 0 and variance 1. Convention is to denote it by Z. If the random variable Z follows a standard normal distribution[n(0, 1)], then Pr( 1 Z 1) = 0.68, Chapter 5 9

Pr( 2 Z 2) = 0.95, Pr( 3 Z 3) = 0.997. pdf of Z, φ(z) = 1 2π exp ( z2 2 ), <Z<. (5.2.2) cdf of Z, Φ(z) =Pr(Z z) = z 1 2π exp ( z2 2 ) dz, <Z<. (5.2.3) Note that, Φ( z) =1 Φ(z). Use Table 3 in the Appendix of FOB to calculate the cdf for specific values of z. For instance, Pr(Z.84) = Φ(.84) = 0.7995, Pr(Z.84) = Φ(.84) = 1 Φ(0.84) = 1 0.7995 = 0.2005, Pr(.12 <Z.24) = Φ(.24) Φ(.12) =.5948 0.5478 = 0.0470, Chapter 5 10

Pr( 1.96 <Z.84) = Φ(.84) Φ( 1.96) =.2005 0.0250 = 0.1755, Pr( 1.96 <Z 1.96) = Φ(1.96) Φ( 1.96) =.9750 0.0250 = 0.9500, Φ(z) can also be calculated using SAS function CDF( normal,z,μ,σ).for example, data mynormal; /*Prob that Z is less than or equal 0 */ pzle0=cdf( normal,0,0,1); /*Pr(Z < 0.84)*/ pzlt_p84=cdf( normal,.84,0,1); /*Pr(.12 < Z <.24)*/ pzgt_p12_lt_p24=cdf( normal,.24,0,1)-cdf( normal,.12,0,1); run; proc print; run; Chapter 5 11

Percentiles of a standard normal distribution The 100pth percentile of a N(0, 1) random variable z p is given by Pr(Z z p )=Φ(z p )=p. We can calculate the percentiles of standard normal distributions from Table 3 (Appendix, FOB). To find the 95th percentile of standard normal distribution, we look for the probability 0.95 in Column A. Corresponding value of x is z.95, the 95th percentile. It is between 1.64 and 1.65, and we take it approximately to be z.95 =1.645. To find the 25th percentile of standard normal distribution, we first notice that (p, the area under the curve is less than 0.5). Therefore, corresponding percentile should be a negative number. What number should it be? Look for the probability 1-0.25 = 0.75 in Column A. The negative of the corresponding value of x is z.25, the 25th percentile. It is Chapter 5 12

between -.68 and -0.67, and we take it approximately to be z.25 =.675. Example SAS codes for standard normal percentiles: data normal_percentiles; /*Find 95th and 25th percentile of N(0,1)*/ Z95=quantile( NORMAL,.95); Z25=quantile( NORMAL,.25); run; proc print; run; 5.3 From Normal to Standard Normal Normal to Standard Normal If X N(μ, σ 2 ), then Z = X μ σ N(0, 1). This result is very important. A consequence is described in the following: Chapter 5 13

Calculating probabilities under normal curve If X N(μ, σ 2 )andz N(0, 1), then ) ) Pr(a <X b) =Pr =Φ ( a μ σ <Z b μ σ ( b μ σ Φ ( ) a μ σ. Another important consequence is that the quantiles (percentiles) of normal distributions can be expressed as a linear function of standard normal percentiles. Chapter 5 14

Calculating percentiles of normal curve If x p and z p denote the 100pth percentiles of N(μ, σ 2 )andn(0, 1), respectively, then, x p = μ + σz p. Example 5.3.1. Suppose that the diastolic blood pressures (DBP) for 35-to-44-year old men are normally distributed with mean 80mm Hg and variance 144mm Hg 2. If persons with DBP between 90 and 100mm Hg inclusive are considered to be moderately hypertensive, what is the probability that a randomly selected person from this population will be a mild hypertensive? Let X = DBP of a 35-to-44-year old man. Then, X N(80, 144). ( 90 80 Pr(90 X 100) = Pr X 80 12 12 = Pr(.883 Z 1.667) = Φ(1.667) Φ(.883) =.9522.7977 = 0.1545. ) 100 80 12 Chapter 5 15

If persons with DBP above 100mm Hg inclusive are considered to be hypertensive, what proportion of persons in this population are hypertensive? ( X 80 Pr(X >100) = Pr > 12 = Pr(Z >1.667) ) 100 80 12 = 1 Φ(1.667) = 1.9522 = 0.0478. Find the upper and lower 5th percentiles of this distribution. x.05 = μ + σz.05 = 80+12 ( 1.645) = 60.3mm Hg. x.95 = μ + σz.95 = 80+12 1.645 = 99.7mm Hg. Chapter 5 16

If a person s DBP belongs to the upper 25% of the distribution in this population he or she will be eligible for a clinical study related to hypertension. What is the minimum DBP level required to be eligible for the study? Find?? such that Pr(X >??) = 0.25. Or, Find x p such that Pr(X x.75 )=0.75. In other words, we need to calculate the 75th percentile of the distribution of X. We know that, x.75 = μ + σz.75 = 80+12 0.675 = 88.1mm Hg. If a 20 patients are selected at random, what is the probability Chapter 5 17

that exactly 5 of them will be hypertensive (DBP > 100mm Hg)? Let Y = # of hypertensives out of 20. Then Y is Binomial with n = 20, p =.0478 (From page 15) Pr(Y =5) = 20 (0.478) 5 (1 0.478) 1 5 5 =.0019. (5.3.1) If a 20 patients are selected at random, what is the probability that no more than 5 of them will be hypertensive (DBP > 100mm Hg)? 5 Pr(Y 5) = 20 (0.478) x (1 0.478) 20 x x x=0 =.9997. (5.3.2) Example SAS codes: data mynormal; Chapter 5 18

/*Pr(90 < X < 100) where X~N(80, 144)*/ pxgt90lt100=cdf( normal,100,80,12)-cdf( normal,90,80,12); /*Pr(X > 100) where X~N(80, 144)*/ pxgt100=1-cdf( normal,100,80,12); /*Pr(5 out of 20 will have X > 100) where X~N(80, 144)*/ p5outof20=pmf( binomial,5,pxgt100,20); plt5outof20=cdf( binomial,5,pxgt100,20); run; proc print; run; /*And for percentiles*/ data normal_percentiles; /*Find 95th and 5th percentile of N(80,144)*/ x95=quantile( NORMAL,.95,80,12); X25=quantile( NORMAL,.05,80,12); run; proc print; run; Chapter 5 19

Example 5.3.2. Continuity correction. Some random variables are usually measured only to the nearest integer. In such cases, while calculating probability that the random variable takes a value in an interval an adjustment for the continuity needs to be made. Suppose that the age distribution of a group of patients being treated for HIV infection is normal with mean 46 years and standard deviation 5 years. What is the probability that a randomly selected individual in this group will be 60 or above? ( ) 59.5 46 Pr(X 60) = 1 Φ 5 = 0.0035. (5.3.3) What is the probability that a randomly selected individual in this group will be below 60? ( ) 59.5 46 Pr(X <60) = Φ 5 = 0.9965. (5.3.4) Chapter 5 20

What is the probability that a randomly selected individual in this group will be above 50 but equal or below 60? Pr(50 <X 60) ( ) ( ) 60.5 46 50.5 46 = Φ Φ 5 5 = 0.1822. (5.3.5) Chapter 5 21

5.4 Covariance and Correlation 5.4.1 Covariance Covariance between two random variables measure how one variable varies relative to the other variable. Recall that the variance of a random variable X is defined as Var(X) =σx 2 = E { (X μ x ) 2} = E(X 2 ) μ 2 x. (5.4.1) If we have two random variables X (height of a person) and Y (weight of the same person), we might be interested in seeing how values of X are associated the values of Y. This leads to the covariance Cov(X) =σ xy = E {(X μ x )(Y μ x )} = E(XY ) μ x μ y. (5.4.2) A positive σ xy indicates that the larger the value of X, the larger the value of Y, indicating a positive relationship. On the other hand, if larger (smaller) values of X are accompanied by smaller (larger) Chapter 5 22

values of Y, the covariance is negative. 5.4.2 Correlation The covariance indicates the direction of the association. Since the covariance can be any value between and, it is usually useless if we want to measure the strength of the association between two random variables. The correlation coefficient is used to measure the strength as well as the direction of the correlation. It is defined as: ρ = Corr(X, Y )= σ xy σ x σ y. (5.4.3) Correlation coefficient ρ (i) 1 ρ 1, (ii) ρ = 0 No linear relationship, (ii) ρ = 1 Perfect negative linear relationship, (ii) ρ =+1 Perfect positive linear relationship. Chapter 5 23

5.4.3 Independence Two random variable X and Y are independent iff Pr(X x Y = y) =Pr(X x), for allx, y. (5.4.4) Given the knowledge about one variable does not change the distribution of the other. Is the distribution of weights for age-group 10-15 same as the distribution of weights for the age-group 30-35? (Age-group and Weight are not independent variables) Is there any reason to believe that the distribution of birthweights differ by the birth month? (Birthweight and birthmonth are independent random variables) Independence and correlation For independent random variables X and Y, (i) σ xy =0, and (ii) ρ =0. Chapter 5 24

On the other hand, if (i) σ xy =0, and (ii) ρ =0, X and Y may or may not be independent. 5.5 Linear combinations of random variables Frequently, we encounter situations where a new variable of interest is created by adding two or more random variables, subtracting one from the other, or taking linear combinations. Example 5.5.1. Let X 1,X 2,X 3,X 4, and X 5 are repeated cholesterol measurements over 5 different times during the day. We would like to know the average cholesterol level for that specific day. Thus the variable we are interested in is, X = X 1 + X 2 + X 3 + X 4 + X 5 5 = 1 5 X 1 + 1 5 X 2 + 1 5 X 3 + 1 5 X 4 + 1 5 X 5. (5.5.1) Chapter 5 25

Example 5.5.2. In the above example one might be interested in the difference between the fist and last cholesterol levels for that specific day. Thus the variable we are interested in is, D = X 5 X 1 =1 X 5 +( 1) X 1. (5.5.2) In general, we would like to know the distribution of a linear combination of random variables of the form Y = c 1 X 1 + c 2 X 2 +...+ c n X n, (5.5.3) where c 1,c 2,...,c n are scalars and X 1,X 2,...,X n are random variables. Expected value of a linear combination of random variables If Y as defined in (5.5.3), and μ j = E(X j ),j =1, 2,...,n,then μ y = E(Y )=c 1 μ 1 + c 2 μ 2 +...+ c n μ n. Thus, the expectation of sum of random variables is the sum of their expectations. That is, E(X 1 + X 2 +...+ X n )=E(X 1 )+E(X 2 )+...+ E(X n ). (5.5.4) Chapter 5 26

The expectation of the difference between two random variables is the difference of their expectations. That is, E(X 1 X 2 )=E(X 1 ) E(X 2 ). (5.5.5) Variance of the sum of two random variables If Y = X 1 + X 2,then σy 2 = Var(Y )=Var(X 1 )+Var(X 2 )+2Cov(X 1,X 2 ). Similarly, Var(X 1 X 2 )=Var(X 1 )+Var(X 2 ) 2Cov(X 1,X 2 ). (5.5.6) If X 1 and X 2 are independent, then Cov(X 1,X 2 ) = 0. And therefore, Var(X 1 + X 2 )=Var(X 1 X 2 )=Var(X 1 )+Var(X 2 ). (5.5.7) Chapter 5 27

Variance of the linear combination of independent random variables If X 1,X 2,...,X n are independent random variables, then Var(Y )=Var(c 1 X 1 + c 2 X 2 +...+ c n X n ) = c 2 1 Var(X 1)+c 2 2 Var(X 2)+...+ c 2 n Var(X n). Example 5.5.3. Suppose X 1 and X 2 respectively denote the preand post- treatment fasting triglyceride levels (mg/dl) for a randomly selected patient. If X 1 N(113, 225), and X 2 N(90, 196),what can we tell about the mean change? Let Y = X 2 X 1 be the change in the fasting triglyceride levels. Then, μ y = E(Y )=E(X 1 X 2 )=E(X 1 ) E(X 2 ) = 113 90 = 23mg/dl. 5.6 Normal approximation to Binomial distribution In many situations normal distribution arises from the sum of independent random variables. A binomial random variable Y with Chapter 5 28

parameter n and p can be treated as a sum of n independent random variables. It is easy to see, as each trial in the binomial experiment can be treated as independent binomial experiment. Let X 1 be the number of success in the 1st trial, X 2 be the number of success in the 2nd trial, and so on. Then, X i,i =1, 2,...,n can take value either 0 (failure) or 1 (success). Each of these variables are referred to as Bernoulli random variables. Theirsum Y = X 1 + X 2 +...+ X n represents the number of successes in n trials. Thus a binomial random variable is the sum of independent Bernoulli random variables. Normal approximation to Binomial distribution If n is large and p is small so that np(1 p) 5, then B(n, p) distribution can be well-approximated by a N(μ = np, σ 2 = npq) distribution. Chapter 5 29

Example 5.6.1. What is the probability that the number of neutrophils will be between 50 and 75 out of 100 white blood cells where the probability that any one cell is a neutrophil is.6? Let X=number of neutrophil out of 100 white blood cells. Then X B(100,.6). Pr(50 X 75) = 75 x=50 100.6 x (1.6) 100 x x = 0.9826. (5.6.1) Now, as np(1 p) =24> 5, this should be well approximated by the N(μ = np =60,σ 2 = npq = 24) distribution: Pr(50 X 75) Φ ( ) ( ) 75.5 60 49.5 60 Φ 24 24 = Φ(3.164) Φ( 2.143) = 0.983. (5.6.2) Chapter 5 30

5.7 Normal approximation to the Poisson distribution Normal approximation to Poisson distribution If μ is large (μ 10), then P (μ) distribution can be well-approximated by a N(μ, σ 2 = μ) distribution. Example 5.7.1. Suppose the number of accidents on the highland park bridge over a year follows a Poisson distribution with mean 20. What is the probability that there will be more than 5 accidents over a 6-month period? Using Poisson distribution with μ = 10, Pr(X >5) = 1 Pr(X 5) = 0.9329. (5.7.1) Using Normal distribution with μ =10andσ 2 = 10, ( ) 5.5 10 Pr(X >5) 1 Φ =0.923. (5.7.2) 10 SOLVE PROBLEMS 5.99-5.101, 5.102-104. Chapter 5 31