Topic 4: Continuous random variables

Similar documents
Topic 4: Continuous random variables

Continuous random variables

Continuous Distributions

MATH Solutions to Probability Exercises

Continuous Random Variables and Continuous Distributions

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

Continuous Random Variables

Continuous random variables

Let X be a continuous random variable, < X < f(x) is the so called probability density function (pdf) if

CHAPTER 6 SOME CONTINUOUS PROBABILITY DISTRIBUTIONS. 6.2 Normal Distribution. 6.1 Continuous Uniform Distribution

Review of Statistics I

Probability Density Functions

Random Variables and Their Distributions

Chapter 5 continued. Chapter 5 sections

3 Continuous Random Variables

1 Review of Probability

Continuous Distributions

Continuous Random Variables

Statistics 3657 : Moment Generating Functions

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

Probability and Distributions

Probability, Random Processes and Inference

Continuous distributions

STAT 509 Section 3.4: Continuous Distributions. Probability distributions are used a bit differently for continuous r.v. s than for discrete r.v. s.

Probability Midterm Exam 2:15-3:30 pm Thursday, 21 October 1999

1 Review of Probability and Distributions

Gamma and Normal Distribuions

Arkansas Tech University MATH 3513: Applied Statistics I Dr. Marcel B. Finan

Review 1: STAT Mark Carpenter, Ph.D. Professor of Statistics Department of Mathematics and Statistics. August 25, 2015

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

3. Probability and Statistics

15 Discrete Distributions

1 Probability and Random Variables

5.2 Continuous random variables

Continuous Random Variables and Probability Distributions

p. 6-1 Continuous Random Variables p. 6-2

Chapter 3. Julian Chan. June 29, 2012

Brief Review of Probability

Probability Distributions for Continuous Variables. Probability Distributions for Continuous Variables

LIST OF FORMULAS FOR STK1100 AND STK1110

Lecture 17: The Exponential and Some Related Distributions

Math 105 Course Outline

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

Slides 8: Statistical Models in Simulation

continuous random variables

1 Solution to Problem 2.1

Expected Values, Exponential and Gamma Distributions

5.6 The Normal Distributions

Chapter 4 Continuous Random Variables and Probability Distributions

STAT 430/510 Probability Lecture 12: Central Limit Theorem and Exponential Distribution

Random Variables. Saravanan Vijayakumaran Department of Electrical Engineering Indian Institute of Technology Bombay

Stat410 Probability and Statistics II (F16)

Definition: A random variable X is a real valued function that maps a sample space S into the space of real numbers R. X : S R

ORF 245 Fundamentals of Statistics Chapter 4 Great Expectations

Preliminary Statistics Lecture 3: Probability Models and Distributions (Outline) prelimsoas.webs.com

RMSC 2001 Introduction to Risk Management

Will Landau. Feb 28, 2013

STAT 430/510: Lecture 10

4.1 Expected value of a discrete random variable. Suppose we shall win $i if the die lands on i. What is our ideal average winnings per toss?

Notes on Continuous Random Variables

Chapter 2. Continuous random variables

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

Chapter 7: Theoretical Probability Distributions Variable - Measured/Categorized characteristic

Exponential, Gamma and Normal Distribuions

Chapter 4: Continuous Random Variables and Probability Distributions

Experimental Design and Statistics - AGA47A

STAT Chapter 5 Continuous Distributions

Exponential & Gamma Distributions

MATH 3510: PROBABILITY AND STATS July 1, 2011 FINAL EXAM

HW7 Solutions. f(x) = 0 otherwise. 0 otherwise. The density function looks like this: = 20 if x [10, 90) if x [90, 100]

Expectation, Variance and Standard Deviation for Continuous Random Variables Class 6, Jeremy Orloff and Jonathan Bloom

This does not cover everything on the final. Look at the posted practice problems for other topics.

CSE 312, 2017 Winter, W.L. Ruzzo. 7. continuous random variables

RMSC 2001 Introduction to Risk Management

Final Solutions Fri, June 8

Continuous Distributions

Chapter 5. Chapter 5 sections

ECON 5350 Class Notes Review of Probability and Distribution Theory

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

S6880 #7. Generate Non-uniform Random Number #1

Common ontinuous random variables

STT 441 Final Exam Fall 2013

Random Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline.

Advanced Herd Management Probabilities and distributions

Math 576: Quantitative Risk Management

Final Exam # 3. Sta 230: Probability. December 16, 2012

Lecture 10: Normal RV. Lisa Yan July 18, 2018

Lesson 100: The Normal Distribution. HL Math - Santowski

MAS3301 / MAS8311 Biostatistics Part II: Survival

Lecture 4. Continuous Random Variables and Transformations of Random Variables

ISyE 3044 Fall 2015 Test #1 Solutions

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 8 10/1/2008 CONTINUOUS RANDOM VARIABLES

Chapter 4: Continuous Probability Distributions

ECON Fundamentals of Probability

Continuous random variables and probability distributions

Definition 1.1 (Parametric family of distributions) A parametric distribution is a set of distribution functions, each of which is determined by speci

Continuous Random Variables

Twelfth Problem Assignment

Definition A random variable X is said to have a Weibull distribution with parameters α and β (α > 0, β > 0) if the pdf of X is

Chapter 4. Continuous Random Variables

Transcription:

Topic 4: Continuous random variables Course 003, 2018 Page 0

Continuous random variables Definition (Continuous random variable): An r.v. X has a continuous distribution if there exists a non-negative function f defined on the real line such that for any interval A, P(X A) = f(x)dx The function f is called the probability density function (PDF) of X. Every PDF must satisfy: A 1. Nonnegative: f(x) 0 2. Integrates to 1: The CDF of X is given by: f(x)dx = 1 x F(x) = f(t)dt A continuous random variable is a random variable with a continuous distribution. Page 1

Continuous r.v.s: caveats and remarks 1. The density is not a probability and it is possible to have f(x) > 1. 2. P(X = x) = 0 for all x. We can compute a probability for X being very close to x by integrating f over an ɛ interval around x. 3. Any function satisfying the properties of the PDF, represents the density of some r.v. 4. For defining probabilities, it doesn t matter whether we include or exclude endpoints: P(a < X < b) = P(a < X b) = P(a X < b) = P(a X b) = b f(x)dx a Page 2

Expectation of a continuous r.v. Definition (Expectation of a continuous random variable): The expected value or mean of a continuous r.v. X with PDF f is E(X) = xf(x)dx Result (Expectation of a function of X): If X is a continuous r.v. X with PDF f and g is a function from R to R, then E(g(X)) = g(x)f(x)dx Page 3

The Logistic distribution The CDF of a logistic distribution is F(x) = 1+e ex x, x R. We differentiate this to get the PDF, f(x) = e x (1+e x ) 2, x R This is a similar to the normal distribution but has a closed-form CDF and is computationally easier. For example: P( 2 < X < 2) = F(2) F( 2) = e2 1 1+e 2 =.76 Page 4

The Uniform distribution Definition (Uniform distribution): A continuous r.v. U has a Uniform distribution on the interval (a, b) if its PDF is 1 f(x; a, b) = b a I (a,b)(x) We denote this by U Unif(a, b) We often write a density in this manner- using the indicator variable to describe its support. This is a valid PDF since the area under the curve is the area of a rectangle with length (b a) 1 and height (b a). The CDF is the accumulated area under the PDF: Graph the PDF and CDF of Unif(0, 1). 0 if x a F(x) = x a b a if a < x b 1 if x b Suppose we have observations from Unif(0, 1). We can transform these into a sample from Ũ Unif(a, b): Ũ = a + (b a)u Page 5

The Probability Integral Transformation Result: Let X be a continuous random variable with the distribution function F and let Y = F(X). Then Y must be uniformly distributed on [0, 1]. The transformation from X to Y is called the probability integral transformation. We know that the distribution function must take values between 0 and 1. If we pick any of these values, y, the y th quantile of the distribution of X will be given by some number x and Pr(Y y) = Pr(X x) = F(x) = y which is the distribution function of a uniform random variable. This result helps us generate random numbers from various distributions, because it allows us to transform a sample from a uniform distribution into a sample from some other distribution provided we can find F 1. Example: Given U Unif(0, 1), log ( U 1 U ) Logistic. (F(x) = 1+e ex x, set this equal to u and solve for x) Page 6

The Normal distribution This symmetric bell-shaped density is widely used because: 1. It captures many types of natural variation quite well: heights-humans, animals and plants, weights, strength of physical materials, the distance from the centre of a target. 2. It has nice mathematical properties: many functions of a set normally distributed random variables have distributions that take simple forms. 3. Central limit theorems are fundamental to statistic inference. The sample mean of a large random sample from any distribution with finite variance is approximately normal. Page 7

The Standard Normal distribution Definition (Standard Normal distribution): A continuous r.v. Z is said to have a standard Normal distribution if its PDF ϕ is given by: ϕ(z) = 1 2π e z2 /2, < z < The CDF is the accumulated area under the PDF: Φ(z) = z ϕ(t)dt = z 1 2π e t2 /2 dt E(Z) = 0, Var(Z) = 1 and X = µ + σz is has a Normal distribution with mean µ and variance σ 2. ( E(µ + σz) = E(µ) + σe(z) = µ and Var(µ + σz) = σ 2 Var(Z) = σ 2 ) Page 8

Approximations of Normal probabilities If X N(µ, σ 2 ) then P( X µ < σ) 0.68 P( X µ < 2σ) 0.95 P( X µ < 3σ) 0.997 Or stated in terms of z: P( Z < 1) 0.68 P( Z < 2) 0.95 P( Z < 3) 0.997 The Normal distribution therefore has very thin tails relative to other commonly used distributions that are symmetric about their mean (logistic, Student s-t, Cauchy). Page 9

The Gamma distribution The gamma function of α is defined as Γ(α) = 0 y α 1 e y dy (1) Definition (Gamma distribution): It is said that a random variable X has a Gamma distribution with parameters α and β (α > 0 and β > 0) if X has a continuous distribution for which the PDF is: f(x; α, β) = βα Γ(α) xα 1 e βx I (0, ) (x) To see that this is a valid PDF: Substitute for y in the expression Γ(α) by y = βx (so dy = βdx ) or Γ(α) = 1 = 0 (xβ) α 1 e xβ βdx 0 β α Γ(α) xα 1 e βx dx E(X) = α β and Var(X) = α β 2 Page 10

Features of the Gamma density The density is decreasing for α 1, and hump-shaped and skewed to the right otherwise. It attains a maximum at x = α 1 β Higher values of α make it more bell-shaped and higher β increases the concentration at lower x values. Page 11

The Exponential distribution We have a special name for a Gamma distribution with α = 1: Definition (Exponential distribution): A continuous random variable X has an exponential distribution with parameter β ( β > 0) if its PDF is: f(x; β) = βe βx I (0, ) (x) We denote this by X Expo(β). We can integrate this to verify that for x > 0, the corresponding CDF is F(x) = 1 e βx The exponential distribution is memoryless: P(X s + t X s) = X s + t X s = e β(s+t) e βs = P(X t) Page 12

The χ 2 distribution Definition (χ 2 distribution): The χ 2 v distribution is the Gamma( v 2, 2 1 ) distribution. Its PDF is therefore f(x; v) = 1 2 v 2 Γ( v 2 ) x v 2 1 e x 2 I (0, ) (x) Notice that for v = 2, the χ 2 density is equivalent to the exponential density with β = 2 1. It is therefore decreasing for this value of v and hump-shaped for other higher values. The χ 2 is especially useful in problems of statistical inference because if we have v independent random variables, X i N(0, 1), the sum v i=1 X 2 i χ2 v Many of the estimators we use in our models fit this case (i.e. they can be expressed as the sum of independent normal variables) Page 13

Gamma applications Survival analysis: The risk of failure at any point t is given by the hazard rate, h(t) = f(t) S(t) where S(t) is the survival function, 1 F(t). The exponential is memoryless and so, if failure hasn t occurred, the object is as good as new and the hazard rate is constant at β. For wear-out effects, α > 1 and for work-hardening effects, α < 1 Relationship to the Poisson distribution: If Y, the number of events in a given time period t has a poisson density with parameter λ, the rate of success is γ = λ t and the time taken to the first breakdown X Gamma(1, γ). Example: A bottling plant breaks down, on average, twice every four weeks, so λ = 2 and the breakdown rate γ = 2 1 per week. The probability of less than three breakdowns in 4 weeks is P(X 3) = 3 e 2 2i i! =.135 +.271 +.271 +.18 =.857 Suppose we wanted the i=0 probability of no breakdown? The time taken until the first break-down, x must therefore 1 be more than four weeks. This P(X 4) = 2 e x 2 dx = e x 2 = 4 e 2 =.135 Income distributions that are uni-modal 4 Page 14