Topic 4: Continuous random variables

Similar documents
Topic 4: Continuous random variables

Continuous random variables

MATH Solutions to Probability Exercises

Continuous Distributions

Let X be a continuous random variable, < X < f(x) is the so called probability density function (pdf) if

Continuous Random Variables and Continuous Distributions

3 Continuous Random Variables

CHAPTER 6 SOME CONTINUOUS PROBABILITY DISTRIBUTIONS. 6.2 Normal Distribution. 6.1 Continuous Uniform Distribution

Continuous random variables

1 Solution to Problem 2.1

Probability and Distributions

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

Random Variables and Their Distributions

Probability Density Functions

Continuous Random Variables

Statistics 3657 : Moment Generating Functions

1 Review of Probability

Review of Statistics I

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

Chapter 5 continued. Chapter 5 sections

Continuous Distributions

1 Review of Probability and Distributions

Continuous Random Variables

Continuous distributions

Probability, Random Processes and Inference

Gamma and Normal Distribuions

Arkansas Tech University MATH 3513: Applied Statistics I Dr. Marcel B. Finan

15 Discrete Distributions

STAT 509 Section 3.4: Continuous Distributions. Probability distributions are used a bit differently for continuous r.v. s than for discrete r.v. s.

Probability Midterm Exam 2:15-3:30 pm Thursday, 21 October 1999

Continuous Random Variables and Probability Distributions

p. 6-1 Continuous Random Variables p. 6-2

3. Probability and Statistics

1 Probability and Random Variables

Lecture 17: The Exponential and Some Related Distributions

Review 1: STAT Mark Carpenter, Ph.D. Professor of Statistics Department of Mathematics and Statistics. August 25, 2015

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

Chapter 3. Julian Chan. June 29, 2012

ORF 245 Fundamentals of Statistics Chapter 4 Great Expectations

Probability Distributions for Continuous Variables. Probability Distributions for Continuous Variables

LIST OF FORMULAS FOR STK1100 AND STK1110

Continuous Distributions

Math 576: Quantitative Risk Management

Random Variables. Saravanan Vijayakumaran Department of Electrical Engineering Indian Institute of Technology Bombay

Chapter 4 Continuous Random Variables and Probability Distributions

Expectation, Variance and Standard Deviation for Continuous Random Variables Class 6, Jeremy Orloff and Jonathan Bloom

Slides 8: Statistical Models in Simulation

5.2 Continuous random variables

Expected Values, Exponential and Gamma Distributions

Stat410 Probability and Statistics II (F16)

continuous random variables

Brief Review of Probability

Common ontinuous random variables

RMSC 2001 Introduction to Risk Management

Definition: A random variable X is a real valued function that maps a sample space S into the space of real numbers R. X : S R

Bell-shaped curves, variance

STAT 430/510: Lecture 10

Math 105 Course Outline

Continuous Random Variables

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

Notes on Continuous Random Variables

5.6 The Normal Distributions

THE QUEEN S UNIVERSITY OF BELFAST

Exponential & Gamma Distributions

Chapter 4: Continuous Probability Distributions

Chapter 4 Multiple Random Variables

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

RMSC 2001 Introduction to Risk Management

Chapter 2. Continuous random variables

STAT Chapter 5 Continuous Distributions

STAT 430/510 Probability Lecture 12: Central Limit Theorem and Exponential Distribution

Exponential, Gamma and Normal Distribuions

S6880 #7. Generate Non-uniform Random Number #1

Chapter 5. Chapter 5 sections

CSE 312, 2017 Winter, W.L. Ruzzo. 7. continuous random variables

ECON 5350 Class Notes Review of Probability and Distribution Theory

SDS 321: Introduction to Probability and Statistics

Preliminary Statistics Lecture 3: Probability Models and Distributions (Outline) prelimsoas.webs.com

APPM/MATH 4/5520 Solutions to Problem Set Two. = 2 y = y 2. e 1 2 x2 1 = 1. (g 1

Definition 1.1 (Parametric family of distributions) A parametric distribution is a set of distribution functions, each of which is determined by speci

STA 256: Statistics and Probability I

Will Landau. Feb 28, 2013

Final Exam # 3. Sta 230: Probability. December 16, 2012

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 8 10/1/2008 CONTINUOUS RANDOM VARIABLES

ARCONES MANUAL FOR THE SOA EXAM P/CAS EXAM 1, PROBABILITY, SPRING 2010 EDITION.

4.1 Expected value of a discrete random variable. Suppose we shall win $i if the die lands on i. What is our ideal average winnings per toss?

Definition A random variable X is said to have a Weibull distribution with parameters α and β (α > 0, β > 0) if the pdf of X is

Chapter 4: Continuous Random Variables and Probability Distributions

Experimental Design and Statistics - AGA47A

2 Functions of random variables

HW7 Solutions. f(x) = 0 otherwise. 0 otherwise. The density function looks like this: = 20 if x [10, 90) if x [90, 100]

Probability Distributions Columns (a) through (d)

This does not cover everything on the final. Look at the posted practice problems for other topics.

MATH 3510: PROBABILITY AND STATS July 1, 2011 FINAL EXAM

Final Solutions Fri, June 8

A Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

Mathematical Statistics 1 Math A 6330

Probability theory for Networks (Part 1) CS 249B: Science of Networks Week 02: Monday, 02/04/08 Daniel Bilar Wellesley College Spring 2008

BMIR Lecture Series on Probability and Statistics Fall, 2015 Uniform Distribution

Transcription:

Topic 4: Continuous random variables Course 3, 216 Page

Continuous random variables Definition (Continuous random variable): An r.v. X has a continuous distribution if there exists a non-negative function f defined on the real line such that for any interval A, P(X A) = f(x)dx The function f is called the probability density function (PDF) of X. Every PDF must satisfy: A 1. Nonnegative: f(x) 2. Integrates to 1: The CDF of X is given by: f(x)dx = 1 x F(x) = f(t)dt A continuous random variable is a random variable with a continuous distribution. Page 1

Continuous r.v.s: caveats and remarks 1. The density is not a probability and it is possible to have f(x) > 1. 2. P(X = x) = for all x. We can compute a probability for X being very close to x by integrating f over an ɛ interval around x. 3. Any function satisfying the properties of the PDF, represents the density of some r.v. 4. It doesn t matter whether we include or exclude endpoints: P(a < X < b) = P(a < X b) = P(a X < b) = P(a X b) = b f(x)dx a Page 2

Expectation of a continuous r.v. Definition (Expectation of a continuous random variable): The expected value or mean of a continuous r.v. X with PDF f is E(X) = xf(x)dx Result (Expectation of a function of X): If X is a continuous r.v. X with PDF f and g is a function from R to R, then E(g(X)) = g(x)f(x)dx Page 3

The Logistic distribution The CDF of a logistic distribution is F(x) = 1+e ex x, x R. We differentiate this to get the PDF, f(x) = e x (1+e x ) 2, x R This is a similar to the normal distribution but has a closed-form CDF and is computationally easier. For example: P( 2 < X < 2) = F(2) F( 2) = e2 1 1+e 2 =.76 Page 4

The Uniform distribution Definition (Uniform distribution): A continuous r.v. U has a Uniform distribution on the interval (a, b) if its PDF is 1 f(x; a, b) = b a I (a,b)(x) We denote this by U Unif(a, b) We often write a density in this manner- using the indicator variable to describe its support. This is a valid PDF since the area under the curve is the area of a rectangle with length (b a) 1 and height (b a). The CDF is the accumulated area under the PDF: Graph the PDF and CDF of Unif(, 1). if x a F(x) = x a b a if a < x b 1 if x b Suppose we have observations from Unif(, 1). We can obtain transform these into a sample from Ũ Unif(a, b): Ũ = a + (b a)u Page 5

The Probability Integral Transformation Result: Let X be a continuous random variable with the distribution function F and let Y = F(X). Then Y must be uniformly distributed on [, 1]. The transformation from X to Y is called the probability integral transformation. We know that the distribution function must take values between and 1. If we pick any of these values, y, the y th quantile of the distribution of X will be given by some number x and Pr(Y y) = Pr(X x) = F(x) = y which is the distribution function of a uniform random variable. This result helps us generate random numbers from various distributions, because it allows us to transform a sample from a uniform distribution into a sample from some other distribution provided we can find F 1. Example: Given U Unif(, 1), log ( U 1 U ) Logistic. (F(x) = 1+e ex x, set this equal to u and solve for x) Page 6

The Normal distribution This symmetric bell-shaped density is widely used because: 1. It captures many types of natural variation quite well: heights-humans, animals and plants, weights, strength of physical materials, the distance from the centre of a target. 2. It has nice mathematical properties: many functions of a set normally distributed random variables have distributions that take simple forms. 3. Central limit theorems are fundamental to statistic inference. The sample mean of a large random sample from any distribution with finite variance is approximately normal. Page 7

The Standard Normal distribution Definition (Standard Normal distribution): A continuous r.v. Z is said to have a standard Normal distribution if its PDF ϕ is given by: ϕ(z) = 1 2π e z2 /2, < z < The CDF is the accumulated area under the PDF: Φ(z) = z ϕ(z)dt = z 1 2π e z2 /2 dt E(Z) =, Var(Z) = 1 and X = µ + σz is has a Normal distribution with mean µ and variance σ 2. (since E(µ + σz) = E(µ) + σe(z) = µ and Var(µ + σz) = σ 2 Var(Z) = σ 2 ) Page 8

Approximations of Normal probabilities If X N(µ, σ 2 ) then P( X µ < σ).68 P( X µ < 2σ).95 P( X µ < 3σ).997 Or stated in terms of z: P( Z < 1).68 P( Z < 2).95 P( Z < 3).997 The Normal distribution therefore has very thin tails relative to other commonly used distributions that are symmetric about their mean (logistic, Student s-t, Cauchy). Page 9

The Gamma function The gamma function is a special mathematical function that is widely used in statistics. The gamma function of α is defined as Γ(α) = y α 1 e y dy (1) If α = 1, Γ(α) = e y dy = e y = 1 If α > 1, we can integrate (??) by parts, setting u = y α 1 and dv = e y and using the formula udv = uv y vdu to get: α 1 e y + (α 1) y α 2 e y dy The first term in the above expression is zero because the exponential function goes to zero faster than any polynomial and we obtain and for any integer α > 1, we have Γ(α) = (α 1)Γ(α 1) Γ(α) = (α 1)(α 2)(α 3)... (3)(2)(1)Γ(1) = (α 1)! Page 1

The Gamma distribution Definition (Gamma distribution): It is said that a random variable X has a Gamma distribution with parameters α and β (α > and β > ) if X has a continuous distribution for which the PDF is: f(x; α, β) = βα Γ(α) xα 1 e βx I (, ) (x) To see that this is a valid PDF, let y = βx in the expression for the Gamma function, Γ(α) = y α 1 e y dy, so dy = βdx and can rewrite Γ(α) as or Γ(α) = 1 = (xβ) α 1 e xβ βdx β α Γ(α) xα 1 e βx dx E(X) = α β Var(X) = α β 2 Page 11

Features of the Gamma density The density is decreasing for α 1, and hump-shaped and skewed to the right otherwise. It attains a maximum at x = α 1 β Higher values of α make it more bell-shaped and higher β increases the concentration at lower x values. Page 12

The Exponential distribution We have a special name for a Gamma distribution with α = 1: Definition (Exponential distribution): A continuous random variable X has an exponential distribution with parameter β ( β > ) if its PDF is: f(x; β) = βe βx I (, ) (x) We denote this by X Expo(β). We can integrate this to verify that for x >, the corresponding CDF is F(x) = 1 e βx The exponential distribution is memoryless: P(X s + t X s) = X s + t X s = e β(s+t) e βs = P(X t) Page 13

The χ 2 distribution Definition (χ 2 distribution): The χ 2 v distribution is the Gamma( v 2, 2 1 ) distribution. Its PDF is therefore f(x; v) = 1 2 v 2 Γ( v 2 ) x v 2 1 e x 2 I (, ) (x) Notice that for v = 2, the χ 2 density is equivalent to the exponential density with β = 2 1. It is therefore decreasing for this value of v and hump-shaped for other higher values. The χ 2 is especially useful in problems of statistical inference because if we have v independent random variables, X i N(, 1), the sum v i=1 X 2 i χ2 v Many of the estimators we use in our models fit this case (i.e. they can be expressed as the sum of independent normal variables) Page 14

Gamma applications Survival analysis: The risk of failure at any point t is given by the hazard rate, h(t) = f(t) S(t) where S(t) is the survival function, 1 F(t). The exponential is memoryless and so, if failure hasn t occurred, the object is as good as new and the hazard rate is constant at β. For wear-out effects, α > 1 and for work-hardening effects, α < 1 Relationship to the Poisson distribution: If Y, the number of events in a given time period t has a poisson density with parameter λ, the rate of success is γ = λ t and the time taken to the first breakdown X Gamma(1, γ). Example: A bottling plant breaks down, on average, twice every four weeks, so λ = 2 and the breakdown rate γ = 2 1 per week. The probability of less than three breakdowns in 4 weeks is P(X 3) = 3 e 2 2i i! =.135 +.271 +.271 +.18 =.857 Suppose we wanted the i= probability of no breakdown? The time taken until the first break-down, x must therefore 1 be more than four weeks. This P(X 4) = 2 e x 2 dx = e x 2 = 4 e 2 =.135 Income distributions that are uni-modal 4 Page 15