Stat 5101 Notes: Brand Name Distributions

Similar documents
Stat 5101 Notes: Brand Name Distributions

Common probability distributionsi Math 217 Probability and Statistics Prof. D. Joyce, Fall 2014

Chapter 5. Chapter 5 sections

Random Variables and Their Distributions

A Few Special Distributions and Their Properties

Chapter 5 continued. Chapter 5 sections

Statistics STAT:5100 (22S:193), Fall Sample Final Exam B

Continuous Random Variables

LIST OF FORMULAS FOR STK1100 AND STK1110

MATH : EXAM 2 INFO/LOGISTICS/ADVICE

STAT Chapter 5 Continuous Distributions

Advanced topics from statistics

Continuous Random Variables and Continuous Distributions

1 Uniform Distribution. 2 Gamma Distribution. 3 Inverse Gamma Distribution. 4 Multivariate Normal Distribution. 5 Multivariate Student-t Distribution

Probability and Distributions

Definition: A random variable X is a real valued function that maps a sample space S into the space of real numbers R. X : S R

Stat 5101 Lecture Notes

3. Probability and Statistics

Department of Large Animal Sciences. Outline. Slide 2. Department of Large Animal Sciences. Slide 4. Department of Large Animal Sciences

15 Discrete Distributions

1 Review of Probability and Distributions

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

Continuous Probability Spaces

STAT:5100 (22S:193) Statistical Inference I

Department of Mathematics

Probability Distributions Columns (a) through (d)

Continuous Distributions

Department of Mathematics

Stat 5101 Lecture Slides: Deck 8 Dirichlet Distribution. Charles J. Geyer School of Statistics University of Minnesota

3 Continuous Random Variables

More on Distribution Function

Review 1: STAT Mark Carpenter, Ph.D. Professor of Statistics Department of Mathematics and Statistics. August 25, 2015

2. Suppose (X, Y ) is a pair of random variables uniformly distributed over the triangle with vertices (0, 0), (2, 0), (2, 1).

1 Presessional Probability

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Things to remember when learning probability distributions:

Probability Theory. Patrick Lam

Probability Background

Continuous Random Variables 1

M2S2 Statistical Modelling

t x 1 e t dt, and simplify the answer when possible (for example, when r is a positive even number). In particular, confirm that EX 4 = 3.

Chapter 7: Special Distributions

Stat410 Probability and Statistics II (F16)

MULTIVARIATE DISTRIBUTIONS

STAT 512 sp 2018 Summary Sheet

Sampling Distributions

Chapter 4 Multiple Random Variables

Some Special Distributions (Hogg Chapter Three)

Common ontinuous random variables

3 Modeling Process Quality

Mathematical Statistics 1 Math A 6330

Probability Review. Yutian Li. January 18, Stanford University. Yutian Li (Stanford University) Probability Review January 18, / 27

Quick Tour of Basic Probability Theory and Linear Algebra

Statistics 1B. Statistics 1B 1 (1 1)

Probability- the good parts version. I. Random variables and their distributions; continuous random variables.

Review of Probability Theory

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

Actuarial Science Exam 1/P

THE QUEEN S UNIVERSITY OF BELFAST

Some Special Distributions (Hogg Chapter Three)

MATH c UNIVERSITY OF LEEDS Examination for the Module MATH2715 (January 2015) STATISTICAL METHODS. Time allowed: 2 hours

Topic 2: Review of Probability Theory

Chapter 1. Sets and probability. 1.3 Probability space

STA 2201/442 Assignment 2

PCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities

Statistical Methods in Particle Physics

Review for the previous lecture

STA 4322 Exam I Name: Introduction to Statistics Theory

Sampling Distributions

CHAPTER 6 SOME CONTINUOUS PROBABILITY DISTRIBUTIONS. 6.2 Normal Distribution. 6.1 Continuous Uniform Distribution

1. Let X be a random variable with probability density function. 1 x < f(x) = 0 otherwise

Statistical Methods in Particle Physics

MAS223 Statistical Inference and Modelling Exercises

Chapter 2. Discrete Distributions

MFM Practitioner Module: Quantitative Risk Management. John Dodson. September 23, 2015

APM 504: Probability Notes. Jay Taylor Spring Jay Taylor (ASU) APM 504 Fall / 65

Continuous random variables

Math Stochastic Processes & Simulation. Davar Khoshnevisan University of Utah

Lecture 6: Special probability distributions. Summarizing probability distributions. Let X be a random variable with probability distribution

Statistics and data analyses

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).

MATH2715: Statistical Methods

Lecture 2: Repetition of probability theory and statistics

STT 441 Final Exam Fall 2013

Probability Distributions

Lecture 2: Conjugate priors

ECON 5350 Class Notes Review of Probability and Distribution Theory

Statistics for scientists and engineers

Random Variables. Definition: A random variable (r.v.) X on the probability space (Ω, F, P) is a mapping

Review of Statistics I

Contents 1. Contents

Random Vectors 1. STA442/2101 Fall See last slide for copyright information. 1 / 30

PROBABILITY AND STATISTICAL INFERENCE

(y 1, y 2 ) = 12 y3 1e y 1 y 2 /2, y 1 > 0, y 2 > 0 0, otherwise.

ARCONES MANUAL FOR THE SOA EXAM P/CAS EXAM 1, PROBABILITY, SPRING 2010 EDITION.

Chapter 3 Common Families of Distributions

Joint Distributions. (a) Scalar multiplication: k = c d. (b) Product of two matrices: c d. (c) The transpose of a matrix:

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Probability. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. August 2014

Formulas for probability theory and linear models SF2941

Transcription:

Stat 5101 Notes: Brand Name Distributions Charles J. Geyer February 14, 2003 1 Discrete Uniform Distribution DiscreteUniform(n). Discrete. Rationale Equally likely outcomes. The interval 1, 2,..., n of the integers. Probability Function f(x) = 1 n, x = 1, 2,..., n E(X) = n + 1 2 var(x) = n2 1 12 2 Uniform Distribution Uniform(a, b). Rationale Continuous analog of the discrete uniform distribution. Parameters Real numbers a and b with a < b. The interval (a, b) of the real numbers. 1

Probability Density Function f(x) = 1 b a, a < x < b E(X) = a + b 2 (b a)2 var(x) = 12 Relation to Other Distributions Beta(1, 1) = Uniform(0, 1). 3 Bernoulli Distribution Bernoulli(p). Discrete. Rationale Any zero-or-one-valued random variable. Parameter Real number 0 p 1. The two-element set {0, 1}. Probability Function f(x) = { p, x = 1 1 p, x = 0 E(X) = p var(x) = p(1 p) Addition Rule If X 1,..., X k are i. i. d. Bernoulli(p) random variables, then X 1 + + X k is a Binomial(k, p) random variable. Relation to Other Distributions Bernoulli(p) = Binomial(1, p). 4 Binomial Distribution Binomial(n, p). 2

Discrete. Rationale Sum of i. i. d. Bernoulli random variables. Parameters Real number 0 p 1. Integer n 1. The interval 0, 1,..., n of the integers. Probability Function f(x) = ( ) n p x (1 p) n x, x x = 0, 1,..., n E(X) = np var(x) = np(1 p) Addition Rule If X 1,..., X k are independent random variables, X i being Binomial(n i, p) distributed, then X 1 + + X k is a Binomial(n 1 + + n k, p) random variable. Normal Approximation If np and n(1 p) are both large, then Binomial(n, p) Normal ( np, np(1 p) ) Poisson Approximation If n is large but np is small, then Binomial(n, p) Poisson(np) Theorem The fact that the probability function sums to one is equivalent to the binomial theorem: for any real numbers a and b n k=0 ( ) n a k b n k = (a + b) n. k Degeneracy If p = 0 the distribution is concentrated at 0. If p = 1 the distribution is concentrated at 1. Relation to Other Distributions Bernoulli(p) = Binomial(1, p). 5 Hypergeometric Distribution Hypergeometric(A, B, n). 3

Discrete. Rationale Sample of size n without replacement from finite population of B zeros and A ones. The interval max(0, n B),..., min(n, A) of the integers. Probability Function ( A B ) f(x) = x)( n x ), x = max(0, n B),..., min(n, A) where ( A+B n E(X) = np var(x) = np(1 p) N n N 1 p = A A + B N = A + B (5.1) Binomial Approximation If n is small compared to either A or B, then where p is given by (5.1). Hypergeometric(n, A, B) Binomial(n, p) Normal Approximation If n is large, but small compared to either A or B, then Hypergeometric(n, A, B) Normal ( np, np(1 p) ) where p is given by (5.1). 6 Poisson Distribution Poisson(µ) Discrete. Rationale Counts in a Poisson process. Parameter Real number µ > 0. 4

The non-negative integers 0, 1,.... Probability Function f(x) = µx x! e µ, x = 0, 1,... E(X) = µ var(x) = µ Addition Rule If X 1,..., X k are independent random variables, X i being Poisson(µ i ) distributed, then X 1 + + X k is a Poisson(µ 1 + + µ k ) random variable. Normal Approximation If µ is large, then Poisson(µ) Normal(µ, µ) Theorem The fact that the probability function sums to one is equivalent to the Maclaurin series for the exponential function: any real numbers x k=0 x k k! = ex. 7 Geometric Distribution Geometric(p). Discrete. Rationales Discrete lifetime of object that does not age. Waiting time or interarrival time in sequence of i. i. d. Bernoulli trials. Inverse sampling. Discrete analog of the exponential distribution. Parameter Real number 0 < p < 1. The non-negative integers 0, 1,.... 5

Probability Function f(x) = p(1 p) x x = 0, 1,... E(X) = 1 p p var(x) = 1 p p 2 Addition Rule If X 1,..., X k are i. i. d. Geometric(p) random variables, then X 1 + + X k is a NegativeBinomial(k, p) random variable. Theorem The fact that the probability function sums to one is equivalent to the geometric series: for any real number s such that 1 < s < 1 s k = 1 1 s. k=0 8 Negative Binomial Distribution NegativeBinomial(r, p). Discrete. Rationale Sum of i. i. d. geometric random variables. Inverse sampling. Parameters Real number 0 p 1. Integer r 1. The non-negative integers 0, 1,.... Probability Function ( ) r + x 1 f(x) = p r (1 p) x, x = 0, 1,... x r(1 p) E(X) = p r(1 p) var(x) = p 2 6

Addition Rule If X 1,..., X k are independent random variables, X i being NegativeBinomial(r i, p) distributed, then X 1 + + X k is a NegativeBinomial(r 1 + + r k, p) random variable. Normal Approximation If r(1 p) is large, then ( r(1 p) NegativeBinomial(r, p) Normal, p ) r(1 p) p 2 Extended Definition The definition makes sense for noninteger r if binomial coefficients are defined by ( ) r r (r 1) (r k + 1) = k k! which for integer r agrees with the standard definition. Also ( ) ( ) r + x 1 r = ( 1) x x x which explains the name negative binomial. (8.1) Theorem The fact that the probability function sums to one is equivalent to the generalized binomial theorem: for any real number s such that 1 < s < 1 and any real number m k=0 ( ) m s k = (1 + s) m. (8.2) k If m is a nonnegative integer, then ( m k ) is zero for k > m, and we get the ordinary binomial theorem. Changing variables from m to m and from s to s and using (8.1) turns (8.2) into ( ) m + k 1 s k = k k=0 ( m k k=0 ) ( s) k = (1 s) m which has a more obvious relationship to the negative binomial density summing to one. 9 Normal Distribution Normal(µ, σ 2 ). 7

Rationale Limiting distribution in the central limit theorem. Parameters Real numbers µ and σ 2 > 0. The real numbers. Probability Density Function f(x) = 1 2πσ e (x µ)2 /2σ 2, < x < E(X) = µ var(x) = σ 2 E{(X µ) 3 } = 0 E{(X µ) 4 } = 3σ 4 Linear Transformations If X is Normal(µ, σ 2 ) distributed, then ax + b is Normal(aµ + b, a 2 σ 2 ) distributed. Addition Rule If X 1,..., X k are independent random variables, X i being Normal(µ i, σi 2) distributed, then X 1 + + X k is a Normal(µ 1 + + µ k, σ1 2 + + σk 2 ) random variable. Theorem The fact that the probability density function integrates to one is equivalent to the integral e z2 /2 dz = 2π Relation to Other Distributions If Z is Normal(0, 1) distributed, then Z 2 is Gamma( 1 2, 1 2 ) distributed. 10 Exponential Distribution Exponential(λ). 8

Rationales Lifetime of object that does not age. Waiting time or interarrival time in Poisson process. Continuous analog of the geometric distribution. Parameter Real number λ > 0. The interval (0, ) of the real numbers. Probability Density Function f(x) = λe λx, 0 < x < Cumulative Distribution Function F (x) = 1 e λx, 0 < x < E(X) = 1 λ var(x) = 1 λ 2 Addition Rule If X 1,..., X k are i. i. d. Exponential(λ) random variables, then X 1 + + X k is a Gamma(k, λ) random variable. Relation to Other Distributions Exponential(λ) = Gamma(1, λ). 11 Gamma Distribution Gamma(α, λ). Rationales Sum of i. i. d. exponential random variables. Conjugate prior for exponential, Poisson, or normal precision family. Parameter Real numbers α > 0 and λ > 0. The interval (0, ) of the real numbers. 9

Probability Density Function f(x) = λα Γ(α) xα 1 e λx, 0 < x < where Γ(α) is defined by (11.1) below. E(X) = α λ var(x) = α λ 2 Addition Rule If X 1,..., X k are independent random variables, X i being Gamma(α i, λ) distributed, then X 1 + + X k is a Gamma(α 1 + + α k, λ) random variable. Normal Approximation If α is large, then ( α Gamma(α, λ) Normal λ, α ) λ 2 Theorem The fact that the probability density function integrates to one is equivalent to the integral 0 x α 1 e λx dx = Γ(α) λ α the case λ = 1 is the definition of the gamma function Γ(α) = Relation to Other Distributions Exponential(λ) = Gamma(1, λ). Chi-Square(ν) = Gamma( ν 2, 1 2 ). 0 x α 1 e x dx (11.1) If X and Y are independent, X is Γ(α, λ) distributed and Y is Γ(β, λ) distributed, then X/(X + Y ) is Beta(α, β) distributed. If Z is Normal(0, 1) distributed, then Z 2 is Gamma( 1 2, 1 2 ) distributed. 10

Facts About Gamma Functions Integration by parts in (11.1) establishes the gamma function recursion formula Γ(α + 1) = αγ(α), α > 0 (11.2) The relationship between the Exponential(λ) and Gamma(1, λ) distributions gives Γ(1) = 1 and the relationship between the Normal(0, 1) and Gamma( 1 2, 1 2 ) distributions gives Γ( 1 2 ) = π Together with the recursion (11.2) these give for any positive integer n Γ(n + 1) = n! and Γ(n + 1 2 ) = ( ( ) n 2) 1 n 3 2 3 2 1 2 π 12 Beta Distribution Beta(α, β). Rationales Ratio of gamma random variables. Conjugate prior for binomial or negative binomial family. Parameter Real numbers α > 0 and β > 0. The interval (0, 1) of the real numbers. Probability Density Function f(x) = where Γ(α) is defined by (11.1) above. Γ(α + β) Γ(α)Γ(β) xα 1 (1 x) β 1 0 < x < 1 E(X) = α α + β αβ var(x) = (α + β) 2 (α + β + 1) 11

Theorem The fact that the probability density function integrates to one is equivalent to the integral 1 Relation to Other Distributions 0 x α 1 (1 x) β 1 dx = Γ(α)Γ(β) Γ(α + β) If X and Y are independent, X is Γ(α, λ) distributed and Y is Γ(β, λ) distributed, then X/(X + Y ) is Beta(α, β) distributed. Beta(1, 1) = Uniform(0, 1). 13 Multinomial Distribution Multinomial(n, p) Discrete. Rationale Parameters Multivariate analog of the binomial distribution. Real vector p in the parameter space { p R k : 0 p i, i = 1,..., k, and } k p i = 1 i=1 (13.1) The set of vectors with integer coordinates { } k S = x Z k : 0 x i, i = 1,..., k, and x i = n i=1 (13.2) Probability Function where f(x) = is called a multinomial coefficient. ( ) n k p xi i x, x S i=1 ( ) n n! = x k i=1 x i! E(X i ) = np i var(x i ) = np i (1 p i ) cov(x i, X j ) = np i p j, i j 12

(Vector Form) where E(X) = np var(x) = nm M = diag(p) pp is the matrix with elements m ij = cov(x i, X j )/n. Addition Rule If X 1,..., X k are independent random vectors, X i being Multinomial(n i, p) distributed, then X 1 + + X k is a Multinomial(n 1 + + n k, p) random variable. Normal Approximation If n is large and p is not near the boundary of the parameter space (13.1), then Multinomial(n, p) Normal(np, nm) Theorem The fact that the probability function sums to one is equivalent to the multinomial theorem: for any vector a of real numbers [ (n ) k ] a xi i = (a 1 + + a k ) n x x S i=1 Degeneracy If there exists a vector a such that Ma = 0, then var(a X) = 0. In particular, the vector u = (1, 1,..., 1) always satisfies Mu = 0, so var(u X) = 0. This is obvious, since u X = k i=1 X i = n by definition of the multinomial distribution, and the variance of a constant is zero. This means a multinomial random vector of dimension k is really of dimension no more than k 1 because it is concentrated on the hyperplane containing the sample space (13.2). Marginal Distributions Every univariate marginal is binomial X i Binomial(n, p i ) Not, strictly speaking marginals, but random vectors formed by collapsing categories are multinomial. If A 1,..., A m is a partition of the set {1,..., k} and Y j = X i, j = 1,..., m i A j q j = p i, j = 1,..., m i A j then the random vector Y has a Multinomial(n, q) distribution. 13

Conditional Distributions If {i 1,..., i m } and {i m+1,..., i k } partition the set {1,..., k}, then the conditional distribution of X i1,..., X im given X im+1,..., X ik is Multinomial(n X im+1 X ik, q), where the parameter vector q has components q j = p ij p i1 + + p im, j = 1,..., m Relation to Other Distributions Each marginal of a multinomial is binomial. If X is Binomial(n, p), then the two-component vector (X, n X) is Multinomial ( n, (p, 1 p) ). 14 Bivariate Normal Distribution See multivariate normal below. Rationales See multivariate normal below. Parameters Real vector µ of dimension 2, real symmetric positive semidefinite matrix M of dimension 2 2 having the form ( ) σ 2 M = 1 ρσ 1 σ 2 ρσ 1 σ 2 σ2 2 where σ 1 > 0, σ 2 > 0 and 1 < ρ < +1. The Euclidean space R 2. Probability Density Function f(x) = 1 2π det(m) 1/2 exp ( 1 2 (x µ) M 1 (x µ) ) ( [ (x1 ) 2 1 = 2π 1 µ 1 exp 1 ρ 2 σ 1 σ 2 2(1 ρ 2 ) σ 1 ( ) ( ) ( ) ]) 2 x1 µ 1 x2 µ 2 x2 µ 2 2ρ +, x R 2 σ 1 σ 2 σ 2 14

(Vector Form) E(X i ) = µ i, i = 1, 2 var(x i ) = σ 2 i, i = 1, 2 cov(x 1, X 2 ) = ρσ 1 σ 2 cor(x 1, X 2 ) = ρ E(X) = µ var(x) = M Linear Transformations See multivariate normal below. Addition Rule See multivariate normal below. Marginal Distributions X i is Normal(µ i, σi 2 ) distributed, i = 1, 2. Conditional Distributions The conditional distribution of X 2 given X 1 is ( Normal µ 2 + ρ σ ) 2 (x 1 µ 1 ), (1 ρ 2 )σ2 2 σ 1 where 15 Multivariate Normal Distribution Normal(µ, M) Rationales Multivariate analog of the univariate normal distribution. Limiting distribution in the multivariate central limit theorem. Parameters Real vector µ of dimension k, real symmetric positive semidefinite matrix M of dimension k k. The Euclidean space R k. Probability Density Function If M is (strictly) positive definite, f(x) = (2π) k/2 det(m) 1/2 exp ( 1 2 (x µ) M 1 (x µ) ), x R k Otherwise there is no density (X is concentrated on a hyperplane). 15

(Vector Form) E(X) = µ var(x) = M Linear Transformations If X is Normal(µ, M) distributed, then AX + b, where A is a constant matrix and b is a constant vector of dimensions such that the matrix multiplication and vector addition make sense, is Normal(Aµ + b, AMA ) distributed. Addition Rule If X 1,..., X k are independent random vectors, X i being Normal(µ i, M i ) distributed, then X 1 + +X k is a Normal(µ 1 + +µ k, M 1 + + M k ) random variable. Degeneracy If there exists a vector a such that Ma = 0, then var(a X) = 0. Partitioned Vectors and Matrices The random vector and parameters are written in partitioned form ( ) X1 X = (15.1a) X 2 ( ) µ1 µ = (15.1b) µ 2 ( ) M11 M M = 12 (15.1c) M 21 M 2 when X 1 consists of the first r elements of X and X 2 of the other k r elements and similarly for µ 1 and µ 2. Marginal Distributions Every marginal of a multivariate normal is normal (univariate or multivariate as the case may be). In partitioned form, the (marginal) distribution of X 1 is Normal(µ 1, M 11 ). Conditional Distributions Every conditional of a multivariate normal is normal (univariate or multivariate as the case may be). In partitioned form, the conditional distribution of X 1 given X 2 is Normal(µ 1 + M 12 M 22 [X 2 µ 2 ], M 11 M 12 M 22 M 21) where the notation M 22 denotes the inverse of the matrix M 22 if the matrix is invertable and otherwise any generalized inverse. 16

16 Chi-Square Distribution Chi-Square(ν) or χ 2 (ν). Rationales Sum of squares of i. i. d. standard normal random variables. Sampling distribution of sample variance when data are i. i. d. normal. Parameter Real number ν > 0 called degrees of freedom. The interval (0, ) of the real numbers. Probability Density Function f(x) = ( 1 2 )ν/2 Γ( ν 2 ) xν/2 1 e x/2, 0 < x <. E(X) = ν var(x) = 2ν Addition Rule If X 1,..., X k are independent random variables, X i being Chi-Square(ν i ) distributed, then X 1 + + X k is a Chi-Square(ν 1 + + ν k ) random variable. Normal Approximation If α is large, then Relation to Other Distributions Chi-Square(ν) = Gamma( ν 2, 1 2 ). Chi-Square(ν) Normal(ν, 2ν) If Z and Y are independent, X is Normal(0, 1) distributed and Y is Chi-Square(ν) distributed, then X/ Y/ν is t(ν) distributed. If X and Y are independent and are Chi-Square(µ) and Chi-Square(ν) distributed, respectively, then (X/µ)/(Y/ν) is F (µ, ν) distributed. 17 Student s t Distribution t(ν). 17

Rationales Sampling distribution of pivotal quantity n(x n µ)/s n when data are i. i. d. normal. Marginal for µ in conjugate prior family for two-parameter normal data. Parameter Real number ν > 0 called degrees of freedom. The real numbers. Probability Density Function f(x) = 1 ν+1 Γ( 2 ) νπ Γ( ν 2 ) If ν > 1, then ( 1 + x 2 ν E(X) = 0. Otherwise the mean does not exist. If ν > 2, then var(x) = Otherwise the variance does not exist. 1 ) (ν+1)/2, < x < + ν ν 2. Normal Approximation If ν is large, then Relation to Other Distributions t(ν) Normal(0, 1) If Z and Y are independent, X is Normal(0, 1) distributed and Y is Chi-Square(ν) distributed, then X/ Y/ν is t(ν) distributed. If X is t(ν) distributed, then X 2 is F (1, ν) distributed. t(1) = Cauchy(0, 1). 18 Snedecor s F Distribution F (µ, ν). 18

Rationale Ratio of sums of squares for normal data (test statistics in regression and analysis of variance). Parameters Real numbers µ > 0 and ν > 0 called numerator degrees of freedom and denominator degrees of freedom, respectively. The interval (0, ) of the real numbers. Probability Density Function f(x) = µ+ν Γ( 2 )µµ/2 ν ν/2 x µ/2+1 Γ( µ 2 )Γ( ν 2 ), 0 < x < + (µx + ν) (µ+ν)/2 If ν > 2, then Otherwise the mean does not exist. E(X) = Relation to Other Distributions ν ν 2. If X and Y are independent and are Chi-Square(µ) and Chi-Square(ν) distributed, respectively, then (X/µ)/(Y/ν) is F (µ, ν) distributed. If X is t(ν) distributed, then X 2 is F (1, ν) distributed. 19 Cauchy Distribution Cauchy(µ, σ). Rationales Very heavy tailed distribution. Counterexample to law of large numbers. Parameters Real numbers µ and σ > 0, called the location and scale parameter, respectively. The real numbers. 19

Probability Density Function f(x) = 1 πσ 1 1 + ( x µ σ ) 2, < x < + No moments exist. Addition Rule If X 1,..., X k are i. i. d. Cauchy(µ, σ) random variables, then X n = (X 1 + + X k )/n is also Cauchy(µ, σ). Relation to Other Distributions t(1) = Cauchy(0, 1). 20