Sampling Distributions

Similar documents
Sampling Distributions

Probability and Distributions

Continuous Distributions

2. The CDF Technique. 1. Introduction. f X ( ).

1 Review of Probability and Distributions

Lecture The Sample Mean and the Sample Variance Under Assumption of Normality

Limiting Distributions

Moments. Raw moment: February 25, 2014 Normalized / Standardized moment:

BMIR Lecture Series on Probability and Statistics Fall 2015 Discrete RVs

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

18.440: Lecture 28 Lectures Review

1 Solution to Problem 2.1

Limiting Distributions

n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2

Stat410 Probability and Statistics II (F16)

APPM/MATH 4/5520 Solutions to Exam I Review Problems. f X 1,X 2. 2e x 1 x 2. = x 2

STA 256: Statistics and Probability I

LIST OF FORMULAS FOR STK1100 AND STK1110

2 Functions of random variables

1.1 Review of Probability Theory

n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2

18.440: Lecture 28 Lectures Review

7.3 The Chi-square, F and t-distributions

THE QUEEN S UNIVERSITY OF BELFAST

, find P(X = 2 or 3) et) 5. )px (1 p) n x x = 0, 1, 2,..., n. 0 elsewhere = 40

Continuous Random Variables and Continuous Distributions

Brief Review of Probability

Lecture 1: August 28

Continuous random variables

Lecture 5: Moment generating functions

Basics of Stochastic Modeling: Part II

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

STA2603/205/1/2014 /2014. ry II. Tutorial letter 205/1/

Continuous Random Variables

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

Probability Density Functions

Things to remember when learning probability distributions:

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics

3 Continuous Random Variables

Exercises and Answers to Chapter 1

Lecture 11. Multivariate Normal theory

Joint p.d.f. and Independent Random Variables

STAT 3610: Review of Probability Distributions

MATH c UNIVERSITY OF LEEDS Examination for the Module MATH2715 (January 2015) STATISTICAL METHODS. Time allowed: 2 hours

MATH 56A: STOCHASTIC PROCESSES CHAPTER 6

[Chapter 6. Functions of Random Variables]

3. Probability and Statistics

Chapter 4 Multiple Random Variables

Math Spring Practice for the Second Exam.

Statistics for scientists and engineers

MAS223 Statistical Inference and Modelling Exercises

Random Variables and Their Distributions

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

Chapter 2 Continuous Distributions

Review for the previous lecture

Exponential Distribution and Poisson Process

Stat 5101 Notes: Brand Name Distributions

CHAPTER 6 SOME CONTINUOUS PROBABILITY DISTRIBUTIONS. 6.2 Normal Distribution. 6.1 Continuous Uniform Distribution

Continuous Distributions

Chapter 3 Common Families of Distributions

ACM 116: Lectures 3 4

Statistics 3657 : Moment Generating Functions

Will Murray s Probability, XXXII. Moment-Generating Functions 1. We want to study functions of them:

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

4. Distributions of Functions of Random Variables

Summary. Ancillary Statistics What is an ancillary statistic for θ? .2 Can an ancillary statistic be a sufficient statistic?

Mathematical Statistics

HW7 Solutions. f(x) = 0 otherwise. 0 otherwise. The density function looks like this: = 20 if x [10, 90) if x [90, 100]

Common ontinuous random variables

Random variables and transform methods

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Probability and Statistics Notes

Probability Distributions Columns (a) through (d)

This does not cover everything on the final. Look at the posted practice problems for other topics.

This exam is closed book and closed notes. (You will have access to a copy of the Table of Common Distributions given in the back of the text.

Continuous Random Variables. and Probability Distributions. Continuous Random Variables and Probability Distributions ( ) ( ) Chapter 4 4.

1 Introduction. P (n = 1 red ball drawn) =

Chapter 7: Special Distributions

Fundamentals of Statistics

Statistics (1): Estimation

Statistics 1B. Statistics 1B 1 (1 1)

Stat 5101 Notes: Brand Name Distributions

Order Statistics and Distributions

Random Vectors and Multivariate Normal Distributions

Lecture 2: Repetition of probability theory and statistics

Some Special Distributions (Hogg Chapter Three)

Continuous Random Variables. and Probability Distributions. Continuous Random Variables and Probability Distributions ( ) ( )

Chapter 5. Chapter 5 sections

Statistics, Data Analysis, and Simulation SS 2015

STAT Chapter 5 Continuous Distributions

BASICS OF PROBABILITY

Chapter 4. Chapter 4 sections

Write your Registration Number, Test Centre, Test Code and the Number of this booklet in the appropriate places on the answersheet.

4. CONTINUOUS RANDOM VARIABLES

Probability- the good parts version. I. Random variables and their distributions; continuous random variables.

Statistics STAT:5100 (22S:193), Fall Sample Final Exam B

Algorithms for Uncertainty Quantification

UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017)

Bivariate Normal Distribution

E[X n ]= dn dt n M X(t). ). What is the mgf? Solution. Found this the other day in the Kernel matching exercise: 1 M X (t) =

Transcription:

Sampling Distributions In statistics, a random sample is a collection of independent and identically distributed (iid) random variables, and a sampling distribution is the distribution of a function of random sample. For example, the average and the variance formula are functions of random sample. The term statistic is referred to an individual function of a random sample, and the understanding of sampling distributions is the major undertaking of statistics. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 1 / 53

Moment generating function. Let X be a random variable. The moment generating function (mgf) of X is given by M(t) = E [ e tx ]. The mgf M(t) is a function of t defined on some open interval (c 0, c 1 ) around 0 with c 0 < 0 < c 1. In particular when X is a continuous random variable having the pdf f (x), the mgf M(t) can be expressed as M(t) = e tx f (x) dx. The most significant property of moment generating function is that the moment generating function uniquely determines the distribution. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 2 / 53

MGF of normal distribution. Suppose that X is a standard normal random variable. Then we can compute the mgf of X as follows. M X (t) = 1 2π = e t2 2 1 2π e tx e x2 2 dx = 1 2π e (x t)2 2 dx = exp e 1 2 (x t)2 + t2 2 ( ) t 2 2 dx Since X = (Y µ)/σ becomes a standard normal random variable and Y = σx + µ, the mgf of Y can be given by ( ) σ M Y (t) = e µt 2 t 2 M X (σt) = exp + µt 2 for < t <. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 3 / 53

MGF of gamma distribution. Suppose now that X is a gamma random variable with parameter (α, λ). Then the mgf M X (t) of X can be computed as follows. M X (t) = = = e tx 0 λ α λ α Γ(α) x α 1 e λx dx = (λ ( t) α Γ(α) ) α λ 1 1 λ t Γ(α) 0 0 λα Γ(α) 0 x α 1 e (λ t)x dx ((λ t)x) α 1 e (λ t)x (λ t)dx ( ) α λ u α 1 e u du = λ t where (λ t)x is substituted by u, and (λ t)dx by du. It should be noted that this mgf M X (t) is defined on the open interval (, λ). Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 4 / 53

Joint moment generating function. Let X and Y be random variables having a joint density function f (x, y). Then we can define the joint moment generating function M(s, t) of X and Y by M(s, t) = E [ e sx +ty ] = e sx+ty f (x, y) dx dy. If X and Y are independent, then the joint mgf M(s, t) becomes M(s, t) = E [ e sx +ty ] = E [ e sx ] E [ e ty ] = M X (s) M Y (t). Moreover, it has been known that if the joint mgf M(s, t) of X and Y is of the form M 1 (s) M 2 (t), then X and Y are independent and have the respective mgf M 1 (s) and M 2 (t). Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 5 / 53

Sum of independent random variables. Let X and Y be independent random variables having the respective probability density functions f X (x) and f Y (y). Then the cumulative distribution function F Z (z) of the random variable Z = X + Y can be given as follows. F Z (z) = P(X + Y z) = = [ z x ] f X (x)f Y (y) dy dx f X (x)f Y (z x) dx, where F Y is the cdf of Y. By differentiating F Z (z), we can obtain the pdf f Z (z) of Z as f Z (z) = d ( ) d dz F Z(z) = f X (x) dz F Y (z x) dx = f X (x)f Y (z x) dx. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 6 / 53

Convolution. The function of z (f g)(z) = f (x)g(z x) dx is called the convolution of f and g. Thus, the pdf f Z (z) for Z = X + Y is given by the convolution of the pdf s f X (x) and f Y (y). Then we can easily verify [ ] M Z (t) = e tz f X (x)f Y (z x) dx dz = M X (t)m Y (t) Therefore, as we will see in the examples below, the construction of moment generating function is much easier in order to find the distribution of the sum of independent random variables. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 7 / 53

Example Let X and Y be independent normal random variables with the respective parameters (µ x, σ 2 x) and (µ y, σ 2 y). Find the distribution for Z = X + Y. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 8 / 53

Example Let X and Y be independent normal random variables with the respective parameters (µ x, σ 2 x) and (µ y, σ 2 y). Find the distribution for Z = X + Y. We can calculate ( σ 2 M Z (t) = M X (t) M Y (t) = exp x t 2 2 ( (σ 2 x + σy)t 2 2 = exp + (µ x + µ y )t 2 ) ( σ 2 y t 2 + µ x t exp 2 ) ) + µ y t which is the mgf of normal distribution with parameter (µ x + µ y, σ 2 x + σ 2 y). Thus, we find that Z is a normal random variable with parameter (µ x + µ y, σ 2 x + σ 2 y). Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 8 / 53

Example Let X and Y be independent gamma random variables with the respective parameters (α 1, λ) and (α 2, λ). Find the distribution for Z = X + Y. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 9 / 53

Example Let X and Y be independent gamma random variables with the respective parameters (α 1, λ) and (α 2, λ). Find the distribution for Z = X + Y. We can calculate M Z (t) = M X (t) M Y (t) = ( ) α1 ( ) α2 ( ) α1 +α λ λ λ 2 = λ t λ t λ t which is the mgf of gamma distribution with parameter (α 1 + α 2, λ). Thus, Z is a gamma random variable with parameter (α 1 + α 2, λ). Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 9 / 53

Chi-square distribution. The gamma distribution f (x) = 1 2 n/2 Γ(n/2) x n/2 1 e x/2, x 0 with parameter (n/2, 1/2) is called the chi-square distribution with n degrees of freedom (df). Since it has only one parameter n, the mgf M(t) of chi-square distribution can be simply expressed as M(t) = ( ) n/2 1/2 = (1 2t) n/2 1/2 t Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 10 / 53

Chi-square distribution of one degree of freedom. Let X be a standard normal random variable, and let Y = X 2 be the square of X. Then we can express the cdf F Y of Y as F Y (t) = P(X 2 t) = P( t X t) = Φ( t) Φ( t) in terms of the standard normal cdf Φ. By differentiating the cdf, we can obtain the pdf f Y as follows. f Y (t) = d dt F Y (t) = 1 2π t 1 2 e t 2 = 1 2 1/2 Γ(1/2) t 1/2 e t/2, where we use Γ(1/2) = π. Thus, the square X 2 of X is the chi-square random variable with 1 degree of freedom. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 11 / 53

Chi-square distribution with n df. Let X 1,..., X n be iid standard normal random variables. Since X 2 i s have the gamma distribution with parameter (1/2, 1/2), the sum Y = n i=1 X 2 i has the gamma distribution with parameter (n/2, 1/2). That is, the sum Y has the chi-square distribution with n degree of freedom. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 12 / 53

Student s t-distribution. Let X be a chi-square random variable with n degrees of freedom, and let Y be a standard normal random variable. Suppose that X and Y are independent. Then the distribution of the quotient Z = Y X /n is called the Student s t-distribution with n degrees of freedom. The pdf f Z (z) is given by f Z (z) = Γ((n + 1)/2) nπγ(n/2) (1 + z 2 /n) (n+1)/2, < z <. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 13 / 53

F-distribution. Let U and V be independent chi-square random variables with r 1 and r 2 degrees of freedom, respectively. Then the distribution of the quotient Z = U/r 1 V /r 2 is called the F-distribution with (r 1, r 2 ) degree of freedom. The pdf f Z (z) is given by f Z (z) = Γ((r ( 1 + r 2 )/2) Γ(r 1 /2)Γ(r 2 /2) (r 1/r 2 ) r1/2 z r 1/2 1 1 + r ) (r1 +r 2 )/2 1 z, 0 < z < r 2 Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 14 / 53

Quotient of two random variables. Let X and Y be independent random variables having the respective pdf s f X (x) and f Y (y). Then the cdf F Z (z) of the quotient can be computed as follows. Z = Y /X F Z (z) = P(Y /X z) = P(Y zx, X < 0) + P(Y zx, X > 0) 0 [ ] [ xz ] = f Y (y) dy f X (x) dx + f Y (y) dy f X (x) dx xz 0 By differentiating, we obtain f Z (z) = d dz F Z(z) = = 0 x f Y (xz)f X (x) dx. [ xf Y (xz)]f X (x) dx + 0 [xf Y (xz)]f X (x) dx Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 15 / 53

Derivation of t-distribution. Let X be a chi-square random variable with n degrees of freedom. Then the pdf of the random variable U = X /n is given by f U (x) = d dx P( X /n x) = d dx F X (nx 2 ) = 2nxf X (nx 2 ) = for t 0; otherwise, f U (t) = 0. n n/2 2 n/2 1 Γ(n/2) x n 1 e nx2 /2 Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 16 / 53

Derivation of t-distribution, continued. Suppose that Y is a standard normal random variable and independent of X. Then the quotient Z = Y X /n has a t-distribution with n degrees of freedom. The pdf f Z (z) can be computed as follows. = = n n/2 x f Y (xz)f U (x) dx = 2 (n 1)/2 x n e (z2 +n)x 2 /2 dx πγ(n/2) 0 1 (1 + z 2 /n) (n+1)/2 t (n 1)/2 e t dt nπγ(n/2) Γ((n + 1)/2) nπγ(n/2) (1 + z 2 /n) (n+1)/2 0 for < z <. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 17 / 53

Derivation of F-distribution. Let U and V be independent chi-square random variables with r 1 and r 2 degrees of freedom, respectively. Then the quotient Z = U/r 1 V /r 2 has F -distribution with (r 1, r 2 ) degree of freedom. By putting Y = U/r 1 and X = V /r 2, we can compute the pdf f Z (z) as follows. = = x f Y (xz)f X (x) dx r r 1/2 1 r r 2/2 2 2 (r 1+r 2 )/2 Γ(r 1 /2)Γ(r 2 /2) z r 1/2 1 r r 1/2 1 r r 2/2 2 Γ(r 1 /2)Γ(r 2 /2) (r 1z + r 2 ) (r 1+r 2 )/2 z r 1/2 1 = Γ((r 1 + r 2 )/2) Γ(r 1 /2)Γ(r 2 /2) (r 1/r 2 ) r 1/2 z r 1/2 1 0 x (r 1+r 2 )/2 1 e (r 1z+r 2 )x/2 dx ( 1 + r 1 r 2 z 0 t (r 1+r 2 )/2 1 e t dt ) (r1 +r 2 )/2, 0 < z <. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 18 / 53

Population and random sample. The data values recorded x 1,..., x n are typically considered as the observed values of iid random variables X 1,..., X n having a common probability distribution f (x). Specifically we may assume that f (x) is a normal distribution with parameter (µ, σ 2 ). In a typical statistical problem it is useful to envisage a population from which the sample should be drawn. A random sample is chosen at random from the population, ensuring that the sample is representative of the population. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 19 / 53

Sample mean. Let X 1,..., X n be iid normal random variables with parameter (µ, σ 2 ). The function of random sample n X = 1 n i=1 X i is called the sample mean. Then we obtain E[ X ] = µ and Var( X ) = σ2 n. Furthermore, X has the normal distribution with parameter (µ, σ 2 /n). Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 20 / 53

Sample variance. The function of random sample S 2 = 1 n 1 n (X i X ) 2 i=1 is called the sample variance. The sample variance S 2 can be rewritten as ( n ) ( n ) S 2 = 1 Xi 2 n n 1 X 2 = 1 (X i µ) 2 n( n 1 X µ) 2. i=1 i=1 In particular, we can obtain E[S 2 ] = 1 n 1 = 1 n 1 ( n ) E[(X i µ) 2 ] ne[( X µ) 2 ] i=1 ( n ) Var(X i ) nvar( X ) = σ 2. i=1 Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 21 / 53

Theorem X has the normal distribution with parameter (µ, σ 2 /n), and X and S 2 are independent. Furthermore, W n 1 = (n 1)S 2 σ 2 = 1 σ 2 n (X i X ) 2 i=1 has the chi-square distribution with (n 1) degrees of freedom. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 22 / 53

Rough explanation of Theorem 3. Assuming that X and S 2 are independent (which is not so easy to prove), we can show that the random variable W n 1 has the chi-square distribution with (n 1) degrees of freedom. By introducing V = 1 σ 2 n i=1 (X i µ) 2 and W 1 = n σ 2 ( X µ) 2, we can write V = W n 1 + W 1. Since W n 1 and W 1 are independent, the mgf M V (t) of V can be expressed as M V (t) = M Wn 1 (t) M W1 (t), where M Wn 1 (t) and M W1 (t) are the mgf s of W n 1 and W 1, respectively. Observe that W 1 has the chi-square distribution with 1 degrees of freedom. Thus, we can obtain M Wn 1 (t) = M V (t) M W1 = (1 2t) n/2 = (1 2t) (n 1)/2 (1 2t) 1/2 which is the mgf of the chi-square distribution with (n 1) degrees of freedom. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 23 / 53

t-statistic. Assuming iid normal random variables X 1,..., X n with parameter (µ, σ 2 ), we can construct a standard normal random variable Z 1 = X µ σ/ and a chi-square random variable W n n 1 of Theorem 3. Furthermore, the function of random sample ) T = X µ S/ n = ( X µ σ/ n (n 1)S 2 /σ 2 n 1 = Z 1 W n 1 n 1 has the t-distribution with (n 1) degrees of freedom, and it is called t-statistic. The statistic S = S 2 is called the sample standard deviation. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 24 / 53

Sampling distributions under normal assumption. 1 The sample mean X is normally distributed with parameter (µ, σ 2 /n), and n( X µ) Z 1 = N(0, 1) σ 2 The statistic W n 1 = 1 n (X σ 2 i X ) 2 2 (n 1)S = χ 2 n 1 i=1 has a chi-square distribution with (n 1) df, and independent of X (Theorem 3). 3 The t-statistic has t-distribution with (n 1) df X µ S/ n = n( X µ) σ (n 1)S 2 /σ 2 n 1 n 1 Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 25 / 53 σ 2 = Z 1 W n 1 t n 1

Critical points. Given a random variable X, the critical point for level α is defined as the value c satisfying P(X > c) = α. 1 Suppose that a random variable X has the chi-square distribution with n degrees of freedom. Then the critical point of chi-square distribution, denoted by χ α,n, is defined by P(X > χ α,n ) = α. 2 Suppose that a random variable Z has the t-distribution with m degrees of freedom. Then the critical point is defined as the value t α,m satisfying P(Z > t α,m ) = α. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 26 / 53

Critical points, continued. Suppose that a random variable Z has the t-distribution with m degrees of freedom. Since the t-distribution is symmetric, the critical point t α,m can be equivalently given by P(Z < t α,m ) = α. Together we can obtain the expression P( t α/2,m Z t α/2,m ) = 1 P(Z < t α/2,m ) P(Z > t α/2,m ) = 1 α. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 27 / 53

Extrema. Let X 1,..., X n be independent and identically distributed random variables (iid random variables) with the common cdf F (x) and the pdf f (x). Then we can define the random variable V = min(x 1,..., X n ) of the minimum of X 1,..., X n. To find the distribution of V, consider the survival function P(V > x) of V and calculate as follows. P(V > x) = P(X 1 > x,..., X n > x) = P(X 1 > x) P(X n > x) = [1 F (x)] n. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 28 / 53

Extrema, continued. Since F V (x) = 1 P(V > x), we obtain the cdf F V (x) and the pdf f V (x) of V F V (x) = 1 [1 F (x)] n and f V (x) = d dx F V (x) = nf (x)[1 F (x)] n 1. In particular, if X 1,..., X n be independent exponential random variables with the common parameter λ, then the minimum min(x 1,..., X n ) has the density (nλ)e (nλ)x which is again the exponential density function with parameter (nλ). Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 29 / 53

Order statistics. Let X 1,..., X n be iid random variables with the common cdf F (x) and the pdf f (x). When we sort X 1,..., X n as X (1) < X (2) < < X (n), the random variable X (k) is called the k-th order statistic. Theorem The pdf of the k-th order statistic X (k) is given by f k (x) = n! (k 1)!(n k)! f (x)f k 1 (x)[1 F (x)] n k. (1.1) Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 30 / 53

Proof of theorem: Infinitesimal probability. If f is continuous at x and δ > 0 is small, then the probability that a random variable X falls into the small interval [x, x + δ] is approximated by δf (x). In differential notation, we can write P(x X x + dx) = f (x)dx. Now consider the k-th order statistic X (k) and the event that x X (k) x + dx. This event occurs when (k 1) of X i s are less than x and (n k) of X i s are greater than x x + dx. Since there are ( n k 1,1,n k) arrangements for Xi s, we can obtain the probability of this event as f k (x)dx = which implies (1.1). n! (k 1)!(n k)! (F (x))k 1 (f (x)dx)[1 F (x)] n k, Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 31 / 53

Beta distribution. When X 1,..., X n be iid uniform random variables on [0, 1], the pdf of the k-th order statistic X (k) becomes f k (x) = n! (k 1)!(n k)! x k 1 (1 x) n k, 0 x 1, which is known as the beta distribution with parameters α = k and β = n k + 1. Note #1 Probability and Statistics II/February 6, 2018 Sampling Distributions 32 / 53

Exercises Note #1 Probability and Statistics II/February 6, 2018 Exercises 33 / 53

Problem Consider a sample X 1,..., X n of normally distributed random variables with variance σ 2 = 5. Suppose that n = 31. 1 What is the value of c for which P(S 2 c) = 0.90? 2 What is the value of c for which P(S 2 c) = 0.95? Note #1 Probability and Statistics II/February 6, 2018 Exercises 34 / 53

Problem Consider a sample X 1,..., X n of normally distributed random variables with mean µ. Suppose that n = 16. 1 What is the value of c for which P( 4( X µ)/s c) = 0.95? 2 What is the value of c for which P( 4( X µ)/s c) = 0.99? Note #1 Probability and Statistics II/February 6, 2018 Exercises 35 / 53

Problem Suppose that X 1,..., X 10 are iid normal random variables with parameters µ = 0 and σ 2 > 0. 1 Let Z = X 1 + X 2 + X 3 + X 4. Then what is the distribution for Z? 2 Let W = X 2 5 + X 2 6 + X 2 7 + X 2 8 + X 2 9 + X 2 10. Then what is the distribution for W σ 2? 3 Describe the distribution for the random variable 4 Find the value c for which P ( ) Z W c = 0.95. Z/2 W /6. Note #1 Probability and Statistics II/February 6, 2018 Exercises 36 / 53

Problem Suppose that system s components are connected in series and have lifetimes that are independent exponential random variables X 1,..., X n with respective parameters λ 1,..., λ n. 1 Let V be the lifetime of the system. Show that V = min(x 1,..., X n ). 2 Show that the random variable V is exponentially distributed with parameter n i=1 λ i. Note #1 Probability and Statistics II/February 6, 2018 Exercises 37 / 53

Problem Each component A i of the following system has lifetime that is iid exponential random variable X i with parameter λ. A 1 A 3 A 5 A 2 A 4 A 6 1 Let V be the lifetime of the system. Show that V = max(min(x 1, X 2 ), min(x 3, X 4 ), min(x 5, X 6 )). 2 Find the cdf and the pdf of the random variable V. Note #1 Probability and Statistics II/February 6, 2018 Exercises 38 / 53

Problem A card contains n chips and has an error-correcting mechanism such that the card still functions if a single chip fails but does not function if two or more chips fail. If each chip has an independent and exponentially distributed lifetime with parameter λ, find the pdf of the card s lifetime. Note #1 Probability and Statistics II/February 6, 2018 Exercises 39 / 53

Problem Suppose that a queue has n servers and that the length of time to complete a job is an exponential random variable. If a job is at the top of the queue and will be handled by the next available server, what is the distribution of the waiting time until service? What is the distribution of the waiting time until service of the next job in the queue? Note #1 Probability and Statistics II/February 6, 2018 Exercises 40 / 53

Optional problems Note #1 Probability and Statistics II/February 6, 2018 Optional problems 41 / 53

Joint distribution of order statistics. Let U and V be iid uniform random variable on [0, 1], and let X = min(u, V ) and Y = max(u, V ). We have discussed that the probability P(x X x + dx) is approximated by f (x) dx. Similarly, we can find the following formula in the case of joint distribution. P(x X x + dx, y Y y + dy) = f (x, y) dx dy. In particular, we can obtain P(x U x + dx, y V y + dy) = dx dy. Note #1 Probability and Statistics II/February 6, 2018 Optional problems 42 / 53

Problem By showing that P(x X x + dx, y Y y + dy) = P(x U x + dx, y V y + dy) + P(x V x + dx, y U y + dy), justify that the joint density function f (x, y) of X and Y is given by { 2 if 0 x y 1; f (x, y) = 0 otherwise. Note #1 Probability and Statistics II/February 6, 2018 Optional problems 43 / 53

Problem Continue Problem 12, and answer the following questions. 1 Find the conditional density f X Y (x y) and f Y X (y x). 2 Find the conditional expectations E(X Y = y) and E(Y X = x). Note #1 Probability and Statistics II/February 6, 2018 Optional problems 44 / 53

Problem Let f (x, y) = {( n x) y x (1 y) n x if x = 0, 1,..., n and 0 y 1; 0 otherwise. 1 Find the marginal density f Y (y) = n x=0 f (x, y) for 0 y 1, and the conditional frequency function f X Y (x θ) given Y = θ. Then identify the distribution for f X Y (x θ). 2 Find the marginal freqency function f X (x) = 1 f (x, y) dy. 3 Given x = k, find the conditional density and f Y X (y k), and identify the distribution for f Y X (y k). In Bayesian statistics (later in this semester) f Y (y) is called a prior distribution of parameter θ of binomial distribution (n, θ), and f Y X (y k) is a posterior distribution of parameter θ provided that we have observed X = k from the binomial distribution. Note #1 Probability and Statistics II/February 6, 2018 Optional problems 45 / 53 0

Answers to exercises Note #1 Probability and Statistics II/February 6, 2018 Answers to exercises 46 / 53

Problem 5. Observe that 30S 2 /5 = 6S 2 has a χ 2 -distribution with 30 degrees of freedom. (a) P(S 2 c) = P(6S 2 6c) = 0.90 implies 6c = χ 2 0.1,30 40.26. Thus, c 6.7. (b) P(S 2 c) = P(6S 2 6c) = 0.95 implies 6c = χ 2 0.05,30 43.77. Thus, c 7.3. Note #1 Probability and Statistics II/February 6, 2018 Answers to exercises 47 / 53

Problem 6. Observe that 4( X µ) S degrees of freedom. (a) c = t 0.025,15 2.13. (b) c = t 0.005,15 2.95. = X µ S/ 16 has a t-distribution with 15 Note #1 Probability and Statistics II/February 6, 2018 Answers to exercises 48 / 53

Problem 7. (a) N(0, 4σ 2 ). (b) Since W /σ 2 = (X 5 /σ) 2 + + (X 10 /σ) 2 and each (X i /σ) has the standard normal distribution, W /σ 2 has the chi-square distribution with 6 degrees of freedom. (c) We can write Z/2 W /6 = Z/(2σ). Since Z/(2σ) is (W /σ2 )/6 independent of W /σ 2, and has the standard normal distribution, Z/2 has the t-distribution with 6 degree of freedom. W /6 (d) Since P ( ) Z W c = P ( Z/2 W /6 6 2 c implies that 6 2 c = t 0.05,6 1.94. Thus, c 1.58. ), the result of (c) Note #1 Probability and Statistics II/February 6, 2018 Answers to exercises 49 / 53

Problem 8. (b) P(V > t) = P(X 1 > t) P(X 1 > t) = e λ1t e λnt = exp ( n i=1 λ i) is the survival function of V. Therefore, V is an exponential random variable with parameter n i=1 λ i. Note #1 Probability and Statistics II/February 6, 2018 Answers to exercises 50 / 53

Problem 9. (b) Let Y 1 = min(x 1, X 2 ), Y 2 = min(x 3, X 4 ), and Y 3 = min(x 5, X 6 ). Then Y 1, Y 2 and Y 3 are iid exponential random variables with parameter 2λ. Since V = Y (3), we obtain f V (t) = 6λe 2λt (1 e 2λt ) 2, t > 0. By integrating it, we can find the cdf F V (t) = (1 e 2λt ) 3. Note #1 Probability and Statistics II/February 6, 2018 Answers to exercises 51 / 53

Problem 10. Let X i denote the lifetime of the i-th chip, and let V denote the lifetime of the card. Then X i s are iid exponential random variables with λ, and V = X (2). Thus, we obtain f V (t) = n(n 1)λ[e λ(n 1)t e λnt ], t > 0. Note #1 Probability and Statistics II/February 6, 2018 Answers to exercises 52 / 53

Problem 11. Each of n servers completes a job independently in an exponentially distributed time with λ. Then it is known that the sojourn time between the jobs is exponential with nλ, and that the sojourn times are independent each other. Therefore, (i) the distribution of the waiting time at the top of the queue is exponential with nλ, and (ii) the distribution of the waiting time for the second in the queue is given by the sum of two iid exponential random variables, that is, the gamma distribution with (2, nλ). Note #1 Probability and Statistics II/February 6, 2018 Answers to exercises 53 / 53