n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2

Similar documents
n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2

Sampling Distributions

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

1 Generating functions

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

Generating and characteristic functions. Generating and Characteristic Functions. Probability generating function. Probability generating function

2 Random Variable Generation

Gaussian vectors and central limit theorem

Formulas for probability theory and linear models SF2941

Lecture 5: Moment generating functions

Sampling Distributions

Solutions to Problem Set 4

STAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Random Variables and Their Distributions

BMIR Lecture Series on Probability and Statistics Fall 2015 Discrete RVs

3. Probability and Statistics

P (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n

FINAL EXAM: Monday 8-10am

7 Random samples and sampling distributions

Brief Review of Probability

1 Review of Probability

Introduction to Probability. Ariel Yadin. Lecture 10. Proposition Let X be a discrete random variable, with range R and density f X.

Week 2. Review of Probability, Random Variables and Univariate Distributions

DO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO

Limiting Distributions

4. Distributions of Functions of Random Variables

Three hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER.

THE QUEEN S UNIVERSITY OF BELFAST

1 Solution to Problem 2.1

18.440: Lecture 28 Lectures Review

Exercises and Answers to Chapter 1

Probability and Distributions

Selected Exercises on Expectations and Some Probability Inequalities

18.440: Lecture 28 Lectures Review

Things to remember when learning probability distributions:

Chapter 5. Chapter 5 sections

PCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities

1 Random Variable: Topics

Twelfth Problem Assignment

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

Expectation. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Analysis of Engineering and Scientific Data. Semester

Random Variables. P(x) = P[X(e)] = P(e). (1)

Lecture Notes 2 Random Variables. Discrete Random Variables: Probability mass function (pmf)

General Random Variables

Basic concepts of probability theory

MAS223 Statistical Inference and Modelling Exercises

t x 1 e t dt, and simplify the answer when possible (for example, when r is a positive even number). In particular, confirm that EX 4 = 3.

Lecture 1: August 28

1.1 Review of Probability Theory

We introduce methods that are useful in:

Expectation. DS GA 1002 Probability and Statistics for Data Science. Carlos Fernandez-Granda

Lecture 17: The Exponential and Some Related Distributions

STAT/MATH 395 PROBABILITY II

8 Laws of large numbers

Continuous random variables

Probability Models. 4. What is the definition of the expectation of a discrete random variable?

Lecture 1 Measure concentration

5. Random Vectors. probabilities. characteristic function. cross correlation, cross covariance. Gaussian random vectors. functions of random vectors

UC Berkeley Department of Electrical Engineering and Computer Sciences. EECS 126: Probability and Random Processes

Basic concepts of probability theory

STAT Chapter 5 Continuous Distributions

CS145: Probability & Computing

Probability and Measure

The Multivariate Normal Distribution. In this case according to our theorem

Chapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University

the convolution of f and g) given by

LIST OF FORMULAS FOR STK1100 AND STK1110

Limiting Distributions

E[X n ]= dn dt n M X(t). ). What is the mgf? Solution. Found this the other day in the Kernel matching exercise: 1 M X (t) =

Chapter 4. Continuous Random Variables 4.1 PDF

Statistics STAT:5100 (22S:193), Fall Sample Final Exam B

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Continuous Random Variables and Continuous Distributions

Statistical Methods in Particle Physics

Lecture Notes 2 Random Variables. Random Variable

Continuous Random Variables

Preliminaries. Probability space

Generation from simple discrete distributions

conditional cdf, conditional pdf, total probability theorem?

APPM/MATH 4/5520 Solutions to Exam I Review Problems. f X 1,X 2. 2e x 1 x 2. = x 2

Weak convergence in Probability Theory A summer excursion! Day 3

Random Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline.

STAT 512 sp 2018 Summary Sheet

ECE 302 Division 2 Exam 2 Solutions, 11/4/2009.

FINAL EXAM: 3:30-5:30pm

Statistics, Data Analysis, and Simulation SS 2015

X n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2)

Spring 2012 Math 541B Exam 1

Hints/Solutions for Homework 3

Susceptible-Infective-Removed Epidemics and Erdős-Rényi random

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

Exponential Distribution and Poisson Process

or E ( U(X) ) e zx = e ux e ivx = e ux( cos(vx) + i sin(vx) ), B X := { u R : M X (u) < } (4)

Chapter 4. Chapter 4 sections

ACM 116: Lectures 3 4

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

Continuous Random Variables

Transcription:

Order statistics Ex. 4. (*. Let independent variables X,..., X n have U(0, distribution. Show that for every x (0,, we have P ( X ( < x and P ( X (n > x as n. Ex. 4.2 (**. By using induction or otherwise, prove (4.6, P(X ( x n! (!(n! F (x 0 y ( y n dy. Ex. 4.3 (*. Derive the density f X( (x from (4.7 by differentiating (4.6. Ex. 4.4 (*. If X has continuous cdf F (, show that Y F (X U(0,. Ex. 4.5 (**. In the case of an n-sample from U(0, distribution, derive (4.8, f X(,X (n (x, y n(n ( F (y F (x n 2 f(xf(yx<y, directly from combinatorics (cf. the second proof of Corollary 4.3, and then use the approach in Remar 4.3. to extend your result to the general case. Ex. 4.6 (*. Prove the density f Rn (r formula (4.9, f Rn (r n(n ( F (z + r F (z n 2 f(zf(z + r dz. Ex. 4.7 (*. Let X β(, m, ie., X has beta distribution with parameters and m. Show that EX VarX m (+m 2 (+m+. +m and Ex. 4.8 (**. Let X ( be the first order variable from an n-sample with density f(, which is positive and continuous on [0, ], and vanishes otherwise. Let, further, f(0 c > 0. For fixed y > 0, show that P ( X ( > y n e cy for large n. Deduce that the distribution of Y n nx ( is approximately Exp(c for large enough n. Ex. 4.9 (**. Let X,..., X n be independent positive random variables whose joint probability density function f( is right-continuous at the origin and satisfies f(0 c > 0. For fixed y > 0, show that P ( X ( > n y e cy for large n. Deduce that the distribution of Y n nx ( is approximately Exp(c for large enough n. Ex. 4.0 (**. Let X, X 2, X 3, X 4 be a sample from U(0,, and let X (, X (2, X (3, X (4 be the corresponding order statistics. Find the pdf for each of the random variables: X (2, X (3 X (, X (4 X (2, and X (3. Ex. 4. (***. Prove the asymptotic independence property of any finite collection of gaps stated in Remar 4.2.. Ex. 4.2 (***. Using induction or otherwise, prove (4.3, describing the joint gaps distribution: for all positive r satisfying n+ r. P ( n+ n ( X r,..., (n+ X r n+ ( r, Ex. 4.3 (*. Let X Exp(,,..., n, be independent with fixed > 0. Denote X 0 min{x,..., X n } and 0 n. Show that for y 0 we have P ( X 0 > y, X 0 X e y 0, ie., the minimum X 0 of the sample satisfies X 0 Exp( 0 and the probability that it coincides with X is proportional to, independently of the value of X 0. Ex. 4.4 (*. If X Exp(, show that for all positive a and b we have P ( X > a + b X > a P(X > b. Ex. 4.5 (**. If X Exp(, Y Exp(µ and Z Exp(ν are independent, show that for every constant a 0 we have P(a + X < Y a < Y +µ ; deduce that P( a + X < min(y, Z a < min(y, Z +µ+ν. Ex. 4.6 (**. Carefully prove Corollary 4.6 and compute EY n and VarY n. Ex. 4.7 (**. Let X (n be the maximum of an n-sample from Exp( distribution. For x R, find the value of P ( X (n log n + x in the limit n.

Ex. 4.8 (*. Let X, X 2 be a sample from a uniform distribution on {, 2, 3, 4, 5}. Find the distribution of X (, the minimum of the sample. Ex. 4.9 (*. Let independent variables X,..., X n be Exp( distributed. Show that for every x > 0, we have P ( X ( x and P ( X (n x as n. Generalise the result to arbitrary distributions on R. Ex. 4.20 (*. Let {X, X 2, X 3, X 4 } be a sample from a distribution with density f(x e 7 x x>7. Find the pdf of the second order variable X (2. Ex. 4.2 (*. Let X and X 2 be independent Exp( random variables. a Show that X ( and X (2 X ( are independent and find their distributions. b Compute E(X (2 X ( x and E(X ( X (2 x 2. Ex. 4.22 (**. Let (X n be an n-sample from Exp( distribution. a Show that the gaps ( ( X n as ined in Lemma 4.0 are independent and find their distribution. b For fixed m n, compute the expectation E(X (m X ( x. Ex. 4.23 (**. If Y Exp(µ and an arbitrary random variable X 0 are independent, show that for every a > 0, P(a + X < Y a < Y Ee µx. Order statistics: optional problems Ex. 4.24 (**. In the context of Ex. 4.8, let {X, X 2} be a sample without replacement from {, 2, 3, 4, 5}. Find the distribution of X (, the minimum of the sample. Ex. 4.25 (*. Let {X, X 2} be an independent sample from Geom(p distribution, P(X > ( p for integer 0. Find the distribution of X (, the minimum of the sample. Ex. 4.26 (**. Let {X, X 2,..., X n} be an n-sample from a distribution with density f(. Show that the joint density of the order variables X (,..., X (n is f X(,...,X (n (x,..., x n n! n f(x x < <x n. Ex. 4.27 (**. Let {X, X 2, X 3} be a sample from U(0,. Find the conditional density f X(,X (3 X (2 (x, z y of X ( and X (3 given that X (2 y. Explain your findings. Ex. 4.28 (**. Let {X, X 2,..., X 00} be a sample from U(0,. Approximate the value of P(X (75 0.8. Ex. 4.29 (**. Let X, X 2,... be independent random variables with cdf F (, and let N > 0 be an integer-valued variable with probability generating function g(, independent of the sequence (X. Find the cdf of max { X, X 2,..., X N }, the maximum of the first N terms in that sequence. Ex. 4.30 (***. Let X ( be the first order variable from an n-sample with density f(, which is positive and continuous on [0, ], and vanishes otherwise. Let, further, f(x cx α for small x > 0 and positive c and α. For y > 0 and β, α+ show that the probability P(X ( > yn β has a well ined limit for large n. What can you deduce about the distribution of the rescaled variable Y n n β X ( for large enough n? Ex. 4.3 (***. Let X,..., X n be independent β(, m-distributed random variables whose joint distribution is given in (4. (with and m. Find δ > 0 such that the distribution of the rescaled variable Y n n δ X ( converges to a well-ined limit as n. Describe the limiting distribution. Ex. 4.32 (****. In the situation of Ex. 4.30, let α < 0. What can you say about possible limiting distribution of the suitably rescaled fisrt order variable, Y n n δ X (, with some δ R? Ex. 4.33 (****. Denote Y n Y n log n; show that the corresponding cdf, P(Y n x, approaches e e x, as n. Deduce that the expectation of the limiting distribution equals γ, the Euler constant, and its variance is π2. 2 6 A distribution with cdf exp { e (x µ/β} is nown as Gumbel distribution (with scale and locations parameters β and µ, resp.. It can be shown that its average is µ + βγ, its variance is π 2 β 2 /6, and its moment generating function equals Γ( βte µt. Ex. 4.34 (****. By using the Weierstrass formula, ( + z e z/ e γz Γ(z +, (where γ is the Euler constant and Γ( is the classical gamma function, Γ(n (n! for integer n > 0 or otherwise, show that the moment generating function Ee tz n of Zn Z n EZ n approaches e γt Γ( t as n (eg., for all t <. Deduce that in that limit Zn + γ is asymptotically Gumbel distributed (with β and µ 0. Ex. 4.35 (***. Let X, X 2,... be independent Exp( random variables; further, let N Poi(ν, independent of the sequence (X, and let X 0 0. Find the distribution of Y max{x 0, X, X 2,..., X N }, the maximum of the first N terms of this sequence, where for N 0 we set Y 0. the Euler constant γ is lim n ( n log n 0.577256649...; 2

Coupling Ex. 5. (*. In the setting of Example 5. show that every convex linear combination of tables T and T 2, ie., each table of the form T α αt + ( αt 2 with α [0, ], gives an example of a coupling of X Ber(p and Y Ber(p 2. Can you find all possible couplings for these variables? Ex. 5.2 (*. Let X 0 be a random variable, and let a 0 be a fixed constant. If Y a + X, is it true that X Y? If Z ax, is it true that X Z? Justify your answer. Ex. 5.3 (*. If X Y and g( is an arbitrary increasing function on R, show that g(x g(y. Ex. 5.4 (*. Generalise the inequality in Example 5.3 to a broader class of functions g( and verify that if X Y, then E(X 2+ E(Y 2+, Ee tx Ee ty with t > 0, Es X Es Y with s > etc. Ex. 5.5 (*. Let ξ U(0, be a standard uniform random variable. For fixed p (0,, ine X ξ<p. Show that X Ber(p, a Bernoulli random variable with parameter p. Now suppose that X ξ<p and Y ξ<p2 for some 0 < p p 2 < and ξ as above. Show that X Y and that P(X Y. Compare your construction to the second table in Example 5.. Ex. 5.6 (**. In the setup of Example 5., show that X Ber(p is stochastically smaller than Y Ber(p 2 (ie., X Y iff p p 2. Further, show that X Y is equivalent to existence of a coupling ( X, Ỹ of X and Y in which these variables are ordered with probablity one, P( X Ỹ. Ex. 5.7 (**. Let X N (µ X, σx 2 and Y N (µ Y, σy 2 be Gaussian r.v. s. a If µ X µ Y but σx 2 σ2 Y, is it true that X Y? b If µ X µ Y but σx 2 σ2 Y, is it true that X Y? Ex. 5.8 (**. Let X Exp( and Y Exp(µ be two exponential random variables. If 0 < µ <, are the variables X and Y stochastically ordered? Justify your answer by proving the result or giving a counter-example. Ex. 5.9 (**. Let X Geom(p and Y Geom(r be two geometric random variables, X Geom(p and Y Geom(r. If 0 < p r <, are the variables X and Y stochastically ordered? Justify your answer by proving the result or giving a counter-example. The gamma distribution Γ(a, has density a Γ(a xa e x x>0 (so that Γ(, is just Exp(. Gamma distributions have the following additive property: if Z Γ(c, and Z 2 Γ(c 2, are independent random variables (with the same, then their sum is also gamma distributed: Z + Z 2 Γ(c + c 2,. Ex. 5.0 (**. Let X Γ(a, and Y Γ(b, be two gamma random variables. If 0 < a b <, are the variables X and Y stochastically ordered? Justify your answer by proving the result or giving a counter-example. Ex. 5. (**. Let measure µ have p.m.f. {p x } x X and let measure ν have p.m.f. {q y } y Y. Show that ( d TV (µ, ν d TV {p}, {q} max A µ(a ν(a 2 z pz q z, where the sum runs over all z X Y. Deduce that d TV (, is a distance between probability measures(ie., it is non-negative, symmetric, and satisfies the triangle inequality such that d TV (µ, ν for all probability measures µ and ν. Ex. 5.2 (**. In the setting of Ex. 5. show that d TV ( {p}, {q} z( pz min(p z, q z z( qz min(p z, q z. Ex. 5.3 (**. If P(, is the maximal coupling of X and Y as ined in Example 5.9 and Remar 5.9., show that P( X Ỹ d TV( X, Y. Ex. 5.4 (**. For fixed p (0,, complete the construction of the maximal coupling of X Ber(p and Y Poi(p as outlined in Example 5.2. Is any of the variables X and Y stochastically dominated by another? Ex. 5.5 (**. Let X Bin(n, n and Y Poi( for some > 0. Show that 2 ( P(X P(Y dtv X, Ỹ 2 /n for every 0. Deduce that for every fixed 0, we have P(X! e as n. 3

Ex. 5.6 (*. Let random variable X have density f(, and let g( be a smooth increasing function with g(x 0 as x. Show that Eg(X dx z<x g (zf(x dz and deduce the integral representation of Eg(X in (5.2, Eg(X 0 g (zp(x > z dz. Ex. 5.7 (**. Let X Bin(, p and Y Bin(2, p 2 with p p 2. By constructing an explicit coupling or otherwise, show that X Y. Ex. 5.8 (***. that X Y. Let X Bin(2, p and Y Bin(2, p 2 with p p 2. Construct an explicit coupling showing Ex. 5.9 (***. Let X Bin(3, p and Y Bin(3, p 2 with 0 < p p 2 <. Construct an explicit coupling showing that X Y. Ex. 5.20 (***. Let X Bin(2, p and Y Bin(4, p with 0 < p <. Construct an explicit coupling showing that X Y. Ex. 5.2 (**. Let m n and 0 < p p 2 <. Show that X Bin(m, p is stochastically smaller than Y Bin(n, p 2. Ex. 5.22 (**. Let X Ber(p and Y Poi(ν. Characterise all pairs (p, ν such that X Y. Is it true that X Y if p ν? Is it possible to have Y X? Ex. 5.23 (**. Let X Bin(n, p and Y Poi(ν. Characterise all pairs (p, ν such that X Y. Is it true that X Y if np ν? Is it possible to have Y X? Ex. 5.24 (**. Prove the additive property of d TV (, for independent sums, (5.7, by following the approach in Remar 5.3.. Ex. 5.25 (***. Generalise your argument in Ex. 5.24 for general sums, and derive an alternative proof of Theorem 5.4. Ex. 5.26 (**. If X, Y, Z are independent with X Y, show that (X + Z (Y + Z. Ex. 5.27 (*. If X, Y, Z are independent with X Y and Y Z, show that X Z. Ex. 5.28 (**. Let X X + X 2 with independent X and X 2, similarly, Y Y + Y 2 with independent Y and Y 2. If X Y and X 2 Y 2, deduce that X Y. Ex. 5.29 (**. Let X Ber(p and Y Ber(p for fixed 0 < p < p < and all 0,,..., n. Show that X X is stochastically smaller than Y Y. 4

Some non-classical limits Ex. 6. (*. Let X n Bin(n, p, where p p(n is such that np (0, as n. Show that for every fixed s R the generating function G n (s E ( ( n s Xn + p(s converges, as n, to GY (s e (s, the generating function of Y Poi(. Deduce that the distribution of X n approaches that of Y in this limit. Ex. 6.2 (*. Show that x + log( x x 2 uniformly in x /2. Ex. 6.3 (**. Use the estimate in Ex. 6.2 to prove Theorem 6.: if X Ber ( p (n,,..., n, are independent with max n p (n 0 and n p(n E ( n X(n (0, as n. Show that the generating function of Y n n X converges pointwise to that of Y Poi( in this limit. Ex. 6.4 (**. Let r (distinguishable balls be placed randomly into n boxes. Write N N (r, n for the number of boxes containing exactly balls. If r/n c as n, then E ( n N c! e and Var ( n N 0 as n. Ex. 6.5 (**. In the setup of Ex. 6.4, let X j be the number of balls in box j. Show that for each integer 0, we have P ( X j c! e c as n. Further, show that for i j, the variables X i and X j are not independent, but in the limit n they become independent Poi(c distributed random variables. Ex. 6.6 (**. Generalise the result in Ex. 6.5 for occupancy numbers of a finite number of boxes. Ex. 6.7 (**. Suppose that each box of cereal contains one of n different coupons. Assume that the coupon in each box is chosen independently and uniformly at random from the n possibilities, and let T n be the number of boxes of cereal one needs to buy before there is at least one of every type of coupon. Show that ET n n n n log n and VarT n n 2 n 2. Ex. 6.8 (**. In the setup of Ex. 6.7, show that for all real a, P(T n n log n + na exp{ e a }, as n. [Hint: If T n >, then balls are placed into n boxes so that at least one box is empty.] Ex. 6.9 (**. In a village of 365 people, what is the probability that all birthdays are represented? Find the answer for 6 and 5. [Hint: In notations of Ex. 6.7, the problem is about evaluating P(T n 365.] Ex. 6.0 (*. Show that the moment generating function of X Geom(p is M X (t Ee tx pe t ( ( pe t and that of Y Exp( is M Y (t t (ined for all t <. Let Z px and deduce that, as p 0, the moment generating function M Z (t converges to that of Y Exp( for each fixed t <. Ex. 6. (**. The running cost of a car between two consecutive services is given by a random variable C with moment generating function (MGF M C (t (and expectation c, where the costs over non-overlapping time intervals are assumed independent and identically distributed. The car is written off before the next service with (small probability p > 0. Show that the number T of services before the car is written off has truncated geometric distribution with MGF M T (t p ( ( pe t and deduce that the total running cost X of the car up to and including the final service has MGF M X (t p ( ( pm C (t. Find the expectation EX of X and show that for small enough p the distribution of X X/EX is close to Exp(. [Hint: Find the MGF of X and follow the approach of Ex. 6.0.] Ex. 6.2 (**. Let r (distinguishable balls be placed randomly into n boxes. Find a constant c > 0 such that with r c n the probability that no two such balls are placed in the same box approaches /e as n. Find a constant b > 0 such that with r b n the probability that no two such balls are placed in the same box approaches /2 as n. Notice that with n 365 and r 23 this is the famous birthday paradox. Ex. 6.3 (***. In the setup of Ex. 6.4, let X j be the number of balls in box j. Show that for every integer j 0, j,..., n, satisfying n j j r we have P ( ( r X,..., X n n n r r! ; 2 ;... ; n! 2!... n!n r. Now suppose that Y j Poi ( r n, j,..., n, are independent. Show that P ( Y j,..., Y n n j Y j r is given by the expression in the last display. Ex. 6.4 (***. Find the probability that in a class of 00 students at least three of them have the same birthday. Ex. 6.5 (****. There are 365 students registered for the first year probability class. Find such that the probability of finding at least students sharing the same birthday is about /2. Ex. 6.6 (***. Let G n,n, N ( n 2, be the collection of all graphs on n vertices connected by N edges. For a fixed constant c R, if N 2 (n log n + cn, show that the probability for a random graph in G n,n to have isolated vertices approached exp{ e c } as n. 5