Stein s Method Applied to Some Statistical Problems

Size: px
Start display at page:

Download "Stein s Method Applied to Some Statistical Problems"

Transcription

1 Stein s Method Applied to Some Statistical Problems Jay Bartroff Borchard Colloquium 2017 Jay Bartroff (USC) Stein s for Stats 4.Jul.17 1 / 36

2 Outline of this talk 1. Stein s Method 2. Bounds to the normal for group sequential statistics with covariates Produce explicit distributional bounds to the limiting normal distribution for repeatedly-computed MLE (ˆθ 1, ˆθ 2,..., ˆθ K ) of a parameter vector θ R p in group sequential trials 3. Concentration inequalities for occupancy models with log-concave marginals How to get bounded, size biased couplings for certain multivariate occupancy models, then use these to get concentration inequalities Joint work with Larry Goldstein (USC) and Ümit Işlak (Boğaziçi University, Istanbul) Jay Bartroff (USC) Stein s for Stats 4.Jul.17 2 / 36

3 Outline of this talk 1. Stein s Method 2. Bounds to the normal for group sequential statistics with covariates Produce explicit distributional bounds to the limiting normal distribution for repeatedly-computed MLE (ˆθ 1, ˆθ 2,..., ˆθ K ) of a parameter vector θ R p in group sequential trials 3. Concentration inequalities for occupancy models with log-concave marginals How to get bounded, size biased couplings for certain multivariate occupancy models, then use these to get concentration inequalities Joint work with Larry Goldstein (USC) and Ümit Işlak (Boğaziçi University, Istanbul) Jay Bartroff (USC) Stein s for Stats 4.Jul.17 2 / 36

4 Outline of this talk 1. Stein s Method 2. Bounds to the normal for group sequential statistics with covariates Produce explicit distributional bounds to the limiting normal distribution for repeatedly-computed MLE (ˆθ 1, ˆθ 2,..., ˆθ K ) of a parameter vector θ R p in group sequential trials 3. Concentration inequalities for occupancy models with log-concave marginals How to get bounded, size biased couplings for certain multivariate occupancy models, then use these to get concentration inequalities Joint work with Larry Goldstein (USC) and Ümit Işlak (Boğaziçi University, Istanbul) Jay Bartroff (USC) Stein s for Stats 4.Jul.17 2 / 36

5 Outline of this talk 1. Stein s Method 2. Bounds to the normal for group sequential statistics with covariates Produce explicit distributional bounds to the limiting normal distribution for repeatedly-computed MLE (ˆθ 1, ˆθ 2,..., ˆθ K ) of a parameter vector θ R p in group sequential trials 3. Concentration inequalities for occupancy models with log-concave marginals How to get bounded, size biased couplings for certain multivariate occupancy models, then use these to get concentration inequalities Joint work with Larry Goldstein (USC) and Ümit Işlak (Boğaziçi University, Istanbul) Jay Bartroff (USC) Stein s for Stats 4.Jul.17 2 / 36

6 Charles Stein Stein s paradox: X in admissable for θ in N p (θ, I) for p 3 James-Stein shrinkage estimator Stein s unbiased risk estimator for MSE Stein s Lemma (1): Covariance estimation Stein s Lemma (2): Sequential sample size tails Stein s method for distributional approximation Jay Bartroff (USC) Stein s for Stats 4.Jul.17 3 / 36

7 Application 1: Group sequential analysis Jay Bartroff (USC) Stein s for Stats 4.Jul.17 4 / 36

8 Application 1: Group sequential analysis Jay Bartroff (USC) Stein s for Stats 4.Jul.17 5 / 36

9 But there are other books on this subject... Jay Bartroff (USC) Stein s for Stats 4.Jul.17 6 / 36

10 Setup Response Y i R of ith patient depends on known covariate vector x i unknown parameter vector θ R p Primary goal: To test a null hypothesis about θ, e.g., H 0 : θ = 0 H 0 : θ j 0 H 0 : at θ = b, some vector a, scalar b Secondary goals: Compute p-values or confidence regions for θ at the end of study Jay Bartroff (USC) Stein s for Stats 4.Jul.17 7 / 36

11 Setup Response Y i R of ith patient depends on known covariate vector x i unknown parameter vector θ R p Primary goal: To test a null hypothesis about θ, e.g., H 0 : θ = 0 H 0 : θ j 0 H 0 : at θ = b, some vector a, scalar b Secondary goals: Compute p-values or confidence regions for θ at the end of study Jay Bartroff (USC) Stein s for Stats 4.Jul.17 7 / 36

12 Setup Response Y i R of ith patient depends on known covariate vector x i unknown parameter vector θ R p Primary goal: To test a null hypothesis about θ, e.g., H 0 : θ = 0 H 0 : θ j 0 H 0 : at θ = b, some vector a, scalar b Secondary goals: Compute p-values or confidence regions for θ at the end of study Jay Bartroff (USC) Stein s for Stats 4.Jul.17 7 / 36

13 Setup: Group sequential analysis For efficiency, ethical, practical, financial reasons, the standard in trials has become group sequential analysis A group sequential trial with at most K groups Group 1: Group 2:. Group K: Y 1,..., Y n1 Y n1 +1,..., Y n2 Y nk 1 +1,..., Y nk Group sequential dominant format for clinical trials since... Beta-Blocker Heart Attack Trial ( BHAT, JAMA 82) Randomized trial of propranolol for heart attack survivors 3837 patients randomized Started June 1978, planned as 4-year study, terminated 8 months early due to observed benefit of propranolol Jay Bartroff (USC) Stein s for Stats 4.Jul.17 8 / 36

14 Setup: Group sequential analysis For efficiency, ethical, practical, financial reasons, the standard in trials has become group sequential analysis A group sequential trial with at most K groups Group 1: Group 2:. Group K: Y 1,..., Y n1 Y n1 +1,..., Y n2 Y nk 1 +1,..., Y nk Group sequential dominant format for clinical trials since... Beta-Blocker Heart Attack Trial ( BHAT, JAMA 82) Randomized trial of propranolol for heart attack survivors 3837 patients randomized Started June 1978, planned as 4-year study, terminated 8 months early due to observed benefit of propranolol Jay Bartroff (USC) Stein s for Stats 4.Jul.17 8 / 36

15 Setup: Group sequential analysis For efficiency, ethical, practical, financial reasons, the standard in trials has become group sequential analysis A group sequential trial with at most K groups Group 1: Group 2:. Group K: Y 1,..., Y n1 Y n1 +1,..., Y n2 Y nk 1 +1,..., Y nk Group sequential dominant format for clinical trials since... Beta-Blocker Heart Attack Trial ( BHAT, JAMA 82) Randomized trial of propranolol for heart attack survivors 3837 patients randomized Started June 1978, planned as 4-year study, terminated 8 months early due to observed benefit of propranolol Jay Bartroff (USC) Stein s for Stats 4.Jul.17 8 / 36

16 Setup: Group sequential analysis Stopping rule related to H 0 likelihood ratio, t-, F-, χ 2 - tests common Of the form: Stop and reject H 0 at stage min{k K : T (Y 1,..., Y nk ) C k } for some statistic T (Y 1,..., Y nk ), often a function of the MLE ˆθ k = ˆθ k (Y 1,..., Y nk ) The joint distribution of needed to ˆθ 1, ˆθ 2,..., ˆθ K choose critical values C k compute p-value at end of study give confidence region for θ at end of study Jay Bartroff (USC) Stein s for Stats 4.Jul.17 9 / 36

17 Setup: Group sequential analysis Stopping rule related to H 0 likelihood ratio, t-, F-, χ 2 - tests common Of the form: Stop and reject H 0 at stage min{k K : T (Y 1,..., Y nk ) C k } for some statistic T (Y 1,..., Y nk ), often a function of the MLE ˆθ k = ˆθ k (Y 1,..., Y nk ) The joint distribution of needed to ˆθ 1, ˆθ 2,..., ˆθ K choose critical values C k compute p-value at end of study give confidence region for θ at end of study Jay Bartroff (USC) Stein s for Stats 4.Jul.17 9 / 36

18 Setup: Group sequential analysis Stopping rule related to H 0 likelihood ratio, t-, F-, χ 2 - tests common Of the form: Stop and reject H 0 at stage min{k K : T (Y 1,..., Y nk ) C k } for some statistic T (Y 1,..., Y nk ), often a function of the MLE ˆθ k = ˆθ k (Y 1,..., Y nk ) The joint distribution of needed to ˆθ 1, ˆθ 2,..., ˆθ K choose critical values C k compute p-value at end of study give confidence region for θ at end of study Jay Bartroff (USC) Stein s for Stats 4.Jul.17 9 / 36

19 Background: Group sequential analysis Jennison & Turnbull (JASA 97) Asymptotic multivariate normal distribution of (ˆθ 1, ˆθ 2,..., ˆθ K ) in a regression setup Y i ind f i (Y i ; x i, θ) nice Asymptotics: n k n k 1 for all k, K fixed E ( θ k ) = θ Independent increments Cov (ˆθ k1, ˆθ k2 ) = Var (ˆθ k2 ) any k 1 k 2 Folk Theorem Normal limit widely (over-)used (software packages, etc.) before Jennison & Turnbull paper Commonly heard: Once n is 5 or so the normal limit kicks in! Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

20 Background: Group sequential analysis Jennison & Turnbull (JASA 97) Asymptotic multivariate normal distribution of (ˆθ 1, ˆθ 2,..., ˆθ K ) in a regression setup Y i ind f i (Y i ; x i, θ) nice Asymptotics: n k n k 1 for all k, K fixed E ( θ k ) = θ Independent increments Cov (ˆθ k1, ˆθ k2 ) = Var (ˆθ k2 ) any k 1 k 2 Folk Theorem Normal limit widely (over-)used (software packages, etc.) before Jennison & Turnbull paper Commonly heard: Once n is 5 or so the normal limit kicks in! Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

21 Background: Group sequential analysis Jennison & Turnbull (JASA 97) Asymptotic multivariate normal distribution of (ˆθ 1, ˆθ 2,..., ˆθ K ) in a regression setup Y i ind f i (Y i ; x i, θ) nice Asymptotics: n k n k 1 for all k, K fixed E ( θ k ) = θ Independent increments Cov (ˆθ k1, ˆθ k2 ) = Var (ˆθ k2 ) any k 1 k 2 Folk Theorem Normal limit widely (over-)used (software packages, etc.) before Jennison & Turnbull paper Commonly heard: Once n is 5 or so the normal limit kicks in! Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

22 Background: Group sequential analysis Jennison & Turnbull (JASA 97) Independent increments Suppose Then Cov (ˆθ k1, ˆθ k2 ) = Var (ˆθ k2 ) any k 1 k 2 H 0 : a t θ = 0, T k = (a t ˆθ k )I k = (a t ˆθ k )[Var (a t ˆθ k )] 1. Cov (T k1, T k2 ) = I k1 I k2 a t Cov (ˆθ k1, ˆθ k2 )a = I k1 I k2 a t Var (ˆθ k2 )a = I k1 I k2 Var (a t ˆθ k2 ) = I k1 I k2 I 1 k 2 = I k1 = Var (T k1 ) Cov (T k1, T k2 T k1 ) = 0 T 1, T 2 T 1,..., T K T K 1 asymptotically independent normals Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

23 Background: Group sequential analysis Jennison & Turnbull (JASA 97) Independent increments Suppose Then Cov (ˆθ k1, ˆθ k2 ) = Var (ˆθ k2 ) any k 1 k 2 H 0 : a t θ = 0, T k = (a t ˆθ k )I k = (a t ˆθ k )[Var (a t ˆθ k )] 1. Cov (T k1, T k2 ) = I k1 I k2 a t Cov (ˆθ k1, ˆθ k2 )a = I k1 I k2 a t Var (ˆθ k2 )a = I k1 I k2 Var (a t ˆθ k2 ) = I k1 I k2 I 1 k 2 = I k1 = Var (T k1 ) Cov (T k1, T k2 T k1 ) = 0 T 1, T 2 T 1,..., T K T K 1 asymptotically independent normals Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

24 Background: Group sequential analysis Jennison & Turnbull (JASA 97) Independent increments Suppose Then Cov (ˆθ k1, ˆθ k2 ) = Var (ˆθ k2 ) any k 1 k 2 H 0 : a t θ = 0, T k = (a t ˆθ k )I k = (a t ˆθ k )[Var (a t ˆθ k )] 1. Cov (T k1, T k2 ) = I k1 I k2 a t Cov (ˆθ k1, ˆθ k2 )a = I k1 I k2 a t Var (ˆθ k2 )a = I k1 I k2 Var (a t ˆθ k2 ) = I k1 I k2 I 1 k 2 = I k1 = Var (T k1 ) Cov (T k1, T k2 T k1 ) = 0 T 1, T 2 T 1,..., T K T K 1 asymptotically independent normals Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

25 Background: Group sequential analysis Jennison & Turnbull (JASA 97) Independent increments Suppose Then Cov (ˆθ k1, ˆθ k2 ) = Var (ˆθ k2 ) any k 1 k 2 H 0 : a t θ = 0, T k = (a t ˆθ k )I k = (a t ˆθ k )[Var (a t ˆθ k )] 1. Cov (T k1, T k2 ) = I k1 I k2 a t Cov (ˆθ k1, ˆθ k2 )a = I k1 I k2 a t Var (ˆθ k2 )a = I k1 I k2 Var (a t ˆθ k2 ) = I k1 I k2 I 1 k 2 = I k1 = Var (T k1 ) Cov (T k1, T k2 T k1 ) = 0 T 1, T 2 T 1,..., T K T K 1 asymptotically independent normals Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

26 New Contributions Extension to group sequential setting of Berry-Esseen bound for multivariate normal limit for smooth functions Anastasiou & Reinert 17: Bounds w/ explicit constants for bounded Wasserstein distance for scalar MLE (K = 1 analysis), Bernoulli. Anastasiou & Ley 17: Bounds for the asymptotic normality of the maximum likelihood estimator using the Delta method, ALEA. Anastasiou 17: Bounds for the normal approximation of the maximum likelihood estimator from m-dependent random variables. Statistics & Probability Letters. Anastasiou & Gaunt 16+: Multivariate normal approximation of the maximum likelihood estimator via the delta method, ArXiv: Anastasiou 15+: Assessing the multivariate normal approximation of the maximum likelihood estimator from high-dimensional, heterogeneous data., ArXiv: Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

27 New Contributions Extension to group sequential setting of Berry-Esseen bound for multivariate normal limit for smooth functions Anastasiou & Reinert 17: Bounds w/ explicit constants for bounded Wasserstein distance for scalar MLE (K = 1 analysis), Bernoulli. Anastasiou & Ley 17: Bounds for the asymptotic normality of the maximum likelihood estimator using the Delta method, ALEA. Anastasiou 17: Bounds for the normal approximation of the maximum likelihood estimator from m-dependent random variables. Statistics & Probability Letters. Anastasiou & Gaunt 16+: Multivariate normal approximation of the maximum likelihood estimator via the delta method, ArXiv: Anastasiou 15+: Assessing the multivariate normal approximation of the maximum likelihood estimator from high-dimensional, heterogeneous data., ArXiv: Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

28 New Contributions Extension to group sequential setting of Berry-Esseen bound for multivariate normal limit for smooth functions Anastasiou & Reinert 17: Bounds w/ explicit constants for bounded Wasserstein distance for scalar MLE (K = 1 analysis), Bernoulli. Anastasiou & Ley 17: Bounds for the asymptotic normality of the maximum likelihood estimator using the Delta method, ALEA. Anastasiou 17: Bounds for the normal approximation of the maximum likelihood estimator from m-dependent random variables. Statistics & Probability Letters. Anastasiou & Gaunt 16+: Multivariate normal approximation of the maximum likelihood estimator via the delta method, ArXiv: Anastasiou 15+: Assessing the multivariate normal approximation of the maximum likelihood estimator from high-dimensional, heterogeneous data., ArXiv: Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

29 New Contributions, Cont d Relaxing independence assumption: Assume log-likelihood of Y k := (Y nk 1 +1,..., Y nk ) is of the form for well-behaved functions f i, g k i G k log f i (Y i, θ) + g k (Y k, θ) g k = 0 gives Jennison & Turnbull s independent setting Some generalized linear mixed models (GLMMs) with random stage effect U k take this form U k = effect due to lab, monitoring board, cohort, etc. Penalized quasi-likelihood (Breslow & Clayton, JASA 93) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

30 New Contributions, Cont d Relaxing independence assumption: Assume log-likelihood of Y k := (Y nk 1 +1,..., Y nk ) is of the form for well-behaved functions f i, g k i G k log f i (Y i, θ) + g k (Y k, θ) g k = 0 gives Jennison & Turnbull s independent setting Some generalized linear mixed models (GLMMs) with random stage effect U k take this form U k = effect due to lab, monitoring board, cohort, etc. Penalized quasi-likelihood (Breslow & Clayton, JASA 93) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

31 New Contributions, Cont d Relaxing independence assumption: Assume log-likelihood of Y k := (Y nk 1 +1,..., Y nk ) is of the form for well-behaved functions f i, g k i G k log f i (Y i, θ) + g k (Y k, θ) g k = 0 gives Jennison & Turnbull s independent setting Some generalized linear mixed models (GLMMs) with random stage effect U k take this form U k = effect due to lab, monitoring board, cohort, etc. Penalized quasi-likelihood (Breslow & Clayton, JASA 93) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

32 New Contributions, Cont d Relaxing independence assumption: Assume log-likelihood of Y k := (Y nk 1 +1,..., Y nk ) is of the form for well-behaved functions f i, g k i G k log f i (Y i, θ) + g k (Y k, θ) g k = 0 gives Jennison & Turnbull s independent setting Some generalized linear mixed models (GLMMs) with random stage effect U k take this form U k = effect due to lab, monitoring board, cohort, etc. Penalized quasi-likelihood (Breslow & Clayton, JASA 93) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

33 GLMM Example: Poisson regression Letting f µ = Po(µ) density, For Y i in kth stage, Y i U k ind f µi where µ i = exp(β t x i + U k ) {U k } iid h λ θ = (β, λ). Then log-likelihood is K log f µi (Y i )h λ (U k )du k = k=1 i G k where µ i = exp(β t x i ). K log f µi (Y i ) + g k (Y k, θ) i Gk k=1 Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

34 GLMM Example: Poisson regression Letting f µ = Po(µ) density, For Y i in kth stage, Y i U k ind f µi where µ i = exp(β t x i + U k ) {U k } iid h λ θ = (β, λ). Then log-likelihood is K log f µi (Y i )h λ (U k )du k = k=1 i G k where µ i = exp(β t x i ). K log f µi (Y i ) + g k (Y k, θ) i Gk k=1 Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

35 GLMM Example: Poisson regression Letting f µ = Po(µ) density, For Y i in kth stage, Y i U k ind f µi where µ i = exp(β t x i + U k ) {U k } iid h λ θ = (β, λ). Then log-likelihood is K log f µi (Y i )h λ (U k )du k = k=1 i G k where µ i = exp(β t x i ). K log f µi (Y i ) + g k (Y k, θ) i Gk k=1 Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

36 GLMM Example: Poisson regression Letting f µ = Po(µ) density, For Y i in kth stage, Y i U k ind f µi where µ i = exp(β t x i + U k ) {U k } iid h λ θ = (β, λ). Then log-likelihood is K log f µi (Y i )h λ (U k )du k = k=1 i G k where µ i = exp(β t x i ). K log f µi (Y i ) + g k (Y k, θ) i Gk k=1 Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

37 Stein s Method for MVN Approximation Generator approach: Barbour 90, Goetze 91 Size biasing: Goldstein & Rinott 96, Rinott & Rotar 96 Zero biasing: Goldstein & Reinert 05 Exchangeable pair: Chatterjee & Meckes 08, Reinert & Röllin 09 Stein couplings: Fang & Röllin 15 Theorem (Reinert & Röllin 09) If W, W R q exchangeable pair with EW = 0, EWW t = Σ PD, and E(W W W ) = ΛW + R with Λ invertible, then for any 3-times differentiable h : R q R, Eh(W ) Eh(Σ 1/2 Z ) a h 2 4 for certain explicit constants a, b, c. + b h c ( h 1 + q 2 Σ 1/2 h 2 ) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

38 Stein s Method for MVN Approximation Generator approach: Barbour 90, Goetze 91 Size biasing: Goldstein & Rinott 96, Rinott & Rotar 96 Zero biasing: Goldstein & Reinert 05 Exchangeable pair: Chatterjee & Meckes 08, Reinert & Röllin 09 Stein couplings: Fang & Röllin 15 Theorem (Reinert & Röllin 09) If W, W R q exchangeable pair with EW = 0, EWW t = Σ PD, and E(W W W ) = ΛW + R with Λ invertible, then for any 3-times differentiable h : R q R, Eh(W ) Eh(Σ 1/2 Z ) a h 2 4 for certain explicit constants a, b, c. + b h c ( h 1 + q 2 Σ 1/2 h 2 ) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

39 Bounds to normal for ˆθ K := (ˆθ 1, ˆθ 2,..., ˆθ K ) Approach: Apply Reinert & Röllin 09 result with W = score function increments to get smooth function bounds to normal. Result In the group sequential setup above, if the Y i are independent or follow GLMMs with the log-likelihood of the kth group data Y k = (Y nk 1 +1,..., Y nk ) of the form i G k log f i (Y i, θ) + g k (Y k, θ), then under regularity conditions on the f i and g k there are a, b, c, d s.t. Eh(J 1/2 (ˆθ K θ K )) Eh(Z ) ak 2 J 1/2 2 h 2 + bk 3 J 1/2 3 h ck J 1/2 ( h 1 + pk 2 ) 2 Σ 1/2 J 1/2 h 2 + d. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

40 Bounds to normal for ˆθ K := (ˆθ 1, ˆθ 2,..., ˆθ K ) Approach: Apply Reinert & Röllin 09 result with W = score function increments to get smooth function bounds to normal. Result In the group sequential setup above, if the Y i are independent or follow GLMMs with the log-likelihood of the kth group data Y k = (Y nk 1 +1,..., Y nk ) of the form i G k log f i (Y i, θ) + g k (Y k, θ), then under regularity conditions on the f i and g k there are a, b, c, d s.t. Eh(J 1/2 (ˆθ K θ K )) Eh(Z ) ak 2 J 1/2 2 h 2 + bk 3 J 1/2 3 h ck J 1/2 ( h 1 + pk 2 ) 2 Σ 1/2 J 1/2 h 2 + d. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

41 Comments on result Bound Eh(J 1/2 (ˆθ K θ K )) Eh(Z ) ak 2 J 1/2 2 h 2 + bk 3 J 1/2 3 h ck J 1/2 ( h 1 + pk 2 ) 2 Σ 1/2 J 1/2 h 2 + d. a, b, c terms directly from Reinert & Röllin 09 bound c term Var(R) in E(W W W ) = ΛW + R, vanishes in independent case d term is from Taylor Series remainders Rate O(1/ n K ) under usual asymptotic n k n k 1 n K γ k (0, 1) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

42 Comments on result Bound Eh(J 1/2 (ˆθ K θ K )) Eh(Z ) ak 2 J 1/2 2 h 2 + bk 3 J 1/2 3 h ck J 1/2 ( h 1 + pk 2 ) 2 Σ 1/2 J 1/2 h 2 + d. a, b, c terms directly from Reinert & Röllin 09 bound c term Var(R) in E(W W W ) = ΛW + R, vanishes in independent case d term is from Taylor Series remainders Rate O(1/ n K ) under usual asymptotic n k n k 1 n K γ k (0, 1) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

43 Sketch of Proof Independent Case Score statistic S i (θ) = θ log f i(y i, θ) R p, W = S i (θ),..., S i (θ) R q, i G1 i G K where q = pk. Fisher Information ( ) J i (θ) = E θ S i(θ) t R p p J(θ 1,..., θ K ) = diag ( n1 J i (θ 1 ),..., i=1 n K i=1 J i (θ K ) ) R q q Σ := Var(W ) = diag J i (θ),..., J i (θ) R q q i G1 i G K Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

44 Sketch of Proof Independent Case Score statistic S i (θ) = θ log f i(y i, θ) R p, W = S i (θ),..., S i (θ) R q, i G1 i G K where q = pk. Fisher Information ( ) J i (θ) = E θ S i(θ) t R p p J(θ 1,..., θ K ) = diag ( n1 J i (θ 1 ),..., i=1 n K i=1 J i (θ K ) ) R q q Σ := Var(W ) = diag J i (θ),..., J i (θ) R q q i G1 i G K Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

45 Sketch of Proof Independent Case Score statistic S i (θ) = θ log f i(y i, θ) R p, W = S i (θ),..., S i (θ) R q, i G1 i G K where q = pk. Fisher Information ( ) J i (θ) = E θ S i(θ) t R p p J(θ 1,..., θ K ) = diag ( n1 J i (θ 1 ),..., i=1 n K i=1 J i (θ K ) ) R q q Σ := Var(W ) = diag J i (θ),..., J i (θ) R q q i G1 i G K Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

46 Sketch of Proof: Exchangeable pair Independent Case 1. Choose i {1,..., n K } uniformly, independent of all else 2. Replace Y i by independent copy Y i (keeping x i ), call result W W, W exchangeable W, W satisfy linearity condition which is easy to check entry-wise E(W W W ) = n 1 K W Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

47 Sketch of Proof: Exchangeable pair Independent Case 1. Choose i {1,..., n K } uniformly, independent of all else 2. Replace Y i by independent copy Y i (keeping x i ), call result W W, W exchangeable W, W satisfy linearity condition which is easy to check entry-wise E(W W W ) = n 1 K W Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

48 Sketch of Proof: Exchangeable pair Independent Case 1. Choose i {1,..., n K } uniformly, independent of all else 2. Replace Y i by independent copy Y i (keeping x i ), call result W W, W exchangeable W, W satisfy linearity condition which is easy to check entry-wise E(W W W ) = n 1 K W Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

49 Sketch of Proof: Exchangeable pair Independent Case 1. Choose i {1,..., n K } uniformly, independent of all else 2. Replace Y i by independent copy Y i (keeping x i ), call result W W, W exchangeable W, W satisfy linearity condition which is easy to check entry-wise E(W W W ) = n 1 K W Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

50 Sketch of Proof: Relating ˆθ K to W Independent Case By standard Taylor series, ˆθ K θ K = J(θ K ) 1 S, where S = ( n1 S i (θ 1 ),..., n K i=1 i=1 S i (θ K ) and θ K R q on line segment connecting ˆθ K, θ K. Then Eh(J 1/2 (ˆθ K θ K )) Eh(Z ) Eh(J 1/2 S) Eh(Z ) + Eh(J 1/2 J(θ K ) 1 S) Eh(J 1/2 S) ) R q Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

51 Sketch of Proof: Relating ˆθ K to W Independent Case By standard Taylor series, ˆθ K θ K = J(θ K ) 1 S, where S = ( n1 S i (θ 1 ),..., n K i=1 i=1 S i (θ K ) and θ K R q on line segment connecting ˆθ K, θ K. Then Eh(J 1/2 (ˆθ K θ K )) Eh(Z ) Eh(J 1/2 S) Eh(Z ) + Eh(J 1/2 J(θ K ) 1 S) Eh(J 1/2 S) ) R q Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

52 Sketch of Proof: Relating ˆθ K to W Independent Case Using S = AW where 1 p 0 p 0 p 1 p 1 p 0 p A = p 1 p 1 p 1st term is, 1 p, 0 p R p p identity and 0 matrices, Eh(J 1/2 S) Eh(Z ) = E h(w ) E h(σ 1/2 Z ) where h(w) = h(j 1/2 Aw), then apply Reinert-Röllin and simplify. 2nd term is bounded by Taylor series arguments. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

53 Sketch of Proof: Relating ˆθ K to W Independent Case Using S = AW where 1 p 0 p 0 p 1 p 1 p 0 p A = p 1 p 1 p 1st term is, 1 p, 0 p R p p identity and 0 matrices, Eh(J 1/2 S) Eh(Z ) = E h(w ) E h(σ 1/2 Z ) where h(w) = h(j 1/2 Aw), then apply Reinert-Röllin and simplify. 2nd term is bounded by Taylor series arguments. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

54 Sketch of Proof: Exchangeable pair GLMM Case 1. Choose i {1,..., n K } uniformly, independent of Y 1,..., Y nk 2. If i in kth group, replace Y i by independent copy Y i with mean ϕ(β t x i + U k ), ϕ 1 = link function (same covariates x i, group effect U k ), call result W W, W exchangeable W, W satisfy linearity condition where R = R(g 1,..., g K ) E(W W W ) = n 1 K W + R Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

55 Sketch of Proof: Exchangeable pair GLMM Case 1. Choose i {1,..., n K } uniformly, independent of Y 1,..., Y nk 2. If i in kth group, replace Y i by independent copy Y i with mean ϕ(β t x i + U k ), ϕ 1 = link function (same covariates x i, group effect U k ), call result W W, W exchangeable W, W satisfy linearity condition where R = R(g 1,..., g K ) E(W W W ) = n 1 K W + R Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

56 Sketch of Proof: Exchangeable pair GLMM Case 1. Choose i {1,..., n K } uniformly, independent of Y 1,..., Y nk 2. If i in kth group, replace Y i by independent copy Y i with mean ϕ(β t x i + U k ), ϕ 1 = link function (same covariates x i, group effect U k ), call result W W, W exchangeable W, W satisfy linearity condition where R = R(g 1,..., g K ) E(W W W ) = n 1 K W + R Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

57 Sketch of Proof: Exchangeable pair GLMM Case 1. Choose i {1,..., n K } uniformly, independent of Y 1,..., Y nk 2. If i in kth group, replace Y i by independent copy Y i with mean ϕ(β t x i + U k ), ϕ 1 = link function (same covariates x i, group effect U k ), call result W W, W exchangeable W, W satisfy linearity condition where R = R(g 1,..., g K ) E(W W W ) = n 1 K W + R Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

58 Sketch of Proof: Exchangeable pair GLMM Case 1. Choose i {1,..., n K } uniformly, independent of Y 1,..., Y nk 2. If i in kth group, replace Y i by independent copy Y i with mean ϕ(β t x i + U k ), ϕ 1 = link function (same covariates x i, group effect U k ), call result W W, W exchangeable W, W satisfy linearity condition where R = R(g 1,..., g K ) E(W W W ) = n 1 K W + R Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

59 Concentration and Coupling Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

60 Application 2: Concentration inequalities for occupancy models with log-concave marginals Main idea: How to get bounded, size biased couplings for certain multivariate occupancy models, then use methods pioneered by Ghosh and Goldstein 11 to get concentration inequalities Concentration Inequalities ( t 2 ) e.g., P (Y µ t) exp 2cµ + ct Widely used in high dimensional statistics machine learning random matrix theory applications: wireless communications, physics,... (See Raginsky & Sason 15) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

61 Application 2: Concentration inequalities for occupancy models with log-concave marginals Main idea: How to get bounded, size biased couplings for certain multivariate occupancy models, then use methods pioneered by Ghosh and Goldstein 11 to get concentration inequalities Concentration Inequalities ( t 2 ) e.g., P (Y µ t) exp 2cµ + ct Widely used in high dimensional statistics machine learning random matrix theory applications: wireless communications, physics,... (See Raginsky & Sason 15) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

62 Setup Occupancy model M = (M α ) M α may be degree count of vertex α in an Erdös-Rényi random graph # of balls in box α in multinomial model # balls of color α in sample from urn of colored balls We consider statistics like m m Y ge = 1{M α d}, Y eq = 1{M α = d} Y ge = α=1 m w α 1{M α d α }, Y eq = α=1 α=1 α=1 m w α 1{M α = d α } Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

63 Setup Occupancy model M = (M α ) M α may be degree count of vertex α in an Erdös-Rényi random graph # of balls in box α in multinomial model # balls of color α in sample from urn of colored balls We consider statistics like m m Y ge = 1{M α d}, Y eq = 1{M α = d} Y ge = α=1 m w α 1{M α d α }, Y eq = α=1 α=1 α=1 m w α 1{M α = d α } Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

64 Setup Occupancy model M = (M α ) M α may be degree count of vertex α in an Erdös-Rényi random graph # of balls in box α in multinomial model # balls of color α in sample from urn of colored balls We consider statistics like m m Y ge = 1{M α d}, Y eq = 1{M α = d} Y ge = α=1 m w α 1{M α d α }, Y eq = α=1 α=1 α=1 m w α 1{M α = d α } Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

65 Setup Occupancy model M = (M α ) M α may be degree count of vertex α in an Erdös-Rényi random graph # of balls in box α in multinomial model # balls of color α in sample from urn of colored balls We consider statistics like m m Y ge = 1{M α d}, Y eq = 1{M α = d} Y ge = α=1 m w α 1{M α d α }, Y eq = α=1 α=1 α=1 m w α 1{M α = d α } Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

66 Some Methods for Concentration Inequalities McDiarmid s Bounded Difference Inequality Y a function with bounded differences of independent inputs Negative Association e.g., Dubashi & Ranjan 98 Certifiable Functions Y a certifiable function of independent inputs Controlling a large enough subset of inputs certifies value of function e.g., McDiarmid & Reed 06 Bounded Size Bias Couplings Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

67 Some Methods for Concentration Inequalities McDiarmid s Bounded Difference Inequality Y a function with bounded differences of independent inputs Negative Association e.g., Dubashi & Ranjan 98 Certifiable Functions Y a certifiable function of independent inputs Controlling a large enough subset of inputs certifies value of function e.g., McDiarmid & Reed 06 Bounded Size Bias Couplings Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

68 Some Methods for Concentration Inequalities McDiarmid s Bounded Difference Inequality Y a function with bounded differences of independent inputs Negative Association e.g., Dubashi & Ranjan 98 Certifiable Functions Y a certifiable function of independent inputs Controlling a large enough subset of inputs certifies value of function e.g., McDiarmid & Reed 06 Bounded Size Bias Couplings Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

69 Some Methods for Concentration Inequalities McDiarmid s Bounded Difference Inequality Y a function with bounded differences of independent inputs Negative Association e.g., Dubashi & Ranjan 98 Certifiable Functions Y a certifiable function of independent inputs Controlling a large enough subset of inputs certifies value of function e.g., McDiarmid & Reed 06 Bounded Size Bias Couplings Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

70 Bounded Size Bias Couplings If there is a coupling Y s of Y with the Y -size bias distribution, i.e., E[Yf (Y )] = µe[f (Y s )] for all f, and Y s Y + c for some c > 0 with probability one, then max {P(Y µ t), P(Y µ t)} b µ,c (t). Ghosh & Goldstein 11: For all t > 0, ) ( P (Y µ t) exp ( t2 t 2 P (Y µ t) exp 2cµ 2cµ + ct b exponential as t. Arratia & Baxendale 13: b µ,c (t) = exp b Poisson as t. ( µ c h ( t µ )), h(x) = (1 + x) log(1 + x) x. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36 ).

71 Bounded Size Bias Couplings If there is a coupling Y s of Y with the Y -size bias distribution, i.e., E[Yf (Y )] = µe[f (Y s )] for all f, and Y s Y + c for some c > 0 with probability one, then max {P(Y µ t), P(Y µ t)} b µ,c (t). Ghosh & Goldstein 11: For all t > 0, ) ( P (Y µ t) exp ( t2 t 2 P (Y µ t) exp 2cµ 2cµ + ct b exponential as t. Arratia & Baxendale 13: b µ,c (t) = exp b Poisson as t. ( µ c h ( t µ )), h(x) = (1 + x) log(1 + x) x. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36 ).

72 Bounded Size Bias Couplings If there is a coupling Y s of Y with the Y -size bias distribution, i.e., E[Yf (Y )] = µe[f (Y s )] for all f, and Y s Y + c for some c > 0 with probability one, then max {P(Y µ t), P(Y µ t)} b µ,c (t). Ghosh & Goldstein 11: For all t > 0, ) ( P (Y µ t) exp ( t2 t 2 P (Y µ t) exp 2cµ 2cµ + ct b exponential as t. Arratia & Baxendale 13: b µ,c (t) = exp b Poisson as t. ( µ c h ( t µ )), h(x) = (1 + x) log(1 + x) x. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36 ).

73 Main Result M = (M α ) α [m], M α lattice log-concave Y ge = w α 1{M α d α }, Y ne = w α 1{M α d α }. α [m] α [m] Main Result (in words) 1. If M is bounded from below and can be closely coupled to a version M having the same distribution conditional on M α = M α + 1, then there is a bounded size biased coupling Y s ge Y ge + C and the above concentration inequalities hold. 2. If M is non-degenerate at (d α ) and can be closely coupled to a version M having the same distribution conditional on M α d α, then there is a bounded size biased coupling Y s ne Y ne + C and the above concentration inequalities hold. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

74 Main Result M = (M α ) α [m], M α lattice log-concave Y ge = w α 1{M α d α }, Y ne = w α 1{M α d α }. α [m] α [m] Main Result (in words) 1. If M is bounded from below and can be closely coupled to a version M having the same distribution conditional on M α = M α + 1, then there is a bounded size biased coupling Y s ge Y ge + C and the above concentration inequalities hold. 2. If M is non-degenerate at (d α ) and can be closely coupled to a version M having the same distribution conditional on M α d α, then there is a bounded size biased coupling Y s ne Y ne + C and the above concentration inequalities hold. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

75 Main Result M = (M α ) α [m], M α lattice log-concave Y ge = w α 1{M α d α }, Y ne = w α 1{M α d α }. α [m] α [m] Main Result (in words) 1. If M is bounded from below and can be closely coupled to a version M having the same distribution conditional on M α = M α + 1, then there is a bounded size biased coupling Y s ge Y ge + C and the above concentration inequalities hold. 2. If M is non-degenerate at (d α ) and can be closely coupled to a version M having the same distribution conditional on M α d α, then there is a bounded size biased coupling Y s ne Y ne + C and the above concentration inequalities hold. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

76 Main Result A few more details on Part 1 M = f (U) where U is some collection of random variables f is measurable Closely coupled means given U k L(V k ) := L(U M α k), there is coupling U + k and constant B such that L(U + k U k) = L(V k M + k,α = M k,α + 1) and Y + k,ge, α Y k,ge, α + B, where Y k,ge, α = β α 1(M k,β d β ). The constant is C = w (B d + 1) where w = max w α, d = max d α. Part 2 is similar. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

77 Main Result A few more details on Part 1 M = f (U) where U is some collection of random variables f is measurable Closely coupled means given U k L(V k ) := L(U M α k), there is coupling U + k and constant B such that L(U + k U k) = L(V k M + k,α = M k,α + 1) and Y + k,ge, α Y k,ge, α + B, where Y k,ge, α = β α 1(M k,β d β ). The constant is C = w (B d + 1) where w = max w α, d = max d α. Part 2 is similar. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

78 Main Result A few more details on Part 1 M = f (U) where U is some collection of random variables f is measurable Closely coupled means given U k L(V k ) := L(U M α k), there is coupling U + k and constant B such that L(U + k U k) = L(V k M + k,α = M k,α + 1) and Y + k,ge, α Y k,ge, α + B, where Y k,ge, α = β α 1(M k,β d β ). The constant is C = w (B d + 1) where w = max w α, d = max d α. Part 2 is similar. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

79 Main Result A few more details on Part 1 M = f (U) where U is some collection of random variables f is measurable Closely coupled means given U k L(V k ) := L(U M α k), there is coupling U + k and constant B such that L(U + k U k) = L(V k M + k,α = M k,α + 1) and Y + k,ge, α Y k,ge, α + B, where Y k,ge, α = β α 1(M k,β d β ). The constant is C = w (B d + 1) where w = max w α, d = max d α. Part 2 is similar. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

80 Main Result Main Ingredient in Proof Incrementing Lemma If M is lattice log-concave then there is π(x, d) [0, 1] such that if M L(M M d) and B M Bern(π(M, d)), then M + B L(M M d + 1). Extension of Goldstein & Penrose 10 for M Binomial, d = 0 Analogous versions for L(M M d) L(M M d 1) L(M) L(M M d) where means coupled to Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

81 Main Result Main Ingredient in Proof Incrementing Lemma If M is lattice log-concave then there is π(x, d) [0, 1] such that if M L(M M d) and B M Bern(π(M, d)), then M + B L(M M d + 1). Extension of Goldstein & Penrose 10 for M Binomial, d = 0 Analogous versions for L(M M d) L(M M d 1) L(M) L(M M d) where means coupled to Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

82 Example 1: Erdös-Rényi random graph m vertices Independent edges with probability p α,β = p β,α [0, 1). Constructing U + k from U k: 1. Selection non-neighbor β of α with probability 2. Add edge connecting β to α p α,β 1 p α,β This affects at most 1 other vertex so B = 1 and Y s ge Y ge + w ( d + 1). Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

83 Example 1: Erdös-Rényi random graph m vertices Independent edges with probability p α,β = p β,α [0, 1). Constructing U + k from U k: 1. Selection non-neighbor β of α with probability 2. Add edge connecting β to α p α,β 1 p α,β This affects at most 1 other vertex so B = 1 and Y s ge Y ge + w ( d + 1). Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

84 Example 1: Erdös-Rényi random graph m vertices Independent edges with probability p α,β = p β,α [0, 1). Constructing U + k from U k: 1. Selection non-neighbor β of α with probability 2. Add edge connecting β to α p α,β 1 p α,β This affects at most 1 other vertex so B = 1 and Y s ge Y ge + w ( d + 1). Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

85 Example 1: Erdös-Rényi random graph Applying this with d α = d, w α = 1, ( t 2 ) P(Y ge µ t) exp 2(d + 1)µ ( t 2 ) exp 2(d + 1)m Compare with McDiarmid s bounded difference inequality: so Y ge = f (X 1,..., X ( m 2) ), X i = 1{edge between vertex pair i}, sup f (X 1,..., X i,..., X X i,x ( m i 2) ) f (X 1,..., X i,..., X ( m 2) ) 2, ( t 2 ) P(Y ge µ t) exp 4m(m 1) New bound an improvement for m > 2d + 3 Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

86 Example 1: Erdös-Rényi random graph Applying this with d α = d, w α = 1, ( t 2 ) P(Y ge µ t) exp 2(d + 1)µ ( t 2 ) exp 2(d + 1)m Compare with McDiarmid s bounded difference inequality: so Y ge = f (X 1,..., X ( m 2) ), X i = 1{edge between vertex pair i}, sup f (X 1,..., X i,..., X X i,x ( m i 2) ) f (X 1,..., X i,..., X ( m 2) ) 2, ( t 2 ) P(Y ge µ t) exp 4m(m 1) New bound an improvement for m > 2d + 3 Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

87 Example 1: Erdös-Rényi random graph Applying this with d α = d, w α = 1, ( t 2 ) P(Y ge µ t) exp 2(d + 1)µ ( t 2 ) exp 2(d + 1)m Compare with McDiarmid s bounded difference inequality: so Y ge = f (X 1,..., X ( m 2) ), X i = 1{edge between vertex pair i}, sup f (X 1,..., X i,..., X X i,x ( m i 2) ) f (X 1,..., X i,..., X ( m 2) ) 2, ( t 2 ) P(Y ge µ t) exp 4m(m 1) New bound an improvement for m > 2d + 3 Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

88 Example 1: Erdös-Rényi random graph Applying this with d α = d, w α = 1, ( t 2 ) P(Y ge µ t) exp 2(d + 1)µ ( t 2 ) exp 2(d + 1)m Compare with McDiarmid s bounded difference inequality: so Y ge = f (X 1,..., X ( m 2) ), X i = 1{edge between vertex pair i}, sup f (X 1,..., X i,..., X X i,x ( m i 2) ) f (X 1,..., X i,..., X ( m 2) ) 2, ( t 2 ) P(Y ge µ t) exp 4m(m 1) New bound an improvement for m > 2d + 3 Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

89 Example 1: Erdös-Rényi random graph Applying this with d α = d, w α = 1, ( t 2 ) P(Y ge µ t) exp 2(d + 1)µ ( t 2 ) exp 2(d + 1)m Compare with McDiarmid s bounded difference inequality: so Y ge = f (X 1,..., X ( m 2) ), X i = 1{edge between vertex pair i}, sup f (X 1,..., X i,..., X X i,x ( m i 2) ) f (X 1,..., X i,..., X ( m 2) ) 2, ( t 2 ) P(Y ge µ t) exp 4m(m 1) New bound an improvement for m > 2d + 3 Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

90 Example 2: Multinomial Counts n balls independently into m boxes Applications in species trapping, linguistics,... # empty boxes proved asymptotically normal by Weiss 58, Rényi 62 in uniform case Englund 81: L bound for # of empty cells, uniform case Dubashi & Ranjan 98: Concentration inequality via NA Penrose 09: L bound for # of isolated balls, uniform and nonuniform cases Bartroff & Goldstein 13: L bound for all d 2, uniform case Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

91 Example 2: Multinomial Counts n balls independently into m boxes Applications in species trapping, linguistics,... # empty boxes proved asymptotically normal by Weiss 58, Rényi 62 in uniform case Englund 81: L bound for # of empty cells, uniform case Dubashi & Ranjan 98: Concentration inequality via NA Penrose 09: L bound for # of isolated balls, uniform and nonuniform cases Bartroff & Goldstein 13: L bound for all d 2, uniform case Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

92 Example 2: Multinomial Counts n balls independently into m boxes Applications in species trapping, linguistics,... # empty boxes proved asymptotically normal by Weiss 58, Rényi 62 in uniform case Englund 81: L bound for # of empty cells, uniform case Dubashi & Ranjan 98: Concentration inequality via NA Penrose 09: L bound for # of isolated balls, uniform and nonuniform cases Bartroff & Goldstein 13: L bound for all d 2, uniform case Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

93 Example 2: Multinomial Counts n balls independently into m boxes Applications in species trapping, linguistics,... # empty boxes proved asymptotically normal by Weiss 58, Rényi 62 in uniform case Englund 81: L bound for # of empty cells, uniform case Dubashi & Ranjan 98: Concentration inequality via NA Penrose 09: L bound for # of isolated balls, uniform and nonuniform cases Bartroff & Goldstein 13: L bound for all d 2, uniform case Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

94 Example 2: Multinomial Counts n balls independently into m boxes Applications in species trapping, linguistics,... # empty boxes proved asymptotically normal by Weiss 58, Rényi 62 in uniform case Englund 81: L bound for # of empty cells, uniform case Dubashi & Ranjan 98: Concentration inequality via NA Penrose 09: L bound for # of isolated balls, uniform and nonuniform cases Bartroff & Goldstein 13: L bound for all d 2, uniform case Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

95 Example 2: Multinomial Counts n balls independently into m boxes Applications in species trapping, linguistics,... # empty boxes proved asymptotically normal by Weiss 58, Rényi 62 in uniform case Englund 81: L bound for # of empty cells, uniform case Dubashi & Ranjan 98: Concentration inequality via NA Penrose 09: L bound for # of isolated balls, uniform and nonuniform cases Bartroff & Goldstein 13: L bound for all d 2, uniform case Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

96 Example 2: Multinomial Counts n balls independently into m boxes Applications in species trapping, linguistics,... # empty boxes proved asymptotically normal by Weiss 58, Rényi 62 in uniform case Englund 81: L bound for # of empty cells, uniform case Dubashi & Ranjan 98: Concentration inequality via NA Penrose 09: L bound for # of isolated balls, uniform and nonuniform cases Bartroff & Goldstein 13: L bound for all d 2, uniform case Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

97 Example 2: Multinomial Counts p α,j = prob. that ball j [n] falls in box α [m] M α = # balls in box α = 1{ball j falls in box α} j [n] Constructing U + k from U k: Choose ball j box α with probability p α,j 1 p α,j and add it to box α. Y s ge, α Y ge, α so B = 0, thus Y s ge Y ge + w Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

98 Example 2: Multinomial Counts p α,j = prob. that ball j [n] falls in box α [m] M α = # balls in box α = 1{ball j falls in box α} j [n] Constructing U + k from U k: Choose ball j box α with probability p α,j 1 p α,j and add it to box α. Y s ge, α Y ge, α so B = 0, thus Y s ge Y ge + w Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

99 Example 2: Multinomial Counts p α,j = prob. that ball j [n] falls in box α [m] M α = # balls in box α = 1{ball j falls in box α} j [n] Constructing U + k from U k: Choose ball j box α with probability p α,j 1 p α,j and add it to box α. Y s ge, α Y ge, α so B = 0, thus Y s ge Y ge + w Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

100 Example 3: Multivariate Hypergeometric Sampling Urn with n = α [m] n α colored balls, n α balls of color α Sample of size s drawn without replacement M α = # balls in sample of color α Applications in sampling (and subsampling) theory, gambling, coupon-collector problems Constructing U + k probability from U k: Select non-α colored ball in sample with n α(j)/n and replace it with α-colored ball, α(j) = color of ball j 1 n α(j) /n Y s ge, α Y ge, α so B = 0, thus Y s ge Y ge + w Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

101 Example 3: Multivariate Hypergeometric Sampling Urn with n = α [m] n α colored balls, n α balls of color α Sample of size s drawn without replacement M α = # balls in sample of color α Applications in sampling (and subsampling) theory, gambling, coupon-collector problems Constructing U + k probability from U k: Select non-α colored ball in sample with n α(j)/n and replace it with α-colored ball, α(j) = color of ball j 1 n α(j) /n Y s ge, α Y ge, α so B = 0, thus Y s ge Y ge + w Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

102 Example 3: Multivariate Hypergeometric Sampling Urn with n = α [m] n α colored balls, n α balls of color α Sample of size s drawn without replacement M α = # balls in sample of color α Applications in sampling (and subsampling) theory, gambling, coupon-collector problems Constructing U + k probability from U k: Select non-α colored ball in sample with n α(j)/n and replace it with α-colored ball, α(j) = color of ball j 1 n α(j) /n Y s ge, α Y ge, α so B = 0, thus Y s ge Y ge + w Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

103 Example 3: Multivariate Hypergeometric Sampling Urn with n = α [m] n α colored balls, n α balls of color α Sample of size s drawn without replacement M α = # balls in sample of color α Applications in sampling (and subsampling) theory, gambling, coupon-collector problems Constructing U + k probability from U k: Select non-α colored ball in sample with n α(j)/n and replace it with α-colored ball, α(j) = color of ball j 1 n α(j) /n Y s ge, α Y ge, α so B = 0, thus Y s ge Y ge + w Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

104 Example 3: Multivariate Hypergeometric Sampling Urn with n = α [m] n α colored balls, n α balls of color α Sample of size s drawn without replacement M α = # balls in sample of color α Applications in sampling (and subsampling) theory, gambling, coupon-collector problems Constructing U + k probability from U k: Select non-α colored ball in sample with n α(j)/n and replace it with α-colored ball, α(j) = color of ball j 1 n α(j) /n Y s ge, α Y ge, α so B = 0, thus Y s ge Y ge + w Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

105 Example 3: Multivariate Hypergeometric Sampling Urn with n = α [m] n α colored balls, n α balls of color α Sample of size s drawn without replacement M α = # balls in sample of color α Applications in sampling (and subsampling) theory, gambling, coupon-collector problems Constructing U + k probability from U k: Select non-α colored ball in sample with n α(j)/n and replace it with α-colored ball, α(j) = color of ball j 1 n α(j) /n Y s ge, α Y ge, α so B = 0, thus Y s ge Y ge + w Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

106 Summary Stein s method applied to produce explicit bounds to the limiting normal distribution for repeatedly-computed MLE (ˆθ 1, ˆθ 2,..., ˆθ K ) of a parameter vector in group sequential trials concentration inequalities for a class of occupancy models with log-concave marginals Many unanswered questions in statistics possibly susceptible to Stein s concentration inequalities for heavy-tailed distributions convergence of empirical measures and dimension reduction methods projections of empirical measures onto subspaces high-dimensional PCA other problems in sequential analysis sequentially stopped test statistics stopping rules for high-dimensional MCMC Thank You! Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

107 Summary Stein s method applied to produce explicit bounds to the limiting normal distribution for repeatedly-computed MLE (ˆθ 1, ˆθ 2,..., ˆθ K ) of a parameter vector in group sequential trials concentration inequalities for a class of occupancy models with log-concave marginals Many unanswered questions in statistics possibly susceptible to Stein s concentration inequalities for heavy-tailed distributions convergence of empirical measures and dimension reduction methods projections of empirical measures onto subspaces high-dimensional PCA other problems in sequential analysis sequentially stopped test statistics stopping rules for high-dimensional MCMC Thank You! Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

108 Summary Stein s method applied to produce explicit bounds to the limiting normal distribution for repeatedly-computed MLE (ˆθ 1, ˆθ 2,..., ˆθ K ) of a parameter vector in group sequential trials concentration inequalities for a class of occupancy models with log-concave marginals Many unanswered questions in statistics possibly susceptible to Stein s concentration inequalities for heavy-tailed distributions convergence of empirical measures and dimension reduction methods projections of empirical measures onto subspaces high-dimensional PCA other problems in sequential analysis sequentially stopped test statistics stopping rules for high-dimensional MCMC Thank You! Jay Bartroff (USC) Stein s for Stats 4.Jul / 36

Stein Couplings for Concentration of Measure

Stein Couplings for Concentration of Measure Stein Couplings for Concentration of Measure Jay Bartroff, Subhankar Ghosh, Larry Goldstein and Ümit Işlak University of Southern California [arxiv:0906.3886] [arxiv:1304.5001] [arxiv:1402.6769] Borchard

More information

Concentration of Measures by Bounded Couplings

Concentration of Measures by Bounded Couplings Concentration of Measures by Bounded Couplings Subhankar Ghosh, Larry Goldstein and Ümit Işlak University of Southern California [arxiv:0906.3886] [arxiv:1304.5001] May 2013 Concentration of Measure Distributional

More information

Stein s Method: Distributional Approximation and Concentration of Measure

Stein s Method: Distributional Approximation and Concentration of Measure Stein s Method: Distributional Approximation and Concentration of Measure Larry Goldstein University of Southern California 36 th Midwest Probability Colloquium, 2014 Stein s method for Distributional

More information

Stein s Method: Distributional Approximation and Concentration of Measure

Stein s Method: Distributional Approximation and Concentration of Measure Stein s Method: Distributional Approximation and Concentration of Measure Larry Goldstein University of Southern California 36 th Midwest Probability Colloquium, 2014 Concentration of Measure Distributional

More information

Submitted to the Brazilian Journal of Probability and Statistics

Submitted to the Brazilian Journal of Probability and Statistics Submitted to the Brazilian Journal of Probability and Statistics Multivariate normal approximation of the maximum likelihood estimator via the delta method Andreas Anastasiou a and Robert E. Gaunt b a

More information

Concentration of Measures by Bounded Size Bias Couplings

Concentration of Measures by Bounded Size Bias Couplings Concentration of Measures by Bounded Size Bias Couplings Subhankar Ghosh, Larry Goldstein University of Southern California [arxiv:0906.3886] January 10 th, 2013 Concentration of Measure Distributional

More information

Interim Monitoring of Clinical Trials: Decision Theory, Dynamic Programming. and Optimal Stopping

Interim Monitoring of Clinical Trials: Decision Theory, Dynamic Programming. and Optimal Stopping Interim Monitoring of Clinical Trials: Decision Theory, Dynamic Programming and Optimal Stopping Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj

More information

Stat 5101 Lecture Notes

Stat 5101 Lecture Notes Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random

More information

Differential Stein operators for multivariate continuous distributions and applications

Differential Stein operators for multivariate continuous distributions and applications Differential Stein operators for multivariate continuous distributions and applications Gesine Reinert A French/American Collaborative Colloquium on Concentration Inequalities, High Dimensional Statistics

More information

Optimising Group Sequential Designs. Decision Theory, Dynamic Programming. and Optimal Stopping

Optimising Group Sequential Designs. Decision Theory, Dynamic Programming. and Optimal Stopping : Decision Theory, Dynamic Programming and Optimal Stopping Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj InSPiRe Conference on Methodology

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

Spring 2012 Math 541B Exam 1

Spring 2012 Math 541B Exam 1 Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote

More information

A Gentle Introduction to Stein s Method for Normal Approximation I

A Gentle Introduction to Stein s Method for Normal Approximation I A Gentle Introduction to Stein s Method for Normal Approximation I Larry Goldstein University of Southern California Introduction to Stein s Method for Normal Approximation 1. Much activity since Stein

More information

Group Sequential Designs: Theory, Computation and Optimisation

Group Sequential Designs: Theory, Computation and Optimisation Group Sequential Designs: Theory, Computation and Optimisation Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj 8th International Conference

More information

Stein s Method and the Zero Bias Transformation with Application to Simple Random Sampling

Stein s Method and the Zero Bias Transformation with Application to Simple Random Sampling Stein s Method and the Zero Bias Transformation with Application to Simple Random Sampling Larry Goldstein and Gesine Reinert November 8, 001 Abstract Let W be a random variable with mean zero and variance

More information

Pubh 8482: Sequential Analysis

Pubh 8482: Sequential Analysis Pubh 8482: Sequential Analysis Joseph S. Koopmeiners Division of Biostatistics University of Minnesota Week 7 Course Summary To this point, we have discussed group sequential testing focusing on Maintaining

More information

MODERATE DEVIATIONS IN POISSON APPROXIMATION: A FIRST ATTEMPT

MODERATE DEVIATIONS IN POISSON APPROXIMATION: A FIRST ATTEMPT Statistica Sinica 23 (2013), 1523-1540 doi:http://dx.doi.org/10.5705/ss.2012.203s MODERATE DEVIATIONS IN POISSON APPROXIMATION: A FIRST ATTEMPT Louis H. Y. Chen 1, Xiao Fang 1,2 and Qi-Man Shao 3 1 National

More information

Wasserstein-2 bounds in normal approximation under local dependence

Wasserstein-2 bounds in normal approximation under local dependence Wasserstein- bounds in normal approximation under local dependence arxiv:1807.05741v1 [math.pr] 16 Jul 018 Xiao Fang The Chinese University of Hong Kong Abstract: We obtain a general bound for the Wasserstein-

More information

Nonconcave Penalized Likelihood with A Diverging Number of Parameters

Nonconcave Penalized Likelihood with A Diverging Number of Parameters Nonconcave Penalized Likelihood with A Diverging Number of Parameters Jianqing Fan and Heng Peng Presenter: Jiale Xu March 12, 2010 Jianqing Fan and Heng Peng Presenter: JialeNonconcave Xu () Penalized

More information

Gaussian Phase Transitions and Conic Intrinsic Volumes: Steining the Steiner formula

Gaussian Phase Transitions and Conic Intrinsic Volumes: Steining the Steiner formula Gaussian Phase Transitions and Conic Intrinsic Volumes: Steining the Steiner formula Larry Goldstein, University of Southern California Nourdin GIoVAnNi Peccati Luxembourg University University British

More information

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3 Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest

More information

Statistics 203: Introduction to Regression and Analysis of Variance Course review

Statistics 203: Introduction to Regression and Analysis of Variance Course review Statistics 203: Introduction to Regression and Analysis of Variance Course review Jonathan Taylor - p. 1/?? Today Review / overview of what we learned. - p. 2/?? General themes in regression models Specifying

More information

Statistical Inference with Monotone Incomplete Multivariate Normal Data

Statistical Inference with Monotone Incomplete Multivariate Normal Data Statistical Inference with Monotone Incomplete Multivariate Normal Data p. 1/4 Statistical Inference with Monotone Incomplete Multivariate Normal Data This talk is based on joint work with my wonderful

More information

Lecture 8: Information Theory and Statistics

Lecture 8: Information Theory and Statistics Lecture 8: Information Theory and Statistics Part II: Hypothesis Testing and I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 23, 2015 1 / 50 I-Hsiang

More information

Previous lecture. Single variant association. Use genome-wide SNPs to account for confounding (population substructure)

Previous lecture. Single variant association. Use genome-wide SNPs to account for confounding (population substructure) Previous lecture Single variant association Use genome-wide SNPs to account for confounding (population substructure) Estimation of effect size and winner s curse Meta-Analysis Today s outline P-value

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random

More information

ML Testing (Likelihood Ratio Testing) for non-gaussian models

ML Testing (Likelihood Ratio Testing) for non-gaussian models ML Testing (Likelihood Ratio Testing) for non-gaussian models Surya Tokdar ML test in a slightly different form Model X f (x θ), θ Θ. Hypothesist H 0 : θ Θ 0 Good set: B c (x) = {θ : l x (θ) max θ Θ l

More information

Mathematics Ph.D. Qualifying Examination Stat Probability, January 2018

Mathematics Ph.D. Qualifying Examination Stat Probability, January 2018 Mathematics Ph.D. Qualifying Examination Stat 52800 Probability, January 2018 NOTE: Answers all questions completely. Justify every step. Time allowed: 3 hours. 1. Let X 1,..., X n be a random sample from

More information

DA Freedman Notes on the MLE Fall 2003

DA Freedman Notes on the MLE Fall 2003 DA Freedman Notes on the MLE Fall 2003 The object here is to provide a sketch of the theory of the MLE. Rigorous presentations can be found in the references cited below. Calculus. Let f be a smooth, scalar

More information

ECE 275A Homework 7 Solutions

ECE 275A Homework 7 Solutions ECE 275A Homework 7 Solutions Solutions 1. For the same specification as in Homework Problem 6.11 we want to determine an estimator for θ using the Method of Moments (MOM). In general, the MOM estimator

More information

Qualifying Exam in Probability and Statistics. https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf

Qualifying Exam in Probability and Statistics. https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part : Sample Problems for the Elementary Section of Qualifying Exam in Probability and Statistics https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part 2: Sample Problems for the Advanced Section

More information

Stein s method and zero bias transformation: Application to CDO pricing

Stein s method and zero bias transformation: Application to CDO pricing Stein s method and zero bias transformation: Application to CDO pricing ESILV and Ecole Polytechnique Joint work with N. El Karoui Introduction CDO a portfolio credit derivative containing 100 underlying

More information

Group Sequential Tests for Delayed Responses. Christopher Jennison. Lisa Hampson. Workshop on Special Topics on Sequential Methodology

Group Sequential Tests for Delayed Responses. Christopher Jennison. Lisa Hampson. Workshop on Special Topics on Sequential Methodology Group Sequential Tests for Delayed Responses Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj Lisa Hampson Department of Mathematics and Statistics,

More information

f(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain

f(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain 0.1. INTRODUCTION 1 0.1 Introduction R. A. Fisher, a pioneer in the development of mathematical statistics, introduced a measure of the amount of information contained in an observaton from f(x θ). Fisher

More information

On bounds in multivariate Poisson approximation

On bounds in multivariate Poisson approximation Workshop on New Directions in Stein s Method IMS (NUS, Singapore), May 18-29, 2015 Acknowledgments Thanks Institute for Mathematical Sciences (IMS, NUS), for the invitation, hospitality and accommodation

More information

Previous lecture. P-value based combination. Fixed vs random effects models. Meta vs. pooled- analysis. New random effects testing.

Previous lecture. P-value based combination. Fixed vs random effects models. Meta vs. pooled- analysis. New random effects testing. Previous lecture P-value based combination. Fixed vs random effects models. Meta vs. pooled- analysis. New random effects testing. Interaction Outline: Definition of interaction Additive versus multiplicative

More information

On the Entropy of Sums of Bernoulli Random Variables via the Chen-Stein Method

On the Entropy of Sums of Bernoulli Random Variables via the Chen-Stein Method On the Entropy of Sums of Bernoulli Random Variables via the Chen-Stein Method Igal Sason Department of Electrical Engineering Technion - Israel Institute of Technology Haifa 32000, Israel ETH, Zurich,

More information

5. Conditional Distributions

5. Conditional Distributions 1 of 12 7/16/2009 5:36 AM Virtual Laboratories > 3. Distributions > 1 2 3 4 5 6 7 8 5. Conditional Distributions Basic Theory As usual, we start with a random experiment with probability measure P on an

More information

Inference in non-linear time series

Inference in non-linear time series Intro LS MLE Other Erik Lindström Centre for Mathematical Sciences Lund University LU/LTH & DTU Intro LS MLE Other General Properties Popular estimatiors Overview Introduction General Properties Estimators

More information

Gauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA

Gauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA JAPANESE BEETLE DATA 6 MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA Gauge Plots TuscaroraLisa Central Madsen Fairways, 996 January 9, 7 Grubs Adult Activity Grub Counts 6 8 Organic Matter

More information

Master s Written Examination - Solution

Master s Written Examination - Solution Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2

More information

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score

More information

Mathematics Qualifying Examination January 2015 STAT Mathematical Statistics

Mathematics Qualifying Examination January 2015 STAT Mathematical Statistics Mathematics Qualifying Examination January 2015 STAT 52800 - Mathematical Statistics NOTE: Answer all questions completely and justify your derivations and steps. A calculator and statistical tables (normal,

More information

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout

More information

BTRY 4090: Spring 2009 Theory of Statistics

BTRY 4090: Spring 2009 Theory of Statistics BTRY 4090: Spring 2009 Theory of Statistics Guozhang Wang September 25, 2010 1 Review of Probability We begin with a real example of using probability to solve computationally intensive (or infeasible)

More information

Lecture 3. Inference about multivariate normal distribution

Lecture 3. Inference about multivariate normal distribution Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates

More information

Multivariate Survival Analysis

Multivariate Survival Analysis Multivariate Survival Analysis Previously we have assumed that either (X i, δ i ) or (X i, δ i, Z i ), i = 1,..., n, are i.i.d.. This may not always be the case. Multivariate survival data can arise in

More information

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.

Unbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it

More information

The purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j.

The purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j. Chapter 9 Pearson s chi-square test 9. Null hypothesis asymptotics Let X, X 2, be independent from a multinomial(, p) distribution, where p is a k-vector with nonnegative entries that sum to one. That

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 59 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

Maximum likelihood estimation

Maximum likelihood estimation Maximum likelihood estimation Guillaume Obozinski Ecole des Ponts - ParisTech Master MVA Maximum likelihood estimation 1/26 Outline 1 Statistical concepts 2 A short review of convex analysis and optimization

More information

MAS223 Statistical Inference and Modelling Exercises

MAS223 Statistical Inference and Modelling Exercises MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,

More information

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Review. DS GA 1002 Statistical and Mathematical Models.   Carlos Fernandez-Granda Review DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing with

More information

Hypothesis Testing. Part I. James J. Heckman University of Chicago. Econ 312 This draft, April 20, 2006

Hypothesis Testing. Part I. James J. Heckman University of Chicago. Econ 312 This draft, April 20, 2006 Hypothesis Testing Part I James J. Heckman University of Chicago Econ 312 This draft, April 20, 2006 1 1 A Brief Review of Hypothesis Testing and Its Uses values and pure significance tests (R.A. Fisher)

More information

Mathematical statistics

Mathematical statistics October 1 st, 2018 Lecture 11: Sufficient statistic Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation

More information

Economics 583: Econometric Theory I A Primer on Asymptotics

Economics 583: Econometric Theory I A Primer on Asymptotics Economics 583: Econometric Theory I A Primer on Asymptotics Eric Zivot January 14, 2013 The two main concepts in asymptotic theory that we will use are Consistency Asymptotic Normality Intuition consistency:

More information

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:. MATHEMATICAL STATISTICS Homework assignment Instructions Please turn in the homework with this cover page. You do not need to edit the solutions. Just make sure the handwriting is legible. You may discuss

More information

11 Survival Analysis and Empirical Likelihood

11 Survival Analysis and Empirical Likelihood 11 Survival Analysis and Empirical Likelihood The first paper of empirical likelihood is actually about confidence intervals with the Kaplan-Meier estimator (Thomas and Grunkmeier 1979), i.e. deals with

More information

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Multilevel Models in Matrix Form Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Today s Lecture Linear models from a matrix perspective An example of how to do

More information

4. Comparison of Two (K) Samples

4. Comparison of Two (K) Samples 4. Comparison of Two (K) Samples K=2 Problem: compare the survival distributions between two groups. E: comparing treatments on patients with a particular disease. Z: Treatment indicator, i.e. Z = 1 for

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth

More information

Ferromagnets and the classical Heisenberg model. Kay Kirkpatrick, UIUC

Ferromagnets and the classical Heisenberg model. Kay Kirkpatrick, UIUC Ferromagnets and the classical Heisenberg model Kay Kirkpatrick, UIUC Ferromagnets and the classical Heisenberg model: asymptotics for a mean-field phase transition Kay Kirkpatrick, Urbana-Champaign June

More information

Methods of evaluating estimators and best unbiased estimators Hamid R. Rabiee

Methods of evaluating estimators and best unbiased estimators Hamid R. Rabiee Stochastic Processes Methods of evaluating estimators and best unbiased estimators Hamid R. Rabiee 1 Outline Methods of Mean Squared Error Bias and Unbiasedness Best Unbiased Estimators CR-Bound for variance

More information

STA 732: Inference. Notes 10. Parameter Estimation from a Decision Theoretic Angle. Other resources

STA 732: Inference. Notes 10. Parameter Estimation from a Decision Theoretic Angle. Other resources STA 732: Inference Notes 10. Parameter Estimation from a Decision Theoretic Angle Other resources 1 Statistical rules, loss and risk We saw that a major focus of classical statistics is comparing various

More information

5 Introduction to the Theory of Order Statistics and Rank Statistics

5 Introduction to the Theory of Order Statistics and Rank Statistics 5 Introduction to the Theory of Order Statistics and Rank Statistics This section will contain a summary of important definitions and theorems that will be useful for understanding the theory of order

More information

1.1 Basis of Statistical Decision Theory

1.1 Basis of Statistical Decision Theory ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 Lecture 1: Introduction Lecturer: Yihong Wu Scribe: AmirEmad Ghassami, Jan 21, 2016 [Ed. Jan 31] Outline: Introduction of

More information

HANDBOOK OF APPLICABLE MATHEMATICS

HANDBOOK OF APPLICABLE MATHEMATICS HANDBOOK OF APPLICABLE MATHEMATICS Chief Editor: Walter Ledermann Volume VI: Statistics PART A Edited by Emlyn Lloyd University of Lancaster A Wiley-Interscience Publication JOHN WILEY & SONS Chichester

More information

Comparing two independent samples

Comparing two independent samples In many applications it is necessary to compare two competing methods (for example, to compare treatment effects of a standard drug and an experimental drug). To compare two methods from statistical point

More information

Networks: Lectures 9 & 10 Random graphs

Networks: Lectures 9 & 10 Random graphs Networks: Lectures 9 & 10 Random graphs Heather A Harrington Mathematical Institute University of Oxford HT 2017 What you re in for Week 1: Introduction and basic concepts Week 2: Small worlds Week 3:

More information

Introduction to Probabilistic Machine Learning

Introduction to Probabilistic Machine Learning Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning

More information

Lecture 1: August 28

Lecture 1: August 28 36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random

More information

STEIN S METHOD, SEMICIRCLE DISTRIBUTION, AND REDUCED DECOMPOSITIONS OF THE LONGEST ELEMENT IN THE SYMMETRIC GROUP

STEIN S METHOD, SEMICIRCLE DISTRIBUTION, AND REDUCED DECOMPOSITIONS OF THE LONGEST ELEMENT IN THE SYMMETRIC GROUP STIN S MTHOD, SMICIRCL DISTRIBUTION, AND RDUCD DCOMPOSITIONS OF TH LONGST LMNT IN TH SYMMTRIC GROUP JASON FULMAN AND LARRY GOLDSTIN Abstract. Consider a uniformly chosen random reduced decomposition of

More information

Distributed Statistical Estimation and Rates of Convergence in Normal Approximation

Distributed Statistical Estimation and Rates of Convergence in Normal Approximation Distributed Statistical Estimation and Rates of Convergence in Normal Approximation Stas Minsker (joint with Nate Strawn) Department of Mathematics, USC July 3, 2017 Colloquium on Concentration inequalities,

More information

Model-Free Knockoffs: High-Dimensional Variable Selection that Controls the False Discovery Rate

Model-Free Knockoffs: High-Dimensional Variable Selection that Controls the False Discovery Rate Model-Free Knockoffs: High-Dimensional Variable Selection that Controls the False Discovery Rate Lucas Janson, Stanford Department of Statistics WADAPT Workshop, NIPS, December 2016 Collaborators: Emmanuel

More information

[y i α βx i ] 2 (2) Q = i=1

[y i α βx i ] 2 (2) Q = i=1 Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation

More information

Statistical Inference On the High-dimensional Gaussian Covarianc

Statistical Inference On the High-dimensional Gaussian Covarianc Statistical Inference On the High-dimensional Gaussian Covariance Matrix Department of Mathematical Sciences, Clemson University June 6, 2011 Outline Introduction Problem Setup Statistical Inference High-Dimensional

More information

Qualifying Exam in Probability and Statistics. https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf

Qualifying Exam in Probability and Statistics. https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part 1: Sample Problems for the Elementary Section of Qualifying Exam in Probability and Statistics https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part 2: Sample Problems for the Advanced Section

More information

Model Selection and Geometry

Model Selection and Geometry Model Selection and Geometry Pascal Massart Université Paris-Sud, Orsay Leipzig, February Purpose of the talk! Concentration of measure plays a fundamental role in the theory of model selection! Model

More information

3.0.1 Multivariate version and tensor product of experiments

3.0.1 Multivariate version and tensor product of experiments ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 Lecture 3: Minimax risk of GLM and four extensions Lecturer: Yihong Wu Scribe: Ashok Vardhan, Jan 28, 2016 [Ed. Mar 24]

More information

Adaptive Designs: Why, How and When?

Adaptive Designs: Why, How and When? Adaptive Designs: Why, How and When? Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj ISBS Conference Shanghai, July 2008 1 Adaptive designs:

More information

Lawrence D. Brown* and Daniel McCarthy*

Lawrence D. Brown* and Daniel McCarthy* Comments on the paper, An adaptive resampling test for detecting the presence of significant predictors by I. W. McKeague and M. Qian Lawrence D. Brown* and Daniel McCarthy* ABSTRACT: This commentary deals

More information

A Few Special Distributions and Their Properties

A Few Special Distributions and Their Properties A Few Special Distributions and Their Properties Econ 690 Purdue University Justin L. Tobias (Purdue) Distributional Catalog 1 / 20 Special Distributions and Their Associated Properties 1 Uniform Distribution

More information

PART I INTRODUCTION The meaning of probability Basic definitions for frequentist statistics and Bayesian inference Bayesian inference Combinatorics

PART I INTRODUCTION The meaning of probability Basic definitions for frequentist statistics and Bayesian inference Bayesian inference Combinatorics Table of Preface page xi PART I INTRODUCTION 1 1 The meaning of probability 3 1.1 Classical definition of probability 3 1.2 Statistical definition of probability 9 1.3 Bayesian understanding of probability

More information

First Year Examination Department of Statistics, University of Florida

First Year Examination Department of Statistics, University of Florida First Year Examination Department of Statistics, University of Florida August 19, 010, 8:00 am - 1:00 noon Instructions: 1. You have four hours to answer questions in this examination.. You must show your

More information

Multivariate Analysis and Likelihood Inference

Multivariate Analysis and Likelihood Inference Multivariate Analysis and Likelihood Inference Outline 1 Joint Distribution of Random Variables 2 Principal Component Analysis (PCA) 3 Multivariate Normal Distribution 4 Likelihood Inference Joint density

More information

Random matrices: Distribution of the least singular value (via Property Testing)

Random matrices: Distribution of the least singular value (via Property Testing) Random matrices: Distribution of the least singular value (via Property Testing) Van H. Vu Department of Mathematics Rutgers vanvu@math.rutgers.edu (joint work with T. Tao, UCLA) 1 Let ξ be a real or complex-valued

More information

Statistics 135 Fall 2008 Final Exam

Statistics 135 Fall 2008 Final Exam Name: SID: Statistics 135 Fall 2008 Final Exam Show your work. The number of points each question is worth is shown at the beginning of the question. There are 10 problems. 1. [2] The normal equations

More information

The Design of a Survival Study

The Design of a Survival Study The Design of a Survival Study The design of survival studies are usually based on the logrank test, and sometimes assumes the exponential distribution. As in standard designs, the power depends on The

More information

Machine Learning 2017

Machine Learning 2017 Machine Learning 2017 Volker Roth Department of Mathematics & Computer Science University of Basel 21st March 2017 Volker Roth (University of Basel) Machine Learning 2017 21st March 2017 1 / 41 Section

More information

Estimating network degree distributions from sampled networks: An inverse problem

Estimating network degree distributions from sampled networks: An inverse problem Estimating network degree distributions from sampled networks: An inverse problem Eric D. Kolaczyk Dept of Mathematics and Statistics, Boston University kolaczyk@bu.edu Introduction: Networks and Degree

More information

Stat 516, Homework 1

Stat 516, Homework 1 Stat 516, Homework 1 Due date: October 7 1. Consider an urn with n distinct balls numbered 1,..., n. We sample balls from the urn with replacement. Let N be the number of draws until we encounter a ball

More information

Independent Increments in Group Sequential Tests: A Review

Independent Increments in Group Sequential Tests: A Review Independent Increments in Group Sequential Tests: A Review KyungMann Kim kmkim@biostat.wisc.edu University of Wisconsin-Madison, Madison, WI, USA July 13, 2013 Outline Early Sequential Analysis Independent

More information

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. (a) If X and Y are independent, Corr(X, Y ) = 0. (b) (c) (d) (e) A consistent estimator must be asymptotically

More information

Part IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015

Part IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015 Part IB Statistics Theorems with proof Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly)

More information

STAT 461/561- Assignments, Year 2015

STAT 461/561- Assignments, Year 2015 STAT 461/561- Assignments, Year 2015 This is the second set of assignment problems. When you hand in any problem, include the problem itself and its number. pdf are welcome. If so, use large fonts and

More information

A Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression

A Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression A Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression Noah Simon Jerome Friedman Trevor Hastie November 5, 013 Abstract In this paper we purpose a blockwise descent

More information

Statistical Methods in HYDROLOGY CHARLES T. HAAN. The Iowa State University Press / Ames

Statistical Methods in HYDROLOGY CHARLES T. HAAN. The Iowa State University Press / Ames Statistical Methods in HYDROLOGY CHARLES T. HAAN The Iowa State University Press / Ames Univariate BASIC Table of Contents PREFACE xiii ACKNOWLEDGEMENTS xv 1 INTRODUCTION 1 2 PROBABILITY AND PROBABILITY

More information

Generalized Linear. Mixed Models. Methods and Applications. Modern Concepts, Walter W. Stroup. Texts in Statistical Science.

Generalized Linear. Mixed Models. Methods and Applications. Modern Concepts, Walter W. Stroup. Texts in Statistical Science. Texts in Statistical Science Generalized Linear Mixed Models Modern Concepts, Methods and Applications Walter W. Stroup CRC Press Taylor & Francis Croup Boca Raton London New York CRC Press is an imprint

More information