Stein s Method Applied to Some Statistical Problems
|
|
- Tobias Woods
- 5 years ago
- Views:
Transcription
1 Stein s Method Applied to Some Statistical Problems Jay Bartroff Borchard Colloquium 2017 Jay Bartroff (USC) Stein s for Stats 4.Jul.17 1 / 36
2 Outline of this talk 1. Stein s Method 2. Bounds to the normal for group sequential statistics with covariates Produce explicit distributional bounds to the limiting normal distribution for repeatedly-computed MLE (ˆθ 1, ˆθ 2,..., ˆθ K ) of a parameter vector θ R p in group sequential trials 3. Concentration inequalities for occupancy models with log-concave marginals How to get bounded, size biased couplings for certain multivariate occupancy models, then use these to get concentration inequalities Joint work with Larry Goldstein (USC) and Ümit Işlak (Boğaziçi University, Istanbul) Jay Bartroff (USC) Stein s for Stats 4.Jul.17 2 / 36
3 Outline of this talk 1. Stein s Method 2. Bounds to the normal for group sequential statistics with covariates Produce explicit distributional bounds to the limiting normal distribution for repeatedly-computed MLE (ˆθ 1, ˆθ 2,..., ˆθ K ) of a parameter vector θ R p in group sequential trials 3. Concentration inequalities for occupancy models with log-concave marginals How to get bounded, size biased couplings for certain multivariate occupancy models, then use these to get concentration inequalities Joint work with Larry Goldstein (USC) and Ümit Işlak (Boğaziçi University, Istanbul) Jay Bartroff (USC) Stein s for Stats 4.Jul.17 2 / 36
4 Outline of this talk 1. Stein s Method 2. Bounds to the normal for group sequential statistics with covariates Produce explicit distributional bounds to the limiting normal distribution for repeatedly-computed MLE (ˆθ 1, ˆθ 2,..., ˆθ K ) of a parameter vector θ R p in group sequential trials 3. Concentration inequalities for occupancy models with log-concave marginals How to get bounded, size biased couplings for certain multivariate occupancy models, then use these to get concentration inequalities Joint work with Larry Goldstein (USC) and Ümit Işlak (Boğaziçi University, Istanbul) Jay Bartroff (USC) Stein s for Stats 4.Jul.17 2 / 36
5 Outline of this talk 1. Stein s Method 2. Bounds to the normal for group sequential statistics with covariates Produce explicit distributional bounds to the limiting normal distribution for repeatedly-computed MLE (ˆθ 1, ˆθ 2,..., ˆθ K ) of a parameter vector θ R p in group sequential trials 3. Concentration inequalities for occupancy models with log-concave marginals How to get bounded, size biased couplings for certain multivariate occupancy models, then use these to get concentration inequalities Joint work with Larry Goldstein (USC) and Ümit Işlak (Boğaziçi University, Istanbul) Jay Bartroff (USC) Stein s for Stats 4.Jul.17 2 / 36
6 Charles Stein Stein s paradox: X in admissable for θ in N p (θ, I) for p 3 James-Stein shrinkage estimator Stein s unbiased risk estimator for MSE Stein s Lemma (1): Covariance estimation Stein s Lemma (2): Sequential sample size tails Stein s method for distributional approximation Jay Bartroff (USC) Stein s for Stats 4.Jul.17 3 / 36
7 Application 1: Group sequential analysis Jay Bartroff (USC) Stein s for Stats 4.Jul.17 4 / 36
8 Application 1: Group sequential analysis Jay Bartroff (USC) Stein s for Stats 4.Jul.17 5 / 36
9 But there are other books on this subject... Jay Bartroff (USC) Stein s for Stats 4.Jul.17 6 / 36
10 Setup Response Y i R of ith patient depends on known covariate vector x i unknown parameter vector θ R p Primary goal: To test a null hypothesis about θ, e.g., H 0 : θ = 0 H 0 : θ j 0 H 0 : at θ = b, some vector a, scalar b Secondary goals: Compute p-values or confidence regions for θ at the end of study Jay Bartroff (USC) Stein s for Stats 4.Jul.17 7 / 36
11 Setup Response Y i R of ith patient depends on known covariate vector x i unknown parameter vector θ R p Primary goal: To test a null hypothesis about θ, e.g., H 0 : θ = 0 H 0 : θ j 0 H 0 : at θ = b, some vector a, scalar b Secondary goals: Compute p-values or confidence regions for θ at the end of study Jay Bartroff (USC) Stein s for Stats 4.Jul.17 7 / 36
12 Setup Response Y i R of ith patient depends on known covariate vector x i unknown parameter vector θ R p Primary goal: To test a null hypothesis about θ, e.g., H 0 : θ = 0 H 0 : θ j 0 H 0 : at θ = b, some vector a, scalar b Secondary goals: Compute p-values or confidence regions for θ at the end of study Jay Bartroff (USC) Stein s for Stats 4.Jul.17 7 / 36
13 Setup: Group sequential analysis For efficiency, ethical, practical, financial reasons, the standard in trials has become group sequential analysis A group sequential trial with at most K groups Group 1: Group 2:. Group K: Y 1,..., Y n1 Y n1 +1,..., Y n2 Y nk 1 +1,..., Y nk Group sequential dominant format for clinical trials since... Beta-Blocker Heart Attack Trial ( BHAT, JAMA 82) Randomized trial of propranolol for heart attack survivors 3837 patients randomized Started June 1978, planned as 4-year study, terminated 8 months early due to observed benefit of propranolol Jay Bartroff (USC) Stein s for Stats 4.Jul.17 8 / 36
14 Setup: Group sequential analysis For efficiency, ethical, practical, financial reasons, the standard in trials has become group sequential analysis A group sequential trial with at most K groups Group 1: Group 2:. Group K: Y 1,..., Y n1 Y n1 +1,..., Y n2 Y nk 1 +1,..., Y nk Group sequential dominant format for clinical trials since... Beta-Blocker Heart Attack Trial ( BHAT, JAMA 82) Randomized trial of propranolol for heart attack survivors 3837 patients randomized Started June 1978, planned as 4-year study, terminated 8 months early due to observed benefit of propranolol Jay Bartroff (USC) Stein s for Stats 4.Jul.17 8 / 36
15 Setup: Group sequential analysis For efficiency, ethical, practical, financial reasons, the standard in trials has become group sequential analysis A group sequential trial with at most K groups Group 1: Group 2:. Group K: Y 1,..., Y n1 Y n1 +1,..., Y n2 Y nk 1 +1,..., Y nk Group sequential dominant format for clinical trials since... Beta-Blocker Heart Attack Trial ( BHAT, JAMA 82) Randomized trial of propranolol for heart attack survivors 3837 patients randomized Started June 1978, planned as 4-year study, terminated 8 months early due to observed benefit of propranolol Jay Bartroff (USC) Stein s for Stats 4.Jul.17 8 / 36
16 Setup: Group sequential analysis Stopping rule related to H 0 likelihood ratio, t-, F-, χ 2 - tests common Of the form: Stop and reject H 0 at stage min{k K : T (Y 1,..., Y nk ) C k } for some statistic T (Y 1,..., Y nk ), often a function of the MLE ˆθ k = ˆθ k (Y 1,..., Y nk ) The joint distribution of needed to ˆθ 1, ˆθ 2,..., ˆθ K choose critical values C k compute p-value at end of study give confidence region for θ at end of study Jay Bartroff (USC) Stein s for Stats 4.Jul.17 9 / 36
17 Setup: Group sequential analysis Stopping rule related to H 0 likelihood ratio, t-, F-, χ 2 - tests common Of the form: Stop and reject H 0 at stage min{k K : T (Y 1,..., Y nk ) C k } for some statistic T (Y 1,..., Y nk ), often a function of the MLE ˆθ k = ˆθ k (Y 1,..., Y nk ) The joint distribution of needed to ˆθ 1, ˆθ 2,..., ˆθ K choose critical values C k compute p-value at end of study give confidence region for θ at end of study Jay Bartroff (USC) Stein s for Stats 4.Jul.17 9 / 36
18 Setup: Group sequential analysis Stopping rule related to H 0 likelihood ratio, t-, F-, χ 2 - tests common Of the form: Stop and reject H 0 at stage min{k K : T (Y 1,..., Y nk ) C k } for some statistic T (Y 1,..., Y nk ), often a function of the MLE ˆθ k = ˆθ k (Y 1,..., Y nk ) The joint distribution of needed to ˆθ 1, ˆθ 2,..., ˆθ K choose critical values C k compute p-value at end of study give confidence region for θ at end of study Jay Bartroff (USC) Stein s for Stats 4.Jul.17 9 / 36
19 Background: Group sequential analysis Jennison & Turnbull (JASA 97) Asymptotic multivariate normal distribution of (ˆθ 1, ˆθ 2,..., ˆθ K ) in a regression setup Y i ind f i (Y i ; x i, θ) nice Asymptotics: n k n k 1 for all k, K fixed E ( θ k ) = θ Independent increments Cov (ˆθ k1, ˆθ k2 ) = Var (ˆθ k2 ) any k 1 k 2 Folk Theorem Normal limit widely (over-)used (software packages, etc.) before Jennison & Turnbull paper Commonly heard: Once n is 5 or so the normal limit kicks in! Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
20 Background: Group sequential analysis Jennison & Turnbull (JASA 97) Asymptotic multivariate normal distribution of (ˆθ 1, ˆθ 2,..., ˆθ K ) in a regression setup Y i ind f i (Y i ; x i, θ) nice Asymptotics: n k n k 1 for all k, K fixed E ( θ k ) = θ Independent increments Cov (ˆθ k1, ˆθ k2 ) = Var (ˆθ k2 ) any k 1 k 2 Folk Theorem Normal limit widely (over-)used (software packages, etc.) before Jennison & Turnbull paper Commonly heard: Once n is 5 or so the normal limit kicks in! Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
21 Background: Group sequential analysis Jennison & Turnbull (JASA 97) Asymptotic multivariate normal distribution of (ˆθ 1, ˆθ 2,..., ˆθ K ) in a regression setup Y i ind f i (Y i ; x i, θ) nice Asymptotics: n k n k 1 for all k, K fixed E ( θ k ) = θ Independent increments Cov (ˆθ k1, ˆθ k2 ) = Var (ˆθ k2 ) any k 1 k 2 Folk Theorem Normal limit widely (over-)used (software packages, etc.) before Jennison & Turnbull paper Commonly heard: Once n is 5 or so the normal limit kicks in! Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
22 Background: Group sequential analysis Jennison & Turnbull (JASA 97) Independent increments Suppose Then Cov (ˆθ k1, ˆθ k2 ) = Var (ˆθ k2 ) any k 1 k 2 H 0 : a t θ = 0, T k = (a t ˆθ k )I k = (a t ˆθ k )[Var (a t ˆθ k )] 1. Cov (T k1, T k2 ) = I k1 I k2 a t Cov (ˆθ k1, ˆθ k2 )a = I k1 I k2 a t Var (ˆθ k2 )a = I k1 I k2 Var (a t ˆθ k2 ) = I k1 I k2 I 1 k 2 = I k1 = Var (T k1 ) Cov (T k1, T k2 T k1 ) = 0 T 1, T 2 T 1,..., T K T K 1 asymptotically independent normals Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
23 Background: Group sequential analysis Jennison & Turnbull (JASA 97) Independent increments Suppose Then Cov (ˆθ k1, ˆθ k2 ) = Var (ˆθ k2 ) any k 1 k 2 H 0 : a t θ = 0, T k = (a t ˆθ k )I k = (a t ˆθ k )[Var (a t ˆθ k )] 1. Cov (T k1, T k2 ) = I k1 I k2 a t Cov (ˆθ k1, ˆθ k2 )a = I k1 I k2 a t Var (ˆθ k2 )a = I k1 I k2 Var (a t ˆθ k2 ) = I k1 I k2 I 1 k 2 = I k1 = Var (T k1 ) Cov (T k1, T k2 T k1 ) = 0 T 1, T 2 T 1,..., T K T K 1 asymptotically independent normals Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
24 Background: Group sequential analysis Jennison & Turnbull (JASA 97) Independent increments Suppose Then Cov (ˆθ k1, ˆθ k2 ) = Var (ˆθ k2 ) any k 1 k 2 H 0 : a t θ = 0, T k = (a t ˆθ k )I k = (a t ˆθ k )[Var (a t ˆθ k )] 1. Cov (T k1, T k2 ) = I k1 I k2 a t Cov (ˆθ k1, ˆθ k2 )a = I k1 I k2 a t Var (ˆθ k2 )a = I k1 I k2 Var (a t ˆθ k2 ) = I k1 I k2 I 1 k 2 = I k1 = Var (T k1 ) Cov (T k1, T k2 T k1 ) = 0 T 1, T 2 T 1,..., T K T K 1 asymptotically independent normals Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
25 Background: Group sequential analysis Jennison & Turnbull (JASA 97) Independent increments Suppose Then Cov (ˆθ k1, ˆθ k2 ) = Var (ˆθ k2 ) any k 1 k 2 H 0 : a t θ = 0, T k = (a t ˆθ k )I k = (a t ˆθ k )[Var (a t ˆθ k )] 1. Cov (T k1, T k2 ) = I k1 I k2 a t Cov (ˆθ k1, ˆθ k2 )a = I k1 I k2 a t Var (ˆθ k2 )a = I k1 I k2 Var (a t ˆθ k2 ) = I k1 I k2 I 1 k 2 = I k1 = Var (T k1 ) Cov (T k1, T k2 T k1 ) = 0 T 1, T 2 T 1,..., T K T K 1 asymptotically independent normals Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
26 New Contributions Extension to group sequential setting of Berry-Esseen bound for multivariate normal limit for smooth functions Anastasiou & Reinert 17: Bounds w/ explicit constants for bounded Wasserstein distance for scalar MLE (K = 1 analysis), Bernoulli. Anastasiou & Ley 17: Bounds for the asymptotic normality of the maximum likelihood estimator using the Delta method, ALEA. Anastasiou 17: Bounds for the normal approximation of the maximum likelihood estimator from m-dependent random variables. Statistics & Probability Letters. Anastasiou & Gaunt 16+: Multivariate normal approximation of the maximum likelihood estimator via the delta method, ArXiv: Anastasiou 15+: Assessing the multivariate normal approximation of the maximum likelihood estimator from high-dimensional, heterogeneous data., ArXiv: Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
27 New Contributions Extension to group sequential setting of Berry-Esseen bound for multivariate normal limit for smooth functions Anastasiou & Reinert 17: Bounds w/ explicit constants for bounded Wasserstein distance for scalar MLE (K = 1 analysis), Bernoulli. Anastasiou & Ley 17: Bounds for the asymptotic normality of the maximum likelihood estimator using the Delta method, ALEA. Anastasiou 17: Bounds for the normal approximation of the maximum likelihood estimator from m-dependent random variables. Statistics & Probability Letters. Anastasiou & Gaunt 16+: Multivariate normal approximation of the maximum likelihood estimator via the delta method, ArXiv: Anastasiou 15+: Assessing the multivariate normal approximation of the maximum likelihood estimator from high-dimensional, heterogeneous data., ArXiv: Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
28 New Contributions Extension to group sequential setting of Berry-Esseen bound for multivariate normal limit for smooth functions Anastasiou & Reinert 17: Bounds w/ explicit constants for bounded Wasserstein distance for scalar MLE (K = 1 analysis), Bernoulli. Anastasiou & Ley 17: Bounds for the asymptotic normality of the maximum likelihood estimator using the Delta method, ALEA. Anastasiou 17: Bounds for the normal approximation of the maximum likelihood estimator from m-dependent random variables. Statistics & Probability Letters. Anastasiou & Gaunt 16+: Multivariate normal approximation of the maximum likelihood estimator via the delta method, ArXiv: Anastasiou 15+: Assessing the multivariate normal approximation of the maximum likelihood estimator from high-dimensional, heterogeneous data., ArXiv: Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
29 New Contributions, Cont d Relaxing independence assumption: Assume log-likelihood of Y k := (Y nk 1 +1,..., Y nk ) is of the form for well-behaved functions f i, g k i G k log f i (Y i, θ) + g k (Y k, θ) g k = 0 gives Jennison & Turnbull s independent setting Some generalized linear mixed models (GLMMs) with random stage effect U k take this form U k = effect due to lab, monitoring board, cohort, etc. Penalized quasi-likelihood (Breslow & Clayton, JASA 93) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
30 New Contributions, Cont d Relaxing independence assumption: Assume log-likelihood of Y k := (Y nk 1 +1,..., Y nk ) is of the form for well-behaved functions f i, g k i G k log f i (Y i, θ) + g k (Y k, θ) g k = 0 gives Jennison & Turnbull s independent setting Some generalized linear mixed models (GLMMs) with random stage effect U k take this form U k = effect due to lab, monitoring board, cohort, etc. Penalized quasi-likelihood (Breslow & Clayton, JASA 93) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
31 New Contributions, Cont d Relaxing independence assumption: Assume log-likelihood of Y k := (Y nk 1 +1,..., Y nk ) is of the form for well-behaved functions f i, g k i G k log f i (Y i, θ) + g k (Y k, θ) g k = 0 gives Jennison & Turnbull s independent setting Some generalized linear mixed models (GLMMs) with random stage effect U k take this form U k = effect due to lab, monitoring board, cohort, etc. Penalized quasi-likelihood (Breslow & Clayton, JASA 93) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
32 New Contributions, Cont d Relaxing independence assumption: Assume log-likelihood of Y k := (Y nk 1 +1,..., Y nk ) is of the form for well-behaved functions f i, g k i G k log f i (Y i, θ) + g k (Y k, θ) g k = 0 gives Jennison & Turnbull s independent setting Some generalized linear mixed models (GLMMs) with random stage effect U k take this form U k = effect due to lab, monitoring board, cohort, etc. Penalized quasi-likelihood (Breslow & Clayton, JASA 93) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
33 GLMM Example: Poisson regression Letting f µ = Po(µ) density, For Y i in kth stage, Y i U k ind f µi where µ i = exp(β t x i + U k ) {U k } iid h λ θ = (β, λ). Then log-likelihood is K log f µi (Y i )h λ (U k )du k = k=1 i G k where µ i = exp(β t x i ). K log f µi (Y i ) + g k (Y k, θ) i Gk k=1 Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
34 GLMM Example: Poisson regression Letting f µ = Po(µ) density, For Y i in kth stage, Y i U k ind f µi where µ i = exp(β t x i + U k ) {U k } iid h λ θ = (β, λ). Then log-likelihood is K log f µi (Y i )h λ (U k )du k = k=1 i G k where µ i = exp(β t x i ). K log f µi (Y i ) + g k (Y k, θ) i Gk k=1 Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
35 GLMM Example: Poisson regression Letting f µ = Po(µ) density, For Y i in kth stage, Y i U k ind f µi where µ i = exp(β t x i + U k ) {U k } iid h λ θ = (β, λ). Then log-likelihood is K log f µi (Y i )h λ (U k )du k = k=1 i G k where µ i = exp(β t x i ). K log f µi (Y i ) + g k (Y k, θ) i Gk k=1 Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
36 GLMM Example: Poisson regression Letting f µ = Po(µ) density, For Y i in kth stage, Y i U k ind f µi where µ i = exp(β t x i + U k ) {U k } iid h λ θ = (β, λ). Then log-likelihood is K log f µi (Y i )h λ (U k )du k = k=1 i G k where µ i = exp(β t x i ). K log f µi (Y i ) + g k (Y k, θ) i Gk k=1 Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
37 Stein s Method for MVN Approximation Generator approach: Barbour 90, Goetze 91 Size biasing: Goldstein & Rinott 96, Rinott & Rotar 96 Zero biasing: Goldstein & Reinert 05 Exchangeable pair: Chatterjee & Meckes 08, Reinert & Röllin 09 Stein couplings: Fang & Röllin 15 Theorem (Reinert & Röllin 09) If W, W R q exchangeable pair with EW = 0, EWW t = Σ PD, and E(W W W ) = ΛW + R with Λ invertible, then for any 3-times differentiable h : R q R, Eh(W ) Eh(Σ 1/2 Z ) a h 2 4 for certain explicit constants a, b, c. + b h c ( h 1 + q 2 Σ 1/2 h 2 ) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
38 Stein s Method for MVN Approximation Generator approach: Barbour 90, Goetze 91 Size biasing: Goldstein & Rinott 96, Rinott & Rotar 96 Zero biasing: Goldstein & Reinert 05 Exchangeable pair: Chatterjee & Meckes 08, Reinert & Röllin 09 Stein couplings: Fang & Röllin 15 Theorem (Reinert & Röllin 09) If W, W R q exchangeable pair with EW = 0, EWW t = Σ PD, and E(W W W ) = ΛW + R with Λ invertible, then for any 3-times differentiable h : R q R, Eh(W ) Eh(Σ 1/2 Z ) a h 2 4 for certain explicit constants a, b, c. + b h c ( h 1 + q 2 Σ 1/2 h 2 ) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
39 Bounds to normal for ˆθ K := (ˆθ 1, ˆθ 2,..., ˆθ K ) Approach: Apply Reinert & Röllin 09 result with W = score function increments to get smooth function bounds to normal. Result In the group sequential setup above, if the Y i are independent or follow GLMMs with the log-likelihood of the kth group data Y k = (Y nk 1 +1,..., Y nk ) of the form i G k log f i (Y i, θ) + g k (Y k, θ), then under regularity conditions on the f i and g k there are a, b, c, d s.t. Eh(J 1/2 (ˆθ K θ K )) Eh(Z ) ak 2 J 1/2 2 h 2 + bk 3 J 1/2 3 h ck J 1/2 ( h 1 + pk 2 ) 2 Σ 1/2 J 1/2 h 2 + d. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
40 Bounds to normal for ˆθ K := (ˆθ 1, ˆθ 2,..., ˆθ K ) Approach: Apply Reinert & Röllin 09 result with W = score function increments to get smooth function bounds to normal. Result In the group sequential setup above, if the Y i are independent or follow GLMMs with the log-likelihood of the kth group data Y k = (Y nk 1 +1,..., Y nk ) of the form i G k log f i (Y i, θ) + g k (Y k, θ), then under regularity conditions on the f i and g k there are a, b, c, d s.t. Eh(J 1/2 (ˆθ K θ K )) Eh(Z ) ak 2 J 1/2 2 h 2 + bk 3 J 1/2 3 h ck J 1/2 ( h 1 + pk 2 ) 2 Σ 1/2 J 1/2 h 2 + d. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
41 Comments on result Bound Eh(J 1/2 (ˆθ K θ K )) Eh(Z ) ak 2 J 1/2 2 h 2 + bk 3 J 1/2 3 h ck J 1/2 ( h 1 + pk 2 ) 2 Σ 1/2 J 1/2 h 2 + d. a, b, c terms directly from Reinert & Röllin 09 bound c term Var(R) in E(W W W ) = ΛW + R, vanishes in independent case d term is from Taylor Series remainders Rate O(1/ n K ) under usual asymptotic n k n k 1 n K γ k (0, 1) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
42 Comments on result Bound Eh(J 1/2 (ˆθ K θ K )) Eh(Z ) ak 2 J 1/2 2 h 2 + bk 3 J 1/2 3 h ck J 1/2 ( h 1 + pk 2 ) 2 Σ 1/2 J 1/2 h 2 + d. a, b, c terms directly from Reinert & Röllin 09 bound c term Var(R) in E(W W W ) = ΛW + R, vanishes in independent case d term is from Taylor Series remainders Rate O(1/ n K ) under usual asymptotic n k n k 1 n K γ k (0, 1) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
43 Sketch of Proof Independent Case Score statistic S i (θ) = θ log f i(y i, θ) R p, W = S i (θ),..., S i (θ) R q, i G1 i G K where q = pk. Fisher Information ( ) J i (θ) = E θ S i(θ) t R p p J(θ 1,..., θ K ) = diag ( n1 J i (θ 1 ),..., i=1 n K i=1 J i (θ K ) ) R q q Σ := Var(W ) = diag J i (θ),..., J i (θ) R q q i G1 i G K Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
44 Sketch of Proof Independent Case Score statistic S i (θ) = θ log f i(y i, θ) R p, W = S i (θ),..., S i (θ) R q, i G1 i G K where q = pk. Fisher Information ( ) J i (θ) = E θ S i(θ) t R p p J(θ 1,..., θ K ) = diag ( n1 J i (θ 1 ),..., i=1 n K i=1 J i (θ K ) ) R q q Σ := Var(W ) = diag J i (θ),..., J i (θ) R q q i G1 i G K Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
45 Sketch of Proof Independent Case Score statistic S i (θ) = θ log f i(y i, θ) R p, W = S i (θ),..., S i (θ) R q, i G1 i G K where q = pk. Fisher Information ( ) J i (θ) = E θ S i(θ) t R p p J(θ 1,..., θ K ) = diag ( n1 J i (θ 1 ),..., i=1 n K i=1 J i (θ K ) ) R q q Σ := Var(W ) = diag J i (θ),..., J i (θ) R q q i G1 i G K Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
46 Sketch of Proof: Exchangeable pair Independent Case 1. Choose i {1,..., n K } uniformly, independent of all else 2. Replace Y i by independent copy Y i (keeping x i ), call result W W, W exchangeable W, W satisfy linearity condition which is easy to check entry-wise E(W W W ) = n 1 K W Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
47 Sketch of Proof: Exchangeable pair Independent Case 1. Choose i {1,..., n K } uniformly, independent of all else 2. Replace Y i by independent copy Y i (keeping x i ), call result W W, W exchangeable W, W satisfy linearity condition which is easy to check entry-wise E(W W W ) = n 1 K W Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
48 Sketch of Proof: Exchangeable pair Independent Case 1. Choose i {1,..., n K } uniformly, independent of all else 2. Replace Y i by independent copy Y i (keeping x i ), call result W W, W exchangeable W, W satisfy linearity condition which is easy to check entry-wise E(W W W ) = n 1 K W Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
49 Sketch of Proof: Exchangeable pair Independent Case 1. Choose i {1,..., n K } uniformly, independent of all else 2. Replace Y i by independent copy Y i (keeping x i ), call result W W, W exchangeable W, W satisfy linearity condition which is easy to check entry-wise E(W W W ) = n 1 K W Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
50 Sketch of Proof: Relating ˆθ K to W Independent Case By standard Taylor series, ˆθ K θ K = J(θ K ) 1 S, where S = ( n1 S i (θ 1 ),..., n K i=1 i=1 S i (θ K ) and θ K R q on line segment connecting ˆθ K, θ K. Then Eh(J 1/2 (ˆθ K θ K )) Eh(Z ) Eh(J 1/2 S) Eh(Z ) + Eh(J 1/2 J(θ K ) 1 S) Eh(J 1/2 S) ) R q Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
51 Sketch of Proof: Relating ˆθ K to W Independent Case By standard Taylor series, ˆθ K θ K = J(θ K ) 1 S, where S = ( n1 S i (θ 1 ),..., n K i=1 i=1 S i (θ K ) and θ K R q on line segment connecting ˆθ K, θ K. Then Eh(J 1/2 (ˆθ K θ K )) Eh(Z ) Eh(J 1/2 S) Eh(Z ) + Eh(J 1/2 J(θ K ) 1 S) Eh(J 1/2 S) ) R q Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
52 Sketch of Proof: Relating ˆθ K to W Independent Case Using S = AW where 1 p 0 p 0 p 1 p 1 p 0 p A = p 1 p 1 p 1st term is, 1 p, 0 p R p p identity and 0 matrices, Eh(J 1/2 S) Eh(Z ) = E h(w ) E h(σ 1/2 Z ) where h(w) = h(j 1/2 Aw), then apply Reinert-Röllin and simplify. 2nd term is bounded by Taylor series arguments. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
53 Sketch of Proof: Relating ˆθ K to W Independent Case Using S = AW where 1 p 0 p 0 p 1 p 1 p 0 p A = p 1 p 1 p 1st term is, 1 p, 0 p R p p identity and 0 matrices, Eh(J 1/2 S) Eh(Z ) = E h(w ) E h(σ 1/2 Z ) where h(w) = h(j 1/2 Aw), then apply Reinert-Röllin and simplify. 2nd term is bounded by Taylor series arguments. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
54 Sketch of Proof: Exchangeable pair GLMM Case 1. Choose i {1,..., n K } uniformly, independent of Y 1,..., Y nk 2. If i in kth group, replace Y i by independent copy Y i with mean ϕ(β t x i + U k ), ϕ 1 = link function (same covariates x i, group effect U k ), call result W W, W exchangeable W, W satisfy linearity condition where R = R(g 1,..., g K ) E(W W W ) = n 1 K W + R Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
55 Sketch of Proof: Exchangeable pair GLMM Case 1. Choose i {1,..., n K } uniformly, independent of Y 1,..., Y nk 2. If i in kth group, replace Y i by independent copy Y i with mean ϕ(β t x i + U k ), ϕ 1 = link function (same covariates x i, group effect U k ), call result W W, W exchangeable W, W satisfy linearity condition where R = R(g 1,..., g K ) E(W W W ) = n 1 K W + R Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
56 Sketch of Proof: Exchangeable pair GLMM Case 1. Choose i {1,..., n K } uniformly, independent of Y 1,..., Y nk 2. If i in kth group, replace Y i by independent copy Y i with mean ϕ(β t x i + U k ), ϕ 1 = link function (same covariates x i, group effect U k ), call result W W, W exchangeable W, W satisfy linearity condition where R = R(g 1,..., g K ) E(W W W ) = n 1 K W + R Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
57 Sketch of Proof: Exchangeable pair GLMM Case 1. Choose i {1,..., n K } uniformly, independent of Y 1,..., Y nk 2. If i in kth group, replace Y i by independent copy Y i with mean ϕ(β t x i + U k ), ϕ 1 = link function (same covariates x i, group effect U k ), call result W W, W exchangeable W, W satisfy linearity condition where R = R(g 1,..., g K ) E(W W W ) = n 1 K W + R Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
58 Sketch of Proof: Exchangeable pair GLMM Case 1. Choose i {1,..., n K } uniformly, independent of Y 1,..., Y nk 2. If i in kth group, replace Y i by independent copy Y i with mean ϕ(β t x i + U k ), ϕ 1 = link function (same covariates x i, group effect U k ), call result W W, W exchangeable W, W satisfy linearity condition where R = R(g 1,..., g K ) E(W W W ) = n 1 K W + R Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
59 Concentration and Coupling Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
60 Application 2: Concentration inequalities for occupancy models with log-concave marginals Main idea: How to get bounded, size biased couplings for certain multivariate occupancy models, then use methods pioneered by Ghosh and Goldstein 11 to get concentration inequalities Concentration Inequalities ( t 2 ) e.g., P (Y µ t) exp 2cµ + ct Widely used in high dimensional statistics machine learning random matrix theory applications: wireless communications, physics,... (See Raginsky & Sason 15) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
61 Application 2: Concentration inequalities for occupancy models with log-concave marginals Main idea: How to get bounded, size biased couplings for certain multivariate occupancy models, then use methods pioneered by Ghosh and Goldstein 11 to get concentration inequalities Concentration Inequalities ( t 2 ) e.g., P (Y µ t) exp 2cµ + ct Widely used in high dimensional statistics machine learning random matrix theory applications: wireless communications, physics,... (See Raginsky & Sason 15) Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
62 Setup Occupancy model M = (M α ) M α may be degree count of vertex α in an Erdös-Rényi random graph # of balls in box α in multinomial model # balls of color α in sample from urn of colored balls We consider statistics like m m Y ge = 1{M α d}, Y eq = 1{M α = d} Y ge = α=1 m w α 1{M α d α }, Y eq = α=1 α=1 α=1 m w α 1{M α = d α } Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
63 Setup Occupancy model M = (M α ) M α may be degree count of vertex α in an Erdös-Rényi random graph # of balls in box α in multinomial model # balls of color α in sample from urn of colored balls We consider statistics like m m Y ge = 1{M α d}, Y eq = 1{M α = d} Y ge = α=1 m w α 1{M α d α }, Y eq = α=1 α=1 α=1 m w α 1{M α = d α } Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
64 Setup Occupancy model M = (M α ) M α may be degree count of vertex α in an Erdös-Rényi random graph # of balls in box α in multinomial model # balls of color α in sample from urn of colored balls We consider statistics like m m Y ge = 1{M α d}, Y eq = 1{M α = d} Y ge = α=1 m w α 1{M α d α }, Y eq = α=1 α=1 α=1 m w α 1{M α = d α } Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
65 Setup Occupancy model M = (M α ) M α may be degree count of vertex α in an Erdös-Rényi random graph # of balls in box α in multinomial model # balls of color α in sample from urn of colored balls We consider statistics like m m Y ge = 1{M α d}, Y eq = 1{M α = d} Y ge = α=1 m w α 1{M α d α }, Y eq = α=1 α=1 α=1 m w α 1{M α = d α } Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
66 Some Methods for Concentration Inequalities McDiarmid s Bounded Difference Inequality Y a function with bounded differences of independent inputs Negative Association e.g., Dubashi & Ranjan 98 Certifiable Functions Y a certifiable function of independent inputs Controlling a large enough subset of inputs certifies value of function e.g., McDiarmid & Reed 06 Bounded Size Bias Couplings Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
67 Some Methods for Concentration Inequalities McDiarmid s Bounded Difference Inequality Y a function with bounded differences of independent inputs Negative Association e.g., Dubashi & Ranjan 98 Certifiable Functions Y a certifiable function of independent inputs Controlling a large enough subset of inputs certifies value of function e.g., McDiarmid & Reed 06 Bounded Size Bias Couplings Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
68 Some Methods for Concentration Inequalities McDiarmid s Bounded Difference Inequality Y a function with bounded differences of independent inputs Negative Association e.g., Dubashi & Ranjan 98 Certifiable Functions Y a certifiable function of independent inputs Controlling a large enough subset of inputs certifies value of function e.g., McDiarmid & Reed 06 Bounded Size Bias Couplings Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
69 Some Methods for Concentration Inequalities McDiarmid s Bounded Difference Inequality Y a function with bounded differences of independent inputs Negative Association e.g., Dubashi & Ranjan 98 Certifiable Functions Y a certifiable function of independent inputs Controlling a large enough subset of inputs certifies value of function e.g., McDiarmid & Reed 06 Bounded Size Bias Couplings Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
70 Bounded Size Bias Couplings If there is a coupling Y s of Y with the Y -size bias distribution, i.e., E[Yf (Y )] = µe[f (Y s )] for all f, and Y s Y + c for some c > 0 with probability one, then max {P(Y µ t), P(Y µ t)} b µ,c (t). Ghosh & Goldstein 11: For all t > 0, ) ( P (Y µ t) exp ( t2 t 2 P (Y µ t) exp 2cµ 2cµ + ct b exponential as t. Arratia & Baxendale 13: b µ,c (t) = exp b Poisson as t. ( µ c h ( t µ )), h(x) = (1 + x) log(1 + x) x. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36 ).
71 Bounded Size Bias Couplings If there is a coupling Y s of Y with the Y -size bias distribution, i.e., E[Yf (Y )] = µe[f (Y s )] for all f, and Y s Y + c for some c > 0 with probability one, then max {P(Y µ t), P(Y µ t)} b µ,c (t). Ghosh & Goldstein 11: For all t > 0, ) ( P (Y µ t) exp ( t2 t 2 P (Y µ t) exp 2cµ 2cµ + ct b exponential as t. Arratia & Baxendale 13: b µ,c (t) = exp b Poisson as t. ( µ c h ( t µ )), h(x) = (1 + x) log(1 + x) x. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36 ).
72 Bounded Size Bias Couplings If there is a coupling Y s of Y with the Y -size bias distribution, i.e., E[Yf (Y )] = µe[f (Y s )] for all f, and Y s Y + c for some c > 0 with probability one, then max {P(Y µ t), P(Y µ t)} b µ,c (t). Ghosh & Goldstein 11: For all t > 0, ) ( P (Y µ t) exp ( t2 t 2 P (Y µ t) exp 2cµ 2cµ + ct b exponential as t. Arratia & Baxendale 13: b µ,c (t) = exp b Poisson as t. ( µ c h ( t µ )), h(x) = (1 + x) log(1 + x) x. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36 ).
73 Main Result M = (M α ) α [m], M α lattice log-concave Y ge = w α 1{M α d α }, Y ne = w α 1{M α d α }. α [m] α [m] Main Result (in words) 1. If M is bounded from below and can be closely coupled to a version M having the same distribution conditional on M α = M α + 1, then there is a bounded size biased coupling Y s ge Y ge + C and the above concentration inequalities hold. 2. If M is non-degenerate at (d α ) and can be closely coupled to a version M having the same distribution conditional on M α d α, then there is a bounded size biased coupling Y s ne Y ne + C and the above concentration inequalities hold. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
74 Main Result M = (M α ) α [m], M α lattice log-concave Y ge = w α 1{M α d α }, Y ne = w α 1{M α d α }. α [m] α [m] Main Result (in words) 1. If M is bounded from below and can be closely coupled to a version M having the same distribution conditional on M α = M α + 1, then there is a bounded size biased coupling Y s ge Y ge + C and the above concentration inequalities hold. 2. If M is non-degenerate at (d α ) and can be closely coupled to a version M having the same distribution conditional on M α d α, then there is a bounded size biased coupling Y s ne Y ne + C and the above concentration inequalities hold. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
75 Main Result M = (M α ) α [m], M α lattice log-concave Y ge = w α 1{M α d α }, Y ne = w α 1{M α d α }. α [m] α [m] Main Result (in words) 1. If M is bounded from below and can be closely coupled to a version M having the same distribution conditional on M α = M α + 1, then there is a bounded size biased coupling Y s ge Y ge + C and the above concentration inequalities hold. 2. If M is non-degenerate at (d α ) and can be closely coupled to a version M having the same distribution conditional on M α d α, then there is a bounded size biased coupling Y s ne Y ne + C and the above concentration inequalities hold. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
76 Main Result A few more details on Part 1 M = f (U) where U is some collection of random variables f is measurable Closely coupled means given U k L(V k ) := L(U M α k), there is coupling U + k and constant B such that L(U + k U k) = L(V k M + k,α = M k,α + 1) and Y + k,ge, α Y k,ge, α + B, where Y k,ge, α = β α 1(M k,β d β ). The constant is C = w (B d + 1) where w = max w α, d = max d α. Part 2 is similar. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
77 Main Result A few more details on Part 1 M = f (U) where U is some collection of random variables f is measurable Closely coupled means given U k L(V k ) := L(U M α k), there is coupling U + k and constant B such that L(U + k U k) = L(V k M + k,α = M k,α + 1) and Y + k,ge, α Y k,ge, α + B, where Y k,ge, α = β α 1(M k,β d β ). The constant is C = w (B d + 1) where w = max w α, d = max d α. Part 2 is similar. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
78 Main Result A few more details on Part 1 M = f (U) where U is some collection of random variables f is measurable Closely coupled means given U k L(V k ) := L(U M α k), there is coupling U + k and constant B such that L(U + k U k) = L(V k M + k,α = M k,α + 1) and Y + k,ge, α Y k,ge, α + B, where Y k,ge, α = β α 1(M k,β d β ). The constant is C = w (B d + 1) where w = max w α, d = max d α. Part 2 is similar. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
79 Main Result A few more details on Part 1 M = f (U) where U is some collection of random variables f is measurable Closely coupled means given U k L(V k ) := L(U M α k), there is coupling U + k and constant B such that L(U + k U k) = L(V k M + k,α = M k,α + 1) and Y + k,ge, α Y k,ge, α + B, where Y k,ge, α = β α 1(M k,β d β ). The constant is C = w (B d + 1) where w = max w α, d = max d α. Part 2 is similar. Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
80 Main Result Main Ingredient in Proof Incrementing Lemma If M is lattice log-concave then there is π(x, d) [0, 1] such that if M L(M M d) and B M Bern(π(M, d)), then M + B L(M M d + 1). Extension of Goldstein & Penrose 10 for M Binomial, d = 0 Analogous versions for L(M M d) L(M M d 1) L(M) L(M M d) where means coupled to Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
81 Main Result Main Ingredient in Proof Incrementing Lemma If M is lattice log-concave then there is π(x, d) [0, 1] such that if M L(M M d) and B M Bern(π(M, d)), then M + B L(M M d + 1). Extension of Goldstein & Penrose 10 for M Binomial, d = 0 Analogous versions for L(M M d) L(M M d 1) L(M) L(M M d) where means coupled to Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
82 Example 1: Erdös-Rényi random graph m vertices Independent edges with probability p α,β = p β,α [0, 1). Constructing U + k from U k: 1. Selection non-neighbor β of α with probability 2. Add edge connecting β to α p α,β 1 p α,β This affects at most 1 other vertex so B = 1 and Y s ge Y ge + w ( d + 1). Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
83 Example 1: Erdös-Rényi random graph m vertices Independent edges with probability p α,β = p β,α [0, 1). Constructing U + k from U k: 1. Selection non-neighbor β of α with probability 2. Add edge connecting β to α p α,β 1 p α,β This affects at most 1 other vertex so B = 1 and Y s ge Y ge + w ( d + 1). Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
84 Example 1: Erdös-Rényi random graph m vertices Independent edges with probability p α,β = p β,α [0, 1). Constructing U + k from U k: 1. Selection non-neighbor β of α with probability 2. Add edge connecting β to α p α,β 1 p α,β This affects at most 1 other vertex so B = 1 and Y s ge Y ge + w ( d + 1). Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
85 Example 1: Erdös-Rényi random graph Applying this with d α = d, w α = 1, ( t 2 ) P(Y ge µ t) exp 2(d + 1)µ ( t 2 ) exp 2(d + 1)m Compare with McDiarmid s bounded difference inequality: so Y ge = f (X 1,..., X ( m 2) ), X i = 1{edge between vertex pair i}, sup f (X 1,..., X i,..., X X i,x ( m i 2) ) f (X 1,..., X i,..., X ( m 2) ) 2, ( t 2 ) P(Y ge µ t) exp 4m(m 1) New bound an improvement for m > 2d + 3 Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
86 Example 1: Erdös-Rényi random graph Applying this with d α = d, w α = 1, ( t 2 ) P(Y ge µ t) exp 2(d + 1)µ ( t 2 ) exp 2(d + 1)m Compare with McDiarmid s bounded difference inequality: so Y ge = f (X 1,..., X ( m 2) ), X i = 1{edge between vertex pair i}, sup f (X 1,..., X i,..., X X i,x ( m i 2) ) f (X 1,..., X i,..., X ( m 2) ) 2, ( t 2 ) P(Y ge µ t) exp 4m(m 1) New bound an improvement for m > 2d + 3 Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
87 Example 1: Erdös-Rényi random graph Applying this with d α = d, w α = 1, ( t 2 ) P(Y ge µ t) exp 2(d + 1)µ ( t 2 ) exp 2(d + 1)m Compare with McDiarmid s bounded difference inequality: so Y ge = f (X 1,..., X ( m 2) ), X i = 1{edge between vertex pair i}, sup f (X 1,..., X i,..., X X i,x ( m i 2) ) f (X 1,..., X i,..., X ( m 2) ) 2, ( t 2 ) P(Y ge µ t) exp 4m(m 1) New bound an improvement for m > 2d + 3 Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
88 Example 1: Erdös-Rényi random graph Applying this with d α = d, w α = 1, ( t 2 ) P(Y ge µ t) exp 2(d + 1)µ ( t 2 ) exp 2(d + 1)m Compare with McDiarmid s bounded difference inequality: so Y ge = f (X 1,..., X ( m 2) ), X i = 1{edge between vertex pair i}, sup f (X 1,..., X i,..., X X i,x ( m i 2) ) f (X 1,..., X i,..., X ( m 2) ) 2, ( t 2 ) P(Y ge µ t) exp 4m(m 1) New bound an improvement for m > 2d + 3 Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
89 Example 1: Erdös-Rényi random graph Applying this with d α = d, w α = 1, ( t 2 ) P(Y ge µ t) exp 2(d + 1)µ ( t 2 ) exp 2(d + 1)m Compare with McDiarmid s bounded difference inequality: so Y ge = f (X 1,..., X ( m 2) ), X i = 1{edge between vertex pair i}, sup f (X 1,..., X i,..., X X i,x ( m i 2) ) f (X 1,..., X i,..., X ( m 2) ) 2, ( t 2 ) P(Y ge µ t) exp 4m(m 1) New bound an improvement for m > 2d + 3 Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
90 Example 2: Multinomial Counts n balls independently into m boxes Applications in species trapping, linguistics,... # empty boxes proved asymptotically normal by Weiss 58, Rényi 62 in uniform case Englund 81: L bound for # of empty cells, uniform case Dubashi & Ranjan 98: Concentration inequality via NA Penrose 09: L bound for # of isolated balls, uniform and nonuniform cases Bartroff & Goldstein 13: L bound for all d 2, uniform case Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
91 Example 2: Multinomial Counts n balls independently into m boxes Applications in species trapping, linguistics,... # empty boxes proved asymptotically normal by Weiss 58, Rényi 62 in uniform case Englund 81: L bound for # of empty cells, uniform case Dubashi & Ranjan 98: Concentration inequality via NA Penrose 09: L bound for # of isolated balls, uniform and nonuniform cases Bartroff & Goldstein 13: L bound for all d 2, uniform case Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
92 Example 2: Multinomial Counts n balls independently into m boxes Applications in species trapping, linguistics,... # empty boxes proved asymptotically normal by Weiss 58, Rényi 62 in uniform case Englund 81: L bound for # of empty cells, uniform case Dubashi & Ranjan 98: Concentration inequality via NA Penrose 09: L bound for # of isolated balls, uniform and nonuniform cases Bartroff & Goldstein 13: L bound for all d 2, uniform case Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
93 Example 2: Multinomial Counts n balls independently into m boxes Applications in species trapping, linguistics,... # empty boxes proved asymptotically normal by Weiss 58, Rényi 62 in uniform case Englund 81: L bound for # of empty cells, uniform case Dubashi & Ranjan 98: Concentration inequality via NA Penrose 09: L bound for # of isolated balls, uniform and nonuniform cases Bartroff & Goldstein 13: L bound for all d 2, uniform case Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
94 Example 2: Multinomial Counts n balls independently into m boxes Applications in species trapping, linguistics,... # empty boxes proved asymptotically normal by Weiss 58, Rényi 62 in uniform case Englund 81: L bound for # of empty cells, uniform case Dubashi & Ranjan 98: Concentration inequality via NA Penrose 09: L bound for # of isolated balls, uniform and nonuniform cases Bartroff & Goldstein 13: L bound for all d 2, uniform case Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
95 Example 2: Multinomial Counts n balls independently into m boxes Applications in species trapping, linguistics,... # empty boxes proved asymptotically normal by Weiss 58, Rényi 62 in uniform case Englund 81: L bound for # of empty cells, uniform case Dubashi & Ranjan 98: Concentration inequality via NA Penrose 09: L bound for # of isolated balls, uniform and nonuniform cases Bartroff & Goldstein 13: L bound for all d 2, uniform case Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
96 Example 2: Multinomial Counts n balls independently into m boxes Applications in species trapping, linguistics,... # empty boxes proved asymptotically normal by Weiss 58, Rényi 62 in uniform case Englund 81: L bound for # of empty cells, uniform case Dubashi & Ranjan 98: Concentration inequality via NA Penrose 09: L bound for # of isolated balls, uniform and nonuniform cases Bartroff & Goldstein 13: L bound for all d 2, uniform case Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
97 Example 2: Multinomial Counts p α,j = prob. that ball j [n] falls in box α [m] M α = # balls in box α = 1{ball j falls in box α} j [n] Constructing U + k from U k: Choose ball j box α with probability p α,j 1 p α,j and add it to box α. Y s ge, α Y ge, α so B = 0, thus Y s ge Y ge + w Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
98 Example 2: Multinomial Counts p α,j = prob. that ball j [n] falls in box α [m] M α = # balls in box α = 1{ball j falls in box α} j [n] Constructing U + k from U k: Choose ball j box α with probability p α,j 1 p α,j and add it to box α. Y s ge, α Y ge, α so B = 0, thus Y s ge Y ge + w Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
99 Example 2: Multinomial Counts p α,j = prob. that ball j [n] falls in box α [m] M α = # balls in box α = 1{ball j falls in box α} j [n] Constructing U + k from U k: Choose ball j box α with probability p α,j 1 p α,j and add it to box α. Y s ge, α Y ge, α so B = 0, thus Y s ge Y ge + w Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
100 Example 3: Multivariate Hypergeometric Sampling Urn with n = α [m] n α colored balls, n α balls of color α Sample of size s drawn without replacement M α = # balls in sample of color α Applications in sampling (and subsampling) theory, gambling, coupon-collector problems Constructing U + k probability from U k: Select non-α colored ball in sample with n α(j)/n and replace it with α-colored ball, α(j) = color of ball j 1 n α(j) /n Y s ge, α Y ge, α so B = 0, thus Y s ge Y ge + w Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
101 Example 3: Multivariate Hypergeometric Sampling Urn with n = α [m] n α colored balls, n α balls of color α Sample of size s drawn without replacement M α = # balls in sample of color α Applications in sampling (and subsampling) theory, gambling, coupon-collector problems Constructing U + k probability from U k: Select non-α colored ball in sample with n α(j)/n and replace it with α-colored ball, α(j) = color of ball j 1 n α(j) /n Y s ge, α Y ge, α so B = 0, thus Y s ge Y ge + w Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
102 Example 3: Multivariate Hypergeometric Sampling Urn with n = α [m] n α colored balls, n α balls of color α Sample of size s drawn without replacement M α = # balls in sample of color α Applications in sampling (and subsampling) theory, gambling, coupon-collector problems Constructing U + k probability from U k: Select non-α colored ball in sample with n α(j)/n and replace it with α-colored ball, α(j) = color of ball j 1 n α(j) /n Y s ge, α Y ge, α so B = 0, thus Y s ge Y ge + w Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
103 Example 3: Multivariate Hypergeometric Sampling Urn with n = α [m] n α colored balls, n α balls of color α Sample of size s drawn without replacement M α = # balls in sample of color α Applications in sampling (and subsampling) theory, gambling, coupon-collector problems Constructing U + k probability from U k: Select non-α colored ball in sample with n α(j)/n and replace it with α-colored ball, α(j) = color of ball j 1 n α(j) /n Y s ge, α Y ge, α so B = 0, thus Y s ge Y ge + w Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
104 Example 3: Multivariate Hypergeometric Sampling Urn with n = α [m] n α colored balls, n α balls of color α Sample of size s drawn without replacement M α = # balls in sample of color α Applications in sampling (and subsampling) theory, gambling, coupon-collector problems Constructing U + k probability from U k: Select non-α colored ball in sample with n α(j)/n and replace it with α-colored ball, α(j) = color of ball j 1 n α(j) /n Y s ge, α Y ge, α so B = 0, thus Y s ge Y ge + w Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
105 Example 3: Multivariate Hypergeometric Sampling Urn with n = α [m] n α colored balls, n α balls of color α Sample of size s drawn without replacement M α = # balls in sample of color α Applications in sampling (and subsampling) theory, gambling, coupon-collector problems Constructing U + k probability from U k: Select non-α colored ball in sample with n α(j)/n and replace it with α-colored ball, α(j) = color of ball j 1 n α(j) /n Y s ge, α Y ge, α so B = 0, thus Y s ge Y ge + w Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
106 Summary Stein s method applied to produce explicit bounds to the limiting normal distribution for repeatedly-computed MLE (ˆθ 1, ˆθ 2,..., ˆθ K ) of a parameter vector in group sequential trials concentration inequalities for a class of occupancy models with log-concave marginals Many unanswered questions in statistics possibly susceptible to Stein s concentration inequalities for heavy-tailed distributions convergence of empirical measures and dimension reduction methods projections of empirical measures onto subspaces high-dimensional PCA other problems in sequential analysis sequentially stopped test statistics stopping rules for high-dimensional MCMC Thank You! Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
107 Summary Stein s method applied to produce explicit bounds to the limiting normal distribution for repeatedly-computed MLE (ˆθ 1, ˆθ 2,..., ˆθ K ) of a parameter vector in group sequential trials concentration inequalities for a class of occupancy models with log-concave marginals Many unanswered questions in statistics possibly susceptible to Stein s concentration inequalities for heavy-tailed distributions convergence of empirical measures and dimension reduction methods projections of empirical measures onto subspaces high-dimensional PCA other problems in sequential analysis sequentially stopped test statistics stopping rules for high-dimensional MCMC Thank You! Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
108 Summary Stein s method applied to produce explicit bounds to the limiting normal distribution for repeatedly-computed MLE (ˆθ 1, ˆθ 2,..., ˆθ K ) of a parameter vector in group sequential trials concentration inequalities for a class of occupancy models with log-concave marginals Many unanswered questions in statistics possibly susceptible to Stein s concentration inequalities for heavy-tailed distributions convergence of empirical measures and dimension reduction methods projections of empirical measures onto subspaces high-dimensional PCA other problems in sequential analysis sequentially stopped test statistics stopping rules for high-dimensional MCMC Thank You! Jay Bartroff (USC) Stein s for Stats 4.Jul / 36
Stein Couplings for Concentration of Measure
Stein Couplings for Concentration of Measure Jay Bartroff, Subhankar Ghosh, Larry Goldstein and Ümit Işlak University of Southern California [arxiv:0906.3886] [arxiv:1304.5001] [arxiv:1402.6769] Borchard
More informationConcentration of Measures by Bounded Couplings
Concentration of Measures by Bounded Couplings Subhankar Ghosh, Larry Goldstein and Ümit Işlak University of Southern California [arxiv:0906.3886] [arxiv:1304.5001] May 2013 Concentration of Measure Distributional
More informationStein s Method: Distributional Approximation and Concentration of Measure
Stein s Method: Distributional Approximation and Concentration of Measure Larry Goldstein University of Southern California 36 th Midwest Probability Colloquium, 2014 Stein s method for Distributional
More informationStein s Method: Distributional Approximation and Concentration of Measure
Stein s Method: Distributional Approximation and Concentration of Measure Larry Goldstein University of Southern California 36 th Midwest Probability Colloquium, 2014 Concentration of Measure Distributional
More informationSubmitted to the Brazilian Journal of Probability and Statistics
Submitted to the Brazilian Journal of Probability and Statistics Multivariate normal approximation of the maximum likelihood estimator via the delta method Andreas Anastasiou a and Robert E. Gaunt b a
More informationConcentration of Measures by Bounded Size Bias Couplings
Concentration of Measures by Bounded Size Bias Couplings Subhankar Ghosh, Larry Goldstein University of Southern California [arxiv:0906.3886] January 10 th, 2013 Concentration of Measure Distributional
More informationInterim Monitoring of Clinical Trials: Decision Theory, Dynamic Programming. and Optimal Stopping
Interim Monitoring of Clinical Trials: Decision Theory, Dynamic Programming and Optimal Stopping Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj
More informationStat 5101 Lecture Notes
Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random
More informationDifferential Stein operators for multivariate continuous distributions and applications
Differential Stein operators for multivariate continuous distributions and applications Gesine Reinert A French/American Collaborative Colloquium on Concentration Inequalities, High Dimensional Statistics
More informationOptimising Group Sequential Designs. Decision Theory, Dynamic Programming. and Optimal Stopping
: Decision Theory, Dynamic Programming and Optimal Stopping Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj InSPiRe Conference on Methodology
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationSpring 2012 Math 541B Exam 1
Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote
More informationA Gentle Introduction to Stein s Method for Normal Approximation I
A Gentle Introduction to Stein s Method for Normal Approximation I Larry Goldstein University of Southern California Introduction to Stein s Method for Normal Approximation 1. Much activity since Stein
More informationGroup Sequential Designs: Theory, Computation and Optimisation
Group Sequential Designs: Theory, Computation and Optimisation Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj 8th International Conference
More informationStein s Method and the Zero Bias Transformation with Application to Simple Random Sampling
Stein s Method and the Zero Bias Transformation with Application to Simple Random Sampling Larry Goldstein and Gesine Reinert November 8, 001 Abstract Let W be a random variable with mean zero and variance
More informationPubh 8482: Sequential Analysis
Pubh 8482: Sequential Analysis Joseph S. Koopmeiners Division of Biostatistics University of Minnesota Week 7 Course Summary To this point, we have discussed group sequential testing focusing on Maintaining
More informationMODERATE DEVIATIONS IN POISSON APPROXIMATION: A FIRST ATTEMPT
Statistica Sinica 23 (2013), 1523-1540 doi:http://dx.doi.org/10.5705/ss.2012.203s MODERATE DEVIATIONS IN POISSON APPROXIMATION: A FIRST ATTEMPT Louis H. Y. Chen 1, Xiao Fang 1,2 and Qi-Man Shao 3 1 National
More informationWasserstein-2 bounds in normal approximation under local dependence
Wasserstein- bounds in normal approximation under local dependence arxiv:1807.05741v1 [math.pr] 16 Jul 018 Xiao Fang The Chinese University of Hong Kong Abstract: We obtain a general bound for the Wasserstein-
More informationNonconcave Penalized Likelihood with A Diverging Number of Parameters
Nonconcave Penalized Likelihood with A Diverging Number of Parameters Jianqing Fan and Heng Peng Presenter: Jiale Xu March 12, 2010 Jianqing Fan and Heng Peng Presenter: JialeNonconcave Xu () Penalized
More informationGaussian Phase Transitions and Conic Intrinsic Volumes: Steining the Steiner formula
Gaussian Phase Transitions and Conic Intrinsic Volumes: Steining the Steiner formula Larry Goldstein, University of Southern California Nourdin GIoVAnNi Peccati Luxembourg University University British
More informationHypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3
Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest
More informationStatistics 203: Introduction to Regression and Analysis of Variance Course review
Statistics 203: Introduction to Regression and Analysis of Variance Course review Jonathan Taylor - p. 1/?? Today Review / overview of what we learned. - p. 2/?? General themes in regression models Specifying
More informationStatistical Inference with Monotone Incomplete Multivariate Normal Data
Statistical Inference with Monotone Incomplete Multivariate Normal Data p. 1/4 Statistical Inference with Monotone Incomplete Multivariate Normal Data This talk is based on joint work with my wonderful
More informationLecture 8: Information Theory and Statistics
Lecture 8: Information Theory and Statistics Part II: Hypothesis Testing and I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 23, 2015 1 / 50 I-Hsiang
More informationPrevious lecture. Single variant association. Use genome-wide SNPs to account for confounding (population substructure)
Previous lecture Single variant association Use genome-wide SNPs to account for confounding (population substructure) Estimation of effect size and winner s curse Meta-Analysis Today s outline P-value
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random
More informationML Testing (Likelihood Ratio Testing) for non-gaussian models
ML Testing (Likelihood Ratio Testing) for non-gaussian models Surya Tokdar ML test in a slightly different form Model X f (x θ), θ Θ. Hypothesist H 0 : θ Θ 0 Good set: B c (x) = {θ : l x (θ) max θ Θ l
More informationMathematics Ph.D. Qualifying Examination Stat Probability, January 2018
Mathematics Ph.D. Qualifying Examination Stat 52800 Probability, January 2018 NOTE: Answers all questions completely. Justify every step. Time allowed: 3 hours. 1. Let X 1,..., X n be a random sample from
More informationDA Freedman Notes on the MLE Fall 2003
DA Freedman Notes on the MLE Fall 2003 The object here is to provide a sketch of the theory of the MLE. Rigorous presentations can be found in the references cited below. Calculus. Let f be a smooth, scalar
More informationECE 275A Homework 7 Solutions
ECE 275A Homework 7 Solutions Solutions 1. For the same specification as in Homework Problem 6.11 we want to determine an estimator for θ using the Method of Moments (MOM). In general, the MOM estimator
More informationQualifying Exam in Probability and Statistics. https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf
Part : Sample Problems for the Elementary Section of Qualifying Exam in Probability and Statistics https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part 2: Sample Problems for the Advanced Section
More informationStein s method and zero bias transformation: Application to CDO pricing
Stein s method and zero bias transformation: Application to CDO pricing ESILV and Ecole Polytechnique Joint work with N. El Karoui Introduction CDO a portfolio credit derivative containing 100 underlying
More informationGroup Sequential Tests for Delayed Responses. Christopher Jennison. Lisa Hampson. Workshop on Special Topics on Sequential Methodology
Group Sequential Tests for Delayed Responses Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj Lisa Hampson Department of Mathematics and Statistics,
More informationf(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain
0.1. INTRODUCTION 1 0.1 Introduction R. A. Fisher, a pioneer in the development of mathematical statistics, introduced a measure of the amount of information contained in an observaton from f(x θ). Fisher
More informationOn bounds in multivariate Poisson approximation
Workshop on New Directions in Stein s Method IMS (NUS, Singapore), May 18-29, 2015 Acknowledgments Thanks Institute for Mathematical Sciences (IMS, NUS), for the invitation, hospitality and accommodation
More informationPrevious lecture. P-value based combination. Fixed vs random effects models. Meta vs. pooled- analysis. New random effects testing.
Previous lecture P-value based combination. Fixed vs random effects models. Meta vs. pooled- analysis. New random effects testing. Interaction Outline: Definition of interaction Additive versus multiplicative
More informationOn the Entropy of Sums of Bernoulli Random Variables via the Chen-Stein Method
On the Entropy of Sums of Bernoulli Random Variables via the Chen-Stein Method Igal Sason Department of Electrical Engineering Technion - Israel Institute of Technology Haifa 32000, Israel ETH, Zurich,
More information5. Conditional Distributions
1 of 12 7/16/2009 5:36 AM Virtual Laboratories > 3. Distributions > 1 2 3 4 5 6 7 8 5. Conditional Distributions Basic Theory As usual, we start with a random experiment with probability measure P on an
More informationInference in non-linear time series
Intro LS MLE Other Erik Lindström Centre for Mathematical Sciences Lund University LU/LTH & DTU Intro LS MLE Other General Properties Popular estimatiors Overview Introduction General Properties Estimators
More informationGauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA
JAPANESE BEETLE DATA 6 MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA Gauge Plots TuscaroraLisa Central Madsen Fairways, 996 January 9, 7 Grubs Adult Activity Grub Counts 6 8 Organic Matter
More informationMaster s Written Examination - Solution
Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2
More informationStatistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach
Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score
More informationMathematics Qualifying Examination January 2015 STAT Mathematical Statistics
Mathematics Qualifying Examination January 2015 STAT 52800 - Mathematical Statistics NOTE: Answer all questions completely and justify your derivations and steps. A calculator and statistical tables (normal,
More informationSTATS 200: Introduction to Statistical Inference. Lecture 29: Course review
STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout
More informationBTRY 4090: Spring 2009 Theory of Statistics
BTRY 4090: Spring 2009 Theory of Statistics Guozhang Wang September 25, 2010 1 Review of Probability We begin with a real example of using probability to solve computationally intensive (or infeasible)
More informationLecture 3. Inference about multivariate normal distribution
Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates
More informationMultivariate Survival Analysis
Multivariate Survival Analysis Previously we have assumed that either (X i, δ i ) or (X i, δ i, Z i ), i = 1,..., n, are i.i.d.. This may not always be the case. Multivariate survival data can arise in
More informationUnbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.
Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it
More informationThe purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j.
Chapter 9 Pearson s chi-square test 9. Null hypothesis asymptotics Let X, X 2, be independent from a multinomial(, p) distribution, where p is a k-vector with nonnegative entries that sum to one. That
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 59 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d
More informationLecture 7 Introduction to Statistical Decision Theory
Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7
More informationMaximum likelihood estimation
Maximum likelihood estimation Guillaume Obozinski Ecole des Ponts - ParisTech Master MVA Maximum likelihood estimation 1/26 Outline 1 Statistical concepts 2 A short review of convex analysis and optimization
More informationMAS223 Statistical Inference and Modelling Exercises
MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,
More informationReview. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda
Review DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing with
More informationHypothesis Testing. Part I. James J. Heckman University of Chicago. Econ 312 This draft, April 20, 2006
Hypothesis Testing Part I James J. Heckman University of Chicago Econ 312 This draft, April 20, 2006 1 1 A Brief Review of Hypothesis Testing and Its Uses values and pure significance tests (R.A. Fisher)
More informationMathematical statistics
October 1 st, 2018 Lecture 11: Sufficient statistic Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation
More informationEconomics 583: Econometric Theory I A Primer on Asymptotics
Economics 583: Econometric Theory I A Primer on Asymptotics Eric Zivot January 14, 2013 The two main concepts in asymptotic theory that we will use are Consistency Asymptotic Normality Intuition consistency:
More informationStatement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.
MATHEMATICAL STATISTICS Homework assignment Instructions Please turn in the homework with this cover page. You do not need to edit the solutions. Just make sure the handwriting is legible. You may discuss
More information11 Survival Analysis and Empirical Likelihood
11 Survival Analysis and Empirical Likelihood The first paper of empirical likelihood is actually about confidence intervals with the Kaplan-Meier estimator (Thomas and Grunkmeier 1979), i.e. deals with
More informationMultilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2
Multilevel Models in Matrix Form Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Today s Lecture Linear models from a matrix perspective An example of how to do
More information4. Comparison of Two (K) Samples
4. Comparison of Two (K) Samples K=2 Problem: compare the survival distributions between two groups. E: comparing treatments on patients with a particular disease. Z: Treatment indicator, i.e. Z = 1 for
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)
More informationMaster s Written Examination
Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth
More informationFerromagnets and the classical Heisenberg model. Kay Kirkpatrick, UIUC
Ferromagnets and the classical Heisenberg model Kay Kirkpatrick, UIUC Ferromagnets and the classical Heisenberg model: asymptotics for a mean-field phase transition Kay Kirkpatrick, Urbana-Champaign June
More informationMethods of evaluating estimators and best unbiased estimators Hamid R. Rabiee
Stochastic Processes Methods of evaluating estimators and best unbiased estimators Hamid R. Rabiee 1 Outline Methods of Mean Squared Error Bias and Unbiasedness Best Unbiased Estimators CR-Bound for variance
More informationSTA 732: Inference. Notes 10. Parameter Estimation from a Decision Theoretic Angle. Other resources
STA 732: Inference Notes 10. Parameter Estimation from a Decision Theoretic Angle Other resources 1 Statistical rules, loss and risk We saw that a major focus of classical statistics is comparing various
More information5 Introduction to the Theory of Order Statistics and Rank Statistics
5 Introduction to the Theory of Order Statistics and Rank Statistics This section will contain a summary of important definitions and theorems that will be useful for understanding the theory of order
More information1.1 Basis of Statistical Decision Theory
ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 Lecture 1: Introduction Lecturer: Yihong Wu Scribe: AmirEmad Ghassami, Jan 21, 2016 [Ed. Jan 31] Outline: Introduction of
More informationHANDBOOK OF APPLICABLE MATHEMATICS
HANDBOOK OF APPLICABLE MATHEMATICS Chief Editor: Walter Ledermann Volume VI: Statistics PART A Edited by Emlyn Lloyd University of Lancaster A Wiley-Interscience Publication JOHN WILEY & SONS Chichester
More informationComparing two independent samples
In many applications it is necessary to compare two competing methods (for example, to compare treatment effects of a standard drug and an experimental drug). To compare two methods from statistical point
More informationNetworks: Lectures 9 & 10 Random graphs
Networks: Lectures 9 & 10 Random graphs Heather A Harrington Mathematical Institute University of Oxford HT 2017 What you re in for Week 1: Introduction and basic concepts Week 2: Small worlds Week 3:
More informationIntroduction to Probabilistic Machine Learning
Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning
More informationLecture 1: August 28
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random
More informationSTEIN S METHOD, SEMICIRCLE DISTRIBUTION, AND REDUCED DECOMPOSITIONS OF THE LONGEST ELEMENT IN THE SYMMETRIC GROUP
STIN S MTHOD, SMICIRCL DISTRIBUTION, AND RDUCD DCOMPOSITIONS OF TH LONGST LMNT IN TH SYMMTRIC GROUP JASON FULMAN AND LARRY GOLDSTIN Abstract. Consider a uniformly chosen random reduced decomposition of
More informationDistributed Statistical Estimation and Rates of Convergence in Normal Approximation
Distributed Statistical Estimation and Rates of Convergence in Normal Approximation Stas Minsker (joint with Nate Strawn) Department of Mathematics, USC July 3, 2017 Colloquium on Concentration inequalities,
More informationModel-Free Knockoffs: High-Dimensional Variable Selection that Controls the False Discovery Rate
Model-Free Knockoffs: High-Dimensional Variable Selection that Controls the False Discovery Rate Lucas Janson, Stanford Department of Statistics WADAPT Workshop, NIPS, December 2016 Collaborators: Emmanuel
More information[y i α βx i ] 2 (2) Q = i=1
Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation
More informationStatistical Inference On the High-dimensional Gaussian Covarianc
Statistical Inference On the High-dimensional Gaussian Covariance Matrix Department of Mathematical Sciences, Clemson University June 6, 2011 Outline Introduction Problem Setup Statistical Inference High-Dimensional
More informationQualifying Exam in Probability and Statistics. https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf
Part 1: Sample Problems for the Elementary Section of Qualifying Exam in Probability and Statistics https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part 2: Sample Problems for the Advanced Section
More informationModel Selection and Geometry
Model Selection and Geometry Pascal Massart Université Paris-Sud, Orsay Leipzig, February Purpose of the talk! Concentration of measure plays a fundamental role in the theory of model selection! Model
More information3.0.1 Multivariate version and tensor product of experiments
ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 Lecture 3: Minimax risk of GLM and four extensions Lecturer: Yihong Wu Scribe: Ashok Vardhan, Jan 28, 2016 [Ed. Mar 24]
More informationAdaptive Designs: Why, How and When?
Adaptive Designs: Why, How and When? Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj ISBS Conference Shanghai, July 2008 1 Adaptive designs:
More informationLawrence D. Brown* and Daniel McCarthy*
Comments on the paper, An adaptive resampling test for detecting the presence of significant predictors by I. W. McKeague and M. Qian Lawrence D. Brown* and Daniel McCarthy* ABSTRACT: This commentary deals
More informationA Few Special Distributions and Their Properties
A Few Special Distributions and Their Properties Econ 690 Purdue University Justin L. Tobias (Purdue) Distributional Catalog 1 / 20 Special Distributions and Their Associated Properties 1 Uniform Distribution
More informationPART I INTRODUCTION The meaning of probability Basic definitions for frequentist statistics and Bayesian inference Bayesian inference Combinatorics
Table of Preface page xi PART I INTRODUCTION 1 1 The meaning of probability 3 1.1 Classical definition of probability 3 1.2 Statistical definition of probability 9 1.3 Bayesian understanding of probability
More informationFirst Year Examination Department of Statistics, University of Florida
First Year Examination Department of Statistics, University of Florida August 19, 010, 8:00 am - 1:00 noon Instructions: 1. You have four hours to answer questions in this examination.. You must show your
More informationMultivariate Analysis and Likelihood Inference
Multivariate Analysis and Likelihood Inference Outline 1 Joint Distribution of Random Variables 2 Principal Component Analysis (PCA) 3 Multivariate Normal Distribution 4 Likelihood Inference Joint density
More informationRandom matrices: Distribution of the least singular value (via Property Testing)
Random matrices: Distribution of the least singular value (via Property Testing) Van H. Vu Department of Mathematics Rutgers vanvu@math.rutgers.edu (joint work with T. Tao, UCLA) 1 Let ξ be a real or complex-valued
More informationStatistics 135 Fall 2008 Final Exam
Name: SID: Statistics 135 Fall 2008 Final Exam Show your work. The number of points each question is worth is shown at the beginning of the question. There are 10 problems. 1. [2] The normal equations
More informationThe Design of a Survival Study
The Design of a Survival Study The design of survival studies are usually based on the logrank test, and sometimes assumes the exponential distribution. As in standard designs, the power depends on The
More informationMachine Learning 2017
Machine Learning 2017 Volker Roth Department of Mathematics & Computer Science University of Basel 21st March 2017 Volker Roth (University of Basel) Machine Learning 2017 21st March 2017 1 / 41 Section
More informationEstimating network degree distributions from sampled networks: An inverse problem
Estimating network degree distributions from sampled networks: An inverse problem Eric D. Kolaczyk Dept of Mathematics and Statistics, Boston University kolaczyk@bu.edu Introduction: Networks and Degree
More informationStat 516, Homework 1
Stat 516, Homework 1 Due date: October 7 1. Consider an urn with n distinct balls numbered 1,..., n. We sample balls from the urn with replacement. Let N be the number of draws until we encounter a ball
More informationIndependent Increments in Group Sequential Tests: A Review
Independent Increments in Group Sequential Tests: A Review KyungMann Kim kmkim@biostat.wisc.edu University of Wisconsin-Madison, Madison, WI, USA July 13, 2013 Outline Early Sequential Analysis Independent
More informationFinal Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.
1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. (a) If X and Y are independent, Corr(X, Y ) = 0. (b) (c) (d) (e) A consistent estimator must be asymptotically
More informationPart IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015
Part IB Statistics Theorems with proof Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly)
More informationSTAT 461/561- Assignments, Year 2015
STAT 461/561- Assignments, Year 2015 This is the second set of assignment problems. When you hand in any problem, include the problem itself and its number. pdf are welcome. If so, use large fonts and
More informationA Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression
A Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression Noah Simon Jerome Friedman Trevor Hastie November 5, 013 Abstract In this paper we purpose a blockwise descent
More informationStatistical Methods in HYDROLOGY CHARLES T. HAAN. The Iowa State University Press / Ames
Statistical Methods in HYDROLOGY CHARLES T. HAAN The Iowa State University Press / Ames Univariate BASIC Table of Contents PREFACE xiii ACKNOWLEDGEMENTS xv 1 INTRODUCTION 1 2 PROBABILITY AND PROBABILITY
More informationGeneralized Linear. Mixed Models. Methods and Applications. Modern Concepts, Walter W. Stroup. Texts in Statistical Science.
Texts in Statistical Science Generalized Linear Mixed Models Modern Concepts, Methods and Applications Walter W. Stroup CRC Press Taylor & Francis Croup Boca Raton London New York CRC Press is an imprint
More information