Likelihood-based inference with missing data under missing-at-random

Size: px
Start display at page:

Download "Likelihood-based inference with missing data under missing-at-random"

Transcription

1 Likelihood-based inference with missing data under missing-at-random Jae-kwang Kim Joint work with Shu Yang Department of Statistics, Iowa State University May 4, 014

2 Outline 1. Introduction. Parametric fractional imputation (PFI) 3. PFI approximation of the observed likelihood 4. Main theory 5. Simulation study 6. Discussion

3 Introduction Let y = (y 1,, y n ) have a joint density f (y; θ). Instead of observing y, we observe y obs, where y = (y obs, y mis ). We are interested in making inference about θ in the presence of missing data. Maximum likelihood estimator ˆθ maximizes the observed log-likelihood l obs (θ) = log f (y; θ)dy mis.

4 Introduction Inference with missing data is usually based on Wald type inference: ˆθ N(θ, Î 1 obs ), where ˆl obs = l obs (θ)/ θ evaluated at θ = ˆθ. We are interested in inference based on Wilks theorem: { } l obs (θ) l obs (ˆθ) χ p, where l obs (θ) is the observed log likelihood and p is the dimension of θ.

5 Basic setup The observed likelihood involves integration: l obs (θ) = log f (y; θ)dy mis = log f obs (y obs ; θ). Monte Carlo approximation is valid in the neighborhood of θ = ˆθ: { } f (y; θ) f obs (y obs ; θ) = E f (y mis y obs ; θ) y obs where y (j) i = 1 M f (y mis y obs ; ˆθ), M j=1 f (y obs, y (j) mis ; θ) f (y (j) mis y obs ; θ)

6 Basic setup Wilk Inference is made using the likelihood ratio test. If l obs (θ) is known, Wilk C.I. can be constructed as { } θ Θ : {l obs (θ) l obs (ˆθ)} χ p(1 α). Computing the observed log likelihood is challenging because the integration over the random variable y i,mis is often intractable. Monte Carlo methods for approximating the observed data likelihood using samples from a distribution near the MLE will fail to compute l obs (θ) correctly when θ is far from the MLE.

7 Parametric fractional imputation (PFI) Main idea of EM algorithm of PFI (Kim, 011) PFI saves the computation associated with MCEM. It uses the importance sampling idea to compute the mean score function where the imputed values are generated only once by importance sampling at the beginning of the EM iteration. PFI doesn t change the imputed values but changes the fractional weights for each EM iteration. It largely reduces the computation burden and ensures the convergence of the EM sequence.

8 Parametric fractional imputation (PFI) (I-step) For missing y i,mis, y (j) i,mis h(y i,mis), for j = 1,..., m. (W-step) Given the current parameter value θ (t), compute S (t) (θ) n m i=1 j=1 w ij (θ (t) )S(θ; y ij ) = 0, (1) where S(θ; y) is the score function of θ, w ij (θ) = and y ij = (y i,obs, y (j) i,mis ). m k=1 f (yij (j) ; θ)/h(yi,mis ) {f (y (k) ik ; θ)/h(yi,mis )}, (M-step) Update parameter ˆθ (t+1) by solving (1). Repeat (W-step) and (M-step) until convergence.

9 Remarks on PFI method (I-step) + (W-step) = E-step Key part is to compute the fractional weights w ij (θ) = m k=1 f (yij (j) ; θ)/h(yi,mis ) {f (y (k) ik ; θ)/h(yi,mis )}. Here, h is often called proposal distribution and f is called target distribution.

10 Parametric fractional imputation (PFI) Wald Inference can be constructed based on ( ) ˆθ N θ, Î 1 obs, where I obs is the approximate information matrix. Issue with Wald inference Wald-type confidence intervals often have poor coverage when the sampling distribution of the MLE is skewed. In such cases, Wilk-type inference is preferred.

11 PFI approximation of likelihood Using the PFI data, we can express m f obs,i (y i,obs ; θ) j=1 {f (y (j) ij ; θ)/h(y = (j) {1/h(y i=1 = m j=1 i,mis )} i,mis )} 1 m j=1 {w ij (θ)/f (y ij ; θ)}. The observed log-likelihood function l obs (θ) can be approximated by n m lobs (θ) = { log w ij (θ)/f (yij ; θ) }. Given the imputed values y ij j=1, only need f (y ij ; θ) and h(y ij ).

12 Main Theory 1 Theorem The imputed observed likelihood ratio statistic for testing H 0 : θ = θ 0 is, { } W 1 = lobs (θ 0) lobs (ˆθ). Under the regularity conditions specified, under the null hypothesis H 0, W 1 χ (p), as m and n.

13 Main Theory Theorem Let θ = (θ 1, θ ), where θ 1 and θ are q 1 and (p q) 1 vectors, respectively. Under the same regularity conditions of Theorem 1, under H 0 : θ 1 = θ 1,0, { } W = lobs (ˆθ (0) ) lobs (ˆθ) χ (q), as m and n, where ˆθ (0) = arg max H0 l obs (θ).

14 Remarks on Theorem There are two models nvolved in Theorem. One is the full model f and the other is the reduced model f 0, the model under null hypothesis. Recall that n m lobs (θ) = log w ij (θ) = i=1 m k=1 j=1 { w ij (θ)/f (y ij ; θ) }. f (yij (j) ; θ)/h(yi,mis ) {f (y (k) ik ; θ)/h(yi,mis )}. Thus, h remains the same but f needs to be changed to f 0 under the reduced model.

15 An example of likelihood ratio test Consider the following bivariate normal distribution ( ) [( ) ( y µ1 σ N, 1 ρσ 1 σ y µ ρσ 1 σ σ )] for i = 1,..., n. We are interested in testing H 0 : µ 1 = µ. Table: Data structure y 1 y Imputation Proposal h( ) H K X {(y (j) L X {(y (j) M X X {(y (j), y (j) h K (y y 1 ) h L (y 1 y ), wij )}m j=1 h M (y 1, y )

16 An example of likelihood ratio test Consider the following bivariate normal distribution ( ) [( ) ( y µ1 σ N, 1 ρσ 1 σ y µ ρσ 1 σ σ )] for i = 1,..., n. We are interested in testing H 0 : µ 1 = µ. Table: Data structure y 1 y Imputation Proposal h( ) H K X {(y (j) L X {(y (j) M X X {(y (j), y (j) h K (y y 1 ) h L (y 1 y ), wij )}m j=1 h M (y 1, y )

17 An example of likelihood ratio test Consider the following bivariate normal distribution ( ) [( ) ( y µ1 σ N, 1 ρσ 1 σ y µ ρσ 1 σ σ )] for i = 1,..., n. We are interested in testing H 0 : µ 1 = µ. Table: Data structure y 1 y Imputation Proposal h( ) H K X {(y (j) L X {(y (j) M X X {(y (j), y (j) h K (y y 1 ) h L (y 1 y ), wij )}m j=1 h M (y 1, y )

18 An example of likelihood ratio test Consider the following bivariate normal distribution ( ) [( ) ( y µ1 σ N, 1 ρσ 1 σ y µ ρσ 1 σ σ )] for i = 1,..., n. We are interested in testing H 0 : µ 1 = µ. Table: Data structure y 1 y Imputation Proposal h( ) H K X {(y (j) L X {(y (j) M X X {(y (j), y (j) h K (y y 1 ) h L (y 1 y ), wij )}m j=1 h M (y 1, y )

19 An example of likelihood ratio test Consider the following bivariate normal distribution ( ) [( ) ( y µ1 σ N, 1 ρσ 1 σ y µ ρσ 1 σ σ )] for i = 1,..., n. We are interested in testing H 0 : µ 1 = µ. Table: Data structure y 1 y Imputation Proposal h( ) H K X {(y (j) L X {(y (j) M X X {(y (j), y (j) h K (y y 1 ) h L (y 1 y ), wij )}m j=1 h M (y 1, y )

20 An example of likelihood ratio test Consider the following bivariate normal distribution ( ) [( ) ( y µ1 σ N, 1 ρσ 1 σ y µ ρσ 1 σ σ )] for i = 1,..., n. We are interested in testing H 0 : µ 1 = µ. Table: Data structure y 1 y Imputation Proposal h( ) H K X {(y (j) L X {(y (j) M X X {(y (j), y (j) h K (y y 1 ) h L (y 1 y ), wij )}m j=1 h M (y 1, y )

21 An example of likelihood ratio test Consider the following bivariate normal distribution ( ) [( ) ( y µ1 σ N, 1 ρσ 1 σ y µ ρσ 1 σ σ )] for i = 1,..., n. We are interested in testing H 0 : µ 1 = µ. Table: Data structure y 1 y Imputation Proposal h( ) H K X {(y (j) L X {(y (j) M X X {(y (j) h K (y y 1 ) h L (y 1 y ), y (j), wij )}m j=1 h M (y 1, y )

22 An example of likelihood ratio test Consider the following bivariate normal distribution ( ) [( ) ( y µ1 σ N, 1 ρσ 1 σ y µ ρσ 1 σ σ )] for i = 1,..., n. We are interested in testing H 0 : µ 1 = µ. Table: Data structure y 1 y Imputation Proposal h( ) H K X {(y (j) L X {(y (j) M X X {(y (j) h K (y y 1 ) h L (y 1 y ), y (j), wij )}m j=1 h M (y 1, y )

23 An example of likelihood ratio test Consider the following bivariate normal distribution ( ) [( ) ( y µ1 σ N, 1 ρσ 1 σ y µ ρσ 1 σ σ )] for i = 1,..., n. We are interested in testing H 0 : µ 1 = µ. Table: Data structure y 1 y Imputation Proposal h( ) H K X {(y (j) L X {(y (j) M X X {(y (j) h K (y y 1 ) h L (y 1 y ), y (j), wij )}m j=1 h M (y 1, y )

24 An example of likelihood ratio test Consider the following bivariate normal distribution ( ) [( ) ( y µ1 σ N, 1 ρσ 1 σ y µ ρσ 1 σ σ )] for i = 1,..., n. We are interested in testing H 0 : µ 1 = µ. Table: Data structure y 1 y Imputation Proposal h( ) H K X {(y (j) L X {(y (j) M X X {(y (j) h K (y y 1 ) h L (y 1 y ), y (j), wij )}m j=1 h M (y 1, y )

25 An example of likelihood ratio test Consider the following bivariate normal distribution ( ) [( ) ( y µ1 σ N, 1 ρσ 1 σ y µ ρσ 1 σ σ )] for i = 1,..., n. We are interested in testing H 0 : µ 1 = µ. Table: Data structure y 1 y Imputation Proposal h( ) H K X {(y (j) L X {(y (j) M X X {(y (j) h K (y y 1 ) h L (y 1 y ), y (j), wij )}m j=1 h M (y 1, y )

26 An example of likelihood ratio test Under the full model, for i K, the fractional weights are given by w ij (θ) = m k=1 f (y (j) i ; θ)/h K (y (j) y ) {f (y (k) i ; θ)/h K (y (k) y )}. Similar for L and M. The MLE under the full model is computed by solving n wij (θ)s(θ; y (j) i ) = 0. i=1 We may use the EM algorithm to obtain the solution. The maximum of the observed likelihood under the full model n m lobs (ˆθ) = log wij (ˆθ)/f (y (j) i ; ˆθ). i=1 j=1

27 An example of likelihood ratio test Let f 0 be the density for the reduced model under H 0 : µ 1 = µ = µ. For i K, the fractional weights are given by w0,ij(θ) f 0 (y (j) i ; θ)/h K (y (j) y ) = m k=1 {f 0(y (k) i ; θ)/h K (y (k) y )}. Similar for L and M. Note that we are using the same imputed values for this computation. The MLE of θ under the reduced model, denoted by ˆθ 0, is obtained by solving n w0,ij(θ)s 0 (θ; y (j) i ) = 0, i=1 where S 0 (θ; y i ) is the score function derived from f 0.

28 An example of likelihood ratio test The maximum of the observed likelihood under the null model is given by n m l0,obs (ˆθ 0 ) = log w 0,ij(ˆθ 0 )/f 0 (y (j) i ; ˆθ 0 ). i=1 The test statistic for testing H 0 : µ 1 = µ is computed from the PFI data as { } W = l0,obs (ˆθ (0) ) lobs (ˆθ). j=1 If W > χ 1,1 α, then we reject the null model.

29 Simulation One Profile likelihood confidence interval y i = + x i + e i, x i N(1, 1), e i N(0, 1), where x i is fully observed, and y i is subjected to missing. δ i iid Bernoulli(0.6). Variable y i is observed if δ i = 1 and y i is missing if δ i = 0. Monte Carlo samples were independently generated B =, 000 times. Constructing 95% confidence interval for β 1 and σ. Two methods: Wald method using asymptotic normality Wilk method using the result of Theorem.

30 Simulation One Table: Monte Carlo length and coverage of the Wald and Wilk confidence intervals for β 1. Wald C.I. Wilk C.I. sample size length coverage length coverage n = n = n =

31 Simulation One Table: Monte Carlo length and coverage of the Wald and Wilk confidence intervals for σ. Wald C.I. Wilk C.I. sample size length coverage length coverage n = n = n = When n=0, about 8% of Monte Carlo samples have negative values for σ with the Wald confidence intervals.

32 Simulation One Sampling distribution of ˆβ n=0 Density n=50 n=100 Density Density

33 Simulation One Sampling distribution of ˆσ n=0 Density n=50 n=100 Density Density

34 Simulation Two Likelihood ratio test Samples of size n = 100 and n = 00 are generated from y i = β 0 + β 1 x + β(( x + ) e i,( where x i = (x, x ) N, 0.1 δ i Bernoulli(0.6) )), e i N(0, 1). (β 0, β 1 ) = (, 1), β changes from 0, 0.1, 0., and 0.3. We are interested in testing the null hypothesis H 0 : β = 0 using the likelihood ratio test (LRT) of Fractional Imputation (FI) the LRT of Multiple Imputation (MI) 1 1 Meng, X.L. and Rubin, D.B. (199). Performing Likelihood Ratio Tests with Multiply-Imputed Data Sets, Biometrika, 79,

35 Simulation Two α = 0.05 α = 0.1 Parameter Value LRT.MI LRT.FI LRT.MI LRT.FI β = β = β = β = Table: Monte Carlo power of the likelihood ratio test (LRT) of Multiple Imputation (MI) and Fractional Imputation (FI) for continuous data with sample size n = 100.

36 Simulation Two α = 0.05 α = 0.1 Parameter Value LRT.MI LRT.FI LRT.MI LRT.FI β = β = β = β = Table: Monte Carlo power of the likelihood ratio test (LRT) of Multiple Imputation (MI) and Fractional Imputation (FI) for continuous data with sample size n = 00.

37 Concluding remarks Parametric fractional imputation provides a completed data with fractional weights, which enables to compute the observed log likelihood function: h(y (j) i,mis ) and f (y ij ; θ). LRT from PFI is more powerful than the Wald test based on the central limit theorem and also than the LRT from MI proposed by Meng and Rubin (199). Extension will be a topic of future research: Model selection criteria can be developed, such as AIC or BIC as considered by Ibrahim et al. (008) and Garcia et al. (010).

38 The end

Fractional Imputation in Survey Sampling: A Comparative Review

Fractional Imputation in Survey Sampling: A Comparative Review Fractional Imputation in Survey Sampling: A Comparative Review Shu Yang Jae-Kwang Kim Iowa State University Joint Statistical Meetings, August 2015 Outline Introduction Fractional imputation Features Numerical

More information

Parametric fractional imputation for missing data analysis

Parametric fractional imputation for missing data analysis 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 Biometrika (????),??,?, pp. 1 15 C???? Biometrika Trust Printed in

More information

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score

More information

Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling

Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling Jae-Kwang Kim 1 Iowa State University June 26, 2013 1 Joint work with Shu Yang Introduction 1 Introduction

More information

Statistical Methods for Handling Missing Data

Statistical Methods for Handling Missing Data Statistical Methods for Handling Missing Data Jae-Kwang Kim Department of Statistics, Iowa State University July 5th, 2014 Outline Textbook : Statistical Methods for handling incomplete data by Kim and

More information

A note on multiple imputation for general purpose estimation

A note on multiple imputation for general purpose estimation A note on multiple imputation for general purpose estimation Shu Yang Jae Kwang Kim SSC meeting June 16, 2015 Shu Yang, Jae Kwang Kim Multiple Imputation June 16, 2015 1 / 32 Introduction Basic Setup Assume

More information

6. Fractional Imputation in Survey Sampling

6. Fractional Imputation in Survey Sampling 6. Fractional Imputation in Survey Sampling 1 Introduction Consider a finite population of N units identified by a set of indices U = {1, 2,, N} with N known. Associated with each unit i in the population

More information

MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30

MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30 MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD Copyright c 2012 (Iowa State University) Statistics 511 1 / 30 INFORMATION CRITERIA Akaike s Information criterion is given by AIC = 2l(ˆθ) + 2k, where l(ˆθ)

More information

Chapter 4: Imputation

Chapter 4: Imputation Chapter 4: Imputation Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Basic Theory for imputation 3 Variance estimation after imputation 4 Replication variance estimation

More information

Shu Yang and Jae Kwang Kim. Harvard University and Iowa State University

Shu Yang and Jae Kwang Kim. Harvard University and Iowa State University Statistica Sinica 27 (2017), 000-000 doi:https://doi.org/10.5705/ss.202016.0155 DISCUSSION: DISSECTING MULTIPLE IMPUTATION FROM A MULTI-PHASE INFERENCE PERSPECTIVE: WHAT HAPPENS WHEN GOD S, IMPUTER S AND

More information

Recent Advances in the analysis of missing data with non-ignorable missingness

Recent Advances in the analysis of missing data with non-ignorable missingness Recent Advances in the analysis of missing data with non-ignorable missingness Jae-Kwang Kim Department of Statistics, Iowa State University July 4th, 2014 1 Introduction 2 Full likelihood-based ML estimation

More information

EM Algorithm II. September 11, 2018

EM Algorithm II. September 11, 2018 EM Algorithm II September 11, 2018 Review EM 1/27 (Y obs, Y mis ) f (y obs, y mis θ), we observe Y obs but not Y mis Complete-data log likelihood: l C (θ Y obs, Y mis ) = log { f (Y obs, Y mis θ) Observed-data

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

2.6.3 Generalized likelihood ratio tests

2.6.3 Generalized likelihood ratio tests 26 HYPOTHESIS TESTING 113 263 Generalized likelihood ratio tests When a UMP test does not exist, we usually use a generalized likelihood ratio test to verify H 0 : θ Θ against H 1 : θ Θ\Θ It can be used

More information

Econ 583 Homework 7 Suggested Solutions: Wald, LM and LR based on GMM and MLE

Econ 583 Homework 7 Suggested Solutions: Wald, LM and LR based on GMM and MLE Econ 583 Homework 7 Suggested Solutions: Wald, LM and LR based on GMM and MLE Eric Zivot Winter 013 1 Wald, LR and LM statistics based on generalized method of moments estimation Let 1 be an iid sample

More information

Pubh 8482: Sequential Analysis

Pubh 8482: Sequential Analysis Pubh 8482: Sequential Analysis Joseph S. Koopmeiners Division of Biostatistics University of Minnesota Week 8 P-values When reporting results, we usually report p-values in place of reporting whether or

More information

Linear Models and Estimation by Least Squares

Linear Models and Estimation by Least Squares Linear Models and Estimation by Least Squares Jin-Lung Lin 1 Introduction Causal relation investigation lies in the heart of economics. Effect (Dependent variable) cause (Independent variable) Example:

More information

Outline of GLMs. Definitions

Outline of GLMs. Definitions Outline of GLMs Definitions This is a short outline of GLM details, adapted from the book Nonparametric Regression and Generalized Linear Models, by Green and Silverman. The responses Y i have density

More information

Maximum Likelihood Tests and Quasi-Maximum-Likelihood

Maximum Likelihood Tests and Quasi-Maximum-Likelihood Maximum Likelihood Tests and Quasi-Maximum-Likelihood Wendelin Schnedler Department of Economics University of Heidelberg 10. Dezember 2007 Wendelin Schnedler (AWI) Maximum Likelihood Tests and Quasi-Maximum-Likelihood10.

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random

More information

Fractional imputation method of handling missing data and spatial statistics

Fractional imputation method of handling missing data and spatial statistics Graduate Theses and Dissertations Graduate College 2014 Fractional imputation method of handling missing data and spatial statistics Shu Yang Iowa State University Follow this and additional works at:

More information

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. (a) If X and Y are independent, Corr(X, Y ) = 0. (b) (c) (d) (e) A consistent estimator must be asymptotically

More information

Introduction An approximated EM algorithm Simulation studies Discussion

Introduction An approximated EM algorithm Simulation studies Discussion 1 / 33 An Approximated Expectation-Maximization Algorithm for Analysis of Data with Missing Values Gong Tang Department of Biostatistics, GSPH University of Pittsburgh NISS Workshop on Nonignorable Nonresponse

More information

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing STAT 135 Lab 5 Bootstrapping and Hypothesis Testing Rebecca Barter March 2, 2015 The Bootstrap Bootstrap Suppose that we are interested in estimating a parameter θ from some population with members x 1,...,

More information

Classical Inference for Gaussian Linear Models

Classical Inference for Gaussian Linear Models Classical Inference for Gaussian Linear Models Surya Tokdar Gaussian linear model Data (z i, Y i ), i = 1,, n, dim(z i ) = p Model Y i = z T i Parameters: β R p, σ 2 > 0 Inference needed on η = a T β IID

More information

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation Yujin Chung November 29th, 2016 Fall 2016 Yujin Chung Lec13: MLE Fall 2016 1/24 Previous Parametric tests Mean comparisons (normality assumption)

More information

Statistical Inference

Statistical Inference Statistical Inference Classical and Bayesian Methods Revision Class for Midterm Exam AMS-UCSC Th Feb 9, 2012 Winter 2012. Session 1 (Revision Class) AMS-132/206 Th Feb 9, 2012 1 / 23 Topics Topics We will

More information

Generalized Linear Models

Generalized Linear Models Generalized Linear Models Lecture 3. Hypothesis testing. Goodness of Fit. Model diagnostics GLM (Spring, 2018) Lecture 3 1 / 34 Models Let M(X r ) be a model with design matrix X r (with r columns) r n

More information

Some New Aspects of Dose-Response Models with Applications to Multistage Models Having Parameters on the Boundary

Some New Aspects of Dose-Response Models with Applications to Multistage Models Having Parameters on the Boundary Some New Aspects of Dose-Response Models with Applications to Multistage Models Having Parameters on the Boundary Bimal Sinha Department of Mathematics & Statistics University of Maryland, Baltimore County,

More information

MS&E 226: Small Data

MS&E 226: Small Data MS&E 226: Small Data Lecture 15: Examples of hypothesis tests (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 32 The recipe 2 / 32 The hypothesis testing recipe In this lecture we repeatedly apply the

More information

Two-phase sampling approach to fractional hot deck imputation

Two-phase sampling approach to fractional hot deck imputation Two-phase sampling approach to fractional hot deck imputation Jongho Im 1, Jae-Kwang Kim 1 and Wayne A. Fuller 1 Abstract Hot deck imputation is popular for handling item nonresponse in survey sampling.

More information

Model comparison and selection

Model comparison and selection BS2 Statistical Inference, Lectures 9 and 10, Hilary Term 2008 March 2, 2008 Hypothesis testing Consider two alternative models M 1 = {f (x; θ), θ Θ 1 } and M 2 = {f (x; θ), θ Θ 2 } for a sample (X = x)

More information

Chapter 7. Hypothesis Testing

Chapter 7. Hypothesis Testing Chapter 7. Hypothesis Testing Joonpyo Kim June 24, 2017 Joonpyo Kim Ch7 June 24, 2017 1 / 63 Basic Concepts of Testing Suppose that our interest centers on a random variable X which has density function

More information

Nonresponse weighting adjustment using estimated response probability

Nonresponse weighting adjustment using estimated response probability Nonresponse weighting adjustment using estimated response probability Jae-kwang Kim Yonsei University, Seoul, Korea December 26, 2006 Introduction Nonresponse Unit nonresponse Item nonresponse Basic strategy

More information

Chapter 4: Constrained estimators and tests in the multiple linear regression model (Part III)

Chapter 4: Constrained estimators and tests in the multiple linear regression model (Part III) Chapter 4: Constrained estimators and tests in the multiple linear regression model (Part III) Florian Pelgrin HEC September-December 2010 Florian Pelgrin (HEC) Constrained estimators September-December

More information

Streamlining Missing Data Analysis by Aggregating Multiple Imputations at the Data Level

Streamlining Missing Data Analysis by Aggregating Multiple Imputations at the Data Level Streamlining Missing Data Analysis by Aggregating Multiple Imputations at the Data Level A Monte Carlo Simulation to Test the Tenability of the SuperMatrix Approach Kyle M Lang Quantitative Psychology

More information

Math 494: Mathematical Statistics

Math 494: Mathematical Statistics Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/

More information

An Efficient Estimation Method for Longitudinal Surveys with Monotone Missing Data

An Efficient Estimation Method for Longitudinal Surveys with Monotone Missing Data An Efficient Estimation Method for Longitudinal Surveys with Monotone Missing Data Jae-Kwang Kim 1 Iowa State University June 28, 2012 1 Joint work with Dr. Ming Zhou (when he was a PhD student at ISU)

More information

Graduate Econometrics I: Maximum Likelihood II

Graduate Econometrics I: Maximum Likelihood II Graduate Econometrics I: Maximum Likelihood II Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Maximum Likelihood

More information

parameter space Θ, depending only on X, such that Note: it is not θ that is random, but the set C(X).

parameter space Θ, depending only on X, such that Note: it is not θ that is random, but the set C(X). 4. Interval estimation The goal for interval estimation is to specify the accurary of an estimate. A 1 α confidence set for a parameter θ is a set C(X) in the parameter space Θ, depending only on X, such

More information

DA Freedman Notes on the MLE Fall 2003

DA Freedman Notes on the MLE Fall 2003 DA Freedman Notes on the MLE Fall 2003 The object here is to provide a sketch of the theory of the MLE. Rigorous presentations can be found in the references cited below. Calculus. Let f be a smooth, scalar

More information

Categorical Data Analysis Chapter 3

Categorical Data Analysis Chapter 3 Categorical Data Analysis Chapter 3 The actual coverage probability is usually a bit higher than the nominal level. Confidence intervals for association parameteres Consider the odds ratio in the 2x2 table,

More information

Biostat 2065 Analysis of Incomplete Data

Biostat 2065 Analysis of Incomplete Data Biostat 2065 Analysis of Incomplete Data Gong Tang Dept of Biostatistics University of Pittsburgh October 20, 2005 1. Large-sample inference based on ML Let θ is the MLE, then the large-sample theory implies

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 05 Full points may be obtained for correct answers to eight questions Each numbered question (which may have several parts) is worth

More information

1 One-way analysis of variance

1 One-way analysis of variance LIST OF FORMULAS (Version from 21. November 2014) STK2120 1 One-way analysis of variance Assume X ij = µ+α i +ɛ ij ; j = 1, 2,..., J i ; i = 1, 2,..., I ; where ɛ ij -s are independent and N(0, σ 2 ) distributed.

More information

Composite Hypotheses and Generalized Likelihood Ratio Tests

Composite Hypotheses and Generalized Likelihood Ratio Tests Composite Hypotheses and Generalized Likelihood Ratio Tests Rebecca Willett, 06 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve

More information

Answer Key for STAT 200B HW No. 8

Answer Key for STAT 200B HW No. 8 Answer Key for STAT 200B HW No. 8 May 8, 2007 Problem 3.42 p. 708 The values of Ȳ for x 00, 0, 20, 30 are 5/40, 0, 20/50, and, respectively. From Corollary 3.5 it follows that MLE exists i G is identiable

More information

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015 STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots March 8, 2015 The duality between CI and hypothesis testing The duality between CI and hypothesis

More information

Theory of Maximum Likelihood Estimation. Konstantin Kashin

Theory of Maximum Likelihood Estimation. Konstantin Kashin Gov 2001 Section 5: Theory of Maximum Likelihood Estimation Konstantin Kashin February 28, 2013 Outline Introduction Likelihood Examples of MLE Variance of MLE Asymptotic Properties What is Statistical

More information

Psychology 282 Lecture #4 Outline Inferences in SLR

Psychology 282 Lecture #4 Outline Inferences in SLR Psychology 282 Lecture #4 Outline Inferences in SLR Assumptions To this point we have not had to make any distributional assumptions. Principle of least squares requires no assumptions. Can use correlations

More information

Loglikelihood and Confidence Intervals

Loglikelihood and Confidence Intervals Stat 504, Lecture 2 1 Loglikelihood and Confidence Intervals The loglikelihood function is defined to be the natural logarithm of the likelihood function, l(θ ; x) = log L(θ ; x). For a variety of reasons,

More information

Comparison between conditional and marginal maximum likelihood for a class of item response models

Comparison between conditional and marginal maximum likelihood for a class of item response models (1/24) Comparison between conditional and marginal maximum likelihood for a class of item response models Francesco Bartolucci, University of Perugia (IT) Silvia Bacci, University of Perugia (IT) Claudia

More information

Mixtures of Rasch Models

Mixtures of Rasch Models Mixtures of Rasch Models Hannah Frick, Friedrich Leisch, Achim Zeileis, Carolin Strobl http://www.uibk.ac.at/statistics/ Introduction Rasch model for measuring latent traits Model assumption: Item parameters

More information

Semiparametric Gaussian Copula Models: Progress and Problems

Semiparametric Gaussian Copula Models: Progress and Problems Semiparametric Gaussian Copula Models: Progress and Problems Jon A. Wellner University of Washington, Seattle European Meeting of Statisticians, Amsterdam July 6-10, 2015 EMS Meeting, Amsterdam Based on

More information

Math 494: Mathematical Statistics

Math 494: Mathematical Statistics Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/

More information

R Package glmm: Likelihood-Based Inference for Generalized Linear Mixed Models

R Package glmm: Likelihood-Based Inference for Generalized Linear Mixed Models R Package glmm: Likelihood-Based Inference for Generalized Linear Mixed Models Christina Knudson, Ph.D. University of St. Thomas user!2017 Reviewing the Linear Model The usual linear model assumptions:

More information

Miscellanea A note on multiple imputation under complex sampling

Miscellanea A note on multiple imputation under complex sampling Biometrika (2017), 104, 1,pp. 221 228 doi: 10.1093/biomet/asw058 Printed in Great Britain Advance Access publication 3 January 2017 Miscellanea A note on multiple imputation under complex sampling BY J.

More information

AN EMPIRICAL LIKELIHOOD RATIO TEST FOR NORMALITY

AN EMPIRICAL LIKELIHOOD RATIO TEST FOR NORMALITY Econometrics Working Paper EWP0401 ISSN 1485-6441 Department of Economics AN EMPIRICAL LIKELIHOOD RATIO TEST FOR NORMALITY Lauren Bin Dong & David E. A. Giles Department of Economics, University of Victoria

More information

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari MS&E 226: Small Data Lecture 11: Maximum likelihood (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 18 The likelihood function 2 / 18 Estimating the parameter This lecture develops the methodology behind

More information

Better Bootstrap Confidence Intervals

Better Bootstrap Confidence Intervals by Bradley Efron University of Washington, Department of Statistics April 12, 2012 An example Suppose we wish to make inference on some parameter θ T (F ) (e.g. θ = E F X ), based on data We might suppose

More information

Introduction to Estimation Methods for Time Series models Lecture 2

Introduction to Estimation Methods for Time Series models Lecture 2 Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:

More information

STAT Financial Time Series

STAT Financial Time Series STAT 6104 - Financial Time Series Chapter 4 - Estimation in the time Domain Chun Yip Yau (CUHK) STAT 6104:Financial Time Series 1 / 46 Agenda 1 Introduction 2 Moment Estimates 3 Autoregressive Models (AR

More information

Asymptotic Tests and Likelihood Ratio Tests

Asymptotic Tests and Likelihood Ratio Tests Asymptotic Tests and Likelihood Ratio Tests Dennis D. Cox Department of Statistics Rice University P. O. Box 1892 Houston, Texas 77251 Email: dcox@stat.rice.edu November 21, 2004 0 1 Chapter 6, Section

More information

Data Integration for Big Data Analysis for finite population inference

Data Integration for Big Data Analysis for finite population inference for Big Data Analysis for finite population inference Jae-kwang Kim ISU January 23, 2018 1 / 36 What is big data? 2 / 36 Data do not speak for themselves Knowledge Reproducibility Information Intepretation

More information

Spatial inference. Spatial inference. Accounting for spatial correlation. Multivariate normal distributions

Spatial inference. Spatial inference. Accounting for spatial correlation. Multivariate normal distributions Spatial inference I will start with a simple model, using species diversity data Strong spatial dependence, Î = 0.79 what is the mean diversity? How precise is our estimate? Sampling discussion: The 64

More information

Linear Mixed Models for Longitudinal Data with Nonrandom Dropouts

Linear Mixed Models for Longitudinal Data with Nonrandom Dropouts Journal of Data Science 4(2006), 447-460 Linear Mixed Models for Longitudinal Data with Nonrandom Dropouts Ahmed M. Gad and Noha A. Youssif Cairo University Abstract: Longitudinal studies represent one

More information

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a

More information

Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone Missing Data

Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone Missing Data Journal of Multivariate Analysis 78, 6282 (2001) doi:10.1006jmva.2000.1939, available online at http:www.idealibrary.com on Inferences on a Normal Covariance Matrix and Generalized Variance with Monotone

More information

ST495: Survival Analysis: Hypothesis testing and confidence intervals

ST495: Survival Analysis: Hypothesis testing and confidence intervals ST495: Survival Analysis: Hypothesis testing and confidence intervals Eric B. Laber Department of Statistics, North Carolina State University April 3, 2014 I remember that one fateful day when Coach took

More information

Introduction Large Sample Testing Composite Hypotheses. Hypothesis Testing. Daniel Schmierer Econ 312. March 30, 2007

Introduction Large Sample Testing Composite Hypotheses. Hypothesis Testing. Daniel Schmierer Econ 312. March 30, 2007 Hypothesis Testing Daniel Schmierer Econ 312 March 30, 2007 Basics Parameter of interest: θ Θ Structure of the test: H 0 : θ Θ 0 H 1 : θ Θ 1 for some sets Θ 0, Θ 1 Θ where Θ 0 Θ 1 = (often Θ 1 = Θ Θ 0

More information

(θ θ ), θ θ = 2 L(θ ) θ θ θ θ θ (θ )= H θθ (θ ) 1 d θ (θ )

(θ θ ), θ θ = 2 L(θ ) θ θ θ θ θ (θ )= H θθ (θ ) 1 d θ (θ ) Setting RHS to be zero, 0= (θ )+ 2 L(θ ) (θ θ ), θ θ = 2 L(θ ) 1 (θ )= H θθ (θ ) 1 d θ (θ ) O =0 θ 1 θ 3 θ 2 θ Figure 1: The Newton-Raphson Algorithm where H is the Hessian matrix, d θ is the derivative

More information

Wald s theorem and the Asimov data set

Wald s theorem and the Asimov data set Wald s theorem and the Asimov data set Eilam Gross & Ofer Vitells ATLAS statistics forum, Dec. 009 1 Outline We have previously guessed that the median significance of many toy MC experiments could be

More information

Lecture 14 More on structural estimation

Lecture 14 More on structural estimation Lecture 14 More on structural estimation Economics 8379 George Washington University Instructor: Prof. Ben Williams traditional MLE and GMM MLE requires a full specification of a model for the distribution

More information

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test. Economics 52 Econometrics Professor N.M. Kiefer LECTURE 1: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING NEYMAN-PEARSON LEMMA: Lesson: Good tests are based on the likelihood ratio. The proof is easy in the

More information

Topic 12 Overview of Estimation

Topic 12 Overview of Estimation Topic 12 Overview of Estimation Classical Statistics 1 / 9 Outline Introduction Parameter Estimation Classical Statistics Densities and Likelihoods 2 / 9 Introduction In the simplest possible terms, the

More information

Lecture 3. Inference about multivariate normal distribution

Lecture 3. Inference about multivariate normal distribution Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates

More information

Semiparametric Gaussian Copula Models: Progress and Problems

Semiparametric Gaussian Copula Models: Progress and Problems Semiparametric Gaussian Copula Models: Progress and Problems Jon A. Wellner University of Washington, Seattle 2015 IMS China, Kunming July 1-4, 2015 2015 IMS China Meeting, Kunming Based on joint work

More information

Advanced Quantitative Methods: maximum likelihood

Advanced Quantitative Methods: maximum likelihood Advanced Quantitative Methods: Maximum Likelihood University College Dublin 4 March 2014 1 2 3 4 5 6 Outline 1 2 3 4 5 6 of straight lines y = 1 2 x + 2 dy dx = 1 2 of curves y = x 2 4x + 5 of curves y

More information

Computer Intensive Methods in Mathematical Statistics

Computer Intensive Methods in Mathematical Statistics Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)

More information

Lecture 8 Inequality Testing and Moment Inequality Models

Lecture 8 Inequality Testing and Moment Inequality Models Lecture 8 Inequality Testing and Moment Inequality Models Inequality Testing In the previous lecture, we discussed how to test the nonlinear hypothesis H 0 : h(θ 0 ) 0 when the sample information comes

More information

Some Curiosities Arising in Objective Bayesian Analysis

Some Curiosities Arising in Objective Bayesian Analysis . Some Curiosities Arising in Objective Bayesian Analysis Jim Berger Duke University Statistical and Applied Mathematical Institute Yale University May 15, 2009 1 Three vignettes related to John s work

More information

Testing for a Global Maximum of the Likelihood

Testing for a Global Maximum of the Likelihood Testing for a Global Maximum of the Likelihood Christophe Biernacki When several roots to the likelihood equation exist, the root corresponding to the global maximizer of the likelihood is generally retained

More information

Quality Control Using Inferential Statistics In Weibull Based Reliability Analyses S. F. Duffy 1 and A. Parikh 2

Quality Control Using Inferential Statistics In Weibull Based Reliability Analyses S. F. Duffy 1 and A. Parikh 2 Quality Control Using Inferential Statistics In Weibull Based Reliability Analyses S. F. Duffy 1 and A. Parikh 2 1 Cleveland State University 2 N & R Engineering www.inl.gov ASTM Symposium on Graphite

More information

The Expectation-Maximization Algorithm

The Expectation-Maximization Algorithm 1/29 EM & Latent Variable Models Gaussian Mixture Models EM Theory The Expectation-Maximization Algorithm Mihaela van der Schaar Department of Engineering Science University of Oxford MLE for Latent Variable

More information

Stat 451 Lecture Notes Monte Carlo Integration

Stat 451 Lecture Notes Monte Carlo Integration Stat 451 Lecture Notes 06 12 Monte Carlo Integration Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 6 in Givens & Hoeting, Chapter 23 in Lange, and Chapters 3 4 in Robert & Casella 2 Updated:

More information

Likelihood-Based Methods

Likelihood-Based Methods Likelihood-Based Methods Handbook of Spatial Statistics, Chapter 4 Susheela Singh September 22, 2016 OVERVIEW INTRODUCTION MAXIMUM LIKELIHOOD ESTIMATION (ML) RESTRICTED MAXIMUM LIKELIHOOD ESTIMATION (REML)

More information

Basics of Modern Missing Data Analysis

Basics of Modern Missing Data Analysis Basics of Modern Missing Data Analysis Kyle M. Lang Center for Research Methods and Data Analysis University of Kansas March 8, 2013 Topics to be Covered An introduction to the missing data problem Missing

More information

(a) (3 points) Construct a 95% confidence interval for β 2 in Equation 1.

(a) (3 points) Construct a 95% confidence interval for β 2 in Equation 1. Problem 1 (21 points) An economist runs the regression y i = β 0 + x 1i β 1 + x 2i β 2 + x 3i β 3 + ε i (1) The results are summarized in the following table: Equation 1. Variable Coefficient Std. Error

More information

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout

More information

Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon

Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon Jianqing Fan Department of Statistics Chinese University of Hong Kong AND Department of Statistics

More information

Can we do statistical inference in a non-asymptotic way? 1

Can we do statistical inference in a non-asymptotic way? 1 Can we do statistical inference in a non-asymptotic way? 1 Guang Cheng 2 Statistics@Purdue www.science.purdue.edu/bigdata/ ONR Review Meeting@Duke Oct 11, 2017 1 Acknowledge NSF, ONR and Simons Foundation.

More information

Optimization. The value x is called a maximizer of f and is written argmax X f. g(λx + (1 λ)y) < λg(x) + (1 λ)g(y) 0 < λ < 1; x, y X.

Optimization. The value x is called a maximizer of f and is written argmax X f. g(λx + (1 λ)y) < λg(x) + (1 λ)g(y) 0 < λ < 1; x, y X. Optimization Background: Problem: given a function f(x) defined on X, find x such that f(x ) f(x) for all x X. The value x is called a maximizer of f and is written argmax X f. In general, argmax X f may

More information

Inference on distributions and quantiles using a finite-sample Dirichlet process

Inference on distributions and quantiles using a finite-sample Dirichlet process Dirichlet IDEAL Theory/methods Simulations Inference on distributions and quantiles using a finite-sample Dirichlet process David M. Kaplan University of Missouri Matt Goldman UC San Diego Midwest Econometrics

More information

Sequential Implementation of Monte Carlo Tests with Uniformly Bounded Resampling Risk

Sequential Implementation of Monte Carlo Tests with Uniformly Bounded Resampling Risk Sequential Implementation of Monte Carlo Tests with Uniformly Bounded Resampling Risk Axel Gandy Department of Mathematics Imperial College London a.gandy@imperial.ac.uk user! 2009, Rennes July 8-10, 2009

More information

Empirical Likelihood Inference for Two-Sample Problems

Empirical Likelihood Inference for Two-Sample Problems Empirical Likelihood Inference for Two-Sample Problems by Ying Yan A thesis presented to the University of Waterloo in fulfillment of the thesis requirement for the degree of Master of Mathematics in Statistics

More information

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided Let us first identify some classes of hypotheses. simple versus simple H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided H 0 : θ θ 0 versus H 1 : θ > θ 0. (2) two-sided; null on extremes H 0 : θ θ 1 or

More information

2017 Financial Mathematics Orientation - Statistics

2017 Financial Mathematics Orientation - Statistics 2017 Financial Mathematics Orientation - Statistics Written by Long Wang Edited by Joshua Agterberg August 21, 2018 Contents 1 Preliminaries 5 1.1 Samples and Population............................. 5

More information

Statistics and econometrics

Statistics and econometrics 1 / 36 Slides for the course Statistics and econometrics Part 10: Asymptotic hypothesis testing European University Institute Andrea Ichino September 8, 2014 2 / 36 Outline Why do we need large sample

More information

Parametric Modelling of Over-dispersed Count Data. Part III / MMath (Applied Statistics) 1

Parametric Modelling of Over-dispersed Count Data. Part III / MMath (Applied Statistics) 1 Parametric Modelling of Over-dispersed Count Data Part III / MMath (Applied Statistics) 1 Introduction Poisson regression is the de facto approach for handling count data What happens then when Poisson

More information