Graduate Econometrics I: Maximum Likelihood II

Size: px
Start display at page:

Download "Graduate Econometrics I: Maximum Likelihood II"

Transcription

1 Graduate Econometrics I: Maximum Likelihood II Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 1/30

2 Outline Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 2/30

3 Outline Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 3/30

4 Principle Assume pairs (y i, x i ) with i = 1,..., n of i.i.d. observations and have a pdf f (y i, x i ; θ) that may be decomposed as : f (y i, x i ; θ) = f (y i x i ; θ)f (x i ; θ), where f (y i x i ; θ) and f (x i ; θ) are the conditional and marginal pdfs, respectively. Definition i) A marginal MLE for θ is a solution ˆθ m,n to the problem : max θ Θ n log f (x i ; θ). ii) A conditional MLE of θ is a solution ˆθ c,n to the problem : max θ Θ i=1 n log f (y i x i ; θ). i=1 Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 4/30

5 Principle We partition θ = (α, β, γ ) as : (α, β ) conditional (β, γ ) marginal. Then : f (y i, x i ; θ) = f (y i x i ; α, β)f (x i ; β, γ). In conditional and marginal models, the notion of cut is important. There is a cut when there are no common parameters. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 5/30

6 Principle Property Suppose that there is a cut so that : f (y i, x i ; θ) = f (y i x i ; α)f (x i ; γ) (α, γ A C). The MLE of θ is ˆθ n = (ˆα c,n, ˆγ m,n). If the parameter of interest is α, we can neglect the marginal distribution. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 6/30

7 Asymptotic Properties We start with some regularity conditions : A1 The ones given previously but adapted to conditional and marginal models. A2 (Y i, X i ) are i.i.d. pairs f (y i x 1,..., x n; θ) = f (y i x i ; θ) f (y i x i ; θ) = f (y j x j ; θ). A3 (X i ) i = 1,..., n are mutually independent. A4 Identification : f (y x; α, β) = f (y x; α 0, β 0 ) x, y { α = α0 β = β 0, f (x; β, γ) = f (x; β 0, γ 0 ) x { γ = γ0 β = β 0. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 7/30

8 Asymptotic Properties Notation : E x( ) = X f (x; β, γ)dx E( x) = θ m = (β, γ ) θ c = (α, β ) Y f (y x; α, β)dy Property Marginal MLE. Under the regularity conditions i) ˆβ m,n and ˆγ m,n exist asymptotically. ii) ˆβ m,n and ˆγ m,n are consistent for β 0 and γ 0. iii) ˆβ m,n and ˆγ m,n are asymptotically normal with : (( ) ( )) ˆβm,n β0 n d N(0, Ĩm 1 (θ ˆγ m,n γ 0 )), 0 ( ) where Ĩ m(θ 0 ) = E x 2 log f (x;β 0,γ 0 ). θ m θ m Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 8/30

9 Asymptotic Properties Property Conditional MLE. Under the regularity conditions, ˆα c,n and ˆβ c,n i) ˆα c,n and ˆβ c,n exist asymptotically. ii) ˆα c,n and ˆβ c,n are consistent for α 0 and β 0. iii) ˆα c,n and ˆβ c,n are asymptotically normal with : ) ( )) ((ˆαc,n α0 n ˆβ d N(0, Ĩc 1 (θ 0 )), c,n ( where Ĩ c(θ 0 ) = E xe β 0 2 log f (y x;α 0,β 0 ) θ c θ c ). Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 9/30

10 Asymptotic Properties Remarks : 1 Ĩ c(θ 0 ) depends on γ 0 since it takes expectation with respect to the true marginal distribution of X. 2 Ĩ m(θ 0 ) and Ĩ c(θ 0 ) are the Fisher information matrices of the marginal and conditional model respectively. They are not I(θ 0 ), the Fisher information matrix of the joint density. Property The estimators ˆα c,n, ˆβ c,n, ˆβ m,n and ˆγ m,n are asymptotically at most as efficient as the MLE ˆα n, ˆβ n and ˆγ n. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 10/30

11 Asymptotic Properties Corollary i) The conditional MLE ˆα c,n is asymptotically as efficient as the ML estimator ˆα n in the following two cases : 1 where there are no β, 2 where ˆα n and ˆβ n are uncorrelated. ii) The marginal MLE ˆγ m,n is asymptotically as efficient as the ML estimator ˆγ n in the following two cases : 1 where there are no β, 2 where ˆβ n and ˆγ n are uncorrelated. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 11/30

12 Asymptotic Properties Finally, Ĩ m(θ 0 ) and Ĩ c(θ 0 ) may be estimated consistently. For example, for Ĩ c(θ 0 ) : or 1 n n i=1 1 n n i=1 2 log l(y i x i ; ˆα c,n, ˆβ c,n) θ c θ c log f (y i x i ; ˆα c,n, ˆβ c,n) log f (y i x i ; ˆα c,n, ˆβ c,n). θ c θ c Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 12/30

13 Outline Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 13/30

14 Principle We now consider the case where the true model is misspecified. The true distribution is : l 0 (y 1,..., y n x 1,..., x n) = n f 0 (y i x i ), where f 0 (y i x i ) does not belong to the specified family, i.e. i=1 f 0 (y i x i ) {f (y x; θ), θ Θ}. θ is going to converge to some value which is not the true but the pseudo-true θ 0. It corresponds to the distribution in the model that is closest to f 0 : and we assume that θ 0 is unique. θ 0 = arg max θ Θ ExE 0 log f (Y X; θ) Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 14/30

15 Principle Definition A Pseudo-MLE (PMLE) ˆθ n of θ is the solution to max θ Θ n log f (Y i X i ; θ). i=1 Regularity conditions : A1 The pairs (Y i, X i ) are i.i.d. A2 Θ is compact. A3 n i=1 log f (y i x i ; θ) is continuous in θ and integrable with respect to the true distribution of (Y i, X i ) for every θ. A4 1 n n i=1 log f (y i x i ; θ) converges uniformly on Θ to E xe 0 log f (Y X; θ). A5 The limit problem max θ Θ E xe 0 log f (Y X; θ) has an unique solution which is θ 0. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 15/30

16 Principle Property Under the assumptions A1-A5, the PMLE ˆθ n converges in probability to θ 0. We now establish the asymptotic normality and the asymptotic variance-covariance function. Regularity conditions : A6 n i=1 log f (y i x i ; θ) is twice continuously differentiable in θ. A7 The matrix ( ) J = E xe 0 2 log f (Y X; θ) θ θ exists and is non-singular. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 16/30

17 Principle Property Under A1-A7, the PMLE is asymptotically normally distributed with : n(ˆθ n θ 0 ) d N(0, J 1 IJ 1 ), where : I = E xe 0 ( log f (Y X; θ 0 ) θ Proof : It relies heavily on the proof of MLE. ) log f (Y X; θ0 ). θ The difference comes from the fact that J and I are NOT equal because the Fisher information matrix cannot be expressed as J. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 17/30

18 Principle In the ML we had : then : l(y ; θ) θ ( ) log l(y ;θ) l(y ; θ) θ θ 2 log l(y ;θ) l(y ; θ) = θ θ Y I = J. = log l(y ; θ) l(y ; θ) θ = 2 log l(y ;θ) l(y ; θ) + θ θ 2 log l(y ;θ) l(y ; θ)dy = θ θ Y log l(y ;θ) log l(y ;θ) l(y ; θ) θ θ log l(y ;θ) θ log l(y ;θ) l(y ;θ) = 0 θ θ log l(y ;θ) l(y ; θ)dy θ Only possible when the assumed density is the true density (MLE). Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 18/30

19 Principle In other words, in MLE : l(y ;θ) dy = log l(y ;θ) l(y ; θ)dy = 0 Y θ Y θ = log l(y ;θ) l Y θ 0 (Y )dy=0 because l(y ; θ) and l 0 (Y ) are asymptotically the same. In PMLE : Y l(y ;θ) dy = log l(y ;θ) l(y ; θ)dy = 0 θ Y θ log l(y ;θ) l Y θ 0 (Y )dy 0 because l(y ; θ) and l 0 (Y ) are not the same. And since the score has not mean equal to zero, the estimators ˆθ n do not converge to θ 0 but to θ 0. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 19/30

20 Outline Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 20/30

21 The model now focuses on the relationship between X and Y via the conditional mean of Y given X : where E x(u i X 1,..., X n) = 0. Y i = m(x i, b 0 ) + u i, Some form of the density has to be proposed. The aim is to find those densities that lead to consistent estimators of b 0 even if they are misspecified. More precisely, we consider a family of densities parametrized by a (conditional) mean m : f (y, m) m M, where M contains all possible values for m(x, b). Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 21/30

22 The conditional distribution of Y i given X i is given by f (y i ; m(x i, b)) and the PMLE for ˆb n is the solution to : max b n log f (y i ; m(x i, b)). i=1 We know that if the model is misspecified, ˆb n converges to a pseudo true value b o, which is the solution of the limit problem : max E xe 0 log f (Y ; m(x, b)), b but in some cases, b 0 = b 0 and hence ˆb n is consistent. The model is assumed to be first order (i.e. mean) identified and differentiable with respect to b. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 22/30

23 Property Under A1-A5, ˆb n is consistent for b 0 for any possible ˆb n, any functional form of m( ) and any conditional distribution, satisfying : Y i = m(x i, b 0 ) + u i, E x(u i X 1,..., X n) = 0, if the pseudo true densities are of the form f (y; m) = exp{a(m) + B(y) + C(m)y}, where m is the mean of f (y; m). Such density is called linear exponential density (LEF). Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 23/30

24 Proof : The PMLE corresponds to the LEF And the limit problem is i.e. or f (y; m) = exp{a(m) + B(y) + C(m)y}. max E xe 0 log f (Y ; m(x, b)), b max E xe 0 [A(m(X, b)) + B(y) + C(m(X, b)y )] b is the solution to this problem is b 0. max E xe 0 [A(m(X, b)) + C(m(X, b)y )] b Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 24/30

25 The Kullback inequality implies : log f (y, m(x, b 0 ))f (y, m(x, b 0 ))dy Y log f (y, m(x, b))f (y, m(x, b 0 ))dy. Y If we denote : D 1 = log f (y, m(x, b))f (y, m(x, b 0 ))dy Y D 2 = log f (y, m(x, b 0 ))f (y, m(x, b 0 ))dy. Y Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 25/30

26 Then D 1 {A(m(X, b)) + B(y) + C(m(X, b))y}f (y, m(x, b 0 ))dy Y = A(m(X, b)) f (y; m(x, b 0 ))dy + Y B(y)f (y; m(x, b 0 ))dy + C(m(X, b)) f (y; m(x, b 0 ))ydy. Y Denoting W = Y B(y)f (y; m(x, b 0))dy and by noting that Y f (y; m(x, b 0))dy = 1 and that Y f (y; m(x, b 0))ydy = m(x, b 0 ), we have : =A(m(X, b)) + W + C(m(X, b))m(x, b 0 ). Y Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 26/30

27 D 2 D 2 = A(m(X, b 0 )) + W + C(m(X, b 0 ))m(x, b 0 ). Therefore the Kullback inequality reduces to : A(m(X, b 0 )) + C(m(X, b 0 ))m(x, b 0 ) A(m(X, b)) + C(m(X, b))m(x, b 0 ), and since b is first order identifiable and b 0 is the PMLE, it follows that the equality is only possible if : m(x, b) = m(x, b 0 ) b = b 0 b 0 = b 0 ˆb n = b 0. And hence b 0 is the value that minimizes the Kullback-Leiber distance. Or it is the value that maximizes the psuedo-likelihood. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 27/30

28 Remarks : 1 2 C(m(X,b)) m = Σ 1 where Σ is the variance-covariance matrix associated with the pseudo true distribution. A(m(X,b)) m + C(m(X,b)) m(x, b) = 0 is the FOC. m 3 Observations need not have the same support as that of the pseudo true distribution. For instance, a pseudo true family of Poisson can be used when the variables y are not positive integer-valued. 4 However, it is crucial to take into account the constraint m(x, b) M. For instance, the Poisson distribution can be used only if the conditional mean m(x, b) is positive valued. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 28/30

29 Property Under A1-A7 a consistent PMLE associated with a LEF is asymptotically normally distributed with n(ˆbn b 0 ) d N(0, J 1 IJ 1 ), where : ( ) m I = E x b Σ 1 ΩΣ 1 m b ( ) m m J = E x b Σ 1. b Ω is the true variance covariance matrix of Y conditional on X, Σ is the conditional variance covariance matrix associated with the chosen pseudo-likelihood. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 29/30

30 Property The variance covariance matrix of a PMLE is at least as large as : K 1 = ( )) m 1 m (E x b Ω 1. b The existence of this lower bound raises the question of whether it is possible to find a consistent estimator of which the asymptotic variance covariance attains the lower bound. This estimator would be asymptotically better than any other consistent PMLE. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 30/30

Graduate Econometrics I: Maximum Likelihood I

Graduate Econometrics I: Maximum Likelihood I Graduate Econometrics I: Maximum Likelihood I Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Maximum Likelihood

More information

Graduate Econometrics I: Unbiased Estimation

Graduate Econometrics I: Unbiased Estimation Graduate Econometrics I: Unbiased Estimation Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Unbiased Estimation

More information

Graduate Econometrics I: Asymptotic Theory

Graduate Econometrics I: Asymptotic Theory Graduate Econometrics I: Asymptotic Theory Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Asymptotic Theory

More information

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score

More information

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. (a) If X and Y are independent, Corr(X, Y ) = 0. (b) (c) (d) (e) A consistent estimator must be asymptotically

More information

Introduction to Estimation Methods for Time Series models Lecture 2

Introduction to Estimation Methods for Time Series models Lecture 2 Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:

More information

Graduate Econometrics I: What is econometrics?

Graduate Econometrics I: What is econometrics? Graduate Econometrics I: What is econometrics? Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: What is econometrics?

More information

Econometrics I, Estimation

Econometrics I, Estimation Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the

More information

MLE and GMM. Li Zhao, SJTU. Spring, Li Zhao MLE and GMM 1 / 22

MLE and GMM. Li Zhao, SJTU. Spring, Li Zhao MLE and GMM 1 / 22 MLE and GMM Li Zhao, SJTU Spring, 2017 Li Zhao MLE and GMM 1 / 22 Outline 1 MLE 2 GMM 3 Binary Choice Models Li Zhao MLE and GMM 2 / 22 Maximum Likelihood Estimation - Introduction For a linear model y

More information

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators Estimation theory Parametric estimation Properties of estimators Minimum variance estimator Cramer-Rao bound Maximum likelihood estimators Confidence intervals Bayesian estimation 1 Random Variables Let

More information

Chapter 3: Maximum Likelihood Theory

Chapter 3: Maximum Likelihood Theory Chapter 3: Maximum Likelihood Theory Florian Pelgrin HEC September-December, 2010 Florian Pelgrin (HEC) Maximum Likelihood Theory September-December, 2010 1 / 40 1 Introduction Example 2 Maximum likelihood

More information

Theory of Maximum Likelihood Estimation. Konstantin Kashin

Theory of Maximum Likelihood Estimation. Konstantin Kashin Gov 2001 Section 5: Theory of Maximum Likelihood Estimation Konstantin Kashin February 28, 2013 Outline Introduction Likelihood Examples of MLE Variance of MLE Asymptotic Properties What is Statistical

More information

Estimation of Dynamic Regression Models

Estimation of Dynamic Regression Models University of Pavia 2007 Estimation of Dynamic Regression Models Eduardo Rossi University of Pavia Factorization of the density DGP: D t (x t χ t 1, d t ; Ψ) x t represent all the variables in the economy.

More information

Statistics and econometrics

Statistics and econometrics 1 / 36 Slides for the course Statistics and econometrics Part 10: Asymptotic hypothesis testing European University Institute Andrea Ichino September 8, 2014 2 / 36 Outline Why do we need large sample

More information

Chapter 1: A Brief Review of Maximum Likelihood, GMM, and Numerical Tools. Joan Llull. Microeconometrics IDEA PhD Program

Chapter 1: A Brief Review of Maximum Likelihood, GMM, and Numerical Tools. Joan Llull. Microeconometrics IDEA PhD Program Chapter 1: A Brief Review of Maximum Likelihood, GMM, and Numerical Tools Joan Llull Microeconometrics IDEA PhD Program Maximum Likelihood Chapter 1. A Brief Review of Maximum Likelihood, GMM, and Numerical

More information

Problem Selected Scores

Problem Selected Scores Statistics Ph.D. Qualifying Exam: Part II November 20, 2010 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. Problem 1 2 3 4 5 6 7 8 9 10 11 12 Selected

More information

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes Maximum Likelihood Estimation Econometrics II Department of Economics Universidad Carlos III de Madrid Máster Universitario en Desarrollo y Crecimiento Económico Outline 1 3 4 General Approaches to Parameter

More information

Mathematical statistics

Mathematical statistics October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter

More information

Likelihood-based inference with missing data under missing-at-random

Likelihood-based inference with missing data under missing-at-random Likelihood-based inference with missing data under missing-at-random Jae-kwang Kim Joint work with Shu Yang Department of Statistics, Iowa State University May 4, 014 Outline 1. Introduction. Parametric

More information

Chapter 4: Asymptotic Properties of the MLE (Part 2)

Chapter 4: Asymptotic Properties of the MLE (Part 2) Chapter 4: Asymptotic Properties of the MLE (Part 2) Daniel O. Scharfstein 09/24/13 1 / 1 Example Let {(R i, X i ) : i = 1,..., n} be an i.i.d. sample of n random vectors (R, X ). Here R is a response

More information

f(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain

f(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain 0.1. INTRODUCTION 1 0.1 Introduction R. A. Fisher, a pioneer in the development of mathematical statistics, introduced a measure of the amount of information contained in an observaton from f(x θ). Fisher

More information

Topic 12 Overview of Estimation

Topic 12 Overview of Estimation Topic 12 Overview of Estimation Classical Statistics 1 / 9 Outline Introduction Parameter Estimation Classical Statistics Densities and Likelihoods 2 / 9 Introduction In the simplest possible terms, the

More information

i=1 h n (ˆθ n ) = 0. (2)

i=1 h n (ˆθ n ) = 0. (2) Stat 8112 Lecture Notes Unbiased Estimating Equations Charles J. Geyer April 29, 2012 1 Introduction In this handout we generalize the notion of maximum likelihood estimation to solution of unbiased estimating

More information

Statistics Ph.D. Qualifying Exam: Part II November 3, 2001

Statistics Ph.D. Qualifying Exam: Part II November 3, 2001 Statistics Ph.D. Qualifying Exam: Part II November 3, 2001 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. 1 2 3 4 5 6 7 8 9 10 11 12 2. Write your

More information

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Article Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Fei Jin 1,2 and Lung-fei Lee 3, * 1 School of Economics, Shanghai University of Finance and Economics,

More information

Advanced Quantitative Methods: maximum likelihood

Advanced Quantitative Methods: maximum likelihood Advanced Quantitative Methods: Maximum Likelihood University College Dublin 4 March 2014 1 2 3 4 5 6 Outline 1 2 3 4 5 6 of straight lines y = 1 2 x + 2 dy dx = 1 2 of curves y = x 2 4x + 5 of curves y

More information

Chapter 3. Point Estimation. 3.1 Introduction

Chapter 3. Point Estimation. 3.1 Introduction Chapter 3 Point Estimation Let (Ω, A, P θ ), P θ P = {P θ θ Θ}be probability space, X 1, X 2,..., X n : (Ω, A) (IR k, B k ) random variables (X, B X ) sample space γ : Θ IR k measurable function, i.e.

More information

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a

More information

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Put your solution to each problem on a separate sheet of paper. Problem 1. (5166) Assume that two random samples {x i } and {y i } are independently

More information

Modification and Improvement of Empirical Likelihood for Missing Response Problem

Modification and Improvement of Empirical Likelihood for Missing Response Problem UW Biostatistics Working Paper Series 12-30-2010 Modification and Improvement of Empirical Likelihood for Missing Response Problem Kwun Chuen Gary Chan University of Washington - Seattle Campus, kcgchan@u.washington.edu

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation University of Pavia Maximum Likelihood Estimation Eduardo Rossi Likelihood function Choosing parameter values that make what one has observed more likely to occur than any other parameter values do. Assumption(Distribution)

More information

Lecture 10 Maximum Likelihood Asymptotics under Non-standard Conditions: A Heuristic Introduction to Sandwiches

Lecture 10 Maximum Likelihood Asymptotics under Non-standard Conditions: A Heuristic Introduction to Sandwiches University of Illinois Department of Economics Spring 2017 Econ 574 Roger Koenker Lecture 10 Maximum Likelihood Asymptotics under Non-standard Conditions: A Heuristic Introduction to Sandwiches Ref: Huber,

More information

Pattern Recognition. Parameter Estimation of Probability Density Functions

Pattern Recognition. Parameter Estimation of Probability Density Functions Pattern Recognition Parameter Estimation of Probability Density Functions Classification Problem (Review) The classification problem is to assign an arbitrary feature vector x F to one of c classes. The

More information

Recent Advances in the analysis of missing data with non-ignorable missingness

Recent Advances in the analysis of missing data with non-ignorable missingness Recent Advances in the analysis of missing data with non-ignorable missingness Jae-Kwang Kim Department of Statistics, Iowa State University July 4th, 2014 1 Introduction 2 Full likelihood-based ML estimation

More information

Econometrics I. Ricardo Mora

Econometrics I. Ricardo Mora Econometrics I Department of Economics Universidad Carlos III de Madrid Master in Industrial Economics and Markets Outline Motivation 1 Motivation 2 3 4 Motivation The Analogy Principle The () is a framework

More information

Mathematical statistics

Mathematical statistics October 1 st, 2018 Lecture 11: Sufficient statistic Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation

More information

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:. MATHEMATICAL STATISTICS Homework assignment Instructions Please turn in the homework with this cover page. You do not need to edit the solutions. Just make sure the handwriting is legible. You may discuss

More information

DA Freedman Notes on the MLE Fall 2003

DA Freedman Notes on the MLE Fall 2003 DA Freedman Notes on the MLE Fall 2003 The object here is to provide a sketch of the theory of the MLE. Rigorous presentations can be found in the references cited below. Calculus. Let f be a smooth, scalar

More information

Chapter 4: Constrained estimators and tests in the multiple linear regression model (Part III)

Chapter 4: Constrained estimators and tests in the multiple linear regression model (Part III) Chapter 4: Constrained estimators and tests in the multiple linear regression model (Part III) Florian Pelgrin HEC September-December 2010 Florian Pelgrin (HEC) Constrained estimators September-December

More information

Lecture 14 More on structural estimation

Lecture 14 More on structural estimation Lecture 14 More on structural estimation Economics 8379 George Washington University Instructor: Prof. Ben Williams traditional MLE and GMM MLE requires a full specification of a model for the distribution

More information

1 One-way analysis of variance

1 One-way analysis of variance LIST OF FORMULAS (Version from 21. November 2014) STK2120 1 One-way analysis of variance Assume X ij = µ+α i +ɛ ij ; j = 1, 2,..., J i ; i = 1, 2,..., I ; where ɛ ij -s are independent and N(0, σ 2 ) distributed.

More information

P n. This is called the law of large numbers but it comes in two forms: Strong and Weak.

P n. This is called the law of large numbers but it comes in two forms: Strong and Weak. Large Sample Theory Large Sample Theory is a name given to the search for approximations to the behaviour of statistical procedures which are derived by computing limits as the sample size, n, tends to

More information

Maximum Likelihood Tests and Quasi-Maximum-Likelihood

Maximum Likelihood Tests and Quasi-Maximum-Likelihood Maximum Likelihood Tests and Quasi-Maximum-Likelihood Wendelin Schnedler Department of Economics University of Heidelberg 10. Dezember 2007 Wendelin Schnedler (AWI) Maximum Likelihood Tests and Quasi-Maximum-Likelihood10.

More information

GARCH Models Estimation and Inference

GARCH Models Estimation and Inference GARCH Models Estimation and Inference Eduardo Rossi University of Pavia December 013 Rossi GARCH Financial Econometrics - 013 1 / 1 Likelihood function The procedure most often used in estimating θ 0 in

More information

Introduction Large Sample Testing Composite Hypotheses. Hypothesis Testing. Daniel Schmierer Econ 312. March 30, 2007

Introduction Large Sample Testing Composite Hypotheses. Hypothesis Testing. Daniel Schmierer Econ 312. March 30, 2007 Hypothesis Testing Daniel Schmierer Econ 312 March 30, 2007 Basics Parameter of interest: θ Θ Structure of the test: H 0 : θ Θ 0 H 1 : θ Θ 1 for some sets Θ 0, Θ 1 Θ where Θ 0 Θ 1 = (often Θ 1 = Θ Θ 0

More information

2 Statistical Estimation: Basic Concepts

2 Statistical Estimation: Basic Concepts Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:

More information

Theory of Statistics.

Theory of Statistics. Theory of Statistics. Homework V February 5, 00. MT 8.7.c When σ is known, ˆµ = X is an unbiased estimator for µ. If you can show that its variance attains the Cramer-Rao lower bound, then no other unbiased

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth

More information

Joint work with Nottingham colleagues Simon Preston and Michail Tsagris.

Joint work with Nottingham colleagues Simon Preston and Michail Tsagris. /pgf/stepx/.initial=1cm, /pgf/stepy/.initial=1cm, /pgf/step/.code=1/pgf/stepx/.expanded=- 10.95415pt,/pgf/stepy/.expanded=- 10.95415pt, /pgf/step/.value required /pgf/images/width/.estore in= /pgf/images/height/.estore

More information

Gauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA

Gauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA JAPANESE BEETLE DATA 6 MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA Gauge Plots TuscaroraLisa Central Madsen Fairways, 996 January 9, 7 Grubs Adult Activity Grub Counts 6 8 Organic Matter

More information

Master s Written Examination - Solution

Master s Written Examination - Solution Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2

More information

Mathematics Ph.D. Qualifying Examination Stat Probability, January 2018

Mathematics Ph.D. Qualifying Examination Stat Probability, January 2018 Mathematics Ph.D. Qualifying Examination Stat 52800 Probability, January 2018 NOTE: Answers all questions completely. Justify every step. Time allowed: 3 hours. 1. Let X 1,..., X n be a random sample from

More information

(θ θ ), θ θ = 2 L(θ ) θ θ θ θ θ (θ )= H θθ (θ ) 1 d θ (θ )

(θ θ ), θ θ = 2 L(θ ) θ θ θ θ θ (θ )= H θθ (θ ) 1 d θ (θ ) Setting RHS to be zero, 0= (θ )+ 2 L(θ ) (θ θ ), θ θ = 2 L(θ ) 1 (θ )= H θθ (θ ) 1 d θ (θ ) O =0 θ 1 θ 3 θ 2 θ Figure 1: The Newton-Raphson Algorithm where H is the Hessian matrix, d θ is the derivative

More information

ECE 275A Homework 7 Solutions

ECE 275A Homework 7 Solutions ECE 275A Homework 7 Solutions Solutions 1. For the same specification as in Homework Problem 6.11 we want to determine an estimator for θ using the Method of Moments (MOM). In general, the MOM estimator

More information

ECE531 Lecture 10b: Maximum Likelihood Estimation

ECE531 Lecture 10b: Maximum Likelihood Estimation ECE531 Lecture 10b: Maximum Likelihood Estimation D. Richard Brown III Worcester Polytechnic Institute 05-Apr-2011 Worcester Polytechnic Institute D. Richard Brown III 05-Apr-2011 1 / 23 Introduction So

More information

Weak convergence of Markov chain Monte Carlo II

Weak convergence of Markov chain Monte Carlo II Weak convergence of Markov chain Monte Carlo II KAMATANI, Kengo Mar 2011 at Le Mans Background Markov chain Monte Carlo (MCMC) method is widely used in Statistical Science. It is easy to use, but difficult

More information

Lecture 1: Introduction

Lecture 1: Introduction Principles of Statistics Part II - Michaelmas 208 Lecturer: Quentin Berthet Lecture : Introduction This course is concerned with presenting some of the mathematical principles of statistical theory. One

More information

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function

More information

1. (Regular) Exponential Family

1. (Regular) Exponential Family 1. (Regular) Exponential Family The density function of a regular exponential family is: [ ] Example. Poisson(θ) [ ] Example. Normal. (both unknown). ) [ ] [ ] [ ] [ ] 2. Theorem (Exponential family &

More information

Exercises Chapter 4 Statistical Hypothesis Testing

Exercises Chapter 4 Statistical Hypothesis Testing Exercises Chapter 4 Statistical Hypothesis Testing Advanced Econometrics - HEC Lausanne Christophe Hurlin University of Orléans December 5, 013 Christophe Hurlin (University of Orléans) Advanced Econometrics

More information

Classical Estimation Topics

Classical Estimation Topics Classical Estimation Topics Namrata Vaswani, Iowa State University February 25, 2014 This note fills in the gaps in the notes already provided (l0.pdf, l1.pdf, l2.pdf, l3.pdf, LeastSquares.pdf). 1 Min

More information

Estimation of Threshold Cointegration

Estimation of Threshold Cointegration Estimation of Myung Hwan London School of Economics December 2006 Outline Model Asymptotics Inference Conclusion 1 Model Estimation Methods Literature 2 Asymptotics Consistency Convergence Rates Asymptotic

More information

PROBABILITY AND INFORMATION THEORY. Dr. Gjergji Kasneci Introduction to Information Retrieval WS

PROBABILITY AND INFORMATION THEORY. Dr. Gjergji Kasneci Introduction to Information Retrieval WS PROBABILITY AND INFORMATION THEORY Dr. Gjergji Kasneci Introduction to Information Retrieval WS 2012-13 1 Outline Intro Basics of probability and information theory Probability space Rules of probability

More information

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation Yujin Chung November 29th, 2016 Fall 2016 Yujin Chung Lec13: MLE Fall 2016 1/24 Previous Parametric tests Mean comparisons (normality assumption)

More information

Posterior Regularization

Posterior Regularization Posterior Regularization 1 Introduction One of the key challenges in probabilistic structured learning, is the intractability of the posterior distribution, for fast inference. There are numerous methods

More information

For iid Y i the stronger conclusion holds; for our heuristics ignore differences between these notions.

For iid Y i the stronger conclusion holds; for our heuristics ignore differences between these notions. Large Sample Theory Study approximate behaviour of ˆθ by studying the function U. Notice U is sum of independent random variables. Theorem: If Y 1, Y 2,... are iid with mean µ then Yi n µ Called law of

More information

STAT 135 Lab 3 Asymptotic MLE and the Method of Moments

STAT 135 Lab 3 Asymptotic MLE and the Method of Moments STAT 135 Lab 3 Asymptotic MLE and the Method of Moments Rebecca Barter February 9, 2015 Maximum likelihood estimation (a reminder) Maximum likelihood estimation Suppose that we have a sample, X 1, X 2,...,

More information

INFORMATION THEORY AND STATISTICS

INFORMATION THEORY AND STATISTICS INFORMATION THEORY AND STATISTICS Solomon Kullback DOVER PUBLICATIONS, INC. Mineola, New York Contents 1 DEFINITION OF INFORMATION 1 Introduction 1 2 Definition 3 3 Divergence 6 4 Examples 7 5 Problems...''.

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Maximum Likelihood Estimation Assume X P θ, θ Θ, with joint pdf (or pmf) f(x θ). Suppose we observe X = x. The Likelihood function is L(θ x) = f(x θ) as a function of θ (with the data x held fixed). The

More information

Brief Review on Estimation Theory

Brief Review on Estimation Theory Brief Review on Estimation Theory K. Abed-Meraim ENST PARIS, Signal and Image Processing Dept. abed@tsi.enst.fr This presentation is essentially based on the course BASTA by E. Moulines Brief review on

More information

EM Algorithm II. September 11, 2018

EM Algorithm II. September 11, 2018 EM Algorithm II September 11, 2018 Review EM 1/27 (Y obs, Y mis ) f (y obs, y mis θ), we observe Y obs but not Y mis Complete-data log likelihood: l C (θ Y obs, Y mis ) = log { f (Y obs, Y mis θ) Observed-data

More information

Nonlinear Error Correction Model and Multiple-Threshold Cointegration May 23, / 31

Nonlinear Error Correction Model and Multiple-Threshold Cointegration May 23, / 31 Nonlinear Error Correction Model and Multiple-Threshold Cointegration Man Wang Dong Hua University, China Joint work with N.H.Chan May 23, 2014 Nonlinear Error Correction Model and Multiple-Threshold Cointegration

More information

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari MS&E 226: Small Data Lecture 11: Maximum likelihood (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 18 The likelihood function 2 / 18 Estimating the parameter This lecture develops the methodology behind

More information

Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory

Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory Statistical Inference Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory IP, José Bioucas Dias, IST, 2007

More information

The properties of L p -GMM estimators

The properties of L p -GMM estimators The properties of L p -GMM estimators Robert de Jong and Chirok Han Michigan State University February 2000 Abstract This paper considers Generalized Method of Moment-type estimators for which a criterion

More information

Statistics Ph.D. Qualifying Exam

Statistics Ph.D. Qualifying Exam Department of Statistics Carnegie Mellon University May 7 2008 Statistics Ph.D. Qualifying Exam You are not expected to solve all five problems. Complete solutions to few problems will be preferred to

More information

A General Overview of Parametric Estimation and Inference Techniques.

A General Overview of Parametric Estimation and Inference Techniques. A General Overview of Parametric Estimation and Inference Techniques. Moulinath Banerjee University of Michigan September 11, 2012 The object of statistical inference is to glean information about an underlying

More information

Information in a Two-Stage Adaptive Optimal Design

Information in a Two-Stage Adaptive Optimal Design Information in a Two-Stage Adaptive Optimal Design Department of Statistics, University of Missouri Designed Experiments: Recent Advances in Methods and Applications DEMA 2011 Isaac Newton Institute for

More information

Primal-dual Covariate Balance and Minimal Double Robustness via Entropy Balancing

Primal-dual Covariate Balance and Minimal Double Robustness via Entropy Balancing Primal-dual Covariate Balance and Minimal Double Robustness via (Joint work with Daniel Percival) Department of Statistics, Stanford University JSM, August 9, 2015 Outline 1 2 3 1/18 Setting Rubin s causal

More information

ECON 3150/4150, Spring term Lecture 6

ECON 3150/4150, Spring term Lecture 6 ECON 3150/4150, Spring term 2013. Lecture 6 Review of theoretical statistics for econometric modelling (II) Ragnar Nymoen University of Oslo 31 January 2013 1 / 25 References to Lecture 3 and 6 Lecture

More information

ENEE 621 SPRING 2016 DETECTION AND ESTIMATION THEORY THE PARAMETER ESTIMATION PROBLEM

ENEE 621 SPRING 2016 DETECTION AND ESTIMATION THEORY THE PARAMETER ESTIMATION PROBLEM c 2007-2016 by Armand M. Makowski 1 ENEE 621 SPRING 2016 DETECTION AND ESTIMATION THEORY THE PARAMETER ESTIMATION PROBLEM 1 The basic setting Throughout, p, q and k are positive integers. The setup With

More information

Lecture 3 September 1

Lecture 3 September 1 STAT 383C: Statistical Modeling I Fall 2016 Lecture 3 September 1 Lecturer: Purnamrita Sarkar Scribe: Giorgio Paulon, Carlos Zanini Disclaimer: These scribe notes have been slightly proofread and may have

More information

Semiparametric Gaussian Copula Models: Progress and Problems

Semiparametric Gaussian Copula Models: Progress and Problems Semiparametric Gaussian Copula Models: Progress and Problems Jon A. Wellner University of Washington, Seattle 2015 IMS China, Kunming July 1-4, 2015 2015 IMS China Meeting, Kunming Based on joint work

More information

Comprehensive Examination Quantitative Methods Spring, 2018

Comprehensive Examination Quantitative Methods Spring, 2018 Comprehensive Examination Quantitative Methods Spring, 2018 Instruction: This exam consists of three parts. You are required to answer all the questions in all the parts. 1 Grading policy: 1. Each part

More information

On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models

On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models Thomas Kneib Institute of Statistics and Econometrics Georg-August-University Göttingen Department of Statistics

More information

The Bivariate Probit Model, Maximum Likelihood Estimation, Pseudo True Parameters and Partial Identification

The Bivariate Probit Model, Maximum Likelihood Estimation, Pseudo True Parameters and Partial Identification ISSN 1440-771X Department of Econometrics and Business Statistics http://business.monash.edu/econometrics-and-business-statistics/research/publications The Bivariate Probit Model, Maximum Likelihood Estimation,

More information

The logistic regression model is thus a glm-model with canonical link function so that the log-odds equals the linear predictor, that is

The logistic regression model is thus a glm-model with canonical link function so that the log-odds equals the linear predictor, that is Example The logistic regression model is thus a glm-model with canonical link function so that the log-odds equals the linear predictor, that is log p 1 p = β 0 + β 1 f 1 (y 1 ) +... + β d f d (y d ).

More information

Bayesian estimation of the discrepancy with misspecified parametric models

Bayesian estimation of the discrepancy with misspecified parametric models Bayesian estimation of the discrepancy with misspecified parametric models Pierpaolo De Blasi University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics workshop ICERM, 17-21 September 2012

More information

Multivariate Regression Analysis

Multivariate Regression Analysis Matrices and vectors The model from the sample is: Y = Xβ +u with n individuals, l response variable, k regressors Y is a n 1 vector or a n l matrix with the notation Y T = (y 1,y 2,...,y n ) 1 x 11 x

More information

Stat 5102 Final Exam May 14, 2015

Stat 5102 Final Exam May 14, 2015 Stat 5102 Final Exam May 14, 2015 Name Student ID The exam is closed book and closed notes. You may use three 8 1 11 2 sheets of paper with formulas, etc. You may also use the handouts on brand name distributions

More information

Chapter 8.8.1: A factorization theorem

Chapter 8.8.1: A factorization theorem LECTURE 14 Chapter 8.8.1: A factorization theorem The characterization of a sufficient statistic in terms of the conditional distribution of the data given the statistic can be difficult to work with.

More information

Various types of likelihood

Various types of likelihood Various types of likelihood 1. likelihood, marginal likelihood, conditional likelihood, profile likelihood, adjusted profile likelihood 2. semi-parametric likelihood, partial likelihood 3. empirical likelihood,

More information

Quick Review on Linear Multiple Regression

Quick Review on Linear Multiple Regression Quick Review on Linear Multiple Regression Mei-Yuan Chen Department of Finance National Chung Hsing University March 6, 2007 Introduction for Conditional Mean Modeling Suppose random variables Y, X 1,

More information

Calibration Estimation of Semiparametric Copula Models with Data Missing at Random

Calibration Estimation of Semiparametric Copula Models with Data Missing at Random Calibration Estimation of Semiparametric Copula Models with Data Missing at Random Shigeyuki Hamori 1 Kaiji Motegi 1 Zheng Zhang 2 1 Kobe University 2 Renmin University of China Econometrics Workshop UNC

More information

Chapter 3 : Likelihood function and inference

Chapter 3 : Likelihood function and inference Chapter 3 : Likelihood function and inference 4 Likelihood function and inference The likelihood Information and curvature Sufficiency and ancilarity Maximum likelihood estimation Non-regular models EM

More information

Semiparametric Gaussian Copula Models: Progress and Problems

Semiparametric Gaussian Copula Models: Progress and Problems Semiparametric Gaussian Copula Models: Progress and Problems Jon A. Wellner University of Washington, Seattle European Meeting of Statisticians, Amsterdam July 6-10, 2015 EMS Meeting, Amsterdam Based on

More information

Likelihood-Based Methods

Likelihood-Based Methods Likelihood-Based Methods Handbook of Spatial Statistics, Chapter 4 Susheela Singh September 22, 2016 OVERVIEW INTRODUCTION MAXIMUM LIKELIHOOD ESTIMATION (ML) RESTRICTED MAXIMUM LIKELIHOOD ESTIMATION (REML)

More information

This paper is not to be removed from the Examination Halls

This paper is not to be removed from the Examination Halls ~~ST104B ZA d0 This paper is not to be removed from the Examination Halls UNIVERSITY OF LONDON ST104B ZB BSc degrees and Diplomas for Graduates in Economics, Management, Finance and the Social Sciences,

More information

Investigation of an Automated Approach to Threshold Selection for Generalized Pareto

Investigation of an Automated Approach to Threshold Selection for Generalized Pareto Investigation of an Automated Approach to Threshold Selection for Generalized Pareto Kate R. Saunders Supervisors: Peter Taylor & David Karoly University of Melbourne April 8, 2015 Outline 1 Extreme Value

More information

Econ 583 Final Exam Fall 2008

Econ 583 Final Exam Fall 2008 Econ 583 Final Exam Fall 2008 Eric Zivot December 11, 2008 Exam is due at 9:00 am in my office on Friday, December 12. 1 Maximum Likelihood Estimation and Asymptotic Theory Let X 1,...,X n be iid random

More information