Graduate Econometrics I: Maximum Likelihood II
|
|
- Domenic Davis
- 5 years ago
- Views:
Transcription
1 Graduate Econometrics I: Maximum Likelihood II Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 1/30
2 Outline Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 2/30
3 Outline Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 3/30
4 Principle Assume pairs (y i, x i ) with i = 1,..., n of i.i.d. observations and have a pdf f (y i, x i ; θ) that may be decomposed as : f (y i, x i ; θ) = f (y i x i ; θ)f (x i ; θ), where f (y i x i ; θ) and f (x i ; θ) are the conditional and marginal pdfs, respectively. Definition i) A marginal MLE for θ is a solution ˆθ m,n to the problem : max θ Θ n log f (x i ; θ). ii) A conditional MLE of θ is a solution ˆθ c,n to the problem : max θ Θ i=1 n log f (y i x i ; θ). i=1 Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 4/30
5 Principle We partition θ = (α, β, γ ) as : (α, β ) conditional (β, γ ) marginal. Then : f (y i, x i ; θ) = f (y i x i ; α, β)f (x i ; β, γ). In conditional and marginal models, the notion of cut is important. There is a cut when there are no common parameters. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 5/30
6 Principle Property Suppose that there is a cut so that : f (y i, x i ; θ) = f (y i x i ; α)f (x i ; γ) (α, γ A C). The MLE of θ is ˆθ n = (ˆα c,n, ˆγ m,n). If the parameter of interest is α, we can neglect the marginal distribution. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 6/30
7 Asymptotic Properties We start with some regularity conditions : A1 The ones given previously but adapted to conditional and marginal models. A2 (Y i, X i ) are i.i.d. pairs f (y i x 1,..., x n; θ) = f (y i x i ; θ) f (y i x i ; θ) = f (y j x j ; θ). A3 (X i ) i = 1,..., n are mutually independent. A4 Identification : f (y x; α, β) = f (y x; α 0, β 0 ) x, y { α = α0 β = β 0, f (x; β, γ) = f (x; β 0, γ 0 ) x { γ = γ0 β = β 0. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 7/30
8 Asymptotic Properties Notation : E x( ) = X f (x; β, γ)dx E( x) = θ m = (β, γ ) θ c = (α, β ) Y f (y x; α, β)dy Property Marginal MLE. Under the regularity conditions i) ˆβ m,n and ˆγ m,n exist asymptotically. ii) ˆβ m,n and ˆγ m,n are consistent for β 0 and γ 0. iii) ˆβ m,n and ˆγ m,n are asymptotically normal with : (( ) ( )) ˆβm,n β0 n d N(0, Ĩm 1 (θ ˆγ m,n γ 0 )), 0 ( ) where Ĩ m(θ 0 ) = E x 2 log f (x;β 0,γ 0 ). θ m θ m Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 8/30
9 Asymptotic Properties Property Conditional MLE. Under the regularity conditions, ˆα c,n and ˆβ c,n i) ˆα c,n and ˆβ c,n exist asymptotically. ii) ˆα c,n and ˆβ c,n are consistent for α 0 and β 0. iii) ˆα c,n and ˆβ c,n are asymptotically normal with : ) ( )) ((ˆαc,n α0 n ˆβ d N(0, Ĩc 1 (θ 0 )), c,n ( where Ĩ c(θ 0 ) = E xe β 0 2 log f (y x;α 0,β 0 ) θ c θ c ). Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 9/30
10 Asymptotic Properties Remarks : 1 Ĩ c(θ 0 ) depends on γ 0 since it takes expectation with respect to the true marginal distribution of X. 2 Ĩ m(θ 0 ) and Ĩ c(θ 0 ) are the Fisher information matrices of the marginal and conditional model respectively. They are not I(θ 0 ), the Fisher information matrix of the joint density. Property The estimators ˆα c,n, ˆβ c,n, ˆβ m,n and ˆγ m,n are asymptotically at most as efficient as the MLE ˆα n, ˆβ n and ˆγ n. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 10/30
11 Asymptotic Properties Corollary i) The conditional MLE ˆα c,n is asymptotically as efficient as the ML estimator ˆα n in the following two cases : 1 where there are no β, 2 where ˆα n and ˆβ n are uncorrelated. ii) The marginal MLE ˆγ m,n is asymptotically as efficient as the ML estimator ˆγ n in the following two cases : 1 where there are no β, 2 where ˆβ n and ˆγ n are uncorrelated. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 11/30
12 Asymptotic Properties Finally, Ĩ m(θ 0 ) and Ĩ c(θ 0 ) may be estimated consistently. For example, for Ĩ c(θ 0 ) : or 1 n n i=1 1 n n i=1 2 log l(y i x i ; ˆα c,n, ˆβ c,n) θ c θ c log f (y i x i ; ˆα c,n, ˆβ c,n) log f (y i x i ; ˆα c,n, ˆβ c,n). θ c θ c Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 12/30
13 Outline Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 13/30
14 Principle We now consider the case where the true model is misspecified. The true distribution is : l 0 (y 1,..., y n x 1,..., x n) = n f 0 (y i x i ), where f 0 (y i x i ) does not belong to the specified family, i.e. i=1 f 0 (y i x i ) {f (y x; θ), θ Θ}. θ is going to converge to some value which is not the true but the pseudo-true θ 0. It corresponds to the distribution in the model that is closest to f 0 : and we assume that θ 0 is unique. θ 0 = arg max θ Θ ExE 0 log f (Y X; θ) Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 14/30
15 Principle Definition A Pseudo-MLE (PMLE) ˆθ n of θ is the solution to max θ Θ n log f (Y i X i ; θ). i=1 Regularity conditions : A1 The pairs (Y i, X i ) are i.i.d. A2 Θ is compact. A3 n i=1 log f (y i x i ; θ) is continuous in θ and integrable with respect to the true distribution of (Y i, X i ) for every θ. A4 1 n n i=1 log f (y i x i ; θ) converges uniformly on Θ to E xe 0 log f (Y X; θ). A5 The limit problem max θ Θ E xe 0 log f (Y X; θ) has an unique solution which is θ 0. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 15/30
16 Principle Property Under the assumptions A1-A5, the PMLE ˆθ n converges in probability to θ 0. We now establish the asymptotic normality and the asymptotic variance-covariance function. Regularity conditions : A6 n i=1 log f (y i x i ; θ) is twice continuously differentiable in θ. A7 The matrix ( ) J = E xe 0 2 log f (Y X; θ) θ θ exists and is non-singular. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 16/30
17 Principle Property Under A1-A7, the PMLE is asymptotically normally distributed with : n(ˆθ n θ 0 ) d N(0, J 1 IJ 1 ), where : I = E xe 0 ( log f (Y X; θ 0 ) θ Proof : It relies heavily on the proof of MLE. ) log f (Y X; θ0 ). θ The difference comes from the fact that J and I are NOT equal because the Fisher information matrix cannot be expressed as J. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 17/30
18 Principle In the ML we had : then : l(y ; θ) θ ( ) log l(y ;θ) l(y ; θ) θ θ 2 log l(y ;θ) l(y ; θ) = θ θ Y I = J. = log l(y ; θ) l(y ; θ) θ = 2 log l(y ;θ) l(y ; θ) + θ θ 2 log l(y ;θ) l(y ; θ)dy = θ θ Y log l(y ;θ) log l(y ;θ) l(y ; θ) θ θ log l(y ;θ) θ log l(y ;θ) l(y ;θ) = 0 θ θ log l(y ;θ) l(y ; θ)dy θ Only possible when the assumed density is the true density (MLE). Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 18/30
19 Principle In other words, in MLE : l(y ;θ) dy = log l(y ;θ) l(y ; θ)dy = 0 Y θ Y θ = log l(y ;θ) l Y θ 0 (Y )dy=0 because l(y ; θ) and l 0 (Y ) are asymptotically the same. In PMLE : Y l(y ;θ) dy = log l(y ;θ) l(y ; θ)dy = 0 θ Y θ log l(y ;θ) l Y θ 0 (Y )dy 0 because l(y ; θ) and l 0 (Y ) are not the same. And since the score has not mean equal to zero, the estimators ˆθ n do not converge to θ 0 but to θ 0. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 19/30
20 Outline Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 20/30
21 The model now focuses on the relationship between X and Y via the conditional mean of Y given X : where E x(u i X 1,..., X n) = 0. Y i = m(x i, b 0 ) + u i, Some form of the density has to be proposed. The aim is to find those densities that lead to consistent estimators of b 0 even if they are misspecified. More precisely, we consider a family of densities parametrized by a (conditional) mean m : f (y, m) m M, where M contains all possible values for m(x, b). Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 21/30
22 The conditional distribution of Y i given X i is given by f (y i ; m(x i, b)) and the PMLE for ˆb n is the solution to : max b n log f (y i ; m(x i, b)). i=1 We know that if the model is misspecified, ˆb n converges to a pseudo true value b o, which is the solution of the limit problem : max E xe 0 log f (Y ; m(x, b)), b but in some cases, b 0 = b 0 and hence ˆb n is consistent. The model is assumed to be first order (i.e. mean) identified and differentiable with respect to b. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 22/30
23 Property Under A1-A5, ˆb n is consistent for b 0 for any possible ˆb n, any functional form of m( ) and any conditional distribution, satisfying : Y i = m(x i, b 0 ) + u i, E x(u i X 1,..., X n) = 0, if the pseudo true densities are of the form f (y; m) = exp{a(m) + B(y) + C(m)y}, where m is the mean of f (y; m). Such density is called linear exponential density (LEF). Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 23/30
24 Proof : The PMLE corresponds to the LEF And the limit problem is i.e. or f (y; m) = exp{a(m) + B(y) + C(m)y}. max E xe 0 log f (Y ; m(x, b)), b max E xe 0 [A(m(X, b)) + B(y) + C(m(X, b)y )] b is the solution to this problem is b 0. max E xe 0 [A(m(X, b)) + C(m(X, b)y )] b Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 24/30
25 The Kullback inequality implies : log f (y, m(x, b 0 ))f (y, m(x, b 0 ))dy Y log f (y, m(x, b))f (y, m(x, b 0 ))dy. Y If we denote : D 1 = log f (y, m(x, b))f (y, m(x, b 0 ))dy Y D 2 = log f (y, m(x, b 0 ))f (y, m(x, b 0 ))dy. Y Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 25/30
26 Then D 1 {A(m(X, b)) + B(y) + C(m(X, b))y}f (y, m(x, b 0 ))dy Y = A(m(X, b)) f (y; m(x, b 0 ))dy + Y B(y)f (y; m(x, b 0 ))dy + C(m(X, b)) f (y; m(x, b 0 ))ydy. Y Denoting W = Y B(y)f (y; m(x, b 0))dy and by noting that Y f (y; m(x, b 0))dy = 1 and that Y f (y; m(x, b 0))ydy = m(x, b 0 ), we have : =A(m(X, b)) + W + C(m(X, b))m(x, b 0 ). Y Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 26/30
27 D 2 D 2 = A(m(X, b 0 )) + W + C(m(X, b 0 ))m(x, b 0 ). Therefore the Kullback inequality reduces to : A(m(X, b 0 )) + C(m(X, b 0 ))m(x, b 0 ) A(m(X, b)) + C(m(X, b))m(x, b 0 ), and since b is first order identifiable and b 0 is the PMLE, it follows that the equality is only possible if : m(x, b) = m(x, b 0 ) b = b 0 b 0 = b 0 ˆb n = b 0. And hence b 0 is the value that minimizes the Kullback-Leiber distance. Or it is the value that maximizes the psuedo-likelihood. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 27/30
28 Remarks : 1 2 C(m(X,b)) m = Σ 1 where Σ is the variance-covariance matrix associated with the pseudo true distribution. A(m(X,b)) m + C(m(X,b)) m(x, b) = 0 is the FOC. m 3 Observations need not have the same support as that of the pseudo true distribution. For instance, a pseudo true family of Poisson can be used when the variables y are not positive integer-valued. 4 However, it is crucial to take into account the constraint m(x, b) M. For instance, the Poisson distribution can be used only if the conditional mean m(x, b) is positive valued. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 28/30
29 Property Under A1-A7 a consistent PMLE associated with a LEF is asymptotically normally distributed with n(ˆbn b 0 ) d N(0, J 1 IJ 1 ), where : ( ) m I = E x b Σ 1 ΩΣ 1 m b ( ) m m J = E x b Σ 1. b Ω is the true variance covariance matrix of Y conditional on X, Σ is the conditional variance covariance matrix associated with the chosen pseudo-likelihood. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 29/30
30 Property The variance covariance matrix of a PMLE is at least as large as : K 1 = ( )) m 1 m (E x b Ω 1. b The existence of this lower bound raises the question of whether it is possible to find a consistent estimator of which the asymptotic variance covariance attains the lower bound. This estimator would be asymptotically better than any other consistent PMLE. Yves Dominicy Graduate Econometrics I: Maximum Likelihood II 30/30
Graduate Econometrics I: Maximum Likelihood I
Graduate Econometrics I: Maximum Likelihood I Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Maximum Likelihood
More informationGraduate Econometrics I: Unbiased Estimation
Graduate Econometrics I: Unbiased Estimation Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Unbiased Estimation
More informationGraduate Econometrics I: Asymptotic Theory
Graduate Econometrics I: Asymptotic Theory Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Asymptotic Theory
More informationStatistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach
Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score
More informationFinal Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.
1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. (a) If X and Y are independent, Corr(X, Y ) = 0. (b) (c) (d) (e) A consistent estimator must be asymptotically
More informationIntroduction to Estimation Methods for Time Series models Lecture 2
Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:
More informationGraduate Econometrics I: What is econometrics?
Graduate Econometrics I: What is econometrics? Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: What is econometrics?
More informationEconometrics I, Estimation
Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the
More informationMLE and GMM. Li Zhao, SJTU. Spring, Li Zhao MLE and GMM 1 / 22
MLE and GMM Li Zhao, SJTU Spring, 2017 Li Zhao MLE and GMM 1 / 22 Outline 1 MLE 2 GMM 3 Binary Choice Models Li Zhao MLE and GMM 2 / 22 Maximum Likelihood Estimation - Introduction For a linear model y
More informationEstimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators
Estimation theory Parametric estimation Properties of estimators Minimum variance estimator Cramer-Rao bound Maximum likelihood estimators Confidence intervals Bayesian estimation 1 Random Variables Let
More informationChapter 3: Maximum Likelihood Theory
Chapter 3: Maximum Likelihood Theory Florian Pelgrin HEC September-December, 2010 Florian Pelgrin (HEC) Maximum Likelihood Theory September-December, 2010 1 / 40 1 Introduction Example 2 Maximum likelihood
More informationTheory of Maximum Likelihood Estimation. Konstantin Kashin
Gov 2001 Section 5: Theory of Maximum Likelihood Estimation Konstantin Kashin February 28, 2013 Outline Introduction Likelihood Examples of MLE Variance of MLE Asymptotic Properties What is Statistical
More informationEstimation of Dynamic Regression Models
University of Pavia 2007 Estimation of Dynamic Regression Models Eduardo Rossi University of Pavia Factorization of the density DGP: D t (x t χ t 1, d t ; Ψ) x t represent all the variables in the economy.
More informationStatistics and econometrics
1 / 36 Slides for the course Statistics and econometrics Part 10: Asymptotic hypothesis testing European University Institute Andrea Ichino September 8, 2014 2 / 36 Outline Why do we need large sample
More informationChapter 1: A Brief Review of Maximum Likelihood, GMM, and Numerical Tools. Joan Llull. Microeconometrics IDEA PhD Program
Chapter 1: A Brief Review of Maximum Likelihood, GMM, and Numerical Tools Joan Llull Microeconometrics IDEA PhD Program Maximum Likelihood Chapter 1. A Brief Review of Maximum Likelihood, GMM, and Numerical
More informationProblem Selected Scores
Statistics Ph.D. Qualifying Exam: Part II November 20, 2010 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. Problem 1 2 3 4 5 6 7 8 9 10 11 12 Selected
More informationMax. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes
Maximum Likelihood Estimation Econometrics II Department of Economics Universidad Carlos III de Madrid Máster Universitario en Desarrollo y Crecimiento Económico Outline 1 3 4 General Approaches to Parameter
More informationMathematical statistics
October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter
More informationLikelihood-based inference with missing data under missing-at-random
Likelihood-based inference with missing data under missing-at-random Jae-kwang Kim Joint work with Shu Yang Department of Statistics, Iowa State University May 4, 014 Outline 1. Introduction. Parametric
More informationChapter 4: Asymptotic Properties of the MLE (Part 2)
Chapter 4: Asymptotic Properties of the MLE (Part 2) Daniel O. Scharfstein 09/24/13 1 / 1 Example Let {(R i, X i ) : i = 1,..., n} be an i.i.d. sample of n random vectors (R, X ). Here R is a response
More informationf(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain
0.1. INTRODUCTION 1 0.1 Introduction R. A. Fisher, a pioneer in the development of mathematical statistics, introduced a measure of the amount of information contained in an observaton from f(x θ). Fisher
More informationTopic 12 Overview of Estimation
Topic 12 Overview of Estimation Classical Statistics 1 / 9 Outline Introduction Parameter Estimation Classical Statistics Densities and Likelihoods 2 / 9 Introduction In the simplest possible terms, the
More informationi=1 h n (ˆθ n ) = 0. (2)
Stat 8112 Lecture Notes Unbiased Estimating Equations Charles J. Geyer April 29, 2012 1 Introduction In this handout we generalize the notion of maximum likelihood estimation to solution of unbiased estimating
More informationStatistics Ph.D. Qualifying Exam: Part II November 3, 2001
Statistics Ph.D. Qualifying Exam: Part II November 3, 2001 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. 1 2 3 4 5 6 7 8 9 10 11 12 2. Write your
More informationLasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices
Article Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Fei Jin 1,2 and Lung-fei Lee 3, * 1 School of Economics, Shanghai University of Finance and Economics,
More informationAdvanced Quantitative Methods: maximum likelihood
Advanced Quantitative Methods: Maximum Likelihood University College Dublin 4 March 2014 1 2 3 4 5 6 Outline 1 2 3 4 5 6 of straight lines y = 1 2 x + 2 dy dx = 1 2 of curves y = x 2 4x + 5 of curves y
More informationChapter 3. Point Estimation. 3.1 Introduction
Chapter 3 Point Estimation Let (Ω, A, P θ ), P θ P = {P θ θ Θ}be probability space, X 1, X 2,..., X n : (Ω, A) (IR k, B k ) random variables (X, B X ) sample space γ : Θ IR k measurable function, i.e.
More informationPh.D. Qualifying Exam Friday Saturday, January 6 7, 2017
Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a
More informationPh.D. Qualifying Exam Friday Saturday, January 3 4, 2014
Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Put your solution to each problem on a separate sheet of paper. Problem 1. (5166) Assume that two random samples {x i } and {y i } are independently
More informationModification and Improvement of Empirical Likelihood for Missing Response Problem
UW Biostatistics Working Paper Series 12-30-2010 Modification and Improvement of Empirical Likelihood for Missing Response Problem Kwun Chuen Gary Chan University of Washington - Seattle Campus, kcgchan@u.washington.edu
More informationMaximum Likelihood Estimation
University of Pavia Maximum Likelihood Estimation Eduardo Rossi Likelihood function Choosing parameter values that make what one has observed more likely to occur than any other parameter values do. Assumption(Distribution)
More informationLecture 10 Maximum Likelihood Asymptotics under Non-standard Conditions: A Heuristic Introduction to Sandwiches
University of Illinois Department of Economics Spring 2017 Econ 574 Roger Koenker Lecture 10 Maximum Likelihood Asymptotics under Non-standard Conditions: A Heuristic Introduction to Sandwiches Ref: Huber,
More informationPattern Recognition. Parameter Estimation of Probability Density Functions
Pattern Recognition Parameter Estimation of Probability Density Functions Classification Problem (Review) The classification problem is to assign an arbitrary feature vector x F to one of c classes. The
More informationRecent Advances in the analysis of missing data with non-ignorable missingness
Recent Advances in the analysis of missing data with non-ignorable missingness Jae-Kwang Kim Department of Statistics, Iowa State University July 4th, 2014 1 Introduction 2 Full likelihood-based ML estimation
More informationEconometrics I. Ricardo Mora
Econometrics I Department of Economics Universidad Carlos III de Madrid Master in Industrial Economics and Markets Outline Motivation 1 Motivation 2 3 4 Motivation The Analogy Principle The () is a framework
More informationMathematical statistics
October 1 st, 2018 Lecture 11: Sufficient statistic Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation
More informationStatement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.
MATHEMATICAL STATISTICS Homework assignment Instructions Please turn in the homework with this cover page. You do not need to edit the solutions. Just make sure the handwriting is legible. You may discuss
More informationDA Freedman Notes on the MLE Fall 2003
DA Freedman Notes on the MLE Fall 2003 The object here is to provide a sketch of the theory of the MLE. Rigorous presentations can be found in the references cited below. Calculus. Let f be a smooth, scalar
More informationChapter 4: Constrained estimators and tests in the multiple linear regression model (Part III)
Chapter 4: Constrained estimators and tests in the multiple linear regression model (Part III) Florian Pelgrin HEC September-December 2010 Florian Pelgrin (HEC) Constrained estimators September-December
More informationLecture 14 More on structural estimation
Lecture 14 More on structural estimation Economics 8379 George Washington University Instructor: Prof. Ben Williams traditional MLE and GMM MLE requires a full specification of a model for the distribution
More information1 One-way analysis of variance
LIST OF FORMULAS (Version from 21. November 2014) STK2120 1 One-way analysis of variance Assume X ij = µ+α i +ɛ ij ; j = 1, 2,..., J i ; i = 1, 2,..., I ; where ɛ ij -s are independent and N(0, σ 2 ) distributed.
More informationP n. This is called the law of large numbers but it comes in two forms: Strong and Weak.
Large Sample Theory Large Sample Theory is a name given to the search for approximations to the behaviour of statistical procedures which are derived by computing limits as the sample size, n, tends to
More informationMaximum Likelihood Tests and Quasi-Maximum-Likelihood
Maximum Likelihood Tests and Quasi-Maximum-Likelihood Wendelin Schnedler Department of Economics University of Heidelberg 10. Dezember 2007 Wendelin Schnedler (AWI) Maximum Likelihood Tests and Quasi-Maximum-Likelihood10.
More informationGARCH Models Estimation and Inference
GARCH Models Estimation and Inference Eduardo Rossi University of Pavia December 013 Rossi GARCH Financial Econometrics - 013 1 / 1 Likelihood function The procedure most often used in estimating θ 0 in
More informationIntroduction Large Sample Testing Composite Hypotheses. Hypothesis Testing. Daniel Schmierer Econ 312. March 30, 2007
Hypothesis Testing Daniel Schmierer Econ 312 March 30, 2007 Basics Parameter of interest: θ Θ Structure of the test: H 0 : θ Θ 0 H 1 : θ Θ 1 for some sets Θ 0, Θ 1 Θ where Θ 0 Θ 1 = (often Θ 1 = Θ Θ 0
More information2 Statistical Estimation: Basic Concepts
Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:
More informationTheory of Statistics.
Theory of Statistics. Homework V February 5, 00. MT 8.7.c When σ is known, ˆµ = X is an unbiased estimator for µ. If you can show that its variance attains the Cramer-Rao lower bound, then no other unbiased
More informationMaster s Written Examination
Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth
More informationJoint work with Nottingham colleagues Simon Preston and Michail Tsagris.
/pgf/stepx/.initial=1cm, /pgf/stepy/.initial=1cm, /pgf/step/.code=1/pgf/stepx/.expanded=- 10.95415pt,/pgf/stepy/.expanded=- 10.95415pt, /pgf/step/.value required /pgf/images/width/.estore in= /pgf/images/height/.estore
More informationGauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA
JAPANESE BEETLE DATA 6 MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA Gauge Plots TuscaroraLisa Central Madsen Fairways, 996 January 9, 7 Grubs Adult Activity Grub Counts 6 8 Organic Matter
More informationMaster s Written Examination - Solution
Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2
More informationMathematics Ph.D. Qualifying Examination Stat Probability, January 2018
Mathematics Ph.D. Qualifying Examination Stat 52800 Probability, January 2018 NOTE: Answers all questions completely. Justify every step. Time allowed: 3 hours. 1. Let X 1,..., X n be a random sample from
More information(θ θ ), θ θ = 2 L(θ ) θ θ θ θ θ (θ )= H θθ (θ ) 1 d θ (θ )
Setting RHS to be zero, 0= (θ )+ 2 L(θ ) (θ θ ), θ θ = 2 L(θ ) 1 (θ )= H θθ (θ ) 1 d θ (θ ) O =0 θ 1 θ 3 θ 2 θ Figure 1: The Newton-Raphson Algorithm where H is the Hessian matrix, d θ is the derivative
More informationECE 275A Homework 7 Solutions
ECE 275A Homework 7 Solutions Solutions 1. For the same specification as in Homework Problem 6.11 we want to determine an estimator for θ using the Method of Moments (MOM). In general, the MOM estimator
More informationECE531 Lecture 10b: Maximum Likelihood Estimation
ECE531 Lecture 10b: Maximum Likelihood Estimation D. Richard Brown III Worcester Polytechnic Institute 05-Apr-2011 Worcester Polytechnic Institute D. Richard Brown III 05-Apr-2011 1 / 23 Introduction So
More informationWeak convergence of Markov chain Monte Carlo II
Weak convergence of Markov chain Monte Carlo II KAMATANI, Kengo Mar 2011 at Le Mans Background Markov chain Monte Carlo (MCMC) method is widely used in Statistical Science. It is easy to use, but difficult
More informationLecture 1: Introduction
Principles of Statistics Part II - Michaelmas 208 Lecturer: Quentin Berthet Lecture : Introduction This course is concerned with presenting some of the mathematical principles of statistical theory. One
More informationTesting Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata
Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function
More information1. (Regular) Exponential Family
1. (Regular) Exponential Family The density function of a regular exponential family is: [ ] Example. Poisson(θ) [ ] Example. Normal. (both unknown). ) [ ] [ ] [ ] [ ] 2. Theorem (Exponential family &
More informationExercises Chapter 4 Statistical Hypothesis Testing
Exercises Chapter 4 Statistical Hypothesis Testing Advanced Econometrics - HEC Lausanne Christophe Hurlin University of Orléans December 5, 013 Christophe Hurlin (University of Orléans) Advanced Econometrics
More informationClassical Estimation Topics
Classical Estimation Topics Namrata Vaswani, Iowa State University February 25, 2014 This note fills in the gaps in the notes already provided (l0.pdf, l1.pdf, l2.pdf, l3.pdf, LeastSquares.pdf). 1 Min
More informationEstimation of Threshold Cointegration
Estimation of Myung Hwan London School of Economics December 2006 Outline Model Asymptotics Inference Conclusion 1 Model Estimation Methods Literature 2 Asymptotics Consistency Convergence Rates Asymptotic
More informationPROBABILITY AND INFORMATION THEORY. Dr. Gjergji Kasneci Introduction to Information Retrieval WS
PROBABILITY AND INFORMATION THEORY Dr. Gjergji Kasneci Introduction to Information Retrieval WS 2012-13 1 Outline Intro Basics of probability and information theory Probability space Rules of probability
More informationBIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation
BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation Yujin Chung November 29th, 2016 Fall 2016 Yujin Chung Lec13: MLE Fall 2016 1/24 Previous Parametric tests Mean comparisons (normality assumption)
More informationPosterior Regularization
Posterior Regularization 1 Introduction One of the key challenges in probabilistic structured learning, is the intractability of the posterior distribution, for fast inference. There are numerous methods
More informationFor iid Y i the stronger conclusion holds; for our heuristics ignore differences between these notions.
Large Sample Theory Study approximate behaviour of ˆθ by studying the function U. Notice U is sum of independent random variables. Theorem: If Y 1, Y 2,... are iid with mean µ then Yi n µ Called law of
More informationSTAT 135 Lab 3 Asymptotic MLE and the Method of Moments
STAT 135 Lab 3 Asymptotic MLE and the Method of Moments Rebecca Barter February 9, 2015 Maximum likelihood estimation (a reminder) Maximum likelihood estimation Suppose that we have a sample, X 1, X 2,...,
More informationINFORMATION THEORY AND STATISTICS
INFORMATION THEORY AND STATISTICS Solomon Kullback DOVER PUBLICATIONS, INC. Mineola, New York Contents 1 DEFINITION OF INFORMATION 1 Introduction 1 2 Definition 3 3 Divergence 6 4 Examples 7 5 Problems...''.
More informationMaximum Likelihood Estimation
Maximum Likelihood Estimation Assume X P θ, θ Θ, with joint pdf (or pmf) f(x θ). Suppose we observe X = x. The Likelihood function is L(θ x) = f(x θ) as a function of θ (with the data x held fixed). The
More informationBrief Review on Estimation Theory
Brief Review on Estimation Theory K. Abed-Meraim ENST PARIS, Signal and Image Processing Dept. abed@tsi.enst.fr This presentation is essentially based on the course BASTA by E. Moulines Brief review on
More informationEM Algorithm II. September 11, 2018
EM Algorithm II September 11, 2018 Review EM 1/27 (Y obs, Y mis ) f (y obs, y mis θ), we observe Y obs but not Y mis Complete-data log likelihood: l C (θ Y obs, Y mis ) = log { f (Y obs, Y mis θ) Observed-data
More informationNonlinear Error Correction Model and Multiple-Threshold Cointegration May 23, / 31
Nonlinear Error Correction Model and Multiple-Threshold Cointegration Man Wang Dong Hua University, China Joint work with N.H.Chan May 23, 2014 Nonlinear Error Correction Model and Multiple-Threshold Cointegration
More informationMS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari
MS&E 226: Small Data Lecture 11: Maximum likelihood (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 18 The likelihood function 2 / 18 Estimating the parameter This lecture develops the methodology behind
More informationParametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory
Statistical Inference Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory IP, José Bioucas Dias, IST, 2007
More informationThe properties of L p -GMM estimators
The properties of L p -GMM estimators Robert de Jong and Chirok Han Michigan State University February 2000 Abstract This paper considers Generalized Method of Moment-type estimators for which a criterion
More informationStatistics Ph.D. Qualifying Exam
Department of Statistics Carnegie Mellon University May 7 2008 Statistics Ph.D. Qualifying Exam You are not expected to solve all five problems. Complete solutions to few problems will be preferred to
More informationA General Overview of Parametric Estimation and Inference Techniques.
A General Overview of Parametric Estimation and Inference Techniques. Moulinath Banerjee University of Michigan September 11, 2012 The object of statistical inference is to glean information about an underlying
More informationInformation in a Two-Stage Adaptive Optimal Design
Information in a Two-Stage Adaptive Optimal Design Department of Statistics, University of Missouri Designed Experiments: Recent Advances in Methods and Applications DEMA 2011 Isaac Newton Institute for
More informationPrimal-dual Covariate Balance and Minimal Double Robustness via Entropy Balancing
Primal-dual Covariate Balance and Minimal Double Robustness via (Joint work with Daniel Percival) Department of Statistics, Stanford University JSM, August 9, 2015 Outline 1 2 3 1/18 Setting Rubin s causal
More informationECON 3150/4150, Spring term Lecture 6
ECON 3150/4150, Spring term 2013. Lecture 6 Review of theoretical statistics for econometric modelling (II) Ragnar Nymoen University of Oslo 31 January 2013 1 / 25 References to Lecture 3 and 6 Lecture
More informationENEE 621 SPRING 2016 DETECTION AND ESTIMATION THEORY THE PARAMETER ESTIMATION PROBLEM
c 2007-2016 by Armand M. Makowski 1 ENEE 621 SPRING 2016 DETECTION AND ESTIMATION THEORY THE PARAMETER ESTIMATION PROBLEM 1 The basic setting Throughout, p, q and k are positive integers. The setup With
More informationLecture 3 September 1
STAT 383C: Statistical Modeling I Fall 2016 Lecture 3 September 1 Lecturer: Purnamrita Sarkar Scribe: Giorgio Paulon, Carlos Zanini Disclaimer: These scribe notes have been slightly proofread and may have
More informationSemiparametric Gaussian Copula Models: Progress and Problems
Semiparametric Gaussian Copula Models: Progress and Problems Jon A. Wellner University of Washington, Seattle 2015 IMS China, Kunming July 1-4, 2015 2015 IMS China Meeting, Kunming Based on joint work
More informationComprehensive Examination Quantitative Methods Spring, 2018
Comprehensive Examination Quantitative Methods Spring, 2018 Instruction: This exam consists of three parts. You are required to answer all the questions in all the parts. 1 Grading policy: 1. Each part
More informationOn the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models
On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models Thomas Kneib Institute of Statistics and Econometrics Georg-August-University Göttingen Department of Statistics
More informationThe Bivariate Probit Model, Maximum Likelihood Estimation, Pseudo True Parameters and Partial Identification
ISSN 1440-771X Department of Econometrics and Business Statistics http://business.monash.edu/econometrics-and-business-statistics/research/publications The Bivariate Probit Model, Maximum Likelihood Estimation,
More informationThe logistic regression model is thus a glm-model with canonical link function so that the log-odds equals the linear predictor, that is
Example The logistic regression model is thus a glm-model with canonical link function so that the log-odds equals the linear predictor, that is log p 1 p = β 0 + β 1 f 1 (y 1 ) +... + β d f d (y d ).
More informationBayesian estimation of the discrepancy with misspecified parametric models
Bayesian estimation of the discrepancy with misspecified parametric models Pierpaolo De Blasi University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics workshop ICERM, 17-21 September 2012
More informationMultivariate Regression Analysis
Matrices and vectors The model from the sample is: Y = Xβ +u with n individuals, l response variable, k regressors Y is a n 1 vector or a n l matrix with the notation Y T = (y 1,y 2,...,y n ) 1 x 11 x
More informationStat 5102 Final Exam May 14, 2015
Stat 5102 Final Exam May 14, 2015 Name Student ID The exam is closed book and closed notes. You may use three 8 1 11 2 sheets of paper with formulas, etc. You may also use the handouts on brand name distributions
More informationChapter 8.8.1: A factorization theorem
LECTURE 14 Chapter 8.8.1: A factorization theorem The characterization of a sufficient statistic in terms of the conditional distribution of the data given the statistic can be difficult to work with.
More informationVarious types of likelihood
Various types of likelihood 1. likelihood, marginal likelihood, conditional likelihood, profile likelihood, adjusted profile likelihood 2. semi-parametric likelihood, partial likelihood 3. empirical likelihood,
More informationQuick Review on Linear Multiple Regression
Quick Review on Linear Multiple Regression Mei-Yuan Chen Department of Finance National Chung Hsing University March 6, 2007 Introduction for Conditional Mean Modeling Suppose random variables Y, X 1,
More informationCalibration Estimation of Semiparametric Copula Models with Data Missing at Random
Calibration Estimation of Semiparametric Copula Models with Data Missing at Random Shigeyuki Hamori 1 Kaiji Motegi 1 Zheng Zhang 2 1 Kobe University 2 Renmin University of China Econometrics Workshop UNC
More informationChapter 3 : Likelihood function and inference
Chapter 3 : Likelihood function and inference 4 Likelihood function and inference The likelihood Information and curvature Sufficiency and ancilarity Maximum likelihood estimation Non-regular models EM
More informationSemiparametric Gaussian Copula Models: Progress and Problems
Semiparametric Gaussian Copula Models: Progress and Problems Jon A. Wellner University of Washington, Seattle European Meeting of Statisticians, Amsterdam July 6-10, 2015 EMS Meeting, Amsterdam Based on
More informationLikelihood-Based Methods
Likelihood-Based Methods Handbook of Spatial Statistics, Chapter 4 Susheela Singh September 22, 2016 OVERVIEW INTRODUCTION MAXIMUM LIKELIHOOD ESTIMATION (ML) RESTRICTED MAXIMUM LIKELIHOOD ESTIMATION (REML)
More informationThis paper is not to be removed from the Examination Halls
~~ST104B ZA d0 This paper is not to be removed from the Examination Halls UNIVERSITY OF LONDON ST104B ZB BSc degrees and Diplomas for Graduates in Economics, Management, Finance and the Social Sciences,
More informationInvestigation of an Automated Approach to Threshold Selection for Generalized Pareto
Investigation of an Automated Approach to Threshold Selection for Generalized Pareto Kate R. Saunders Supervisors: Peter Taylor & David Karoly University of Melbourne April 8, 2015 Outline 1 Extreme Value
More informationEcon 583 Final Exam Fall 2008
Econ 583 Final Exam Fall 2008 Eric Zivot December 11, 2008 Exam is due at 9:00 am in my office on Friday, December 12. 1 Maximum Likelihood Estimation and Asymptotic Theory Let X 1,...,X n be iid random
More information