Estimation: Part 2. Chapter GREG estimation

Similar documents
Efficient nonresponse weighting adjustment using estimated response probability

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Econ Statistical Properties of the OLS estimator. Sanjaya DeSilva

Conditional and unconditional models in modelassisted estimation of finite population totals

Economics 130. Lecture 4 Simple Linear Regression Continued

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.

e i is a random error

Matrix Approximation via Sampling, Subspace Embedding. 1 Solving Linear Systems Using SVD

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

ECONOMICS 351*-A Mid-Term Exam -- Fall Term 2000 Page 1 of 13 pages. QUEEN'S UNIVERSITY AT KINGSTON Department of Economics

Chapter 7 Generalized and Weighted Least Squares Estimation. In this method, the deviation between the observed and expected values of

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

Generalized Linear Methods

Lecture Notes on Linear Regression

Small Area Interval Estimation

Composite Hypotheses testing

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix

Lecture 3 Stat102, Spring 2007

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family

Explaining the Stein Paradox

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010

A note on regression estimation with unknown population size

MMA and GCMMA two methods for nonlinear optimization

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

However, since P is a symmetric idempotent matrix, of P are either 0 or 1 [Eigen-values

Parametric fractional imputation for missing data analysis

Improved Class of Ratio Estimators for Finite Population Variance

Solutions Homework 4 March 5, 2018

Linear Approximation with Regularization and Moving Least Squares

Nonparametric model calibration estimation in survey sampling

APPENDIX A Some Linear Algebra

Lecture 4 Hypothesis Testing

MATH 281A: Homework #6

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands

ANOMALIES OF THE MAGNITUDE OF THE BIAS OF THE MAXIMUM LIKELIHOOD ESTIMATOR OF THE REGRESSION SLOPE

Exam. Econometrics - Exam 1

First Year Examination Department of Statistics, University of Florida

n α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0

Goodness of fit and Wilks theorem

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)

Computation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models

LINEAR REGRESSION ANALYSIS. MODULE VIII Lecture Indicator Variables

ON A DETERMINATION OF THE INITIAL FUNCTIONS FROM THE OBSERVED VALUES OF THE BOUNDARY FUNCTIONS FOR THE SECOND-ORDER HYPERBOLIC EQUATION

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal

Nonparametric Regression Estimation. of Finite Population Totals. under Two-Stage Sampling

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA

Asymptotics of the Solution of a Boundary Value. Problem for One-Characteristic Differential. Equation Degenerating into a Parabolic Equation

princeton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg

STAT 3008 Applied Regression Analysis

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

Systems of Equations (SUR, GMM, and 3SLS)

x i1 =1 for all i (the constant ).

Lecture 4: Universal Hash Functions/Streaming Cont d

Appendix for Causal Interaction in Factorial Experiments: Application to Conjoint Analysis

Interval Estimation in the Classical Normal Linear Regression Model. 1. Introduction

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Lossy Compression. Compromise accuracy of reconstruction for increased compression.

18.1 Introduction and Recap

USE OF DOUBLE SAMPLING SCHEME IN ESTIMATING THE MEAN OF STRATIFIED POPULATION UNDER NON-RESPONSE

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models

Econometrics of Panel Data

where I = (n x n) diagonal identity matrix with diagonal elements = 1 and off-diagonal elements = 0; and σ 2 e = variance of (Y X).

Bias-correction under a semi-parametric model for small area estimation

Solutions to exam in SF1811 Optimization, Jan 14, 2015

Random Partitions of Samples

A general class of estimators for the population mean using multi-phase sampling with the non-respondents

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models

REPLICATION VARIANCE ESTIMATION UNDER TWO-PHASE SAMPLING IN THE PRESENCE OF NON-RESPONSE

Joint Statistical Meetings - Biopharmaceutical Section

8/25/17. Data Modeling. Data Modeling. Data Modeling. Patrice Koehl Department of Biological Sciences National University of Singapore

Convergence of random processes

Stat 642, Lecture notes for 01/27/ d i = 1 t. n i t nj. n j

CSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography

Supplementary material: Margin based PU Learning. Matrix Concentration Inequalities

Lecture 21: Numerical methods for pricing American type derivatives

A Note on Test of Homogeneity Against Umbrella Scale Alternative Based on U-Statistics

SIO 224. m(r) =(ρ(r),k s (r),µ(r))

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

Classification as a Regression Problem

Chapter 4: Regression With One Regressor

Lecture 10: May 6, 2013

Some modelling aspects for the Matlab implementation of MMA

Chapter 8 Indicator Variables

Errors for Linear Systems

PHYS 705: Classical Mechanics. Calculus of Variations II

Statistics for Managers Using Microsoft Excel/SPSS Chapter 13 The Simple Linear Regression Model and Correlation

Introduction to Algorithms

Chapter 11: Simple Linear Regression and Correlation

princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 3: Large deviations bounds and applications Lecturer: Sanjeev Arora

Chapter 3 Describing Data Using Numerical Measures

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1

Improvement in Estimating the Population Mean Using Exponential Estimator in Simple Random Sampling

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 8 Luca Trevisan February 17, 2016

Statistics II Final Exam 26/6/18

Lecture 2: Prelude to the big shrink

Estimation of the Mean of Truncated Exponential Distribution

Transcription:

Chapter 9 Estmaton: Part 2 9. GREG estmaton In Chapter 8, we have seen that the regresson estmator s an effcent estmator when there s a lnear relatonshp between y and x. In ths chapter, we generalzed the concept of regresson estmaton to a more general class of models for developng model-asssted estmaton. To motvate the proposed estmator, we frst ntroduce dfferent estmator. Suppose that a proxy value of y, denoted by y 0), throughout the populaton. The dfferent estmator of Y = y usng y 0) Ŷ dff =,,y0) s defned as y 0 + y y 0 ). 9.) The dfference estmator s unbased regardless of the choce of y 0). The varance of the dfference estmator s Var ) ) d Ŷ dff = Var where d = y y 0. Under SRS, the varance s 2 n n ) d d ) 2.

2 CHAPTER 9. ESTIMATIO: PART 2 Thus, the dfference estmator s unbased regardless of y 0) but ts varances are dfferent for dfferent choce of y 0). If y 0) s close to the true value of y, then the varance wll be small. In practce, we do not know y 0). Instead, we use auxlary varable x to construct a model for y and develop y 0) from the model. That s, we assume that the fnte populaton s a random sample from a superpopulaton model that has generated the current fnte populaton. One of the commonly used superpopulaton model s E ζ y ) = x β 9.2) Cov ζ y,y j ) = c σ 2 f = j 0 f j 9.3) where c = cx ) s a known functon of x and β are σ 2 are unknown parameters. Model 9.3) s often called the generalzed regresson GREG) model. Under the GREG model n 9.3), the model-based estmator s defned to be the sum of the predcted values n the GREG model. That s, we defne Ŷ GREG = ŷ 9.4) where ŷ = x ˆβ and ˆβ = ) x x c x y. 9.5) c ote that ˆβ converges to the followng fnte populaton quantty B = x x ) c x y, 9.6) c whch s the best estmator of β under Census. The model-based estmator n 9.4) s developed under the GERG model n 9.3). Thus, t s often called GREG estmator. The followng lemma shows that the GREG estmator n 9.4) s asymptotcally equvalent to the dfference estmator n 9.).

9.. GREG ESTIMATIO 3 Lemma 9.. Under some regularty condtons, f c satsfes c = λ x for some λ, the GREG estmator n 9.4) satsfes where B s defned n 9.5). Ŷ GREG = x B + Proof. If c = λ x holds for some λ, and we can wrte Ŷ GREG = ŷ = λ x x c ˆβ = λ x y c y x π B) 9.7) = λ x y = c y x ˆβ + y x ˆβ ) = Ŷ HT + X ˆX HT ) ˆβ = Ŷ HT + X ˆX HT ) B + X ˆX HT ) ˆβ B ) and the last term s neglgble under some regularty condtons. and By 9.7), we can establsh that E Ŷ GREG ) = Y 9.8) Var ) Ŷ GREG = Var π y x B ). 9.9) If E = y x B were observed, the varance 9.9) would be estmated by Var Ŷ GREG ) = j A j j E E j.

4 CHAPTER 9. ESTIMATIO: PART 2 If we use e = y x ˆβ nstead of E, a consstent varance estmator s computed by ˆV GREG = j A Example 9.. Rakng rato estmaton) j j e e j. 9.0) Suppose that we have I J groups or cells. Cell counts j are not known. Margnal counts = J j= j and j = I j are known. In ths case, we may consder the followng two-way addtve model E ζ Y k ) = α + β j V ζ Y k ) = σ 2 where α, β j, and σ 2 are unknown parameters. Defne δ jk = f k U j 0 otherwse. Unfortunately, we do not observe δ jk n the populaton. Let x k = δ,k,δ 2 k,,δ I k,δ k,δ 2k,,δ Jk ) and we know k= x k. The GREG estmator n ths case can be wrtten as Ŷ GREG = g A)y where g A) = + x k k= k A x k π k ) k A ) x k x k x. π k Unfortunately, we cannot compute the nverse of k A π k x k x k because ts rank s I + J, whch s not full rank. Thus, there s no unque soluton ˆB to π x x ˆB = π x y.

9.2. OPTIMAL ESTIMATIO 5 The goal s to fnd g ka = g k A) such that g ka δ,k = k A π k g ka δ jk = k A π k δ k, =,2,,I 9.) k= δ jk, j =,2,,J. 9.2) k= One way to obtan the soluton to 9.) and 9.2) s to solve the equatons teratvely as follows:. Start wth g 0) ka =. 2. For δ k =, g t+) ka = g t) k= δ k ka k A g t) ka δ. k/π k It satsfes 9.), but not necessarly satsfy 9.2). 3. For δ jk =, g t+2) ka = g t+) ka k= δ jk k A g t+) ka δ jk /π k. It satsfes 9.2), but not necessarly satsfy 9.). 4. Set t t + 2 and go to Step 2. Contnue untl convergence. Such computaton method s called rakng rato estmaton and was frst consdered by Demng and Stephan 940) n the Census applcaton. See also Devlle, et al. 993). 9.2 Optmal Estmaton So far, we have dscussed a class of model-asssted estmators that mprove the effcency of the HT estmator by ncorporatng the auxlary nformaton. In ths secton, we dscuss some optmalty of the GREG estmator under some class of estmators. We frst show the nonexstence of the UMVUE Unformly Mnmum

6 CHAPTER 9. ESTIMATIO: PART 2 Varance Unbased Estmator) n a strctly desgn-based sense, whch was orgnally dscussed by Godambe and Josh 965) and then also proved by Basu 97) wth a smpler proof. Theorem 9.. Let any noncensus desgn wth π k > 0 k =,2,,) be gven. Then no unformly mnmum varance estmator exsts n the class of all unbased estmators of Y = y. Proof. For gven value y = y,y 2,,y ), consder the followng dfference estmator Ŷ dff = y + y y ) 9.3) whch s unbased regardless of y = y,y 2,,y ) and ts varance s zero when y = y. ow, n order for an unbased estmator Ŷ to be an UMVUE, t should satsfy Var Ŷ ) Var Ŷ dff ), y. ow, snce Var ) Ŷ dff = 0 for y = y, t means Var Ŷ ) = 0 for y = y. Snce we can choose any arbtrary y, t means that Var Ŷ ) = 0 for all y, whch holds only for the census. The above theorem shows that t s mpossble to fnd the best estmator among the class of unbased estmator n the sense of mnmzng the desgn varance. To overcome ths problem, we relax the objectve functon for comparng the effcency of the estmators by allowng for the superpopulaton model nto consderaton. Specfcally, we wll consder the expected value of the desgn varance under the superpopulaton model. Such varance s called antcpated varance and s formally defned as follows. Defnton 9.. Antcpated varance of ˆθ s defned by AV ˆθ ) = E ζ Var ) p ˆθ, where subscrpt ζ denotes the dstrbuton wth respect to the superpopulaton model and subscrpt p denotes the dstrbuton generated by the samplng mechansm.

9.2. OPTIMAL ESTIMATIO 7 Lemma 9.2. Let ˆθ be desgn-unbased for θ. The antcpated varance of ˆθ can be wrtten as AV ˆθ ) = E ) p Varζ ˆθ +Var ) p Eζ ˆθ Var ζ θ ). 9.4) Proof. Snce ˆθ s desgn-unbased for θ, we can wrte AV ˆθ ) = E ζ Var ) ) 2 p ˆθ = E ζ E p ˆθ θ. Thus, AV ˆθ ) = ) 2 E p E ζ ˆθ θ = 2 E p E ζ ˆθ E ζ ˆθ) + E ζ ˆθ) E ζ θ ) + E ζ θ ) θ and AV ˆθ ) = E p E ζ ˆθ E ζ ˆθ) 2 + Ep Eζ ˆθ) E ζ θ ) 2 2 + Ep Eζ θ ) θ +2E p ˆθ E ζ ˆθ) ) ) E ζ θ ) θ and the remanng cross product terms are zero. Snce E p ˆθ E ζ ˆθ) ) = ) E ζ θ ) θ and 2E p ˆθ E ζ ˆθ) ) ) E ζ θ ) θ = ) 2 Eζ θ ) θ Ep ˆθ E ζ ˆθ) ) = 2 ) 2 E ζ θ ) θ, we obtan AV ˆθ ) = E p E ζ ˆθ E ζ ˆθ) 2 + Ep Eζ ˆθ) E ζ θ ) 2 2 Eζ θ ) θ whch proves 9.4). The followng theorem presents the lower bound of the antcpated varance for a desgn unbased estmator.

8 CHAPTER 9. ESTIMATIO: PART 2 Theorem 9.2. Let y be ndependently dstrbuted n the superpopulaton model. The antcpate varance of any desgn-unbased estmator of Y satsfes E ζ Var Ŷ ) ) Var π ζ y ). 9.5) Proof. Wrte Ŷ as Ŷ = Ŷ HT + R where Ŷ HT s the HT estmator of Y. Snce Ŷ s desgn unbased, we have E R) = 0 and, for fxed j U, ow, snce 0 = E R) = pa)ra) A A = A A ; j A pa)ra) + A A ; j / A pa)ra). V ζ Ŷ ) = Vζ ŶHT ) +Vζ R) + 2Cov ζ ŶHT,R ), we obtan E p Covζ ŶHT,R ) = E p [ Eζ ŶHT E ζ ŶHT )) R ] = E p [ j U y j E ζ y j ) ) I j E ζ R ] y j E ζ y j ) ) = E ζ E I j A)RA) j U y j E ζ y j ) ) = E ζ j U RA) pa) = j U = 0, E ζ y j E ζ y j ) ) A A ; j A A A ; j / A RA) pa) where the last equalty follows because A A ; j / A RA) pa) does not depend on

9.2. OPTIMAL ESTIMATIO 9 y j. Thus, E p Vζ Ŷ ) = Ep Vζ ŶHT ) + Ep Vζ R) ) E p Vζ ŶHT ) y I = E p = E p V ζ σ 2I π 2 = σ 2 and we have AV Ŷ ) = E p V ζ Ŷ ) +Vp E ζ Ŷ ) Vζ Y ) E p V ζ Ŷ ) Vζ Y ) σ 2 U σ 2. The lower bound n 9.5) s the lower bound of the antcpated varance of any unbased estmator. The lower bound was frst dscovered by Godambe and Josh 965) and s often called Godambe-Josh lower bound. For a fx-ed sze probablty samplng desgn, the Godambe-Josh lower bound s mnmzed when Var ζ y ) /2. 9.6) To show ths, we mnmze Var ζ y )/ subject to = n. The soluton can be obtaned by applyng Cauchy -Schwarz nequalty to get σ 2 wth equalty when 9.6) holds. / ) σ 2 The followng theorem, whch was frst proved by Isak and Fuller 982), shows that the GREG estmator acheves the Godambe-Josh lower bound asymptotcally.

0 CHAPTER 9. ESTIMATIO: PART 2 Theorem 9.3. Suppose that ζ s a superpopulaton model wth y s ndependent and E ζ y ) = x β and V ζ y ) = c σ 2. Then, the antcpated varance of the GREG estmator n 9.4) asymptotcally attans the Godambe-Josh lower bound f c = λ x for some λ. Proof. By Lemma 9., when c = λ x for some λ, the GREG estmator s asymptotcally equvalent to the dfference estmator n 9.7). Thus, E ζ Var ŶGREG ). = Eζ. = E ζ = j= j= j ) y x B y j x j B j ) y x β y j x j β ) c σ 2 whch s equal to the Gobambe-Josh lower bound under the superpopulaton model. Reference Basu, D. 97). An essay on the logcal foundatons of survey samplng, part one. In V.P. Godambe and D.A. Sprott, Eds., Foundatons of Statstcal Inference. Toronto: Holt, Rnehart & Wnston, 203-242. Demng, W. E. and Stephan, F. F. 940). On a least squares adjustment of a sampled frequency table when the expected margnal totals are known. The Annals of Mathematcal Statstcs,, 427-444. Devlle, J.C., Särndal, C.E., and Sautory, O. 993). Generalzed rakng procedure n survey samplng, Journal of the Amercan Statstcal Assocaton 88, 03-020. Godambe, V.P. and Josh, V.M. 965). Admssblty and Bayes estmaton n samplng fnte populatons,. Annals of Mathematcal Statstcs 36, 707-722.

9.2. OPTIMAL ESTIMATIO Isak, C. and Fuller, W.A. 982). Survey desgn under the regresson superpopulaton model. Journal of the Amercan Statstcal Assocaton 78, 7-23.