Chapter 2. Section Section 2.9. J. Kim (ISU) Chapter 2 1 / 26. Design-optimal estimator under stratified random sampling
|
|
- Constance Hunter
- 5 years ago
- Views:
Transcription
1 Chapter 2 Section Section 2.9 J. Kim (ISU) Chapter 2 1 / Regression and stratification Design-optimal estimator under stratified random sampling where (Ŝxxh, Ŝxyh) ˆβ opt = ( x st, ȳ st ) = ( H ȳ reg = ȳ st + ( x N x) ˆβ opt Wh 2 (1 f h) n 1 h Ŝxxh ) 1 H Wh 2 (1 f h) n 1 h Ŝxyh n h = (n h 1) 1 (x hj x h ) (x hj x h, y hj ȳ h ) j=1 H W h ( x h, ȳ h ). J. Kim (ISU) Chapter 2 2 / 26
2 2.4 Regression and stratification Note that H n h ˆβ opt = K h (x hj x h ) (x hj x h ) j=1 1 H n h j=1 where K h = Wh 2(1 f h)n 1 h (n h 1) 1 = W 2 h (1 f h )n 2 h On the other hand, H n h ˆβ GREG = j=1 hj (x hj x h ) (x hj x h ) π 1 1 H n h j=1 K h (x hj x h ) y hj, = (1 f h )π 2 hi. π 1 hj (x hj x h ) y hj. Roughly speaking, ˆβ opt is the first part of the slope for the regression of π 1 hi y i on π 1 hi x i and z i, where z i is a vector of stratum indicator functions. J. Kim (ISU) Chapter 2 3 / Regression and stratification Given ˆβ, consider a regression estimator under stratified sampling ȳ st,reg = ȳ st + ( x N x st ) ˆβ. Write y hi = x hi β h + e hi, e hi ( 0, σe,h) 2. The large-sample variance of the regression estimator is V (ȳ st,reg ) = H Wh 2 (1 f h) n 1 h σ2 a,h where σ 2 a,h = σ2 e,h + (β h β N ) Σ xx,h (β h β N ), Σ xx,h = V {x hi }, and β N is the probability limit of ˆβ. J. Kim (ISU) Chapter 2 4 / 26
3 2.4 Regression and stratification Example Two estimators of β: ˆβ wls = ( X D w X ) 1 X D w y ˆβ opt = ( X D 2 w X ) 1 X D 2 w y, where D w is a diagonal matrix with diagonal elements equal to W h n 1 h for units in stratum h. Probability limits: β ols,n = p lim ˆβ wls = ( X NX N ) 1 X Ny N β opt,n = p lim ˆβ opt = ( X ND w,n X N ) 1 X ND w,n y N J. Kim (ISU) Chapter 2 5 / Regression and stratification Example (Cont d) For example, assume H = 2 with W 1 = 0.15 and W 2 = Stratum parameters: σ 2 x,h = { 4.3 if h = if h = 2, β 1,h = Population regression coefficients (under n 1 = n 2 ) { 3.0 if h = if h = 2 β obs,n = H W hσ 2 xh β 1h H W hσ 2 xh = β opt,n = H W 2 h σ2 xh β 1h H W 2 h σ2 xh = J. Kim (ISU) Chapter 2 6 / 26
4 2.4 Regression and stratification Example (Cont d) To compare the variances, assume that σ 2 e,h = { 24 if h = if h = 2. Stratum variances of the residuals from ˆβ ols. σ 2 a,h = { ( ) 2 (4.3) + 24 = if h = 1 ( ) 2 (0.6) = if h = 2 Stratum variances of the residuals from ˆβ opt. σ 2 a,h = { ( ) 2 (4.3) + 24 = if h = 1 ( ) 2 (0.6) = if h = 2 J. Kim (ISU) Chapter 2 7 / Regression and stratification Example (Cont d) (Under n h =constant,) the large-sample variances of the regression estimator satisfy and n h V {ȳ st,reg,wls } = (0.15) 2 ( ) + (0.85) 2 (1.5485) = n h V {ȳ st,reg,opt } = (0.15) 2 ( ) + (0.85) 2 (0.8106) = Roughly speaking, β ols,n minimizes h W hσa,h 2 while β opt,n minimizes h W h 2n 1 h σ2 a,h, where σ2 ah = E { (y hi x hi β) 2}. J. Kim (ISU) Chapter 2 8 / 26
5 2.4 Regression and stratification If x h,n = N 1 Nh h i=1 x hi are available then we can construct a separate regression estimator ȳ s,reg = N W h x h,n ˆβ h where ˆβ h = { nh i=1 ( x hi x hn ) ( x hi x hn ) } 1 n h ( x hi x hn ) y hi. i=1 Because the weights are the same within each stratum, the GREG type estimator is the same as the design-optimal estimator when the separate regression estimation is used. Bias can be sizable if n h are small in some strata. J. Kim (ISU) Chapter 2 9 / Regression for two-stage samples Basic Setup Two-stage cluster sampling 1 Stage One: select n clusters 2 Stage Two: Within the selected cluster i, select m i second-stage units (from the M i units). π (ij) : the inclusion probability of selecting element j in primary sampling unit i. (π (ij) = π 1i π 2j i ) The analysis unit is the element, not the cluster. Thus, we want to construct weights for the sample elements. J. Kim (ISU) Chapter 2 10 / 26
6 2.6 Regression for two-stage samples Basic Setup Two-types of auxiliary information x ij : element level auxiliary information z i : cluster level auxiliary information Want to incorporate the auxiliary information. I j A i w ij x ij = I j A i w ij z i = i U I M i j=1 i U I z i x ij J. Kim (ISU) Chapter 2 11 / Regression for two-stage samples Approach 1 Construct z ij from z i and apply the regression weighting method using (x ij, z ij ) in the sample. Use z ij = z im 1 i π 2j i. Note that j A i π 1 2j i z ij = z i and so E π 1 1i π 1 2j i z ij = E j A i I I π 1 1i z i = i UI z i. J. Kim (ISU) Chapter 2 12 / 26
7 2.6 Regression for two-stage samples Approach 2 : design-consistent model-based approach Model for the two-stage sample y ij = x ij β + u ij u ij = b i + e ij where b i iid(0, σ 2 b ), e ij iid(o, σ 2 e), and e ij is independent of b k for all i, j, k. Writing u i = (u i1,, u im ), we have u i (0, Σ uu ) where Σ uu = I m σ 2 e + J m J mσ 2 a. For illustration, see Example J. Kim (ISU) Chapter 2 13 / Calibration Minimize ω Vω s.t. ω X = x N (ω Vω)(aX V 1 Xa ) (ω Xa ) 2 with equality iff ω V 1/2 ax V 1/2 ω ax V 1 ω = kax V 1, k : constant ω X = kax V 1 X & x N (X V 1 X) 1 = ka ω = x N (X V 1 X) 1 X V 1 ω Vω x N (X V 1 X) 1 x N Note Minimize V ξ (ω y) s.t. E ξ (ω y) = E(ȳ N ). J. Kim (ISU) Chapter 2 14 / 26
8 Alternative Minimization Lemma α : given n-dimensional vector Let ω a = arg min ω ω Vω s.t ω X = x N Let ω b = arg min ω (ω α) V(ω α) s.t ω X = x N If V α C(X), then ω a = ω b. Proof : (ω α) V(ω α) = ω Vω α Vω ω Vα + α Vα = ω Vω λ X ω ω Xλ + α Vα where V α = Xλ = ω Vω 2λ x N + α Vα ω X = x N If α = D 1 π J n, then V α C(X) is the condition for design consistency in Corollary J. Kim (ISU) Chapter 2 15 / 26 General Objective Function min G(ω i, α i ) s.t. ω i x i = x N Lagrange multiplier method g(ω i, α i ) λ x i = 0 where g(ω i, α i ) = G ω i ω i = g 1 (λ x i, α i ) where λ is from g 1 (λ x i, α i )x i = x N J. Kim (ISU) Chapter 2 16 / 26
9 GREG Estimator min Q(ω, d) = d 1 i (ω i d i )q i + λ x i = 0 ω i = d i + λ d i x i/q i ω i x i = ( ) 2 ωi d i 1 q i s.t. d i d i x i + λ d i x ix i /q i ω i x i = x N. λ = ( x N x HT )( d i x ix i /q i ) 1 w i = d i + ( x N x HT )( d i x ix i /q i ) 1 d i x i/q i J. Kim (ISU) Chapter 2 17 / 26 Other Objective Functions Pseudo empirical likelihood Q(ω, d) = d i log Kullback-Leibler distance: Q(ω, d) = ω i log ( ωi d i ( ωi d i ), ω i = d i /(1 + x i λ) ), ω i = d i exp(x i λ) J. Kim (ISU) Chapter 2 18 / 26
10 Theorem Deville and Särndal (1992) Theorem Let G(ω, α) be a continuous convex function with a first derivative that is zero for ω = α. Under some regularity conditions, the solution ω i that minimizes G(ω i, α i ) s.t. ω i x i = x N satisfies ω i y i = α i y i + ( x N x α ) ˆβ + O p (n 1 ) where ˆβ = ( x i x i/φ ii ) 1 x i y i/φ ii and φ ii = 2 G(α i, α i )/ ω 2 i. J. Kim (ISU) Chapter 2 19 / 26 Proof of Theorem Using the Lagrange multiplier method and Taylor linearization, ω i = ω i (λ) = g 1 (λ x i, α i ) where g(ω i, α i ) = G/ ω i. By assumption, g 1 (0, α i ) = α i. Define Û(λ) = ω ix i x N and let ˆλ satisfy Û(ˆλ) = 0. By Taylor 0 = Û(ˆλ) = Û(0) + Û(0) λ (ˆλ 0) + O p (n 1 ). Because Û(0) = α ix i and where g (α i, α i ) = Û(0) λ 2 ω 2 i = 1 g (α i, α i ) x ix i = x ix i /φ ii, G(ω i, α i ) = φ ii. ωi =α i J. Kim (ISU) Chapter 2 20 / 26
11 Proof of Theorem 2.7.1, continued ȳ cal (ˆλ) = ω i (ˆλ)y i [ ] ȳcal (0) = ȳ cal (0) + (ˆλ 0) + O p (n 1 ) λ = [ ] [ ] 1 x i α i y i + y i x i x i ( x N x α ) + O p (n 1 ). φ ii φ ii J. Kim (ISU) Chapter 2 21 / Weight Bounds ω i = d i + d i λ x i /c i can take negative values (or take very large values) Add L 1 ω i L 2 to ω i x i = x N. Approaches 1 Huang and Fuller: 2 Husain (1969) Q(w i, d i ) = d i Ψ ( wi d i ), Ψ : Huber function min ω ω + γ(ω X x N ) Σ 1 x x (ω X x N ) for some γ 3 Other methods, quadratic programming. J. Kim (ISU) Chapter 2 22 / 26
12 2.9 Maximum likelihood and raking ratio Basic Setup Two-way (r c) categorical data a km = n km n, [ nkm p km = E n p k, p m : known k = 1, 2,, r, m = 1, 2,, c ] We are interested in estimating p km. Constraints: ˆp km = p k ˆp km = p m m k J. Kim (ISU) Chapter 2 23 / 26 Maximum likelihood approach Multinomial Likelihood r c a km log(p km ) k=1 m=1 Lagrangian Multiplier Method r c r a km log(p km ) + p km p k ) + k=1 m=1 c p km = m=1 λ r+m ( r k=1 a km λ k + λ r+m ( c λ k k=1 m=1 ) p km p m J. Kim (ISU) Chapter 2 24 / 26
13 Raking ratio method Deming & Stephan (1940) idea: Approximate r k=1 m=1 c a km log(p km ). = r k=1 m=1 c {a km log(a km ) + (p km a km ) a 1 km (p km a km ) 2} Thus, maximizing r c k=1 m=1 a kmlog(p km ) is asymptotically equivalent to minimizing r c k=1 m=1 a 1 km (p km a km ) 2. If there is only one set of constraints, c p km = p k k = 1,, r, m=1 then the solution to minimizing c m=1 a 1 km (p km a km ) 2 s.t. the constraint is p k p km = a km c m=1 a. km J. Kim (ISU) Chapter 2 25 / 26 Raking ratio method (Cont d) For the two sets of constraints, c p km = p k k = 1,, r m=1 r p km = p m m = 1,, c. k=1 p (t+1) km = p (t) km p k, p (t+2) c m=1 p(t) km km = p (t+1) km p m r k=1 p(t+1) km J. Kim (ISU) Chapter 2 26 / 26
Empirical Likelihood Methods
Handbook of Statistics, Volume 29 Sample Surveys: Theory, Methods and Inference Empirical Likelihood Methods J.N.K. Rao and Changbao Wu (February 14, 2008, Final Version) 1 Likelihood-based Approaches
More informationAdvanced Topics in Survey Sampling
Advanced Topics in Survey Sampling Jae-Kwang Kim Wayne A Fuller Pushpal Mukhopadhyay Department of Statistics Iowa State University World Statistics Congress Short Course July 23-24, 2015 Kim & Fuller
More informationINSTRUMENTAL-VARIABLE CALIBRATION ESTIMATION IN SURVEY SAMPLING
Statistica Sinica 24 (2014), 1001-1015 doi:http://dx.doi.org/10.5705/ss.2013.038 INSTRUMENTAL-VARIABLE CALIBRATION ESTIMATION IN SURVEY SAMPLING Seunghwan Park and Jae Kwang Kim Seoul National Univeristy
More informationChapter 5: Models used in conjunction with sampling. J. Kim, W. Fuller (ISU) Chapter 5: Models used in conjunction with sampling 1 / 70
Chapter 5: Models used in conjunction with sampling J. Kim, W. Fuller (ISU) Chapter 5: Models used in conjunction with sampling 1 / 70 Nonresponse Unit Nonresponse: weight adjustment Item Nonresponse:
More informationCombining data from two independent surveys: model-assisted approach
Combining data from two independent surveys: model-assisted approach Jae Kwang Kim 1 Iowa State University January 20, 2012 1 Joint work with J.N.K. Rao, Carleton University Reference Kim, J.K. and Rao,
More informationEmpirical Likelihood Methods for Sample Survey Data: An Overview
AUSTRIAN JOURNAL OF STATISTICS Volume 35 (2006), Number 2&3, 191 196 Empirical Likelihood Methods for Sample Survey Data: An Overview J. N. K. Rao Carleton University, Ottawa, Canada Abstract: The use
More informationChapter 8: Estimation 1
Chapter 8: Estimation 1 Jae-Kwang Kim Iowa State University Fall, 2014 Kim (ISU) Ch. 8: Estimation 1 Fall, 2014 1 / 33 Introduction 1 Introduction 2 Ratio estimation 3 Regression estimator Kim (ISU) Ch.
More informationData Integration for Big Data Analysis for finite population inference
for Big Data Analysis for finite population inference Jae-kwang Kim ISU January 23, 2018 1 / 36 What is big data? 2 / 36 Data do not speak for themselves Knowledge Reproducibility Information Intepretation
More informationCalibration estimation in survey sampling
Calibration estimation in survey sampling Jae Kwang Kim Mingue Park September 8, 2009 Abstract Calibration estimation, where the sampling weights are adjusted to make certain estimators match known population
More informationChapter 4. Replication Variance Estimation. J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 1 / 28
Chapter 4 Replication Variance Estimation J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 1 / 28 Jackknife Variance Estimation Create a new sample by deleting one observation n 1 n n ( x (k) x) 2 = x (k) = n
More informationVariance Estimation for Calibration to Estimated Control Totals
Variance Estimation for Calibration to Estimated Control Totals Siyu Qing Coauthor with Michael D. Larsen Associate Professor of Statistics Tuesday, 11/05/2013 2 Outline A. Background B. Calibration Technique
More informationWeighted Least Squares
Weighted Least Squares The standard linear model assumes that Var(ε i ) = σ 2 for i = 1,..., n. As we have seen, however, there are instances where Var(Y X = x i ) = Var(ε i ) = σ2 w i. Here w 1,..., w
More informationAn Efficient Estimation Method for Longitudinal Surveys with Monotone Missing Data
An Efficient Estimation Method for Longitudinal Surveys with Monotone Missing Data Jae-Kwang Kim 1 Iowa State University June 28, 2012 1 Joint work with Dr. Ming Zhou (when he was a PhD student at ISU)
More informationGeneralized Pseudo Empirical Likelihood Inferences for Complex Surveys
The Canadian Journal of Statistics Vol.??, No.?,????, Pages???-??? La revue canadienne de statistique Generalized Pseudo Empirical Likelihood Inferences for Complex Surveys Zhiqiang TAN 1 and Changbao
More informationCalibration estimation using exponential tilting in sample surveys
Calibration estimation using exponential tilting in sample surveys Jae Kwang Kim February 23, 2010 Abstract We consider the problem of parameter estimation with auxiliary information, where the auxiliary
More informationSimple design-efficient calibration estimators for rejective and high-entropy sampling
Biometrika (202), 99,, pp. 6 C 202 Biometrika Trust Printed in Great Britain Advance Access publication on 3 July 202 Simple design-efficient calibration estimators for rejective and high-entropy sampling
More informationProblems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B
Simple Linear Regression 35 Problems 1 Consider a set of data (x i, y i ), i =1, 2,,n, and the following two regression models: y i = β 0 + β 1 x i + ε, (i =1, 2,,n), Model A y i = γ 0 + γ 1 x i + γ 2
More informationModification and Improvement of Empirical Likelihood for Missing Response Problem
UW Biostatistics Working Paper Series 12-30-2010 Modification and Improvement of Empirical Likelihood for Missing Response Problem Kwun Chuen Gary Chan University of Washington - Seattle Campus, kcgchan@u.washington.edu
More informationFractional Imputation in Survey Sampling: A Comparative Review
Fractional Imputation in Survey Sampling: A Comparative Review Shu Yang Jae-Kwang Kim Iowa State University Joint Statistical Meetings, August 2015 Outline Introduction Fractional imputation Features Numerical
More informationSTAT 100C: Linear models
STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix
More informationGraduate Econometrics I: Maximum Likelihood I
Graduate Econometrics I: Maximum Likelihood I Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Maximum Likelihood
More information6. Fractional Imputation in Survey Sampling
6. Fractional Imputation in Survey Sampling 1 Introduction Consider a finite population of N units identified by a set of indices U = {1, 2,, N} with N known. Associated with each unit i in the population
More informationPropensity score adjusted method for missing data
Graduate Theses and Dissertations Graduate College 2013 Propensity score adjusted method for missing data Minsun Kim Riddles Iowa State University Follow this and additional works at: http://lib.dr.iastate.edu/etd
More informationNonresponse weighting adjustment using estimated response probability
Nonresponse weighting adjustment using estimated response probability Jae-kwang Kim Yonsei University, Seoul, Korea December 26, 2006 Introduction Nonresponse Unit nonresponse Item nonresponse Basic strategy
More informationthe error term could vary over the observations, in ways that are related
Heteroskedasticity We now consider the implications of relaxing the assumption that the conditional variance Var(u i x i ) = σ 2 is common to all observations i = 1,..., n In many applications, we may
More informationA comparison of stratified simple random sampling and sampling with probability proportional to size
A comparison of stratified simple random sampling and sampling with probability proportional to size Edgar Bueno Dan Hedlin Per Gösta Andersson Department of Statistics Stockholm University Introduction
More informationA comparison of stratified simple random sampling and sampling with probability proportional to size
A comparison of stratified simple random sampling and sampling with probability proportional to size Edgar Bueno Dan Hedlin Per Gösta Andersson 1 Introduction When planning the sampling strategy (i.e.
More informationSTAT5044: Regression and Anova. Inyoung Kim
STAT5044: Regression and Anova Inyoung Kim 2 / 51 Outline 1 Matrix Expression 2 Linear and quadratic forms 3 Properties of quadratic form 4 Properties of estimates 5 Distributional properties 3 / 51 Matrix
More informationRegression and Statistical Inference
Regression and Statistical Inference Walid Mnif wmnif@uwo.ca Department of Applied Mathematics The University of Western Ontario, London, Canada 1 Elements of Probability 2 Elements of Probability CDF&PDF
More informationModelling Non-linear and Non-stationary Time Series
Modelling Non-linear and Non-stationary Time Series Chapter 2: Non-parametric methods Henrik Madsen Advanced Time Series Analysis September 206 Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September
More informationHeteroskedasticity. We now consider the implications of relaxing the assumption that the conditional
Heteroskedasticity We now consider the implications of relaxing the assumption that the conditional variance V (u i x i ) = σ 2 is common to all observations i = 1,..., In many applications, we may suspect
More informationQuantitative Analysis of Financial Markets. Summary of Part II. Key Concepts & Formulas. Christopher Ting. November 11, 2017
Summary of Part II Key Concepts & Formulas Christopher Ting November 11, 2017 christopherting@smu.edu.sg http://www.mysmu.edu/faculty/christophert/ Christopher Ting 1 of 16 Why Regression Analysis? Understand
More informationStatistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach
Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score
More informationMaximum Likelihood (ML) Estimation
Econometrics 2 Fall 2004 Maximum Likelihood (ML) Estimation Heino Bohn Nielsen 1of32 Outline of the Lecture (1) Introduction. (2) ML estimation defined. (3) ExampleI:Binomialtrials. (4) Example II: Linear
More informationJong-Min Kim* and Jon E. Anderson. Statistics Discipline Division of Science and Mathematics University of Minnesota at Morris
Jackknife Variance Estimation of the Regression and Calibration Estimator for Two 2-Phase Samples Jong-Min Kim* and Jon E. Anderson jongmink@morris.umn.edu Statistics Discipline Division of Science and
More informationREPLICATION VARIANCE ESTIMATION FOR THE NATIONAL RESOURCES INVENTORY
REPLICATION VARIANCE ESTIMATION FOR THE NATIONAL RESOURCES INVENTORY J.D. Opsomer, W.A. Fuller and X. Li Iowa State University, Ames, IA 50011, USA 1. Introduction Replication methods are often used in
More informationGraybill Conference Poster Session Introductions
Graybill Conference Poster Session Introductions 2013 Graybill Conference in Modern Survey Statistics Colorado State University Fort Collins, CO June 10, 2013 Small Area Estimation with Incomplete Auxiliary
More informationFractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling
Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling Jae-Kwang Kim 1 Iowa State University June 26, 2013 1 Joint work with Shu Yang Introduction 1 Introduction
More informationGraduate Econometrics I: Asymptotic Theory
Graduate Econometrics I: Asymptotic Theory Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Asymptotic Theory
More informationEmpirical Likelihood Methods for Two-sample Problems with Data Missing-by-Design
1 / 32 Empirical Likelihood Methods for Two-sample Problems with Data Missing-by-Design Changbao Wu Department of Statistics and Actuarial Science University of Waterloo (Joint work with Min Chen and Mary
More informationChapter 5 Matrix Approach to Simple Linear Regression
STAT 525 SPRING 2018 Chapter 5 Matrix Approach to Simple Linear Regression Professor Min Zhang Matrix Collection of elements arranged in rows and columns Elements will be numbers or symbols For example:
More informationNonlinear Signal Processing ELEG 833
Nonlinear Signal Processing ELEG 833 Gonzalo R. Arce Department of Electrical and Computer Engineering University of Delaware arce@ee.udel.edu May 5, 2005 8 MYRIAD SMOOTHERS 8 Myriad Smoothers 8.1 FLOM
More informationChapter 3: Maximum Likelihood Theory
Chapter 3: Maximum Likelihood Theory Florian Pelgrin HEC September-December, 2010 Florian Pelgrin (HEC) Maximum Likelihood Theory September-December, 2010 1 / 40 1 Introduction Example 2 Maximum likelihood
More informationInference for High Dimensional Robust Regression
Department of Statistics UC Berkeley Stanford-Berkeley Joint Colloquium, 2015 Table of Contents 1 Background 2 Main Results 3 OLS: A Motivating Example Table of Contents 1 Background 2 Main Results 3 OLS:
More informationSimple Linear Regression: The Model
Simple Linear Regression: The Model task: quantifying the effect of change X in X on Y, with some constant β 1 : Y = β 1 X, linear relationship between X and Y, however, relationship subject to a random
More informationSome General Types of Tests
Some General Types of Tests We may not be able to find a UMP or UMPU test in a given situation. In that case, we may use test of some general class of tests that often have good asymptotic properties.
More informationEstimation of change in a rotation panel design
Int. Statistical Inst.: Proc. 58th World Statistical Congress, 2011, Dublin (Session CPS028) p.4520 Estimation of change in a rotation panel design Andersson, Claes Statistics Sweden S-701 89 Örebro, Sweden
More informationCombining Non-probability and Probability Survey Samples Through Mass Imputation
Combining Non-probability and Probability Survey Samples Through Mass Imputation Jae-Kwang Kim 1 Iowa State University & KAIST October 27, 2018 1 Joint work with Seho Park, Yilin Chen, and Changbao Wu
More informationWeight calibration and the survey bootstrap
Weight and the survey Department of Statistics University of Missouri-Columbia March 7, 2011 Motivating questions 1 Why are the large scale samples always so complex? 2 Why do I need to use weights? 3
More informationGMM Estimation and Testing
GMM Estimation and Testing Whitney Newey July 2007 Idea: Estimate parameters by setting sample moments to be close to population counterpart. Definitions: β : p 1 parameter vector, with true value β 0.
More informationWeighted Least Squares
Weighted Least Squares The standard linear model assumes that Var(ε i ) = σ 2 for i = 1,..., n. As we have seen, however, there are instances where Var(Y X = x i ) = Var(ε i ) = σ2 w i. Here w 1,..., w
More informationTime Series Analysis
Time Series Analysis hm@imm.dtu.dk Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby 1 Outline of the lecture Regression based methods, 1st part: Introduction (Sec.
More informationHeteroskedasticity. Part VII. Heteroskedasticity
Part VII Heteroskedasticity As of Oct 15, 2015 1 Heteroskedasticity Consequences Heteroskedasticity-robust inference Testing for Heteroskedasticity Weighted Least Squares (WLS) Feasible generalized Least
More informationPart 1.) We know that the probability of any specific x only given p ij = p i p j is just multinomial(n, p) where p k1 k 2
Problem.) I will break this into two parts: () Proving w (m) = p( x (m) X i = x i, X j = x j, p ij = p i p j ). In other words, the probability of a specific table in T x given the row and column counts
More informationIntroduction to Survey Data Integration
Introduction to Survey Data Integration Jae-Kwang Kim Iowa State University May 20, 2014 Outline 1 Introduction 2 Survey Integration Examples 3 Basic Theory for Survey Integration 4 NASS application 5
More informationWeek 3: The EM algorithm
Week 3: The EM algorithm Maneesh Sahani maneesh@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit University College London Term 1, Autumn 2005 Mixtures of Gaussians Data: Y = {y 1... y N } Latent
More information8. Hypothesis Testing
FE661 - Statistical Methods for Financial Engineering 8. Hypothesis Testing Jitkomut Songsiri introduction Wald test likelihood-based tests significance test for linear regression 8-1 Introduction elements
More informationProblem Set #6: OLS. Economics 835: Econometrics. Fall 2012
Problem Set #6: OLS Economics 835: Econometrics Fall 202 A preliminary result Suppose we have a random sample of size n on the scalar random variables (x, y) with finite means, variances, and covariance.
More informationRecent Advances in the analysis of missing data with non-ignorable missingness
Recent Advances in the analysis of missing data with non-ignorable missingness Jae-Kwang Kim Department of Statistics, Iowa State University July 4th, 2014 1 Introduction 2 Full likelihood-based ML estimation
More informationSTAT 100C: Linear models
STAT 100C: Linear models Arash A. Amini April 27, 2018 1 / 1 Table of Contents 2 / 1 Linear Algebra Review Read 3.1 and 3.2 from text. 1. Fundamental subspace (rank-nullity, etc.) Im(X ) = ker(x T ) R
More informationRelaxed linearized algorithms for faster X-ray CT image reconstruction
Relaxed linearized algorithms for faster X-ray CT image reconstruction Hung Nien and Jeffrey A. Fessler University of Michigan, Ann Arbor The 13th Fully 3D Meeting June 2, 2015 1/20 Statistical image reconstruction
More informationWeighted Least Squares
Weighted Least Squares ST 430/514 Recall the linear regression equation E(Y ) = β 0 + β 1 x 1 + β 2 x 2 + + β k x k We have estimated the parameters β 0, β 1, β 2,..., β k by minimizing the sum of squared
More informationQuantile regression and heteroskedasticity
Quantile regression and heteroskedasticity José A. F. Machado J.M.C. Santos Silva June 18, 2013 Abstract This note introduces a wrapper for qreg which reports standard errors and t statistics that are
More informationCombining multiple observational data sources to estimate causal eects
Department of Statistics, North Carolina State University Combining multiple observational data sources to estimate causal eects Shu Yang* syang24@ncsuedu Joint work with Peng Ding UC Berkeley May 23,
More informationMax. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes
Maximum Likelihood Estimation Econometrics II Department of Economics Universidad Carlos III de Madrid Máster Universitario en Desarrollo y Crecimiento Económico Outline 1 3 4 General Approaches to Parameter
More informationModel Assisted Survey Sampling
Carl-Erik Sarndal Jan Wretman Bengt Swensson Model Assisted Survey Sampling Springer Preface v PARTI Principles of Estimation for Finite Populations and Important Sampling Designs CHAPTER 1 Survey Sampling
More informationThe R package sampling, a software tool for training in official statistics and survey sampling
The R package sampling, a software tool for training in official statistics and survey sampling Yves Tillé 1 and Alina Matei 2 1 Institute of Statistics, University of Neuchâtel, Switzerland yves.tille@unine.ch
More informationChapter 9: Hypothesis Testing Sections
Chapter 9: Hypothesis Testing Sections 9.1 Problems of Testing Hypotheses 9.2 Testing Simple Hypotheses 9.3 Uniformly Most Powerful Tests Skip: 9.4 Two-Sided Alternatives 9.6 Comparing the Means of Two
More informationChapter 1: A Brief Review of Maximum Likelihood, GMM, and Numerical Tools. Joan Llull. Microeconometrics IDEA PhD Program
Chapter 1: A Brief Review of Maximum Likelihood, GMM, and Numerical Tools Joan Llull Microeconometrics IDEA PhD Program Maximum Likelihood Chapter 1. A Brief Review of Maximum Likelihood, GMM, and Numerical
More informationChapter 3: Element sampling design: Part 1
Chapter 3: Element sampling design: Part 1 Jae-Kwang Kim Fall, 2014 Simple random sampling 1 Simple random sampling 2 SRS with replacement 3 Systematic sampling Kim Ch. 3: Element sampling design: Part
More informationCSCI5654 (Linear Programming, Fall 2013) Lectures Lectures 10,11 Slide# 1
CSCI5654 (Linear Programming, Fall 2013) Lectures 10-12 Lectures 10,11 Slide# 1 Today s Lecture 1. Introduction to norms: L 1,L 2,L. 2. Casting absolute value and max operators. 3. Norm minimization problems.
More informationEconometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018
Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate
More informationMEI Exam Review. June 7, 2002
MEI Exam Review June 7, 2002 1 Final Exam Revision Notes 1.1 Random Rules and Formulas Linear transformations of random variables. f y (Y ) = f x (X) dx. dg Inverse Proof. (AB)(AB) 1 = I. (B 1 A 1 )(AB)(AB)
More informationWeighting in survey analysis under informative sampling
Jae Kwang Kim and Chris J. Skinner Weighting in survey analysis under informative sampling Article (Accepted version) (Refereed) Original citation: Kim, Jae Kwang and Skinner, Chris J. (2013) Weighting
More informationMultiple Linear Regression for the Supervisor Data
for the Supervisor Data Rating 40 50 60 70 80 90 40 50 60 70 50 60 70 80 90 40 60 80 40 60 80 Complaints Privileges 30 50 70 40 60 Learn Raises 50 70 50 70 90 Critical 40 50 60 70 80 30 40 50 60 70 80
More informationREPLICATION VARIANCE ESTIMATION FOR TWO-PHASE SAMPLES
Statistica Sinica 8(1998), 1153-1164 REPLICATION VARIANCE ESTIMATION FOR TWO-PHASE SAMPLES Wayne A. Fuller Iowa State University Abstract: The estimation of the variance of the regression estimator for
More informationDensity estimation Nonparametric conditional mean estimation Semiparametric conditional mean estimation. Nonparametrics. Gabriel Montes-Rojas
0 0 5 Motivation: Regression discontinuity (Angrist&Pischke) Outcome.5 1 1.5 A. Linear E[Y 0i X i] 0.2.4.6.8 1 X Outcome.5 1 1.5 B. Nonlinear E[Y 0i X i] i 0.2.4.6.8 1 X utcome.5 1 1.5 C. Nonlinearity
More informationGraduate Econometrics Lecture 4: Heteroskedasticity
Graduate Econometrics Lecture 4: Heteroskedasticity Department of Economics University of Gothenburg November 30, 2014 1/43 and Autocorrelation Consequences for OLS Estimator Begin from the linear model
More informationEconomics 582 Random Effects Estimation
Economics 582 Random Effects Estimation Eric Zivot May 29, 2013 Random Effects Model Hence, the model can be re-written as = x 0 β + + [x ] = 0 (no endogeneity) [ x ] = = + x 0 β + + [x ] = 0 [ x ] = 0
More informationModels, Testing, and Correction of Heteroskedasticity. James L. Powell Department of Economics University of California, Berkeley
Models, Testing, and Correction of Heteroskedasticity James L. Powell Department of Economics University of California, Berkeley Aitken s GLS and Weighted LS The Generalized Classical Regression Model
More informationLet us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided
Let us first identify some classes of hypotheses. simple versus simple H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided H 0 : θ θ 0 versus H 1 : θ > θ 0. (2) two-sided; null on extremes H 0 : θ θ 1 or
More informationLecture 6: Discrete Choice: Qualitative Response
Lecture 6: Instructor: Department of Economics Stanford University 2011 Types of Discrete Choice Models Univariate Models Binary: Linear; Probit; Logit; Arctan, etc. Multinomial: Logit; Nested Logit; GEV;
More informationCOS513: FOUNDATIONS OF PROBABILISTIC MODELS LECTURE 9: LINEAR REGRESSION
COS513: FOUNDATIONS OF PROBABILISTIC MODELS LECTURE 9: LINEAR REGRESSION SEAN GERRISH AND CHONG WANG 1. WAYS OF ORGANIZING MODELS In probabilistic modeling, there are several ways of organizing models:
More informationSensitivity of GLS estimators in random effects models
of GLS estimators in random effects models Andrey L. Vasnev (University of Sydney) Tokyo, August 4, 2009 1 / 19 Plan Plan Simulation studies and estimators 2 / 19 Simulation studies Plan Simulation studies
More informationThe outline for Unit 3
The outline for Unit 3 Unit 1. Introduction: The regression model. Unit 2. Estimation principles. Unit 3: Hypothesis testing principles. 3.1 Wald test. 3.2 Lagrange Multiplier. 3.3 Likelihood Ratio Test.
More informationEFFICIENT REPLICATION VARIANCE ESTIMATION FOR TWO-PHASE SAMPLING
Statistica Sinica 13(2003), 641-653 EFFICIENT REPLICATION VARIANCE ESTIMATION FOR TWO-PHASE SAMPLING J. K. Kim and R. R. Sitter Hankuk University of Foreign Studies and Simon Fraser University Abstract:
More informationF. Jay Breidt Colorado State University
Model-assisted survey regression estimation with the lasso 1 F. Jay Breidt Colorado State University Opening Workshop on Computational Methods in Social Sciences SAMSI August 2013 This research was supported
More informationWhat if we want to estimate the mean of w from an SS sample? Let non-overlapping, exhaustive groups, W g : g 1,...G. Random
A Course in Applied Econometrics Lecture 9: tratified ampling 1. The Basic Methodology Typically, with stratified sampling, some segments of the population Jeff Wooldridge IRP Lectures, UW Madison, August
More informationLecture 14 Simple Linear Regression
Lecture 4 Simple Linear Regression Ordinary Least Squares (OLS) Consider the following simple linear regression model where, for each unit i, Y i is the dependent variable (response). X i is the independent
More informationAccounting for Complex Sample Designs via Mixture Models
Accounting for Complex Sample Designs via Finite Normal Mixture Models 1 1 University of Michigan School of Public Health August 2009 Talk Outline 1 2 Accommodating Sampling Weights in Mixture Models 3
More informationMaximum Likelihood Estimation
Maximum Likelihood Estimation Merlise Clyde STA721 Linear Models Duke University August 31, 2017 Outline Topics Likelihood Function Projections Maximum Likelihood Estimates Readings: Christensen Chapter
More informationNew Developments in Econometrics Lecture 9: Stratified Sampling
New Developments in Econometrics Lecture 9: Stratified Sampling Jeff Wooldridge Cemmap Lectures, UCL, June 2009 1. Overview of Stratified Sampling 2. Regression Analysis 3. Clustering and Stratification
More informationUnit roots in vector time series. Scalar autoregression True model: y t 1 y t1 2 y t2 p y tp t Estimated model: y t c y t1 1 y t1 2 y t2
Unit roots in vector time series A. Vector autoregressions with unit roots Scalar autoregression True model: y t y t y t p y tp t Estimated model: y t c y t y t y t p y tp t Results: T j j is asymptotically
More informationLecture 3: More on regularization. Bayesian vs maximum likelihood learning
Lecture 3: More on regularization. Bayesian vs maximum likelihood learning L2 and L1 regularization for linear estimators A Bayesian interpretation of regularization Bayesian vs maximum likelihood fitting
More informationB y t = γ 0 + Γ 1 y t + ε t B(L) y t = γ 0 + ε t ε t iid (0, D) D is diagonal
Structural VAR Modeling for I(1) Data that is Not Cointegrated Assume y t =(y 1t,y 2t ) 0 be I(1) and not cointegrated. That is, y 1t and y 2t are both I(1) and there is no linear combination of y 1t and
More informationSome Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model
Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model 1. Introduction Varying-coefficient partially linear model (Zhang, Lee, and Song, 2002; Xia, Zhang, and Tong, 2004;
More informationThe regression model with one fixed regressor cont d
The regression model with one fixed regressor cont d 3150/4150 Lecture 4 Ragnar Nymoen 27 January 2012 The model with transformed variables Regression with transformed variables I References HGL Ch 2.8
More informationWISE International Masters
WISE International Masters ECONOMETRICS Instructor: Brett Graham INSTRUCTIONS TO STUDENTS 1 The time allowed for this examination paper is 2 hours. 2 This examination paper contains 32 questions. You are
More informationFinite Population Sampling and Inference
Finite Population Sampling and Inference A Prediction Approach RICHARD VALLIANT ALAN H. DORFMAN RICHARD M. ROYALL A Wiley-Interscience Publication JOHN WILEY & SONS, INC. New York Chichester Weinheim Brisbane
More informationIntroductory Econometrics
Based on the textbook by Wooldridge: : A Modern Approach Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna December 11, 2012 Outline Heteroskedasticity
More information