Chapter 4. Replication Variance Estimation. J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 1 / 28
|
|
- Monica Bradford
- 5 years ago
- Views:
Transcription
1 Chapter 4 Replication Variance Estimation J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 1 / 28
2 Jackknife Variance Estimation Create a new sample by deleting one observation n 1 n n ( x (k) x) 2 = x (k) = n x x k n 1 x (k) x = x k x n 1 1 n(n 1) n (x k x) 2 = n 1 sx 2 J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 2 / 28
3 Alternative Jackknife Weights n x (k) ψ = ψx k + (1 ψ) x (k) x (k) ψ x = ψ(x k x) + (1 ψ)( x (k) x) x (k) 1 ψ ψ x = (ψ n 1 )(x k x) = ( nψ 1 n 1 )(x k x) ψ (nψ x)2 1)2 n = (n 1) 2 (x k x) 2 ( x (k) If (nψ 1) 2 = n 1 n n (1 f ), then ( x (k) ψ x)2 = 1 n (1 f )s2 x. J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 3 / 28
4 Random Group Jackknife n = mb : m groups of size b, x 1,..., x m x = 1 m x i m i=1 ˆV ( x) = 1 m 1 m 1 x (k) n = n x b x k n b x (k) b x = n x b x k n b ˆV RGJK ( x) m 1 m Unbiased but d.f. = m 1. m ( x (k) m ( x i x) 2 i=1 = m x x k m 1 x = 1 m 1 ( x k x) b x) 2 = 1 m(m 1) m ( x k x) 2 J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 4 / 28
5 Theorem Theorem F N = {y 1,..., y N } : sequence of finite population y i iid(µ y, σ 2 y ) with finite 4 + δ moments g( ) : continuous function with continuous first derivative at µ y n 1 n n {g(ȳ (k) ) g(ȳ)} 2 = [g (ȳ)] 2 ˆV (ȳ) + o p (n 1 ) where ˆV (ȳ) = n 1 s 2 and g (ȳ) = g(ȳ) ȳ. J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 5 / 28
6 Proof of Theorem By a Taylor linearization, g(ȳ (k) ) = g(ȳ)+g (ȳ k )(ȳ (k) ȳ) = g(ȳ)+g (ȳ)(ȳ (k) ȳ)+r nk (ȳ (k) ȳ) for some ȳ k B δ k (ȳ), δ k = ȳ (k) ȳ where R nk = g (ȳ k ) g (ȳ). Thus, n { g(ȳ (k) ) g(ȳ)} 2 = = n { g (ȳk )} 2 (ȳ (k) ȳ) 2 n { g } 2 (ȳ) + R nk (ȳ (k) ȳ) 2 (i) max 1 k n ȳk ȳ 0 in probability. (ii) max 1 k n g (ȳk ) g (ȳ) 0 in probability. n 1 n n {g(ȳ (k) ) g(ȳ)} 2 = [g (ȳ)] 2 ˆV (ȳ) + o p (n 1 ) J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 6 / 28
7 Proof of (i) Pr[max ȳ (k) ȳ > ɛ] k for some c. max k ȳ (k) ȳ 0 in probability. max k ȳ k ȳ max k ȳ (k) ȳ 0 n Pr[ ȳ (k) ȳ > ɛ] n ɛ 2 E[ ȳ (k) ȳ 2 ] cv (ȳ) 0 J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 7 / 28
8 Proof of (ii) Since g (ȳ) g (µ y ) in probability, it suffices to show that max g (ȳk ) g (µ y ) = o p (1). k Now, note that { Pr max g (ȳk ) g (µ y ) } > ɛ k { +Pr max k Pr{max k ȳk µ y > δ} } ȳk µ y < δ, max g (ȳk ) g (µ y ) > ɛ k holds for any δ > 0. Since g (y) is continuous at y = µ y, we can choose δ such that the second term equals to zero. The first term converges to zero by (i). J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 8 / 28
9 Remainder with Second Derivatives If g( ) has continuous second derivatives at µ y, then n 1 (n 1) n [g(ȳ (k) ) g(ȳ)] 2 = [g (ȳ)] 2 ˆV (ȳ) + O p (n 2 ). Proof : [g(ȳ (k) ) g(ȳ)] 2 = [g (ȳ k )]2 (ȳ (k) ȳ) 2 g (ȳk )2 = [g (ȳ)] 2 + 2[g (ȳk (ȳk k ȳ) [g (ȳk )]2 = [g (ȳ)] 2 + K 1 ȳk ȳ for some K 1. Thus, since ȳ k ȳ = O p(n 1 ), we have the result. J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 9 / 28
10 Jackknife Often Larger than Taylor ˆR = x 1 ȳ ˆR (k) ˆR = [ x (k) ] 1 ȳ (k) x 1 ȳ = [ x (k) ] 1 (y k ˆRx k )(n 1) 1 ˆV JK ( ˆR) = n 1 n ( ˆR (k) ˆR) 2 n 1 n = [ x (k) ] 2 (y k n(n 1) ˆRx k ) 2 vs. ˆVL ( ˆR) = 1 n(n 1) E[( x (k) ) 2 ] E[( x) 2 ] n ( x) 2 (y k ˆRx k ) 2 J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 10 / 28
11 Quantiles ξ p = Q(p) = F 1 (p), p (0, 1) where F (y) is cdf ˆξ p = ˆQ(p) = inf { ˆF (y) p}, p = 0.5 for median y ( ) 1 ˆF (y) = w i w i I (y i y) i A i A To reduce the bias, use interpolation: ˆξ p = ˆξ p0 + ˆξ p1 ˆξ p0 ˆF (ˆξ p1 ) ˆF (ˆξ p0 ) {p ˆF (ˆξ p0 )} where ˆξ p1 = inf x1,...,x n {x; ˆF (x) p} & ˆξ p0 = sup x1,...,x n {x; ˆF (x) p} J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 11 / 28
12 Test Inversion for Quantile C.I. Construct acceptance region for H 0 : p = p 0 {p 0 2[ ˆV (ˆp 0 )] 1/2, p 0 + 2[ ˆV (ˆp 0 )] 1/2 } Invert p-interval to give C.I. for ξ p0 { ˆQ(p 0 2[ ˆV (ˆp 0 )] 1/2 ), ˆQ(p 0 + 2[ ˆV (ˆp 0 )] 1/2 )} J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 12 / 28
13 p ζ p Figure: Plot of CDF J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 13 / 28
14 ζ p p Figure: Inverse CDF J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 14 / 28
15 Bahadur representation Let ˆF (x) be an unbiased estimator of F (x), the population CDF of X. For given p (0, 1), we can define ζ p = F 1 (p) to be the p-th population quantile of X. Let ˆζ p = ˆF 1 (p) be the sample estimator of ζ p using ˆF. Also, define ˆp = ˆF (ζ p ). We are interested in applying a Taylor linearization of ˆζ p around p = ˆp. J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 15 / 28
16 Bahadur representation (Cont d) Note that, by Taylor linearization, ζ p = ˆF 1 (ˆp) = ˆF 1 (p) + d ˆF 1 (p) (ˆp p) dp = ˆF 1 (p) + df 1 (p) (ˆp p) dp = ˆζ p + 1 (ˆp p), f (ζ p ) where the last equality follows from dζ p /dp = (dp/dζ p ) 1 with p = F (ζ p ). Therefore, we have which is first established by Bahadur (1966). ˆζ p = ζp + 1 (p ˆp) (1) f (ζ p ) J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 16 / 28
17 From (1), we can use V (ˆζ p ) = {f (ζ p )} 2 V (ˆp), which requires knowledge of density function f ( ), or we need to estimate it. Define Q(p) = F 1 (p). Since 1 f (ζ p ) = dq(p) = dp Q(p + h) Q(p h), p + h (p h) we can use Q ( ˆp + 2 ) ( V (ˆp) Q ˆp 2 ) V (ˆp) ˆp + 2 V (ˆp) {ˆp 2 V (ˆp)} := ˆγ p to estimate 1/f (ζ p ). This is essentially the idea behind the test inversion method of Woodruff. J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 17 / 28
18 Variance Estimation for Quantiles Bahadur Representation ( ) p(1 p) n(ˆξ p ξ p ) N 0, [F (ξ p )] 2 (SRS) V [F (ˆξ b )]. = ( ) F 2 V (ˆξ p ) ξ ˆQ (p) = [ ˆF (ξ p )] 1 = ˆQ(ˆp + 2 V (ˆp)) ˆQ(ˆp 2 V (ˆp)) (ˆp + 2 V (ˆp)) (ˆp 2 V (ˆp)) ˆV (ˆξ p ) = ˆγ 2 ˆV { ˆF (ξ p )} =: ˆγ J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 18 / 28
19 Variance Estimation for Quantiles (2) 1 Jackknife variance estimator is not consistent. Median ˆθ = 0.5(x m + x m+1 ) for n=2m even ˆV JK = 0.25(n 1)[x m+1 x m ] 2 2 Bootstrap and BRR are O.K. 3 Smoothed quantile ˆξ p = ˆξ p0 + ˆγ[p ˆF (ˆξ p0 )] ˆξ (k) p = ˆξ p0 + ˆγ[p ˆF (k) (ˆξ p0 )] ˆV JK (ˆξ p ) = k c k (ˆξ (k) p ˆξ p ) 2 J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 19 / 28
20 Two-Phase Samples ȳ 2p,reg = ȳ 2p,REE = ȳ 2π + ( x 1π x 2π ) ˆβ 2 ˆβ 2 = w 2i (x i x 2π ) (x i x 2π ) i A 2 w 1 2i = π 1i π 2i 1i, w 1i = π 1 1i 1 i A 2 w 2i (x i x 2π ) y i ˆV JK (ȳ 2p,REE ) = L [ ] c k ȳ (k) 2 2p,REE ȳ 2p,REE J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 20 / 28
21 Two-Phase Samples where w (k) 2i ˆβ (k) 2 = = (w (k) 1i )(π 1 2i 1i ) x (k) 1π = ( x (k) 2π, ȳ (k) 2π ) = w (k) i A1 1i w (k) i A2 w (k) 2i (x i x (k) 2π ) (x i x (k) i A2 2i 2π ) 1 1 i A1 w (k) 1i x i w (k) 2i (x i, y i ) i A2 1 w (k) 2i (x i x (k) 2π ) (y i ȳ (k) 2π ) i A 2 J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 21 / 28
22 Theorem Kim, Navarro, Fuller (2006 JASA), Assumptions (i) Second phase is stratified with π 2i 1i constant within group (ii) K L < Nn 1 π 1i < K U for some K L &K U (iii) V { ˆT 1y F} K M V { ˆT 1y,SRS } where ˆT 1y = i A 1 π 1 1i y i (iv) nv { ˆT 1y F} = N N i=1 j=1 Ω ijy i y j where N i=1 Ω ij = O(N 1 ) ( ) 2 (v) E ˆV 1 ( ˆT 1y ) F V ( ˆT 1y F) 1 = o(1) (vi) E{[c k ( ˆT (k) 1y ˆT 1y ) 2 ] 2 F} < K L L 2 [V ( ˆT 1y )] 2 J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 22 / 28
23 Theorem Kim, Navarro, Fuller (2006 JASA), Result ˆV {ȳ 2p,reg } = V (ȳ 2p,reg F) 1 N 2 where κ 2i = π 2i 1i N i=1 κ 1 2i (1 κ 2i )e 2 i + o p (n 1 ) e i = y i ȳ N (x i x)β N [ N ] 1 N β N = (x i x N ) (x i x N ) (x i x N ) y i. i=1 i=1 J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 23 / 28
24 Sketched Proof Extend definition of a i : { 1 if i A2, if i A a i = 1 0 if i / A 2, if i A 1 a i defined throughout the population. Writing ȳ 2p,reg = z 1π + ( x 1π ū 1π ) ˆβ where ( z 1π, ū 1π ) = ( i A 1 a i π 1 2i ) 1 i A 1 π 1 2i (z i, u i ), z i = a i y i, u i = a i x i : conditional on a = (a 1,, a N ), ȳ 2p,reg is a smooth function of ( z 1π, x 1π, ū 1π, ˆβ) ˆβ = π 1 2i (u i ū 1π ) (u i ū 1π ) i A1 1 i A 1 π 1 2i (u i ū 1π ) z i. J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 24 / 28
25 Sketched Proof (Cont d) Variance of ȳ 2p,reg : V [ȳ 2p,reg F] = E[V (ȳ 2p,reg a, F) F] + V [E(ȳ 2p,reg a, F) F] Replication variance estimation: L ȳ (k) 2p,reg = z (k) (k) 1π + ( x 1π ū(k) 1π ) ˆβ (k) c k (ȳ (k) 2p,reg ȳ 2p,reg ) 2 = E[V (ȳ 2p,reg a, F) F] + o p (n 1 ) ˆV (ȳ reg ) = L c k (ȳ (k) 2p,reg ȳ 2p,reg ) 2 where ˆV (ȳ 2p,reg a) is used to emphasize that it is a function of a. =: ˆV (ȳ 2p,reg a), J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 25 / 28
26 Sketched Proof (Cont d) 1 By condition (v), ˆV (ȳ 2p,reg a) = V [ȳ 2p,reg a, F] + o p (n 1 ) conditional on a, F. 2 Show ˆV (ȳ 2p,reg a) = E{V [ [ȳ 2p,reg a, F] F} ] + o p (n 1 ) ˆV (ȳ2p,reg a) Need to show that V = o(1) V (ȳ 2p,reg a, F) 3 ˆV (ȳ 2p,reg a) = V (ȳ 2p,reg F) V (E(ȳ 2p,reg a, F) F) + o p (n 1 ȳ 2p,reg = ȳ 2π + ( x 1π x 2π ) ˆβ = ȳ 2π + ( x 1π x 2π )β N + O p (n 1 ) = ē 2π + x 1π β N + O p (n 1 ) E[ȳ 2p,reg a, F] = 1 N κ 1 2i a i e i + x 1π β N N V [E[ȳ 2p,reg a, F] F] = 1 i=1 N κ 1 2i (1 κ 2i )e 2 i N 2 J. Kim, W. Fuller (ISU) Chapter i=1 4 7/31/11 26 / 28
27 Remark (i) The bias term of ˆV {ȳ 2p,reg } is of order O(N 1 ) and can be negligible if n/n is negligible. (ii) The bias term can be unbiasedly estimated by N 2 i A 2 w 1i π 2 2i 1i (1 π 2i 1i)ê 2 i. (iii) For some designs and replication procedures, it is possible to modify the weights to reduce the bias. (See the textbook.) J. Kim, W. Fuller (ISU) Chapter 4 7/31/11 27 / 28
Chapter 8: Estimation 1
Chapter 8: Estimation 1 Jae-Kwang Kim Iowa State University Fall, 2014 Kim (ISU) Ch. 8: Estimation 1 Fall, 2014 1 / 33 Introduction 1 Introduction 2 Ratio estimation 3 Regression estimator Kim (ISU) Ch.
More informationChapter 5: Models used in conjunction with sampling. J. Kim, W. Fuller (ISU) Chapter 5: Models used in conjunction with sampling 1 / 70
Chapter 5: Models used in conjunction with sampling J. Kim, W. Fuller (ISU) Chapter 5: Models used in conjunction with sampling 1 / 70 Nonresponse Unit Nonresponse: weight adjustment Item Nonresponse:
More informationChapter 4: Imputation
Chapter 4: Imputation Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Basic Theory for imputation 3 Variance estimation after imputation 4 Replication variance estimation
More informationPAijpam.eu NEW H 1 (Ω) CONFORMING FINITE ELEMENTS ON HEXAHEDRA
International Journal of Pure and Applied Mathematics Volume 109 No. 3 2016, 609-617 ISSN: 1311-8080 (printed version); ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu doi: 10.12732/ijpam.v109i3.10
More informationCombining data from two independent surveys: model-assisted approach
Combining data from two independent surveys: model-assisted approach Jae Kwang Kim 1 Iowa State University January 20, 2012 1 Joint work with J.N.K. Rao, Carleton University Reference Kim, J.K. and Rao,
More informationA Resampling Method on Pivotal Estimating Functions
A Resampling Method on Pivotal Estimating Functions Kun Nie Biostat 277,Winter 2004 March 17, 2004 Outline Introduction A General Resampling Method Examples - Quantile Regression -Rank Regression -Simulation
More informationCluster Sampling 2. Chapter Introduction
Chapter 7 Cluster Sampling 7.1 Introduction In this chapter, we consider two-stage cluster sampling where the sample clusters are selected in the first stage and the sample elements are selected in the
More informationAn Efficient Estimation Method for Longitudinal Surveys with Monotone Missing Data
An Efficient Estimation Method for Longitudinal Surveys with Monotone Missing Data Jae-Kwang Kim 1 Iowa State University June 28, 2012 1 Joint work with Dr. Ming Zhou (when he was a PhD student at ISU)
More informationAdvanced Topics in Survey Sampling
Advanced Topics in Survey Sampling Jae-Kwang Kim Wayne A Fuller Pushpal Mukhopadhyay Department of Statistics Iowa State University World Statistics Congress Short Course July 23-24, 2015 Kim & Fuller
More informationChapter 2. Section Section 2.9. J. Kim (ISU) Chapter 2 1 / 26. Design-optimal estimator under stratified random sampling
Chapter 2 Section 2.4 - Section 2.9 J. Kim (ISU) Chapter 2 1 / 26 2.4 Regression and stratification Design-optimal estimator under stratified random sampling where (Ŝxxh, Ŝxyh) ˆβ opt = ( x st, ȳ st )
More informationBahadur representations for bootstrap quantiles 1
Bahadur representations for bootstrap quantiles 1 Yijun Zuo Department of Statistics and Probability, Michigan State University East Lansing, MI 48824, USA zuo@msu.edu 1 Research partially supported by
More informationLocal Polynomial Regression
VI Local Polynomial Regression (1) Global polynomial regression We observe random pairs (X 1, Y 1 ),, (X n, Y n ) where (X 1, Y 1 ),, (X n, Y n ) iid (X, Y ). We want to estimate m(x) = E(Y X = x) based
More informationA measurement error model approach to small area estimation
A measurement error model approach to small area estimation Jae-kwang Kim 1 Spring, 2015 1 Joint work with Seunghwan Park and Seoyoung Kim Ouline Introduction Basic Theory Application to Korean LFS Discussion
More informationBootstrap inference for the finite population total under complex sampling designs
Bootstrap inference for the finite population total under complex sampling designs Zhonglei Wang (Joint work with Dr. Jae Kwang Kim) Center for Survey Statistics and Methodology Iowa State University Jan.
More informationAsymmetric least squares estimation and testing
Asymmetric least squares estimation and testing Whitney Newey and James Powell Princeton University and University of Wisconsin-Madison January 27, 2012 Outline ALS estimators Large sample properties Asymptotic
More informationData Integration for Big Data Analysis for finite population inference
for Big Data Analysis for finite population inference Jae-kwang Kim ISU January 23, 2018 1 / 36 What is big data? 2 / 36 Data do not speak for themselves Knowledge Reproducibility Information Intepretation
More informationImputation for Missing Data under PPSWR Sampling
July 5, 2010 Beijing Imputation for Missing Data under PPSWR Sampling Guohua Zou Academy of Mathematics and Systems Science Chinese Academy of Sciences 1 23 () Outline () Imputation method under PPSWR
More informationIEOR E4703: Monte-Carlo Simulation
IEOR E4703: Monte-Carlo Simulation Output Analysis for Monte-Carlo Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com Output Analysis
More informationTwo-phase sampling approach to fractional hot deck imputation
Two-phase sampling approach to fractional hot deck imputation Jongho Im 1, Jae-Kwang Kim 1 and Wayne A. Fuller 1 Abstract Hot deck imputation is popular for handling item nonresponse in survey sampling.
More informationCombining Non-probability and Probability Survey Samples Through Mass Imputation
Combining Non-probability and Probability Survey Samples Through Mass Imputation Jae-Kwang Kim 1 Iowa State University & KAIST October 27, 2018 1 Joint work with Seho Park, Yilin Chen, and Changbao Wu
More informationChapter 7. Hypothesis Testing
Chapter 7. Hypothesis Testing Joonpyo Kim June 24, 2017 Joonpyo Kim Ch7 June 24, 2017 1 / 63 Basic Concepts of Testing Suppose that our interest centers on a random variable X which has density function
More informationBasis Penalty Smoothers. Simon Wood Mathematical Sciences, University of Bath, U.K.
Basis Penalty Smoothers Simon Wood Mathematical Sciences, University of Bath, U.K. Estimating functions It is sometimes useful to estimate smooth functions from data, without being too precise about the
More informationInference on distributions and quantiles using a finite-sample Dirichlet process
Dirichlet IDEAL Theory/methods Simulations Inference on distributions and quantiles using a finite-sample Dirichlet process David M. Kaplan University of Missouri Matt Goldman UC San Diego Midwest Econometrics
More information14 Multiple Linear Regression
B.Sc./Cert./M.Sc. Qualif. - Statistics: Theory and Practice 14 Multiple Linear Regression 14.1 The multiple linear regression model In simple linear regression, the response variable y is expressed in
More information6. Fractional Imputation in Survey Sampling
6. Fractional Imputation in Survey Sampling 1 Introduction Consider a finite population of N units identified by a set of indices U = {1, 2,, N} with N known. Associated with each unit i in the population
More informationVariance Estimation for Calibration to Estimated Control Totals
Variance Estimation for Calibration to Estimated Control Totals Siyu Qing Coauthor with Michael D. Larsen Associate Professor of Statistics Tuesday, 11/05/2013 2 Outline A. Background B. Calibration Technique
More informationPh.D. Qualifying Exam Friday Saturday, January 6 7, 2017
Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a
More informationChapter 3: Element sampling design: Part 1
Chapter 3: Element sampling design: Part 1 Jae-Kwang Kim Fall, 2014 Simple random sampling 1 Simple random sampling 2 SRS with replacement 3 Systematic sampling Kim Ch. 3: Element sampling design: Part
More informationEstimation of Parameters and Variance
Estimation of Parameters and Variance Dr. A.C. Kulshreshtha U.N. Statistical Institute for Asia and the Pacific (SIAP) Second RAP Regional Workshop on Building Training Resources for Improving Agricultural
More informationVARIANCE ESTIMATION FOR NEAREST NEIGHBOR IMPUTATION FOR U.S. CENSUS LONG FORM DATA
Submitted to the Annals of Applied Statistics VARIANCE ESTIMATION FOR NEAREST NEIGHBOR IMPUTATION FOR U.S. CENSUS LONG FORM DATA By Jae Kwang Kim, Wayne A. Fuller and William R. Bell Iowa State University
More informationIntroduction to Survey Data Integration
Introduction to Survey Data Integration Jae-Kwang Kim Iowa State University May 20, 2014 Outline 1 Introduction 2 Survey Integration Examples 3 Basic Theory for Survey Integration 4 NASS application 5
More informationPaper Review: Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties by Jianqing Fan and Runze Li (2001)
Paper Review: Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties by Jianqing Fan and Runze Li (2001) Presented by Yang Zhao March 5, 2010 1 / 36 Outlines 2 / 36 Motivation
More informationPh.D. Qualifying Exam Friday Saturday, January 3 4, 2014
Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Put your solution to each problem on a separate sheet of paper. Problem 1. (5166) Assume that two random samples {x i } and {y i } are independently
More informationREPLICATION VARIANCE ESTIMATION FOR TWO-PHASE SAMPLES
Statistica Sinica 8(1998), 1153-1164 REPLICATION VARIANCE ESTIMATION FOR TWO-PHASE SAMPLES Wayne A. Fuller Iowa State University Abstract: The estimation of the variance of the regression estimator for
More informationNonparametric Regression. Badr Missaoui
Badr Missaoui Outline Kernel and local polynomial regression. Penalized regression. We are given n pairs of observations (X 1, Y 1 ),...,(X n, Y n ) where Y i = r(x i ) + ε i, i = 1,..., n and r(x) = E(Y
More information5.3 LINEARIZATION METHOD. Linearization Method for a Nonlinear Estimator
Linearization Method 141 properties that cover the most common types of complex sampling designs nonlinear estimators Approximative variance estimators can be used for variance estimation of a nonlinear
More informationChapter 7: Model Assessment and Selection
Chapter 7: Model Assessment and Selection DD3364 April 20, 2012 Introduction Regression: Review of our problem Have target variable Y to estimate from a vector of inputs X. A prediction model ˆf(X) has
More informationJong-Min Kim* and Jon E. Anderson. Statistics Discipline Division of Science and Mathematics University of Minnesota at Morris
Jackknife Variance Estimation for Two Samples after Imputation under Two-Phase Sampling Jong-Min Kim* and Jon E. Anderson jongmink@mrs.umn.edu Statistics Discipline Division of Science and Mathematics
More informationSupplement to Quantile-Based Nonparametric Inference for First-Price Auctions
Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions Vadim Marmer University of British Columbia Artyom Shneyerov CIRANO, CIREQ, and Concordia University August 30, 2010 Abstract
More informationCh 3: Multiple Linear Regression
Ch 3: Multiple Linear Regression 1. Multiple Linear Regression Model Multiple regression model has more than one regressor. For example, we have one response variable and two regressor variables: 1. delivery
More informationNon-parametric bootstrap mean squared error estimation for M-quantile estimates of small area means, quantiles and poverty indicators
Non-parametric bootstrap mean squared error estimation for M-quantile estimates of small area means, quantiles and poverty indicators Stefano Marchetti 1 Nikos Tzavidis 2 Monica Pratesi 3 1,3 Department
More informationL-statistics for Repeated Measurements Data With Application to. Trimmed Means, Quantiles and Tolerance Intervals
L-statistics for Repeated Measurements Data With Application to Trimmed Means, Quantiles and Tolerance Intervals Houssein I. Assaad Department of Statistics Texas A&M University College Station, TX 77843-3143,
More informationA Practitioner s Guide to Cluster-Robust Inference
A Practitioner s Guide to Cluster-Robust Inference A. C. Cameron and D. L. Miller presented by Federico Curci March 4, 2015 Cameron Miller Cluster Clinic II March 4, 2015 1 / 20 In the previous episode
More informationCoefficient of Determination
Coefficient of Determination ST 430/514 The coefficient of determination, R 2, is defined as before: R 2 = 1 SS E (yi ŷ i ) = 1 2 SS yy (yi ȳ) 2 The interpretation of R 2 is still the fraction of variance
More informationReliable Inference in Conditions of Extreme Events. Adriana Cornea
Reliable Inference in Conditions of Extreme Events by Adriana Cornea University of Exeter Business School Department of Economics ExISta Early Career Event October 17, 2012 Outline of the talk Extreme
More informationEFFICIENT REPLICATION VARIANCE ESTIMATION FOR TWO-PHASE SAMPLING
Statistica Sinica 13(2003), 641-653 EFFICIENT REPLICATION VARIANCE ESTIMATION FOR TWO-PHASE SAMPLING J. K. Kim and R. R. Sitter Hankuk University of Foreign Studies and Simon Fraser University Abstract:
More informationResampling and the Bootstrap
Resampling and the Bootstrap Axel Benner Biostatistics, German Cancer Research Center INF 280, D-69120 Heidelberg benner@dkfz.de Resampling and the Bootstrap 2 Topics Estimation and Statistical Testing
More informationGoodness-of-fit tests for the cure rate in a mixture cure model
Biometrika (217), 13, 1, pp. 1 7 Printed in Great Britain Advance Access publication on 31 July 216 Goodness-of-fit tests for the cure rate in a mixture cure model BY U.U. MÜLLER Department of Statistics,
More information1 General problem. 2 Terminalogy. Estimation. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ).
Estimation February 3, 206 Debdeep Pati General problem Model: {P θ : θ Θ}. Observe X P θ, θ Θ unknown. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ). Examples: θ = (µ,
More informationITERATED BOOTSTRAP PREDICTION INTERVALS
Statistica Sinica 8(1998), 489-504 ITERATED BOOTSTRAP PREDICTION INTERVALS Majid Mojirsheibani Carleton University Abstract: In this article we study the effects of bootstrap iteration as a method to improve
More informationAdaptive Piecewise Polynomial Estimation via Trend Filtering
Adaptive Piecewise Polynomial Estimation via Trend Filtering Liubo Li, ShanShan Tu The Ohio State University li.2201@osu.edu, tu.162@osu.edu October 1, 2015 Liubo Li, ShanShan Tu (OSU) Trend Filtering
More informationMonte Carlo Study on the Successive Difference Replication Method for Non-Linear Statistics
Monte Carlo Study on the Successive Difference Replication Method for Non-Linear Statistics Amang S. Sukasih, Mathematica Policy Research, Inc. Donsig Jang, Mathematica Policy Research, Inc. Amang S. Sukasih,
More informationFractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling
Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling Jae-Kwang Kim 1 Iowa State University June 26, 2013 1 Joint work with Shu Yang Introduction 1 Introduction
More informationBOOTSTRAPPING SAMPLE QUANTILES BASED ON COMPLEX SURVEY DATA UNDER HOT DECK IMPUTATION
Statistica Sinica 8(998), 07-085 BOOTSTRAPPING SAMPLE QUANTILES BASED ON COMPLEX SURVEY DATA UNDER HOT DECK IMPUTATION Jun Shao and Yinzhong Chen University of Wisconsin-Madison Abstract: The bootstrap
More informationInference based on robust estimators Part 2
Inference based on robust estimators Part 2 Matias Salibian-Barrera 1 Department of Statistics University of British Columbia ECARES - Dec 2007 Matias Salibian-Barrera (UBC) Robust inference (2) ECARES
More informationSlides 12: Output Analysis for a Single Model
Slides 12: Output Analysis for a Single Model Objective: Estimate system performance via simulation. If θ is the system performance, the precision of the estimator ˆθ can be measured by: The standard error
More informationLecture 6 Multiple Linear Regression, cont.
Lecture 6 Multiple Linear Regression, cont. BIOST 515 January 22, 2004 BIOST 515, Lecture 6 Testing general linear hypotheses Suppose we are interested in testing linear combinations of the regression
More informationExperimental variogram of the residuals in the universal kriging (UK) model
Experimental variogram of the residuals in the universal kriging (UK) model Nicolas DESASSIS Didier RENARD Technical report R141210NDES MINES ParisTech Centre de Géosciences Equipe Géostatistique 35, rue
More informationWeighting in survey analysis under informative sampling
Jae Kwang Kim and Chris J. Skinner Weighting in survey analysis under informative sampling Article (Accepted version) (Refereed) Original citation: Kim, Jae Kwang and Skinner, Chris J. (2013) Weighting
More informationChemistry 532 Problem Set 7 Spring 2012 Solutions
Chemistry 53 Problem Set 7 Spring 01 Solutions 1. The study of the time-independent Schrödinger equation for a one-dimensional particle subject to the potential function leads to the differential equation
More informationSTAT5044: Regression and Anova
STAT5044: Regression and Anova Inyoung Kim 1 / 49 Outline 1 How to check assumptions 2 / 49 Assumption Linearity: scatter plot, residual plot Randomness: Run test, Durbin-Watson test when the data can
More informationMultiple Linear Regression
Multiple Linear Regression ST 430/514 Recall: a regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates).
More informationSmall Domains Estimation and Poverty Indicators. Carleton University, Ottawa, Canada
Small Domains Estimation and Poverty Indicators J. N. K. Rao Carleton University, Ottawa, Canada Invited paper for presentation at the International Seminar Population Estimates and Projections: Methodologies,
More informationNonconcave Penalized Likelihood with A Diverging Number of Parameters
Nonconcave Penalized Likelihood with A Diverging Number of Parameters Jianqing Fan and Heng Peng Presenter: Jiale Xu March 12, 2010 Jianqing Fan and Heng Peng Presenter: JialeNonconcave Xu () Penalized
More informationHigher-Order Confidence Intervals for Stochastic Programming using Bootstrapping
Preprint ANL/MCS-P1964-1011 Noname manuscript No. (will be inserted by the editor) Higher-Order Confidence Intervals for Stochastic Programming using Bootstrapping Mihai Anitescu Cosmin G. Petra Received:
More informationInference For High Dimensional M-estimates: Fixed Design Results
Inference For High Dimensional M-estimates: Fixed Design Results Lihua Lei, Peter Bickel and Noureddine El Karoui Department of Statistics, UC Berkeley Berkeley-Stanford Econometrics Jamboree, 2017 1/49
More informationSTAT 512 sp 2018 Summary Sheet
STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}
More informationIntroduction The framework Bias and variance Approximate computation of leverage Empirical evaluation Discussion of sampling approach in big data
Discussion of sampling approach in big data Big data discussion group at MSCS of UIC Outline 1 Introduction 2 The framework 3 Bias and variance 4 Approximate computation of leverage 5 Empirical evaluation
More informationLecture 13: Subsampling vs Bootstrap. Dimitris N. Politis, Joseph P. Romano, Michael Wolf
Lecture 13: 2011 Bootstrap ) R n x n, θ P)) = τ n ˆθn θ P) Example: ˆθn = X n, τ n = n, θ = EX = µ P) ˆθ = min X n, τ n = n, θ P) = sup{x : F x) 0} ) Define: J n P), the distribution of τ n ˆθ n θ P) under
More informationOn the asymptotic normality and variance estimation of nondifferentiable survey estimators
On the asymptotic normality and variance estimation of nondifferentiable survey estimators Jianqiang.C Wang Visiting Assistant Professor, Department of Statistics Colorado State University Jean Opsomer
More informationStatistics 135 Fall 2008 Final Exam
Name: SID: Statistics 135 Fall 2008 Final Exam Show your work. The number of points each question is worth is shown at the beginning of the question. There are 10 problems. 1. [2] The normal equations
More informationRecent Advances in the analysis of missing data with non-ignorable missingness
Recent Advances in the analysis of missing data with non-ignorable missingness Jae-Kwang Kim Department of Statistics, Iowa State University July 4th, 2014 1 Introduction 2 Full likelihood-based ML estimation
More informationModel-free prediction intervals for regression and autoregression. Dimitris N. Politis University of California, San Diego
Model-free prediction intervals for regression and autoregression Dimitris N. Politis University of California, San Diego To explain or to predict? Models are indispensable for exploring/utilizing relationships
More informationEstimating Estimable Functions of β. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 17
Estimating Estimable Functions of β Copyright c 202 Dan Nettleton (Iowa State University) Statistics 5 / 7 The Response Depends on β Only through Xβ In the Gauss-Markov or Normal Theory Gauss-Markov Linear
More informationParametric fractional imputation for missing data analysis
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 Biometrika (????),??,?, pp. 1 15 C???? Biometrika Trust Printed in
More informationBANA 7046 Data Mining I Lecture 2. Linear Regression, Model Assessment, and Cross-validation 1
BANA 7046 Data Mining I Lecture 2. Linear Regression, Model Assessment, and Cross-validation 1 Shaobo Li University of Cincinnati 1 Partially based on Hastie, et al. (2009) ESL, and James, et al. (2013)
More information4 Multiple Linear Regression
4 Multiple Linear Regression 4. The Model Definition 4.. random variable Y fits a Multiple Linear Regression Model, iff there exist β, β,..., β k R so that for all (x, x 2,..., x k ) R k where ε N (, σ
More informationNonresponse weighting adjustment using estimated response probability
Nonresponse weighting adjustment using estimated response probability Jae-kwang Kim Yonsei University, Seoul, Korea December 26, 2006 Introduction Nonresponse Unit nonresponse Item nonresponse Basic strategy
More informationSimple and Multiple Linear Regression
Sta. 113 Chapter 12 and 13 of Devore March 12, 2010 Table of contents 1 Simple Linear Regression 2 Model Simple Linear Regression A simple linear regression model is given by Y = β 0 + β 1 x + ɛ where
More informationUNIVERSITY OF WISCONSIN DEPARTMENT OF BIOSTATISTICS AND MEDICAL INFORMATICS
UNIVERSITY OF WISCONSIN DEPARTMENT OF BIOSTATISTICS AND MEDICAL INFORMATICS Technical Report August, 25 # 19 INFERENCE UNDER RIGHT CENSORING FOR TRANSFORMATION MODELS WITH A CHANGE- POINT BASED ON A COVARIATE
More informationRecap. HW due Thursday by 5 pm Next HW coming on Thursday Logistic regression: Pr(G = k X) linear on the logit scale Linear discriminant analysis:
1 / 23 Recap HW due Thursday by 5 pm Next HW coming on Thursday Logistic regression: Pr(G = k X) linear on the logit scale Linear discriminant analysis: Pr(G = k X) Pr(X G = k)pr(g = k) Theory: LDA more
More informationlarge number of i.i.d. observations from P. For concreteness, suppose
1 Subsampling Suppose X i, i = 1,..., n is an i.i.d. sequence of random variables with distribution P. Let θ(p ) be some real-valued parameter of interest, and let ˆθ n = ˆθ n (X 1,..., X n ) be some estimate
More information1. Variance stabilizing transformations; Box-Cox Transformations - Section. 2. Transformations to linearize the model - Section 5.
Ch. 5: Transformations and Weighting 1. Variance stabilizing transformations; Box-Cox Transformations - Section 5.2; 5.4 2. Transformations to linearize the model - Section 5.3 3. Weighted regression -
More informationESTIMATION AND OUTPUT ANALYSIS (L&K Chapters 9, 10) Review performance measures (means, probabilities and quantiles).
ESTIMATION AND OUTPUT ANALYSIS (L&K Chapters 9, 10) Set up standard example and notation. Review performance measures (means, probabilities and quantiles). A framework for conducting simulation experiments
More informationThe Finite Element Method
The Finite Element Method 3D Problems Heat Transfer and Elasticity Read: Chapter 14 CONTENTS Finite element models of 3-D Heat Transfer Finite element model of 3-D Elasticity Typical 3-D Finite Elements
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationCh 2: Simple Linear Regression
Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component
More informationCentral Limit Theorem ( 5.3)
Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately
More informationIt should be unbiased, or approximately unbiased. Variance of the variance estimator should be small. That is, the variance estimator is stable.
Chapter 10 Variace Estimatio 10.1 Itroductio Variace estimatio is a importat practical problem i survey samplig. Variace estimates are used i two purposes. Oe is the aalytic purpose such as costructig
More informationNonparametric Small Area Estimation Using Penalized Spline Regression
Nonparametric Small Area Estimation Using Penalized Spline Regression 0verview Spline-based nonparametric regression Nonparametric small area estimation Prediction mean squared error Bootstrapping small
More informationBootstrap, Jackknife and other resampling methods
Bootstrap, Jackknife and other resampling methods Part III: Parametric Bootstrap Rozenn Dahyot Room 128, Department of Statistics Trinity College Dublin, Ireland dahyot@mee.tcd.ie 2005 R. Dahyot (TCD)
More informationSpring 2012 Math 541B Exam 1
Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote
More informationAccounting for Population Uncertainty in Covariance Structure Analysis
Accounting for Population Uncertainty in Structure Analysis Boston College May 21, 2013 Joint work with: Michael W. Browne The Ohio State University matrix among observed variables are usually implied
More information41903: Introduction to Nonparametrics
41903: Notes 5 Introduction Nonparametrics fundamentally about fitting flexible models: want model that is flexible enough to accommodate important patterns but not so flexible it overspecializes to specific
More informationQualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama
Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama Instructions This exam has 7 pages in total, numbered 1 to 7. Make sure your exam has all the pages. This exam will be 2 hours
More informationThe Nonparametric Bootstrap
The Nonparametric Bootstrap The nonparametric bootstrap may involve inferences about a parameter, but we use a nonparametric procedure in approximating the parametric distribution using the ECDF. We use
More informationEstimation of Complex Small Area Parameters with Application to Poverty Indicators
1 Estimation of Complex Small Area Parameters with Application to Poverty Indicators J.N.K. Rao School of Mathematics and Statistics, Carleton University (Joint work with Isabel Molina from Universidad
More informationEmpirical Likelihood Methods for Two-sample Problems with Data Missing-by-Design
1 / 32 Empirical Likelihood Methods for Two-sample Problems with Data Missing-by-Design Changbao Wu Department of Statistics and Actuarial Science University of Waterloo (Joint work with Min Chen and Mary
More informationProceedings of the 2014 Winter Simulation Conference A. Tolk, S. Y. Diallo, I. O. Ryzhov, L. Yilmaz, S. Buckley, and J. A. Miller, eds.
Proceedings of the 2014 Winter Simulation Conference A. Tolk, S. Y. Diallo, I. O. Ryzhov, L. Yilmaz, S. Buckley, and J. A. Miller, eds. CONSTRUCTING CONFIDENCE INTERVALS FOR A QUANTILE USING BATCHING AND
More informationAsymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands
Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands Elizabeth C. Mannshardt-Shamseldin Advisor: Richard L. Smith Duke University Department
More information