Nonparametric Estimation of Regression Functions In the Presence of Irrelevant Regressors

Size: px
Start display at page:

Download "Nonparametric Estimation of Regression Functions In the Presence of Irrelevant Regressors"

Transcription

1 Nonparametric Estimation of Regression Functions In the Presence of Irrelevant Regressors Peter Hall, Qi Li, Jeff Racine 1 Introduction Nonparametric techniques robust to functional form specification. curse of dimensionality and the number of discrete cells. The conventional frequency estimation method. Treatment effect, Hahn (1998), Hirano, Imbens and Ridder (2003). Smooth the discrete variables, Aitchison and Aitken (1976). Bandwidth selection: Data-driven methods. Irrelevant covariates (regressors) should be removed. Testing procedure not powerful and not efficient. The data-driven cross-validation method. Automatic removal of irrelevant variables. 1

2 2 Regression Models with Irrelevant Regressors X d i, q 1 vector of discrete variables. Xi c Rp continuous regressors. Xis d, the sth component of Xd i, Xis d takes c s 2 different values. X i = (Xi d, Xc i ), We want to estimate E(Y i X i ) The first p 1 (1 p 1 p) components of X c and the first q 1 (0 q 1 q) components of X d are relevant regressors. Let X be relevant components of X and let X = X \ { X} denote the irrelevant components of X. We assume that ( X, Y ) is independent of X (2.1) (2.1) implies that E(Y X) = E(Y X), (2.2) So that Y = g(x) + error, Y i = ḡ( X i ) + u i, (2.3) E(u i X i ) = 0. 2

3 For 1 s q, define l(x d is, x d s, λ s ) = { 1 if X d is = x d s, λ s if Xis d xd s. (2.4) The product kernel for x d = (x d 1,..., x d q) is given by K d (x d, X d i ) = q l(xis, d x d s, λ s ). s=1 0 λ s 1. When λ s = 1, K d (x d, X d i ) is unrelated to (xd s, X d is ) x d s is smoothed out. For the continuous variables x c = (x c 1,..., x c p), the product kernel is K c (x c, X c i ) = p s=1 1 h s K ( x c s X c is h s where K is a symmetric, univariate density function, 0 < h s <. ), Note that when h s =, x c s is smoothed out. K(x, X i ) = K c (x c, Xi c)kd (x d, Xi d ). We estimate E(Y X = x) by ĝ(x) = n 1 n i=1 Y ik(x, X i ) n 1 n i=1 K(x, X i). (2.5) We choose (h, λ) = (h 1,..., h p, λ 1,..., λ q ) by minimizing the crossvalidation function given by CV (h, λ) = 1 n (Y i ĝ i (X i )) 2 w(x i ), (2.6) n i=1 3

4 where ĝ i (X i ) = n j i Y jk(x i, X j )/ n j i K(X i, X j ) is the leave-one-out kernel estimator of E(Y i X i ), and w(.) is a weight function. Define σ 2 ( x) = E(u 2 i X i = x) and let S c denote the support of w. We also assume that g, f and σ 2 have two continuous derivatives; w is continuous, nonnegative and has compact support; f is bounded away from zero for x = (x c, x d ) S. (2.7) We impose the following conditions on the bandwidth and kernel functions. Define H = ( p 1 s=1 h s) p s=p 1 +1 min(h s, 1). (2.8) Letting 0 < ɛ < 1/(p + 4) and for some constant c > 0, n ɛ 1 H n ɛ ; n c < h s < n c for all s = 1,..., p ; the kernel K is a symmetric, compactly supported, Hölder-continuous probability density; and K(0) > K(δ) for all δ > 0. (2.9) The above conditions basically ask that each h s does not converge to zero, or to infinity, too fast, and that nh 1... h p1. We expect that, as n, the smoothing parameters associated with the relevant regressors will converge to zero, while those associated with the irrelevant regressors will not. It would be convenient to further assume that h s 0 for s = 1,..., p 1, and that λ s 0 for s = 1,..., q 1, however, the relevant components are known a priori. 4

5 Therefore, we shall assume the following condition holds. Defining µ g ( x) = E[ĝ(x) ˆf(x)]/E[ ˆf(x)] 1, we assume that supp w [ µ g(x) ḡ( x)] 2 w( x) f( x)d x, as a function of h 1,..., h p1, and λ 1,..., λ q1 vanishes if and only if all of the smoothing parameters vanish, (2.10) where w is a weight function defined in (2.16) below. Define an indicator function q I s (v d, x d ) = I(vs d x d s) t s,t=1 I(v d s = x d t ). (2.11) Note that I s (v d, x d ) = 1 if and only if v d and x d differ in their sth component only. The leading term of CV is κ p 1 σ 2 ( x) w(x) nh 1... h R(x) f( x)dx (2.12) p1 ( q 1 [ + λ s Is ( v d, x d ) { ḡ( x c, v d ) ḡ( x) } f( x c, v d ) ] κ 2 p 1 s=1 s=1 h 2 s v d {ḡss ( x) f( x) + 2 f s ( x)ḡ s ( x) }) 2 f( x) 1 w( x)d x, (2.13) 1 Note that µ g ( x) does not depend on x, nor does it depend on (h p1+1,..., p, λ q1+1,..., λ q ), because the components in the numerator related to the irrelevant variables cancel with those in the denominator. 5

6 with κ = K(v) 2 dv, κ 2 = K(v)v 2 dv, and where R(x) = R(x, h p1 +1,..., h p, λ q1 +1,..., λ q ) is given by R(x) = ν 2(x) ν 1 (x), (2.14) where for j = 1, 2, [ p ν j (x) = E s=p 1 +1 h 1 s K ( X c is x c ) q s h s s=q 1 +1 λ I(Xd is xd s) s ] j, (2.15) and w( x) = f( x)w(x)d x, (2.16) where, again, we define d x = x d d x c. Irrelevant variable x appears in R. By Hölder s inequality, R(x) 1 for all choices of x, h p1 +1,..., h p, and λ q1 +1,..., λ q. Also, R 1 as hs (p s p) and λ s 1 (q s q). One needs to select h s (s = p 1 + 1,..., p) and λ s (s = q 1 + 1,..., q) to minimize R. The only smoothing parameter values for which R(x, h p1 +1,..., h p, λ q1 +1,..., λ q ) = 1 are h s = for p s p, and λ s = 1 for q s q. Define Z n (x) = p s=p 1 +1 K ( x c s X d is ) q s X d h s s=q 1 +1 λi(xd is s ). If at least one h s is finite (for p s p), or one λ s < 1 (for q s q), then by K(0) > K(δ) for all δ > 0, we know that Var(Z n ) = E[Z 2 n] [E(Z n )] 2 > 0 so that R = E[Z 2 n]/[e(z n )] 2 > 1. 6

7 Only when all h s = and all λ s = 1, we have Z n K(0) p p 1 and Var(Z n ) = 0. R = 1 only in this case. Smoothing parameters related to irrelevant regressors must all converge to their upper extremities, so that R(x) 1 as n for all x S. Replacing R(x) by 1, the first term of (2.12) becomes κ p 1 σ 2 ( x) nh 1... h p1 w(x)d x, (2.17) a s = h s n 1/(q 1+4) and b s = λ s n 2/(q 1+4), then (2.12) (with (2.17) as its first term since R(x) 1) becomes n 4/(q 1+4) χ(a 1,..., a p1, b 1,..., b q1 ), where χ(a, b) = κ 2 p 1 s=1 a 2 s κ p σ 2 ( ( x) q 1 w( x)d x + b s I s ( v d, x d ) (ḡ( x c, v d ) ḡ( x) ) f( x c, v d ) a 1... a p1 s=1 v d (ḡss ( x) f( x) + 2 f s ( x)ḡ s ( x) )) 2 f( x) 1 w( x)d x. (2.18) Let a 0 1,..., a 0 p 1, b 0 1,..., b 0 q 1 denote values of a 1,..., a p1, b 1,..., b q1 that minimize χ subject to each of them being nonnegative. We require that Each a 0 s is positive and each b 0 s nonnegative, all are finite and uniquely defined. (2.19) Theorem 2.1 Assume conditions (2.7), (2.9), (2.10), and (2.19) hold, and let ĥ1,..., ĥp, ˆλ 1,..., ˆλ q denote the smoothing parameters that minimize CV. 7

8 Then n 1/(p1+4) ĥ s a 0 s in probability for 1 s p 1, P (ĥs > C) 1 for p s p and for all C > 0, n 2/(p 1+4)ˆλs b 0 s in probability for 1 s q 1, ˆλ s 1 in probability for q s q, (2.20) and n 4/(p 1+4) CV (ĥ1,..., ĥp, ˆλ 1,..., ˆλ q ) inf χ in probability. The asymptotic normality result for ĝ(x). Theorem 2.2 Under the same conditions as in Theorem 2.1, for x = (x c, x d ) S = S c S d, then ( nĥ1... ĥp 1 ) 1/2 [ĝ(x) ḡ( x) q 1 s=1 ] p 1 B 1s ( x)ˆλ 2 s B 2s ( x)ĥ2 d s N(0, Ω( x)), s=1 (2.21) where B 1s ( x) = v I d s ( v d, x d ) ( ḡ( x c, v d ) ḡ( x) ) f( x c, v d ) f( x) 1, B 2s ( x) = ( ) 1 2 κ 2 ḡ ss ( x) + 2 f s ( x)ḡ s ( x), and Ω( x) = κ f( x) p 1 σ2 ( x)/ f( x) are the asymptotic bias and variance terms respectively. 8

9 The Presence of Ordered Discrete Regressors If x d s takes c s different values, {0, 1,..., c s 1}, and is ordered, We suggest the use K d (x d s, v d s ) = λ xd s v d s s. (2.22) λ s [0, 1]. Aitchison and Aitken (1976,p.29) suggested using K d (x d s, v d s, λ s ) = ( ) cs λ t t s(1 λ s ) c s t where x d s v d s = t (0 t c s ), and ( ) cs = c t s!/[t!(c s t)!]. It can not smooth out irrelevant regressors, 9

10 Cross-Validation and Estimation of Conditional Densities Peter Hall, Jeff Racine, Qi Li Let f(x, y) denote the joint pdf of (X, Y ), and m(x) the marginal pdf X, and g(y x) = f(x, y)/m(x). We estimate g(y x) = f(x, y)/m(x) by ĝ(y x) = ˆf(x, y) ˆm(x). (2.23) Using the notation dz = z d dz c, a weighted integrated square difference between ĝ( ) and g( ) is I n = [ĝ(y x) g(y x)] 2 m(x) dz = [ĝ(y x)] 2 m(x) dz 2 ĝ(y x)g(y x)m(x) dz + [g(y x)] 2 m(x) dz I 1n 2I 2n + I 3n. (2.24) I 3n is independent of (h, λ). Define Ĝ(x) = [ ˆf(x, y)] 2 dy = n 2 K Xi,xK Xj,x K Yi,yK Yj,ydy i j = n 2 K Xi,xK Xj,xK (2) Y i,y j, (2.25) i j where K (2) Y i,y j = K Yi,yK Yj,ydy y d KYi,yK Yj,ydy c is the second order convolution kernel, K Yi,y = W Yi,yL Yi,y. 10

11 Using Eq. (2.25), we have I 1n = [ĝ(y x)] 2 m(x) dz = = [ ˆf(x, y)] 2 dy [ ˆm(x)] 2 m(x) dx Ĝ(x) [ ˆm(x)] 2m(x) dx = E Ĝ(X) X[ [ ˆm(X)] 2], (2.26) where E X ( ) denotes the expectation with respect to X only. Also, I 2n = = ĝ(y x)g(y x)m(x) dz = ] [ ˆf(x, y) ˆm(x) f(x, y) dx = E Z [ ˆf(Z) ˆm(X) ĝ(y x)f(x, y) dxdy ] where E Z denotes the expectation with respect to Z only., (2.27) Choosing m(x) as the weighting function give I 1n and I 2n some simple forms. { } Ĝ(X) I 1n 2I 2n = E X [ ˆm(X)] 2 [ ] ˆf(X, Y ) 2E Z, (2.28) ˆm(X) Use the leave-one-out estimators for ˆf(X l, Y l ) and ˆm(X l ), also use a leaveone-out estimator for G(X l ): Ĝ l (X l ) = n 2 K Xi,X l K Xj,X l K (2) Y i,y j. (2.29) i l j l CV (h, λ) def = 1 n n l=1 Ĝ l (X l ) [ ˆm l (X l )] 2 2 n n l=1 We will choose (λ, h) to minimize CV (h, λ). 11 ˆf l (X l, Y l ) ˆm l (X l ). (2.30)

12 Theorem 2.3 Assume some regularity conditions, and let ĥ 1,..., ĥ p, ˆλ 1,..., ˆλ q denote the smoothing parameters that minimize CV. Then n 1/(p1+5) ĥ s a 0 s in probability for 0 s p 1, P (ĥs > C) 1 for p s p and for all C > 0, n 2/(p 1+5)ˆλs b 0 s in probability for 1 s q 1, ˆλ s 1 in probability for q s q, (2.31) and n 4/(p 1+5) CV (ĥ0, ĥ1,..., ĥp, ˆλ 1,..., ˆλ q ) inf χ p (.) in probability. The asymptotic normality result. Theorem 2.4 Under some regularity conditions, we have nĥp (ĝ(y x) g(y x) ĥ2 B 1 (z) ˆλB 2 (z)) N(0, Ω(z)) in distribution, where B 1 (z) = (1/2)(1/m(z))tr[ 2 f(z)][ w(v)v 2 dv], B 2 (z) = (1/m(z)) z d,d z, z =1 [f(zc, z d ) f(z c, z d )], and Ω(z) = [f(z)/m 2 (x)][ W 2 (v)dv] ( 2 is with respect to z c ). 12

13 In the conditional probability paper, we define the the kernel function for discrete variables via for 1 s q, { 1 l(xis, d x d λs if Xis d s, λ s ) = = xd s, λ s c s 1 if Xis d xd s. λ s [0, (c s 1)/c s ]. (2.32) If x d s {0, 1}, a dummy variable with c s = 2. Then l(x d is, x d s, λ s ) = { 1 λ2 if X d is = xd s, λ s if X d is xd s. (2.33) λ s [0, 1/2]. The reason for this is that the kernel weights add up to one. x d S L(x d, X d d i, λ) = 1. It is needed in unconditional density estimation. This is not needed in regression estimation, or the conditional covariates. The relationship between the two λ s are Relative weight is λ/1 = λ earlier. Now the relative weight is λ/[(c 1)(1 λ)]. For example, if c = 2 and λ = 0.2 in the second expression, then the equivalent λ in the first expression should be λ = 0.2/[(1)(1 0.2)] = 0.2/0.8 =

14 3 Applications Two empirical applications. Out-of-sample predictive performance. 3.1 Count Survival Data - Veteran s Administration Lung Cancer Trial The data set is from Kalbfleisch and Prentice (1980, pg ), which models survival in days of cancer patients with six discrete explanatory variables being treatment type, cell type, Karnofsky score, months from diagnosis, age in years, and prior therapy. The dataset contains 137 observations, and the number of cells greatly exceed the number of observations. Model the expected survival in days. The estimation sample n 1 = 132, and the prediction sample is n 2 = 5. The out-of-sample MSE = n 1 2 n2 i=1 (y i ŷ i ) 2. Randomly split the sample into two sub-samples of size n 1 and n times. Table 1: Veteran s Lung Cancer Data Mean MSEnp Model Median MSE Mean MSE Mean MSE NONP 9, , % OLS 13, , % POISSON 15, , % 14

15 Median values of the CV smoothing parameters over the 100 sample splits. Table 2: Summary of Median Smoothing Parameters [number of categories parentheses] λ 1 [2] λ 2 [4] λ 3 [12] λ 4 [28] λ 5 [40] λ 6 [2] The relevant explanatory variables for predicting survival are cancer cell-type (λ 2 ), Karnofsky score 2 (λ 3 ), and months from diagnosis (λ 4 ). The remaining regressors are smoothed out by cross-validation. 2 The Karnofsky score measures patient performance of activities of daily living. The score has proven useful not only to follow the course of the illness (usually progressive deficit and ultimately death), but also a prognosticator: patients with the highest (best) Karnofsky scores at the time of tumor diagnosis have the best survival and quality of life over the course of their illness. 15

16 3.2 Count Data - Fair s 1977 Extramarital Affairs Data We consider Fair s (1978) dataset which models how often an individual engaged in extramarital affairs during the past year. This often cited paper with a complete description of the data is located at The dataset contains 601 observations and has eight discrete explanatory variables given by sex, age, number of years married, have children or not, how religious, level of education, occupation, and how they rate their marriage. Estimation sample n 1 = 500, prediction sample n 2 = 101. The goal is to model the expected number of affairs. We split the sample into two sub-sample of size n 1 and n 2 to compute the out-of-sample squared prediction errors, and repeat this 100 times. Median and mean out-of-sample squared prediction errors by different estimation methods. Table 3: Fair s 1977 Extramarital Affairs Data Mean MSEnp Model Median MSE Mean MSE Mean MSE NONP % OLS % PROBIT % TOBIT % POISSON % NP MSPE is 69% to 82% of those by various Para. methods. 16

17 Median values of CV smoothing parameters over the 100 sample splits. Table 4: Summary of Median Smoothing Parameters [number of categories in parentheses] λ 1 [2] λ 2 [9] λ 3 [8] λ 4 [2] λ 5 [5] λ 6 [7] λ 7 [7] λ 8 [5] The most relevant explanatory variables are age (λ 2 ) and how a person rates their marriage (λ 8 ), while irrrelevant variables are number of years married (λ 3 ), having children or not (λ 4 ), and level of education (λ 6 ) appear to be irrelevant for prediction of number of affairs, and the remaining variables appear to have some relevance again from a cross-validation perspective. In comparison, using a Tobit model, Fair obtained a t-statistic of 3.63 for the number of years married, while our nonparametric method removes this variable from the resulting estimate. The out-of-sample prediction results suggest that the Tobit model is likely to be misspecified. 17

Nonparametric Estimation of Regression Functions with Both Categorical and Continuous Data

Nonparametric Estimation of Regression Functions with Both Categorical and Continuous Data Nonparametric Estimation of Regression Functions with Both Categorical and Continuous Data Jeff Racine Department of Economics, University of South Florida Tampa, FL 33620, USA Qi Li Department of Economics,

More information

41903: Introduction to Nonparametrics

41903: Introduction to Nonparametrics 41903: Notes 5 Introduction Nonparametrics fundamentally about fitting flexible models: want model that is flexible enough to accommodate important patterns but not so flexible it overspecializes to specific

More information

Efficient Estimation of Average Treatment Effects With Mixed Categorical and Continuous Data

Efficient Estimation of Average Treatment Effects With Mixed Categorical and Continuous Data Efficient Estimation of Average Treatment Effects With Mixed Categorical and Continuous Data Qi Li Department of Economics Texas A&M University Bush Academic Building West College Station, TX 77843-48

More information

Local Polynomial Regression

Local Polynomial Regression VI Local Polynomial Regression (1) Global polynomial regression We observe random pairs (X 1, Y 1 ),, (X n, Y n ) where (X 1, Y 1 ),, (X n, Y n ) iid (X, Y ). We want to estimate m(x) = E(Y X = x) based

More information

Nonparametric Econometrics

Nonparametric Econometrics Applied Microeconometrics with Stata Nonparametric Econometrics Spring Term 2011 1 / 37 Contents Introduction The histogram estimator The kernel density estimator Nonparametric regression estimators Semi-

More information

Density estimation Nonparametric conditional mean estimation Semiparametric conditional mean estimation. Nonparametrics. Gabriel Montes-Rojas

Density estimation Nonparametric conditional mean estimation Semiparametric conditional mean estimation. Nonparametrics. Gabriel Montes-Rojas 0 0 5 Motivation: Regression discontinuity (Angrist&Pischke) Outcome.5 1 1.5 A. Linear E[Y 0i X i] 0.2.4.6.8 1 X Outcome.5 1 1.5 B. Nonlinear E[Y 0i X i] i 0.2.4.6.8 1 X utcome.5 1 1.5 C. Nonlinearity

More information

Error distribution function for parametrically truncated and censored data

Error distribution function for parametrically truncated and censored data Error distribution function for parametrically truncated and censored data Géraldine LAURENT Jointly with Cédric HEUCHENNE QuantOM, HEC-ULg Management School - University of Liège Friday, 14 September

More information

Nonparametric Econometrics in R

Nonparametric Econometrics in R Nonparametric Econometrics in R Philip Shaw Fordham University November 17, 2011 Philip Shaw (Fordham University) Nonparametric Econometrics in R November 17, 2011 1 / 16 Introduction The NP Package R

More information

Nonparametric Methods

Nonparametric Methods Nonparametric Methods Michael R. Roberts Department of Finance The Wharton School University of Pennsylvania July 28, 2009 Michael R. Roberts Nonparametric Methods 1/42 Overview Great for data analysis

More information

Time Series and Forecasting Lecture 4 NonLinear Time Series

Time Series and Forecasting Lecture 4 NonLinear Time Series Time Series and Forecasting Lecture 4 NonLinear Time Series Bruce E. Hansen Summer School in Economics and Econometrics University of Crete July 23-27, 2012 Bruce Hansen (University of Wisconsin) Foundations

More information

Density estimators for the convolution of discrete and continuous random variables

Density estimators for the convolution of discrete and continuous random variables Density estimators for the convolution of discrete and continuous random variables Ursula U Müller Texas A&M University Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract

More information

Gibbs Sampling in Latent Variable Models #1

Gibbs Sampling in Latent Variable Models #1 Gibbs Sampling in Latent Variable Models #1 Econ 690 Purdue University Outline 1 Data augmentation 2 Probit Model Probit Application A Panel Probit Panel Probit 3 The Tobit Model Example: Female Labor

More information

Non-parametric Inference and Resampling

Non-parametric Inference and Resampling Non-parametric Inference and Resampling Exercises by David Wozabal (Last update 3. Juni 2013) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend

More information

Modelling Non-linear and Non-stationary Time Series

Modelling Non-linear and Non-stationary Time Series Modelling Non-linear and Non-stationary Time Series Chapter 2: Non-parametric methods Henrik Madsen Advanced Time Series Analysis September 206 Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September

More information

Nonparametric Regression

Nonparametric Regression Nonparametric Regression Econ 674 Purdue University April 8, 2009 Justin L. Tobias (Purdue) Nonparametric Regression April 8, 2009 1 / 31 Consider the univariate nonparametric regression model: where y

More information

Sparse Nonparametric Density Estimation in High Dimensions Using the Rodeo

Sparse Nonparametric Density Estimation in High Dimensions Using the Rodeo Outline in High Dimensions Using the Rodeo Han Liu 1,2 John Lafferty 2,3 Larry Wasserman 1,2 1 Statistics Department, 2 Machine Learning Department, 3 Computer Science Department, Carnegie Mellon University

More information

Nonparametric Regression Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction

Nonparametric Regression Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction Tine Buch-Kromann Univariate Kernel Regression The relationship between two variables, X and Y where m(

More information

Local regression I. Patrick Breheny. November 1. Kernel weighted averages Local linear regression

Local regression I. Patrick Breheny. November 1. Kernel weighted averages Local linear regression Local regression I Patrick Breheny November 1 Patrick Breheny STA 621: Nonparametric Statistics 1/27 Simple local models Kernel weighted averages The Nadaraya-Watson estimator Expected loss and prediction

More information

CALCULATING DEGREES OF FREEDOM IN MULTIVARIATE LOCAL POLYNOMIAL REGRESSION. 1. Introduction

CALCULATING DEGREES OF FREEDOM IN MULTIVARIATE LOCAL POLYNOMIAL REGRESSION. 1. Introduction CALCULATING DEGREES OF FREEDOM IN MULTIVARIATE LOCAL POLYNOMIAL REGRESSION NADINE MCCLOUD AND CHRISTOPHER F. PARMETER Abstract. The matrix that transforms the response variable in a regression to its predicted

More information

ECO Class 6 Nonparametric Econometrics

ECO Class 6 Nonparametric Econometrics ECO 523 - Class 6 Nonparametric Econometrics Carolina Caetano Contents 1 Nonparametric instrumental variable regression 1 2 Nonparametric Estimation of Average Treatment Effects 3 2.1 Asymptotic results................................

More information

ECON 721: Lecture Notes on Nonparametric Density and Regression Estimation. Petra E. Todd

ECON 721: Lecture Notes on Nonparametric Density and Regression Estimation. Petra E. Todd ECON 721: Lecture Notes on Nonparametric Density and Regression Estimation Petra E. Todd Fall, 2014 2 Contents 1 Review of Stochastic Order Symbols 1 2 Nonparametric Density Estimation 3 2.1 Histogram

More information

Statistical Properties of Numerical Derivatives

Statistical Properties of Numerical Derivatives Statistical Properties of Numerical Derivatives Han Hong, Aprajit Mahajan, and Denis Nekipelov Stanford University and UC Berkeley November 2010 1 / 63 Motivation Introduction Many models have objective

More information

Robustness to Parametric Assumptions in Missing Data Models

Robustness to Parametric Assumptions in Missing Data Models Robustness to Parametric Assumptions in Missing Data Models Bryan Graham NYU Keisuke Hirano University of Arizona April 2011 Motivation Motivation We consider the classic missing data problem. In practice

More information

Model-free prediction intervals for regression and autoregression. Dimitris N. Politis University of California, San Diego

Model-free prediction intervals for regression and autoregression. Dimitris N. Politis University of California, San Diego Model-free prediction intervals for regression and autoregression Dimitris N. Politis University of California, San Diego To explain or to predict? Models are indispensable for exploring/utilizing relationships

More information

Practice Examination # 3

Practice Examination # 3 Practice Examination # 3 Sta 23: Probability December 13, 212 This is a closed-book exam so do not refer to your notes, the text, or any other books (please put them on the floor). You may use a single

More information

Introduction to Regression

Introduction to Regression Introduction to Regression p. 1/97 Introduction to Regression Chad Schafer cschafer@stat.cmu.edu Carnegie Mellon University Introduction to Regression p. 1/97 Acknowledgement Larry Wasserman, All of Nonparametric

More information

Pairwise rank based likelihood for estimating the relationship between two homogeneous populations and their mixture proportion

Pairwise rank based likelihood for estimating the relationship between two homogeneous populations and their mixture proportion Pairwise rank based likelihood for estimating the relationship between two homogeneous populations and their mixture proportion Glenn Heller and Jing Qin Department of Epidemiology and Biostatistics Memorial

More information

2 Functions of random variables

2 Functions of random variables 2 Functions of random variables A basic statistical model for sample data is a collection of random variables X 1,..., X n. The data are summarised in terms of certain sample statistics, calculated as

More information

Local linear multiple regression with variable. bandwidth in the presence of heteroscedasticity

Local linear multiple regression with variable. bandwidth in the presence of heteroscedasticity Local linear multiple regression with variable bandwidth in the presence of heteroscedasticity Azhong Ye 1 Rob J Hyndman 2 Zinai Li 3 23 January 2006 Abstract: We present local linear estimator with variable

More information

Generated Covariates in Nonparametric Estimation: A Short Review.

Generated Covariates in Nonparametric Estimation: A Short Review. Generated Covariates in Nonparametric Estimation: A Short Review. Enno Mammen, Christoph Rothe, and Melanie Schienle Abstract In many applications, covariates are not observed but have to be estimated

More information

Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes

Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract Convergence

More information

Inverse Statistical Learning

Inverse Statistical Learning Inverse Statistical Learning Minimax theory, adaptation and algorithm avec (par ordre d apparition) C. Marteau, M. Chichignoud, C. Brunet and S. Souchet Dijon, le 15 janvier 2014 Inverse Statistical Learning

More information

Motivational Example

Motivational Example Motivational Example Data: Observational longitudinal study of obesity from birth to adulthood. Overall Goal: Build age-, gender-, height-specific growth charts (under 3 year) to diagnose growth abnomalities.

More information

13 Endogeneity and Nonparametric IV

13 Endogeneity and Nonparametric IV 13 Endogeneity and Nonparametric IV 13.1 Nonparametric Endogeneity A nonparametric IV equation is Y i = g (X i ) + e i (1) E (e i j i ) = 0 In this model, some elements of X i are potentially endogenous,

More information

Econ 2148, fall 2017 Instrumental variables II, continuous treatment

Econ 2148, fall 2017 Instrumental variables II, continuous treatment Econ 2148, fall 2017 Instrumental variables II, continuous treatment Maximilian Kasy Department of Economics, Harvard University 1 / 35 Recall instrumental variables part I Origins of instrumental variables:

More information

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model Minimum Hellinger Distance Estimation in a Semiparametric Mixture Model Sijia Xiang 1, Weixin Yao 1, and Jingjing Wu 2 1 Department of Statistics, Kansas State University, Manhattan, Kansas, USA 66506-0802.

More information

Statistics 100A Homework 5 Solutions

Statistics 100A Homework 5 Solutions Chapter 5 Statistics 1A Homework 5 Solutions Ryan Rosario 1. Let X be a random variable with probability density function a What is the value of c? fx { c1 x 1 < x < 1 otherwise We know that for fx to

More information

Non-parametric bootstrap mean squared error estimation for M-quantile estimates of small area means, quantiles and poverty indicators

Non-parametric bootstrap mean squared error estimation for M-quantile estimates of small area means, quantiles and poverty indicators Non-parametric bootstrap mean squared error estimation for M-quantile estimates of small area means, quantiles and poverty indicators Stefano Marchetti 1 Nikos Tzavidis 2 Monica Pratesi 3 1,3 Department

More information

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics Jiti Gao Department of Statistics School of Mathematics and Statistics The University of Western Australia Crawley

More information

Lecture Notes 15 Prediction Chapters 13, 22, 20.4.

Lecture Notes 15 Prediction Chapters 13, 22, 20.4. Lecture Notes 15 Prediction Chapters 13, 22, 20.4. 1 Introduction Prediction is covered in detail in 36-707, 36-701, 36-715, 10/36-702. Here, we will just give an introduction. We observe training data

More information

finite-sample optimal estimation and inference on average treatment effects under unconfoundedness

finite-sample optimal estimation and inference on average treatment effects under unconfoundedness finite-sample optimal estimation and inference on average treatment effects under unconfoundedness Timothy Armstrong (Yale University) Michal Kolesár (Princeton University) September 2017 Introduction

More information

5 Operations on Multiple Random Variables

5 Operations on Multiple Random Variables EE360 Random Signal analysis Chapter 5: Operations on Multiple Random Variables 5 Operations on Multiple Random Variables Expected value of a function of r.v. s Two r.v. s: ḡ = E[g(X, Y )] = g(x, y)f X,Y

More information

Adaptive Kernel Estimation of The Hazard Rate Function

Adaptive Kernel Estimation of The Hazard Rate Function Adaptive Kernel Estimation of The Hazard Rate Function Raid Salha Department of Mathematics, Islamic University of Gaza, Palestine, e-mail: rbsalha@mail.iugaza.edu Abstract In this paper, we generalized

More information

Kernel density estimation

Kernel density estimation Kernel density estimation Patrick Breheny October 18 Patrick Breheny STA 621: Nonparametric Statistics 1/34 Introduction Kernel Density Estimation We ve looked at one method for estimating density: histograms

More information

Spatially Smoothed Kernel Density Estimation via Generalized Empirical Likelihood

Spatially Smoothed Kernel Density Estimation via Generalized Empirical Likelihood Spatially Smoothed Kernel Density Estimation via Generalized Empirical Likelihood Kuangyu Wen & Ximing Wu Texas A&M University Info-Metrics Institute Conference: Recent Innovations in Info-Metrics October

More information

4 Riesz Kernels. Since the functions k i (ξ) = ξ i. are bounded functions it is clear that R

4 Riesz Kernels. Since the functions k i (ξ) = ξ i. are bounded functions it is clear that R 4 Riesz Kernels. A natural generalization of the Hilbert transform to higher dimension is mutiplication of the Fourier Transform by homogeneous functions of degree 0, the simplest ones being R i f(ξ) =

More information

Optimal bandwidth selection for the fuzzy regression discontinuity estimator

Optimal bandwidth selection for the fuzzy regression discontinuity estimator Optimal bandwidth selection for the fuzzy regression discontinuity estimator Yoichi Arai Hidehiko Ichimura The Institute for Fiscal Studies Department of Economics, UCL cemmap working paper CWP49/5 Optimal

More information

ECE Lecture #9 Part 2 Overview

ECE Lecture #9 Part 2 Overview ECE 450 - Lecture #9 Part Overview Bivariate Moments Mean or Expected Value of Z = g(x, Y) Correlation and Covariance of RV s Functions of RV s: Z = g(x, Y); finding f Z (z) Method : First find F(z), by

More information

Chapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables

Chapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables Chapter 2 Some Basic Probability Concepts 2.1 Experiments, Outcomes and Random Variables A random variable is a variable whose value is unknown until it is observed. The value of a random variable results

More information

Semiparametric Estimation of a Sample Selection Model in the Presence of Endogeneity

Semiparametric Estimation of a Sample Selection Model in the Presence of Endogeneity Semiparametric Estimation of a Sample Selection Model in the Presence of Endogeneity Jörg Schwiebert Abstract In this paper, we derive a semiparametric estimation procedure for the sample selection model

More information

Introduction. Linear Regression. coefficient estimates for the wage equation: E(Y X) = X 1 β X d β d = X β

Introduction. Linear Regression. coefficient estimates for the wage equation: E(Y X) = X 1 β X d β d = X β Introduction - Introduction -2 Introduction Linear Regression E(Y X) = X β +...+X d β d = X β Example: Wage equation Y = log wages, X = schooling (measured in years), labor market experience (measured

More information

On the Robust Modal Local Polynomial Regression

On the Robust Modal Local Polynomial Regression International Journal of Statistical Sciences ISSN 683 5603 Vol. 9(Special Issue), 2009, pp 27-23 c 2009 Dept. of Statistics, Univ. of Rajshahi, Bangladesh On the Robust Modal Local Polynomial Regression

More information

Exercises and Answers to Chapter 1

Exercises and Answers to Chapter 1 Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean

More information

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) 1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For

More information

Computational treatment of the error distribution in nonparametric regression with right-censored and selection-biased data

Computational treatment of the error distribution in nonparametric regression with right-censored and selection-biased data Computational treatment of the error distribution in nonparametric regression with right-censored and selection-biased data Géraldine Laurent 1 and Cédric Heuchenne 2 1 QuantOM, HEC-Management School of

More information

Contents 1. Contents

Contents 1. Contents Contents 1 Contents 6 Distributions of Functions of Random Variables 2 6.1 Transformation of Discrete r.v.s............. 3 6.2 Method of Distribution Functions............. 6 6.3 Method of Transformations................

More information

Lecture 25: Review. Statistics 104. April 23, Colin Rundel

Lecture 25: Review. Statistics 104. April 23, Colin Rundel Lecture 25: Review Statistics 104 Colin Rundel April 23, 2012 Joint CDF F (x, y) = P [X x, Y y] = P [(X, Y ) lies south-west of the point (x, y)] Y (x,y) X Statistics 104 (Colin Rundel) Lecture 25 April

More information

Math 5588 Final Exam Solutions

Math 5588 Final Exam Solutions Math 5588 Final Exam Solutions Prof. Jeff Calder May 9, 2017 1. Find the function u : [0, 1] R that minimizes I(u) = subject to u(0) = 0 and u(1) = 1. 1 0 e u(x) u (x) + u (x) 2 dx, Solution. Since the

More information

Nonparametric Function Estimation with Infinite-Order Kernels

Nonparametric Function Estimation with Infinite-Order Kernels Nonparametric Function Estimation with Infinite-Order Kernels Arthur Berg Department of Statistics, University of Florida March 15, 2008 Kernel Density Estimation (IID Case) Let X 1,..., X n iid density

More information

Quantile Regression for Residual Life and Empirical Likelihood

Quantile Regression for Residual Life and Empirical Likelihood Quantile Regression for Residual Life and Empirical Likelihood Mai Zhou email: mai@ms.uky.edu Department of Statistics, University of Kentucky, Lexington, KY 40506-0027, USA Jong-Hyeon Jeong email: jeong@nsabp.pitt.edu

More information

Introduction to Regression

Introduction to Regression Introduction to Regression Chad M. Schafer May 20, 2015 Outline General Concepts of Regression, Bias-Variance Tradeoff Linear Regression Nonparametric Procedures Cross Validation Local Polynomial Regression

More information

Independent and conditionally independent counterfactual distributions

Independent and conditionally independent counterfactual distributions Independent and conditionally independent counterfactual distributions Marcin Wolski European Investment Bank M.Wolski@eib.org Society for Nonlinear Dynamics and Econometrics Tokyo March 19, 2018 Views

More information

Adaptive Nonparametric Density Estimators

Adaptive Nonparametric Density Estimators Adaptive Nonparametric Density Estimators by Alan J. Izenman Introduction Theoretical results and practical application of histograms as density estimators usually assume a fixed-partition approach, where

More information

Expectation and Variance

Expectation and Variance Expectation and Variance August 22, 2017 STAT 151 Class 3 Slide 1 Outline of Topics 1 Motivation 2 Expectation - discrete 3 Transformations 4 Variance - discrete 5 Continuous variables 6 Covariance STAT

More information

Flexible Estimation of Treatment Effect Parameters

Flexible Estimation of Treatment Effect Parameters Flexible Estimation of Treatment Effect Parameters Thomas MaCurdy a and Xiaohong Chen b and Han Hong c Introduction Many empirical studies of program evaluations are complicated by the presence of both

More information

Simple and Efficient Improvements of Multivariate Local Linear Regression

Simple and Efficient Improvements of Multivariate Local Linear Regression Journal of Multivariate Analysis Simple and Efficient Improvements of Multivariate Local Linear Regression Ming-Yen Cheng 1 and Liang Peng Abstract This paper studies improvements of multivariate local

More information

A Discontinuity Test for Identification in Nonparametric Models with Endogeneity

A Discontinuity Test for Identification in Nonparametric Models with Endogeneity A Discontinuity Test for Identification in Nonparametric Models with Endogeneity Carolina Caetano 1 Christoph Rothe 2 Nese Yildiz 1 1 Department of Economics 2 Department of Economics University of Rochester

More information

Additive Isotonic Regression

Additive Isotonic Regression Additive Isotonic Regression Enno Mammen and Kyusang Yu 11. July 2006 INTRODUCTION: We have i.i.d. random vectors (Y 1, X 1 ),..., (Y n, X n ) with X i = (X1 i,..., X d i ) and we consider the additive

More information

Truncation and Censoring

Truncation and Censoring Truncation and Censoring Laura Magazzini laura.magazzini@univr.it Laura Magazzini (@univr.it) Truncation and Censoring 1 / 35 Truncation and censoring Truncation: sample data are drawn from a subset of

More information

Lectures on Elementary Probability. William G. Faris

Lectures on Elementary Probability. William G. Faris Lectures on Elementary Probability William G. Faris February 22, 2002 2 Contents 1 Combinatorics 5 1.1 Factorials and binomial coefficients................. 5 1.2 Sampling with replacement.....................

More information

STAT 450: Statistical Theory. Distribution Theory. Reading in Casella and Berger: Ch 2 Sec 1, Ch 4 Sec 1, Ch 4 Sec 6.

STAT 450: Statistical Theory. Distribution Theory. Reading in Casella and Berger: Ch 2 Sec 1, Ch 4 Sec 1, Ch 4 Sec 6. STAT 45: Statistical Theory Distribution Theory Reading in Casella and Berger: Ch 2 Sec 1, Ch 4 Sec 1, Ch 4 Sec 6. Basic Problem: Start with assumptions about f or CDF of random vector X (X 1,..., X p

More information

Let X and Y be two real valued stochastic variables defined on (Ω, F, P). Theorem: If X and Y are independent then. . p.1/21

Let X and Y be two real valued stochastic variables defined on (Ω, F, P). Theorem: If X and Y are independent then. . p.1/21 Multivariate transformations The remaining part of the probability course is centered around transformations t : R k R m and how they transform probability measures. For instance tx 1,...,x k = x 1 +...

More information

Introduction to Regression

Introduction to Regression Introduction to Regression David E Jones (slides mostly by Chad M Schafer) June 1, 2016 1 / 102 Outline General Concepts of Regression, Bias-Variance Tradeoff Linear Regression Nonparametric Procedures

More information

Introduction to Regression

Introduction to Regression Introduction to Regression Chad M. Schafer cschafer@stat.cmu.edu Carnegie Mellon University Introduction to Regression p. 1/100 Outline General Concepts of Regression, Bias-Variance Tradeoff Linear Regression

More information

Final Exam # 3. Sta 230: Probability. December 16, 2012

Final Exam # 3. Sta 230: Probability. December 16, 2012 Final Exam # 3 Sta 230: Probability December 16, 2012 This is a closed-book exam so do not refer to your notes, the text, or any other books (please put them on the floor). You may use the extra sheets

More information

Ultra High Dimensional Variable Selection with Endogenous Variables

Ultra High Dimensional Variable Selection with Endogenous Variables 1 / 39 Ultra High Dimensional Variable Selection with Endogenous Variables Yuan Liao Princeton University Joint work with Jianqing Fan Job Market Talk January, 2012 2 / 39 Outline 1 Examples of Ultra High

More information

Homework 10 (due December 2, 2009)

Homework 10 (due December 2, 2009) Homework (due December, 9) Problem. Let X and Y be independent binomial random variables with parameters (n, p) and (n, p) respectively. Prove that X + Y is a binomial random variable with parameters (n

More information

ESTIMATING AVERAGE TREATMENT EFFECTS: REGRESSION DISCONTINUITY DESIGNS Jeff Wooldridge Michigan State University BGSE/IZA Course in Microeconometrics

ESTIMATING AVERAGE TREATMENT EFFECTS: REGRESSION DISCONTINUITY DESIGNS Jeff Wooldridge Michigan State University BGSE/IZA Course in Microeconometrics ESTIMATING AVERAGE TREATMENT EFFECTS: REGRESSION DISCONTINUITY DESIGNS Jeff Wooldridge Michigan State University BGSE/IZA Course in Microeconometrics July 2009 1. Introduction 2. The Sharp RD Design 3.

More information

Michael Lechner Causal Analysis RDD 2014 page 1. Lecture 7. The Regression Discontinuity Design. RDD fuzzy and sharp

Michael Lechner Causal Analysis RDD 2014 page 1. Lecture 7. The Regression Discontinuity Design. RDD fuzzy and sharp page 1 Lecture 7 The Regression Discontinuity Design fuzzy and sharp page 2 Regression Discontinuity Design () Introduction (1) The design is a quasi-experimental design with the defining characteristic

More information

Integral approximation by kernel smoothing

Integral approximation by kernel smoothing Integral approximation by kernel smoothing François Portier Université catholique de Louvain - ISBA August, 29 2014 In collaboration with Bernard Delyon Topic of the talk: Given ϕ : R d R, estimation of

More information

What s New in Econometrics? Lecture 14 Quantile Methods

What s New in Econometrics? Lecture 14 Quantile Methods What s New in Econometrics? Lecture 14 Quantile Methods Jeff Wooldridge NBER Summer Institute, 2007 1. Reminders About Means, Medians, and Quantiles 2. Some Useful Asymptotic Results 3. Quantile Regression

More information

New Local Estimation Procedure for Nonparametric Regression Function of Longitudinal Data

New Local Estimation Procedure for Nonparametric Regression Function of Longitudinal Data ew Local Estimation Procedure for onparametric Regression Function of Longitudinal Data Weixin Yao and Runze Li The Pennsylvania State University Technical Report Series #0-03 College of Health and Human

More information

4.1 The Expectation of a Random Variable

4.1 The Expectation of a Random Variable STAT 42 Lecture Notes 93 4. The Expectation of a Random Variable This chapter begins the discussion of properties of random variables. The focus of this chapter is on expectations of random variables.

More information

Smooth simultaneous confidence bands for cumulative distribution functions

Smooth simultaneous confidence bands for cumulative distribution functions Journal of Nonparametric Statistics, 2013 Vol. 25, No. 2, 395 407, http://dx.doi.org/10.1080/10485252.2012.759219 Smooth simultaneous confidence bands for cumulative distribution functions Jiangyan Wang

More information

Estimating a frontier function using a high-order moments method

Estimating a frontier function using a high-order moments method 1/ 16 Estimating a frontier function using a high-order moments method Gilles STUPFLER (University of Nottingham) Joint work with Stéphane GIRARD (INRIA Rhône-Alpes) and Armelle GUILLOU (Université de

More information

ASM Study Manual for Exam P, Second Edition By Dr. Krzysztof M. Ostaszewski, FSA, CFA, MAAA Errata

ASM Study Manual for Exam P, Second Edition By Dr. Krzysztof M. Ostaszewski, FSA, CFA, MAAA Errata ASM Study Manual for Exam P, Second Edition By Dr. Krzysztof M. Ostaszewski, FSA, CFA, MAAA (krzysio@krzysio.net) Errata Effective July 5, 3, only the latest edition of this manual will have its errata

More information

On variable bandwidth kernel density estimation

On variable bandwidth kernel density estimation JSM 04 - Section on Nonparametric Statistics On variable bandwidth kernel density estimation Janet Nakarmi Hailin Sang Abstract In this paper we study the ideal variable bandwidth kernel estimator introduced

More information

Understanding Regressions with Observations Collected at High Frequency over Long Span

Understanding Regressions with Observations Collected at High Frequency over Long Span Understanding Regressions with Observations Collected at High Frequency over Long Span Yoosoon Chang Department of Economics, Indiana University Joon Y. Park Department of Economics, Indiana University

More information

Final Exam. Economics 835: Econometrics. Fall 2010

Final Exam. Economics 835: Econometrics. Fall 2010 Final Exam Economics 835: Econometrics Fall 2010 Please answer the question I ask - no more and no less - and remember that the correct answer is often short and simple. 1 Some short questions a) For each

More information

Chapter 8. Quantile Regression and Quantile Treatment Effects

Chapter 8. Quantile Regression and Quantile Treatment Effects Chapter 8. Quantile Regression and Quantile Treatment Effects By Joan Llull Quantitative & Statistical Methods II Barcelona GSE. Winter 2018 I. Introduction A. Motivation As in most of the economics literature,

More information

Class 8 Review Problems solutions, 18.05, Spring 2014

Class 8 Review Problems solutions, 18.05, Spring 2014 Class 8 Review Problems solutions, 8.5, Spring 4 Counting and Probability. (a) Create an arrangement in stages and count the number of possibilities at each stage: ( ) Stage : Choose three of the slots

More information

Econ 582 Nonparametric Regression

Econ 582 Nonparametric Regression Econ 582 Nonparametric Regression Eric Zivot May 28, 2013 Nonparametric Regression Sofarwehaveonlyconsideredlinearregressionmodels = x 0 β + [ x ]=0 [ x = x] =x 0 β = [ x = x] [ x = x] x = β The assume

More information

Robustifying Trial-Derived Treatment Rules to a Target Population

Robustifying Trial-Derived Treatment Rules to a Target Population 1/ 39 Robustifying Trial-Derived Treatment Rules to a Target Population Yingqi Zhao Public Health Sciences Division Fred Hutchinson Cancer Research Center Workshop on Perspectives and Analysis for Personalized

More information

Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix

Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix Yingying Dong and Arthur Lewbel California State University Fullerton and Boston College July 2010 Abstract

More information

1 Exercises for lecture 1

1 Exercises for lecture 1 1 Exercises for lecture 1 Exercise 1 a) Show that if F is symmetric with respect to µ, and E( X )

More information

Stat 704 Data Analysis I Probability Review

Stat 704 Data Analysis I Probability Review 1 / 39 Stat 704 Data Analysis I Probability Review Dr. Yen-Yi Ho Department of Statistics, University of South Carolina A.3 Random Variables 2 / 39 def n: A random variable is defined as a function that

More information

Nonparametric confidence intervals. for receiver operating characteristic curves

Nonparametric confidence intervals. for receiver operating characteristic curves Nonparametric confidence intervals for receiver operating characteristic curves Peter G. Hall 1, Rob J. Hyndman 2, and Yanan Fan 3 5 December 2003 Abstract: We study methods for constructing confidence

More information

NADARAYA WATSON ESTIMATE JAN 10, 2006: version 2. Y ik ( x i

NADARAYA WATSON ESTIMATE JAN 10, 2006: version 2. Y ik ( x i NADARAYA WATSON ESTIMATE JAN 0, 2006: version 2 DATA: (x i, Y i, i =,..., n. ESTIMATE E(Y x = m(x by n i= ˆm (x = Y ik ( x i x n i= K ( x i x EXAMPLES OF K: K(u = I{ u c} (uniform or box kernel K(u = u

More information

Optimal bandwidth selection for differences of nonparametric estimators with an application to the sharp regression discontinuity design

Optimal bandwidth selection for differences of nonparametric estimators with an application to the sharp regression discontinuity design Optimal bandwidth selection for differences of nonparametric estimators with an application to the sharp regression discontinuity design Yoichi Arai Hidehiko Ichimura The Institute for Fiscal Studies Department

More information

Week 9 The Central Limit Theorem and Estimation Concepts

Week 9 The Central Limit Theorem and Estimation Concepts Week 9 and Estimation Concepts Week 9 and Estimation Concepts Week 9 Objectives 1 The Law of Large Numbers and the concept of consistency of averages are introduced. The condition of existence of the population

More information