Nonparametric Econometrics
|
|
- Godfrey Preston
- 5 years ago
- Views:
Transcription
1 Applied Microeconometrics with Stata Nonparametric Econometrics Spring Term / 37
2 Contents Introduction The histogram estimator The kernel density estimator Nonparametric regression estimators Semi- and nonparametric extensions Summary and references 2 / 37
3 Introduction Up to now, all regression models relied on functions (or densities) which depend on an unknown finite-dimensional parameter. A finite-dimensional parameter is an element of R q with q N. For example, linear regression models use an additive combination of the covariates (x β). Nonlinear regression models specify a known function of (a linear index of) the covariates. ML theory is based on a assumption about the density of the data, which depends on a finite dimensional parameter. If the functional form or the distributional assumption is wrong, the parameter estimators of these models are inconsistent, however. To circumvent this kind of misspecification problem, nonparametric methods can be used, which do not impose assumptions on some functional form. In the following, nonparametric density and regression estimators will be described, which are the base for most nonparametric econometric models. 3 / 37
4 The Histogram Estimator Consider a sample {X i } n density f X (x). of a random variable X with unknown First, the support of X is divided in K intervals [a j 1, a j ) with min(x ) = a 0 < a 1 <... < a K = max(x ). The histogram estimator for the probability that X takes a value in the interval [a j 1, a j ) is then defined by Pr(X [a j 1, a j )) = 1 n 1{X i [a j 1, a j )}. n This estimator can be viewed as a (crude) approximation of f X at some point x [a j 1, a j ). Furthermore, the histogram estimator is a step-function even if f X is continuous (see the examples of the next slide). A solution is to use smaller intervals [a j 1, a j ), which converge in the limit to a point. This approach is called a local estimation approach, and is the basic idea of the kernel density estimator. 4 / 37
5 The Histogram Estimator The graphs show histograms of the logarithm of wage (from the data set Mroz.dta). The histogram of the top left panel divides the data in 20 classes (the default value of the histogram command of Stata), that of the top right panel uses 30 classes, and those of the bottom left and right panels use 50 and 100 classes, respectively. For the histogram of the bottom right panel, a normal density fitted to the data is added (blue line). Note that the normal density does not seem to fit the data well. 5 / 37
6 The Kernel Density Estimator Consider now the local estimation of a density function. For some intuition of the kernel density estimator, note first that the following holds by the definition of a derivative: f (x) = d dx F (x) = lim h 0 F (x + h) F (x h) 2h The numerator of the last term can be rewritten as F (x + h) F (x h) = Pr(X (x h, x + h)). Define now the following kernel function: { 1/2 if z 1, k(z) = 0 otherwise. The kernel function can be viewed as a weighting function. From this definition of k(z), it follows that ( ) Xi x k = 1 h 2 1{X i (x h, x + h)}. See the next slides for a proof of this claim. 6 / 37
7 The Kernel Density Estimator Consider first the following general results: For some a and ε > 0, a < ε a < ε and a < ε, ε < a < ε. The first claim holds as a a < ε a < ε, a a < ε a < ε. By a < ε ε < a, and by combining the inequalities a < ε and ε < a, the second claim ( ε < a < ε) follows. Next, as h is positive, it holds that X i x h < 1 X i x < h. By the general inequality just derived (with a = X i x and ε = h), it holds that X i x < h h < X i x < h x h < X i < x + h. 7 / 37
8 The Kernel Density Estimator It follows therefore that ( ) { Xi x 1/2 if Xi x h x h < X k = i < x + h, h 0 otherwise, = 1 2 1{X i (x h, x + h)}. Now, recall that for some event A, Pr(A) = E[1{A}], where the expectation can be estimated by Ê[1{A}] = 1 n 1{A i }. n Using these results, the kernel density estimator of f X (x) is given by ˆf (x) = 1 n ( ) Xi x k. nh h The factor h 1 originates from the definition of f (x) as the derivative of the cdf F (x). 8 / 37
9 The Kernel Density Estimator The parameter h is called the bandwidth (or smoothing parameter or window width) of the kernel density estimator. The bandwidth is the most important parameter of the kernel density estimation problem. Note that ˆf (x) is the value of the density of X at some given point x, that is, it is a local estimate. To obtain an estimate of f for each value of the support of X, estimates for several points x supp(x ) are necessary. The kernel used above is called the uniform kernel. Different kernel functions may be used, as long as they have certain properties (which are listed later). Examples for kernel functions are the standard normal density, i.e., k(z) = φ(z), or the Epanechnikov kernel, which is defined as { 3 k(z) = 4 (1 z2 ) if z 1, 0 otherwise. The choice of the kernel function is of minor importance in practice. The choice of h has a much larger impact on the estimation results. 9 / 37
10 The Kernel Density Estimator The graphs show kernel density estimates of the same variable as in the previous example. The top left panel shows the estimates using the default bandwidth of the kdensity command. The graph in the top right panel adds a normal density function (red line) to the default kernel density estimate. The bottom left and right panels use a smaller and a larger bandwidth, respectively (compared to the default value of the kdensity command). 10 / 37
11 The Kernel Density Estimator Consider now the following general properties which have to hold for a kernel function. A kernel function k integrates to one, i.e., k(v)dv = 1, is symmetric, k(v) = k( v), and has a finite second moment, i.e., v 2 k(v)dv = κ 2 <. Note that by the symmetry condition it holds that vk(v)dv = 0. Every kernel function with these properties can be used for a kernel-based local estimation method. 11 / 37
12 The Kernel Density Estimator To proceed, consider first the definition of the mean squared error. ) The mean squared error MSE(ˆθ of an estimator ˆθ of θ is defined as ) [ ) ] [ 2 [ˆθ] [ˆθ] ) ] 2 MSE (ˆθ = E (ˆθ θ = E (ˆθ E + E θ [ ]) 2 [ˆθ]) ( [ˆθ] ) ( [ˆθ] ) ] 2 = E (ˆθ E [ˆθ + 2 (ˆθ E E θ + E θ [ [ˆθ]) ] 2 ( [ˆθ] 2 = E (ˆθ E + E θ), as it holds for the cross-product term that [ˆθ]) ( [ˆθ] )] ( [ˆθ] ) E [(ˆθ E E θ = E θ E [(ˆθ E [ˆθ])] = 0 The MSE can be rewritten as ) ) ( (ˆθ)) 2 MSE (ˆθ = Var (ˆθ + bias, ) [ˆθ] where the bias term is defined by bias (ˆθ = E θ. MSE convergence implies convergence in probability. 12 / 37
13 The Kernel Density Estimator To derive asymptotic properties of the kernel density estimator, the following assumptions are used. The data {X i } n is an iid sample. This is a standard assumption. The true density f (x) is three times differentiable, where the derivatives are denoted by f (s) (x), with s {1, 2, 3}. The derivatives of the density are used for a Taylor series expansion. The kernel function satisfies the assumptions stated previously. By these properties, some elements of the Taylor series expansion are equal to zero. The point x at which the the density is estimated is an interior point of the support of X. For an element at the boundary of the support of X, certain terms of the expansion are unequal to zero, which leads to a larger bias. For n, h 0 and nh. This assumption means that the bandwidth h converges slower to zero than n tends to infinity. As n, observations of a smaller environment of x are used for estimation (local approach). 13 / 37
14 The Kernel Density Estimator To describe the results, the order notation is useful. In the following, let a n and b n denote some sequences, and consider the case of n. Boundedness can be described by the O( )-notation: a n = O(1) a n < C for some positive constant C (and n ). More general, a n = O(b n ) a n b n < C. Convergence to zero is denoted by a n = o(b n ) a n /b n 0. A special case is b n 1, i.e., a n = o(1) a n 0. This notation is used to abbreviate terms which converge faster to zero or to some constant than the other remaining terms. For random variables X n, X n = O p (1) denotes Pr( X n > M) ε for n. X n p 0 is denoted by Xn = o p (1). O p (Y n ) and o p (Y n ) are defined similarly as O(Y n ) and o(y n ). 14 / 37
15 The Kernel Density Estimator Now, consider the MSE of the kernel density estimator at point x, which is given by ) ) ( 2 MSE (ˆf (x) = Var (ˆf (x) + bias (ˆf (x))). First, the bias term is analyzed, which is equal to [ 1 n ( ) ] Xi x E k f (x) = 1 [ ( )] nh h h E X1 x k f (x). h This equality follows by the identical distribution of the sample (which is expressed by setting E[X i ] = E[X 1 ] for all i): [ 1 n ( ) ] Xi x E k = 1 n [ ( )] Xi x E k nh h nh h = 1 n [ ( )] X1 x E k = 1 [ ( )] nh h nh ne X1 x k h = 1 ( )] [k h E X1 x. h 15 / 37
16 The Kernel Density Estimator Next, by the definition of the expectation, the bias can be written as [ ( )] 1 h E X1 x k f (x) = 1 ( ) x1 x k f (x 1 )dx 1 f (x) h h h Consider now the following change of the integration variable: x 1 x h = v x 1 = x + vh. Using this, the bias can be rewritten as follows (see the next slide for a derivation): 1 k(v)f (x + vh)hdv f (x) = k(v)f (x + vh)dv f (x). h Note that by this reformulation, the properties of the kernel function can be used to derive the properties of the bias term. Furthermore, a suitable Taylor expansion of f (x + vh) leads to an expression which is interpretable using the basic assumptions stated above. 16 / 37
17 The Kernel Density Estimator Consider now the following equality just stated: ( ) 1 x1 x k f (x 1 )dx 1 = 1 k(v)f (x + vh)hdv. h h h To see this, the formula for integration by substitution is used: For suitable functions w and g it holds that g(b) g(a) w(x)dx = b a w(g(t))g (t)dt. For the problem at hand, w(x 1 ) = k((x 1 x)/h)f (x 1 ) and g(v) = x + vh (= x 1 ). Let Ī = and I = be the upper and lower bounds of the integral. Therefore, Ī = g(b) and hence b = g 1 (Ī ) = (Ī x)/h. For Ī = (and similarly for I = ), b = g 1 (Ī ) =. For finite Ī and/or I, boundaries may change due to the substitution. Using g (v) = h, the equality claimed above follows. 17 / 37
18 The Kernel Density Estimator A Taylor expansion of a m-times differentiable function g(x) at some point x 0 is given by g(x) = g(x 0 ) + g (1) (x 0 )(x x 0 ) + 1 2! g (2) (x 0 )(x x 0 ) m! g (m) (ξ)(x x 0 ) m, where g (j) (x 0 ) = j g(x)/( x) j x=x0 and ξ lies between x and x 0. Now, consider a Taylor expansion of the density f (x + vh) at the point x: f (x + vh) = f (x) + f (1) (x)(x + vh x) f (2) (x)(x + vh x) = f (x) + f (1) (x)vh f (2) (x)v 2 h For the next term in the expansion it holds that 1 3! f (3) (x)v 3 h 3 = O(h 3 ), as f (3) (x) and v 3 are bounded (by assumption). 18 / 37
19 The Kernel Density Estimator Now insert f (x + vh) = f (x) + f (1) (x)vh f (2) (x)v 2 h 2 + O(h 3 ) into the bias expression: k(v)f (x + vh)dv f (x) ( = k(v) f (x) + f (1) (x)vh + 1 ) 2 f (2) (x)v 2 h 2 + O(h 3 ) By using this expansion, the term f (x + vh) can be expressed by terms involving x alone, which enables the computation of the integral with respect to v. dv f (x). The bias is therefore equal to k(v)f (x + vh)dv f (x) = f (x) k(v)dv + f (1) (x)h vk(v)dv f (2) (x)h 2 v 2 k(v)dv + O(h 3 ) f (x) 19 / 37
20 The Kernel Density Estimator As it holds by the properties of a kernel function that k(v)dv = 1 and vk(v)dv = 0, the bias is equal to ) bias (ˆf (x) = h2 2 f (2) (x) v 2 k(v)dv + O(h 3 ). A similar expression can be derived for the variance of the kernel density estimator at point x. Combining these results, the MSE of the kernel density estimator is given by ) ( ) ( MSE (ˆf (x) = h4 2 κ 2 f (2) κf (x) (x) + + o h ), 4 nh nh where κ = k 2 (v)dv and κ 2 = v 2 k(v)dv. From this result pointwise consistency of the kernel density estimator follows (i.e., for a given point x). 20 / 37
21 The Kernel Density Estimator To compute the kernel density estimator, a value of h is needed. It can be shown that the MSE of ˆf (x) is minimized by the following bandwidth: ( ) 1/5 κf (x) h opt = n 1/5. (κ 2 f (2) (x)) 2 This expression depends on κ and κ 2, which can be computed by knowledge of the used kernel function, and on the true density function f (x) and its second derivative, which are unknown. One simple possibility of this problem is to use a pilot bandwidth h pilot to estimate f (x) and f (2) (x) nonparametrically. Of course, this bandwidth needs also to be chosen. A solution of this is to assume (for deriving a pilot bandwidth) that f (x) is a normally distributed with variance σ 2. Then one can derive h pilot 1.06σn 1/5 (Silverman s rule of thumb). This expression can be used to estimate the expressions needed to derive h opt. Often h pilot is used directly as h opt. Various other methods for bandwidth choice exist. 21 / 37
22 The Kernel Density Estimator Kernel density estimators can also be used to estimate multivariate densities. Given a sample {X i } n, where X i R q and q > 1, the density f (x) = f (x 1, x 2,..., x q ) can be estimated by 1 n ( ) Xi x ˆf (x) = K nh 1... h q h 1 n ( ) ( ) Xi1 x 1 Xiq x q = k k. nh 1... h q h 1 h q where K( ) is called a product kernel, and the functions k( ) are univariate kernels (as used previously). Under some assumptions, it can be shown that ) q nh1... h q (ˆf (x) f (x) κ2 hs 2 f ss (x) 2 s=1 Here, f ss is some derivative of the density f. Note that an asymptotic bias term occurs here. d N (0, κ q f (x)). 22 / 37
23 The Kernel Density Estimator ˆf (x) is (for univariate and multivariate estimators alike) a consistent estimator of f (x). From the expression of the asymptotic distribution it follows that the speed of convergence of nonparametric kernel density estimates is much smaller than for parametric estimators, and decreases with an increasing number of variables, i.e., with increasing q. To understand the concept of convergence speed, consider some parametric estimator ˆθ, for which ˆθ p θ 0 (that is, ˆθ θ 0 = o P (1)). Multiplying this difference by n leads to an expression which does not tend to infinity or to zero, but to a random variable. n is called the speed of convergence of the estimator. As can be seen from the expression of the previous slide, the kernel density estimator has a much slower speed, as nh 1... h q (or nh in the univariate case) is smaller than n, as the bandwidths converge to zero. In practice, this means that nonparametric methods should only be applied if large samples are available. 23 / 37
24 Nonparametric Regression Consider the general regression model y i = g(x i ) + ε i, where g(x) = E[y X = x], and no assumptions on the form of g(x) are imposed. The conditional expectation of y is defined by E[y X = x] = yf Y X (y x)dy = y f Y,X (y, x) dy = f X (x) where the second equality follows by Bayes law. The denominator f X (x) can be estimated by a nonparametric density estimator. Consider now the following kernel estimator of f y,x (y, x): 1 n ( ) ( ) Xi x yi y ˆf Y,X (y, x) = K k, nh 0 h 1... h q h h 0 yfy,x (y, x)dy f X (x) where K( ) is the product kernel for multivariate covariates defined earlier., 24 / 37
25 Nonparametric Regression Consider now the numerator of the estimator of g(x), i.e., consider yfy,x (y, x)dy with f Y,X replaced by the estimator ˆf Y,X just defined: yˆf Y,X (y, x)dy = = 1 n ( ) ( ) Xi x yi y y K k dy nh 0 h 1... h q h h 0 1 n ( ) ( ) Xi x yi y K yk dy. nh 0 h 1... h q h h 0 By a change of variables ((y i y)/h 0 = v y = y i vh 0 ), the right hand side of this expression can be rewritten as 1 n ( ) Xi x K (y i vh 0 )k(v)h 0 dv. nh 0 h 1... h q h As k(v)dv = 1 and vk(v)dv = 0, this simplifies to 1 n ( ) Xi x K y i. nh 1... h q h 25 / 37
26 Nonparametric Regression Using this result, the estimator of g(x) can be rewritten as: n yˆfy,x (y, x)dy ĝ(x) = = K ( X i ) x h yi n ˆf X (x) K ( X i ). x h This estimator is called local constant regression or Nadaraya- Watson estimator. Note that ĝ(x) is a weighted average of y: n ĝ(x) = ω i y i where the weights ω i are defined by K ( X i ) x h ω i = ( ), n j=1 K Xj x h and have the properties ω i 0 and n ω i = 1. The nonparametric regression estimator can be viewed as a weighted average of y, where the weights depend on the distance of the covariates X i to the point x at which ĝ(x) is estimated. 26 / 37
27 Nonparametric Regression Under some assumptions, the following holds: ( ) q nh1... h q ĝ(x) g(x) hs 2 B s (x) s=1 Again, an asymptotic bias term occurs. d N ( 0, κq σ 2 ) (x). f (x) The bandwidth parameter(s) can be determined by several methods, for example by the plug-in approach or by cross-validation. ĝ(x) can be expressed as the solution of an optimization problem: n ( ) ĝ(x) = arg min (y i µ) 2 Xi x K. µ h That is, the nonparametric regression estimate at point x is the constant of a weighted OLS regression without further covariates. The covariates appear only in the kernel function, i.e., they are only used for computing the weights K( ). Better asymptotic properties may be obtained by more general models, which will be discussed next. 27 / 37
28 Nonparametric Regression Consider as a first extension of the local constant regression model the local linear regression model. Here, a local linear instead of a local constant function is used to approximate the unknown g(x). The local linear regression at a point x is determined by the following optimization problem: ( ) â(x) n ( ˆδ(x) = = arg min yi a (X ˆb(x) i x) b ) ( 2 Xi x k a,b h To derive a closed form expression of the estimators, define the following [n n] and [n (k + 1)] dimensional matrices: ( ( ) ( )) X1 x Xn x K x = diag k,..., k, h h X x = (1, (X i x) ),...,n. ). The optimization problem can be restated as ˆδ(x) = arg min δ(x) (y X xδ(x)) K x (y X x δ(x)). 28 / 37
29 Nonparametric Regression The solution can be expressed explicitely as ˆδ(x) = (X xk x X x ) 1 X xk x y. The estimate ĝ(x) of the mean function at point x is equal to â(x). A further generalization is to approximate g(x) locally by a polynomial of order p. For univariate X, the objective function of the local parameters δ(x) = (δ 0 (x),..., δ p (x)) is given by n ( ( ˆδ(x) = arg min yi δ 0 (X i x)δ 1... (X i x) p ) 2k Xi x δ p δ h In this case, the covariate matrix X x is defined as: 1 X 1 x (X 1 x) 2... (X 1 x) p 1 X 2 x (X 2 x) 2... (X 2 x) p X x = X n x (X n x) 2... (X n x) p. ). 29 / 37
30 Nonparametric Regression With K x as defined above, the parameter estimators are given by ˆδ(x) = ( X xk x X x ) 1X x K x y. Again, ĝ(x) is equal to the (weighted) constant of the polynomial expression above, i.e., ĝ(x) = ˆδ 0 (x), where δ 0 (x) is the first element of the vector δ(x). The additional parameters ˆδ s for s = 1,..., p are estimators of the derivatives of the regression function, i.e., δ s (x) = 1 s! s g(x) x s. In fact, the polynomial expression used above corresponds to a Taylor approximation of g(x), which explains the factor 1 s!. When estimating the s th derivative of g(x), p s should be odd to lower the bias of the estimate. The advantage of local polynomial regression models are smaller bias terms compared to local constant regressions. Extensions for multivariate covariates exist, which have a somewhat involved notation. 30 / 37
31 Nonparametric Regression For all nonparametric methods presented here (kernel density and local regression estimators), it is assumed that the covariates are continuous. Therefore, indicator variables and discrete ordered variables can not be used directly for nonparametric regression. A possibility to use multivariate nonparametric estimators in the presence of discrete variables is to divide the data set into cells of different values of the discrete variables and to compute the nonparametric estimators using the remaining continuous variables. As an example, consider the regression estimate for a dependent variable y and two covariates x 1 and x 2. If x 1 and x 2 are continuous, ˆm(y x 1, x 2 ) is a function with two (continuous) arguments. Assume now that x 2 is an indicator variable. The procedure above computes two regressions ˆm(y x 1, x 2 = 1) and ˆm(y x 1, x 2 = 0). The problem of this approach is that the number of observations for each cell can be quite low when there are several discrete variables. If the data set contains only discrete variables, the nonparametric regressions correspond to the cell means of y. 31 / 37
32 Semi- and Nonparametric Extensions The convergence rate of nonparametric regression estimators decreases with an increasing number of regressors (curse of dimensionality). A solution of this problem is to impose some further structure. One possibility for such restrictions are additive models, which model the conditional mean of the dependent variable by a sum of univariate unknown functions of the covariates. A second class of models which circumvent the curse of dimensionality are semiparametric models, which assume a parametric specification for a part of the model. Consider first the basic additive nonparametric model: E[y x] = g 1 (x 1 ) + g 2 (x 2 ) g k (x k ). Here, the rate of convergence is equal to that of an univariate nonparametric regression, which is faster as in the case of a general multivariate nonparametric model. The unknown functions g j (x j ) can be estimated by a backfitting algorithm, for example. 32 / 37
33 Semi- and Nonparametric Extensions A second possibility to circumvent the curse of dimensionality is to impose some parametric structure for a part of the model. One example of this model class is the partial linear model, which uses an unknown function g(x 2 ) to specify the true model as y = x 1β + g(x 2 ) + ε. To estimate the model, consider first the expectation of y given x 2 : E[y x 2 ] = E[x 1 x 2 ] β + g(x 2 ). Now, the following difference does not contain g(x 2 ) and can be estimated by OLS to obtain an estimator of β: ( y Ê[y x 2] = x 1 Ê[x 1 x 2 ]) β + ε. The nonparametric part can be estimated by ĝ(x 2 ) = Ê[y x 2] Ê[x 1 x 2 ] ˆβ. A further example is the single-index model, which uses a linear index of X and an unknown function g( ) to specify the conditional mean as E[y x] = g(x β). 33 / 37
34 Review and Definitions Parametric models depend on a finite dimensional parameter, which is an element of R q for q N. Nonparametric models are based on one (or several) parameter(s) of infinite dimension, that is, the parameter(s) cannot be represented by a finite dimensional value, i.e., by a scalar (or a vector thereof). The parameter of a nonparametric model is a function, like a conditional expectation (i.e., a regression) or a probability density. Nonparametric models weaken functional form assumptions (like the linearity assumptions of the OLS model). Nonlinear models can also be estimated by OLS or by NLS, however. The difference to nonparametric models is that the structure of the nonlinearity is given by the parametric form, i.e., by assumption. A broad class of nonparametric models (all of those presented here) are based on a local estimation approach, where the estimation is based mainly on observations near to the point of interest. There are also nonparametric models which are not based on a local approach (e.g., the Nelson-Aalen or Kaplan-Meier estimators). 34 / 37
35 Review and Definitions Semiparametric estimators depend on finite as well as on infinite dimensional parameters. Examples are the partial linear and singles index models presented previously. Two-step estimators with nonparametric first-step estimates can also be viewed as semiparametric estimators. The parametric part of these estimators is the (scalar) second-step estimator, which is usually an average of the nonparametric first-step estimates evaluated for all observations of the sample. Examples are nonparametric matching or reweighting evaluation estimators (see the following lecture on evalution methods). These estimators are averages of nonparametrically estimated functions. Nevertheless, they are n-consistent, as they have the structure of an U-statistic. For more on U-statistics, see Pagan and Ullah, Nonparametric Econometrics, Cambridge University Press 1999, p. 358, or Powell, Stock, and Stoker (Econometrica, 1989). 35 / 37
36 Summary Parametric specifications of regression equations and density functions may lead to inconsistent estimates. Nonparametric models circumvent the need for distributional or functional form assumptions. A large number of nonparametric methods are based on the local estimation approach. Density functions can be estimated by kernel density estimators. A basic local nonparametric regression estimator is the local constant (or Nadaraya-Watson) regression estimator. The local linear and local polynomial regression models are generalizations of the Nadaraya-Watson regression estimator and improve the asymptotic properties. Several semi- and nonparametric extensions of the basic models were proposed to address the curse of dimensionality. 36 / 37
37 Basic and Additional References Basic references: Cameron and Trivedi (2005), ch. 9, Cameron and Trivedi (2009), sec General textbooks on nonparametric methods: Q. Li and J. S. Racine, Nonparametric Econometrics, Princeton University Press, A. Pagan and A. Ullah, Nonparametric Econometrics, Cambridge University Press, D. Ruppert, M.P. Wand, and R.J. Carroll, Semiparametric Regression, Cambridge University Press, A. Yatchew, Semiparametric Regression for the Applied Econometrician, Cambridge University Press, Survey articles: J. DiNardo and J. L. Tobias, Nonparametric Density and Regression Estimation, Journal of Economic Perspectives, vol. 15(4), Fall 2001, p A. Yatchew, Nonparametric Regression Techniques in Economics, Journal of Economic Literature, Vol. 36 (1998), p / 37
Density estimation Nonparametric conditional mean estimation Semiparametric conditional mean estimation. Nonparametrics. Gabriel Montes-Rojas
0 0 5 Motivation: Regression discontinuity (Angrist&Pischke) Outcome.5 1 1.5 A. Linear E[Y 0i X i] 0.2.4.6.8 1 X Outcome.5 1 1.5 B. Nonlinear E[Y 0i X i] i 0.2.4.6.8 1 X utcome.5 1 1.5 C. Nonlinearity
More informationNonparametric Regression
Nonparametric Regression Econ 674 Purdue University April 8, 2009 Justin L. Tobias (Purdue) Nonparametric Regression April 8, 2009 1 / 31 Consider the univariate nonparametric regression model: where y
More informationNonparametric Methods
Nonparametric Methods Michael R. Roberts Department of Finance The Wharton School University of Pennsylvania July 28, 2009 Michael R. Roberts Nonparametric Methods 1/42 Overview Great for data analysis
More information41903: Introduction to Nonparametrics
41903: Notes 5 Introduction Nonparametrics fundamentally about fitting flexible models: want model that is flexible enough to accommodate important patterns but not so flexible it overspecializes to specific
More informationNonparametric Density Estimation
Nonparametric Density Estimation Econ 690 Purdue University Justin L. Tobias (Purdue) Nonparametric Density Estimation 1 / 29 Density Estimation Suppose that you had some data, say on wages, and you wanted
More informationLocal Polynomial Regression
VI Local Polynomial Regression (1) Global polynomial regression We observe random pairs (X 1, Y 1 ),, (X n, Y n ) where (X 1, Y 1 ),, (X n, Y n ) iid (X, Y ). We want to estimate m(x) = E(Y X = x) based
More informationTime Series and Forecasting Lecture 4 NonLinear Time Series
Time Series and Forecasting Lecture 4 NonLinear Time Series Bruce E. Hansen Summer School in Economics and Econometrics University of Crete July 23-27, 2012 Bruce Hansen (University of Wisconsin) Foundations
More informationNonparametric Regression. Changliang Zou
Nonparametric Regression Institute of Statistics, Nankai University Email: nk.chlzou@gmail.com Smoothing parameter selection An overall measure of how well m h (x) performs in estimating m(x) over x (0,
More informationEcon 582 Nonparametric Regression
Econ 582 Nonparametric Regression Eric Zivot May 28, 2013 Nonparametric Regression Sofarwehaveonlyconsideredlinearregressionmodels = x 0 β + [ x ]=0 [ x = x] =x 0 β = [ x = x] [ x = x] x = β The assume
More informationNonparametric Function Estimation with Infinite-Order Kernels
Nonparametric Function Estimation with Infinite-Order Kernels Arthur Berg Department of Statistics, University of Florida March 15, 2008 Kernel Density Estimation (IID Case) Let X 1,..., X n iid density
More informationLocal linear multiple regression with variable. bandwidth in the presence of heteroscedasticity
Local linear multiple regression with variable bandwidth in the presence of heteroscedasticity Azhong Ye 1 Rob J Hyndman 2 Zinai Li 3 23 January 2006 Abstract: We present local linear estimator with variable
More informationDay 4A Nonparametrics
Day 4A Nonparametrics A. Colin Cameron Univ. of Calif. - Davis... for Center of Labor Economics Norwegian School of Economics Advanced Microeconometrics Aug 28 - Sep 2, 2017. Colin Cameron Univ. of Calif.
More informationAn introduction to nonparametric and semi-parametric econometric methods
An introduction to nonparametric and semi-parametric econometric methods Robert Breunig Australian National University robert.breunig@anu.edu.au http://econrsss.anu.edu.au/staff/breunig/course_bb.htm March
More informationIntroduction. Linear Regression. coefficient estimates for the wage equation: E(Y X) = X 1 β X d β d = X β
Introduction - Introduction -2 Introduction Linear Regression E(Y X) = X β +...+X d β d = X β Example: Wage equation Y = log wages, X = schooling (measured in years), labor market experience (measured
More informationPreface. 1 Nonparametric Density Estimation and Testing. 1.1 Introduction. 1.2 Univariate Density Estimation
Preface Nonparametric econometrics has become one of the most important sub-fields in modern econometrics. The primary goal of this lecture note is to introduce various nonparametric and semiparametric
More informationEconomics 583: Econometric Theory I A Primer on Asymptotics
Economics 583: Econometric Theory I A Primer on Asymptotics Eric Zivot January 14, 2013 The two main concepts in asymptotic theory that we will use are Consistency Asymptotic Normality Intuition consistency:
More informationEstimation of Treatment Effects under Essential Heterogeneity
Estimation of Treatment Effects under Essential Heterogeneity James Heckman University of Chicago and American Bar Foundation Sergio Urzua University of Chicago Edward Vytlacil Columbia University March
More informationDay 3B Nonparametrics and Bootstrap
Day 3B Nonparametrics and Bootstrap c A. Colin Cameron Univ. of Calif.- Davis Frontiers in Econometrics Bavarian Graduate Program in Economics. Based on A. Colin Cameron and Pravin K. Trivedi (2009,2010),
More informationEconomics 582 Random Effects Estimation
Economics 582 Random Effects Estimation Eric Zivot May 29, 2013 Random Effects Model Hence, the model can be re-written as = x 0 β + + [x ] = 0 (no endogeneity) [ x ] = = + x 0 β + + [x ] = 0 [ x ] = 0
More informationDo Markov-Switching Models Capture Nonlinearities in the Data? Tests using Nonparametric Methods
Do Markov-Switching Models Capture Nonlinearities in the Data? Tests using Nonparametric Methods Robert V. Breunig Centre for Economic Policy Research, Research School of Social Sciences and School of
More informationEconomics 620, Lecture 19: Introduction to Nonparametric and Semiparametric Estimation
Economics 620, Lecture 19: Introduction to Nonparametric and Semiparametric Estimation Nicholas M. Kiefer Cornell University Professor N. M. Kiefer (Cornell University) Lecture 19: Nonparametric Analysis
More informationECON 721: Lecture Notes on Nonparametric Density and Regression Estimation. Petra E. Todd
ECON 721: Lecture Notes on Nonparametric Density and Regression Estimation Petra E. Todd Fall, 2014 2 Contents 1 Review of Stochastic Order Symbols 1 2 Nonparametric Density Estimation 3 2.1 Histogram
More informationSingle Index Quantile Regression for Heteroscedastic Data
Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University JSM, 2015 E. Christou, M. G. Akritas (PSU) SIQR JSM, 2015
More informationStatistics: Learning models from data
DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial
More informationLecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16)
Lecture: Simultaneous Equation Model (Wooldridge s Book Chapter 16) 1 2 Model Consider a system of two regressions y 1 = β 1 y 2 + u 1 (1) y 2 = β 2 y 1 + u 2 (2) This is a simultaneous equation model
More informationA nonparametric method of multi-step ahead forecasting in diffusion processes
A nonparametric method of multi-step ahead forecasting in diffusion processes Mariko Yamamura a, Isao Shoji b a School of Pharmacy, Kitasato University, Minato-ku, Tokyo, 108-8641, Japan. b Graduate School
More informationNon-parametric Inference and Resampling
Non-parametric Inference and Resampling Exercises by David Wozabal (Last update 3. Juni 2013) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend
More informationModelling Non-linear and Non-stationary Time Series
Modelling Non-linear and Non-stationary Time Series Chapter 2: Non-parametric methods Henrik Madsen Advanced Time Series Analysis September 206 Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September
More informationECON 3150/4150, Spring term Lecture 6
ECON 3150/4150, Spring term 2013. Lecture 6 Review of theoretical statistics for econometric modelling (II) Ragnar Nymoen University of Oslo 31 January 2013 1 / 25 References to Lecture 3 and 6 Lecture
More informationNonparametric Density Estimation (Multidimension)
Nonparametric Density Estimation (Multidimension) Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction Tine Buch-Kromann February 19, 2007 Setup One-dimensional
More informationRank Estimation of Partially Linear Index Models
Rank Estimation of Partially Linear Index Models Jason Abrevaya University of Texas at Austin Youngki Shin University of Western Ontario October 2008 Preliminary Do not distribute Abstract We consider
More informationIntroduction to Regression
Introduction to Regression p. 1/97 Introduction to Regression Chad Schafer cschafer@stat.cmu.edu Carnegie Mellon University Introduction to Regression p. 1/97 Acknowledgement Larry Wasserman, All of Nonparametric
More informationECO Class 6 Nonparametric Econometrics
ECO 523 - Class 6 Nonparametric Econometrics Carolina Caetano Contents 1 Nonparametric instrumental variable regression 1 2 Nonparametric Estimation of Average Treatment Effects 3 2.1 Asymptotic results................................
More informationNonparametric Regression. Badr Missaoui
Badr Missaoui Outline Kernel and local polynomial regression. Penalized regression. We are given n pairs of observations (X 1, Y 1 ),...,(X n, Y n ) where Y i = r(x i ) + ε i, i = 1,..., n and r(x) = E(Y
More informationCURRENT STATUS LINEAR REGRESSION. By Piet Groeneboom and Kim Hendrickx Delft University of Technology and Hasselt University
CURRENT STATUS LINEAR REGRESSION By Piet Groeneboom and Kim Hendrickx Delft University of Technology and Hasselt University We construct n-consistent and asymptotically normal estimates for the finite
More informationNonparametric Estimation of Regression Functions In the Presence of Irrelevant Regressors
Nonparametric Estimation of Regression Functions In the Presence of Irrelevant Regressors Peter Hall, Qi Li, Jeff Racine 1 Introduction Nonparametric techniques robust to functional form specification.
More informationGenerated Covariates in Nonparametric Estimation: A Short Review.
Generated Covariates in Nonparametric Estimation: A Short Review. Enno Mammen, Christoph Rothe, and Melanie Schienle Abstract In many applications, covariates are not observed but have to be estimated
More informationAPEC 8212: Econometric Analysis II
APEC 8212: Econometric Analysis II Instructor: Paul Glewwe Spring, 2014 Office: 337a Ruttan Hall (formerly Classroom Office Building) Phone: 612-625-0225 E-Mail: pglewwe@umn.edu Class Website: http://faculty.apec.umn.edu/pglewwe/apec8212.html
More informationFrom Histograms to Multivariate Polynomial Histograms and Shape Estimation. Assoc Prof Inge Koch
From Histograms to Multivariate Polynomial Histograms and Shape Estimation Assoc Prof Inge Koch Statistics, School of Mathematical Sciences University of Adelaide Inge Koch (UNSW, Adelaide) Poly Histograms
More informationExam C Solutions Spring 2005
Exam C Solutions Spring 005 Question # The CDF is F( x) = 4 ( + x) Observation (x) F(x) compare to: Maximum difference 0. 0.58 0, 0. 0.58 0.7 0.880 0., 0.4 0.680 0.9 0.93 0.4, 0.6 0.53. 0.949 0.6, 0.8
More informationIntroduction to Nonparametric and Semiparametric Estimation. Good when there are lots of data and very little prior information on functional form.
1 Introduction to Nonparametric and Semiparametric Estimation Good when there are lots of data and very little prior information on functional form. Examples: y = f(x) + " (nonparametric) y = z 0 + f(x)
More informationNonparametric Modal Regression
Nonparametric Modal Regression Summary In this article, we propose a new nonparametric modal regression model, which aims to estimate the mode of the conditional density of Y given predictors X. The nonparametric
More informationRegression #3: Properties of OLS Estimator
Regression #3: Properties of OLS Estimator Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #3 1 / 20 Introduction In this lecture, we establish some desirable properties associated with
More informationPenalized Splines, Mixed Models, and Recent Large-Sample Results
Penalized Splines, Mixed Models, and Recent Large-Sample Results David Ruppert Operations Research & Information Engineering, Cornell University Feb 4, 2011 Collaborators Matt Wand, University of Wollongong
More informationSingle Index Quantile Regression for Heteroscedastic Data
Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University SMAC, November 6, 2015 E. Christou, M. G. Akritas (PSU) SIQR
More informationEC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)
1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For
More informationDESIGN-ADAPTIVE MINIMAX LOCAL LINEAR REGRESSION FOR LONGITUDINAL/CLUSTERED DATA
Statistica Sinica 18(2008), 515-534 DESIGN-ADAPTIVE MINIMAX LOCAL LINEAR REGRESSION FOR LONGITUDINAL/CLUSTERED DATA Kani Chen 1, Jianqing Fan 2 and Zhezhen Jin 3 1 Hong Kong University of Science and Technology,
More informationOn the Robust Modal Local Polynomial Regression
International Journal of Statistical Sciences ISSN 683 5603 Vol. 9(Special Issue), 2009, pp 27-23 c 2009 Dept. of Statistics, Univ. of Rajshahi, Bangladesh On the Robust Modal Local Polynomial Regression
More informationSome Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model
Some Theories about Backfitting Algorithm for Varying Coefficient Partially Linear Model 1. Introduction Varying-coefficient partially linear model (Zhang, Lee, and Song, 2002; Xia, Zhang, and Tong, 2004;
More informationModel-free prediction intervals for regression and autoregression. Dimitris N. Politis University of California, San Diego
Model-free prediction intervals for regression and autoregression Dimitris N. Politis University of California, San Diego To explain or to predict? Models are indispensable for exploring/utilizing relationships
More informationTeruko Takada Department of Economics, University of Illinois. Abstract
Nonparametric density estimation: A comparative study Teruko Takada Department of Economics, University of Illinois Abstract Motivated by finance applications, the objective of this paper is to assess
More informationNonparametric Regression Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction
Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction Tine Buch-Kromann Univariate Kernel Regression The relationship between two variables, X and Y where m(
More informationECON Introductory Econometrics. Lecture 16: Instrumental variables
ECON4150 - Introductory Econometrics Lecture 16: Instrumental variables Monique de Haan (moniqued@econ.uio.no) Stock and Watson Chapter 12 Lecture outline 2 OLS assumptions and when they are violated Instrumental
More informationIntroduction to Estimation Methods for Time Series models Lecture 2
Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:
More informationSome Monte Carlo Evidence for Adaptive Estimation of Unit-Time Varying Heteroscedastic Panel Data Models
Some Monte Carlo Evidence for Adaptive Estimation of Unit-Time Varying Heteroscedastic Panel Data Models G. R. Pasha Department of Statistics, Bahauddin Zakariya University Multan, Pakistan E-mail: drpasha@bzu.edu.pk
More informationLecture 3: Statistical Decision Theory (Part II)
Lecture 3: Statistical Decision Theory (Part II) Hao Helen Zhang Hao Helen Zhang Lecture 3: Statistical Decision Theory (Part II) 1 / 27 Outline of This Note Part I: Statistics Decision Theory (Classical
More informationQuantile Regression for Panel Data Models with Fixed Effects and Small T : Identification and Estimation
Quantile Regression for Panel Data Models with Fixed Effects and Small T : Identification and Estimation Maria Ponomareva University of Western Ontario May 8, 2011 Abstract This paper proposes a moments-based
More information1 Glivenko-Cantelli type theorems
STA79 Lecture Spring Semester Glivenko-Cantelli type theorems Given i.i.d. observations X,..., X n with unknown distribution function F (t, consider the empirical (sample CDF ˆF n (t = I [Xi t]. n Then
More informationMinimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model.
Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model By Michael Levine Purdue University Technical Report #14-03 Department of
More informationconditional cdf, conditional pdf, total probability theorem?
6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random
More informationUltra High Dimensional Variable Selection with Endogenous Variables
1 / 39 Ultra High Dimensional Variable Selection with Endogenous Variables Yuan Liao Princeton University Joint work with Jianqing Fan Job Market Talk January, 2012 2 / 39 Outline 1 Examples of Ultra High
More informationSEMIPARAMETRIC APPLICATIONS IN ECONOMIC GROWTH. Mustafa Koroglu. A Thesis presented to The University of Guelph
SEMIPARAMETRIC APPLICATIONS IN ECONOMIC GROWTH by Mustafa Koroglu A Thesis presented to The University of Guelph In partial fulfilment of requirements for the degree of Doctor of Philosophy in Economics
More informationIntroduction to Regression
Introduction to Regression David E Jones (slides mostly by Chad M Schafer) June 1, 2016 1 / 102 Outline General Concepts of Regression, Bias-Variance Tradeoff Linear Regression Nonparametric Procedures
More informationQuantile methods. Class Notes Manuel Arellano December 1, Let F (r) =Pr(Y r). Forτ (0, 1), theτth population quantile of Y is defined to be
Quantile methods Class Notes Manuel Arellano December 1, 2009 1 Unconditional quantiles Let F (r) =Pr(Y r). Forτ (0, 1), theτth population quantile of Y is defined to be Q τ (Y ) q τ F 1 (τ) =inf{r : F
More informationGoodness-of-fit tests for the cure rate in a mixture cure model
Biometrika (217), 13, 1, pp. 1 7 Printed in Great Britain Advance Access publication on 31 July 216 Goodness-of-fit tests for the cure rate in a mixture cure model BY U.U. MÜLLER Department of Statistics,
More informationBoundary Correction Methods in Kernel Density Estimation Tom Alberts C o u(r)a n (t) Institute joint work with R.J. Karunamuni University of Alberta
Boundary Correction Methods in Kernel Density Estimation Tom Alberts C o u(r)a n (t) Institute joint work with R.J. Karunamuni University of Alberta November 29, 2007 Outline Overview of Kernel Density
More informationNew Local Estimation Procedure for Nonparametric Regression Function of Longitudinal Data
ew Local Estimation Procedure for onparametric Regression Function of Longitudinal Data Weixin Yao and Runze Li The Pennsylvania State University Technical Report Series #0-03 College of Health and Human
More informationIntroduction to machine learning and pattern recognition Lecture 2 Coryn Bailer-Jones
Introduction to machine learning and pattern recognition Lecture 2 Coryn Bailer-Jones http://www.mpia.de/homes/calj/mlpr_mpia2008.html 1 1 Last week... supervised and unsupervised methods need adaptive
More informationNonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix
Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix Yingying Dong and Arthur Lewbel California State University Fullerton and Boston College July 2010 Abstract
More informationSTAT 512 sp 2018 Summary Sheet
STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}
More informationMotivational Example
Motivational Example Data: Observational longitudinal study of obesity from birth to adulthood. Overall Goal: Build age-, gender-, height-specific growth charts (under 3 year) to diagnose growth abnomalities.
More informationEstimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators
Estimation theory Parametric estimation Properties of estimators Minimum variance estimator Cramer-Rao bound Maximum likelihood estimators Confidence intervals Bayesian estimation 1 Random Variables Let
More informationConfidence intervals for kernel density estimation
Stata User Group - 9th UK meeting - 19/20 May 2003 Confidence intervals for kernel density estimation Carlo Fiorio c.fiorio@lse.ac.uk London School of Economics and STICERD Stata User Group - 9th UK meeting
More informationLocal regression I. Patrick Breheny. November 1. Kernel weighted averages Local linear regression
Local regression I Patrick Breheny November 1 Patrick Breheny STA 621: Nonparametric Statistics 1/27 Simple local models Kernel weighted averages The Nadaraya-Watson estimator Expected loss and prediction
More informationIntroduction to Regression
Introduction to Regression Chad M. Schafer May 20, 2015 Outline General Concepts of Regression, Bias-Variance Tradeoff Linear Regression Nonparametric Procedures Cross Validation Local Polynomial Regression
More informationAdditive Isotonic Regression
Additive Isotonic Regression Enno Mammen and Kyusang Yu 11. July 2006 INTRODUCTION: We have i.i.d. random vectors (Y 1, X 1 ),..., (Y n, X n ) with X i = (X1 i,..., X d i ) and we consider the additive
More informationNonparametric Identication of a Binary Random Factor in Cross Section Data and
. Nonparametric Identication of a Binary Random Factor in Cross Section Data and Returns to Lying? Identifying the Effects of Misreporting When the Truth is Unobserved Arthur Lewbel Boston College This
More information13 Endogeneity and Nonparametric IV
13 Endogeneity and Nonparametric IV 13.1 Nonparametric Endogeneity A nonparametric IV equation is Y i = g (X i ) + e i (1) E (e i j i ) = 0 In this model, some elements of X i are potentially endogenous,
More informationOn variable bandwidth kernel density estimation
JSM 04 - Section on Nonparametric Statistics On variable bandwidth kernel density estimation Janet Nakarmi Hailin Sang Abstract In this paper we study the ideal variable bandwidth kernel estimator introduced
More informationCross-fitting and fast remainder rates for semiparametric estimation
Cross-fitting and fast remainder rates for semiparametric estimation Whitney K. Newey James M. Robins The Institute for Fiscal Studies Department of Economics, UCL cemmap working paper CWP41/17 Cross-Fitting
More informationPeter Hoff Minimax estimation October 31, Motivation and definition. 2 Least favorable prior 3. 3 Least favorable prior sequence 11
Contents 1 Motivation and definition 1 2 Least favorable prior 3 3 Least favorable prior sequence 11 4 Nonparametric problems 15 5 Minimax and admissibility 18 6 Superefficiency and sparsity 19 Most of
More informationEstimation of partial effects in non-linear panel data models
Estimation of partial effects in non-linear panel data models by Jason Abrevaya and Yu-Chin Hsu This version: September 2011 ABSTRACT Nonlinearity and heterogeneity complicate the estimation and interpretation
More informationA Bootstrap Test for Conditional Symmetry
ANNALS OF ECONOMICS AND FINANCE 6, 51 61 005) A Bootstrap Test for Conditional Symmetry Liangjun Su Guanghua School of Management, Peking University E-mail: lsu@gsm.pku.edu.cn and Sainan Jin Guanghua School
More informationUNIVERSITY OF CALIFORNIA Spring Economics 241A Econometrics
DEPARTMENT OF ECONOMICS R. Smith, J. Powell UNIVERSITY OF CALIFORNIA Spring 2006 Economics 241A Econometrics This course will cover nonlinear statistical models for the analysis of cross-sectional and
More informationGaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012
Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature
More informationIdentification and Estimation of Partially Linear Censored Regression Models with Unknown Heteroscedasticity
Identification and Estimation of Partially Linear Censored Regression Models with Unknown Heteroscedasticity Zhengyu Zhang School of Economics Shanghai University of Finance and Economics zy.zhang@mail.shufe.edu.cn
More informationNonparametric Density Estimation
Nonparametric Density Estimation Advanced Econometrics Douglas G. Steigerwald UC Santa Barbara D. Steigerwald (UCSB) Density Estimation 1 / 20 Overview Question of interest has wage inequality among women
More informationEconomics 536 Lecture 7. Introduction to Specification Testing in Dynamic Econometric Models
University of Illinois Fall 2016 Department of Economics Roger Koenker Economics 536 Lecture 7 Introduction to Specification Testing in Dynamic Econometric Models In this lecture I want to briefly describe
More informationRobust covariance estimation for quantile regression
1 Robust covariance estimation for quantile regression J. M.C. Santos Silva School of Economics, University of Surrey UK STATA USERS GROUP, 21st MEETING, 10 Septeber 2015 2 Quantile regression (Koenker
More informationSemiparametric Models and Estimators
Semiparametric Models and Estimators Whitney Newey October 2007 Semiparametric Models Data: Z 1,Z 2,... i.i.d. Model: F aset of pdfs. Correct specification: pdf f 0 of Z i in F. Parametric model: F = {f(z
More informationON SOME TWO-STEP DENSITY ESTIMATION METHOD
UNIVESITATIS IAGELLONICAE ACTA MATHEMATICA, FASCICULUS XLIII 2005 ON SOME TWO-STEP DENSITY ESTIMATION METHOD by Jolanta Jarnicka Abstract. We introduce a new two-step kernel density estimation method,
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationMax. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes
Maximum Likelihood Estimation Econometrics II Department of Economics Universidad Carlos III de Madrid Máster Universitario en Desarrollo y Crecimiento Económico Outline 1 3 4 General Approaches to Parameter
More informationA Probability Review
A Probability Review Outline: A probability review Shorthand notation: RV stands for random variable EE 527, Detection and Estimation Theory, # 0b 1 A Probability Review Reading: Go over handouts 2 5 in
More informationMinimum Hellinger Distance Estimation in a. Semiparametric Mixture Model
Minimum Hellinger Distance Estimation in a Semiparametric Mixture Model Sijia Xiang 1, Weixin Yao 1, and Jingjing Wu 2 1 Department of Statistics, Kansas State University, Manhattan, Kansas, USA 66506-0802.
More informationNew Developments in Econometrics Lecture 16: Quantile Estimation
New Developments in Econometrics Lecture 16: Quantile Estimation Jeff Wooldridge Cemmap Lectures, UCL, June 2009 1. Review of Means, Medians, and Quantiles 2. Some Useful Asymptotic Results 3. Quantile
More informationIntroduction to Estimation Methods for Time Series models. Lecture 1
Introduction to Estimation Methods for Time Series models Lecture 1 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 1 SNS Pisa 1 / 19 Estimation
More informationIntroduction to Regression
Introduction to Regression Chad M. Schafer cschafer@stat.cmu.edu Carnegie Mellon University Introduction to Regression p. 1/100 Outline General Concepts of Regression, Bias-Variance Tradeoff Linear Regression
More informationLECTURE 2 LINEAR REGRESSION MODEL AND OLS
SEPTEMBER 29, 2014 LECTURE 2 LINEAR REGRESSION MODEL AND OLS Definitions A common question in econometrics is to study the effect of one group of variables X i, usually called the regressors, on another
More information12 - Nonparametric Density Estimation
ST 697 Fall 2017 1/49 12 - Nonparametric Density Estimation ST 697 Fall 2017 University of Alabama Density Review ST 697 Fall 2017 2/49 Continuous Random Variables ST 697 Fall 2017 3/49 1.0 0.8 F(x) 0.6
More information