Consistent Parameter Estimation for Conditional Moment Restrictions

Size: px
Start display at page:

Download "Consistent Parameter Estimation for Conditional Moment Restrictions"

Transcription

1 Consistent Parameter Estimation for Conditional Moment Restrictions Shih-Hsun Hsu Department of Economics National Taiwan University Chung-Ming Kuan Institute of Economics Academia Sinica Preliminary; Please do not cite or quote. Author for correspondence: Chung-Ming Kuan, Institute of Economics, Academia Sinica, Taipei 115, Taiwan; The research support from the National Science Council of the Republic of China (NSC H ) is gratefully acknowledged.

2 Abstract In estimating conditional moment restrictions, a well known difficulty is that the estimator based on a set of implied unconditional moments may lose its consistency when the parameters are not globally identified. In this paper, we consider a continuum of unconditional moments that are equivalent to the postulated conditional moments and project them along the exponential Fourier series. By constructing an objective function using these Fourier coefficients, a consistent GMM estimator is readily obtained. Our estimator compares favorably with that of Domínguez and Lobato (2004, Econometrica) in that our method utilizes the full continuum of unconditional moments. We establish the asymptotic properties of the proposed estimator. It is also shown that an efficient estimator can be derived from the proposed consistent estimator via a Newton-Raphson step. Our simulations demonstrate that the proposed estimator outperforms that of Domínguez and Lobato (2004) in terms of bias, standard error and mean squared error. JEL classification: C12, C22 Keywords: conditional moment restrictions, Fourier coefficients, generically comprehensive function, GMM, parameter identification

3 1 Introduction Many economic and econometric models can be characterized in terms of conditional moment restrictions. Estimating the parameters in such restrictions thus plays a major role in empirical studies. A typical estimation approach is to find a finite set of unconditional moment restrictions implied by the conditional moment restrictions and apply a suitable estimation method, such as the generalized method of moment (GMM) of Hansen (1982) or the empirical likelihood method of Qin and Lawless (1994). The instrumental-variable estimation method for regression models is a leading example. Yet there also exist estimation methods based directly on the conditional moments, e.g., Ai and Chen (2003) and Kitamura, Tripathi, and Ahn (2004). A crucial assumption in the former apporach is that the parameters in the conditional moment restrictions can be gloabally identified by the implied, unconditional restrictions. With this assumption, the consistency of an estimator can be easily established under suitable regularity conditions. Much research interest therefore focuses on estimator efficiency, e.g., Chamberlain (1987), Newey (1990, 1993), Carrasco and Florens (2000), and Donald, Imbens, and Newey (2003). Domínguez and Lobato (2004), however, pointed out that the assumption of global identifiability may fail to hold in nonlinear models. It is well known that, except in some special cases, conditional moment restrictions are equivalent to infinitely many unconditional restrictions in general (Bierens, 1982, 1990; Chamberlain, 1987; Stinchcombe and White, 1998). The identification problem may arise when one arbitrarily selects a finite number of unconditional moments for estimation. Domínguez and Lobato (2004) also demonstrated that this may happen even when the unconditional moments are based on the so-called optimal instruments. When the parameters of interest are not identified, the corresponding (GMM) estimator is inconsistent. To circumvent the potential non-identification and inconsistency problems, Domínguez and Lobato (2004) proposed a different approach to estimating conditional moment restrictions. They considered a continuum of unconditional moment restrictions that are equivalent to the conditional restrictions, where the unconditional moments depend on the instruments generated from the indicator function. There are some disadvantages of their approach. In view of Stinchcombe and White (1998), any generically comprehensively revealing function, such as the exponential and logistic functions, can also induce equivalent, unconditional moment restrictions. Compared with these functions, 1

4 the indicator function, which takes only the values one and zero, may not well present the information in the conditioning variables. Moreover, Domínguez and Lobato (2004) did not utilize the full continuum of unconditional restrictions but included only a finite number of them in estimation. This may result in further efficiency loss, as argued by Carrasco and Florens (2000). In this paper, we propose a new approach for consistent estimation of conditional moment restrictions. We also consider a continuum of unconditional moments that are equivalent to the postulated conditional moments. Instead of dealing with these moments directly, we project them along the exponential Fourier series. By constructing an objective function using these Fourier coefficients, a consistent GMM estimator is readily obtained. Because all unconditional moments are incorporated into each Fourier coefficient, our estimation method in effect utilizes all possible information. Moreover, the remote Fourier coefficients provide little information on the parameters of interest and may be excluded from the objective function. The objective function is hence relatively simple while still involving the full continuum of moment conditions. By choosing the exponential function to generate instruments, the resulting unconditional moments and the Fourier coefficients have analytic forms. These properties greatly facilitate estimation in practice. Compared with Carrasco and Florens (2000), our method is also simpler because we do not require preliminary estimation of the eigenvalues and eigen-functions of a covariance operator, nor do we need to determine a regularization parameter. In this paper, we show that under quite general conditions, the proposed estimator is consistent and asymptotically normally distributed. It is also shown that an efficient estimator is readily obtained from the proposed consistent estimator via a Newton-Raphson step, as in, e.g., Newey (1990) and Domínguez and Lobato (2004). Our simulations demonstrate that, under various settings, the proposed estimator compares favorably with that of Domínguez and Lobato (2004) in terms of bias, standard error and mean squared error. This paper is organized as follows. The proposed approach is introduced in Section 2. The asymptotic properties of the proposed estimator are established in Section 3. We report Monte Carlo simulation results in Section 4 and discuss some extensions of the proposed method. Section 6 concludes this paper. All proofs are deferred to the Appendix. 2

5 2 The Proposed Estimation Approach We are primarily interested in estimating θ o, the q 1 vector of unknown parameters, in the following conditional moment restriction: IE[h(Y, θ o ) X] = 0, w.p.1, (1) where h is a p 1 vector of functions of the variable Y (r 1) and the parameter vector θ (q 1), and X is an m 1 vector of conditioning variables. It is well known that (1) is equivalent to the unconditional moment restriction: IE[h(Y, θ o )f(x)] = 0, (2) for all measurable function f. Here, each f(x) may be interpreted as an instrument that helps to identify θ o. Clearly, estimation based on all measurable functions of X is practically infeasible. It is thus typical to select a finite number of unconditional moments in (2) and estimate θ o using, say, the GMM method. However, Domínguez and Lobato (2004) showed that if these restrictions are not chosen properly, there is no guarantee that θ o can be identified globally. This identification problem in turn renders the associated GMM estimator inconsistent. In light of Bierens (1982, 1990) and Stinchcombe and White (1998), the parameters of interest can be identified when the unconditional moments are based on a class of instruments that can span an appropriate space of functions of X. Domínguez and Lobato (2004) chose the indicator function to generate a continuum of instruments in (2), i.e., f(x) = 1(X τ ) = m 1(X j τ j ), j=1 with 1(B) the indicator function of the event B. The indicator function is not the only choice of functions that can generate the desired instruments; any generically comprehensively revealing function as defined in Stinchcombe and White (1998) will also do. Letting A denote the affine transformation such that A(X, τ ) = τ 0 + m j=1 X jτ j. Stinchcombe and White (1998) showed that, for a real analytic function G that is not a polynomial, (1) is equivalent to IE [ h(y, θ o )G ( A(X, τ ) )] = 0, for almost all τ T R m+1, (3) 3

6 where T has a nonempty interior and could be a small set in R m+1. Observe that (3) is a continuum of unconditional moment restrictions indexed by the parameter τ in T, cf. (2). In particular, the function G may be the exponential function (Bierens, 1982, 1990) or the logistic function (White, 1989). In this paper, we propose an alternative to constructing a consistent estimator of θ o that improves on the approach of Domínguez and Lobato (2004). We first consider the case where X is univariate, denoted as X. Without loss of generality, we transform X to a bounded random variable using a generalized logistic mapping: X = exp (X) c + exp (X), c > 0, so that 0 < X < 1/c. restriction (1) is equivalent to Because this mapping is one-to-one, the conditional moment IE[h(Y, θ o ) X] = 0, w.p.1; (4) see e.g., Bierens (1994, Theorem 3.2.1). For a real function G that is analytic but not a polynomial, Stinchcombe and White (1998) showed that the unonditional restriction equivalent to (4) is IE [ h(y, θ o )G ( A( X, τ ) )] = 0, for almost all τ T R 2. (5) Then, θ o can be globally identified via an L 2 -norm of (5): θ o = argmin θ Θ = argmin θ Θ IE [ h(y, θ)g ( A( X, ) )] 2 T L 2 IE [ h(y, θ o )G ( A( X, τ ) )] (6) 2 dp (τ ), where P (τ ) is a distribution function of τ in T. For the choice of G, we set G ( A( X, τ ) ) = exp ( A( X, τ ) ), as in Bierens (1982, 1990). This choice has some advantages over the indicator function. First, while the indicator function takes only the values one and zero, the exponential function is more flexible and hence may better present the information in the conditioning variables. That is, the exponential function may generate better instruments for identifying θ o. Second, the exponential function is generically comprehensively revealing, but the indicator function is not. By the genericity property, we can focus on the instruments indexed by τ in a small set T. By contrast, one must deal with the instruments generated 4

7 by the indicator function for all possible τ R. Moreover, in the class of generically comprehensively revealing function, exp ( A( X, τ ) ) and exp(τ X) only differ by a constant and hence play the same role in function approximation (Stinchcombe and White, 1998). Choosing the exponential function thus reduces the dimension of integration in (6) by one, i.e., T R. As far as implementation is concerned, the exponential function results in an analytic form for the objective function, which greatly facilitates estimation in practice. Apart from the choice of the instrument-generating function G, our approach differs from that of Domínguez and Lobato (2004) in an important respect. Instead of directly working on the unconditional moments, we project IE[h(Y, θ) exp(τ X)] along the exponential Fourier series: IE [ h(y, θ) exp(τ X) ] = k= C k (θ) exp (ikτ), where C k (θ) is a Fourier coefficient such that C k (θ) = 1 2π π π = 1 2π IE IE [ h(y, θ) exp(τ X) ] exp ( ikτ) dτ [ π h(y, θ) exp(τ X) exp( ikτ) dτ π ], k = 0, ±1, ±2,... Note that each C k (θ) incorporates the continuum of the original instruments exp(τ X) into a new instrument with the following analytic form: ϕ k ( X) = 1 2π π π exp(τ X) exp( ikτ) dτ = 1 2π ( 1) k [exp (π X) exp( π X)] ( X ik) = ( 1)k 2 sinh (π X) 2π( X ik), where sinh(w) = (exp (w) exp ( w))/2. Setting dp (τ) = dτ and T = [ π, π] in (6), we have from Parseval s Theorem that IE [ h(y, θ)g( X) ] 2 L 2 = 2π k= Ck (θ) 2 = k= It follows from (6) that θ o can be globally identified as θ o = argmin θ Θ k= IE [ h(y, θ)ϕ k ( X) ] 2. IE [ h(y, θ)ϕ k ( X) ] 2. (7) 5

8 That is, θ o is identified using a countable collection of unconditional moments. By Bessel s inequality, C k (θ) 0 as k tends to infinity, so that the moment condition IE [ h(y, θ)ϕ k ( X) ] becomes less informative for identifying θ o when k is large. This suggests that there is no need to include all the Fourier coefficients in estimating θ o. As such, we may replace IE [ h(y, θ)ϕ k ( X) ] with its sample counterpart and compute an estimator of θ o as θ(k T ) = argmin θ Θ = argmin θ Θ K T k= K T K T k= K T 1 T 1 T T h(y t, θ)ϕ k ( x t ) t=1 T t=1 h(y t, θ) ( 1)k 2 sinh (π x t ) 2π( xt ik) where K T depends on T but is less than T, and y t and x t are the realizations of Y and X, respectively. This is in fact a GMM estimator based on (2K T + 1) unconditional moments and the identity weighting matrix. The estimator θ(k T ) now can be computed quite straightforwardly using a numerical optimization algorithm. By choosing the indicator function to generate the desired instruments, Domínguez and Lobato (2004) showed that θ o can be identified as [ ] θ o = argmin IE h(y, θ)1(x ) 2 θ Θ = argmin θ Θ R L 2 IE [ h(y, θ)1(x τ) ] 2 dp (τ), where P (τ) is a distribution function of τ R, cf. (6) and (7). Setting P (τ) as P X (τ), the distribution function of X, it is easy to obtain the sample counterpart of the L 2 -norm above. The estimator proposed by Dimínguez and Lobato (2004) is then θ DL (T ) = argmin θ Θ T k=1 ( 1 T 2 2, (8) 2 T h(y t, θ)1(x t τ k )), (9) t=1 where τ k = x k, k = 1,..., T, are just the sample observations. This is also a GMM estimator based on T unconditional moments induced by the indicator function. Note that when P X (τ) is used in the L 2 -norm, θ DL (T ) utilizes only a finite number of the unconditional moments. By contrast, the proposed estimator (8) depends on a few new instruments ϕ k, yet each ϕ k utlizes the full continuum of the original instruments required for identifying θ o. 6

9 Carrasco and Florens (2000) also proposed an efficient GMM estimation method based on a continuum of unconditional moments. Adapting to the current context, their approach amounts to projecting the continuum of the induced unconditional moments along the eigen-functions of a covariance operator. This approach thus requires preliminary estimation of a covariance operator and its eigenvalues and eigen-functions. Also, to ensure the invertibility of the estimated covariance operator, a regularization parameter must be set by researchers. As a result, their method is not easy to implement and may be arbitrary in practice. 3 Asymptotic Properties In this section, we establish the asymptotic properties of the proposed estimator θ(k T ). For notation simplicity, we define µ k (θ) = IE[h(Y, θ)ϕ k ( X)] and m T,k (θ) = 1 T T h(y t, θ)ϕ k ( x t ). t=1 When integration and differentiation can be interchanged, let θ µ k (θ) = IE [ θ h(y, θ)ϕ k ( X) ], θ m T,k (θ) = 1 T T θ h(y t, θ)ϕ k ( x t ). t=1 Our maintained assumption is: θ o is the unique solution to IE[h(Y, θ) X] = 0. We impose the following conditions that are quite standard in the GMM literature and establish a lemma concerning the limiting behavior of the sum of m T,k (θ) µ k (θ). [A1] The observed data (y t, x t ), t = 1,..., T, are independent realizations of (Y, X). [A2] Θ R q is compact such that for each θ Θ, h(, θ) is measurable, and for each y R r, h(y, ) is continuous on Θ. [A3] IE[sup θ Θ h(y, θ) 2 ] <. Lemma 3.1 Suppose that [A1] [A3] hold. Then for any ε > 0, K T IP m T,k (θ) µ k (θ) 2 ε (2K T + 1) T ε k= K T uniformly in θ Θ, where is some real number. 7

10 This result shows that, as long as K T = o(t ), K T k= K T m T,k (θ) µ k (θ) 2 IP 0, and hence K T implies, uniformly for all θ in Θ, K T k= K T m T,k (θ) 2 IP k= µ k (θ) 2. (10) Given the maintained assumption, θ o = argmin θ Θ k= µ k(θ) 2, and θ(k T ) = argmin θ Θ K T k= K T m T,k (θ) 2. (11) The consistency result now follows from (10) and Theorem 2.1 of Newey and McFadden (1994). Theorem 3.2 (Consistency) Given [A1] [A3], suppose that K T as T such that K T = o(t ). Then θ(k T ) IP θ o. In what follows, for a complex number (function) f, let f denote its complex conjugate and Re(f) and Im(f) denote its real and imaginary parts, respectively. For the vector of complex functions f, its complex conjugate, real part and imaginary part are defined elementwise. For two p 1 vectors of functions f and g in L 2 [ π, π], their inner product is defined as f, g = π π f(τ) g(τ) dτ = π π i=1 p f i (τ)ḡ i (τ) dτ. Given (11), θ(k T ) satisfies the first order condition: K T 0 = θ m T,k ( θ(kt ) ) mt,k ( θ(kt ) ) + θ m T,k ( θ(kt ) ) mt,k ( θ(kt ) ) k= K T = K T k= K T 2 Re ( θ m T,k ( θ(kt ) ) mt,k ( θ(kt ) )). A mean value expansion of m T,k ( θ(kt ) ) about θ o gives m T,k ( θ(kt ) ) = m T,k (θ o ) + θ m T,k ( θ )( θ(kt ) θ o ), 8

11 where θ is on the line segment joining θ(k T ) and θ o which may be different for each row of θ m T,k ( θ ). It follows that K T k= K T 2 Re ( θ m T,k ( θ(kt ) ) [ mt,k (θ o ) + θ m T,k ( θ )( θ(kt ) θ o )] ) = 0. (12) Passing to the limit, we have, by the multiplication theorem (e.g., Stuart, 1961), K T k= K T Re K T k= K T Re ( θ m T,k (θ o ) θ m T,k (θ o ) ) IP ( θ m T,k (θ o ) ) T m T,k (θ o ) = k= k= The normalized estimator can then be expressed as T ( θ(kt ) θ o ) = [ k= θ µ k (θ o ) θ µ k (θ o ) ] 1 [ k= θ µ k (θ o )µ k (θ o ), θ µ k (θ o ) T m T,k (θ o ) + o IP (1). θ µ k (θ o ) T m T,k (θ o ) ] + o IP (1). Asymptotic normality then follows if the term in the first brackets is a finite valued, positive definite matrix, and the term in the second brackets obeys a central limit theorem. We impose the following conditions. [A4] θ o is in the interior of Θ. [A5] For each y, h(y, ) is continuously differentiable in a neighborhood N of θ o such that IE [ sup θ N θ h(y, θ) 2] <, where is a matrix norm. [A6] The q q matrix M q, with the (i, j)-th element [ IE θi h(y, θ o ) exp( X) ] [, IE θj h(y, θ o ) exp( X) ], is symmetric and positive definite. [A7] As T goes to infinity, 1 T T t=1 h(y t, θ o ) exp ( x t ) D Z, (13) 9

12 where Z is a Gaussian random element in L 2 [ π, π] that has a zero mean and the covariance operator K such that [ Kf(τ ) = IE h(y, θ o ) exp( X), ( f h(y, θ o ) exp( X) )], for any f (p 1) in L 2 [ π, π]. The covariance operator K in [A8] can be expressed as ( p ) π Kf(τ ) = κ ji (τ, s)f i (s) ds, with the kernel i=1 π j=1,...,p κ ji (τ, s) = IE [ h j (Y, θ o ) exp(τ X)h i (Y, θ o ) exp(s X) ], j, i = 1,..., p. It can also be seen that Z, f has a normal distribution with mean zero and variance Kf, f. Note that the functional convergence result in [A8] is the same as the Assumption 11 in Carrasco and Florens (2000). One may, of course, impose more primitive conditions on h and the data that ensure such convergence result. For example, [A8] would hold when h(y t, θ o ) exp( x t ) are i.i.d. with bounded second moment. More discussions and results on functional convergence can be found in Chen and White (1998) and Carrasco and Florens (2000). Let Ω q denote a q q matrix with the (i, j)-th element [ IE θi h(y, θ o ) exp( X) ] [, K IE θj h(y, θ o ) exp( X) ]. We are ready to present the limiting distribution of the normalized estimator. Theorem 3.3 (Asymptotic Normality) Given Assumptions [A1] [A7], suppose that K T as T such that K T = o(t 1/2 ). Then T ( θ(kt ) θ o ) where V = M 1 q Ω q M 1 q. D N (0, V), 10

13 4 Simulations In this section we examine the finite-sample performance of the proposed estimator via Monte Carlo simulations. We compare the performance of the nonlinear least squares (NLS) estimator: θ NLS = argmin θ Θ 1 T T h(y t, θ) 2. t=1 the estimator of Domínguez and Lobato (2004): θ DL in (9), and the proposed estimator: θ(k T ) with K T = 5 in (8). We consider three performance criteria: bias, standard error (SE), and mean squared error (MSE). In all experiments, we compute the parameter estimate using the GAUSS optimization procedure, OPTMUM, with the BFGS algorithm; for each optimization, three initial values are randomly drawn from the standard normal distribution. The number of replications is It should be noted that neither θ DL nor θ(k T ) is an efficient estimator here. 4.1 The Experiment in Domínguez and Lobato (2004) Following Domínguez and Lobato (2004), we consider a simple nonlinear model: Y = θ 2 ox + θ o X 2 + ɛ, ɛ N (0, 1), where θ o = 5/4. The associated conditional moment restriction is IE[ɛ X] = 0. In this case, in addition to θ NLS, θ DL and θ(k T ), we also consider the GMM estimator with the feasible optimal instrument: 2θX + X 2. The resulting GMM estimator is θ OPIV = argmin θ Θ ( 1 T 2 T (y t θ 2 x t θx 2 t )(2θx t + x 2 t )). t=1 We consider two cases in which X follows N (0, 1) distribution and N (1, 1) distribution, respectively. Domínguez and Lobato (2004) showed that in the latter case, θ o can not be identified using the feasible optimal instrument because θ = 5/4 and θ = 3 also satisfy the implied, unconditional moment restriction. When X is N (0, 1), θ o = 5/4 is the only real solution to the unconditional moment restriction implied by the feasible optimal instrument. 1 The simulation results are summarized in Tables 1. It can be seen that in both cases, the NLS estimator is the best estimator in terms of all 3 criteria, and 1 The other two solutions are complex: ± i. 11

14 θ OPIV has the worst performance with severe biases and larger standard errors. As to the comparison between θ DL and θ(k T ), when X N (1, 1), the performance of θ DL is dominated by that of θ(k T ) under all 3 criteria. When X N (0, 1), θ(k T ) outperforms θ DL under all 3 criteria only for the samples T = 50 and Model with An Endogenous Explanatory Variable We extend the previous experiment to the cases where the explanatory variable is endogenous. The model specification is: { Y = θ 2 o Z + θ o Z 2 + ɛ, Z = X + ν [ ɛ ν ] N ( 0, [ 1 ρ ρ 1 where θ o = 5/4, and X N (0, 1) and is independent of ɛ and ν. In this specification, X can be treated as an instrumental variable of Z and satisfies the conditional moment restriction IE[ɛ X] = 0. ] ) In Table 2, we summarize the results for ρ = 0.01, 0.1, 0.3, 0.5, 0.7, 0.9 and the sample sizes 50, 100 and 200., It can be seen that the NLS estimator has more severe bias (and smaller SE) when ρ becomes larger. It is also quite remarkable that the proposed estimator θ(k T ) outperforms θ DL under all 3 criteria for any ρ and any sample size. The bias of θ(k T ) is also less than that of ˆθ NLS in almost all cases, except for the case that ρ = 0.01 and T = 100. Moreover, the proposed estimator may outperform the NLS estimator in terms of MSE when ρ is not too small. For example, when ρ 0.3, ˆθ(K T ) has a smaller MSE than the NLS estimator for T = 100 and Noisy disturbance In this experiment, we examine the effect of the disturbance variance on the performance of various estimators. The model is again Y = θox 2 + θ o X 2 + ɛ, ɛ N (0, σ 2 ), where θ o = 5/4, X is a uniform random variable on ( 1, 1), and σ 2 =0.1, 1, 4, 9, and 16. The results are summarized in Table 3. We first observe that the bias, SE and MSE of these estimators all increase as σ 2 increases. In all cases, the proposed estimator is the one with the smallest bias, while the estimtor of Domínguez and Lobato (2004) has the largest bias. In terms of the MSE, the proposed estimator, in general, dominates ˆθ DL and may also outperform the NLS estimator when σ 2 is not too large. For example, for the 12

15 sample T = 100, the proposed estimator has smaller MSE than does the NLS estimator (ˆθ DL ) when σ 2 1 (σ 2 9). For T = 200, the proposed estimator has smaller MSE than does the NLS estimator when σ 2 4 but dominates ˆθ DL in all cases. 4.4 The Proposed Estimator with Various K T In this experiment, we examine how the choice of K T may affect the performance of ˆθ(K T ). The model specification is the same as that in Section 4.2 where the explanatory variable is endogenous. We consider the cases that ρ equals 0.01, 0.1, 0.5 and 0.9, and the sample T = 50, 100, 200 and 400. We simulate ˆθ(K T ) with K T = 1, 2,..., 10, 15, 20, 30, The results are in Table to Table For the three performance criteria (bias, SE and MSE), we also calculate their percentage changes when K T increase. In Table 4 1 1, for instance, when K T increases from 1 to 2, the bias increases %, SE decreases % and MSE decreases %. The performance of ˆθ NLS and ˆθ DL are also given for comparison. It is easily seen that, when K T increases, SE decreases in all cases while the bias also decreases in most cases (except in Tables 4-1-1, and 4-3-2). The percentage changes are typically small. In most case, the percentage change of the bias is less 0.1% when K T is greater than 5. These results suggest that for the propposed estimator, the first few Fourier coefficients indeed contain the most information for identifying θ o. 5 Extensions In this section we discuss two extensions of the previous results. We first discuss how to obtain an efficient estimator from the proposed consistent estimator via a single Newton- Raphson iterative step. We also extend the model with a univariate conditioning variable to the model with multiple conditioning variables. 5.1 Efficient Estimator We follow Newey (1990, 1993) and Dominguez and Lobato (2004) to construct a twostep efficient estimator. Let Q T (θ, K T ) be the objective function for the efficient GMM estimator and θ Q T (θ, K T ) and θθ Q T (θ, K T ) be its gradient vector and the Hessian matrix, respectively. Since θ(k T ) is consistent for θ o, it will be in a neighborhood of θ o for sufficiently large T. Therefore, a Newton-Raphson step from θ(k T ) will lead to a 13

16 consistent and efficient estimator: θ e (K T ) = θ(k [ 1 T ) θθ Q T (ˆθ(K T ), K T )] θ Q T (ˆθ(K T ), K T ). Note that this approach is valid provided that θ o Q T (θ, K T ). can be identified by the limit of 5.2 Multiple Conditioning Variables Consider the conditional moment restriction IE[h(Y, θ o ) X] = 0 with multiple conditioning variables X. This is equivalent to IE[h(Y, θ o ) exp(τ X)] = 0, almost all τ [ π, π] m R m. Let S := {k = [k 1 k 2... k m ] R m } with k i = 0, ±1, ±2,, ±. The Fourier representation of the unconditional moment is IE[h(Y, θ) exp(τ X)] = C k (θ) exp (ik τ ), where for k S, with C k (θ) = 1 (2π) m = [ π,π] m IE k S [h(y, θ) exp(τ X) ] exp ( ik τ ) dτ 1 [ (2π) m/2 IE h(y, θ)ϕ k1 ( X 1 )ϕ k2 ( X 2 ) ϕ km ( X m ) ϕ kj ( X j ) = 1 2π π π ], exp(τ j Xj ) exp ( ik j τ j ) dτ j = ( 1)kj 2 sinh (π X j ), 2π( Xj ik j ) as in the model with a univariate conditioning variable. Let ϕ k ( X) = ϕ k1 ( X 1 )ϕ k2 ( X 2 ) ϕ km ( X m ). Again by Parseval s Theorem, we have IE[h(Y, θ) exp( X)] 2 = (2π) m Ck (θ) 2 = [ 2 IE h(y, θ)ϕ k ( X)]. L 2 k S k S It follows that θ o can be globally identified as [ 2 θ o = argmin IE h(y, θ)ϕ k ( X)], (14) θ Θ k S Then, a consistent estimator of θ o can be computed as 2 1 T ˆθ(K T ) = argmin h(y θ Θ T t, θ)ϕ k ( x t ). (15) k S(K T ) t=1 14

17 Appendix Proof of Lemma 3.1: Let be a generic constant which varies in different cases and η tk = h(y t, θ)ϕ k ( x t ) IE [ h(y, θ)ϕ k ( X) ], for t = 1,..., T and k = 0,..., K T. First note that as X is bounded, so is sinh(π X). Thus, ϕ 0 ( X) is bounded, and with [A3], IE [ η t0 2] IE [ h(y, θ) 2 ϕ 0 ( X) 2]. For k 1, it is easy to see ϕ k ( X) / k. It follows that IE [ η tk 2] IE [ h(y, θ) 2 ϕ k ( X) 2] /k 2. By [A1], the data are i.i.d., and K T 1 T 2 η T tk = 1 k= K T IE t=1 T T 2 t=1 IE [ ηt0 2] + 2 T 2 ( K T 1 ) T k 2 T. k=1 K T k=1 t=1 T IE [ ηtk 2] By the implication rule and the generalized Chebyshev inequality, we have K T IP 1 T 2 K T η T tk ε IP 1 T 2 ε η T tk 2K k= K T t=1 k= K T t=1 T + 1 K T 1 T 2 η T tk 2K T + 1 ε k= K T IE (2K T + 1). T ε This result holds uniformly in θ because the bound does not depend on θ. t=1 Proof of Theorem 3.2: K T k= K T m T,k (θ) 2 IP We have seen from the text that Lemma 3.1 implies k= m k (θ) 2, uniformly for all θ in Θ. It follows from Theorem 2.1 of Newey and West (1994) that the solution to the left-hand side, ˆθ(K T ), must converge in probability to the unique solution to the right-hand side, θ o. 15

18 References Ai, C. and X. Chen (2003). Efficient estimation of models with conditional moment restrictions containing unknown functions, Econometrica, 71, Bierens, H. J. (1982). Consistent model specification test Journal of Econometrics, 26, Bierens, H. J. (1990). A consistent conditional moment test of functional form, Econometrica, 58, Bierens, H. J. (1994). Topics in Advanced Econometrics, New York: Cambridge University Press. Chen, X. and H. White (1998). Central limit and functional central limit theorems for Hilbert-valued dependent heterogeneous arrays with applications, Econometric Theory, 14, Carrasco, M. and J.- P. Florens (2000). Generalization of GMM to a continuum of moment conditions, Econometric Theory, 16, Chamberlain, G. (1987). Asymptotic efficiency in estimation with conditional moment restrictions, Journal of Econometrics, 34, Domínguez, M. A. and I. N. Lobato (2004). Consistent estimation of models defined by conditional moment restrictions, Econometrica, 72, Donald, S. G., G. W. Imbens, and W. K. Newey (2003). Empirical likelihood estimation and consistent tests with conditional moment restrictions, Journal of Econometrics, 117, Hansen, L. P. (1982). Large sample properties of generalized method of moments estimators, Econometrica, 50, Kitamura, Y., G. Tripathi, and H. Ahn (2004). Empirical likelihood-based inference in conditional moment restriction models, Econometrica, 72, Newey, W. K. (1990). Efficient instrumental variables estimation of nonlinear models, Econometrica, 58, Newey, W. K. (1993). Efficient estimation of models with conditional moment restrictions, in G. Maddala, C. Rao, and H. Vinod (eds.), Handbook of Statistics, 11, 16

19 pp Elsevier Science. Qin, J. and J. Lawless (1994). Empirical likelihood and generalized estimating equations, Annals of Statistics, 22, Stinchcombe, M. B. and H. White (1998). Consistent specification testing with nuisance parameters present only under the alternative, Econometric Theory, 14, Stuart, R. D. (1961). An Introduction to Fourier Analysis, New York: Halsted Press. White, H. (1989). An additional hidden unit test for neglected nonlinearity, Proceedings of the International Joint Conference on Neural Networks, Vol. 2, pp New York: IEEE Press. 17

20 Table 1: Example in Domínguez and Lobato (2004). Sample X N (0, 1) X N (1, 1) T Estimator Bias SE MSE Bias SE MSE 50 ˆθNLS ˆθ DL ˆθ(K T ) ˆθ OPIV ˆθNLS ˆθ DL ˆθ(K T ) ˆθ OPIV ˆθNLS ˆθ DL ˆθ(K T ) ˆθ OPIV

21 Table 2: Models with an endogenous explanatory variable. T = 50 T = 100 T = 200 ρ Est. Bias SE MSE Bias SE MSE Bias SE MSE 0.01 ˆθNLS ˆθ DL ˆθ(K T ) ˆθNLS ˆθ DL ˆθ(K T ) ˆθNLS ˆθ DL ˆθ(K T ) ˆθNLS ˆθ DL ˆθ(K T ) ˆθNLS ˆθ DL ˆθ(K T ) ˆθNLS ˆθ DL ˆθ(K T )

22 Table 3: Models with different disturbance variances. T = 50 T = 100 T = 200 σ 2 Est. Bias SE MSE Bias SE MSE Bias SE MSE σ 2 = 0.1 ˆθNLS ˆθ DL ˆθ(K T ) σ 2 = 1 ˆθNLS ˆθ DL ˆθ(K T ) σ 2 = 4 ˆθNLS ˆθ DL ˆθ(K T ) σ 2 = 9 ˆθNLS ˆθ DL ˆθ(K T ) σ 2 = 16 ˆθNLS ˆθ DL ˆθ(K T )

23 Table : The performance of ˆθ(K T ) with various K T (ρ = 0.01, T = 50). K T Bias Bias(+%) SE SE(+%) MSE MSE(+%) ˆθ NLSE ˆθ DL

24 Table : The performance of ˆθ(K T ) with various K T (ρ = 0.01, T = 100). K T Bias Bias(+%) SE SE(+%) MSE MSE(+%) ˆθ NLSE ˆθ DL

25 Table : The performance of ˆθ(K T ) with various K T (ρ = 0.01, T = 200). K T Bias Bias(+%) SE SE(+%) MSE MSE(+%) ˆθ NLSE ˆθ DL

26 Table : The performance of ˆθ(K T ) with various K T (ρ = 0.01, T = 400). K T Bias Bias(+%) SE SE(+%) MSE MSE(+%) ˆθ NLSE ˆθ DL

27 Table : The performance of ˆθ(K T ) with various K T (ρ = 0.1, T = 50). K T Bias Bias(+%) SE SE(+%) MSE MSE(+%) ˆθ NLSE ˆθ DL

28 Table : The performance of ˆθ(K T ) with various K T (ρ = 0.1, T = 100). K T Bias Bias(+%) SE SE(+%) MSE MSE(+%) ˆθ NLSE ˆθ DL

29 Table : The performance of ˆθ(K T ) with various K T (ρ = 0.1, T = 200). K T Bias Bias(+%) SE SE(+%) MSE MSE(+%) ˆθ NLSE ˆθ DL

30 Table : The performance of ˆθ(K T ) with various K T (ρ = 0.1, T = 400). K T Bias Bias(+%) SE SE(+%) MSE MSE(+%) ˆθ NLSE ˆθ DL

31 Table : The performance of ˆθ(K T ) with various K T (ρ = 0.5, T = 50). K T Bias Bias(+%) SE SE(+%) MSE MSE(+%) ˆθ NLSE ˆθ DL

32 Table : The performance of ˆθ(K T ) with various K T (ρ = 0.5, T = 100). K T Bias Bias(+%) SE SE(+%) MSE MSE(+%) ˆθ NLSE ˆθ DL

33 Table : The performance of ˆθ(K T ) with various K T (ρ = 0.5, T = 200). K T Bias Bias(+%) SE SE(+%) MSE MSE(+%) ˆθ NLSE ˆθ DL

34 Table : The performance of ˆθ(K T ) with various K T (ρ = 0.5, T = 400). K T Bias Bias(+%) SE SE(+%) MSE MSE(+%) ˆθ NLSE ˆθ DL

35 Table : The performance of ˆθ(K T ) with various K T (ρ = 0.9, T = 50). K T Bias Bias(+%) SE SE(+%) MSE MSE(+%) ˆθ NLSE ˆθ DL

36 Table : The performance of ˆθ(K T ) with various K T (ρ = 0.9, T = 100). K T Bias Bias(+%) SE SE(+%) MSE MSE(+%) ˆθ NLSE ˆθ DL

37 Table : The performance of ˆθ(K T ) with various K T (ρ = 0.9, T = 200). K T Bias Bias(+%) SE SE(+%) MSE MSE(+%) ˆθ NLSE ˆθ DL

38 Table : The performance of ˆθ(K T ) with various K T (ρ = 0.9, T = 400). K T Bias Bias(+%) SE SE(+%) MSE MSE(+%) ˆθ NLSE ˆθ DL

Estimation of Conditional Moment Restrictions. without Assuming Parameter Identifiability in the Implied Unconditional Moments

Estimation of Conditional Moment Restrictions. without Assuming Parameter Identifiability in the Implied Unconditional Moments Estimation of Conditional Moment Restrictions without Assuming Parameter Identifiability in the Implied Unconditional Moments Shih-Hsun Hsu Department of Economics National Chengchi University Chung-Ming

More information

A Local Generalized Method of Moments Estimator

A Local Generalized Method of Moments Estimator A Local Generalized Method of Moments Estimator Arthur Lewbel Boston College June 2006 Abstract A local Generalized Method of Moments Estimator is proposed for nonparametrically estimating unknown functions

More information

LECTURE ON HAC COVARIANCE MATRIX ESTIMATION AND THE KVB APPROACH

LECTURE ON HAC COVARIANCE MATRIX ESTIMATION AND THE KVB APPROACH LECURE ON HAC COVARIANCE MARIX ESIMAION AND HE KVB APPROACH CHUNG-MING KUAN Institute of Economics Academia Sinica October 20, 2006 ckuan@econ.sinica.edu.tw www.sinica.edu.tw/ ckuan Outline C.-M. Kuan,

More information

The properties of L p -GMM estimators

The properties of L p -GMM estimators The properties of L p -GMM estimators Robert de Jong and Chirok Han Michigan State University February 2000 Abstract This paper considers Generalized Method of Moment-type estimators for which a criterion

More information

Panel Threshold Regression Models with Endogenous Threshold Variables

Panel Threshold Regression Models with Endogenous Threshold Variables Panel Threshold Regression Models with Endogenous Threshold Variables Chien-Ho Wang National Taipei University Eric S. Lin National Tsing Hua University This Version: June 29, 2010 Abstract This paper

More information

Spring 2017 Econ 574 Roger Koenker. Lecture 14 GEE-GMM

Spring 2017 Econ 574 Roger Koenker. Lecture 14 GEE-GMM University of Illinois Department of Economics Spring 2017 Econ 574 Roger Koenker Lecture 14 GEE-GMM Throughout the course we have emphasized methods of estimation and inference based on the principle

More information

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Article Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Fei Jin 1,2 and Lung-fei Lee 3, * 1 School of Economics, Shanghai University of Finance and Economics,

More information

A Regularization Approach to the Minimum Distance Estimation : Application to Structural Macroeconomic Estimation Using IRFs

A Regularization Approach to the Minimum Distance Estimation : Application to Structural Macroeconomic Estimation Using IRFs A Regularization Approach to the Minimum Distance Estimation : Application to Structural Macroeconomic Estimation Using IRFs Senay SOKULLU September 22, 2011 Abstract This paper considers the problem of

More information

Efficient Estimation of Non-Linear Dynamic Panel Data Models with Application to Smooth Transition Models

Efficient Estimation of Non-Linear Dynamic Panel Data Models with Application to Smooth Transition Models Efficient Estimation of Non-Linear Dynamic Panel Data Models with Application to Smooth Transition Models Tue Gørgens Christopher L. Skeels Allan H. Würtz February 16, 2008 Preliminary and incomplete.

More information

Generalized Method of Moment

Generalized Method of Moment Generalized Method of Moment CHUNG-MING KUAN Department of Finance & CRETA National Taiwan University June 16, 2010 C.-M. Kuan (Finance & CRETA, NTU Generalized Method of Moment June 16, 2010 1 / 32 Lecture

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

COMPARISON OF GMM WITH SECOND-ORDER LEAST SQUARES ESTIMATION IN NONLINEAR MODELS. Abstract

COMPARISON OF GMM WITH SECOND-ORDER LEAST SQUARES ESTIMATION IN NONLINEAR MODELS. Abstract Far East J. Theo. Stat. 0() (006), 179-196 COMPARISON OF GMM WITH SECOND-ORDER LEAST SQUARES ESTIMATION IN NONLINEAR MODELS Department of Statistics University of Manitoba Winnipeg, Manitoba, Canada R3T

More information

An Encompassing Test for Non-Nested Quantile Regression Models

An Encompassing Test for Non-Nested Quantile Regression Models An Encompassing Test for Non-Nested Quantile Regression Models Chung-Ming Kuan Department of Finance National Taiwan University Hsin-Yi Lin Department of Economics National Chengchi University Abstract

More information

Artificial Neural Networks

Artificial Neural Networks Artificial Neural Networks Chung-Ming Kuan Institute of Economics Academia Sinica Abstract Artificial neural networks (ANNs) constitute a class of flexible nonlinear models designed to mimic biological

More information

Closest Moment Estimation under General Conditions

Closest Moment Estimation under General Conditions Closest Moment Estimation under General Conditions Chirok Han Victoria University of Wellington New Zealand Robert de Jong Ohio State University U.S.A October, 2003 Abstract This paper considers Closest

More information

Econometrics I, Estimation

Econometrics I, Estimation Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the

More information

Is there an optimal weighting for linear inverse problems?

Is there an optimal weighting for linear inverse problems? Is there an optimal weighting for linear inverse problems? Jean-Pierre FLORENS Toulouse School of Economics Senay SOKULLU University of Bristol October 9, 205 Abstract This paper considers linear equations

More information

ON ILL-POSEDNESS OF NONPARAMETRIC INSTRUMENTAL VARIABLE REGRESSION WITH CONVEXITY CONSTRAINTS

ON ILL-POSEDNESS OF NONPARAMETRIC INSTRUMENTAL VARIABLE REGRESSION WITH CONVEXITY CONSTRAINTS ON ILL-POSEDNESS OF NONPARAMETRIC INSTRUMENTAL VARIABLE REGRESSION WITH CONVEXITY CONSTRAINTS Olivier Scaillet a * This draft: July 2016. Abstract This note shows that adding monotonicity or convexity

More information

Generated Covariates in Nonparametric Estimation: A Short Review.

Generated Covariates in Nonparametric Estimation: A Short Review. Generated Covariates in Nonparametric Estimation: A Short Review. Enno Mammen, Christoph Rothe, and Melanie Schienle Abstract In many applications, covariates are not observed but have to be estimated

More information

Statistical Properties of Numerical Derivatives

Statistical Properties of Numerical Derivatives Statistical Properties of Numerical Derivatives Han Hong, Aprajit Mahajan, and Denis Nekipelov Stanford University and UC Berkeley November 2010 1 / 63 Motivation Introduction Many models have objective

More information

ECON 3150/4150, Spring term Lecture 6

ECON 3150/4150, Spring term Lecture 6 ECON 3150/4150, Spring term 2013. Lecture 6 Review of theoretical statistics for econometric modelling (II) Ragnar Nymoen University of Oslo 31 January 2013 1 / 25 References to Lecture 3 and 6 Lecture

More information

Flexible Estimation of Treatment Effect Parameters

Flexible Estimation of Treatment Effect Parameters Flexible Estimation of Treatment Effect Parameters Thomas MaCurdy a and Xiaohong Chen b and Han Hong c Introduction Many empirical studies of program evaluations are complicated by the presence of both

More information

Inverse problems in statistics

Inverse problems in statistics Inverse problems in statistics Laurent Cavalier (Université Aix-Marseille 1, France) Yale, May 2 2011 p. 1/35 Introduction There exist many fields where inverse problems appear Astronomy (Hubble satellite).

More information

Notes on Asymptotic Theory: Convergence in Probability and Distribution Introduction to Econometric Theory Econ. 770

Notes on Asymptotic Theory: Convergence in Probability and Distribution Introduction to Econometric Theory Econ. 770 Notes on Asymptotic Theory: Convergence in Probability and Distribution Introduction to Econometric Theory Econ. 770 Jonathan B. Hill Dept. of Economics University of North Carolina - Chapel Hill November

More information

Introduction to the Mathematical and Statistical Foundations of Econometrics Herman J. Bierens Pennsylvania State University

Introduction to the Mathematical and Statistical Foundations of Econometrics Herman J. Bierens Pennsylvania State University Introduction to the Mathematical and Statistical Foundations of Econometrics 1 Herman J. Bierens Pennsylvania State University November 13, 2003 Revised: March 15, 2004 2 Contents Preface Chapter 1: Probability

More information

High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data

High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data High Dimensional Empirical Likelihood for Generalized Estimating Equations with Dependent Data Song Xi CHEN Guanghua School of Management and Center for Statistical Science, Peking University Department

More information

The Bayesian Approach to Multi-equation Econometric Model Estimation

The Bayesian Approach to Multi-equation Econometric Model Estimation Journal of Statistical and Econometric Methods, vol.3, no.1, 2014, 85-96 ISSN: 2241-0384 (print), 2241-0376 (online) Scienpress Ltd, 2014 The Bayesian Approach to Multi-equation Econometric Model Estimation

More information

1 Estimation of Persistent Dynamic Panel Data. Motivation

1 Estimation of Persistent Dynamic Panel Data. Motivation 1 Estimation of Persistent Dynamic Panel Data. Motivation Consider the following Dynamic Panel Data (DPD) model y it = y it 1 ρ + x it β + µ i + v it (1.1) with i = {1, 2,..., N} denoting the individual

More information

Supplemental Material for KERNEL-BASED INFERENCE IN TIME-VARYING COEFFICIENT COINTEGRATING REGRESSION. September 2017

Supplemental Material for KERNEL-BASED INFERENCE IN TIME-VARYING COEFFICIENT COINTEGRATING REGRESSION. September 2017 Supplemental Material for KERNEL-BASED INFERENCE IN TIME-VARYING COEFFICIENT COINTEGRATING REGRESSION By Degui Li, Peter C. B. Phillips, and Jiti Gao September 017 COWLES FOUNDATION DISCUSSION PAPER NO.

More information

Regularized Empirical Likelihood as a Solution to the No Moment Problem: The Linear Case with Many Instruments

Regularized Empirical Likelihood as a Solution to the No Moment Problem: The Linear Case with Many Instruments Regularized Empirical Likelihood as a Solution to the No Moment Problem: The Linear Case with Many Instruments Pierre Chaussé November 29, 2017 Abstract In this paper, we explore the finite sample properties

More information

Vasil Khalidov & Miles Hansard. C.M. Bishop s PRML: Chapter 5; Neural Networks

Vasil Khalidov & Miles Hansard. C.M. Bishop s PRML: Chapter 5; Neural Networks C.M. Bishop s PRML: Chapter 5; Neural Networks Introduction The aim is, as before, to find useful decompositions of the target variable; t(x) = y(x, w) + ɛ(x) (3.7) t(x n ) and x n are the observations,

More information

Department of Econometrics and Business Statistics

Department of Econometrics and Business Statistics ISSN 440-77X Australia Department of Econometrics and Business Statistics http://wwwbusecomonasheduau/depts/ebs/pubs/wpapers/ The Asymptotic Distribution of the LIML Estimator in a artially Identified

More information

Closest Moment Estimation under General Conditions

Closest Moment Estimation under General Conditions Closest Moment Estimation under General Conditions Chirok Han and Robert de Jong January 28, 2002 Abstract This paper considers Closest Moment (CM) estimation with a general distance function, and avoids

More information

ECO Class 6 Nonparametric Econometrics

ECO Class 6 Nonparametric Econometrics ECO 523 - Class 6 Nonparametric Econometrics Carolina Caetano Contents 1 Nonparametric instrumental variable regression 1 2 Nonparametric Estimation of Average Treatment Effects 3 2.1 Asymptotic results................................

More information

The Uniform Weak Law of Large Numbers and the Consistency of M-Estimators of Cross-Section and Time Series Models

The Uniform Weak Law of Large Numbers and the Consistency of M-Estimators of Cross-Section and Time Series Models The Uniform Weak Law of Large Numbers and the Consistency of M-Estimators of Cross-Section and Time Series Models Herman J. Bierens Pennsylvania State University September 16, 2005 1. The uniform weak

More information

Identification and Estimation of Partially Linear Censored Regression Models with Unknown Heteroscedasticity

Identification and Estimation of Partially Linear Censored Regression Models with Unknown Heteroscedasticity Identification and Estimation of Partially Linear Censored Regression Models with Unknown Heteroscedasticity Zhengyu Zhang School of Economics Shanghai University of Finance and Economics zy.zhang@mail.shufe.edu.cn

More information

Economics 536 Lecture 7. Introduction to Specification Testing in Dynamic Econometric Models

Economics 536 Lecture 7. Introduction to Specification Testing in Dynamic Econometric Models University of Illinois Fall 2016 Department of Economics Roger Koenker Economics 536 Lecture 7 Introduction to Specification Testing in Dynamic Econometric Models In this lecture I want to briefly describe

More information

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature

More information

CALCULATION METHOD FOR NONLINEAR DYNAMIC LEAST-ABSOLUTE DEVIATIONS ESTIMATOR

CALCULATION METHOD FOR NONLINEAR DYNAMIC LEAST-ABSOLUTE DEVIATIONS ESTIMATOR J. Japan Statist. Soc. Vol. 3 No. 200 39 5 CALCULAION MEHOD FOR NONLINEAR DYNAMIC LEAS-ABSOLUE DEVIAIONS ESIMAOR Kohtaro Hitomi * and Masato Kagihara ** In a nonlinear dynamic model, the consistency and

More information

BTRY 4090: Spring 2009 Theory of Statistics

BTRY 4090: Spring 2009 Theory of Statistics BTRY 4090: Spring 2009 Theory of Statistics Guozhang Wang September 25, 2010 1 Review of Probability We begin with a real example of using probability to solve computationally intensive (or infeasible)

More information

Iterative Bias Correction Procedures Revisited: A Small Scale Monte Carlo Study

Iterative Bias Correction Procedures Revisited: A Small Scale Monte Carlo Study Discussion Paper: 2015/02 Iterative Bias Correction Procedures Revisited: A Small Scale Monte Carlo Study Artūras Juodis www.ase.uva.nl/uva-econometrics Amsterdam School of Economics Roetersstraat 11 1018

More information

A PARAMETRIC MODEL FOR DISCRETE-VALUED TIME SERIES. 1. Introduction

A PARAMETRIC MODEL FOR DISCRETE-VALUED TIME SERIES. 1. Introduction tm Tatra Mt. Math. Publ. 00 (XXXX), 1 10 A PARAMETRIC MODEL FOR DISCRETE-VALUED TIME SERIES Martin Janžura and Lucie Fialová ABSTRACT. A parametric model for statistical analysis of Markov chains type

More information

Research Article Optimal Portfolio Estimation for Dependent Financial Returns with Generalized Empirical Likelihood

Research Article Optimal Portfolio Estimation for Dependent Financial Returns with Generalized Empirical Likelihood Advances in Decision Sciences Volume 2012, Article ID 973173, 8 pages doi:10.1155/2012/973173 Research Article Optimal Portfolio Estimation for Dependent Financial Returns with Generalized Empirical Likelihood

More information

ECE531 Lecture 10b: Maximum Likelihood Estimation

ECE531 Lecture 10b: Maximum Likelihood Estimation ECE531 Lecture 10b: Maximum Likelihood Estimation D. Richard Brown III Worcester Polytechnic Institute 05-Apr-2011 Worcester Polytechnic Institute D. Richard Brown III 05-Apr-2011 1 / 23 Introduction So

More information

Regularization in Neural Networks

Regularization in Neural Networks Regularization in Neural Networks Sargur Srihari 1 Topics in Neural Network Regularization What is regularization? Methods 1. Determining optimal number of hidden units 2. Use of regularizer in error function

More information

Hilbert Spaces. Hilbert space is a vector space with some extra structure. We start with formal (axiomatic) definition of a vector space.

Hilbert Spaces. Hilbert space is a vector space with some extra structure. We start with formal (axiomatic) definition of a vector space. Hilbert Spaces Hilbert space is a vector space with some extra structure. We start with formal (axiomatic) definition of a vector space. Vector Space. Vector space, ν, over the field of complex numbers,

More information

Self-Concordant Barrier Functions for Convex Optimization

Self-Concordant Barrier Functions for Convex Optimization Appendix F Self-Concordant Barrier Functions for Convex Optimization F.1 Introduction In this Appendix we present a framework for developing polynomial-time algorithms for the solution of convex optimization

More information

Testing Overidentifying Restrictions with Many Instruments and Heteroskedasticity

Testing Overidentifying Restrictions with Many Instruments and Heteroskedasticity Testing Overidentifying Restrictions with Many Instruments and Heteroskedasticity John C. Chao, Department of Economics, University of Maryland, chao@econ.umd.edu. Jerry A. Hausman, Department of Economics,

More information

Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses

Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses Ann Inst Stat Math (2009) 61:773 787 DOI 10.1007/s10463-008-0172-6 Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses Taisuke Otsu Received: 1 June 2007 / Revised:

More information

GMM-based inference in the AR(1) panel data model for parameter values where local identi cation fails

GMM-based inference in the AR(1) panel data model for parameter values where local identi cation fails GMM-based inference in the AR() panel data model for parameter values where local identi cation fails Edith Madsen entre for Applied Microeconometrics (AM) Department of Economics, University of openhagen,

More information

Mark Gales October y (x) x 1. x 2 y (x) Inputs. Outputs. x d. y (x) Second Output layer layer. layer.

Mark Gales October y (x) x 1. x 2 y (x) Inputs. Outputs. x d. y (x) Second Output layer layer. layer. University of Cambridge Engineering Part IIB & EIST Part II Paper I0: Advanced Pattern Processing Handouts 4 & 5: Multi-Layer Perceptron: Introduction and Training x y (x) Inputs x 2 y (x) 2 Outputs x

More information

Estimation of the Conditional Variance in Paired Experiments

Estimation of the Conditional Variance in Paired Experiments Estimation of the Conditional Variance in Paired Experiments Alberto Abadie & Guido W. Imbens Harvard University and BER June 008 Abstract In paired randomized experiments units are grouped in pairs, often

More information

Bootstrapping Heteroskedasticity Consistent Covariance Matrix Estimator

Bootstrapping Heteroskedasticity Consistent Covariance Matrix Estimator Bootstrapping Heteroskedasticity Consistent Covariance Matrix Estimator by Emmanuel Flachaire Eurequa, University Paris I Panthéon-Sorbonne December 2001 Abstract Recent results of Cribari-Neto and Zarkos

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Chapter 8 Maximum Likelihood Estimation 8. Consistency If X is a random variable (or vector) with density or mass function f θ (x) that depends on a parameter θ, then the function f θ (X) viewed as a function

More information

GMM Estimation and Testing

GMM Estimation and Testing GMM Estimation and Testing Whitney Newey July 2007 Idea: Estimate parameters by setting sample moments to be close to population counterpart. Definitions: β : p 1 parameter vector, with true value β 0.

More information

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Gerdie Everaert 1, Lorenzo Pozzi 2, and Ruben Schoonackers 3 1 Ghent University & SHERPPA 2 Erasmus

More information

Effi cient Estimation Using the Characteristic Function

Effi cient Estimation Using the Characteristic Function Effi cient Estimation Using the Characteristic Function Marine Carrasco University of Montreal Rachidi Kotchoni African School of Economics First draft: January 010 his version: January 015 Abstract he

More information

A Bootstrap Test for Conditional Symmetry

A Bootstrap Test for Conditional Symmetry ANNALS OF ECONOMICS AND FINANCE 6, 51 61 005) A Bootstrap Test for Conditional Symmetry Liangjun Su Guanghua School of Management, Peking University E-mail: lsu@gsm.pku.edu.cn and Sainan Jin Guanghua School

More information

Specification Test on Mixed Logit Models

Specification Test on Mixed Logit Models Specification est on Mixed Logit Models Jinyong Hahn UCLA Jerry Hausman MI December 1, 217 Josh Lustig CRA Abstract his paper proposes a specification test of the mixed logit models, by generalizing Hausman

More information

Locally Robust Semiparametric Estimation

Locally Robust Semiparametric Estimation Locally Robust Semiparametric Estimation Victor Chernozhukov Juan Carlos Escanciano Hidehiko Ichimura Whitney K. Newey The Institute for Fiscal Studies Department of Economics, UCL cemmap working paper

More information

Asymptotic inference for a nonstationary double ar(1) model

Asymptotic inference for a nonstationary double ar(1) model Asymptotic inference for a nonstationary double ar() model By SHIQING LING and DONG LI Department of Mathematics, Hong Kong University of Science and Technology, Hong Kong maling@ust.hk malidong@ust.hk

More information

Nonparametric Identication of a Binary Random Factor in Cross Section Data and

Nonparametric Identication of a Binary Random Factor in Cross Section Data and . Nonparametric Identication of a Binary Random Factor in Cross Section Data and Returns to Lying? Identifying the Effects of Misreporting When the Truth is Unobserved Arthur Lewbel Boston College This

More information

Empirical Risk Minimization

Empirical Risk Minimization Empirical Risk Minimization Fabrice Rossi SAMM Université Paris 1 Panthéon Sorbonne 2018 Outline Introduction PAC learning ERM in practice 2 General setting Data X the input space and Y the output space

More information

Can we do statistical inference in a non-asymptotic way? 1

Can we do statistical inference in a non-asymptotic way? 1 Can we do statistical inference in a non-asymptotic way? 1 Guang Cheng 2 Statistics@Purdue www.science.purdue.edu/bigdata/ ONR Review Meeting@Duke Oct 11, 2017 1 Acknowledge NSF, ONR and Simons Foundation.

More information

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines Maximilian Kasy Department of Economics, Harvard University 1 / 37 Agenda 6 equivalent representations of the

More information

Linear Models in Machine Learning

Linear Models in Machine Learning CS540 Intro to AI Linear Models in Machine Learning Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu We briefly go over two linear models frequently used in machine learning: linear regression for, well, regression,

More information

Asymptotic Distributions of Instrumental Variables Statistics with Many Instruments

Asymptotic Distributions of Instrumental Variables Statistics with Many Instruments CHAPTER 6 Asymptotic Distributions of Instrumental Variables Statistics with Many Instruments James H. Stock and Motohiro Yogo ABSTRACT This paper extends Staiger and Stock s (1997) weak instrument asymptotic

More information

Calibration Estimation for Semiparametric Copula Models under Missing Data

Calibration Estimation for Semiparametric Copula Models under Missing Data Calibration Estimation for Semiparametric Copula Models under Missing Data Shigeyuki Hamori 1 Kaiji Motegi 1 Zheng Zhang 2 1 Kobe University 2 Renmin University of China Economics and Economic Growth Centre

More information

MATH 205C: STATIONARY PHASE LEMMA

MATH 205C: STATIONARY PHASE LEMMA MATH 205C: STATIONARY PHASE LEMMA For ω, consider an integral of the form I(ω) = e iωf(x) u(x) dx, where u Cc (R n ) complex valued, with support in a compact set K, and f C (R n ) real valued. Thus, I(ω)

More information

7 Semiparametric Estimation of Additive Models

7 Semiparametric Estimation of Additive Models 7 Semiparametric Estimation of Additive Models Additive models are very useful for approximating the high-dimensional regression mean functions. They and their extensions have become one of the most widely

More information

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,

More information

DSGE Methods. Estimation of DSGE models: GMM and Indirect Inference. Willi Mutschler, M.Sc.

DSGE Methods. Estimation of DSGE models: GMM and Indirect Inference. Willi Mutschler, M.Sc. DSGE Methods Estimation of DSGE models: GMM and Indirect Inference Willi Mutschler, M.Sc. Institute of Econometrics and Economic Statistics University of Münster willi.mutschler@wiwi.uni-muenster.de Summer

More information

Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions

Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions Vadim Marmer University of British Columbia Artyom Shneyerov CIRANO, CIREQ, and Concordia University August 30, 2010 Abstract

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Chapter 7 Maximum Likelihood Estimation 7. Consistency If X is a random variable (or vector) with density or mass function f θ (x) that depends on a parameter θ, then the function f θ (X) viewed as a function

More information

On the Power of Tests for Regime Switching

On the Power of Tests for Regime Switching On the Power of Tests for Regime Switching joint work with Drew Carter and Ben Hansen Douglas G. Steigerwald UC Santa Barbara May 2015 D. Steigerwald (UCSB) Regime Switching May 2015 1 / 42 Motivating

More information

Department of Economics. Working Papers

Department of Economics. Working Papers 10ISSN 1183-1057 SIMON FRASER UNIVERSITY Department of Economics Working Papers 08-07 "A Cauchy-Schwarz inequality for expectation of matrices" Pascal Lavergne November 2008 Economics A Cauchy-Schwarz

More information

Asymptotic distributions of the quadratic GMM estimator in linear dynamic panel data models

Asymptotic distributions of the quadratic GMM estimator in linear dynamic panel data models Asymptotic distributions of the quadratic GMM estimator in linear dynamic panel data models By Tue Gørgens Chirok Han Sen Xue ANU Working Papers in Economics and Econometrics # 635 May 2016 JEL: C230 ISBN:

More information

A General Overview of Parametric Estimation and Inference Techniques.

A General Overview of Parametric Estimation and Inference Techniques. A General Overview of Parametric Estimation and Inference Techniques. Moulinath Banerjee University of Michigan September 11, 2012 The object of statistical inference is to glean information about an underlying

More information

Unconditional Quantile Regression with Endogenous Regressors

Unconditional Quantile Regression with Endogenous Regressors Unconditional Quantile Regression with Endogenous Regressors Pallab Kumar Ghosh Department of Economics Syracuse University. Email: paghosh@syr.edu Abstract This paper proposes an extension of the Fortin,

More information

NUCLEAR NORM PENALIZED ESTIMATION OF INTERACTIVE FIXED EFFECT MODELS. Incomplete and Work in Progress. 1. Introduction

NUCLEAR NORM PENALIZED ESTIMATION OF INTERACTIVE FIXED EFFECT MODELS. Incomplete and Work in Progress. 1. Introduction NUCLEAR NORM PENALIZED ESTIMATION OF IERACTIVE FIXED EFFECT MODELS HYUNGSIK ROGER MOON AND MARTIN WEIDNER Incomplete and Work in Progress. Introduction Interactive fixed effects panel regression models

More information

On prediction and density estimation Peter McCullagh University of Chicago December 2004

On prediction and density estimation Peter McCullagh University of Chicago December 2004 On prediction and density estimation Peter McCullagh University of Chicago December 2004 Summary Having observed the initial segment of a random sequence, subsequent values may be predicted by calculating

More information

Semiparametric Estimation of Partially Linear Dynamic Panel Data Models with Fixed Effects

Semiparametric Estimation of Partially Linear Dynamic Panel Data Models with Fixed Effects Semiparametric Estimation of Partially Linear Dynamic Panel Data Models with Fixed Effects Liangjun Su and Yonghui Zhang School of Economics, Singapore Management University School of Economics, Renmin

More information

Specification Test for Instrumental Variables Regression with Many Instruments

Specification Test for Instrumental Variables Regression with Many Instruments Specification Test for Instrumental Variables Regression with Many Instruments Yoonseok Lee and Ryo Okui April 009 Preliminary; comments are welcome Abstract This paper considers specification testing

More information

13 Endogeneity and Nonparametric IV

13 Endogeneity and Nonparametric IV 13 Endogeneity and Nonparametric IV 13.1 Nonparametric Endogeneity A nonparametric IV equation is Y i = g (X i ) + e i (1) E (e i j i ) = 0 In this model, some elements of X i are potentially endogenous,

More information

Improving linear quantile regression for

Improving linear quantile regression for Improving linear quantile regression for replicated data arxiv:1901.0369v1 [stat.ap] 16 Jan 2019 Kaushik Jana 1 and Debasis Sengupta 2 1 Imperial College London, UK 2 Indian Statistical Institute, Kolkata,

More information

is a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications.

is a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications. Stat 811 Lecture Notes The Wald Consistency Theorem Charles J. Geyer April 9, 01 1 Analyticity Assumptions Let { f θ : θ Θ } be a family of subprobability densities 1 with respect to a measure µ on a measurable

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 57 Table of Contents 1 Sparse linear models Basis Pursuit and restricted null space property Sufficient conditions for RNS 2 / 57

More information

Joint Estimation of Risk Preferences and Technology: Further Discussion

Joint Estimation of Risk Preferences and Technology: Further Discussion Joint Estimation of Risk Preferences and Technology: Further Discussion Feng Wu Research Associate Gulf Coast Research and Education Center University of Florida Zhengfei Guan Assistant Professor Gulf

More information

Time Series and Forecasting Lecture 4 NonLinear Time Series

Time Series and Forecasting Lecture 4 NonLinear Time Series Time Series and Forecasting Lecture 4 NonLinear Time Series Bruce E. Hansen Summer School in Economics and Econometrics University of Crete July 23-27, 2012 Bruce Hansen (University of Wisconsin) Foundations

More information

Least Squares Estimation of a Panel Data Model with Multifactor Error Structure and Endogenous Covariates

Least Squares Estimation of a Panel Data Model with Multifactor Error Structure and Endogenous Covariates Least Squares Estimation of a Panel Data Model with Multifactor Error Structure and Endogenous Covariates Matthew Harding and Carlos Lamarche January 12, 2011 Abstract We propose a method for estimating

More information

Econometric Analysis of Cross Section and Panel Data

Econometric Analysis of Cross Section and Panel Data Econometric Analysis of Cross Section and Panel Data Jeffrey M. Wooldridge / The MIT Press Cambridge, Massachusetts London, England Contents Preface Acknowledgments xvii xxiii I INTRODUCTION AND BACKGROUND

More information

Statistical Data Mining and Machine Learning Hilary Term 2016

Statistical Data Mining and Machine Learning Hilary Term 2016 Statistical Data Mining and Machine Learning Hilary Term 2016 Dino Sejdinovic Department of Statistics Oxford Slides and other materials available at: http://www.stats.ox.ac.uk/~sejdinov/sdmml Naïve Bayes

More information

On the econometrics of the Koyck model

On the econometrics of the Koyck model On the econometrics of the Koyck model Philip Hans Franses and Rutger van Oest Econometric Institute, Erasmus University Rotterdam P.O. Box 1738, NL-3000 DR, Rotterdam, The Netherlands Econometric Institute

More information

Lecture 1: Supervised Learning

Lecture 1: Supervised Learning Lecture 1: Supervised Learning Tuo Zhao Schools of ISYE and CSE, Georgia Tech ISYE6740/CSE6740/CS7641: Computational Data Analysis/Machine from Portland, Learning Oregon: pervised learning (Supervised)

More information

Outline Lecture 2 2(32)

Outline Lecture 2 2(32) Outline Lecture (3), Lecture Linear Regression and Classification it is our firm belief that an understanding of linear models is essential for understanding nonlinear ones Thomas Schön Division of Automatic

More information

GMM and SMM. 1. Hansen, L Large Sample Properties of Generalized Method of Moments Estimators, Econometrica, 50, p

GMM and SMM. 1. Hansen, L Large Sample Properties of Generalized Method of Moments Estimators, Econometrica, 50, p GMM and SMM Some useful references: 1. Hansen, L. 1982. Large Sample Properties of Generalized Method of Moments Estimators, Econometrica, 50, p. 1029-54. 2. Lee, B.S. and B. Ingram. 1991 Simulation estimation

More information

Location Properties of Point Estimators in Linear Instrumental Variables and Related Models

Location Properties of Point Estimators in Linear Instrumental Variables and Related Models Location Properties of Point Estimators in Linear Instrumental Variables and Related Models Keisuke Hirano Department of Economics University of Arizona hirano@u.arizona.edu Jack R. Porter Department of

More information

Semiparametric Estimation of Partially Linear Dynamic Panel Data Models with Fixed Effects

Semiparametric Estimation of Partially Linear Dynamic Panel Data Models with Fixed Effects Semiparametric Estimation of Partially Linear Dynamic Panel Data Models with Fixed Effects Liangjun Su and Yonghui Zhang School of Economics, Singapore Management University School of Economics, Renmin

More information

Generalized Method of Moments (GMM) Estimation

Generalized Method of Moments (GMM) Estimation Econometrics 2 Fall 2004 Generalized Method of Moments (GMM) Estimation Heino Bohn Nielsen of29 Outline of the Lecture () Introduction. (2) Moment conditions and methods of moments (MM) estimation. Ordinary

More information

Chapter 6: Derivative-Based. optimization 1

Chapter 6: Derivative-Based. optimization 1 Chapter 6: Derivative-Based Optimization Introduction (6. Descent Methods (6. he Method of Steepest Descent (6.3 Newton s Methods (NM (6.4 Step Size Determination (6.5 Nonlinear Least-Squares Problems

More information