Should We Go One Step Further? An Accurate Comparison of One-Step and Two-Step Procedures in a Generalized Method of Moments Framework

Size: px
Start display at page:

Download "Should We Go One Step Further? An Accurate Comparison of One-Step and Two-Step Procedures in a Generalized Method of Moments Framework"

Transcription

1 Sould We Go One Step Furter? An Accurate Comparison of One-Step and wo-step Procedures in a Generalized Metod of Moments Framework Jungbin Hwang and Yixiao Sun Department of Economics, University of California, San Diego Preliminary. Please do not circulate. Abstract According to te conventional asymptotic teory, te two-step Generalized Metod of Moments (GMM) estimator and test perform as least as well as te one-step estimator and test in large samples. e conventional asymptotics teory, as elegant and convenient as it is, completely ignores te estimation uncertainty in te weigting matrix, and as a result it may not re ect nite sample situations well. In tis paper, we employ te xed-smooting asymptotic teory tat accounts for te estimation uncertainty, and compare te performance of te one-step and two-step procedures in tis more accurate asymptotic framework. We sow te two-step procedure outperforms te one-step procedure only wen te bene t of using te optimal weigting matrix outweigs te cost of estimating it. is qualitative message applies to bot te asymptotic variance comparison and power comparison of te associated tests. A Monte Carlo study lends support to our asymptotic results. JEL Classi cation: C2, C32 Keywords: Asymptotic E ciency, Fixed-smooting Asymptotics, Heteroskedasticity and Autocorrelation Robust, Increasing-smooting Asymptotics, Asymptotic Mixed Normality, Nonstandard Asymptotics, wo-step GMM Estimation Introduction E ciency is one of te most important problems in statistics and econometrics. In te widelyused GMM framework, it is standard practice to employ a two-step procedure to improve te e ciency of te GMM estimator and te power of te associated tests. e two-step procedure requires te estimation of a weigting matrix. According to te Hansen (982), te optimal weigting matrix is te asymptotic variance of te (scaled) sample moment conditions. For time series data, wic is our focus ere, te optimal weigting matrix is usually referred to as te j6wang@ucsd.edu, yisun@ucsd.edu. Correspondence to: Department of Economics, University of California, San Diego, 95 Gilman Drive, La Jolla, CA

2 long run variance (LRV) of te moment conditions. o be completely general, we often estimate te LRV using te nonparametric kernel or series metod. Under te conventional asymptotics, bot te one-step and two-step GMM estimators are asymptotically normal. In general, te two-step GMM estimator as a smaller asymptotic variance. Statistical tests based on te two-step estimator are also asymptotically more powerful tan tose based on te one-step estimator. A driving force beind tese results is tat te two-step estimator and te associated tests ave te same asymptotic properties as te corresponding ones wen te optimal weigting matrix is known. However, given tat te optimal weigting matrix is estimated nonparametrically in te time series setting, tere is large estimation uncertainty. A good approximation to te distributions of te two-step estimator and te associated tests sould re ect tis relatively ig estimation uncertainty. One of te goals of tis paper is to compare te asymptotic properties of te one-step and twostep procedures wen te estimation uncertainty in te weiging matrix is accounted for. ere are two ways to capture te estimation uncertainty. One is to use te ig order conventional asymptotic teory under wic te amount of nonparametric smooting in te LRV estimator increases wit te sample size but at a slower rate. Wile te estimation uncertainty vanises in te rst order asymptotics, we expect it to remain in ig order asymptotics. e second way is to use an alternative asymptotic approximation tat can capture te estimation uncertainty even wit just a rst-order asymptotics. o tis end, we consider a limiting tougt experiment in wic te amount of nonparametric smooting is eld xed as te sample size increases. is leads to te so-called xed-smooting asymptotics in te recent literature. In tis paper, we employ te xed-smooting asymptotics to compare te one-step and twostep procedures. For te one-step procedure, te LRV estimator is used in computing te standard errors, leading to te popular eteroskedasticity and autocorrelation robust (HAR) standard errors. See, for example, Newey and West (987) and Andrews (99). For te two-step procedure, te LRV estimator not only appears in te standard error estimation but also plays te role of te optimal weigting matrix in te second-step GMM criterion function. Under te xed-smooting asymptotics, te weigting matrix converges to a random matrix. As a result, te second-step GMM estimator is not asymptotically normal but rater asymptotically mixed normal. e asymptotic mixed normality re ects te estimation uncertainty of te GMM weigting matrix and is expected to be closer to te nite sample distribution of te second-step GMM estimator. In a recent paper, Sun (24b) sows tat bot te one-step and two-step test statistics are asymptotically pivotal under tis new asymptotic teory. So a nuisance-parameter-free comparison of te one-step and two-step tests is possible. Comparing te one-step and two-step procedures under te new asymptotics is fundamentally di erent from tat under te conventional asymptotics. Under te new asymptotics, te twostep procedure outperforms te one-step procedure only wen te bene t of using te optimal weigting matrix outweigs te cost of estimating it. is qualitative message applies to bot te asymptotic variance comparison and te power comparison of te associated tests. is is in sarp contrast wit te conventional asymptotics were te cost of estimating te optimal weigting matrix is completely ignored. Since te new asymptotic approximation is more accurate tan te conventional asymptotic approximation, comparing te two procedures under tis new asymptotics will give an onest assessment of teir relative merits. is is con rmed by a Monte In tis paper, te one-step estimator refers to te rst-step estimator in te typical two-step GMM framework. It does not refer to te continuous updating GMM estimator tat involves only one step. We use te terms one-step and rst-step intercangingly. 2

3 Carlo study. ere is a large and growing literature on te xed-smooting asymptotics. For kernel LRV variance estimators, te xed-smooting asymptotics is te so-called te xed-b asymptotics rst studied by Kiefer, Vogelsang and Bunzel (22) and Kiefer and Vogelsang (22a, 22b, 25) in te econometrics literature. For oter studies, see, for example, Jansson (24), Sun, Pillips and Jin (28), Sun and Pillips (29), Gonçlaves and Vogelsang (2), and Zang and Sao (23) in te time series setting; Bester, Conley, Hansen and Vogelsang (24) and Sun and Kim (24) in te spatial setting; and Gonçalves (2), Kim and Sun (23), and Vogelsang (22) in te panel data setting. For OS LRV variance estimators, te xed-smooting asymptotics is te socalled xed-k asymptotics. For its teoretical development and related simulation evidence, see, for example, Pillips (25), Müller (27), and Sun (2, 23). e approximation approaces in some oter papers can also be regarded as special cases of te xed-smooting asymptotics. is includes, among oters, Ibragimov and Müller (2), Sao (2) and Bester, Conley, and Hansen (2). e xed-smooting asymptotics can be regarded as a convenient device to obtain some ig order terms under te conventional increasing-smooting asymptotics. e rest of te paper is organized as follows. e next section presents a simple overidenti ed GMM framework. Section 3 compares te two procedures from te perspective of point estimation. Section 4 compares tem from te testing perspective. Section 5 extends te ideas to a general GMM framework. Section 6 reports simulation evidence on te nite sample performances of te two procedures. Proofs of te main teorems are provided in te Appendix. A word on notation: for a symmetric matrix A; A =2 (or A =2 ) is a matrix square root of A suc tat A =2 A =2 = A: Note tat A =2 does not ave to be symmetric. We will specify A =2 explicitly wen it is not symmetric. If not speci ed, A =2 is a symmetric matrix square root of A based on its eigendecomposition. By de nition, A =2 = A =2 and so A = (A =2 ) (A =2 ): For matrices A and B; we use A B to signify tat A B is positive (semi)de nite. We use and O intercangeably to denote a matrix of zeros wose dimension may be di erent at di erent occurrences. For two random variables X and Y; we use X? Y to indicate tat X and Y are independent. 2 A Simple Overidenti ed GMM Framework o illustrate te basic ideas of tis paper, we consider a simple overidenti ed time series model of te form: y t = + u t ; y t 2 R d ; y 2t = u 2t ; y 2t 2 R q () for t = ; :::; were 2 R d is te parameter of interest and te vector process u t := (u t ; u 2t ) is stationary wit mean zero. We allow u t to ave autocorrelation of unknown forms so tat te long run variance of u t : X = lrvar(u t ) = Eu t u t j j= 3

4 takes a general form. However, for simplicity, we assume tat var(u t ) = 2 I d for te moment 2. Our model is just a location model. We initially consider a general GMM framework but later nd out tat our points can be made more clearly in te simple location model. In fact, te simple location model may be regarded as a limiting experiment in a general GMM framework. Embedding te location model in a GMM framework, te moment conditions are E(y t ) q = were y t = (y t ; y 2t ). Let g () = p P t= y t p P t= y 2t! : en a GMM estimator of can be de ned as ^GMM = arg min g () W g () for some positive de nite weigting matrix W : Writing W W W = 2 W 2 W 22 were W is a d d matrix and W 22 is a q q matrix, ten it is easy to sow tat ; ^GMM = X t= (y t W y 2t ) for W = W 2 W 22 : ere are at least two di erent coices of W. First, we can take W to be te identity matrix W = I m for m = d + q: In tis case, W = and te GMM estimator ^ is simply ^ = X y t : t= Second, we can take W to be te optimal weigting matrix W =. Wit tis coice, we obtain te GMM estimator: ~ 2 = X (y t y 2t ) ; t= were = 2 22 is te long run regression coe cient matrix. Wile ^ completely ignores te information in fy 2t g ; ~ 2 takes advantage of tis information. 2 If for any 2 > ; we can let V V 2 var (u t) = V := 6= 2 I V 2 V d 22 V =2 = (V 2) =2 V 2 (V 22) =2 (V 22) =2 en V =2 (y t; y 2t) can be written as a location model wose error variance is te identity matrix I d : e estimation uncertainty in estimating V will not a ect our asymptotic results.! : 4

5 were Under some moment and mixing conditions, we ave p ^ d =) N(; ) and p 2 = : ~2 d =) N (; 2 ) ; So Asym Var( ~ 2 ) < Asym Var(^ ) unless 2 =. is is a well known result in te literature. Since we do not know in practice, ~ 2 is infeasible. However, given te feasible estimator ^ ; we can estimate and construct a feasible version of ~ 2 : e common two-step estimation strategy is as follows. i) Estimate te long run covariance matrix by ^ := ^ (^u) = X X Q ( s ; t ) s= t= ^u t! X ^u = ^u s! X ^u = were ^u t = (y t ^ ; y 2t ). ii) Obtain te feasible two-step estimator ^ 2 = P t= y t ^y2t were ^ = ^ 2 ^ 22. In te above de nition of ^, Q (r; s) is a symmetric weigting function tat depends on te smooting parameter : For conventional kernel estimators, Q (r; s) = k ((r s) =b) and we take = =b: For te ortonormal series (OS) estimators, Q (r; s) = K P K j= j (r) j (s) and we take = K; were j (r) are ortonormal basis functions on L 2 [; ] satisfying R j (r) dr = : We parametrize in suc a way so tat indicates te level or amount of smooting for bot types of LRV estimators. Note tat we use te demeaned process f^u t P = ^u g in constructing ^ (^u) : For te location model, ^ (^u) is numerically identical to ^ (u) were te unknown error process fu t g is used. e moment estimation uncertainty is re ected in te demeaning operation. Had we known te true value of and ence te true moment process fu t g ; we would not need to demean fu t g. Wile ~ 2 is asymptotically more e cient tan ^ ; is ^ 2 necessarily more e cient tan ^ and in wat sense? Is te Wald test based on ^ 2 necessary more powerful tan tat based on ^? One of te objectives of tis paper is to address tese questions. 3 A ale of wo Asymptotics: Point Estimation We rst consider te conventional asymptotics were! as! but at a slower rate, i.e., =! : e asymptotic direction is represented by te curved arrow in Figure. Sun (24a, 24b) calls tis type of asymptotics te Increasing-smooting Asymptotics, as increases wit te sample size. Under tis type of asymptotics and some regularity conditions, we ave ^! p : It can ten be sown tat ^ 2 is asymptotically equivalent to ~ 2, i.e., p ( ~ 2 ^2 ) = o p (). As a direct consequence, we ave p d ^ =) N(; ); p d ^2 =) N ; : So ^ 2 is still asymptotically more e cient tan ^ : 5

6 Figure : Conventional Asymptotics (increasing smooting) vs New Asymptotics ( xed smooting) e conventional asymptotics, as elegant and convenient as it is, does not re ect te nite sample situations well. Under tis type of asymptotics, we essentially approximate te distribution of ^ by te degenerate distribution concentrating on. at is, we completely ignore te estimation uncertainty in ^: e degenerate approximation is too optimistic, as ^ is a nonparametric estimator, wic by de nition can ave ig variation in nite samples. o obtain a more accurate distributional approximation of p (^ 2 ); we could develop a ig order increasing-smooting asymptotics tat re ects te estimation uncertainty in ^. is is possible but requires strong assumptions tat cannot be easily veri ed. In addition, it is also tecnically callenging and tedious to rigorously justify te ig order asymptotic teory. Instead of ig order asymptotic teory under te conventional asymptotics, we adopt te type of asymptotics tat olds xed as! : is asymptotic beavior of and is illustrated by te arrow pointing to te rigt in Figure. Given tat is xed, we follow Sun (24a, 24b) and call tis type of asymptotics te Fixed-smooting Asymptotics. is type of asymptotics takes te sampling variability of ^ into consideration. Sun (23, 24a) as sown tat te xed-smooting asymptotic distribution is ig order correct under te conventional increasing-smooting asymptotics. So te xed-smooting asymptotics can be regarded as a convenient device to obtain some ig order terms under te conventional increasing-smooting asymptotics. o establis te xed-smooting asymptotics, we maintain Assumption on te kernel function and basis functions. 6

7 Assumption (i) For kernel HAR variance estimators, te kernel function k () satis es te following condition: for any b 2 (; ] and, k b (x) and k (x) are symmetric, continuous, piecewise monotonic, and piecewise continuously di erentiable on [ ; ]. (ii) For te OS HAR variance estimator, te basis functions j () are piecewise monotonic, continuously di erentiable and ortonormal in L 2 [; ] and R j (x) dx = : Assumption on te kernel function is very mild. It includes many commonly used kernel functions suc as te Bartlett kernel, Parzen kernel, and Quadratic Spectral (QS) kernel. De ne Q (r; s) = Q (r; s) Z Q (; s)d wic is a centered version of Q (r; s); and Z Q (r; )d + Z Z Q ( ; 2 )d d 2 ~ = X X Q ( s ; t )^u t^u s: s= t= Assumption ensures tat ~ and ^ are asymptotically equivalent. Furtermore, under tis assumption, Sun (24a) sows tat, for bot kernel LRV and OS LRV estimation, te centered weigting function Q (r; s) satis es : Q (r; s) = X j j (r) j (s) j= were f j (r)g is a sequence of continuously di erentiable functions satisfying R j(r)dr = and te series on te rigt and side converges to Q (r; s) absolutely and uniformly over (r; s) 2 [; ] [; ]. e representation can be regarded as a spectral decomposition of te compact Fredolm operator wit kernel Q (r; s) : See Sun (24a) for more discussion. Now, letting () := and using te basis functions f j ()g j= in te series representation of te weigting function, we make te following assumptions. Assumption 2 e vector process fu t g t= satis es: a) =2 P t= j(t= )u t converges weakly to a continuous distribution, jointly over j = ; ; :::; J for every xed J; b) For every xed J and x 2 R m,! X P p j ( t )u t x for j = ; ; ::; J = t=! X P =2 p j ( t )e t x for j = ; ; :::; J + o() as! t= were! =2 = =2 2 2 =2 22 =2 > 22 is a matrix square root of = P j= Eu tu t j and e t s iid N(; I m ) 7

8 Assumption 3 P j= k Eu tu t j k< : Proposition Let Assumptions 3 old. As! for a xed, we ave: (a) ^ =) d were = =2 ~ =2 := ; ;2 ;2 ;22 Z Z ~; ~ = Q (r; s)db ~ m(r)db m (s) ;2 := ~ ;2 ;22 ~ and B m () is a standard Brownian motion of dimension m = d + q; (b) p d ^2 =) I d ; =2 B m () were = (; d; q) := ;2 ;22 is independent of B m () : Conditional on, te asymptotic distribution of p ^2 is a normal distribution wit variance V 2 = I d ; 2 Id 2 22 = : at is, p (^ 2 ) is asymptotically mixed-normal rater tan normal. Since V = = almost surely, te feasible estimator ^ 2 as a large variation tan te infeasible estimator ~ 2 : is is consistent wit our intuition. Under te xed-smooting asymptotics, we still ave p d (^ ) =) N(; ) as ^ pdoes not depend on te smooting parameter : So te asymptotic (conditional) variances of (^ ) and p (^ 2 ) are bot larger tan in terms of matrix de niteness. However, it is not straigtforward to directly compare tem. o evaluate te relative magnitude of te two (conditional) asymptotic variances, we de ne ~ := ~ (; d; q) := ~ ;2 ~ ;22 ; (2) wic does not depend on any nuisance parameter but depends on ; d; q: For notational economy, we sometimes suppress tis dependence. Direct calculations sow tat Using tis, we ave, for any conforming vector a : = =2 2 ~ = : (3) Ea (V 2 ) a = Ea 22 a a a = a =2 2 E~ ~ = a =2 2 =2 2 a a a E ~ ~ = ( =2 2 ) i ( =2 2 ) a: (4) e following lemma gives a caracterization of E ~ (; d; q) ~ (; d; q) : 8

9 Lemma 2 For any d ; we ave E ~ (; d; q) ~ (; d; q) = Ejj ~ (; ; q) jj 2 I d : Let On te basis of te above lemma, we de ne g(; q) := Ejj~ (; ; q)jj 2 + Ejj ~ 2 (; ): (; ; q)jj2 = =2 2 ( =2 22 ) 2 R dq ; wic is te long run correlation matrix between u t and u 2t. Using (4) and rewriting it in terms of ; we obtain te following proposition. Proposition 3 Let Assumptions 3 old. If g(; q)i d is positive (negative) semide nite, ten ^ 2 as a larger (smaller) asymptotic variance tan ^ under te xed-smooting asymptotics. In te proof of te proposition, we sow tat te coices of te matrix square roots =2 and =2 22 are innocuous. o use te proposition, we need only to ceck weter te condition is satis ed for a given coice of te matrix square roots. If = ; ten g(; q)i d olds trivially. In tis case, te asymptotic variance of ^ 2 is larger tan te asymptotic variance of ^ : Intuitively, wen te long run correlation is zero, tere is no information tat can be explored to improve e ciency. If we insist on using te long run correlation matrix in attempt to improve te e ciency, we may end up wit a less e cient estimator, due to te noise in estimating te zero long run correlation matrix. On te oter and, if = I d after some possible rotation, wic olds wen te long run variation of u t is almost perfectly predicted by u 2t ; ten g(; q)i d <. In tis case, it is wortwile estimating te long run variance and using it to improve te e ciency ^ 2 : ere are many scenarios in between were te matrix g(; q)i d is inde nite. In tis case we may still be able to access te de niteness of te matrix along some directions; see eorem. e case wit d = is simplest. In tis case, if > g(; q); ten ^ 2 is asymptotically more e cient tan ^ : Oterwise, it is not asymptotically less e cient. In te case of kernel LRV estimation, it is ard to obtain an analytical expression for Ejj ~ (; ; q)jj 2 and ence g(; q), altoug we can always simulate it numerically. e tresold g(; q) depends on te smooting parameter = =b and te degree of overidenti cation q: ables 3 report te simulated values of g(; q) for b = : : : : :2 and q = 5. It is clear tat g(; q) increases wit q and decreases wit te smooting parameter = =b. Wen te OS LRV estimation is used, we do not need to simulate g(; q) as we can obtain a closed form expression for it. Using tis expression, we can prove te following corollary. Corollary 4 Let Assumptions 3 old. Consider te case of OS LRV estimation. If q K I d is positive (negative) semide nite, ten ^ 2 as a larger (smaller) asymptotic variance tan ^ under te xed-smooting asymptotics. A necessary and su cient condition for q=(k )I d to be positive semide nite is tat te largest eigenvalue of is smaller tan q=(k ): e eigenvalues of are te squared long run correlation coe cients between c u t and c 2 u 2t for some c and c 2 ; i.e., te square long run canonical correlation coe cients between u t and u 2t : So if te largest squared long run 9

10 canonical correlation is smaller tan q=(k ); ten ^ 2 as a larger asymptotic variance tan ^ : Similarly, if te smallest squared long run canonical correlation is greater tan q=(k ); ten ^ 2 as a smaller asymptotic variance tan ^ : Since ^ 2 is not asymptotically normal, asymptotic variance comparison does not paint te wole picture. o compare te asymptotic distributions of ^ and ^ 2, we consider te case tat d = q = and K = 4 as an example: Figure 2 represents te sapes of probability density functions in te case of OS LRV estimation wen ( ; 2 2 ; 22) = (; :; ). In tis case, = :9. e rst grap sows p (^ ) a N(; ) (red solid line) and p (^ 2 ) a N(; :9) (blue dotted line) under te conventional asymptotics. e conventional limiting distributions for p (^ ) and p (^ 2 ) are bot normal but te latter as a smaller variance, so te asymptotic e ciency of ^ 2 is always guaranteed. However, tis is not true in te second grap of Figure 2, wic represents te limiting distributions under te xed-smooting asymptotics. Here, as before te red solid line represents N(; ) but te blue dotted line represents te mixed normal distribution MN[; :9( + ~ 2 )] wit ~ = P K i= i 2i = P K i= 2 2i s MN(; = 2 K ): is can be obtained by using (4) wit a = : More speci cally, V 2 = =2 2 ~ ~ ( =2 2 ) = :9 ~ 2 + :9 = :9( + ~ 2 ): (5) Comparing tese two di erent families of distributions, we nd tat te asymptotic distribution of ^ 2 as fatter tail tan tat of ^ : e asymptotic variance of ^ 2 is :9( + =2) = :35 wic is larger tan te asymptotic variance of ^. More generally, wen d = q = wit ( ; 2 2 ; 22) = (; 2 ; ); we can use (4) to obtain te asymptotic variance of ^ 2 as follows: + 2 E ~ 2 = 2 + K = 2 + E ~ 2 : = 2 K K 2 So, for any coice of K, te asymptotic variance of ^ 2 decreases as 2 increases. Under te quadratic loss function, te asymptotic relative e ciency of ^ 2 relative to ^ is tus equal to 2 (K ) = (K 2) : e second factor (K ) = (K 2) > re ects te cost of aving to estimate : Wen 2 = =(K ); ^ and ^ 2 ave te same asymptotic variance. For a given coice of K, if 2 =(K ), we prefer to use ^. is is not captured by te conventional asymptotics, wic always cooses ^ 2 over ^. 4 A ale of wo Asymptotics: Hypotesis esting We are interested in testing te null ypotesis H : R = r against te local alternative H : R = r + = p for some p d matrix R and p vectors r and : Nonlinear restrictions can be converted into linear ones using te Delta metod. We construct te following two Wald statistics: W := (R^ r) R^ R (R^ r) W 2 := (R^ 2 r) R^ 2 R i (R^2 r)

11 Conventional Asymptoptic Distributions One step estimator: N(,) wo step estimator: N(,.9) Fixed smootign Asymptotic Distributions One step estimator: N(,) wo step estimator: MN(,.9(+ 2 )) Figure 2: Limiting distributions of ^ and ^ 2 based on te OS LRV estimator wit K = 4: were ^ 2 = ^ ^2 ^ 22 2 : Wen p = and te alternative is one sided, we can construct te following two t statistics: : = p R^ r pr^ R (6) p R^ 2 r 2 : = pr^ 2 R : (7) No matter weter te test is based on ^ or ^ 2 ; we ave to employ te long run covariance estimator ^. De ne te p p matrices and 2 according to = R R and 2 2 = R 2 R : In oter words, and 2 are matrix square roots of R R and R 2 R respectively.

12 Under te conventional increasing-smooting asymptotics, it is straigtforward to sow tat under H : R = r + = p : W d =) 2 p( Wen = ; we obtain te null distributions: 2 ), W 2 d =) 2 p( 2 d =) N( ; ), 2 d =) N( 2 ; ): W ; W 2 d =) 2 d p and ; 2 =) N(; ): So under te conventional increasing-smooting asymptotics, te null limiting distributions of W and W 2 are identical. Since 2 2 2, under te conventional asymptotics, te local asymptotic power function of te test based on W 2 is iger tan tat based on W : e key driving force beind te conventional asymptotics is tat we approximate te distribution of ^ by te degenerate distribution concentrating on : e degenerate approximation does not re ect te nite sample distribution well. As in te previous section, we employ te xed-smooting asymptotics to derive more accurate distributional approximations. Let and C pp = C qq = Z Z Z Z Q (r; s)db p(r)db p (s) ; C pq = Q (r; s)db q(r)db q (s) ; C qp = C pq Z Z D pp = C pp C pq C qq C pq 2 ), Q (r; s)db p(r)db q (s) were B p () 2 R p and B q () 2 R q are independent standard Brownian motion processes. Proposition 5 Let Assumptions 3 old. As! for a xed, we ave, under H : R = r + = p : d (a) W =) W ( 2 ) were (b) W 2 (c) (d) 2 d =) W 2 ( W (kk 2 ) = [B p () + ] C pp [B p () + ] for 2 R p : (8) 2 2 ) were W 2 (kk 2 ) = B p () C pq Cqq B q () + D pp d =) d =) 2 2 Bp () C pq C qq B q () + : (9) := Bp ()+ p = Cpp for p = : := Bp () C pq Cqq B q () + 2 p = Dpp for p = : In Proposition 5, we use te notation W (kk 2 ), wic implies tat te rigt and side of (8) depends on only troug kk 2 : is is true, because for any ortogonal matrix H : [B p () + ] C pp [B p () + ] = [HB p () + H] HC pp H [HB p () + H] d = [B p () + H] C pp [B p () + H] : 2

13 If we coose H = (= kk ; ~ H) for some ~ H suc tat H is ortogonal, ten [B p () + ] C pp [B p () + ] d = [B p () + kk e p ] C pp [B p () + kk e p ] : So te distribution of [B p () + ] C pp [B p () + ] depends on only troug kk : Similarly, te distribution of te rigt and side of (9) depends only on kk 2 : Wen = ; we obtain te limiting distributions of W ; W 2 ; and 2 under te null ypotesis: W W 2 2 d =) W := W () = B p () C pp B p () ; d =) W 2 := W 2 () = B p () C pq C d =) := () = B p ()= p C pp ; qq B q () D pp d =) 2 := 2 () = B p () C pq C qq B q () = p D pp : Bp () C pq C qq B q () ; ese distributions are di erent from tose under te conventional asymptotics. For W and ; te di erence lies in te random scaling factor C pp or p C pp : e random scaling factor captures te estimation uncertainty of te LRV estimator. For W 2 and 2 ; tere is an additional di erence embodied by te random location sift C pq Cqq B q () wit a consequent cange in te random scaling factor. e proposition below provides some caracterization of te two limiting distributions W and W 2 : Proposition 6 For any x > ; te following old: (a) W 2 () rst-order stocastically dominates (FSD) W () in tat P [W 2 () x] > P [W () x] : i i (b) P W (kk 2 ) x strictly increases wit kk 2 and lim kk! P W (kk 2 ) x = : i i (c) P W 2 (kk 2 ) x strictly increases wit kk 2 and lim kk! P W 2 (kk 2 ) x = : Proposition 6(a) is intuitive. W 2 FSD W because W 2 FSD B p () Dpp B p (), wic in turn FSD B p () Cpp B p () wic is just W. According to a property of te rst-order stocastic dominance, we ave W 2 d = W + W e for some W e > : Intuitively, W 2 sifts some of te probability mass of W to te rigt. A direct implication is tat te asymptotic critical values for W 2 are larger tan te corresponding ones for W : e di erence in critical values as implications on te power properties of te two tests. For x > ; we ave P ( > x) = 2 P W x 2 and P ( 2 > x) = 2 P W 2 x 2. It ten follows from Proposition 6(a) tat P ( 2 > x) P ( > x) for x > : So for a onesided test wit te alternative H : R > r; critical values from 2 are larger tan tose from 3

14 : Similarly, we ave P ( 2 < x) P ( < x) for x < : is implies tat for a one-sided test wit te alternative H : R < r; critical values from 2 are smaller tan tose from : Let W and W 2 be te ( ) quantile from te distributions W and W 2 ; respectively. e local asymptotic power functions of te two tests are 2 := 2 ; K; p; q; = P W ( i 2 ) > W ; 2 2 := 2 2 ; K; p; q; = P W 2 ( i 2 ) > W 2 : 2 Wile ; we also ave W 2 > W : e e ects of te critical values and te noncentrality parameter move in opposite directions. It is not straigtforward to compare te two power functions. However, Proposition 6 suggests tat if te di erence in te noncentrality parameters is large enoug to o set te increase in critical values, ten te two-step test based on W 2 will be more powerful. o evaluate 2 2 ; we de ne 2 2 R = R R =2 (R2 ) =2 22 ; () wic is te long run correlation matrix R between Ru t and u 2t : In terms of R 2 R pq we ave = R R R R R R = I p R 222 2R i n = Ip R R = ( Ip R =2) R R R Ip R =2 R > : I p o So te di erence in te noncentrality parameters depends on te matrix R R : Let R = P min(p;q) i= i a i b i be te singular value decomposition of R were fa i g and fb i g are ortonormal vectors in R p and R q respectively. Sorted in te descending order, 2 i are te (squared) long run canonical correlation coe cients between Ru t and u 2t : en = min(p;q) X i= 2 i 2 i a i 2 : Consider a special case tat 2 := max p i= 2 i approaces. If a 6= ; ten jj2 jj 2 2 and ence jj2 jj 2 approaces as 2 approaces from below. is case appens wen te second block of moment conditions as very ig long run prediction power for te rst block. In tis case, we expect te W 2 test to be more powerful, as lim j j! 2 ( 2 2 ) = : Consider anoter special case tat max i 2 i = ; i.e., R is a matrix of zeros. In tis case, te second block of moment conditions contains no additional information, and we ave 2 2 = 2 : In tis case, we expect te W 2 test to be less powerful. It follows from Proposition 6(b)(c) tat for any, tere exists a unique () := (; ; p; q; ) suc tat () = 2 ( ()) : 4

15 en 2 ( 2 2 ) < ( 2 ) if and only if 2 2 ( 2 ) 2 = ( 2 ) Ip R =2 R R R 2 f 2 < ( 2 ) 2 : But 2 i Ip I p R =2 R ; were (; ; p; q; ) f () := f(; ; p; q; ) = : (; ; p; q; ) So te power comparison depends on weter f( 2 )I p R R is positive semide nite. Proposition 7 Let Assumptions 3 old and de ne := 2 = (R R ). If f(; ; p; q; )I p R R is positive (negative) semide nite, ten te two-step test based on W 2 as a lower (iger) local asymptotic power tan te one-step test based on W under te xedsmooting asymptotics. Note tat R R = R R =2 R R R R =2 : o pin down R R ; we need to coose a matrix square root of R R : Following te arguments in te proof of Proposition 3, it is easy to see tat we can use any matrix square root. In particular, we can use te principal matrix square root. Wen p = ; wic is of ultimate importance in empirical studies, R R is equal to te sum of te squared long run canonical correlation coe cients. In tis case, f(; ; p; q; ) is te tresold value of R R for assessing te relative e ciency of te two tests. More speci cally, wen R R > f(; ; p; q; ), te two-step W 2 test is more powerful tan te one-step W test. Oterwise, te two-step W 2 test is less powerful. Proposition 7 is in parallel wit Proposition 3. e qualitative messages of tese two propositions are te same wen te long run correlation is ig enoug, we sould estimate and exploit it to reduce te variation of our point estimator and improve te power of te associated tests. However, te tresolds are di erent quantitatively. e two propositions fully caracterize te tresold for eac criterion under consideration. Proposition 8 Consider te case of OS LRV estimation. For any 2 R +, we ave () > 2 () and ence (; ; p; q; ) > and f(; ; p; q; ) > : Proposition 8 is intuitive. Wen tere is no long run correlation between Ru t and u 2t ; we ave 2 2 = 2 : In tis case, te two-step W 2 test is necessarily less powerful. e proof uses te teory of uniformly most powerful invariant tests and te teory of complete and su cient statistics. It is an open question weter te same strategy can be adopted to prove Proposition 8 in te case of kernel LRV estimation. Our extensive numerical work supports tat (; ; p; q; ) > and f(; ; p; q; ) > continue to old in te kernel case. It is not easy to give an analytical expression for f(; ; p; q; ) but we can compute it numerically witout any di culty. In ables 4 and 5, we consider te case of OS LRV estimation and compute te values of f(; K; p; q; ) for = ; K = 8; ; 2; 4, p = 4 and q = 3: Similar to te asymptotic variance comparison, we nd tat tese tresold values increase as te degree of over-identi cation increases and decrease as te smooting parameter K increases. Note tat te continuity of two power functions (; K; p; q; ) and 2 (; K; p; q; ) guarantees 5

16 tat for 2 [; ] te two-step W 2 test as a lower local asymptotic power tan te one-step W test if te largest eigenvalue of R R is smaller tan min 2[;] f(; K; p; q; ). on te oter and, for 2 [; ] te two-step W 2 test as a iger local asymptotic power if te smallest eigenvalue of R R is greater tan max f(; K; p; q; ). For te case of kernel LRV estimator, results not reported ere sow tat f(; ; p; q; ) increases wit q and decreases wit. is is entirely analogous to te case of OS LRV estimation. 5 General Overidenti ed GMM Framework In tis section, we consider te general GMM framework. e parameter of interest is a d vector 2 R d. Let v t 2 R dv denote te vector of observations at time t. We assume tat is te true value, an interior point of te parameter space. e moment conditions E f(v t ; ) = ; t = ; 2; :::; : old if and only if = were f (v t ; ) is an m vector of continuously di erentiable functions. e process f (v t ; ) may exibit autocorrelation of unknown forms. We assume tat m d and tat te rank of E[@ f (v t ; ) =@ ] is equal to d: at is, we consider a model tat is possibly over-identi ed wit te degree of over-identi cation q = m d: 5. One-Step and wo-step Estimation and Inference De ne te m m contemporaneous covariance matrix and te LRV matrix as: = E f(v t ; ) f(v t ; ) and = X j= j were j = E f(v t ; ) f(v t j ; ) : Let g t () = p tx f(v j ; ): j= Given a simple positive-de nite weigting matrix W tat does not depend on any unknown parameter, we can obtain an initial GMM estimator of as ^ = arg min g () W g (): For example, we may set W equal to I m. In te case of IV regression, we may set W equal to Z Z = were Z is te matrix of te instruments. Using or as te weigting matrix, we obtain te following two (infeasible) GMM estimators: ~ : = arg min () g (); 2 () ~ 2 : = arg min () g (): 2 (2) For te estimator ~ ; we use te contemporaneous covariance matrix as te weigting matrix and ignore all te serial dependency in te moment vector process ff(v t ; )g t=. In contrast to tis procedure, te second estimator ~ 2 accounts for te long run dependency. e feasible 6

17 versions of tese two estimators ^ and ^ 2 can be naturally de ned by replacing and wit teir estimates est (^ ) and est (^ ) were est () : = est () : = X f(v t ; ) f(v t ; ) ; (3) t= X s= t= X Q ( s ; t ) f(v t ; ) f(v s ; ) : (4) o test te null ypotesis H : R = r against H : R = r + = p ; we construct two di erent Wald statistics as follows: W : = (R^ r) n R ^V R o (R^ r); (5) W 2 : = (R^ 2 r) n R ^V 2 R o (R^2 r); were ^V = ^V 2 = G est (^ ) G i G est (^ ) est (^ ) est (^ ) G i G est (^ ) G i (6) G2 est (^ 2 ) G i 2 and G = X f(v t ; ; G2 = =^ X f(v t ; : =^2 ese are te standard Wald test statistics in te GMM framework. o compare te two estimators ^ and ^ 2 and associated tests, we maintain te standard assumptions below. Assumption 4 As! ; ^ = +o p () ; ^ = +o p () ; ^ 2 = +o p () for an interior point 2 : Assumption 5 De ne G t () = = tx f(v t ; for t and G () = : For any = + o p (); te following olds: (a) plim! G[r ] ( ) = r G uniformly in r were G = G( ) and G() = E@ f(v t ; )=@ ; (b) est ( ) p! > : Wit tese assumptions and some mild conditions, te standard GMM teory gives us p (^ ) = p X t= G G i G f(vt ; ) + o p (): Under te xed-smooting asymptotics, Sun (24b) establises te representation: p (^2 ) = p X t= G G i G f(v t ; ) + o p (); 7

18 were is de ned in te similar way as in Proposition : = =2 ~ =2 : Due to te complicated structure of two transformed moment vector processes, it is not straigtforward to compare te asymptotic distributions of ^ and ^ 2 as in Sections 3 and 4. o confront tis callenge, we let G = U V mm md dd be a singular value decomposition (SVD) of G, were = A dd ; O dq ; A is a d d diagonal matrix and O is a matrix of zeros. Also, we de ne f (v t ; ) = (f (v t ; ); f 2 (v t ; )) := U f(vt ; ) 2 R m ; were f (v t; ) 2 R d and f2 (v t; ) 2 R q are te rotated moment conditions. e variance and long run variance matrices of ff (v t ; )g are := U U = and := U U. o convert te variance matrix into an identity matrix, we de ne te normalized moment conditions below: were f(v t ; ) = f (v t ; ) ; f 2 (v t ; ) := ( =2 ) f (v t ; ) =2 = ( 2 )=2 2 ( 22 ) =2 ( 22 )=2! : (7) More speci cally, f (v t ; ) : = ( 2) =2 f (v t ; ) i 2 ( 22) f2 (v t ; ) 2 R d ; f 2 (v t ; ) : = ( 22) =2 f 2 (v t ; ) 2 R q : en te contemporaneous variance of te time series ff(v t ; )g is I m and te long run variance is := ( =2 ) ( =2 ) : Lemma 9 Let Assumptions 5 old by replacing u t wit f(v t ; ) in Assumptions 2 and 3. en as! for a xed ; ( 2) =2 AV p (^ ) = p ( 2) =2 AV p (^ 2 ) = p X t= X t= f (v t ; ) + o p () were := ;2 ;22 is te same as in Proposition. d =) N(; ) (8) [f (v t ; ) f 2 (v t ; )] + o p () (9) d =) MN ;

19 e asymptotic normality and mixed normality in te lemma are not new. e limiting distribution of p (^ ) is available from te standard GMM teory wile tat of p (^ 2 ) as been recently establised in Sun (24b). It is te representation tat is new and insigtful. e representation casts te two estimators in te same form and enables us to directly compare teir asymptotic properties. It follows from te proof of te lemma tat ( 2) =2 AV p ( ~ 2 ) = p X t= [f (v t ; ) f 2 (v t ; )] + o p () were = 2 22 as de ned before. So te di erence between te feasible and infeasible twostep GMM estimators lies in te uncertainty in estimating : Wile te true value of appears in te asymptotic distribution of te infeasible estimator ~ 2 ; te xed-smooting limit of te implied estimator ^ := ^ 2 ^ 22 appears in tat of te feasible estimator ^ 2 : It is important to point out tat te estimation uncertainty in te wole weigting matrix est matters only troug tat in ^: If we let (u t ; u 2t ) = (f (v t ; ); f 2 (v t ; )); ten te rigt and sides of (8) and (9) are exactly te same as wat we would obtain in te location model. e location model, as simple as it is, as implications for general settings from an asymptotic point of view. More speci cally, de ne y t = ( 2) =2 AV + u t ; y 2t = u 2t ; were u t = f (v t ; ) and u 2t = f 2 (v t ; ). e estimation and inference problems in te GMM setting are asymptotically equivalent to tose in te above simple location model wit fy t ; y 2t g as te observations. o present our next teorem, we transform R into ~ R using wic as te same dimension as R: We let Z Z Z ~ (; p; q) = Q (r; s)db p(r)db q (s) wic is compatible wit te de nition in (2). We de ne = =2 2 = R dq and R = ~R = RV A ( 2) =2 ; (2) Z Q (r; s)db q(r)db q (s) ; ~R ~ R =2 ~R2 = R pq : Wile is te long run correlation matrix between f (v t ; ) and f 2 (v t ; ), R is te long run correlation matrix between Rf ~ (v t ; ) and f 2 (v t ; ). Finally, we rede ne te noncentrality parameter : = R( G G) ( G G)( G G) R i ; (2) wic is te noncentrality parameter based on te rst-step test. For te location model considered before, te above de nition of R is identical to tat in (). In tat case, we ave G = (I d ; O dq ) and so U = I m, A = I d and V = I d : Given te assumption tat = = I m ; wic implies tat 2 = I d; we ave ~ R = R: In addition, te noncentrality parameter reduces to = (R R ) as de ned in Proposition 7. 9

20 eorem Let te assumptions in Lemma 9 old. (a) If g(; q) I d is positive (negative) semide nite, ten ^ 2 as a larger (smaller) asymptotic variance tan ^ as! for a xed : (b) If g(; q) I p R R is positive (negative) semide nite, ten R^ 2 as a larger (smaller) asymptotic variance tan R^ as! for a xed : (c) If f (; ; p; q; ) I p R R is positive (negative) semide nite, ten te test based on W 2 is asymptotically less (more) powerful tan tat based on W ; as! for a xed : eorem (a) is a special case of eorem (b) wit R = I d. We single out tis case in order to compare it wit Proposition 3. It is reassuring to see tat te results are te same. e only di erence is tat in te general GMM case we need to rotate and standardize te original moment conditions before computing te long run correlation matrix. Part (b) can also be applied to a general location model wit a nonscalar error variance, in wic case ~ R = R ( 2 )=2. eorem (c) reduces to Proposition 7 in te case of te location model. 5.2 wo-step GMM Estimation and Inference wit a Working Weigting Matrix In te previous subsection, we employ two speci c weigting matrices te variance and long run variance estimators. In tis subsection, we consider a general weigting matrix W (^ ); wic may depend on te initial estimator ^ and te sample size ; leading to yet anoter GMM estimator: i ^a = arg min g () W (^ ) g () 2 were te subscript a signi es anoter or alternative. An example of W (^ ) is te implied LRV matrix wen we employ a simple approximating parametric model to capture te dynamics in te moment process. We could also use te general LRV estimator but we coose a large so tat te variation in W (^ ) is small. In te kernel LRV estimation, tis amounts to including only autocovariances of low orders in constructing W (^ ): We assume tat W (^ )! p W, a positive de nite nonrandom matrix under te xed-smooting asymptotics. W may not be equal to te variance or long run variance of te moment process. We call W (^ ) a working weigting matrix. is is in te same spirit of using a working correlation matrix rater tan a true correlation matrix in te generalized estimating equations (GEE) setting. See, for example, Liang and Zeger (986). In parallel to (5), we construct te test statistic W a := (R^ a r) n R ^V a R o (R^a r); were, for G a = P f(v t ; )=@ ; =^a ^V a is de ned according to ^V a = G a W (^ a ) G a i G a W (^ a ) est ^a W (^ a ) G a i G a W (^ a ) G a i ; wic is a standard variance estimator for ^ a : De ne W = U W U and W = =2 W ( =2 ) W W = 2 W 2 W 22 2

21 and a = W 2 W 22 : Using te same argument for proving Lemma 9, we can sow tat ( 2) =2 A V p (^ a ) = p X t= [f (v t ; ) a f 2 (v t ; )] + o p (): (22) e above representation is te same as tat in (9) except tat is now replaced by a : De ne ~ a according to a = =2~ 2 a = : at is, ~ a = =2 2 ( a ) =2 22 = =2 2 a =2 22 I d =2 : (23) e relationsip between a and ~ a is entirely analogous to tat between and ~ ; see (3). In fact, if we let ~W = =2 W =2 ~W W2 ~ = ~W 2 W22 ~ ; ten ~ a = W ~ 2 W ~ 22 ; wic is te long run regression matrix implied by te normalized weigting matrix W ~ : Let V a be te long run variance of R ~ [f (v t ; ) a f 2 (v t ; )] and a;r = Va =2 ~R (2 a 22 )i =2 22 ; be te long run correlation matrix between R ~ [f (v t ; ) R = I d ; a;r reduces to a f 2 (v t ; )] and f 2 (v t ; ) : Wen Let a = + a 22 a a 2 2 =2 a (2 a 22 ) =2 22 : ~R a = RV A ( 2) =2 =2 2 = R ~ =2 2 and D ~ =2 R = ~Ra R ~ ~Ra a = ~R2 R ~ =2 ~R =2 2 : ~D R as te same dimension as ~ R a and R: By construction, ~ DR ~ D R = I p, tat is, eac row of ~ D R is a unit vector in R d and te rows of ~ D R are ortogonal to eac oter. e matrix ~ D R signi es te ortogonal directions embodied in ~ R a : Wit tese additional notations, we are ready to state te teorem below. eorem Let te assumptions in Lemma 9 old. Assume furter tat W (^ )! p positive de nite nonrandom matrix. (a) If E ~ (; d; q) ~ (; d; q) a ~ ~ a is positive (negative) semide nite, ten ^ 2 as a larger (smaller) asymptotic variance tan ^ a ; as! for a xed : (b) If E ~ (; p; q) ~ (; p; q) ~ DR ~ a ~ a ~ D R is positive (negative) semide nite, ten R^ 2 as a larger (smaller) asymptotic variance tan R^ a ; as! for a xed : (c) If g(; d) I d a a is positive (negative) semide nite, ten ^ 2 as a larger (smaller) asymptotic variance tan ^ a ; as! for a xed : (d) If g(; d)i p a;r a;r is positive (negative) semide nite, ten R^ 2 as a larger (smaller) asymptotic variance tan R^ a ; as! for a xed : (e) Let a = jjva =2 jj 2 : If f ( a ; ; p; q; ) I p a;r a;r ten te test based on W 2! for a xed : W, a is positive (negative) semide nite, is asymptotically less (more) powerful tan tat based on W a ; as 2

22 eorem (a) is intuitive. Wile E ~ (; d; q) ~ (; d; q) can be regarded as te variance in ation arising from estimating te long run regression matrix, ~ ~ a a can be regarded as te bias e ect from not using a true long run regression matrix. Wen te variance in ation e ect dominates te bias e ect, te two-step estimator ^ 2 will ave a larger asymptotic variance tan ^a : In te special case wen W = ; we obtain te infeasible two-step estimator ~ 2 and ~ a = : In tis case, E ~ (; d; q) ~ (; d; q) ~ ~ a a olds trivially and te feasible two-step estimator ^2 as iger variation tan its infeasible version ~ 2 : In anoter special case wen W = ; we obtain te rst step estimator ^ in wic case a = and by (23), ~ a = (I d ) =2 : Hence ~ a ~ a = I d =2 I d =2 : e condition in eorem (a) reduces to weter g(; q) I q not. eorem (a) tus reduces to eorem (a). is positive semide nite or eorem (b) is a rotated version of eorem (a). Given tat ~ D R ~ (; d; q) d = ~ (; p; q); te condition in eorem (a) implies tat in eorem (b). So part (b) of te teorem is stronger tan part (a) wen R is not a square matrix. Wen W = ; we ave a = and so ~D R ~ a ~ a ~ D R = = = ~R2 R ~ =2 i i ~R =2 ~a ~ 2 ~R =2 ~R2 ~ =2 a 2 R ~R2 R ~ =2 ~R (a ) 22 f( a )g R ~ i R2 ~ R ~ =2 ~R2 ~ R =2 ~R ~ R i ~ R2 ~ R =2 = I p R R =2 R R Ip R =2 R : As a result, te condition E ~ (; p; q) ~ (; p; q) D ~ R a ~ ~ ~ a DR in eorem (b) is equivalent to E ~ (; p; q) ~ (; p; q) I p R =2 R R R Ip R =2 R ; wic can be reduced to te same condition in eorem (b). eorem (c)-(e) is entirely analogous to eorem (a)-(c). e only di erence is tat te second block of moment conditions is removed from te rst block using te implied matrix coe cient a before computing te long run correlation coe cient. o understand (e), we can see tat te e ective moment conditions beind R^ a are: Ef a (v t ; ) = for f a (v t ; ) = RV A ( 2) =2 [f (v t ; ) a f 2 (v t ; )] : R^ a uses te information in Ef 2 (v t ; ) = to some extent, but it ignores te residual information tat is still potentially available from Ef 2 (v t ; ) = : In contrast, R^ 2 attempts to explore te residual information. If tere is no long run correlation between f a (v t ; ) and f 2 (v t ; ) ; i.e., ar = ; ten all te information in Ef 2 (v t ; ) = as been fully captured by te e ective moment conditions underlying R^ a : As a result, te test based on R^ a necessarily outperforms tat based on R^ 2 : If te long run correlation ar is large enoug in te sense tat is given in (e), te test based on R^ 2 could be more powerful tan tat based on R^ a in large samples. 22

23 5.3 Practical Recommendation eorems and give teoretical comparisons of various procedures. But tere are some unknown quantities in te two eorem. In practice, given te set of moment conditions Ef(v t ; ) = and te data fv t g ; suppose tat we want to test H : R = r against R 6= r for some R 2 R pd We can follow te steps below to evaluate te relative merit of one-step and two-step testing procedures.. Compute te initial estimator ^ P = arg min t= f(v t ; ) 2 : 2. One te basis of ^ ; use a data-driven metod to select te smooting parameter: Denote te data-driven value by ^: 3. Compute est (^ ) and est (^ ) using te formulae in (4) and smooting parameter f(v t;) 4. Compute G (^ ) = P t= =^ and its singular value decomposition were ^ = ( ^A dd ; O dq ) and ^A dd is diagonal. ^U ^ ^V 5. Estimate te variance and te long run variance of te rotated moment processes by ^ := ^U est (^ ) ^U and ^ := ^U est (^ ) ^U. 6. Compute te normalized LRV estimator: ^ ^ = (^ =2 ) ^ (^ =2 ) := ^ 2 ^2 ^22 were ^ =2 = ^ 2 =2 ^ 2 ^ 22 =2 ^ 22 =2 C A : (24) 7. Let ~ R est = R ^V ^A (^ 2 )=2 and compute ^ R^ R = ( ~ R est ^ ~ R est ) =2i ~ Rest ^2 ^ 22 ^ 2 ~ R est i ( ~ R est ^ ~ R est ) =2i and its largest eigenvalue v max ^ R^ R and smallest eigenvalue vmin ^ R^ R : 8. Coose te value of o suc tat P 2 p () > p = 75%. We may also coose a value of to re ect scienti c interest or economic signi cance. 9. (a) If v min ^ R^ R > f o ; ^; p; q;, ten we use te second-step testing procedure based on W 2 : (b) If v max ^ R^ R < f o ; ^; p; q; ; ten we use te rst-step testing procedure based on W : (c) If eiter te condition in (a) or in (b) is satis ed, ten we use te rst-step testing procedure based on W a : 23

24 6 Simulation Evidence is section compares te nite sample performances of one-step and two-step estimators and tests using te xed-smooting approximation. 6. Point Estimation We consider te location model given in () wit te true parameter value = (; :::; ) 2 R d but we allow for a nonscalar error variance. e error fu t g follows a VAR() process: u i t = u i t u i 2t = u i 2t u j 2t j= + p q qx + e i 2t for i = ; :::; q + ei t for i = ; :::; d (25) were e i t iid N(; ) across i and t, ei 2t iid N(; ) across i and t; and fe t ; t = ; 2; :::; g are independent of fe 2t ; t = ; 2; :::; g : Let u t := (u t ; u 2t ) 2 R m, ten u t = u t + e t were mm = I d p q J d;q I q! ; e t = et e 2t iid N (; I m ) : Direct calculations give us te expressions for te long run and contemporaneous variances of fu t g as = = X Eu t u t j = (I m ) (I m ) I ( ) 2 d + 2 ( ) 3 p q J q;d I ( ) 2 q J ( ) 4 d ( ) 3 p q J d;q A and = var(u t ) = I 2 d + 2 (+ 2 ) J ( 2 ) 3 d pq J ( 2 ) 2 dq p q ( 2 ) 2 J qd 2 I q C A ; were J d;q is te d q matrix of ones. In our simulation, we x te value of at :75 i so tat eac time series is reasonably persistent. Let u t = ( 2 ) =2 u t 2 ( 22 ) u 2t and u 2t = ( 22 ) =2 u 2t and be te long run correlation matrix of u t = (u t ; u 2t ) : Wit some algebraic manipulations, we ave = d + ( 2 ) 2 J d : 2 So te maximum eigenvalue of is given by v max ( ) = + ( 2 ) 2 =(d 2 ), wic is also te only nonzero eigenvalue. Obviously, v max ( ) is increasing in 2 for any given and d: Given and d, we coose te value of to get a di erent value of v max ( ): We consider v max ( ) = ; :9; :8; :::; :9; :99 wic are obtained by setting 2 = v max ( )( 2 ) 2 = (d( v max ( )) : 24

Should We Go One Step Further? An Accurate Comparison of One-step and Two-step Procedures in a Generalized Method of Moments Framework

Should We Go One Step Further? An Accurate Comparison of One-step and Two-step Procedures in a Generalized Method of Moments Framework Should We Go One Step Further? An Accurate Comparison of One-step and wo-step Procedures in a Generalized Method of Moments Framework Jungbin Hwang and Yixiao Sun Department of Economics, University of

More information

Simple and Powerful GMM Over-identi cation Tests with Accurate Size

Simple and Powerful GMM Over-identi cation Tests with Accurate Size Simple and Powerful GMM Over-identi cation ests wit Accurate Size Yixiao Sun and Min Seong Kim Department of Economics, University of California, San Diego is version: August, 2 Abstract e paper provides

More information

7 Semiparametric Methods and Partially Linear Regression

7 Semiparametric Methods and Partially Linear Regression 7 Semiparametric Metods and Partially Linear Regression 7. Overview A model is called semiparametric if it is described by and were is nite-dimensional (e.g. parametric) and is in nite-dimensional (nonparametric).

More information

Financial Econometrics Prof. Massimo Guidolin

Financial Econometrics Prof. Massimo Guidolin CLEFIN A.A. 2010/2011 Financial Econometrics Prof. Massimo Guidolin A Quick Review of Basic Estimation Metods 1. Were te OLS World Ends... Consider two time series 1: = { 1 2 } and 1: = { 1 2 }. At tis

More information

A New Approach to Robust Inference in Cointegration

A New Approach to Robust Inference in Cointegration A New Approach to Robust Inference in Cointegration Sainan Jin Guanghua School of Management, Peking University Peter C. B. Phillips Cowles Foundation, Yale University, University of Auckland & University

More information

CS522 - Partial Di erential Equations

CS522 - Partial Di erential Equations CS5 - Partial Di erential Equations Tibor Jánosi April 5, 5 Numerical Di erentiation In principle, di erentiation is a simple operation. Indeed, given a function speci ed as a closed-form formula, its

More information

A Flexible Nonparametric Test for Conditional Independence

A Flexible Nonparametric Test for Conditional Independence A Flexible Nonparametric Test for Conditional Independence Meng Huang Freddie Mac Yixiao Sun and Halbert Wite UC San Diego August 5, 5 Abstract Tis paper proposes a nonparametric test for conditional independence

More information

The derivative function

The derivative function Roberto s Notes on Differential Calculus Capter : Definition of derivative Section Te derivative function Wat you need to know already: f is at a point on its grap and ow to compute it. Wat te derivative

More information

Volume 29, Issue 3. Existence of competitive equilibrium in economies with multi-member households

Volume 29, Issue 3. Existence of competitive equilibrium in economies with multi-member households Volume 29, Issue 3 Existence of competitive equilibrium in economies wit multi-member ouseolds Noriisa Sato Graduate Scool of Economics, Waseda University Abstract Tis paper focuses on te existence of

More information

Stationary Gaussian Markov Processes As Limits of Stationary Autoregressive Time Series

Stationary Gaussian Markov Processes As Limits of Stationary Autoregressive Time Series Stationary Gaussian Markov Processes As Limits of Stationary Autoregressive Time Series Lawrence D. Brown, Pilip A. Ernst, Larry Sepp, and Robert Wolpert August 27, 2015 Abstract We consider te class,

More information

Differentiation in higher dimensions

Differentiation in higher dimensions Capter 2 Differentiation in iger dimensions 2.1 Te Total Derivative Recall tat if f : R R is a 1-variable function, and a R, we say tat f is differentiable at x = a if and only if te ratio f(a+) f(a) tends

More information

Accurate Asymptotic Approximation in the Optimal GMM Frame. Stochastic Volatility Models

Accurate Asymptotic Approximation in the Optimal GMM Frame. Stochastic Volatility Models Accurate Asymptotic Approximation in the Optimal GMM Framework with Application to Stochastic Volatility Models Yixiao Sun UC San Diego January 21, 2014 GMM Estimation Optimal Weighting Matrix Two-step

More information

CAE Working Paper # A New Asymptotic Theory for Heteroskedasticity-Autocorrelation Robust Tests. Nicholas M. Kiefer and Timothy J.

CAE Working Paper # A New Asymptotic Theory for Heteroskedasticity-Autocorrelation Robust Tests. Nicholas M. Kiefer and Timothy J. CAE Working Paper #05-08 A New Asymptotic Theory for Heteroskedasticity-Autocorrelation Robust Tests by Nicholas M. Kiefer and Timothy J. Vogelsang January 2005. A New Asymptotic Theory for Heteroskedasticity-Autocorrelation

More information

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley Time Series Models and Inference James L. Powell Department of Economics University of California, Berkeley Overview In contrast to the classical linear regression model, in which the components of the

More information

Optimal Bandwidth Selection in Heteroskedasticity-Autocorrelation Robust Testing

Optimal Bandwidth Selection in Heteroskedasticity-Autocorrelation Robust Testing Optimal Bandwidth Selection in Heteroskedasticity-Autocorrelation Robust Testing Yixiao Sun Department of Economics University of California, San Diego Peter C. B. Phillips Cowles Foundation, Yale University,

More information

GMM-based inference in the AR(1) panel data model for parameter values where local identi cation fails

GMM-based inference in the AR(1) panel data model for parameter values where local identi cation fails GMM-based inference in the AR() panel data model for parameter values where local identi cation fails Edith Madsen entre for Applied Microeconometrics (AM) Department of Economics, University of openhagen,

More information

Combining functions: algebraic methods

Combining functions: algebraic methods Combining functions: algebraic metods Functions can be added, subtracted, multiplied, divided, and raised to a power, just like numbers or algebra expressions. If f(x) = x 2 and g(x) = x + 2, clearly f(x)

More information

Differential Calculus (The basics) Prepared by Mr. C. Hull

Differential Calculus (The basics) Prepared by Mr. C. Hull Differential Calculus Te basics) A : Limits In tis work on limits, we will deal only wit functions i.e. tose relationsips in wic an input variable ) defines a unique output variable y). Wen we work wit

More information

1. Questions (a) through (e) refer to the graph of the function f given below. (A) 0 (B) 1 (C) 2 (D) 4 (E) does not exist

1. Questions (a) through (e) refer to the graph of the function f given below. (A) 0 (B) 1 (C) 2 (D) 4 (E) does not exist Mat 1120 Calculus Test 2. October 18, 2001 Your name Te multiple coice problems count 4 points eac. In te multiple coice section, circle te correct coice (or coices). You must sow your work on te oter

More information

Polynomial Interpolation

Polynomial Interpolation Capter 4 Polynomial Interpolation In tis capter, we consider te important problem of approximatinga function fx, wose values at a set of distinct points x, x, x,, x n are known, by a polynomial P x suc

More information

Higher Derivatives. Differentiable Functions

Higher Derivatives. Differentiable Functions Calculus 1 Lia Vas Higer Derivatives. Differentiable Functions Te second derivative. Te derivative itself can be considered as a function. Te instantaneous rate of cange of tis function is te second derivative.

More information

5.1 We will begin this section with the definition of a rational expression. We

5.1 We will begin this section with the definition of a rational expression. We Basic Properties and Reducing to Lowest Terms 5.1 We will begin tis section wit te definition of a rational epression. We will ten state te two basic properties associated wit rational epressions and go

More information

Math 2921, spring, 2004 Notes, Part 3. April 2 version, changes from March 31 version starting on page 27.. Maps and di erential equations

Math 2921, spring, 2004 Notes, Part 3. April 2 version, changes from March 31 version starting on page 27.. Maps and di erential equations Mat 9, spring, 4 Notes, Part 3. April version, canges from Marc 3 version starting on page 7.. Maps and di erential equations Horsesoe maps and di erential equations Tere are two main tecniques for detecting

More information

SECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY

SECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY (Section 3.2: Derivative Functions and Differentiability) 3.2.1 SECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY LEARNING OBJECTIVES Know, understand, and apply te Limit Definition of te Derivative

More information

CAE Working Paper # Fixed-b Asymptotic Approximation of the Sampling Behavior of Nonparametric Spectral Density Estimators

CAE Working Paper # Fixed-b Asymptotic Approximation of the Sampling Behavior of Nonparametric Spectral Density Estimators CAE Working Paper #06-04 Fixed-b Asymptotic Approximation of the Sampling Behavior of Nonparametric Spectral Density Estimators by Nigar Hashimzade and Timothy Vogelsang January 2006. Fixed-b Asymptotic

More information

REVIEW LAB ANSWER KEY

REVIEW LAB ANSWER KEY REVIEW LAB ANSWER KEY. Witout using SN, find te derivative of eac of te following (you do not need to simplify your answers): a. f x 3x 3 5x x 6 f x 3 3x 5 x 0 b. g x 4 x x x notice te trick ere! x x g

More information

Polynomial Interpolation

Polynomial Interpolation Capter 4 Polynomial Interpolation In tis capter, we consider te important problem of approximating a function f(x, wose values at a set of distinct points x, x, x 2,,x n are known, by a polynomial P (x

More information

Poisson Equation in Sobolev Spaces

Poisson Equation in Sobolev Spaces Poisson Equation in Sobolev Spaces OcMountain Dayligt Time. 6, 011 Today we discuss te Poisson equation in Sobolev spaces. It s existence, uniqueness, and regularity. Weak Solution. u = f in, u = g on

More information

4. The slope of the line 2x 7y = 8 is (a) 2/7 (b) 7/2 (c) 2 (d) 2/7 (e) None of these.

4. The slope of the line 2x 7y = 8 is (a) 2/7 (b) 7/2 (c) 2 (d) 2/7 (e) None of these. Mat 11. Test Form N Fall 016 Name. Instructions. Te first eleven problems are wort points eac. Te last six problems are wort 5 points eac. For te last six problems, you must use relevant metods of algebra

More information

Bootstrap confidence intervals in nonparametric regression without an additive model

Bootstrap confidence intervals in nonparametric regression without an additive model Bootstrap confidence intervals in nonparametric regression witout an additive model Dimitris N. Politis Abstract Te problem of confidence interval construction in nonparametric regression via te bootstrap

More information

A = h w (1) Error Analysis Physics 141

A = h w (1) Error Analysis Physics 141 Introduction In all brances of pysical science and engineering one deals constantly wit numbers wic results more or less directly from experimental observations. Experimental observations always ave inaccuracies.

More information

Bootstrap prediction intervals for Markov processes

Bootstrap prediction intervals for Markov processes arxiv: arxiv:0000.0000 Bootstrap prediction intervals for Markov processes Li Pan and Dimitris N. Politis Li Pan Department of Matematics University of California San Diego La Jolla, CA 92093-0112, USA

More information

New Distribution Theory for the Estimation of Structural Break Point in Mean

New Distribution Theory for the Estimation of Structural Break Point in Mean New Distribution Teory for te Estimation of Structural Break Point in Mean Liang Jiang Singapore Management University Xiaou Wang Te Cinese University of Hong Kong Jun Yu Singapore Management University

More information

Kernel Density Based Linear Regression Estimate

Kernel Density Based Linear Regression Estimate Kernel Density Based Linear Regression Estimate Weixin Yao and Zibiao Zao Abstract For linear regression models wit non-normally distributed errors, te least squares estimate (LSE will lose some efficiency

More information

Math 161 (33) - Final exam

Math 161 (33) - Final exam Name: Id #: Mat 161 (33) - Final exam Fall Quarter 2015 Wednesday December 9, 2015-10:30am to 12:30am Instructions: Prob. Points Score possible 1 25 2 25 3 25 4 25 TOTAL 75 (BEST 3) Read eac problem carefully.

More information

Consider a function f we ll specify which assumptions we need to make about it in a minute. Let us reformulate the integral. 1 f(x) dx.

Consider a function f we ll specify which assumptions we need to make about it in a minute. Let us reformulate the integral. 1 f(x) dx. Capter 2 Integrals as sums and derivatives as differences We now switc to te simplest metods for integrating or differentiating a function from its function samples. A careful study of Taylor expansions

More information

Generic maximum nullity of a graph

Generic maximum nullity of a graph Generic maximum nullity of a grap Leslie Hogben Bryan Sader Marc 5, 2008 Abstract For a grap G of order n, te maximum nullity of G is defined to be te largest possible nullity over all real symmetric n

More information

MAT 145. Type of Calculator Used TI-89 Titanium 100 points Score 100 possible points

MAT 145. Type of Calculator Used TI-89 Titanium 100 points Score 100 possible points MAT 15 Test #2 Name Solution Guide Type of Calculator Used TI-89 Titanium 100 points Score 100 possible points Use te grap of a function sown ere as you respond to questions 1 to 8. 1. lim f (x) 0 2. lim

More information

1watt=1W=1kg m 2 /s 3

1watt=1W=1kg m 2 /s 3 Appendix A Matematics Appendix A.1 Units To measure a pysical quantity, you need a standard. Eac pysical quantity as certain units. A unit is just a standard we use to compare, e.g. a ruler. In tis laboratory

More information

Inference about Clustering and Parametric. Assumptions in Covariance Matrix Estimation

Inference about Clustering and Parametric. Assumptions in Covariance Matrix Estimation Inference about Clustering and Parametric Assumptions in Covariance Matrix Estimation Mikko Packalen y Tony Wirjanto z 26 November 2010 Abstract Selecting an estimator for the variance covariance matrix

More information

Symmetry Labeling of Molecular Energies

Symmetry Labeling of Molecular Energies Capter 7. Symmetry Labeling of Molecular Energies Notes: Most of te material presented in tis capter is taken from Bunker and Jensen 1998, Cap. 6, and Bunker and Jensen 2005, Cap. 7. 7.1 Hamiltonian Symmetry

More information

POLYNOMIAL AND SPLINE ESTIMATORS OF THE DISTRIBUTION FUNCTION WITH PRESCRIBED ACCURACY

POLYNOMIAL AND SPLINE ESTIMATORS OF THE DISTRIBUTION FUNCTION WITH PRESCRIBED ACCURACY APPLICATIONES MATHEMATICAE 36, (29), pp. 2 Zbigniew Ciesielski (Sopot) Ryszard Zieliński (Warszawa) POLYNOMIAL AND SPLINE ESTIMATORS OF THE DISTRIBUTION FUNCTION WITH PRESCRIBED ACCURACY Abstract. Dvoretzky

More information

HOMEWORK HELP 2 FOR MATH 151

HOMEWORK HELP 2 FOR MATH 151 HOMEWORK HELP 2 FOR MATH 151 Here we go; te second round of omework elp. If tere are oters you would like to see, let me know! 2.4, 43 and 44 At wat points are te functions f(x) and g(x) = xf(x)continuous,

More information

Mass Lumping for Constant Density Acoustics

Mass Lumping for Constant Density Acoustics Lumping 1 Mass Lumping for Constant Density Acoustics William W. Symes ABSTRACT Mass lumping provides an avenue for efficient time-stepping of time-dependent problems wit conforming finite element spatial

More information

The Priestley-Chao Estimator

The Priestley-Chao Estimator Te Priestley-Cao Estimator In tis section we will consider te Pristley-Cao estimator of te unknown regression function. It is assumed tat we ave a sample of observations (Y i, x i ), i = 1,..., n wic are

More information

2.11 That s So Derivative

2.11 That s So Derivative 2.11 Tat s So Derivative Introduction to Differential Calculus Just as one defines instantaneous velocity in terms of average velocity, we now define te instantaneous rate of cange of a function at a point

More information

SECTION 1.10: DIFFERENCE QUOTIENTS LEARNING OBJECTIVES

SECTION 1.10: DIFFERENCE QUOTIENTS LEARNING OBJECTIVES (Section.0: Difference Quotients).0. SECTION.0: DIFFERENCE QUOTIENTS LEARNING OBJECTIVES Define average rate of cange (and average velocity) algebraically and grapically. Be able to identify, construct,

More information

NUMERICAL DIFFERENTIATION. James T. Smith San Francisco State University. In calculus classes, you compute derivatives algebraically: for example,

NUMERICAL DIFFERENTIATION. James T. Smith San Francisco State University. In calculus classes, you compute derivatives algebraically: for example, NUMERICAL DIFFERENTIATION James T Smit San Francisco State University In calculus classes, you compute derivatives algebraically: for example, f( x) = x + x f ( x) = x x Tis tecnique requires your knowing

More information

Chapter 1. GMM: Basic Concepts

Chapter 1. GMM: Basic Concepts Chapter 1. GMM: Basic Concepts Contents 1 Motivating Examples 1 1.1 Instrumental variable estimator....................... 1 1.2 Estimating parameters in monetary policy rules.............. 2 1.3 Estimating

More information

Robust Average Derivative Estimation. February 2007 (Preliminary and Incomplete Do not quote without permission)

Robust Average Derivative Estimation. February 2007 (Preliminary and Incomplete Do not quote without permission) Robust Average Derivative Estimation Marcia M.A. Scafgans Victoria inde-wals y February 007 (Preliminary and Incomplete Do not quote witout permission) Abstract. Many important models, suc as index models

More information

Lecture XVII. Abstract We introduce the concept of directional derivative of a scalar function and discuss its relation with the gradient operator.

Lecture XVII. Abstract We introduce the concept of directional derivative of a scalar function and discuss its relation with the gradient operator. Lecture XVII Abstract We introduce te concept of directional derivative of a scalar function and discuss its relation wit te gradient operator. Directional derivative and gradient Te directional derivative

More information

More on generalized inverses of partitioned matrices with Banachiewicz-Schur forms

More on generalized inverses of partitioned matrices with Banachiewicz-Schur forms More on generalized inverses of partitioned matrices wit anaciewicz-scur forms Yongge Tian a,, Yosio Takane b a Cina Economics and Management cademy, Central University of Finance and Economics, eijing,

More information

Gradient Descent etc.

Gradient Descent etc. 1 Gradient Descent etc EE 13: Networked estimation and control Prof Kan) I DERIVATIVE Consider f : R R x fx) Te derivative is defined as d fx) = lim dx fx + ) fx) Te cain rule states tat if d d f gx) )

More information

Copyright c 2008 Kevin Long

Copyright c 2008 Kevin Long Lecture 4 Numerical solution of initial value problems Te metods you ve learned so far ave obtained closed-form solutions to initial value problems. A closedform solution is an explicit algebriac formula

More information

Exam 1 Review Solutions

Exam 1 Review Solutions Exam Review Solutions Please also review te old quizzes, and be sure tat you understand te omework problems. General notes: () Always give an algebraic reason for your answer (graps are not sufficient),

More information

Function Composition and Chain Rules

Function Composition and Chain Rules Function Composition and s James K. Peterson Department of Biological Sciences and Department of Matematical Sciences Clemson University Marc 8, 2017 Outline 1 Function Composition and Continuity 2 Function

More information

3.1 Extreme Values of a Function

3.1 Extreme Values of a Function .1 Etreme Values of a Function Section.1 Notes Page 1 One application of te derivative is finding minimum and maimum values off a grap. In precalculus we were only able to do tis wit quadratics by find

More information

Comment on HAC Corrections for Strongly Autocorrelated Time Series by Ulrich K. Müller

Comment on HAC Corrections for Strongly Autocorrelated Time Series by Ulrich K. Müller Comment on HAC Corrections for Strongly Autocorrelated ime Series by Ulrich K. Müller Yixiao Sun Department of Economics, UC San Diego May 2, 24 On the Nearly-optimal est Müller applies the theory of optimal

More information

Differentiation: Our First View

Differentiation: Our First View Capter 6 Differentiation: Our First View We are now ready to reàect on a particular application of its of functions namely, te derivative of a function. Tis view will focus on te derivative of real-valued

More information

Preface. Here are a couple of warnings to my students who may be here to get a copy of what happened on a day that you missed.

Preface. Here are a couple of warnings to my students who may be here to get a copy of what happened on a day that you missed. Preface Here are my online notes for my course tat I teac ere at Lamar University. Despite te fact tat tese are my class notes, tey sould be accessible to anyone wanting to learn or needing a refreser

More information

Chapter 2 Limits and Continuity

Chapter 2 Limits and Continuity 4 Section. Capter Limits and Continuity Section. Rates of Cange and Limits (pp. 6) Quick Review.. f () ( ) () 4 0. f () 4( ) 4. f () sin sin 0 4. f (). 4 4 4 6. c c c 7. 8. c d d c d d c d c 9. 8 ( )(

More information

MA455 Manifolds Solutions 1 May 2008

MA455 Manifolds Solutions 1 May 2008 MA455 Manifolds Solutions 1 May 2008 1. (i) Given real numbers a < b, find a diffeomorpism (a, b) R. Solution: For example first map (a, b) to (0, π/2) and ten map (0, π/2) diffeomorpically to R using

More information

232 Calculus and Structures

232 Calculus and Structures 3 Calculus and Structures CHAPTER 17 JUSTIFICATION OF THE AREA AND SLOPE METHODS FOR EVALUATING BEAMS Calculus and Structures 33 Copyrigt Capter 17 JUSTIFICATION OF THE AREA AND SLOPE METHODS 17.1 THE

More information

ERROR BOUNDS FOR THE METHODS OF GLIMM, GODUNOV AND LEVEQUE BRADLEY J. LUCIER*

ERROR BOUNDS FOR THE METHODS OF GLIMM, GODUNOV AND LEVEQUE BRADLEY J. LUCIER* EO BOUNDS FO THE METHODS OF GLIMM, GODUNOV AND LEVEQUE BADLEY J. LUCIE* Abstract. Te expected error in L ) attimet for Glimm s sceme wen applied to a scalar conservation law is bounded by + 2 ) ) /2 T

More information

1 The concept of limits (p.217 p.229, p.242 p.249, p.255 p.256) 1.1 Limits Consider the function determined by the formula 3. x since at this point

1 The concept of limits (p.217 p.229, p.242 p.249, p.255 p.256) 1.1 Limits Consider the function determined by the formula 3. x since at this point MA00 Capter 6 Calculus and Basic Linear Algebra I Limits, Continuity and Differentiability Te concept of its (p.7 p.9, p.4 p.49, p.55 p.56). Limits Consider te function determined by te formula f Note

More information

A SHORT INTRODUCTION TO BANACH LATTICES AND

A SHORT INTRODUCTION TO BANACH LATTICES AND CHAPTER A SHORT INTRODUCTION TO BANACH LATTICES AND POSITIVE OPERATORS In tis capter we give a brief introduction to Banac lattices and positive operators. Most results of tis capter can be found, e.g.,

More information

A h u h = f h. 4.1 The CoarseGrid SystemandtheResidual Equation

A h u h = f h. 4.1 The CoarseGrid SystemandtheResidual Equation Capter Grid Transfer Remark. Contents of tis capter. Consider a grid wit grid size and te corresponding linear system of equations A u = f. Te summary given in Section 3. leads to te idea tat tere migt

More information

Regularized Regression

Regularized Regression Regularized Regression David M. Blei Columbia University December 5, 205 Modern regression problems are ig dimensional, wic means tat te number of covariates p is large. In practice statisticians regularize

More information

Continuous Stochastic Processes

Continuous Stochastic Processes Continuous Stocastic Processes Te term stocastic is often applied to penomena tat vary in time, wile te word random is reserved for penomena tat vary in space. Apart from tis distinction, te modelling

More information

Quantum Numbers and Rules

Quantum Numbers and Rules OpenStax-CNX module: m42614 1 Quantum Numbers and Rules OpenStax College Tis work is produced by OpenStax-CNX and licensed under te Creative Commons Attribution License 3.0 Abstract Dene quantum number.

More information

Bob Brown Math 251 Calculus 1 Chapter 3, Section 1 Completed 1 CCBC Dundalk

Bob Brown Math 251 Calculus 1 Chapter 3, Section 1 Completed 1 CCBC Dundalk Bob Brown Mat 251 Calculus 1 Capter 3, Section 1 Completed 1 Te Tangent Line Problem Te idea of a tangent line first arises in geometry in te context of a circle. But before we jump into a discussion of

More information

A New Diagnostic Test for Cross Section Independence in Nonparametric Panel Data Model

A New Diagnostic Test for Cross Section Independence in Nonparametric Panel Data Model e University of Adelaide Scool of Economics Researc Paper No. 2009-6 October 2009 A New Diagnostic est for Cross Section Independence in Nonparametric Panel Data Model Jia Cen, Jiti Gao and Degui Li e

More information

Introduction to Derivatives

Introduction to Derivatives Introduction to Derivatives 5-Minute Review: Instantaneous Rates and Tangent Slope Recall te analogy tat we developed earlier First we saw tat te secant slope of te line troug te two points (a, f (a))

More information

The Laplace equation, cylindrically or spherically symmetric case

The Laplace equation, cylindrically or spherically symmetric case Numerisce Metoden II, 7 4, und Übungen, 7 5 Course Notes, Summer Term 7 Some material and exercises Te Laplace equation, cylindrically or sperically symmetric case Electric and gravitational potential,

More information

lecture 26: Richardson extrapolation

lecture 26: Richardson extrapolation 43 lecture 26: Ricardson extrapolation 35 Ricardson extrapolation, Romberg integration Trougout numerical analysis, one encounters procedures tat apply some simple approximation (eg, linear interpolation)

More information

MATH745 Fall MATH745 Fall

MATH745 Fall MATH745 Fall MATH745 Fall 5 MATH745 Fall 5 INTRODUCTION WELCOME TO MATH 745 TOPICS IN NUMERICAL ANALYSIS Instructor: Dr Bartosz Protas Department of Matematics & Statistics Email: bprotas@mcmasterca Office HH 36, Ext

More information

f a h f a h h lim lim

f a h f a h h lim lim Te Derivative Te derivative of a function f at a (denoted f a) is f a if tis it exists. An alternative way of defining f a is f a x a fa fa fx fa x a Note tat te tangent line to te grap of f at te point

More information

1 Calculus. 1.1 Gradients and the Derivative. Q f(x+h) f(x)

1 Calculus. 1.1 Gradients and the Derivative. Q f(x+h) f(x) Calculus. Gradients and te Derivative Q f(x+) δy P T δx R f(x) 0 x x+ Let P (x, f(x)) and Q(x+, f(x+)) denote two points on te curve of te function y = f(x) and let R denote te point of intersection of

More information

LIMITS AND DERIVATIVES CONDITIONS FOR THE EXISTENCE OF A LIMIT

LIMITS AND DERIVATIVES CONDITIONS FOR THE EXISTENCE OF A LIMIT LIMITS AND DERIVATIVES Te limit of a function is defined as te value of y tat te curve approaces, as x approaces a particular value. Te limit of f (x) as x approaces a is written as f (x) approaces, as

More information

University Mathematics 2

University Mathematics 2 University Matematics 2 1 Differentiability In tis section, we discuss te differentiability of functions. Definition 1.1 Differentiable function). Let f) be a function. We say tat f is differentiable at

More information

Optimal Mechanism with Budget Constraint Bidders

Optimal Mechanism with Budget Constraint Bidders Optimal Mecanism wit Budget Constraint Bidders Alexei Bulatov Sergei Severinov Tis draft: November 6 Abstract Te paper deals wit te optimal mecanism design for selling to buyers wo ave commonly known budget

More information

Simple Estimators for Semiparametric Multinomial Choice Models

Simple Estimators for Semiparametric Multinomial Choice Models Simple Estimators for Semiparametric Multinomial Choice Models James L. Powell and Paul A. Ruud University of California, Berkeley March 2008 Preliminary and Incomplete Comments Welcome Abstract This paper

More information

Multiplicity, Overtaking and Convergence in the Lucas Two-Sector Growth Model by José Ramón Ruiz-Tamarit* DOCUMENTO DE TRABAJO

Multiplicity, Overtaking and Convergence in the Lucas Two-Sector Growth Model by José Ramón Ruiz-Tamarit* DOCUMENTO DE TRABAJO Multiplicity, Overtaking and Convergence in te Lucas Two-Sector Growt Model by José Ramón Ruiz-Tamarit* DOCUMENTO DE TRABAJO 22-17 July 22 * Universitat de València Los Documentos de Trabajo se distribuyen

More information

Bandwidth Selection in Nonparametric Kernel Testing

Bandwidth Selection in Nonparametric Kernel Testing Te University of Adelaide Scool of Economics Researc Paper No. 2009-0 January 2009 Bandwidt Selection in Nonparametric ernel Testing Jiti Gao and Irene Gijbels Bandwidt Selection in Nonparametric ernel

More information

Time (hours) Morphine sulfate (mg)

Time (hours) Morphine sulfate (mg) Mat Xa Fall 2002 Review Notes Limits and Definition of Derivative Important Information: 1 According to te most recent information from te Registrar, te Xa final exam will be eld from 9:15 am to 12:15

More information

Average Rate of Change

Average Rate of Change Te Derivative Tis can be tougt of as an attempt to draw a parallel (pysically and metaporically) between a line and a curve, applying te concept of slope to someting tat isn't actually straigt. Te slope

More information

GMM estimation of spatial panels

GMM estimation of spatial panels MRA Munich ersonal ReEc Archive GMM estimation of spatial panels Francesco Moscone and Elisa Tosetti Brunel University 7. April 009 Online at http://mpra.ub.uni-muenchen.de/637/ MRA aper No. 637, posted

More information

Sin, Cos and All That

Sin, Cos and All That Sin, Cos and All Tat James K. Peterson Department of Biological Sciences and Department of Matematical Sciences Clemson University Marc 9, 2017 Outline Sin, Cos and all tat! A New Power Rule Derivatives

More information

THE IMPLICIT FUNCTION THEOREM

THE IMPLICIT FUNCTION THEOREM THE IMPLICIT FUNCTION THEOREM ALEXANDRU ALEMAN 1. Motivation and statement We want to understand a general situation wic occurs in almost any area wic uses matematics. Suppose we are given number of equations

More information

3.4 Worksheet: Proof of the Chain Rule NAME

3.4 Worksheet: Proof of the Chain Rule NAME Mat 1170 3.4 Workseet: Proof of te Cain Rule NAME Te Cain Rule So far we are able to differentiate all types of functions. For example: polynomials, rational, root, and trigonometric functions. We are

More information

Derivatives of Exponentials

Derivatives of Exponentials mat 0 more on derivatives: day 0 Derivatives of Eponentials Recall tat DEFINITION... An eponential function as te form f () =a, were te base is a real number a > 0. Te domain of an eponential function

More information

Material for Difference Quotient

Material for Difference Quotient Material for Difference Quotient Prepared by Stepanie Quintal, graduate student and Marvin Stick, professor Dept. of Matematical Sciences, UMass Lowell Summer 05 Preface Te following difference quotient

More information

THE STURM-LIOUVILLE-TRANSFORMATION FOR THE SOLUTION OF VECTOR PARTIAL DIFFERENTIAL EQUATIONS. L. Trautmann, R. Rabenstein

THE STURM-LIOUVILLE-TRANSFORMATION FOR THE SOLUTION OF VECTOR PARTIAL DIFFERENTIAL EQUATIONS. L. Trautmann, R. Rabenstein Worksop on Transforms and Filter Banks (WTFB),Brandenburg, Germany, Marc 999 THE STURM-LIOUVILLE-TRANSFORMATION FOR THE SOLUTION OF VECTOR PARTIAL DIFFERENTIAL EQUATIONS L. Trautmann, R. Rabenstein Lerstul

More information

MVT and Rolle s Theorem

MVT and Rolle s Theorem AP Calculus CHAPTER 4 WORKSHEET APPLICATIONS OF DIFFERENTIATION MVT and Rolle s Teorem Name Seat # Date UNLESS INDICATED, DO NOT USE YOUR CALCULATOR FOR ANY OF THESE QUESTIONS In problems 1 and, state

More information

Quantum Mechanics Chapter 1.5: An illustration using measurements of particle spin.

Quantum Mechanics Chapter 1.5: An illustration using measurements of particle spin. I Introduction. Quantum Mecanics Capter.5: An illustration using measurements of particle spin. Quantum mecanics is a teory of pysics tat as been very successful in explaining and predicting many pysical

More information

ECONOMETRICS FIELD EXAM Michigan State University May 9, 2008

ECONOMETRICS FIELD EXAM Michigan State University May 9, 2008 ECONOMETRICS FIELD EXAM Michigan State University May 9, 2008 Instructions: Answer all four (4) questions. Point totals for each question are given in parenthesis; there are 00 points possible. Within

More information

Let s Fix It: Fixed-b Asymptotics versus Small-b Asymptotics in Heteroskedasticity and Autocorrelation Robust Inference

Let s Fix It: Fixed-b Asymptotics versus Small-b Asymptotics in Heteroskedasticity and Autocorrelation Robust Inference Let s Fix It: Fixed-b Asymtotics versus Small-b Asymtotics in Heteroskedasticity and Autocorrelation Robust Inference Yixiao Sun Deartment of Economics, University of California, San Diego June 5, 3 Abstract

More information

5 Ordinary Differential Equations: Finite Difference Methods for Boundary Problems

5 Ordinary Differential Equations: Finite Difference Methods for Boundary Problems 5 Ordinary Differential Equations: Finite Difference Metods for Boundary Problems Read sections 10.1, 10.2, 10.4 Review questions 10.1 10.4, 10.8 10.9, 10.13 5.1 Introduction In te previous capters we

More information

Chapter 1. Density Estimation

Chapter 1. Density Estimation Capter 1 Density Estimation Let X 1, X,..., X n be observations from a density f X x. Te aim is to use only tis data to obtain an estimate ˆf X x of f X x. Properties of f f X x x, Parametric metods f

More information

A.P. CALCULUS (AB) Outline Chapter 3 (Derivatives)

A.P. CALCULUS (AB) Outline Chapter 3 (Derivatives) A.P. CALCULUS (AB) Outline Capter 3 (Derivatives) NAME Date Previously in Capter 2 we determined te slope of a tangent line to a curve at a point as te limit of te slopes of secant lines using tat point

More information