POWER MAXIMIZATION AND SIZE CONTROL IN HETEROSKEDASTICITY AND AUTOCORRELATION ROBUST TESTS WITH EXPONENTIATED KERNELS

Similar documents
Optimal Bandwidth Selection in Heteroskedasticity-Autocorrelation Robust Testing

A New Approach to Robust Inference in Cointegration

POWER MAXIMIZATION AND SIZE CONTROL IN HETEROSKEDASTICITY AND AUTOCORRELATION ROBUST TESTS WITH EXPONENTIATED KERNELS

Optimal Bandwidth Selection in Heteroskedasticity-Autocorrelation Robust Testing

Comment on HAC Corrections for Strongly Autocorrelated Time Series by Ulrich K. Müller

CAE Working Paper # Fixed-b Asymptotic Approximation of the Sampling Behavior of Nonparametric Spectral Density Estimators

Testing Weak Convergence Based on HAR Covariance Matrix Estimators

CAE Working Paper # A New Asymptotic Theory for Heteroskedasticity-Autocorrelation Robust Tests. Nicholas M. Kiefer and Timothy J.

LONG RUN VARIANCE ESTIMATION AND ROBUST REGRESSION TESTING USING SHARP ORIGIN KERNELS WITH NO TRUNCATION

Inference about Clustering and Parametric. Assumptions in Covariance Matrix Estimation

Serial Correlation Robust LM Type Tests for a Shift in Trend

Fixed-b Inference for Testing Structural Change in a Time Series Regression

OPTIMAL BANDWIDTH CHOICE FOR INTERVAL ESTIMATION IN GMM REGRESSION. Yixiao Sun and Peter C.B. Phillips. May 2008

Estimation and Inference of Linear Trend Slope Ratios

GMM-based inference in the AR(1) panel data model for parameter values where local identi cation fails

LECTURE ON HAC COVARIANCE MATRIX ESTIMATION AND THE KVB APPROACH

Testing for a Trend with Persistent Errors

OPTIMAL BANDWIDTH SELECTION IN HETEROSKEDASTICITY-AUTOCORRELATION ROBUST TESTING. YIAIAO SUN, PETER C. B. PHILLIPS, and SAINAN JIN

Likelihood Ratio Based Test for the Exogeneity and the Relevance of Instrumental Variables

HAR Inference: Recommendations for Practice

Should We Go One Step Further? An Accurate Comparison of One-step and Two-step Procedures in a Generalized Method of Moments Framework

An estimate of the long-run covariance matrix, Ω, is necessary to calculate asymptotic

Robust testing of time trend and mean with unknown integration order errors

Testing for Regime Switching: A Comment

SIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER. Donald W. K. Andrews. August 2011

The Size-Power Tradeoff in HAR Inference

Problem set 1 - Solutions

Heteroskedasticity- and Autocorrelation-Robust Inference or Three Decades of HAC and HAR: What Have We Learned?

Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity

Testing in GMM Models Without Truncation

Heteroskedasticity and Autocorrelation Consistent Standard Errors

11. Bootstrap Methods

Comparing Nested Predictive Regression Models with Persistent Predictors

Panel Data. March 2, () Applied Economoetrics: Topic 6 March 2, / 43

Cointegration Tests Using Instrumental Variables Estimation and the Demand for Money in England

Weak - Convergence: Theory and Applications

13 Endogeneity and Nonparametric IV

Robust Con dence Intervals in Nonlinear Regression under Weak Identi cation

Economics 620, Lecture 19: Introduction to Nonparametric and Semiparametric Estimation

On Size and Power of Heteroskedasticity and Autocorrelation Robust Tests

Block Bootstrap HAC Robust Tests: The Sophistication of the Naive Bootstrap

1 The Multiple Regression Model: Freeing Up the Classical Assumptions

Introduction to Nonparametric and Semiparametric Estimation. Good when there are lots of data and very little prior information on functional form.

GMM estimation of spatial panels

Accurate Asymptotic Approximation in the Optimal GMM Frame. Stochastic Volatility Models

Supplemental Material for KERNEL-BASED INFERENCE IN TIME-VARYING COEFFICIENT COINTEGRATING REGRESSION. September 2017

Simple Estimators for Semiparametric Multinomial Choice Models

Powerful Trend Function Tests That are Robust to Strong Serial Correlation with an Application to the Prebish Singer Hypothesis

Simple Estimators for Monotone Index Models

Choice of Spectral Density Estimator in Ng-Perron Test: Comparative Analysis

HAC Corrections for Strongly Autocorrelated Time Series

Finite sample distributions of nonlinear IV panel unit root tests in the presence of cross-section dependence

Discussion Paper Series

Robust Unit Root and Cointegration Rank Tests for Panels and Large Systems *

Parametric Inference on Strong Dependence

When is it really justifiable to ignore explanatory variable endogeneity in a regression model?

Economics 241B Estimation with Instruments

Cointegration and the joint con rmation hypothesis

Inference on a Structural Break in Trend with Fractionally Integrated Errors

ECONOMETRICS FIELD EXAM Michigan State University May 9, 2008

Applications of Subsampling, Hybrid, and Size-Correction Methods

Estimation and Inference with Weak Identi cation

A New Asymptotic Theory for Heteroskedasticity-Autocorrelation Robust Tests

The Impact of a Hausman Pretest on the Size of a Hypothesis Test: the Panel Data Case

GMM, HAC estimators, & Standard Errors for Business Cycle Statistics

Markov-Switching Models with Endogenous Explanatory Variables. Chang-Jin Kim 1

Simple and Powerful GMM Over-identi cation Tests with Accurate Size

Tests for Cointegration, Cobreaking and Cotrending in a System of Trending Variables

SIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER. Donald W. K. Andrews. August 2011 Revised March 2012

Estimating the Number of Common Factors in Serially Dependent Approximate Factor Models

A New Approach to the Asymptotics of Heteroskedasticity-Autocorrelation Robust Testing

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria

Supplemental Material 1 for On Optimal Inference in the Linear IV Model

Department of Economics, UCSD UC San Diego

Chapter 2. Dynamic panel data models

GMM based inference for panel data models

Bootstrapping Long Memory Tests: Some Monte Carlo Results

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

Föreläsning /31

Chapter 1. GMM: Basic Concepts

THE ERROR IN REJECTION PROBABILITY OF SIMPLE AUTOCORRELATION ROBUST TESTS

7 Semiparametric Methods and Partially Linear Regression

On the Long-Run Variance Ratio Test for a Unit Root

x i = 1 yi 2 = 55 with N = 30. Use the above sample information to answer all the following questions. Show explicitly all formulas and calculations.

Estimation and Inference with Weak, Semi-strong, and Strong Identi cation

Economics 620, Lecture 13: Time Series I

Economics Division University of Southampton Southampton SO17 1BJ, UK. Title Overlapping Sub-sampling and invariance to initial conditions

ECONOMET RICS P RELIM EXAM August 24, 2010 Department of Economics, Michigan State University

Robust Standard Errors in Transformed Likelihood Estimation of Dynamic Panel Data Models with Cross-Sectional Heteroskedasticity

AN AUTOREGRESSIVE SPECTRAL DENSITY ESTIMATOR AT FREQUENCY ZERO FOR NONSTATIONARITY TESTS

Taylor series - Solutions

A Bootstrap Test for Conditional Symmetry

Inference for Parameters Defined by Moment Inequalities: A Recommended Moment Selection Procedure. Donald W.K. Andrews and Panle Jia

Testing Linear Restrictions: cont.

GMM Based Tests for Locally Misspeci ed Models

On the Power of Tests for Regime Switching

ECON2285: Mathematical Economics

ESSAYS ON TIME SERIES ECONOMETRICS. Cheol-Keun Cho

Measuring robustness

A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED

Transcription:

POWER MAXIMIZAION AND SIZE CONROL IN HEEROSKEDASICIY AND AUOCORRELAION ROBUS ESS WIH EXPONENIAED KERNELS By Yixiao Sun, Peter C. B. Phillips and Sainan Jin January COWLES FOUNDAION DISCUSSION PAPER NO. 749 COWLES FOUNDAION FOR RESEARCH IN ECONOMICS YALE UNIVERSIY Box 88 New Haven, Connecticut 65-88

Power Maximization and Size Control in Heteroskedasticity and Autocorrelation Robust ests with Exponentiated Kernels Yixiao Sun Department of Economics University of California, San Diego Peter C. B. Phillips Yale University, University of Auckland, Singapore Management University & University of Southampton Sainan Jin School of Economics Singapore Management University, August 4, 9 he authors gratefully acknowledge partial research support from NSF under Grant No. SES- 75443 (Sun), SES 6-4786 (Phillips), and NSFC under grant No. 75 and 76 (Jin).

ABSRAC Using the power kernels of Phillips, Sun and Jin (6, 7), we examine the large sample asymptotic properties of the t-test for di erent choices of power parameter (). We show that the nonstandard xed- limit distributions of the t-statistic provide more accurate approximations to the nite sample distributions than the conventional large- limit distribution. We prove that the second-order corrected critical value based on an asymptotic expansion of the nonstandard limit distribution is also secondorder correct under the large- asymptotics. As a further contribution, we propose a new practical procedure for selecting the test-imal power parameter that addresses the central concern of hypothesis testing: the selected power parameter is test-imal in the sense that it minimizes the type II error while controlling for the type I error. A plug-in procedure for implementing the test-imal power parameter is suggested. Simulations indicate that the new test is as accurate in size as the nonstandard test of Kiefer and Vogelsang (a, b; KV), and yet it does not incur the power loss that often hurts the performance of the latter test. he new test therefore combines the advantages of the KV test and the standard (MSE imal) HAC test while avoiding their main disadvantages (power loss and size distortion, respectively). he results complement recent work by Sun, Phillips and Jin (8) on conventional and b HAC testing. JEL Classi cation: C3; C4; C; C5 Keywords: Asymptotic expansion, HAC estimation, Long run variance, Loss function, Optimal smoothing parameter, Power kernel, Power maximization, Size control, ype I error, ype II error.

Introduction Seeking to robustify inference, many practical methods in econometrics now make use of heteroskedasticity and autocorrelation consistent (HAC) covariance matrix estimates. Most commonly used HAC estimates are formulated using conventional kernel smoothing techniques (for an overview, see den Haan and Levin (997)), although quite di erent approaches like wavelets (Hong and Lee ()) and direct regression methods (Phillips (5)) have recently been explored. While appealing in terms of their asymptotic properties, consistent HAC estimates provide only asymptotic robustness in econometric testing and nite sample performance is known to be unsatisfactory in many cases, but especially when there is strong autocorrelation in the data. HAC estimates are then biased downwards and the associated tests are liberal-biased. hese size distortions in testing are often substantial and have been discussed extensively in recent work (e.g., Kiefer and Vogelsang (5, hereafter KV (5)) and Sul, Phillips and Choi (5)). o address the size distortion problem, Kiefer, Vogelsang and Bunzel (, hereafter KVB), Kiefer and Vogelsang (a, b, hereafter KV) and Vogelsang (3) suggested setting the bandwidth equal to the sample size (i.e. M = ) in the construction of the long run variance (LRV) estimation. While the resulting LRV estimate is inconsistent, the associated test statistic is asymptotically nuisance parameter free and critical values can be simulated from its nonstandard asymptotic distribution. KV show by simulation that the nonstandard test has better size properties than the conventional asymptotic normal or chi-squared test. However the size improvement comes at the cost of a clear power reduction. In order to reduce the power loss while maintaining the good size properties of the KVB test, KV (5) set the bandwidth to be proportional to the sample size ( ), i.e. M = b for some b (; ]: heir approach is equivalent to contracting traditional kernels k() to get k b (x) = k(x=b) and using the contracted kernels k b () in the LRV estimation without truncation. In other work, Phillips, Sun and Jin (6, 7; PSJ) obtain a new class of kernels by exponentiating the conventional kernels using k () = (k()) ; where is a power exponent parameter. For nite b and ; both contracted and exponentiated kernels are designed to reduce, but not totally eliminate, the randomness of the denominator in the t-ratio and in doing so help to improve the power of the t-test. he parameter b in the KV approach and in the PSJ approach are smoothing parameters that play an important role in balancing size distortion and potential power gain. In other words, the smoothing parameters entail some inherent tradeo between type I and type II errors. Both KV and PSJ suggest plug-in procedures that rely on conventional asymptotic theories to select b or : More speci cally, the plug-in selection of b or suggested in these papers minimizes the asymptotic mean squared errors (AMSE) of the underlying LRV estimator. In theory, such formulations ensure that the selected values b! and! as! ; and thus the xed b or xed asymptotic theory is not applicable. However, to maintain good size properties or smaller type I errors in practical testing, KV and PSJ propose using critical values from the xed b or xed asymptotic distributions, which are nonstandard, and treating the estimated b or as xed even though they are delivered by asymptotic

formulae. PSJ justify this hybrid procedure on the basis of simulations that show the resulting tests have better size properties than tests that use standard normal asymptotic critical values. However, there are two remaining problems with this procedure. First, the rationale for good test size based on such plug-in values of is not rigorously established. Second, the AMSE-imal choice of is not necessarily imal in the context of hypothesis testing. A primary contribution of the present paper is to provide analytic solutions to both of these problems by means of asymptotic expansions that provide higher order information about the type I and type II errors. his approach provides a rigorous justi cation for the recommended procedure. We consider both rst order and second order power kernels from PSJ (6, 7)), which have been found to work very well in simulations and empirical applications (see Ray and Savin (8), Ray, Savin, and iwari (9)). o investigate the size properties of the PSJ test, we consider the Gaussian location model, which is used also in Jansson (4) and SPJ (8). Asymptotic expansions developed here reveal that the PSJ statistic is closer to its limit distribution when is xed than when increases with : More speci cally, the error in rejection probability (ERP) of the t-test based on the nonstandard limiting distribution is of order O(= ) while that based on the standard normal is O(= q=q+ ): his result relates to similar results in Jansson and SPJ who showed that the ERP of the KVB test is of order O(log = ) and O (= ), respectively, while the ERP of the conventional test using the Bartlett kernel is at most O(= p ); as shown in Velasco and Robinson (). hese ndings therefore provide theoretical support for the simulation results reported in PSJ (6, 7), Ray and Savin (8), Ray, Savin, and iwari (9). he PSJ test, which is based on the nonstandard limiting distribution, is not very convenient to use in practice as the critical values have to be simulated. o design an easy-to-implement test, we develop an expansion of the nonstandard limiting distribution about its limiting chi-squared distribution. A Cornish-Fisher type expansion then leads to second-order corrected critical values. We nd that the corrected critical values provide good approximations to the actual critical values of the nonstandard distribution. he PSJ test based on the corrected critical values has the advantage of being easily implemented and does not require the use of tables of nonstandard distributions. o show that the hybrid PSJ test using a plug-in exponent and nonstandard critical value is generally less size-distorted than the conventional test, we develop a higher order asymptotic expansion of the nite sample distribution of the t-statistic as and go to in nity simultaneously. It is shown that the corrected critical values obtained from the asymptotic expansion of the nonstandard distribution are also second order correct under the conventional limiting theory. his nding provides a theoretical explanation for the size improvement of the hybrid PSJ test compared to the conventional test. Combining the standard t-statistic and the high order corrected critical values, we obtain a new t -test whose type I and type II errors can be approximately measured

using the above asymptotic expansions. he type I and type II errors depend on the power parameter used in the HAC estimation. Following Sun (9), we propose to choose the power parameter to minimize the type II error subject to the constraint that the type I error is bounded. he bound is de ned to be ; where is the nominal type I error and > is a parameter that captures the user s tolerance on the discrepancy between the nominal and true type I errors. he parameter is allowed to be sample size dependent. For a smaller sample size, we may have a higher tolerance while for larger sample sizes we may have lower tolerance. he new procedure addresses the central concern of classical hypothesis testing, viz. maximizing power subject to controlling size. For convenience, we refer to the resulting as the test-imal : he test-imal is fundamentally di erent from the MSE-imal that applies when minimizing the asymptotic mean squared error of the corresponding HAC variance estimate (c.f., PSJ (6, 7)). When the tolerance factor is small enough, the test-imal is of smaller order than the MSE-imal. he test imal can even be O () for certain choice of the tolerance factor : he theory provides some theoretic justi cation for the use of xed rules in econometric testing. For typical economic time series and with a high tolerance for the type I error, the test-imal can be also considerably larger than the MSE-imal : Simulation results show that the new plug-in procedure suggested in the present paper works remarkably well in nite samples. In related work, SPJ (8) considered conventional kernels and selected the bandwidth to minimize a loss function formed from a weighted average of type I and type II errors. he present paper di ers from SPJ in two aspects. First, while SPJ used the contracted kernel method, an exponentiated kernel approach is used here. An earlier version of the present paper followed SPJ and employed a loss function methodology to select the power parameter, nding that for both LRV estimation and hypothesis testing, the nite sample performance of the exponentiated kernel method is similar to and sometimes better than that of the contracted kernel method. So, exponentiated kernels appear to have some natural advantages. Second, the procedure for selecting the smoothing parameter is di erent in the present paper. While SPJ selected the smoothing parameter to minimize loss based on a weighted average of type I and type II errors, we minimize the type II error after controlling for the type I error. In e ect, the loss function here is implicitly de ned with an endogenous weight given by the Lagrange multiplier while the loss function in SPJ is explicitly de ned and thus requires a user-chosen weight. his requirement can be regarded as a drawback of the explicit loss function approach, especially when it is hard to evaluate the relative importance of type I and type II errors. Furthermore, the implicit loss function approach used here is more in line with the standard econometrics testing literature where size control is often a top priority. he rest of the paper is organized as follows. Section overviews the class of power kernels that are used in the present paper and reviews some rst order limit theory for Wald type tests as! with the power parameter xed and as!. Section 3 derives an exact distribution theory using operational techniques. 3

Section 4 develops an asymptotic expansion of the nonstandard limit distribution under both the null and alternative hypotheses as the power parameter! : Section 5 develops comparable nite sample expansions of the statistic as! for a xed and as both! and!. Section 6 proposes a selection rule for that is suitable for implementation in semiparametric testing. Section 7 reports some simulation evidence on the performance of the new procedure. Section 8 concludes. Proofs and additional technical results are in the Appendix. HAR Inference for the Mean Consider the location model: y t = + u t ; t = ; ; :::; ; () with u t autocorrelated and possibly heteroskedastic, and E(u t ) = : o test a hypothesis about ; we consider the OLS estimator ^ = Y = P t= y t: Recentering and normalizing gives us p ( ^ ) = = S ; () where S t = P t = u : We impose the following convenient high level condition (e.g., KVB, PSJ (6, 7), and Jansson (4)). Assumption A he partial sum process S [ r] satis es the functional law = S [ r] )!W (r); r [; ] ; where! is the long run variance of u t and W (r) is the standard Brownian motion. hus, p ( ^ ) )!W () = N(;! ). Let ^u = y ^ be the demeaned time series and de ne the corresponding partial sum process ^S t = P t = ^u. Under A, we have = ^S[ r] )!V (r); r [; ] ; where V is a standard Brownian bridge process. When u t is stationary, the long run variance of u t is! = + P j= (j); where (j) = E(u t u t j ): Conventional approach to estimating! typically involves smoothing and truncation lag covariances using kernel-based nonparametric HAC estimators. HAC estimates of! typically have the form ^! (M) = j= + P j ( k( j M )^(j); ^(j) = t= ^u t+j ^u t for j P t= j+ ^u t+j ^u t for j < ; (3) where ^(j) are sample covariances, k() is some kernel function, M is a bandwidth parameter. Consistency of ^! (M) requires M grows with the sample size but at a slower rate so that M = o( ) (e.g. Andrews (99), Andrews and Monahan (99), Hansen (99), Newey and West (987,994), de Jong and Davidson ()). Jansson () provides a recent overview and weak conditions for consistency of such estimates. o test the null H : = against H : 6= ; the standard nonparametrically studentized t-ratio statistic is of the form t^!(m) = = ( ^ )=^!(M); (4) 4

which is asymptotically standard normal. ests based on t^!(m) and critical values from the standard normal are subject to size distortion, especially when there is strong autocorrelation in the time series. In a series of papers, KVB and KV propose the use of kernel-based estimators of! in which M is set equal to the sample size or proportional to ; taking the form M = b. hese (so called xed-b) estimates are inconsistent and tend to random quantities instead of! ; so the limit distribution of (4) is no longer standard normal. Nonetheless, use of these estimates results in valid asymptotically similar tests. In related work, PSJ (6, 7) propose the use of estimates of! based on power kernels without truncation. he power kernels were constructed by taking an arbitrary integer power of conventional kernels. In this paper, we consider the power kernels k (x) = (k BAR (x)) ; (k P R (x)) ; or (k QS (x)) where ( jxj); jxj k BAR (x) = 8 < k P R (x) = : ; jxj > 6x + 6 jxj 3 ; for jxj =; ( jxj) 3 ; for = jxj ; ; otherwise. k QS (x) = 5 sin(6x=5) x 6x=5 cos (6x=5) are the Bartlett, Parzen and Quadratic Spectral (QS) kernels respectively. kernels have a linear or quadratic expansion at the origin: hese k (x) = gx q + o (x q ) ; as x! +; (5) where g = ; q = for the Bartlett kernel, g = 6; q = for the Parzen kernel and g = 8 =5; q = for the QS kernel. For convenience, we call the exponentiated Bartlett kernels the rst order power kernels and the exponentiated Parzen and QS kernels the second order power kernels. Using k in (3) and letting M = gives HAC estimates of the form ^! = he associated t statistic is given by j= + k j ^(j): (6) t (^! ) = = ( ^ )=^! : (7) When the power parameter is xed as! ; PSJ (6, 7) show that under A, ^! )! ; where = R R k (r s)dv (r)dv (s). he associated t -statistic has the nonstandard limit distribution t (^! ) ) W () = ; (8) under the null and t (^! ) ) ( + W ()) = ; (9) 5

under the local alternative H : = + c = ; where = c=!: When goes to at a certain rate, PSJ (6, 7) further show that ^! is consistent. In this case, the t -statistic has conventional normal limits: under the null t (^! ) ) W () = d N(; ); and under the local alternative t (^! ) ) + W (): hus, the t -statistic has nonstandard limit distributions arising from the random limit of the HAC estimate ^! when is xed as! ; just as the KVB and KV tests do. However, as increases, the e ect of this randomness diminishes, and when! the limit distributions approach those of standard regression tests with consistent HAC estimates. he mechanism we develop for making improvements in size without sacri cing much power, is to use a test statistic constructed with ^! and to employ critical values that are second-order corrected. he correction is rst obtained using an accurate but simple asymptotic expansion of the nonstandard distribution about its limiting chi-squared distribution that applies as!. his expansion is developed in Section 4. he correction is further justi ed by an asymptotic expansion of the nite sample distribution in Section 5. 3 Probability Densities of the Nonstandard Limit Distribution and the Finite Sample Distribution his section develops some useful formulae for the probability densities of the xed limit theory and the exact distribution of the test statistic. First note that in the limit theory of the t-ratio test, W () is independent of, so the conditional distribution of W () = variance given is normal with zero mean and : We can write = (V) where the process V has probability measure P (V) : he pdf of t = W () = Z p t (z) = can then be written in the mixed normal form as (V)> N ; dp (V) : () For the nite sample distribution of t = t (^! ), we assume that u t is a Gaussian process. Since u t is in general autocorrelated, p ( ^ ) and ^! are statistically dependent. o nd the exact nite sample distribution of the t-statistic, we decompose ^ and ^! into statistically independent components. Let u = ( ; :::u ) ; y = (y ; :::; y ); l = (; :::; ) and = var(u): hen the GLS estimator of is ~ = l l l y and ^ = ~ + l l l ~u; () where ~u = (I l l l l )u; which is statistically independent of ~ : herefore the t-statistic can be written as t = p ( ~ ) + ^! (^u) l ~u p ^! (~u) : () 6

It is easy to see that ^u = I l (l l ) l u = I l (l l ) l ~u: In consequence, the conditional distribution of t given ~u is l N p ~u ^! (~u) ; l l! (^! (~u)) : (3) Letting P (~u) be the probability measure of ~u; we deduce that the probability density of t is Z l p t (z) = N p ~u ^! (~u) ; l l! (^! (~u)) dp (~u) ( l = E N p ~u ^! (~u) ; l l!) (^! (~u)) ; (4) which is a mean and variance mixture of normal distributions. Using ~u s N ; l l l l and employing operational techniques along the lines developed in Phillips (993), we can write expression (4) in the form " l p t (z) = N p @q ^! (@q) ; l l! Z (^! (@q)) e z ~u dp (~u)# (5) q= " l = N p @q ^! (@q) ; l l! n (^! (@q)) e q l (l l ) o # l q : his provides a general expression for the nite sample distribution of the test statistic t under Gaussianity. 4 Expansion of the Nonstandard Limit heory Asymptotic expansions of the limit distributions given in (8) and (9) can be obtained as the power parameter! : Moreover, these expansions may be developed for both the central and noncentral chi-squared limit distributions that apply when! ; corresponding to the null and alternative hypotheses. he approach aded here is to work with the expansion of the noncentral distribution and is therefore closely related to the approach used in SPJ (8) for studying standard and b type tests. Let G = G(; ) be n the cdf of a non-central o ( ) variate with noncentrality parameter ( =, then P + W ()) z = P o n( + W ()) z = E G ( z ) : An expansion of E G ( z ) can be developed in terms of the moments of where = E ( ) and = var ( ) : In particular, we have EG ( z ) = G ( z ) + G ( z )E ( ) z 4 + 6 E h G ( z ) ( ) 3 z 6i + O q= n E ( ) 4o ; (6) 7

as! ; where the O () term holds uniformly for any z [M l ; M u ] R + and M l and M u may be chosen arbitrarily small and large, respectively. It is easy to see that = R R k (r; s)dw (r)dw (s); where k(r; s) is de ned by k (r; s) = k (r s) Z k (r t)dt Z k ( s)d + Z Z k (t )dtd: Since k(r; s) is a positive semi-de nite kernel (see Sun (4) for a proof), it can be represented as k(r; s) = P i= nfn(r)f n(s); by Mercer s theorem, where n > are the eigenvalues of the kernel and fn(r) are the corresponding eigenfunctions, i.e. nfn(s) = R k (r; s)fn(r)dr: Using this representation, we can write as = P n= nzn; where Z n s iidn(; ). herefore, the characteristic function of is given by (t) = E e it( ) = e it n= f i ntg = : Let ; ; 3,... be the cumulants of : hen Z Z my =, m = m (m )! ::: @ k( j ; j+ ) A d d m ; (7) where = m+ : hese calculations enable us to develop an asymptotic expansion of E G ( z ) as the power parameter! : A full series expansion is possible using this method, but we only require the leading term in the expansion in what follows. Ignoring the technical details, we have, up to smaller order terms n o ( P + W ()) = z = G (z ) z G z Z Z k (r s)drds Z Z + z 4 G (z ) k (r s)drds : (8) R R Depending on the value of q; the integral k (r s)drds has di erent expansions as! : For rst order power kernels, direct calculation yields Z Z k (r s)drds = + O : For second order power kernels, we use the Laplace approximation and obtain Z Z k (r s)drds = j= = + O g : It is clear that the rate of decay of the integral depends on the local behavior of the kernel function at the origin. Plugging these expressions into (8), we obtain the following theorem. 8

n o ( = heorem Let F (z) := P + W ()) z distribution. hen as! ; be the nonstandard limiting where F (z) = G (z ) + =q c q L q (G ; z) + O L q (G ; z) = G (z )z 4 G (z )z I(q = ) + = c q = gi(q = ) + I(q = ); g =q n G (z )z 4 p G (z )z o I(q = ); and the remainder O () term holds uniformly for any z [M l ; M u ] with < M l < M u < : When = ; we have F (z) = D(z ) + c q =q L q (D; z) + O =q (9) where D() = G () is the CDF of distribution. For any (; ); let z ; R +, z R + ; such that F (z ; ) = ; D(z ) = : hen, using a Cornish-Fisher type expansion, we obtain the following corollary. Corollary Second order corrected critical values based on the expansion (9) are as follows: (i) For the rst order power kernel z ; = z + 4 (5z + z 3 ) + O ; () (ii) For the second order power kernel z ; = z + ( = g p! + z + 4 p ) 4 z3 + O () where z is the nominal critical value from the standard normal distribution. Consider as an example the case where = :5; z = :96 and P (W () (:96) ) = :95: hus, for a two-sided t (^! ) test, the corrected critical values at the 5% level for the Bartlett, Parzen and QS kernels are ; = :96 + 4:335 ; z; P AR = :96 + :93 p ; z QS ; = :96 + 3:95 p ; () z BAR respectively. hese are also the critical value for the one-sided test (>) at the.5% level. Similarly, the corrected critical values for = : are given by ; = :645 + :676 ; z; P AR = :645 + :375 p ; z QS ; = :645 + :85 p : (3) z BAR 9

We can evaluate the accuracy of the approximate critical values by comparing them with the exact ones obtained via simulations. In all three cases, the second order corrected critical values are remarkably close to the exact ones. Details are available upon request. Since the limiting distributions (8) and (9) are valid for general regression models under certain conditions on the regressors (see PSJ (6, 7)), the corrected critical value z ; may be used for hypothesis testing in a general regression framework. De ne X = j K (z) = e = z j = e z= j j! (j + =) j+= z j= which is positive for all z and : When 6= and the corrected critical values are used, we can establish the local asymptotic power in the following corollary. Corollary 3 he local asymptotic power satis es n o ( P + W ()) = > z ; = G (z) c q zk 4 z =q + O( =q ) (4) Corollary 3 shows that the asymptotic test power increases monotonically with when is large. Fig. graphs the function f(z ; ) = zk 4 z against di erent values of z and : For any given local alternative hypothesis, it is apparent that the function f(z ; ) is monotonically increasing in z : hus, the power increase due to the choice of a large increases with the con dence level : Further, the function obtains its maximum around = for a given critical value, implying that the power improvement from choosing a large is greatest when the local alternative is in an intermediate neighborhood of the null. 5 Expansions of the Finite Sample Distribution Following SPJ (8), we now proceed to develop an asymptotic expansion of the nite sample distribution of the t-statistic in a simple location model. We start with the following weak dependence condition. Assumption A u t is a mean zero stationary Gaussian process with P h= h j (h)j < ; where (h) = Eu t u t h : In what follows, we develop an asymptotic expansion of P f p ( ^ )=^! zg for ^! = ^! and for local alternatives of the form = + c= p : A complicating factor in the development is that p ( ^ ) and ^! are in general statistically dependent due to the autocorrelation structure of u t. o overcome this di culty, we decompose ^ and ^! into statistically independent components as in Section 3. After some manipulation, we obtain n p o F ; (z) := P ^ =^! z = E G (z & ) + O ; (5)

f.8.6.4. 4 3.5 3.5 δ.5.5.5.5 z α.75 Figure : he graph of f(z ; ) = z 4 K z as a function of z and. uniformly over z R +, where & := (^!=! ) converges weakly to, and! := p var ( ^ ). Since ^! = ^u W ^u = u A W A u; where W is with (j; s)-th element k ((j s)= ) and A = I l l =, & is a quadratic form in a Gaussian vector. o evaluate E G (z & ) ; we proceed to compute the cumulants of & for := E&. It is easy to show that the characteristic function of & is given by (t) = I it A W A =! exp f it g ; where = E(uu ) and the cumulant generating function is ln ( (t)) = log det I it A W A! it := X m= (it) m m; ; (6) m! where the m; are the cumulants of & : It follows from (6) that ; = and m; = m (m )! m! m race [( A W A ) m ] for m : (7)

By proving that m; is close to m in the precise sense given in Lemma 9 in the appendix, we can establish the following theorem, which gives the order of magnitude of the error in the nonstandard limit distribution of the t-statistic as! with xed : heorem 4 Let A hold. If R k (v)dv < = 6z ; then as! with xed : F ; (z) = F (z) + O ; (8) he requirement R k (v)dv < = 6z on is a technical condition in the proof. It can be relaxed but at the cost of more tedious calculations. F ; (z) gives the power of the test under the alternative hypothesis H : 6=. heorem 4 indicates that when is xed the power of the test can be approximated by F (z) with an error of order O(= ): Under the null hypothesis H : =, = ; heorem 4 shows that for xed the ERP for tests using critical values obtained from the nonstandard limit distribution of W () = is O : his rate is faster than the rate for conventional tests based on consistent HAC estimates. Combined with heorem, heorem 4 characterizes the size and power properties of the test under the sequential limit in which goes to in nity rst for a xed and then as goes to in nity. Under this sequential limit theory, the size distortion of the t-test based on the corrected critical values is n p o P ^ =^! z ; = O =q + O and the corresponding local asymptotic power is n p o P ^ =^! > z ; = G (z) c q zk 4 z =q +O =q +O : o evaluate the order of size distortion, we have to compare the orders of magnitude of =q and =: Such a comparison jeopardizes the sequential nature of the limiting directions and calls for a higher order approximation that allows! and! simultaneously. A corollary to this observation is that xed asymptotics do not provide an internally consistent framework for selecting the imal power parameter. he next theorem establishes the required higher order expansion. heorem 5 Let A hold. If = + = q! as! ; then F ; (z) = G (z ) + c q =q L q (G ; z) gd G (z )z q + o q + O + =q ; (9) where d =! P h= + jhjq (h).

Under the local alternative hypothesis H : j j = c= p ; the power of the test based on the corrected critical values is F ; (z ; ): heorem 5 shows that F ; (z ; ) can be approximated by G (z )+c q =q L q (G ; z) gd G (z )z ( q ) ; and the approximation error is of order o ( q ) + O + =q : Under the null hypothesis, = and G () = D(); so F ; (z) = D(z ) + c q =q L q (D; z) gd D (z )z q + o q + O + =q : (3) Importantly, the leading term D(z ) + c q =q L q (D; z) in this expansion is the same as the corresponding expansion of the limit distribution F (z) given in (9). A direct implication is that the corrected critical values from the nonstandard limiting distribution are also second order correct under the conventional large asymptotics. More speci cally, up to smaller order terms, F ; (z ; ) = gd D (z)z ( q ) : By adjusting the critical values, we eliminate the term c q q =q L (D; z), which re- ects the randomness of the standard error estimator. So if the remaining term gd D (z)z ( q ) is of smaller order than the eliminated term, the use of the corrected critical values given in Corollary should reduce the size distortion. When increases with ; the remaining term gd D (z)z ( q ) approximately measures the size distortion in tests based on the corrected critical values, or equivalently those based on the nonstandard limit theory, at least to order O( =q ). If critical values from the standard normal are used, then the ERP is given by the O =q and O( q ) terms. o obtain the best rate of convergence of the ERP to zero, we set = O( q =q+ ) to balance these two terms and the resulting rate of convergence is of order O( q=q+ ): As we discussed before, this rate is larger than that of the nonstandard test by an order of magnitude. One might argue that this comparison is not meaningful as the respective orders of the ERP are obtained under di erent asymptotic speci cations of : one for growing and the other one for xed : o sharpen the comparison, we can compare the ERP under the same asymptotic speci cation as in SPJ (8). When is xed, the ERP of the standard normal test contains an O() term, which makes the ERP larger by an order of magnitude than the ERP of the nonstandard test. When grows with ; the ERP of the nonstandard test contains fewer terms than that of the standard normal test. his is usually regarded as an advantage in the econometrics literature. If we set = O q =q+ which is the AMSE-imal rate for the exponent, then the ERP of the standard normal test is of order O ( q ) = O( q=(q+) ): In this case, the ERP of the nonstandard test or test using the second order corrected critical values is also of order O( q=(q+) ): herefore, compared with the standard normal test, the hybrid procedure suggested in PSJ (6, 7) does not reduce the ERP by an order of magnitude. However, the hybrid procedure eliminates the O( =q ) term from the ERP and thus reduces the over-rejection that is commonly seen in economic applications. his explains the better nite sample performances found in the simulation studies of PSJ (6, 7), Ray and Savin (8), and Ray, Savin, and iwari (9). 3

For convenience, we refer to the test based the t -statistic t (^! ) and the second order corrected critical values as the t -test. We formalize the results on the size distortion and local power expansion of the t -test in the following corollary. Corollary 6 Let A hold. If = + = q!! as!, then (a) the size distortion of the t -test is gd D (z)z q + o q + O + =q : (3) (b) under the local alternative, the power of the t -test is G (z ) c q z 4 K z =q + gd G (z )z q + o q + O + =q : Just as standard limit theory, the nonstandard limit theory does not address the bias problem due to nonparametric smoothing in LRV estimation. Comparing (3) with (4), we get an additional term, re ecting the asymptotic bias of the LRV estimator. For economic time series, it is typical that d > : So this additional term also increases monotonically with. Of course, size distortion tends to increase with as well. In the next section, we propose a procedure to choose that maximizes the local asymptotic power while controlling for the size. 6 Optimal Exponent Choice When estimating the long run variance, PSJ (6, 7) show there is an imal choice of which minimizes the asymptotic mean squared error of the estimator and give an imal expansion rate of O q =(q+) for in terms of the sample size : he present paper attempts to provide a new approach for imal exponent selection that addresses the central concern of classical hypothesis testing, which can be expressed as maximizing power subject to controlling size. Using (3) and ignoring the higher order terms, the type I error of the t -test can be measured by e I = + gd D (z)z q : Similarly, from (3), the type II error of the t -test can be measured by (3) e II = G (z ) + c q z 4 K z =q gd G (z )z q : We choose to minimize the type II error while controlling for the type I error. More speci cally, we solve min e II ; s:t: e I where is a constant greater than. Ideally, the type I error is less than or equal to the nominal type I error : In nite samples, there is always some approximation error and we allow for some discrepancy by introducing the tolerance factor : For example, when = 5% and = :; we aim to control the type I error such that it 4

is not greater than 6%. Note that may depend on the sample size : For a larger sample size, we may require to take smaller values. he solution to the minimization problem depends on the sign of d : When d < ; the constraint e I is not binding and we have the unconstrained minimization problem = min e II ; whose solution for the imal is = c q zk z! q q+ qgd G q q+ : (33) (z ) When d > ; the constraint e I may be binding and we have to use the Kuhn-ucker theorem to search for the imum. Let be the Lagrange multiplier, and de ne L(; ) = G (z) + c q zk 4 z =q gd G (z )z q (34) + + gd D (z )z q : It is easy to show that at the imal ; the constraint e I is indeed binding and > : Hence, the imal is = ( ) gd D (z)z q ; (35) and the corresponding Lagrange multiplier is: = G (z ) D (z) + c qk z gd D (z) q q [( ) ] +=q z =q+ : Formulae (33) and (35) can be written collectively in the form = c q zk z qgd D (z) G (z )! q q+ q q+ ; where 8 < = : ; if d < G (z ) D (z) + cqk (z)[gd D (z )] q z =q+ ; if d > q[( )] +=q (36) If we ignore the nonessential constant, then the function L(; ) is a weighted sum of the type I and type II errors with the weight given by the imal Lagrange multiplier. When d < ; the type I error is expected to be capped by the nominal type I error. As a result, the imal Lagrange multiplier is zero and we assign all weight to the type II error. his weighting scheme might be justi ed by the argument that it is worthwhile to take advantage of the extra reduction in the type II error without in ating the type I error. When d > ; the type I error is expected to larger than the nominal type I error. he constraint on the type I error is binding 5

and the Lagrange multiplier is positive. In this case, the loss function is a genuine weighted sum of type I and type II errors. As the tolerance parameter decreases toward, the weight attached to the type I error increases. When the nonparametric bias, as measured by d ; is positive, the imal grows with at the rate of q =(q+) ; which is slower than q =(q+), the MSEimal expansion rate. When the nonparametric bias is negative, which is typical for economic time series, the expansion rate of the imal depends on the rate at which approaches. When! at a rate faster than q, the Lagrange multiplier increases with the sample size at a rate faster than q : In this case, the imal is bounded. Fixed rules may then be interpreted as assigning increasingly larger weight to the type I error. his gives us a practical interpretation of xed rules in terms of the permitted tolerance of the type I error. When! at a rate slower than q q+ ; the Lagrange multiplier is bounded and the imal expands with q q+ ; at a rate faster than q =(q+) : In particular, when! at the rate of the imal expands with at the MSE-imal rate q =(q+) : So when! q at a rate faster than q+ ; the imal has a smaller order of magnitude than the MSE-imal regardless of the direction of the nonparametric bias. When = O((= qp ) q =(q+) ); the size distortion of the t -test is of order O(( ) q=(q+) ): It is apparent that the size distortion becomes smaller if a larger weight is implicitly assigned to the type I error. In particular, when is nite, the size distortion is of order O( q=(q+) ); which is larger than O ( q ), the size distortion for the case s q : It is obvious that the use of involves some tradeo between two elements in the loss function (34). Note that even when is nite, the size distortion is smaller than O( q=(q+) ); which is the size distortion for the conventional t-test using the AMSE imal ; that is when is set to be O( q =(q+) ): he formula for involves the unknown parameter d ; which could be estimated nonparametrically or by a standard plug-in procedure based on a simple model like an AR(). See Andrews (99) and Newey and West (994). Both methods achieve a valid order of magnitude and the procedure is obviously analogous to conventional data-driven methods for HAC estimation. o sum up, the test-imal that maximizes the local asymptotic power while preserving size in large samples is fundamentally di erent from the MSE-imal : he test-imal depends on the sign of the nonparametric bias and the permitted tolerance for the type I error while the MSE-imal does not. When the permitted tolerance becomes su ciently small, the test-imal is of smaller order than the MSE-imal : 7 Simulation Evidence his section presents some simulation evidence on the performance of the t -test based on the plug-in implementation of the test-imal power parameter. We consider the 6

simple location model with Gaussian ARMA(,) errors: where y t = + c= p + u t ; u t = u t + " t + " t ; " t s iidn(; ): (37) Note that the long run variance of u t is ( + ) = ( ) : On the basis of the long run variance, we consider the following c values: c = ( + ) = ( ) for [; 5]: We consider three sets of parameter con gurations for and : AR() : (; ) = (:9; ); (:6; ); (:3; ); ( :3; ); ( :6; ); ( :9; ) MA() : (; ) = (; :9); (; :6); (; :3); (; :3); (; :6); (; :9) ARMA(; ) : (; ) = ( :6; :3); (:3; :6); (:3; :3); (; ):(:6; :3); ( :3; :6) and write u t s ARMA[; ]: We consider three sample sizes =, and 5: For each data generating process, we obtain an estimate ^ of the AR coe cient by tting an AR() model to the demeaned time series. Given the estimate ^; d can be estimated by ^d = ^=( ^) when q = and ^d = ^=( ^ ) when q = : We consider the following tolerance factors = : and :: We set the signi cance level to be = % and the corresponding nominal critical value for the two sided test is z = :645: o compute the test-imal power parameter, we need to choose for the local alternative hypothesis. In principle, we can estimate by OLS but the OLS estimator is not consistent. Accordingly, we follow standard practice in the imal testing literature and propose to select such that the rst order power is 75%, that is, solves G (:645) = 75%: he solution is = :39; which lies in the intermediate range of alternative hypotheses presented in the gures below. Although this choice of may not match the true under the local alternative, Monte Carlo experiments show that it delivers a test with good size and power properties. For each choice of ; we obtain ^ and use it to construct the LRV estimate and corresponding t -statistic. We reject the null hypothesis if jt j is larger than the corrected critical values given in (3). Using, replications, we compute the empirical type I error (when = and c = ). For comparative purposes, we also compute the empirical type I error when the power parameter is the imal one that minimizes the asymptotic mean squared error of the LRV estimate. he formulae for this power parameter are given in PSJ (6, 7) and plug-in versions are ^ MSE = g p B @ 6 ^ ^ 4 C A =5 8=5 for the second order power kernels and! 3 =3 ^ MSE = 4 ( ^ ) 4 ^ =3 5 : 7

for the rst order power kernels. In addition, we include the nonstandard test proposed by KV(a) which sets to be one and uses nonstandard critical values. We refer to the three testing procedures as OP, MSE and KV respectively. ables -3 report the empirical type I error for the ARMA(,) error with sample size =, tolerance parameter = : and signi cance level = %: Several patterns emerge. First, when the empirical type I error is around or greater than the nominal type I error, the new plug-in procedure incurs a signi cantly smaller type I error than the conventional plug-in procedure. In other cases, the type I errors are more or less the same for the two plug-in procedures. Second, compared with the KV procedure, the new t -test has similar size distortion except when the error process is highly persistent and second order kernels are used. Since the bandwidth is set equal to the sample size, the KV procedure is designed to achieve the smallest possible size distortion. Given this observation, we can conclude that the new t -test succeeds in controlling for the type I error. hird, the new t -test based on the exponentiated Bartlett kernel has the smallest size distortion in an overall sense. his is true even when the error autocorrelation is very high. he above qualitative observations remain valid for other con gurations such as di erent sample sizes and di erent values of. All else being equal, the size distortion of the new t -test for = : is slightly larger than that for = :: his is expected as we have a higher tolerance for the type I error when the value of is larger. Figures -4 present nite sample power under AR() errors for the three kernels considered. We compute power using the % empirical nite sample critical values obtained from the null distribution. So the nite sample power is size-adjusted and power comparisons are meaningful. he parameter con guration is the same as those for ables -3 except the DGP is generated under local alternatives. wo observations can be drawn from these gures. First, the power of the new t -test is consistently higher than the KV test. he power di erence is larger for the second order kernels than for rst order kernels. Second, the new t -test with test-imal exponent is as powerful as the conventional MSE-based test. For a given power parameter ; using nonstandard critical values or second order corrected critical values improves the size accuracy of the test but at the cost of clear power reduction. Figures -4 show that we can employ the test-imal to compensate for the power loss. o sum up, the new t -test achieves the same degree of size accuracy as the nonstandard KV test and yet maintains the power of the conventional t-test. Rather than reporting all of the remaining gures for other con gurations, we present two representative gures. Figure 5 presents the power curves under the MA() error and with the exponentiated Bartlett kernel while Figure 6 presents the power curves under the ARMA(,) error and with the exponentiated Parzen kernel. he basic qualitative observations remain the same: the new t -test is as powerful as the standard t-test and much more powerful than the nonstandard KV test. It is important to point out the KV test considered here does not use any smoothing parameter. he power of the KV test may be improved if the bandwidth is set proportional to the sample size (KV 5). We may use the idea presented here and asymptotic expansions in SPJ (8) to develop a test imal procedure to select 8

the proportional factor. his extension is straightforward and we do not pursue it here. 8 Conclusion Pursuing the same the line of research taken in SPJ (8) to improve econometric testing where there is nonparametric studentization, the present paper employs the power kernel approach of Phillips, Sun and Jin (6, 7) and proposes a power parameter choice that maximizes the local asymptotic power while controlling for the asymptotic size of the test. his new selection criterion is fundamentally di erent from the MSE criterion for the point estimation of the long run variance. Depending on the permitted tolerance on the type I error, the expansion rate of the test-imal power parameter may be larger or smaller than the MSE-imal power parameter. he xed power parameter rule can be interpreted as exerting increasingly tight control on the type I error. Monte Carlo experiments show that the size of the new t -test is as accurate as the nonstandard KV test with the bandwidth equal to the sample size. On the other hand, unlike the KV test, which is generally less powerful than the standard t-test, the t -test based on the test-imal power parameter is as powerful as the standard t-test. We have therefore introduced a new test that combines the good elements of the two existing tests but without inheriting their main drawbacks. 9

able : Finite Sample Sizes of Di erent esting Procedures Under AR() errors ( = ; = :; = :) [; ] [.9,] [.6,] [.3,] [-.3,] [-.6,] [-.9,] Bartlett OP.399.59.44.75.583. MSE.3675.98.538.866.847.85 KV.96.59.38.893.794.46 Parzen OP.85.7.69.89.89.88 MSE.396.77.36.866.846.85 KV.36.985.955.944.98.858 QS OP.5.5.67.89.89.88 MSE.377.74.356.867.847.84 KV.8.9.99.99.955.86 Note: OP: t -test with test-imal rho; MSE: t-test with mse imal ; KV: t-test with = and nonstandard critical values from KV (). able : Finite Sample Sizes of Di erent esting Procedures Under MA() errors ( = ; = :; = :) [; ] [,.9] [,.6] [,.3] [,-.3] [,-.6] [,-.9] Bartlett OP.97.934.979.638.66. MSE.47.444.346.67.75. KV.5..995.85.477. Parzen OP.4.996.98.76.345. MSE.34.73.9.63.7. KV.954.964.95.93.84.9 QS OP.989.986.977.756.343. MSE.35.63.5.69.7. KV.995.989..96.86.7 Note: OP: t -test with test-imal rho; MSE: t-test with mse imal ; KV: t-test with = and nonstandard critical values from KV ().

able 3: Finite Sample Sizes of Di erent esting Procedures Under ARMA() errors ( = ; = :; = :) [; ] [-.6,.3] [.3,-.6] [.3,.3] [,] [.6,-.3] [-.3,.6] Bartlett OP.84.364.954.976.45.953 MSE.978.34.6.6.933.53 KV.97.673.55.955..986 Parzen OP.933.448.54.97.37.99 MSE.987.76.45.4.83.5 KV.944.85.973.95.96.947 QS OP.933.44.39.97.34.95 MSE.988.7.48.4.84. KV.983.856.7.989..994 Note: OP: t -test with test-imal rho; MSE: t-test with mse imal ; KV: t-test with = and nonstandard critical values from KV ().

power power power power power power (a) ARMA[.9,] (b) ARMA[.6,].8.8.6.6.4. mse 3 4 5 δ.4. mse 3 4 5 δ (c) ARMA[.3,] (d) ARMA[.3,].8.8.6.6.4. mse 3 4 5 δ.4. mse 3 4 5 δ (e) ARMA[.6,] (f ) ARMA[.9,].8.8.6.6.4. mse 3 4 5 δ.4. mse 3 4 5 δ Figure : Size-adjusted power for di erent testing procedures under AR() errors with exponentiated Bartlett kernel

power power power power power power (a) ARMA[.9,] (b) ARMA[.6,].8.8.6.6.4. m se 3 4 5 δ.4. m se 3 4 5 δ (c) ARMA[.3,] (d) ARMA[.3,].8.8.6.6.4. m se 3 4 5 δ.4. m se 3 4 5 δ (e) ARMA[.6,] (f) ARMA[.9,].8.8.6.6.4. m se 3 4 5 δ.4. m se 3 4 5 δ Figure 3: Size-adjusted power for di erent testing procedures under AR() errors with exponentiated Parzen kernel 3

power power power power power power (a) ARMA[.9,] (b) ARMA[.6,].8.8.6.6.4. m se 3 4 5 δ.4. m se 3 4 5 δ (c) ARMA[.3,] (d) ARMA[.3,].8.8.6.6.4. m se 3 4 5 δ.4. m se 3 4 5 δ (e) ARMA[.6,] (f) ARMA[.9,].8.8.6.6.4. m se 3 4 5 δ.4. m se 3 4 5 δ Figure 4: Size-adjusted power for di erent testing procedures under AR() errors with exponentiated QS kernel 4

power power power power power power (a) ARMA[,.9] (b) ARMA[,.6].8.8.6.6.4. m se 3 4 5 δ.4. m se 3 4 5 δ (c) ARMA[,.3] (d) ARMA[,.3].8.8.6.6.4. m se 3 4 5 δ.4. m se 3 4 5 δ (e) ARMA[,.6] (f) ARMA[,.9].8.8.6.6.4. m se 3 4 5 δ.4. m se 3 4 5 δ Figure 5: Size-adjusted power for di erent testing procedures under MA() errors with exponentiated Bartlett kernels 5

power power power power power power (a) ARMA[.6,.3] (b) ARMA[.3,.6].8.8.6.6.4. mse 3 4 5 δ.4. mse 3 4 5 δ (c) ARMA[.3,.3] (d) ARMA[,].8.8.6.6.4. mse 3 4 5 δ.4. mse 3 4 5 δ (e) ARMA[.6,.3] (f) ARMA[.3,.6].8.8.6.6.4. mse 3 4 5 δ.4. mse 3 4 5 δ Figure 6: Size-adjusted power for di erent testing procedures under ARMA(,) errors with exponentiated Parzen kernels6

9 Appendix 9. echnical Lemmas and Supplements Lemma 7 For quadratic power kernels, as! ; we have (a) R k (x) dx = O p ; (b) R = ( x)k (x) dx = g + O : Proof of Lemma 7. We prove part (b) only as part (a) follows from similar but easier arguments. For both the Parzen and QS kernels, we have for any > ; there exists := () > such that log k(x) () for x : herefore, the contribution of the interval x satis es Z ( x)k (x) dx = Z exp f log k (x) + log( exp [ ( ) ()] Z x)g dx k (x) dx exp [ ( ) ()] : (38) We now deal with the integral from to : Both the Parzen and QS kernels exhibit quadratic behavior around the origin in the sense that k (x) = gx + o x ; as x! for some g > ; which implies log k(x) = gx + o(x ): So, for any given " >, we can determine > such that log k(x) + gx "x ; for x : In consequence, Z ( x) exp (g + ")x dx Note that Z ( x) exp (g + ")x dx = In view of Z Z Z Z ( x)k (x) dx ( x) exp (g ")x dx: ( x) exp (g + ")x dx Z ( x) exp (g + ")x dx = p p (g + ") (g + ") ; ( x) exp (g + ")x : and Z ( x) exp (g + ")x Z dx exp ( )(g + ")x exp (g + ")x (j xj) dx exp ( )(g + ") Z j xj exp (g + ")x dx = O exp ( )g ; 7