A Reference Guide for Statistical Tests Useful in Finance (Preliminary)
|
|
- Prosper Townsend
- 5 years ago
- Views:
Transcription
1 A Reference Guide for Statistical Tests Useful in Finance (Preliminary) Market Pattern Research, Inc. modified from the book The Strategic Analysis of Financial Markets Volume 1: Framework Steven D. Moffitt, Ph.D. April 14, 2018 Contents 1 A Reference Guide to Statistical Tests for Financial Analysis 3 2 A Note on Nonparametric Tests 4 3 Statistical Tests for Gaussianity The Anderson-Darling Test of Normality (AD Test) The Cramer-von Mises Test (CvM) of Normality The Shapiro-Francia Test of Normality Pearson s Chi-square Goodness-of-Fit Test of Normality The Lilliefors Test of Normality The Jarque-Bera Test of Normality The D Agostino Test of Normality The Shapiro-Wilk Test of Normality Testing for Randomness Nonparametric Tests for Randomness: Runs Tests The Wald Wolfowitz Runs Test The Up and Down Test The Bartels Rank Test Nonparametric Tests for Randomness: Trend Tests The Cox-Stuart Test The Difference Sign Test The Mann Test for Trend The Dietz-Kileen Multivariate Test for Trend
2 5 Nonparametric Tests for Comparing Distributions A Comparison of Distributions Test: The two-sample test of Kolmogorov-Smirnov (KS test) Subset Location Tests: Wilcoxon Rank-Sum (k = 2) and Kruskal-Wallis (k 2) Location Tests The Ansari-Bradley Test of Equality of Variances The Fligner-Kileen Test of Scale The Mood Test of Scale Time Series Tests General Tests Parametric Tests for Random Walks A Unit Root Test: The Augmented Dickey-Fuller Test (ADF-Test)) A Unit Root Test: The Phillips-Perron Test (PP-Test) Variance-Ratio Tests The Lo-MacKinlay Test Chen-Deo Test Chow-Denning Test Wald Test Wright and Joint Wright Tests
3 This article was abstracted from material presented in a forthcoming twovolume series, The Strategic Analysis of Financial Markets by Steven D. Moffitt, Ph.D. in May, 2017 by World Scientific. Those books analyze the strategies used by investors and traders, deconstruct the associated market games in ways that uncover a commonality of structure, and apply statistics, psychology and gambling logic to identify winning (and losing) strategies. 1 A Reference Guide to Statistical Tests for Financial Analysis Though seldom acknowledged, the statistical analysis of price series is a problematic part of empirical finance. The reason is that prices arise from unknown processes that on empirical grounds alone, are unlikely to be stationary since they exhibit outliers and volatility intermittency. But probably 99% of all statistical methods for time series require stationarity. Clearly, traditional statistics should be applied to price series with caution. This paper serves the purpose of presenting methods that detect exploitable stochastic price behavior such as non-randomness, trends, the propensity to produce outliers, etc., which are more akin to data mining and pattern recognition than to traditional statistical estimation and hypothesis testing. The first Section presents tests of the null hypothesis that a distribution is univariate or multivariate Gaussian. Applying these tests to price series almost always results in rejection of Gaussianity, but the nature of that rejection can reveal things about statistical behavior and is useful for component decomposition methods like independent component analysis. The second Section presents tests of the Hypothesis of Randomness for time series, which includes nonparametric tests for runs and for monotonic (not necessarily linear) trends. The third Section presents non-parametric methods for comparing distributions, with the notably exception of bootstrap methods. In the fourth Section, various time series tests are presented: (1) for i.i.d. data, (2) for martingales and martingale difference series,(3) for serial correlation and (4) for random walks. The presentation is semi-formal; each test has its null and alternative hypotheses stated formally, accompanied by an informal discussion of the test s strengths and weaknesses. References are included for those who want additional information or proofs. A number of tests were omitted, mostly because their statement required material deemed too specialized. Four major instances of such omission are (1) monte carlo tests, (2) tests involving wavelets, (3) spectral tests for independence or martingales, and (4) random matrix methods for covariance matrices. Of these omissions, (1) is the most serious, followed closely by (2). 3
4 2 A Note on Nonparametric Tests Nonparametric or Distribution-free statistical tests are ones that are valid for large families of distributions, e.g. for all continuous distributions (ones with a density function). Many such tests use the signs of observations or their ranks in a sample. Without going into the theory of these tests, we point out some obvious properties of tests involving signs or ranks. For ranks, the data are replaced by their order in a sample. For example, the ranks for sample data x 1 = 5, x 2 = 1, x 3 = 2, and x 4 = 10 are R(x 1 ) = 3, R(x 2 ) = 1, R(x 3 ) = 2 and R(x 4 ) = 4. If a test is based on ranks, then it is invariant with respect to all strictly increasing transformations of the underlying data. Thus is one wants to test for a monotone trend in a time series, for example, rank tests are particularly appropriate. Here s an example of a nonparametric test using signs. The statistic is the number of consecutive +1 s in differences sgn(x t+1 X t ), where sgn(x) = 1 if x > 0, 0 if x = 0 and 1 if x < 0 (Bartels Test, Section 4.1.2). This test is also invariant with respect to all strictly increasing transformations of the underlying data. As another example, sgn(x i med(x)) (med(x) is the median of the sample) is invariant with respect to strictly increasing transformations. On the other hand, sgn(x i mean(x)) is not. We also note an extremely important characteristic of sign and rank tests they are robust against outliers. Thus if a few large returns in a long series are doubled, it will generally change tests based on ranks little. We prefer nonparametric tests for because they might perform better on nonstationary data and because small sample sizes are typically not a problem. On the other hand, parametric tests have two things that can go wrong: (1) the chosen parametric family doesn t fit the data, e.g. the data aren t Gaussian but Gaussian models are assumed, and (2) the observations or their transformations don t form a random sample. Nonparametric tests suffer only from the second problem, which is the one of primary interest in trading systems development. 3 Statistical Tests for Gaussianity Gaussian distributions occupy a central place in financial theory for several reasons. They are limiting distributions for random samples of distributions with finite variances, they are minimum information distributions among those that have finite variances, their affine transformations are again Gaussian, and they are infinitely divisible, meaning that for any Gaussian random variable X and any positive integer n, there is a (Gaussian) distribution Y whose n-fold convolution has the same distribution as X. Certainly, financial time series would be much simpler if all distributions were Gaussian, but the reality is far different. In general, tests based on Gaussian distributions should be avoided in trading system analysis, a notable exception being independent component analysis. Eight tests of the null hypotheses that a sample or time series is distributed 4
5 as i.i.d. Gaussians are presented in this Section. They are N.1: The Anderson-Darling Test of Normality (AD Test) N.2: The Cramer-von Mises Test (CvM) of Normality N.3: The Shapiro-Francia Test of Normality N.4: The Pearson Chi-square Goodness-of-Fit Test of Normality N.5: The Lilliefors Test of Normality N.6: The Jarque-Bera Test of Normality N.7: The D Agostino Test of Normality N.8: The Shapiro-Wilk Test of Normality 3.1 The Anderson-Darling Test of Normality (AD Test) Let F be an unknown continuous distribution function for which the mean µ and variance σ 2 exist, and let Φ(x; µ, σ 2 ) be the cumulative normal distribution with mean µ and variance σ 2. Null Hypothesis: H 0 : F = Φ(µ, σ 2 ). Alt. Hypothesis: H a : F Φ(µ, σ 2 ). Let X 1, X 2,... X n be a random sample from F and let ˆF be its empirical distribution function ˆF (x) = 1 n I(X i x), n i=1 where I(c) = 1 if c is true and I(c) = 0 is c is false. Then the Anderson-Darling test statistic is ( n ˆF (x) Φ(x; µ, σ 2 )) 2 [Φ(x; µ, σ 2 )][1 Φ(x; µ, σ 2 )] dφ(x; µ, σ2 ). (1) A formula can be developed from (1), as follows. Assuming that the mean and variance of the X i are unknown, standardize the X i s, Z i = (X i X)/S X, where X is the sample mean and S X is the sample standard deviation, and order the Z i s from lower to highest, calling the results Y i : Define A 2 as A 2 = n 1 n Y 1 < Y 2 < < Y n. n [(2i 1)ln(Φ(Y i )) (2(n i) + 1)ln(1 Φ(Y i ))], i=1 and let A 2 = A 2 (1 + 4/n 25/n 2 ) (2) 5
6 be a correction to A 2 for small samples (see Shorack & Wellner, [Shorack and Wellner, 2009]). The theoretical asymptotic distribution 1 of A 2 has been tabulated, and the null hypothesis of normality will be rejected if A 2 exceeds the tabulated value. Note that the test statistic (1) depends only on the empirical distribution function (EDF) 2, as do several other test criteria, and therefore, are nonparametric tests. Other tests in the EDF family are the Cramer-von Mises test and the Lilliefors test. The rationale for (1) is as follows. If U i is a random sample from N(0, 1), then Φ 1 (U i ) will be a random sample from the uniform distribution on (0, 1). Thus by standardizing the X i to Y i, the distribution of Φ 1 (Y i ) will asymptotically also have a uniform distribution under the null hypothesis. One can then perform a theoretical calculation of the distribution of (1) assuming that Φ is replaced by the cumulative distribution of an uniform distribution and ˆF is the empirical cumulative distribution for an uniform distribution. Simplified Explanation: Note that test statistic (1) will be large if there are too many outliers relative to a normal distribution, since the denominator [Φ(x)][1 Φ(x)] approaches zero quickly beyond x = ±3. Thus the AD is an excellent choice if there are outliers in the data, as there usually are in financial returns series. But distribution is asymptotic, that is, holds exactly only in the limit as n. Reference: Anderson & Darling [Anderson and Darling, 1952, Anderson and Darling, 1954], Shorack & Wellner [Shorack and Wellner, 2009] and Stephens [Stephens, 1974]. 3.2 The Cramer-von Mises Test (CvM) of Normality Let F be an unknown continuous distribution function for which the mean µ and variance σ 2 exist, and let Φ(µ, σ 2 ) be the cumulative normal distribution with mean µ and variance σ 2. Null Hypothesis: H 0 : F = Φ(µ, σ 2 ). Alt. Hypothesis: H a : F Φ(µ, σ 2 ). Let X 1, X 2,... X n be a random sample from F and let ˆF be its empirical distribution function ˆF (x) = 1 i=n I(X i x), n i=1 where I(c) = 1 if c is true and I(c) = 0 is c is false. Then the Cramer-von Mises test statistic is n 1 That is, the distribution as n. 2 And implicity, the normal distribution ( ˆF (x) Φ(x; µ, σ 2 )) 2 dφ(x; µ, σ 2 ). (3) 6
7 A formula can be developed from (3), as follows, assuming that the mean and variance of the X i are unknown. Standardize the X i s, Z i = (X i X)/s X, where X is the sample mean and s X is the sample standard deviation and order the Z i s from least to greatest. Define W 2 as W 2 = 1 n 12n + [ ] 2 2i 1 2n Φ(Z i), (4) i=1 The theoretical distribution of W 2 has been tabulated, and the null hypothesis of normality will be rejected if W 2 exceeds the tabulated value. Note that the test statistic (3) depends only on the empirical distribution function (EDF) 3, as do Anderson-Darling test and the Lilliefors test. The rationale for (3) is as follows. If U i is a random sample from N(0, 1), then Φ 1 (U i ) will be a random sample from the uniform distribution on (0, 1). Thus by standardizing and ordering the X i, the distribution of Φ 1 (Z i ) will asymptotically also have a uniform distribution under the null hypothesis. One can then perform a theoretical calculation of the distribution of (3) assuming that Φ is replaced by the cumulative distribution of an uniform distribution and ˆF is the empirical cumulative distribution for an uniform distribution. Simplified Explanation: The test statistic (3) will be large if differences from a normal in the center or in the extremes receive enough weight. Thus the CvM is an excellent choice relative to the AD test for testing differences in the center of the distribution. Reference: Darling [Darling, 1957]. 3.3 The Shapiro-Francia Test of Normality Let F be an unknown continuous distribution function for which the mean µ and variance σ 2 exist, and let Φ(µ, σ 2 ) be the cumulative normal distribution with mean µ and variance σ 2. Null Hypothesis: H 0 : F = Φ(µ, σ 2 ). Alt. Hypothesis: H a : F Φ(µ, σ 2 ). The Shapiro-Francia test statistic for random sample {X i } n 1 uses the order statistics {X (i) } n 1 to form a test for normality. Order statistics for a sample have the same values as {X i } n 1, but ordered from lowest to highest. 4 Thus X (1) is the smallest value in {X i } n 1, X (2) is the next smallest, X (n) being the largest. 3 And implicity, the normal distribution 4 The ordering of order statistics are unique only for continuous distributions, since the probability of ties is zero. 7
8 With this notation, the Shapiro-Wilk test statistic is W = ( n i=1 b ix (i) ) 2 n i=1 (X i X) 2 (5) where X is the sample mean, with m (b 1, b 2,..., b n ) = (m m) 1/2 m = ( E[X (1) ], E[X (2) ],..., E[X (n) ] ). Since a closed form of the W -test s distribution is unknown, the percentiles have been estimated using Monte Carlo methods. Simplified Explanation: The Shapiro-Francia test compares the order statistics from a random sample to those expected in a normal distribution having the same mean as the sample distribution, but with an assumption of independence. The test is almost the same as the Shapiro-Wilk W test; it differs only in the use of order statistic weights b i instead of a i (see the Shapiro-Wilk description in this Section.) Compared with the Shapiro-Wilk, it is more sensitive to alternatives that are continuous and symmetric with high kurtosis, to ones that are near normal and to ones that are discrete and skewed. But it is less sensitive than W on alternatives that are continuous and skewed with high kurtosis, and on ones that are discrete and symmetric. Since market returns in general are right-skewed and have high kurtosis, the Shapiro-Wilk should be preferred over the Shapiro-Francia on this data. Reference: Shapiro & Francia (1972) [Shapiro and Francia, 1972] and Sarkadi (1975) [Sarkadi, 1975]. 3.4 Pearson s Chi-square Goodness-of-Fit Test of Normality Let F be an unknown continuous distribution function for which the mean µ and variance σ 2 exist, and let Φ(µ, σ 2 ) be the cumulative normal distribution with mean µ and variance σ 2. Null Hypothesis: H 0 : F = Φ(µ, σ 2 ). Alt. Hypothesis: H a : F Φ(µ, σ 2 ). The Pearson χ 2 test is a mainstay of beginning statistics; for a sample that is binned into K classes, it has the form χ 2 = i=k i=1 (O i E i ) 2 E i (6) 8
9 where the i th term of the sum applies to all observations in the i th bin, O i is the observed count in bin i and E i is the expected count (generally non-integer) for bin i. Under conditions of the null hypothesis, the statistic will have an asymptotic χ 2 distribution with K 2 degrees of freedom. Simplified Explanation: The χ 2 assumes that data from a sample are put into discrete bins, each of which has a positive probability of occurrence. It compares using statistic 6 the observed counts in the bins with those expected. The resulting test has K p degrees of freedom, where p is the number of parameters that must be estimated to calculate expected values. For the normal, p = 2, since a mean and variance must be calculated to determine bin probabilities. Reference: Spiegel (2000) [Spiegel et al., 2000]. 3.5 The Lilliefors Test of Normality Let F be an unknown continuous distribution function for which the mean µ and variance σ 2 exist, and let Φ(µ, σ 2 ) be the cumulative normal distribution with mean µ and variance σ 2. Null Hypothesis: H 0 : F = Φ(µ, σ 2 ). Alt. Hypothesis: H a : F Φ(µ, σ 2 ). For a random sample {X i } n 1 having sample mean X and sample standard deviation s X, the Lilliefors test uses the statistic L = sup ˆF (x) Φ(x; X, s 2 x), (7) x where ˆF is the empirical distribution of the random sample and Φ(x; µ, σ 2 ) is the cumulative distribution of a Gaussian distribution with mean µ and variance σ 2, X. This statistic is the same as the Kolmogorov-Smirnov one-sample test (KS test) except that the mean and variance are estimated in the L-test but assumed known in the KS-test. This parameter estimation changes the distribution of L as compared to KS. 5 At this writing, the asymptotic L distribution has been tabulated only using Monte Carlo methods. Simplified Explanation: The Lilliefors statistic measures the maximum absolute difference between the empirical distribution function and a Gaussian having the same mean and standard deviation. The test will be quite sensitive to deviations from normality, but will be inferior to the Anderson-Darling test in the presence of outliers. Reference: Lilliefors [Lilliefors, 1967]. 5 An asymptotic closed form distribution is known for the one-sample KS-test. 9
10 3.6 The Jarque-Bera Test of Normality Let F be an unknown continuous distribution function for which the mean µ and variance σ 2 exist, and let Φ(µ, σ 2 ) be the cumulative normal distribution with mean µ and variance σ 2. Null Hypothesis: H 0 : F = Φ(µ, σ 2 ). Alt. Hypothesis: H a : F Φ(µ, σ 2 ). The Jarque-Bera tests whether a distribution s skewness and kurtosis match those of a Gaussian distribution. For a random sample {X i } n 1 having sample mean X JB = n (U ) 6 (V 3), (8) where and 1 n 4 i=1 U = (X i X) 3 ( 1 n 4 i=1 (X i X) 2) 3/2 1 n 4 i=1 V = (X i X) 4 ( 1 n 4 i=1 (X i X) 2) 2. Under the null hypothesis of normality, U = 0 and V = 3. For large samples, n 200, the distribution of JB is approximately χ 2 with 2 degrees of freedom. Simplified Explanation: The terms U and V are, respectively, estimates of the sample s skew and kurtosis. For a normal distribution, the kurtosis equals 3, thus the appearance of that term in JB. The weights given U and V in the expression defining JB ensure that it is asymptotically χ 2 with 2 d.f. under the null hyothesis. I would not recommend using this test because it is not powerful against non-normal distributions with skew and kurtotis that match a normal. Reference: Jarque & Bera [Jarque and Bera, 1987]. 3.7 The D Agostino Test of Normality Let F be an unknown continuous distribution function for which the mean µ and variance σ 2 exist, and let Φ(µ, σ 2 ) be the cumulative normal distribution with mean µ and variance σ 2. Null Hypothesis: H 0 : F = Φ(µ, σ 2 ). Alt. Hypothesis: H a : F Φ(µ, σ 2 ). The D Agostino K 2 -test statistic for a random sample {X i } n 1 uses Cornish- Fisher expansions to derive an approximate test based on the first four moments of the sample distribution. The test statistic has the form K 2 = Z 1 (g 1 ) 2 + Z 2 (g 2 ) 2 (9) 10
11 where Z 1 (g 1 ) 2 and Z 2 (g 2 ) 2 have lengthy algebraic expressions. See the references for definitions. Like the Jarque-Bera, K 2 tests for skew and kurtosis that match a normal. Simplified Explanation: The K 2 test uses a Cornish-Fisher expansion to approximate a distribution based on the first four moments of a normal. Like the Jarque-Bera test, it is not recommended because any non-normal distribution that matches the first four moments of a normal will not be rejected. Reference: D Agostino [D Agostino, 1971]. 3.8 The Shapiro-Wilk Test of Normality Let F be an unknown continuous distribution function for which the mean µ and variance σ 2 exist, and let Φ(µ, σ 2 ) be the cumulative normal distribution with mean µ and variance σ 2. Null Hypothesis: H 0 : F = Φ(µ, σ 2 ). Alt. Hypothesis: H a : F Φ(µ, σ 2 ). The Shapiro-Wilk test statistic for random sample {X i } n 1 uses the order statistics {X (i) } n 1 to form a test for normality. Order statistics for a sample have the same values as {X i } n 1, but ordered from lowest to highest. 6 Thus X (1) is the smallest value in {X i } n 1, X (2) is the next smallest, X (n) being the largest. With this notation, the Shapiro-Wilk test statistic is W = ( n i=1 a ix (i) ) 2 n i=1 (X i X) 2 (10) where X is the sample mean, with and m V 1 (a 1, a 2,..., a n ) = (m V 2 m) 1/2 m = ( E[X (1) ], E[X (2) ],..., E[X (n) ] ). u 11 u u 1n u 21 u u 2n V =..... u1n u n1 u n2... u nn where u ij = cov(x (i), X (j) ). Since a closed form of the W -test s distribution is unknown, the percentiles have been estimated using Monte Carlo methods. 6 The ordering of order statistics are unique only for continuous distributions since the probability of ties is zero. 11
12 Simplified Explanation: The Shapiro-Wilk test compares the order statistics from a random sample to those expected in a normal distribution having the same mean and variance as the sample distribution. The test may be interpreted as the Pearson correlation coefficient between the ordered observations and the weights a i used in the numerator, so that the test in effect measures the straightness of the line in a normal Q-Q plot. Among the several tests of normality presented above, the Shapiro-Wilk ranks favorably with the Anderson-Darling test and is generally better than the others. Reference: Shapiro & Wilk [Shapiro and Wilk, 1965] and Royston [Royston, 1982, Royston, 1995]. 4 Testing for Randomness The difference between statistics for random samples and statistics for time series is the assumption in the latter of dependence in the variables. In time series models such as ARMA, ARCH and GARCH, decompositions that have a deterministic part and a white noise part are produced (assuming that those models fit, of course). Trading system development has the similar objective of transforming time series of stochastically dependent prices into time series of uncorrelated trades. Therefore it is important to have statistical tools that can test for randomness, that is, which test that a time series of trades behaves like a random sample with positive expected value. An hypothesis that goes a long way toward testing for random sample behavior is the Hypothesis of Randomness. Let X 1, X 2,..., X T,... be the variables of a time series and let P T be the set of all permutations on integers 1,..., T. The Hypothesis of Randomness is met if all permutations of variables in the series have the same distribution: H 0 :(X 1, X 2,..., X T ) D = (X σ(1), X σ(2),..., X σ(t ) ) (11) for any permutation σ P T. where the = D notation means having the same distribution. In the following, we use the notation sgn(x) for +1 x > 0 sgn(x) = 0 x = 0 1 x < Nonparametric Tests for Randomness: Runs Tests In this Section, we present three nonparametric runs tests of the Hypothesis of Randomness for continuous data. Runs of this section are of two types: (1) runs against the median, and (2) ascending/descending runs. Runs against the median are maximal numbers of consecutive observations that are above or below the median. For example, with 4 as the median of the 10 values 12
13 1, 2.3, 3, 3.5, 8, 9, 1, 1.5, 8, 2.7, there are 5 runs below and above the median of lengths 4,2,2,1,1. We note that a test statistic counting the number of runs above or below the median does not depend on the underlying (continuous) distribution and is therefore distribution-free or non-parametric. If there are too many or too few runs against the median, then the series is clearly not uncorrelated. On the other hand, ascending (descending) runs are consecutive increasing (decreasing) values. Under the Hypothesis of Randomness a testic consisting of the number of increasing runs is distribution-free and has a calculable distribution. Too many or too few ascending runs indicates dependence among variables. In the above data, there were ascending runs (using positive or negative differences) of lengths 5 and 2 and descend runs of lengths 1 and The Wald Wolfowitz Runs Test {X t } T 1 are continuous random variables with joint density f(x 1, X 2,..., X T ). Null Hypothesis: H 0 : The Hypothesis of Randomness. Alt. Hypothesis: H a : Series {X t } has either too many or too few runs (1-sided) or either (two-sided.) The following procedure is used to calculate the test statistic: 1. The median or some other cutoff, m x, of the sample x 1, x 2,...x T is calculated, 2. A new sample y 1, y 2,..., y T is formed, where y i = +1, if x i m x and y i = 1, if x i < m x, 3. The number of runs u in the sequence of y s is counted, where a run is a maximal consecutive number of +1 s or 1 s. 4. The value u is then compared to a statistical table and a significance level assigned. It is also possible to conduct runs tests with percentiles other than the median; see the references for details. Simplified Explanation: The runs test examines a series to see if there are too many or too few consecutive observations above the cutoff and below the cutoff, i.e., there are too many or too few runs. If a series is slowly varying and stays above the cutoff for long stretches, or is rapidly oscillating above and below the cutoff, then the null hypothesis will be rejected. Reference: Gibbons & Chakraborti [Gibbons and Chakraborti, 2010], Siegel & Castellan [Siegel and Castellan, 1998] The Up and Down Test {X t } T 1 are continuous random variables with joint density f(x 1, X 2,..., X T ). Null Hypothesis: H 0 : The Hypothesis of Randomness. Alt. Hypothesis: H a : Series {sgn(x t+1 X t )} has either too many or too few runs of +1 s. 13
14 The test statistic is the number of runs of consecutive +1 s of the signed differences Y t = sgn(x t+1 X t ), t = 1, 2,..., T 1. Under the Hypothesis of Randomness, one can calculate finite sample distributions, as in Gibbons & Chakraborti [Gibbons and Chakraborti, 2010]. Letting m be the number of non-zero Y t, it can be shown that the test statistic is asymptotically normal with mean (2m 1)/3 and variance (16m 29)/90. Simplified Explanation: The Up and Down Test determines if there are too many or too few consecutive ascending runs in the data. It is not a test for global trend, though, since a sawtooth pattern that goes nowhere has no long-term trend. Therefore the Up and Down Test rejects the Hypothesis of Randomness, but not a global test of runs for the entire series. In a long series, the Up and Down Test detects local trends. Reference: Gibbons & Chakraborti [Gibbons and Chakraborti, 2010] The Bartels Rank Test {X t } T 1 are continuous random variables with joint density f(x 1, X 2,..., X T ). Null Hypothesis: H 0 : The Hypothesis of Randomness Alt. Hypothesis: H a : The sequential variance of ranks is not equal to the usual variance of ranks. The test statistic is based on a parametric test originally suggested by John von Neumann, but substitutes ranks for observations. The Bartels statistic U is U = (R t R t+1 ) 2 / (R i (t + 1)/2) 2 where R t = rank(x t ), t = 1,..., T. It is known that the distribution of (U 2)/σ, where σ 2 = [4(T 2)(5T 2 2T 9)]/[5T (T + 1)(T 1) 2 ]. is asyptotically normal N(0, 1) under the Hypothesis of Randomness. If for example, the data are positively serially correlated, then U will be small and if negatively correlated, large compared to the average random arrangement. Simplified Explanation: The Bartels test is based on a test originally suggested by von Neumann, in which the ratio of a variance determinded by sequential differences is compared to the variance formed from deviations from the mean (ignoring a norming constant). In the von Neumann test, the asymptotic sampling distribution for a normal distribution can be calculated. The Bartels test substitutes ranks for the actual observations, and forms the ratio of the variance calculated from consecutive ranks R t R t+1 to that based on deviations from the mean R t (T + 1)/2. The reference below shows that the rank version has an asymptotically normal distribution, with mean and variance indicated above. In simulations run by Bartels against serially correlated alternatives, this test was more sensitive that the runs up and down test. Reference: Bartels [Bartels, 1982]. 14
15 4.2 Nonparametric Tests for Randomness: Trend Tests Another important deviation from randomness is a global trend. In elementary theory of time series, trends are treated simply and naïvely, as either random walks with drift or trend stationary processes. In either case there is a constant push up or down augmented by noise. But what if one wants to test for trend without restricted models of these types, as can occur when the drift is not constant or the process is nonlinear? This Section presents several nonparametric tests that can be used to detect such trends The Cox-Stuart Test Null Hypothesis: Alt. Hypothesis: Let f(x 1,..., x T ) be unknown density function for (X 1,..., X T ). H 0 : The Hypothesis of Randomness H a : There is either an excess number of positive (or negative) sequential values in the first half compared to the second. Data is split into two halves with the middle observation dropped if sample size is odd, giving equal numbers m in each half. Calling X t and Y t, t = 1, 2,..., m the respective observations in the two halves, the test statistic is t=m t=1 sgn(y t X t ). Under the null hypothesis, this statistic (known as a sign test) has a binomial distribution B(m, 1/2), i.e. with p = 0.5. When the value significantly exceeds m/2, there is evidence of an upward trend, and when it is significantly below, evidence of a downward trend. Simplified Explanation: The Cox-Stuart Test is a simple but not very powerful test. It compares the ranks of two halves of a sample using only the signs of pairwise equential ifferences. It s easy to calculate, but not appropriate if one wants to determine the degree of monotonicity in the data. Reference: Cox & Stuart [D. R. Cox, 1955] The Difference Sign Test Null Hypothesis: Let f(x 1,..., x T ) be unknown density function for(x 1,..., X T ). H 0 : The Hypothesis of Randomness Null Hypothesis: H 0 : (X 1,..., X T ) = D (X σ(1),..., X σ(t ) ), for all σ P T. Alt. Hypothesis: H aa : P [X t > X t 1 ] > 1/2 for all t or H ad : P [X t > X t 1 ] < 1/2 for all t. 15
16 This test counts the number of positive differences among sample data and performs a sign test. Under the null hypothesis, the number of positive differences X t X t 1 will be distributed as a binomial B(T 1, 1/2). Simplified Explanation: The test statistic is the number of positive differences which has under the null hypothesis a binomial distribution B(T 1, 1/2). To be effective, it needs a large sample. Reference: Moore & Wallis [Geoffrey H. Moore, 1943] The Mann Test for Trend The data have joint density f(x 1, x 2,..., x T ). Null Hypothesis: H 0 : The Hypothesis of Randomness. Alt. Hypothesis: H a : There is a monotone trend, P [X t2 > X t1 ] for t 2 > t 1. The original Mann test is S = t=t 1 t=1 s=t s=t+1 sgn(x s X t ). Under the null Hypothesis of Randomness, E[S] = 0 and the variance is p=g V ar(s) = [T (T 1)(2T + 5) t p (t p 1)(2t p + 5)]/18, where the number of groups of observations that are tied is g and the number of ties in each group p is t p. Asymptotically, Z = p=1 S V ar(s) has a N(0, 1) distribution. In small samples of T < 30 the test S 1 S > 0 V ar(s) Z = 0 S = 0 S+1 S < 0 V ar(s) which applies a continuity correction, is recommended. Simplified Explanation: When the Mann Test is positive and large, there is evidence of an upward trend; when highly negative, a downward trend. The two previous tests, the Cox-Stuart and the Difference Sign Test as well as this one all tests for monotone, not necessarily linear, trends. Of these three tests, the Mann is definitely the best. Moreover, as we see below in the Dietz- Killeen multivariate generalization of the Mann Test, short term correlations can be partially removed, allowing the dectection of an underlying trend not due to those correlations. Reference: Hirsh & Slack [Hirsh and Slack, 1984], Kendall [Kendall, 1975], Mann [Mann, 1945]. 16
17 4.2.4 The Dietz-Kileen Multivariate Test for Trend Null Hypothesis: Alt. Hypothesis: The p-variate time series X t = (X t1, X t2,..., X tp has unknown joint density f(x 1, x 2,..., x p ). H 0 : The Hypothesis of Randomness H a : There is a monotone trend in one or more of the p variables. The test statistic is complicated to describe; the interested reader is referred to the cited references. Under H 0 and full rank of a covariance matrix, the test statistic will be distributed as a χ 2 (p) distribution. Simplified Explanation: When the Test is positive and large, one rejects the hypothesis of no trend in any variable. Directionality is not measured some variables can have ascending trends, some descending and rejection will occur anyway. This test is useful in trading systems testing, as follows. One forms a multivariate series by taking every two or three trades and calling that vector an observation. Then perform the Dietz-Kileen procedure tests for monotone trends adjusted for serial or three-fold correlation. Under tests of this method by Hirsh and Slack [Hirsh and Slack, 1984], it proved quite effective if the serial correlation does not exceed 0.6. Thus the Dietz-Kileen test can address a common question about trading systems is serial correlation the cause of a trend or not? Reference: Dietz & Kileen [E. Jacquelin Dietz, 1981], [Hirsh et al., 1982], Mann [Mann, 1945]. 5 Nonparametric Tests for Comparing Distributions Because returns distributions have been shown to have systematic departures from normality, it is more appropriate to test i.i.d., white noise and random walk hypotheses using nonparametric or semiparametric methods. Below are several of the more common tests of this type. 5.1 A Comparison of Distributions Test: The two-sample test of Kolmogorov-Smirnov (KS test) One univariate random sample from continous distribution F 1 (x), the other from continuous distribution F 2 (x). Null Hypothesis: H 0 : F 1 (x) = F 2 (x) for all x R. Alt. Hypothesis: H a : F 1 (x) F 2 (x) > 0 for some x R. The test statistic uses the empirical distribution functions for the two samples. An empirical distribution ˆF (x) for a univariate sample {X i } i=1 i=n is defined 17
18 as ˆF (x) = # of X is in the sample that are x. (12) n The KS test statistic is just the largest absolute difference between the empirical distribution functions. KS = max x R ˆF 1 (x) ˆF 2 (x) (13) The strength of the KS test is that its asymptotic distribution does not depend on the hypothesized common distribution function. We omit discussion of the Cramer-von Mises [Anderson, 1962] and Anderson-Darling [Scholz and Stephens, 1987] tests, which are similar the KS test, but have power against different alternatives. Simplified Explanation: This test detects a worst case difference between two distribution functions. When sample sizes are large, it is quite sensitive to small differences in two distributions. Like the Kruskal-Wallis test and other tests that compare distributions, however, it is questionable for time series because the null hypothesis ignores the sample s original order. Reference: Hollander & Wolfe [Hollander and Wolfe, 1999]. 5.2 Subset Location Tests: Wilcoxon Rank-Sum (k = 2) and Kruskal-Wallis (k 2) Location Tests Assuming that there are k groups of approximately the same size, where the i th group has indices m i to m i, 0 = m 0 < m 1 < < m k = T For each i, j = 1, 2,..., k, define A ij = P [X i > X j ] P [X i = X j ] and for each i = 1, 2,..., k, let n i = m i m i 1 be the size of the i th group. Null Hypothesis: H 0 : For each i = 1, 2,..., k, j=k n j j=1 T A i,j = 0.5. Alt. Hypothesis: H a : For at least one i = 1, 2,..., k, j=k j=1 n j T A i,j 0.5. This rather convoluted statement of the null and alternative hypotheses can be reduced to the statement that the null hypothesis H 0 must be invariant under the conversion of returns to ranks. A weaker condition which ensures that this null hypothesis holds is 18
19 Let A i be the indices of group i. For each j A 1, let X j have the form X j = µ + e j, and for each i > 1 and each j A i, let X j = µ + i + e j. Assume that e l for l = 1, 2,..., T are a random sample from a continuous distribution. Null Hypothesis: H0 : 2 = 3 = = k = 0. Alt. Hypothesis: Ha : At least one i 0 (one-sided test is available for k = 2). The model assumes that each group s distribution is merely shifted from the others; the null hypothesis H 0 hypothesis is then true if all shifts are 0. Historically, the Wilcoxon test (k = 2) was proposed first and was generalized later to several groups. Thus 1. For k = 2, the test is known as the Wilcoxon Rank-Sum Test. 2. For k > 2, the test is known as the Kruskal-Wallis Test. The Wilcoxon and Kruskal-Wallis tests convert the combined samples to ranks and then applies a small or large-sample ANOVA to the groups of ranks. Simplified Explanation: The Kruskal-Wallis test is identical to the Wilcoxon Rank-Sum test when k = 2. Most users of these tests consider that they compare the medians of the distributions of the k groups and the null hypothesis will be rejected when there is evidence that not all of these medians are equal. The Kruskal-Wallis test, however, is not powerful against most differences that exist in returns data for two reasons: (1) it converts the data to ranks, which neglects their magnitudes, and (2) it ignores the time sequence of the data so that serial dependencies are ignored. Reference: Vargha & Delaney [Vargha and Delaney, 1998]; Hollander & Wolfe [Hollander and Wolfe, 1999]. NOTE: The following three tests, the Ansari-Bradley, Fligner-Kileen and Mood, all have the same null hypothesis for testing the scale of k groups, but have different test statistics. However, the default R-language implementations of the Ansari-Bradley and Mood tests supports k = 2 only, while the Fligner- Killeen test supports k 2. For simplicity, all null hypotheses are stated for two groups only. 19
20 5.3 The Ansari-Bradley Test of Equality of Variances Let F be an unknown continuous distribution function, let {X 1i } n1 1 be a random sample with d.f. F (x 1 m) and let {X 2i } n2 1 be an independent r.s. with d.f. F ((x 2 m)/s), where m is a common unknown location for the two distributions and where s is an unknown scale parameter. Null Hypothesis: H 0 : s = 1. Alt. Hypothesis: H a : s 1, or the distributions are not representable as described in the model. The test statistic ranks the combined N = n 1 + n 2 observations from the n 1 of X 1 and the n 2 of X 2 and assigns a score of a i = (N + 1)/2 i N (14) to the i th in the ranking. Thus the first and last have the smallest scores of 1, the next smallest and next largest, scores of 2 and so on. Designating the indices in the ranking corresponding to the X 1 sample as i 1, i 2,..., i n1, the test statistic is Z = j=n 1 j=1 a ij. (15) Under the null hypothesis, the distribution of Z is described in [Hollander and Wolfe, 1999]; for small samples, a formula is available and for large samples a normal approximation is available. It is advisable to calculate medians for each sample separately and to subtact them from their sample prior to performing the test. Note that this test, unlike the parametric F -test, does not require that means or variances of either sample exist. Note also that the R-language implementation of the Ansari-Bradley test ansari.test does not median-correct the data, and that correction should be performed prior to the test. Simplified Explanation: The Ansari-Bradley test relies on the fact that under the null distribution, all permutations of data are equally likely. On the other hand, if one sample has the same location but a different variance, then more of its scores will be near the extremes of the ranks, or more toward the middle of the ranks. In these cases, (15) will be either too large, or too small, respectively, compared to a typical case under the null hypothesis. Before applying this test, each sample s median should be subtracted from its observations. A word should be added contrasting this test with the Fligner-Kileen and Mood tests. The difference among these tests is the method of scoring; 7 in all other formal respects they are the same. Among the three scale tests, this one has the smallest weights for large observations, so that unlike the others, it is less sensitive to 7 Scores are the values assigned to the observations and then summed up to form the test statistic in the Ansari-Bradley, they are the a i of formula (14). 20
21 very large observations, but more sensitive to very small observations. As such, it should not perform as well on market data that has outliers. Like the Kruskal- Wallis test and other distributional comparison tests, it is questionable for time series because the null hypothesis ignores the sample s original order. Reference: Hollander & Wolfe [Hollander and Wolfe, 1999]. 5.4 The Fligner-Kileen Test of Scale Null Hypothesis: H 0 : s = 1. Alt. Hypothesis: Let F be an unknown continuous distribution function, random sample X 1 have distribution F (x 1 m) and independent random sample X 2 have distribution F ((x 2 m)/s), where m is a common unknown location for the distributions and s is an unknown scale parameter. H a : s 1, or the distributions are not representable as described in the model. Let N = n 1 +n 2 be the total number of observations, where n 1 is the number in the X 1 sample and n 2 is the number in the X 2 sample. The Fligner-Kileen test assigns scores a ij to observations x ij as follows: (1) observations in each group are centered separately using means or medians, (2) the centered data are combined into one sample and ranked from lowest to highest, yielding ranks r ij, 8 and (3) the score for the j th observation in the i th group, i = 1, 2 is assigned the score ( ) 1 a ij = Φ r ij (16) 2(N + 1) where Φ is the standard normal distribution and Φ 1 is its inverse, i.e. maps probabilites to quantiles. Scores are then assigned as in steps (2) and (3) above and the test statistic where χ 2 1 = n 1(Ā1 ā) + n 2 (Ā2 ā) V 2, (17) Ā i = 1 a ij, n i j ā = (n 1 A 1 + n 2 A 2 )/N, and i=2 V 2 i=1 j = (a ij ā) 2. N 1 As the notation indicates, χ 2 1 is distributed as a χ 2 distribution with 1 degree of freedom. Simplified Explanation: The Fligner-Kileen test relies on the fact that under the null distribution, all permutations of data are equally likely. If one sample 8 For the handling of ties, see the references below. 21
22 has the same location but a different variance, then more of its scores will be near the extremes of the ranks, or more toward the middle of the ranks. In either case the χ 2 1 statistic will be either too large or too small compared to typical ones under the null hypothesis. By default, the R-language implementation centers the data with medians. A word should be added contrasting this test with the Ansari-Bradley and Mood tests. The difference among these tests is the method of scoring; in all other formal respects they are the same. This test has the relatively largest weights for extreme observations, so that among the others it should be most sensitive to extreme observations; as such, it should perform better than the others on a sample with more outliers. Like the Kruskal-Wallis test and other distributional comparison tests, however, it is questionable for time series because the null hypothesis ignores the sample s original order. Reference: Conover, Johnson & Johnson [Conover et al., 1981]. 5.5 The Mood Test of Scale Null Hypothesis: H 0 : s = 1. Alt. Hypothesis: Let F be an unknown continuous distribution function, random sample X 1 have distribution F (x 1 m and independent random sample X 2 have distribution F ((x 2 m)/s), where m is a common unknown location for the distributions and s is an unknown scale parameter. H a : s 1, or the distributions are not representable as described in the model. The Mood test has the form (15) where r ij is the rank of the (ij) th observation, i = 1, 2, j = 1, 2,..., n i, and with scores ( a ij = r ij N + 1 ) 2 (18) 2 and ā i = j=n i j=1 a ij. In other words, it differs from the Ansari-Bradley test only in the form of its scores a ij. The critical values under the null hypothesis are given by a formula in small samples, and by a normal approximation in large samples. Simplified Explanation: The test statistic is based on the fact that under the null distribution, all permutations of data are equally likely. If one sample has the same location but a different variance, then more of its scores will be near the extremes of the ranks, or more toward the middle of the ranks. A word should be added contrasting this test with the Ansari-Bradley and Fligner- Kileen tests. The difference among these tests is the method of scoring; in all other formal respects they are the same. On large observations, this test has larger scores than Ansari-Bradley, but less than Fligner-Kileen, so that it will have intermediate sensitivity to outliers. By default, the R-language implementation does not center the data, so this should be done before analysis. Like the Kruskal-Wallis test and other distributional comparison tests, however, 22
23 it is questionable for time series because the null hypothesis ignores the sample s original order. Reference: Conover, Johnson & Jonhnson [Conover et al., 1981]. 6 Time Series Tests 6.1 General Tests (G.1) The Brock-Dechert-Schenkman (BDS) i.i.d. Test Null Hypothesis: Alt Hypothesis: Arbitrary data {X t } (of infinite extent above and below) H 0 : X i are i.i.d. from continuous distribution F (x). H a : X i are from a continuous distributions but not i.i.d.. This test is complicated to describe, but the basic idea is that an i.i.d. series X 1, X 2,..., X T has the property that P [ X t X s < ɛ] 2 = P [ X t X s < ɛ]p [ X t 1 X s 1 < ɛ] (19) for all s, t and ɛ > 0. Thus lagged pairs (X t, X t 1 ) must satisfy this relationship, and by extending the above relationship to sequences of length m, one can obtain analogous relationships among m-length sequences. The test itself uses a statistic which is an estimate of the correlation dimension of the series for an account of how the BDS is developed, see references on chaos or the original articles cited below. The BDS test is one of the few tests for i.i.d. robust against nonlinear and chaotic time series. It is valid for the standardized residuals of all ARIMA, ARCH and GARCH-family models provided the sample size is sufficient (> 500). Simplified Explanation: The BDS test is one of the few i.i.d. tests that is robust against nonlinear and chaotic time series. In general, the BDS test is appropriate for ARIMA and ARCH standardized residuals, but not for small samples from a GARCH model. Reference: Brock, Dechert & Scheinkman [Brock et al., 1996]; LeBaron [LeBaron, 1997], de Lima [de Lima, 1996]. (G.2) The KPSS Test of Stationarity 23
24 Null Hypothesis: Alt Hypothesis: X t = βt + Y t + U t, where U t is i.i.d white noise, Y t = Y t 1 + V t and X t is an observed series, β is a constant, V t is i.i.d. and V t N(0, σv). 2 H0 1 : σv 2 = 0 for trend stationarity and H0 2 : β = 0 for level stationarity. Ha 1 : σv 2 > 0 for stationarity (the series has a unit root.) Ha 2 : β 0 for level stationarity. Trend stationarity refers to a series with β 0, and level stationarity to one with β = 0. The KPSS null hypothesis of stationarity stipulates (1) that there is no random walk term Y t = Y t 1 +V t embedded in the observed series X t, and (2) that the time series has the same probability generating mechanism at all time periods. When a random walk term is present, the series will have a unit root, and therefore, will not be stationary. This test is a one-sided, right-tailed test, so rejection of occurs when the statistic is large. The level stationarity null hypothesis is satisfied when the long term trend of the series is zero. In this case, the test is two-sided, being positive when the series is increasing and negative when it is decreasing. Simplified Explanation: The KPSS test is typically used to detect the existence of an embedded random walk in the series, guaranteeing that ARMA models are not appropriate without a data transformation such as differencing. And if a series is not level stationary, then it is certainly not i.i.d.. Thus rejection of the stationary null hypothesis means that the series is not i.i.d.. The problem with this test is that there are other deviations from i.i.d. besides an embedded random walk, and this test may not be powerful against such alternatives. Reference: Kwiatkowski, Phillips and Schmidt [Kwiatkowski et al., 1991]. (G.3) The Terasvirta Neural Network Test For Neglected Nonlinearity This test is too complicated to described in detail here. The idea is to build a neural network with hidden processing units and to use these to test that the data have non-zero linearity. See the Terasvirta reference for details. Simplified Explanation: The Terasvirta tests the null hypothesis of i.i.d. against the alternative that there is nonlinear dependence of X t on X t 1, X t 2,.... If the null hypothesis is rejected, then there is evidence that the series not only fails to be i.i.d., but the relationship is not linear. Reference: Terasvirta, Lin & Granger [Tersvirta et al., 1993]. (G.4) The White Test for Neglected Nonlinearity 24
A nonparametric test for seasonal unit roots
Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna To be presented in Innsbruck November 7, 2007 Abstract We consider a nonparametric test for the
More informationGlossary. The ISI glossary of statistical terms provides definitions in a number of different languages:
Glossary The ISI glossary of statistical terms provides definitions in a number of different languages: http://isi.cbs.nl/glossary/index.htm Adjusted r 2 Adjusted R squared measures the proportion of the
More informationBCT Lecture 3. Lukas Vacha.
BCT Lecture 3 Lukas Vacha vachal@utia.cas.cz Stationarity and Unit Root Testing Why do we need to test for Non-Stationarity? The stationarity or otherwise of a series can strongly influence its behaviour
More informationRecall the Basics of Hypothesis Testing
Recall the Basics of Hypothesis Testing The level of significance α, (size of test) is defined as the probability of X falling in w (rejecting H 0 ) when H 0 is true: P(X w H 0 ) = α. H 0 TRUE H 1 TRUE
More informationUnit 14: Nonparametric Statistical Methods
Unit 14: Nonparametric Statistical Methods Statistics 571: Statistical Methods Ramón V. León 8/8/2003 Unit 14 - Stat 571 - Ramón V. León 1 Introductory Remarks Most methods studied so far have been based
More informationMultivariate Time Series: VAR(p) Processes and Models
Multivariate Time Series: VAR(p) Processes and Models A VAR(p) model, for p > 0 is X t = φ 0 + Φ 1 X t 1 + + Φ p X t p + A t, where X t, φ 0, and X t i are k-vectors, Φ 1,..., Φ p are k k matrices, with
More informationTHE ROYAL STATISTICAL SOCIETY HIGHER CERTIFICATE
THE ROYAL STATISTICAL SOCIETY 004 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE PAPER II STATISTICAL METHODS The Society provides these solutions to assist candidates preparing for the examinations in future
More informationVolatility. Gerald P. Dwyer. February Clemson University
Volatility Gerald P. Dwyer Clemson University February 2016 Outline 1 Volatility Characteristics of Time Series Heteroskedasticity Simpler Estimation Strategies Exponentially Weighted Moving Average Use
More informationStatistical Procedures for Testing Homogeneity of Water Quality Parameters
Statistical Procedures for ing Homogeneity of Water Quality Parameters Xu-Feng Niu Professor of Statistics Department of Statistics Florida State University Tallahassee, FL 3306 May-September 004 1. Nonparametric
More informationNonlinear Time Series
Nonlinear Time Series Recall that a linear time series {X t } is one that follows the relation, X t = µ + i=0 ψ i A t i, where {A t } is iid with mean 0 and finite variance. A linear time series is stationary
More informationIf we want to analyze experimental or simulated data we might encounter the following tasks:
Chapter 1 Introduction If we want to analyze experimental or simulated data we might encounter the following tasks: Characterization of the source of the signal and diagnosis Studying dependencies Prediction
More informationProf. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis
Introduction to Time Series Analysis 1 Contents: I. Basics of Time Series Analysis... 4 I.1 Stationarity... 5 I.2 Autocorrelation Function... 9 I.3 Partial Autocorrelation Function (PACF)... 14 I.4 Transformation
More informationNonparametric Statistics. Leah Wright, Tyler Ross, Taylor Brown
Nonparametric Statistics Leah Wright, Tyler Ross, Taylor Brown Before we get to nonparametric statistics, what are parametric statistics? These statistics estimate and test population means, while holding
More informationStatistical Hypothesis Testing with SAS and R
Statistical Hypothesis Testing with SAS and R Statistical Hypothesis Testing with SAS and R Dirk Taeger Institute for Prevention and Occupational Medicine of the German Social Accident Insurance, Institute
More informationAN EMPIRICAL LIKELIHOOD RATIO TEST FOR NORMALITY
Econometrics Working Paper EWP0401 ISSN 1485-6441 Department of Economics AN EMPIRICAL LIKELIHOOD RATIO TEST FOR NORMALITY Lauren Bin Dong & David E. A. Giles Department of Economics, University of Victoria
More information5.5.3 Statistical Innovative Trend Test Application Crossing Trend Analysis Methodology Rational Concept...
Contents 1 Introduction.... 1 1.1 General... 1 1.2 Trend Definition and Analysis... 3 1.2.1 Conceptual and Visual Trends.... 4 1.2.2 Mathematical Trend.... 7 1.2.3 Statistical Trend.... 9 1.3 Trend in
More informationOne-Sample Numerical Data
One-Sample Numerical Data quantiles, boxplot, histogram, bootstrap confidence intervals, goodness-of-fit tests University of California, San Diego Instructor: Ery Arias-Castro http://math.ucsd.edu/~eariasca/teaching.html
More informationTopic 4 Unit Roots. Gerald P. Dwyer. February Clemson University
Topic 4 Unit Roots Gerald P. Dwyer Clemson University February 2016 Outline 1 Unit Roots Introduction Trend and Difference Stationary Autocorrelations of Series That Have Deterministic or Stochastic Trends
More informationPSY 307 Statistics for the Behavioral Sciences. Chapter 20 Tests for Ranked Data, Choosing Statistical Tests
PSY 307 Statistics for the Behavioral Sciences Chapter 20 Tests for Ranked Data, Choosing Statistical Tests What To Do with Non-normal Distributions Tranformations (pg 382): The shape of the distribution
More informationStatistical Inference: Estimation and Confidence Intervals Hypothesis Testing
Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing 1 In most statistics problems, we assume that the data have been generated from some unknown probability distribution. We desire
More informationNonstationary Time Series:
Nonstationary Time Series: Unit Roots Egon Zakrajšek Division of Monetary Affairs Federal Reserve Board Summer School in Financial Mathematics Faculty of Mathematics & Physics University of Ljubljana September
More informationTextbook Examples of. SPSS Procedure
Textbook s of IBM SPSS Procedures Each SPSS procedure listed below has its own section in the textbook. These sections include a purpose statement that describes the statistical test, identification of
More informationResearch Article A Nonparametric Two-Sample Wald Test of Equality of Variances
Advances in Decision Sciences Volume 211, Article ID 74858, 8 pages doi:1.1155/211/74858 Research Article A Nonparametric Two-Sample Wald Test of Equality of Variances David Allingham 1 andj.c.w.rayner
More informationEcon 423 Lecture Notes: Additional Topics in Time Series 1
Econ 423 Lecture Notes: Additional Topics in Time Series 1 John C. Chao April 25, 2017 1 These notes are based in large part on Chapter 16 of Stock and Watson (2011). They are for instructional purposes
More informationCHAPTER 17 CHI-SQUARE AND OTHER NONPARAMETRIC TESTS FROM: PAGANO, R. R. (2007)
FROM: PAGANO, R. R. (007) I. INTRODUCTION: DISTINCTION BETWEEN PARAMETRIC AND NON-PARAMETRIC TESTS Statistical inference tests are often classified as to whether they are parametric or nonparametric Parameter
More informationLecture 5: Unit Roots, Cointegration and Error Correction Models The Spurious Regression Problem
Lecture 5: Unit Roots, Cointegration and Error Correction Models The Spurious Regression Problem Prof. Massimo Guidolin 20192 Financial Econometrics Winter/Spring 2018 Overview Stochastic vs. deterministic
More informationSAS/STAT 14.1 User s Guide. Introduction to Nonparametric Analysis
SAS/STAT 14.1 User s Guide Introduction to Nonparametric Analysis This document is an individual chapter from SAS/STAT 14.1 User s Guide. The correct bibliographic citation for this manual is as follows:
More informationFinancial Econometrics and Quantitative Risk Managenent Return Properties
Financial Econometrics and Quantitative Risk Managenent Return Properties Eric Zivot Updated: April 1, 2013 Lecture Outline Course introduction Return definitions Empirical properties of returns Reading
More informationHypothesis testing:power, test statistic CMS:
Hypothesis testing:power, test statistic The more sensitive the test, the better it can discriminate between the null and the alternative hypothesis, quantitatively, maximal power In order to achieve this
More informationMultivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8]
1 Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8] Insights: Price movements in one market can spread easily and instantly to another market [economic globalization and internet
More informationApplication of Variance Homogeneity Tests Under Violation of Normality Assumption
Application of Variance Homogeneity Tests Under Violation of Normality Assumption Alisa A. Gorbunova, Boris Yu. Lemeshko Novosibirsk State Technical University Novosibirsk, Russia e-mail: gorbunova.alisa@gmail.com
More information* Tuesday 17 January :30-16:30 (2 hours) Recored on ESSE3 General introduction to the course.
Name of the course Statistical methods and data analysis Audience The course is intended for students of the first or second year of the Graduate School in Materials Engineering. The aim of the course
More informationAnalyzing Small Sample Experimental Data
Analyzing Small Sample Experimental Data Session 2: Non-parametric tests and estimators I Dominik Duell (University of Essex) July 15, 2017 Pick an appropriate (non-parametric) statistic 1. Intro to non-parametric
More informationA comparison study of the nonparametric tests based on the empirical distributions
통계연구 (2015), 제 20 권제 3 호, 1-12 A comparison study of the nonparametric tests based on the empirical distributions Hyo-Il Park 1) Abstract In this study, we propose a nonparametric test based on the empirical
More informationA Monte Carlo Simulation of the Robust Rank- Order Test Under Various Population Symmetry Conditions
Journal of Modern Applied Statistical Methods Volume 12 Issue 1 Article 7 5-1-2013 A Monte Carlo Simulation of the Robust Rank- Order Test Under Various Population Symmetry Conditions William T. Mickelson
More informationNonlinear Characterization of Activity Dynamics in Online Collaboration Websites
Nonlinear Characterization of Activity Dynamics in Online Collaboration Websites Tiago Santos 1 Simon Walk 2 Denis Helic 3 1 Know-Center, Graz, Austria 2 Stanford University 3 Graz University of Technology
More informationLecture 6a: Unit Root and ARIMA Models
Lecture 6a: Unit Root and ARIMA Models 1 2 Big Picture A time series is non-stationary if it contains a unit root unit root nonstationary The reverse is not true. For example, y t = cos(t) + u t has no
More informationIntroduction and Descriptive Statistics p. 1 Introduction to Statistics p. 3 Statistics, Science, and Observations p. 5 Populations and Samples p.
Preface p. xi Introduction and Descriptive Statistics p. 1 Introduction to Statistics p. 3 Statistics, Science, and Observations p. 5 Populations and Samples p. 6 The Scientific Method and the Design of
More informationSolutions exercises of Chapter 7
Solutions exercises of Chapter 7 Exercise 1 a. These are paired samples: each pair of half plates will have about the same level of corrosion, so the result of polishing by the two brands of polish are
More informationModeling Hydrologic Chanae
Modeling Hydrologic Chanae Statistical Methods Richard H. McCuen Department of Civil and Environmental Engineering University of Maryland m LEWIS PUBLISHERS A CRC Press Company Boca Raton London New York
More informationContents. Acknowledgments. xix
Table of Preface Acknowledgments page xv xix 1 Introduction 1 The Role of the Computer in Data Analysis 1 Statistics: Descriptive and Inferential 2 Variables and Constants 3 The Measurement of Variables
More informationIntroduction to Nonparametric Analysis (Chapter)
SAS/STAT 9.3 User s Guide Introduction to Nonparametric Analysis (Chapter) SAS Documentation This document is an individual chapter from SAS/STAT 9.3 User s Guide. The correct bibliographic citation for
More informationAPPLIED TIME SERIES ECONOMETRICS
APPLIED TIME SERIES ECONOMETRICS Edited by HELMUT LÜTKEPOHL European University Institute, Florence MARKUS KRÄTZIG Humboldt University, Berlin CAMBRIDGE UNIVERSITY PRESS Contents Preface Notation and Abbreviations
More informationLecture Slides. Elementary Statistics. by Mario F. Triola. and the Triola Statistics Series
Lecture Slides Elementary Statistics Tenth Edition and the Triola Statistics Series by Mario F. Triola Slide 1 Chapter 13 Nonparametric Statistics 13-1 Overview 13-2 Sign Test 13-3 Wilcoxon Signed-Ranks
More informationLecture Slides. Section 13-1 Overview. Elementary Statistics Tenth Edition. Chapter 13 Nonparametric Statistics. by Mario F.
Lecture Slides Elementary Statistics Tenth Edition and the Triola Statistics Series by Mario F. Triola Slide 1 Chapter 13 Nonparametric Statistics 13-1 Overview 13-2 Sign Test 13-3 Wilcoxon Signed-Ranks
More informationPhysics 509: Non-Parametric Statistics and Correlation Testing
Physics 509: Non-Parametric Statistics and Correlation Testing Scott Oser Lecture #19 Physics 509 1 What is non-parametric statistics? Non-parametric statistics is the application of statistical tests
More informationChapte The McGraw-Hill Companies, Inc. All rights reserved.
er15 Chapte Chi-Square Tests d Chi-Square Tests for -Fit Uniform Goodness- Poisson Goodness- Goodness- ECDF Tests (Optional) Contingency Tables A contingency table is a cross-tabulation of n paired observations
More informationDistribution-Free Procedures (Devore Chapter Fifteen)
Distribution-Free Procedures (Devore Chapter Fifteen) MATH-5-01: Probability and Statistics II Spring 018 Contents 1 Nonparametric Hypothesis Tests 1 1.1 The Wilcoxon Rank Sum Test........... 1 1. Normal
More informationTesting for non-stationarity
20 November, 2009 Overview The tests for investigating the non-stationary of a time series falls into four types: 1 Check the null that there is a unit root against stationarity. Within these, there are
More informationEmpirical Power of Four Statistical Tests in One Way Layout
International Mathematical Forum, Vol. 9, 2014, no. 28, 1347-1356 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/imf.2014.47128 Empirical Power of Four Statistical Tests in One Way Layout Lorenzo
More informationAsymptotic Statistics-VI. Changliang Zou
Asymptotic Statistics-VI Changliang Zou Kolmogorov-Smirnov distance Example (Kolmogorov-Smirnov confidence intervals) We know given α (0, 1), there is a well-defined d = d α,n such that, for any continuous
More informationTesting for Normality
Testing for Normality For each mean and standard deviation combination a theoretical normal distribution can be determined. This distribution is based on the proportions shown below. This theoretical normal
More informationAnalysis of 2x2 Cross-Over Designs using T-Tests
Chapter 234 Analysis of 2x2 Cross-Over Designs using T-Tests Introduction This procedure analyzes data from a two-treatment, two-period (2x2) cross-over design. The response is assumed to be a continuous
More informationA nonparametric two-sample wald test of equality of variances
University of Wollongong Research Online Faculty of Informatics - Papers (Archive) Faculty of Engineering and Information Sciences 211 A nonparametric two-sample wald test of equality of variances David
More informationCHAPTER 2 SIMPLE LINEAR REGRESSION
CHAPTER 2 SIMPLE LINEAR REGRESSION 1 Examples: 1. Amherst, MA, annual mean temperatures, 1836 1997 2. Summer mean temperatures in Mount Airy (NC) and Charleston (SC), 1948 1996 Scatterplots outliers? influential
More informationNAG Library Chapter Introduction. G08 Nonparametric Statistics
NAG Library Chapter Introduction G08 Nonparametric Statistics Contents 1 Scope of the Chapter.... 2 2 Background to the Problems... 2 2.1 Parametric and Nonparametric Hypothesis Testing... 2 2.2 Types
More informationContents 1. Contents
Contents 1 Contents 1 One-Sample Methods 3 1.1 Parametric Methods.................... 4 1.1.1 One-sample Z-test (see Chapter 0.3.1)...... 4 1.1.2 One-sample t-test................. 6 1.1.3 Large sample
More informationContents Kruskal-Wallis Test Friedman s Two-way Analysis of Variance by Ranks... 47
Contents 1 Non-parametric Tests 3 1.1 Introduction....................................... 3 1.2 Advantages of Non-parametric Tests......................... 4 1.3 Disadvantages of Non-parametric Tests........................
More informationE 4160 Autumn term Lecture 9: Deterministic trends vs integrated series; Spurious regression; Dickey-Fuller distribution and test
E 4160 Autumn term 2016. Lecture 9: Deterministic trends vs integrated series; Spurious regression; Dickey-Fuller distribution and test Ragnar Nymoen Department of Economics, University of Oslo 24 October
More informationNon-Parametric Statistics: When Normal Isn t Good Enough"
Non-Parametric Statistics: When Normal Isn t Good Enough" Professor Ron Fricker" Naval Postgraduate School" Monterey, California" 1/28/13 1 A Bit About Me" Academic credentials" Ph.D. and M.A. in Statistics,
More informationNonparametric statistic methods. Waraphon Phimpraphai DVM, PhD Department of Veterinary Public Health
Nonparametric statistic methods Waraphon Phimpraphai DVM, PhD Department of Veterinary Public Health Measurement What are the 4 levels of measurement discussed? 1. Nominal or Classificatory Scale Gender,
More information5 Introduction to the Theory of Order Statistics and Rank Statistics
5 Introduction to the Theory of Order Statistics and Rank Statistics This section will contain a summary of important definitions and theorems that will be useful for understanding the theory of order
More informationFor the full text of this licence, please go to:
This item was submitted to Loughborough s Institutional Repository by the author and is made available under the following Creative Commons Licence conditions. For the full text of this licence, please
More informationG. S. Maddala Kajal Lahiri. WILEY A John Wiley and Sons, Ltd., Publication
G. S. Maddala Kajal Lahiri WILEY A John Wiley and Sons, Ltd., Publication TEMT Foreword Preface to the Fourth Edition xvii xix Part I Introduction and the Linear Regression Model 1 CHAPTER 1 What is Econometrics?
More informationNon-parametric tests, part A:
Two types of statistical test: Non-parametric tests, part A: Parametric tests: Based on assumption that the data have certain characteristics or "parameters": Results are only valid if (a) the data are
More informationTime Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY
Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY & Contents PREFACE xiii 1 1.1. 1.2. Difference Equations First-Order Difference Equations 1 /?th-order Difference
More informationE 4101/5101 Lecture 9: Non-stationarity
E 4101/5101 Lecture 9: Non-stationarity Ragnar Nymoen 30 March 2011 Introduction I Main references: Hamilton Ch 15,16 and 17. Davidson and MacKinnon Ch 14.3 and 14.4 Also read Ch 2.4 and Ch 2.5 in Davidson
More informationSmall Sample Properties of Alternative Tests for Martingale Difference Hypothesis
Small Sample Properties of Alternative Tests for Martingale Difference Hypothesis Amélie Charles, Olivier Darné, Jae Kim To cite this version: Amélie Charles, Olivier Darné, Jae Kim. Small Sample Properties
More informationTesting for Normality
Testing for Normality For each mean and standard deviation combination a theoretical normal distribution can be determined. This distribution is based on the proportions shown below. This theoretical normal
More informationUnivariate, Nonstationary Processes
Univariate, Nonstationary Processes Jamie Monogan University of Georgia March 20, 2018 Jamie Monogan (UGA) Univariate, Nonstationary Processes March 20, 2018 1 / 14 Objectives By the end of this meeting,
More informationWhen is a copula constant? A test for changing relationships
When is a copula constant? A test for changing relationships Fabio Busetti and Andrew Harvey Bank of Italy and University of Cambridge November 2007 usetti and Harvey (Bank of Italy and University of Cambridge)
More informationON SMALL SAMPLE PROPERTIES OF PERMUTATION TESTS: INDEPENDENCE BETWEEN TWO SAMPLES
ON SMALL SAMPLE PROPERTIES OF PERMUTATION TESTS: INDEPENDENCE BETWEEN TWO SAMPLES Hisashi Tanizaki Graduate School of Economics, Kobe University, Kobe 657-8501, Japan e-mail: tanizaki@kobe-u.ac.jp Abstract:
More informationNearest-Neighbor Forecasts Of U.S. Interest Rates
1 Nearest-Neighbor Forecasts Of U.S. Interest Rates John Barkoulas 1 Department of Economics University of Tennessee 534 Stokely Management Center Knoxville, TN 37996 Christopher F. Baum Department of
More informationApplication of Homogeneity Tests: Problems and Solution
Application of Homogeneity Tests: Problems and Solution Boris Yu. Lemeshko (B), Irina V. Veretelnikova, Stanislav B. Lemeshko, and Alena Yu. Novikova Novosibirsk State Technical University, Novosibirsk,
More informationResearch Article The Laplace Likelihood Ratio Test for Heteroscedasticity
International Mathematics and Mathematical Sciences Volume 2011, Article ID 249564, 7 pages doi:10.1155/2011/249564 Research Article The Laplace Likelihood Ratio Test for Heteroscedasticity J. Martin van
More informationStochastic Processes
Stochastic Processes Stochastic Process Non Formal Definition: Non formal: A stochastic process (random process) is the opposite of a deterministic process such as one defined by a differential equation.
More informationInferential Statistics
Inferential Statistics Eva Riccomagno, Maria Piera Rogantin DIMA Università di Genova riccomagno@dima.unige.it rogantin@dima.unige.it Part G Distribution free hypothesis tests 1. Classical and distribution-free
More informationTESTING FOR NORMALITY IN THE LINEAR REGRESSION MODEL: AN EMPIRICAL LIKELIHOOD RATIO TEST
Econometrics Working Paper EWP0402 ISSN 1485-6441 Department of Economics TESTING FOR NORMALITY IN THE LINEAR REGRESSION MODEL: AN EMPIRICAL LIKELIHOOD RATIO TEST Lauren Bin Dong & David E. A. Giles Department
More informationTime Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley
Time Series Models and Inference James L. Powell Department of Economics University of California, Berkeley Overview In contrast to the classical linear regression model, in which the components of the
More informationDr. Maddah ENMG 617 EM Statistics 10/12/12. Nonparametric Statistics (Chapter 16, Hines)
Dr. Maddah ENMG 617 EM Statistics 10/12/12 Nonparametric Statistics (Chapter 16, Hines) Introduction Most of the hypothesis testing presented so far assumes normally distributed data. These approaches
More informationH 2 : otherwise. that is simply the proportion of the sample points below level x. For any fixed point x the law of large numbers gives that
Lecture 28 28.1 Kolmogorov-Smirnov test. Suppose that we have an i.i.d. sample X 1,..., X n with some unknown distribution and we would like to test the hypothesis that is equal to a particular distribution
More informationDoes k-th Moment Exist?
Does k-th Moment Exist? Hitomi, K. 1 and Y. Nishiyama 2 1 Kyoto Institute of Technology, Japan 2 Institute of Economic Research, Kyoto University, Japan Email: hitomi@kit.ac.jp Keywords: Existence of moments,
More informationNon-parametric Tests for Complete Data
Non-parametric Tests for Complete Data Non-parametric Tests for Complete Data Vilijandas Bagdonavičius Julius Kruopis Mikhail S. Nikulin First published 2011 in Great Britain and the United States by
More informationDiagnostic Test for GARCH Models Based on Absolute Residual Autocorrelations
Diagnostic Test for GARCH Models Based on Absolute Residual Autocorrelations Farhat Iqbal Department of Statistics, University of Balochistan Quetta-Pakistan farhatiqb@gmail.com Abstract In this paper
More informationMultivariate Regression
Multivariate Regression The so-called supervised learning problem is the following: we want to approximate the random variable Y with an appropriate function of the random variables X 1,..., X p with the
More informationNon-parametric (Distribution-free) approaches p188 CN
Week 1: Introduction to some nonparametric and computer intensive (re-sampling) approaches: the sign test, Wilcoxon tests and multi-sample extensions, Spearman s rank correlation; the Bootstrap. (ch14
More informationSimulating Uniform- and Triangular- Based Double Power Method Distributions
Journal of Statistical and Econometric Methods, vol.6, no.1, 2017, 1-44 ISSN: 1792-6602 (print), 1792-6939 (online) Scienpress Ltd, 2017 Simulating Uniform- and Triangular- Based Double Power Method Distributions
More informationEconometric Methods for Panel Data
Based on the books by Baltagi: Econometric Analysis of Panel Data and by Hsiao: Analysis of Panel Data Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies
More informationSequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process
Applied Mathematical Sciences, Vol. 4, 2010, no. 62, 3083-3093 Sequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process Julia Bondarenko Helmut-Schmidt University Hamburg University
More informationMA Advanced Econometrics: Applying Least Squares to Time Series
MA Advanced Econometrics: Applying Least Squares to Time Series Karl Whelan School of Economics, UCD February 15, 2011 Karl Whelan (UCD) Time Series February 15, 2011 1 / 24 Part I Time Series: Standard
More informationRank-Based Methods. Lukas Meier
Rank-Based Methods Lukas Meier 20.01.2014 Introduction Up to now we basically always used a parametric family, like the normal distribution N (µ, σ 2 ) for modeling random data. Based on observed data
More informationNON-PARAMETRIC STATISTICS * (http://www.statsoft.com)
NON-PARAMETRIC STATISTICS * (http://www.statsoft.com) 1. GENERAL PURPOSE 1.1 Brief review of the idea of significance testing To understand the idea of non-parametric statistics (the term non-parametric
More informationModule 9: Nonparametric Statistics Statistics (OA3102)
Module 9: Nonparametric Statistics Statistics (OA3102) Professor Ron Fricker Naval Postgraduate School Monterey, California Reading assignment: WM&S chapter 15.1-15.6 Revision: 3-12 1 Goals for this Lecture
More informationunadjusted model for baseline cholesterol 22:31 Monday, April 19,
unadjusted model for baseline cholesterol 22:31 Monday, April 19, 2004 1 Class Level Information Class Levels Values TRETGRP 3 3 4 5 SEX 2 0 1 Number of observations 916 unadjusted model for baseline cholesterol
More informationarxiv:math/ v1 [math.pr] 9 Sep 2003
arxiv:math/0309164v1 [math.pr] 9 Sep 003 A NEW TEST FOR THE MULTIVARIATE TWO-SAMPLE PROBLEM BASED ON THE CONCEPT OF MINIMUM ENERGY G. Zech and B. Aslan University of Siegen, Germany August 8, 018 Abstract
More informationECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications
ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications Yongmiao Hong Department of Economics & Department of Statistical Sciences Cornell University Spring 2019 Time and uncertainty
More informationANALYSIS OF VARIANCE OF BALANCED DAIRY SCIENCE DATA USING SAS
ANALYSIS OF VARIANCE OF BALANCED DAIRY SCIENCE DATA USING SAS Ravinder Malhotra and Vipul Sharma National Dairy Research Institute, Karnal-132001 The most common use of statistics in dairy science is testing
More informationFinancial Econometrics
Financial Econometrics Nonlinear time series analysis Gerald P. Dwyer Trinity College, Dublin January 2016 Outline 1 Nonlinearity Does nonlinearity matter? Nonlinear models Tests for nonlinearity Forecasting
More information401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.
401 Review Major topics of the course 1. Univariate analysis 2. Bivariate analysis 3. Simple linear regression 4. Linear algebra 5. Multiple regression analysis Major analysis methods 1. Graphical analysis
More informationChapter 15: Nonparametric Statistics Section 15.1: An Overview of Nonparametric Statistics
Section 15.1: An Overview of Nonparametric Statistics Understand Difference between Parametric and Nonparametric Statistical Procedures Parametric statistical procedures inferential procedures that rely
More information