A Reference Guide for Statistical Tests Useful in Finance (Preliminary)

Size: px
Start display at page:

Download "A Reference Guide for Statistical Tests Useful in Finance (Preliminary)"

Transcription

1 A Reference Guide for Statistical Tests Useful in Finance (Preliminary) Market Pattern Research, Inc. modified from the book The Strategic Analysis of Financial Markets Volume 1: Framework Steven D. Moffitt, Ph.D. April 14, 2018 Contents 1 A Reference Guide to Statistical Tests for Financial Analysis 3 2 A Note on Nonparametric Tests 4 3 Statistical Tests for Gaussianity The Anderson-Darling Test of Normality (AD Test) The Cramer-von Mises Test (CvM) of Normality The Shapiro-Francia Test of Normality Pearson s Chi-square Goodness-of-Fit Test of Normality The Lilliefors Test of Normality The Jarque-Bera Test of Normality The D Agostino Test of Normality The Shapiro-Wilk Test of Normality Testing for Randomness Nonparametric Tests for Randomness: Runs Tests The Wald Wolfowitz Runs Test The Up and Down Test The Bartels Rank Test Nonparametric Tests for Randomness: Trend Tests The Cox-Stuart Test The Difference Sign Test The Mann Test for Trend The Dietz-Kileen Multivariate Test for Trend

2 5 Nonparametric Tests for Comparing Distributions A Comparison of Distributions Test: The two-sample test of Kolmogorov-Smirnov (KS test) Subset Location Tests: Wilcoxon Rank-Sum (k = 2) and Kruskal-Wallis (k 2) Location Tests The Ansari-Bradley Test of Equality of Variances The Fligner-Kileen Test of Scale The Mood Test of Scale Time Series Tests General Tests Parametric Tests for Random Walks A Unit Root Test: The Augmented Dickey-Fuller Test (ADF-Test)) A Unit Root Test: The Phillips-Perron Test (PP-Test) Variance-Ratio Tests The Lo-MacKinlay Test Chen-Deo Test Chow-Denning Test Wald Test Wright and Joint Wright Tests

3 This article was abstracted from material presented in a forthcoming twovolume series, The Strategic Analysis of Financial Markets by Steven D. Moffitt, Ph.D. in May, 2017 by World Scientific. Those books analyze the strategies used by investors and traders, deconstruct the associated market games in ways that uncover a commonality of structure, and apply statistics, psychology and gambling logic to identify winning (and losing) strategies. 1 A Reference Guide to Statistical Tests for Financial Analysis Though seldom acknowledged, the statistical analysis of price series is a problematic part of empirical finance. The reason is that prices arise from unknown processes that on empirical grounds alone, are unlikely to be stationary since they exhibit outliers and volatility intermittency. But probably 99% of all statistical methods for time series require stationarity. Clearly, traditional statistics should be applied to price series with caution. This paper serves the purpose of presenting methods that detect exploitable stochastic price behavior such as non-randomness, trends, the propensity to produce outliers, etc., which are more akin to data mining and pattern recognition than to traditional statistical estimation and hypothesis testing. The first Section presents tests of the null hypothesis that a distribution is univariate or multivariate Gaussian. Applying these tests to price series almost always results in rejection of Gaussianity, but the nature of that rejection can reveal things about statistical behavior and is useful for component decomposition methods like independent component analysis. The second Section presents tests of the Hypothesis of Randomness for time series, which includes nonparametric tests for runs and for monotonic (not necessarily linear) trends. The third Section presents non-parametric methods for comparing distributions, with the notably exception of bootstrap methods. In the fourth Section, various time series tests are presented: (1) for i.i.d. data, (2) for martingales and martingale difference series,(3) for serial correlation and (4) for random walks. The presentation is semi-formal; each test has its null and alternative hypotheses stated formally, accompanied by an informal discussion of the test s strengths and weaknesses. References are included for those who want additional information or proofs. A number of tests were omitted, mostly because their statement required material deemed too specialized. Four major instances of such omission are (1) monte carlo tests, (2) tests involving wavelets, (3) spectral tests for independence or martingales, and (4) random matrix methods for covariance matrices. Of these omissions, (1) is the most serious, followed closely by (2). 3

4 2 A Note on Nonparametric Tests Nonparametric or Distribution-free statistical tests are ones that are valid for large families of distributions, e.g. for all continuous distributions (ones with a density function). Many such tests use the signs of observations or their ranks in a sample. Without going into the theory of these tests, we point out some obvious properties of tests involving signs or ranks. For ranks, the data are replaced by their order in a sample. For example, the ranks for sample data x 1 = 5, x 2 = 1, x 3 = 2, and x 4 = 10 are R(x 1 ) = 3, R(x 2 ) = 1, R(x 3 ) = 2 and R(x 4 ) = 4. If a test is based on ranks, then it is invariant with respect to all strictly increasing transformations of the underlying data. Thus is one wants to test for a monotone trend in a time series, for example, rank tests are particularly appropriate. Here s an example of a nonparametric test using signs. The statistic is the number of consecutive +1 s in differences sgn(x t+1 X t ), where sgn(x) = 1 if x > 0, 0 if x = 0 and 1 if x < 0 (Bartels Test, Section 4.1.2). This test is also invariant with respect to all strictly increasing transformations of the underlying data. As another example, sgn(x i med(x)) (med(x) is the median of the sample) is invariant with respect to strictly increasing transformations. On the other hand, sgn(x i mean(x)) is not. We also note an extremely important characteristic of sign and rank tests they are robust against outliers. Thus if a few large returns in a long series are doubled, it will generally change tests based on ranks little. We prefer nonparametric tests for because they might perform better on nonstationary data and because small sample sizes are typically not a problem. On the other hand, parametric tests have two things that can go wrong: (1) the chosen parametric family doesn t fit the data, e.g. the data aren t Gaussian but Gaussian models are assumed, and (2) the observations or their transformations don t form a random sample. Nonparametric tests suffer only from the second problem, which is the one of primary interest in trading systems development. 3 Statistical Tests for Gaussianity Gaussian distributions occupy a central place in financial theory for several reasons. They are limiting distributions for random samples of distributions with finite variances, they are minimum information distributions among those that have finite variances, their affine transformations are again Gaussian, and they are infinitely divisible, meaning that for any Gaussian random variable X and any positive integer n, there is a (Gaussian) distribution Y whose n-fold convolution has the same distribution as X. Certainly, financial time series would be much simpler if all distributions were Gaussian, but the reality is far different. In general, tests based on Gaussian distributions should be avoided in trading system analysis, a notable exception being independent component analysis. Eight tests of the null hypotheses that a sample or time series is distributed 4

5 as i.i.d. Gaussians are presented in this Section. They are N.1: The Anderson-Darling Test of Normality (AD Test) N.2: The Cramer-von Mises Test (CvM) of Normality N.3: The Shapiro-Francia Test of Normality N.4: The Pearson Chi-square Goodness-of-Fit Test of Normality N.5: The Lilliefors Test of Normality N.6: The Jarque-Bera Test of Normality N.7: The D Agostino Test of Normality N.8: The Shapiro-Wilk Test of Normality 3.1 The Anderson-Darling Test of Normality (AD Test) Let F be an unknown continuous distribution function for which the mean µ and variance σ 2 exist, and let Φ(x; µ, σ 2 ) be the cumulative normal distribution with mean µ and variance σ 2. Null Hypothesis: H 0 : F = Φ(µ, σ 2 ). Alt. Hypothesis: H a : F Φ(µ, σ 2 ). Let X 1, X 2,... X n be a random sample from F and let ˆF be its empirical distribution function ˆF (x) = 1 n I(X i x), n i=1 where I(c) = 1 if c is true and I(c) = 0 is c is false. Then the Anderson-Darling test statistic is ( n ˆF (x) Φ(x; µ, σ 2 )) 2 [Φ(x; µ, σ 2 )][1 Φ(x; µ, σ 2 )] dφ(x; µ, σ2 ). (1) A formula can be developed from (1), as follows. Assuming that the mean and variance of the X i are unknown, standardize the X i s, Z i = (X i X)/S X, where X is the sample mean and S X is the sample standard deviation, and order the Z i s from lower to highest, calling the results Y i : Define A 2 as A 2 = n 1 n Y 1 < Y 2 < < Y n. n [(2i 1)ln(Φ(Y i )) (2(n i) + 1)ln(1 Φ(Y i ))], i=1 and let A 2 = A 2 (1 + 4/n 25/n 2 ) (2) 5

6 be a correction to A 2 for small samples (see Shorack & Wellner, [Shorack and Wellner, 2009]). The theoretical asymptotic distribution 1 of A 2 has been tabulated, and the null hypothesis of normality will be rejected if A 2 exceeds the tabulated value. Note that the test statistic (1) depends only on the empirical distribution function (EDF) 2, as do several other test criteria, and therefore, are nonparametric tests. Other tests in the EDF family are the Cramer-von Mises test and the Lilliefors test. The rationale for (1) is as follows. If U i is a random sample from N(0, 1), then Φ 1 (U i ) will be a random sample from the uniform distribution on (0, 1). Thus by standardizing the X i to Y i, the distribution of Φ 1 (Y i ) will asymptotically also have a uniform distribution under the null hypothesis. One can then perform a theoretical calculation of the distribution of (1) assuming that Φ is replaced by the cumulative distribution of an uniform distribution and ˆF is the empirical cumulative distribution for an uniform distribution. Simplified Explanation: Note that test statistic (1) will be large if there are too many outliers relative to a normal distribution, since the denominator [Φ(x)][1 Φ(x)] approaches zero quickly beyond x = ±3. Thus the AD is an excellent choice if there are outliers in the data, as there usually are in financial returns series. But distribution is asymptotic, that is, holds exactly only in the limit as n. Reference: Anderson & Darling [Anderson and Darling, 1952, Anderson and Darling, 1954], Shorack & Wellner [Shorack and Wellner, 2009] and Stephens [Stephens, 1974]. 3.2 The Cramer-von Mises Test (CvM) of Normality Let F be an unknown continuous distribution function for which the mean µ and variance σ 2 exist, and let Φ(µ, σ 2 ) be the cumulative normal distribution with mean µ and variance σ 2. Null Hypothesis: H 0 : F = Φ(µ, σ 2 ). Alt. Hypothesis: H a : F Φ(µ, σ 2 ). Let X 1, X 2,... X n be a random sample from F and let ˆF be its empirical distribution function ˆF (x) = 1 i=n I(X i x), n i=1 where I(c) = 1 if c is true and I(c) = 0 is c is false. Then the Cramer-von Mises test statistic is n 1 That is, the distribution as n. 2 And implicity, the normal distribution ( ˆF (x) Φ(x; µ, σ 2 )) 2 dφ(x; µ, σ 2 ). (3) 6

7 A formula can be developed from (3), as follows, assuming that the mean and variance of the X i are unknown. Standardize the X i s, Z i = (X i X)/s X, where X is the sample mean and s X is the sample standard deviation and order the Z i s from least to greatest. Define W 2 as W 2 = 1 n 12n + [ ] 2 2i 1 2n Φ(Z i), (4) i=1 The theoretical distribution of W 2 has been tabulated, and the null hypothesis of normality will be rejected if W 2 exceeds the tabulated value. Note that the test statistic (3) depends only on the empirical distribution function (EDF) 3, as do Anderson-Darling test and the Lilliefors test. The rationale for (3) is as follows. If U i is a random sample from N(0, 1), then Φ 1 (U i ) will be a random sample from the uniform distribution on (0, 1). Thus by standardizing and ordering the X i, the distribution of Φ 1 (Z i ) will asymptotically also have a uniform distribution under the null hypothesis. One can then perform a theoretical calculation of the distribution of (3) assuming that Φ is replaced by the cumulative distribution of an uniform distribution and ˆF is the empirical cumulative distribution for an uniform distribution. Simplified Explanation: The test statistic (3) will be large if differences from a normal in the center or in the extremes receive enough weight. Thus the CvM is an excellent choice relative to the AD test for testing differences in the center of the distribution. Reference: Darling [Darling, 1957]. 3.3 The Shapiro-Francia Test of Normality Let F be an unknown continuous distribution function for which the mean µ and variance σ 2 exist, and let Φ(µ, σ 2 ) be the cumulative normal distribution with mean µ and variance σ 2. Null Hypothesis: H 0 : F = Φ(µ, σ 2 ). Alt. Hypothesis: H a : F Φ(µ, σ 2 ). The Shapiro-Francia test statistic for random sample {X i } n 1 uses the order statistics {X (i) } n 1 to form a test for normality. Order statistics for a sample have the same values as {X i } n 1, but ordered from lowest to highest. 4 Thus X (1) is the smallest value in {X i } n 1, X (2) is the next smallest, X (n) being the largest. 3 And implicity, the normal distribution 4 The ordering of order statistics are unique only for continuous distributions, since the probability of ties is zero. 7

8 With this notation, the Shapiro-Wilk test statistic is W = ( n i=1 b ix (i) ) 2 n i=1 (X i X) 2 (5) where X is the sample mean, with m (b 1, b 2,..., b n ) = (m m) 1/2 m = ( E[X (1) ], E[X (2) ],..., E[X (n) ] ). Since a closed form of the W -test s distribution is unknown, the percentiles have been estimated using Monte Carlo methods. Simplified Explanation: The Shapiro-Francia test compares the order statistics from a random sample to those expected in a normal distribution having the same mean as the sample distribution, but with an assumption of independence. The test is almost the same as the Shapiro-Wilk W test; it differs only in the use of order statistic weights b i instead of a i (see the Shapiro-Wilk description in this Section.) Compared with the Shapiro-Wilk, it is more sensitive to alternatives that are continuous and symmetric with high kurtosis, to ones that are near normal and to ones that are discrete and skewed. But it is less sensitive than W on alternatives that are continuous and skewed with high kurtosis, and on ones that are discrete and symmetric. Since market returns in general are right-skewed and have high kurtosis, the Shapiro-Wilk should be preferred over the Shapiro-Francia on this data. Reference: Shapiro & Francia (1972) [Shapiro and Francia, 1972] and Sarkadi (1975) [Sarkadi, 1975]. 3.4 Pearson s Chi-square Goodness-of-Fit Test of Normality Let F be an unknown continuous distribution function for which the mean µ and variance σ 2 exist, and let Φ(µ, σ 2 ) be the cumulative normal distribution with mean µ and variance σ 2. Null Hypothesis: H 0 : F = Φ(µ, σ 2 ). Alt. Hypothesis: H a : F Φ(µ, σ 2 ). The Pearson χ 2 test is a mainstay of beginning statistics; for a sample that is binned into K classes, it has the form χ 2 = i=k i=1 (O i E i ) 2 E i (6) 8

9 where the i th term of the sum applies to all observations in the i th bin, O i is the observed count in bin i and E i is the expected count (generally non-integer) for bin i. Under conditions of the null hypothesis, the statistic will have an asymptotic χ 2 distribution with K 2 degrees of freedom. Simplified Explanation: The χ 2 assumes that data from a sample are put into discrete bins, each of which has a positive probability of occurrence. It compares using statistic 6 the observed counts in the bins with those expected. The resulting test has K p degrees of freedom, where p is the number of parameters that must be estimated to calculate expected values. For the normal, p = 2, since a mean and variance must be calculated to determine bin probabilities. Reference: Spiegel (2000) [Spiegel et al., 2000]. 3.5 The Lilliefors Test of Normality Let F be an unknown continuous distribution function for which the mean µ and variance σ 2 exist, and let Φ(µ, σ 2 ) be the cumulative normal distribution with mean µ and variance σ 2. Null Hypothesis: H 0 : F = Φ(µ, σ 2 ). Alt. Hypothesis: H a : F Φ(µ, σ 2 ). For a random sample {X i } n 1 having sample mean X and sample standard deviation s X, the Lilliefors test uses the statistic L = sup ˆF (x) Φ(x; X, s 2 x), (7) x where ˆF is the empirical distribution of the random sample and Φ(x; µ, σ 2 ) is the cumulative distribution of a Gaussian distribution with mean µ and variance σ 2, X. This statistic is the same as the Kolmogorov-Smirnov one-sample test (KS test) except that the mean and variance are estimated in the L-test but assumed known in the KS-test. This parameter estimation changes the distribution of L as compared to KS. 5 At this writing, the asymptotic L distribution has been tabulated only using Monte Carlo methods. Simplified Explanation: The Lilliefors statistic measures the maximum absolute difference between the empirical distribution function and a Gaussian having the same mean and standard deviation. The test will be quite sensitive to deviations from normality, but will be inferior to the Anderson-Darling test in the presence of outliers. Reference: Lilliefors [Lilliefors, 1967]. 5 An asymptotic closed form distribution is known for the one-sample KS-test. 9

10 3.6 The Jarque-Bera Test of Normality Let F be an unknown continuous distribution function for which the mean µ and variance σ 2 exist, and let Φ(µ, σ 2 ) be the cumulative normal distribution with mean µ and variance σ 2. Null Hypothesis: H 0 : F = Φ(µ, σ 2 ). Alt. Hypothesis: H a : F Φ(µ, σ 2 ). The Jarque-Bera tests whether a distribution s skewness and kurtosis match those of a Gaussian distribution. For a random sample {X i } n 1 having sample mean X JB = n (U ) 6 (V 3), (8) where and 1 n 4 i=1 U = (X i X) 3 ( 1 n 4 i=1 (X i X) 2) 3/2 1 n 4 i=1 V = (X i X) 4 ( 1 n 4 i=1 (X i X) 2) 2. Under the null hypothesis of normality, U = 0 and V = 3. For large samples, n 200, the distribution of JB is approximately χ 2 with 2 degrees of freedom. Simplified Explanation: The terms U and V are, respectively, estimates of the sample s skew and kurtosis. For a normal distribution, the kurtosis equals 3, thus the appearance of that term in JB. The weights given U and V in the expression defining JB ensure that it is asymptotically χ 2 with 2 d.f. under the null hyothesis. I would not recommend using this test because it is not powerful against non-normal distributions with skew and kurtotis that match a normal. Reference: Jarque & Bera [Jarque and Bera, 1987]. 3.7 The D Agostino Test of Normality Let F be an unknown continuous distribution function for which the mean µ and variance σ 2 exist, and let Φ(µ, σ 2 ) be the cumulative normal distribution with mean µ and variance σ 2. Null Hypothesis: H 0 : F = Φ(µ, σ 2 ). Alt. Hypothesis: H a : F Φ(µ, σ 2 ). The D Agostino K 2 -test statistic for a random sample {X i } n 1 uses Cornish- Fisher expansions to derive an approximate test based on the first four moments of the sample distribution. The test statistic has the form K 2 = Z 1 (g 1 ) 2 + Z 2 (g 2 ) 2 (9) 10

11 where Z 1 (g 1 ) 2 and Z 2 (g 2 ) 2 have lengthy algebraic expressions. See the references for definitions. Like the Jarque-Bera, K 2 tests for skew and kurtosis that match a normal. Simplified Explanation: The K 2 test uses a Cornish-Fisher expansion to approximate a distribution based on the first four moments of a normal. Like the Jarque-Bera test, it is not recommended because any non-normal distribution that matches the first four moments of a normal will not be rejected. Reference: D Agostino [D Agostino, 1971]. 3.8 The Shapiro-Wilk Test of Normality Let F be an unknown continuous distribution function for which the mean µ and variance σ 2 exist, and let Φ(µ, σ 2 ) be the cumulative normal distribution with mean µ and variance σ 2. Null Hypothesis: H 0 : F = Φ(µ, σ 2 ). Alt. Hypothesis: H a : F Φ(µ, σ 2 ). The Shapiro-Wilk test statistic for random sample {X i } n 1 uses the order statistics {X (i) } n 1 to form a test for normality. Order statistics for a sample have the same values as {X i } n 1, but ordered from lowest to highest. 6 Thus X (1) is the smallest value in {X i } n 1, X (2) is the next smallest, X (n) being the largest. With this notation, the Shapiro-Wilk test statistic is W = ( n i=1 a ix (i) ) 2 n i=1 (X i X) 2 (10) where X is the sample mean, with and m V 1 (a 1, a 2,..., a n ) = (m V 2 m) 1/2 m = ( E[X (1) ], E[X (2) ],..., E[X (n) ] ). u 11 u u 1n u 21 u u 2n V =..... u1n u n1 u n2... u nn where u ij = cov(x (i), X (j) ). Since a closed form of the W -test s distribution is unknown, the percentiles have been estimated using Monte Carlo methods. 6 The ordering of order statistics are unique only for continuous distributions since the probability of ties is zero. 11

12 Simplified Explanation: The Shapiro-Wilk test compares the order statistics from a random sample to those expected in a normal distribution having the same mean and variance as the sample distribution. The test may be interpreted as the Pearson correlation coefficient between the ordered observations and the weights a i used in the numerator, so that the test in effect measures the straightness of the line in a normal Q-Q plot. Among the several tests of normality presented above, the Shapiro-Wilk ranks favorably with the Anderson-Darling test and is generally better than the others. Reference: Shapiro & Wilk [Shapiro and Wilk, 1965] and Royston [Royston, 1982, Royston, 1995]. 4 Testing for Randomness The difference between statistics for random samples and statistics for time series is the assumption in the latter of dependence in the variables. In time series models such as ARMA, ARCH and GARCH, decompositions that have a deterministic part and a white noise part are produced (assuming that those models fit, of course). Trading system development has the similar objective of transforming time series of stochastically dependent prices into time series of uncorrelated trades. Therefore it is important to have statistical tools that can test for randomness, that is, which test that a time series of trades behaves like a random sample with positive expected value. An hypothesis that goes a long way toward testing for random sample behavior is the Hypothesis of Randomness. Let X 1, X 2,..., X T,... be the variables of a time series and let P T be the set of all permutations on integers 1,..., T. The Hypothesis of Randomness is met if all permutations of variables in the series have the same distribution: H 0 :(X 1, X 2,..., X T ) D = (X σ(1), X σ(2),..., X σ(t ) ) (11) for any permutation σ P T. where the = D notation means having the same distribution. In the following, we use the notation sgn(x) for +1 x > 0 sgn(x) = 0 x = 0 1 x < Nonparametric Tests for Randomness: Runs Tests In this Section, we present three nonparametric runs tests of the Hypothesis of Randomness for continuous data. Runs of this section are of two types: (1) runs against the median, and (2) ascending/descending runs. Runs against the median are maximal numbers of consecutive observations that are above or below the median. For example, with 4 as the median of the 10 values 12

13 1, 2.3, 3, 3.5, 8, 9, 1, 1.5, 8, 2.7, there are 5 runs below and above the median of lengths 4,2,2,1,1. We note that a test statistic counting the number of runs above or below the median does not depend on the underlying (continuous) distribution and is therefore distribution-free or non-parametric. If there are too many or too few runs against the median, then the series is clearly not uncorrelated. On the other hand, ascending (descending) runs are consecutive increasing (decreasing) values. Under the Hypothesis of Randomness a testic consisting of the number of increasing runs is distribution-free and has a calculable distribution. Too many or too few ascending runs indicates dependence among variables. In the above data, there were ascending runs (using positive or negative differences) of lengths 5 and 2 and descend runs of lengths 1 and The Wald Wolfowitz Runs Test {X t } T 1 are continuous random variables with joint density f(x 1, X 2,..., X T ). Null Hypothesis: H 0 : The Hypothesis of Randomness. Alt. Hypothesis: H a : Series {X t } has either too many or too few runs (1-sided) or either (two-sided.) The following procedure is used to calculate the test statistic: 1. The median or some other cutoff, m x, of the sample x 1, x 2,...x T is calculated, 2. A new sample y 1, y 2,..., y T is formed, where y i = +1, if x i m x and y i = 1, if x i < m x, 3. The number of runs u in the sequence of y s is counted, where a run is a maximal consecutive number of +1 s or 1 s. 4. The value u is then compared to a statistical table and a significance level assigned. It is also possible to conduct runs tests with percentiles other than the median; see the references for details. Simplified Explanation: The runs test examines a series to see if there are too many or too few consecutive observations above the cutoff and below the cutoff, i.e., there are too many or too few runs. If a series is slowly varying and stays above the cutoff for long stretches, or is rapidly oscillating above and below the cutoff, then the null hypothesis will be rejected. Reference: Gibbons & Chakraborti [Gibbons and Chakraborti, 2010], Siegel & Castellan [Siegel and Castellan, 1998] The Up and Down Test {X t } T 1 are continuous random variables with joint density f(x 1, X 2,..., X T ). Null Hypothesis: H 0 : The Hypothesis of Randomness. Alt. Hypothesis: H a : Series {sgn(x t+1 X t )} has either too many or too few runs of +1 s. 13

14 The test statistic is the number of runs of consecutive +1 s of the signed differences Y t = sgn(x t+1 X t ), t = 1, 2,..., T 1. Under the Hypothesis of Randomness, one can calculate finite sample distributions, as in Gibbons & Chakraborti [Gibbons and Chakraborti, 2010]. Letting m be the number of non-zero Y t, it can be shown that the test statistic is asymptotically normal with mean (2m 1)/3 and variance (16m 29)/90. Simplified Explanation: The Up and Down Test determines if there are too many or too few consecutive ascending runs in the data. It is not a test for global trend, though, since a sawtooth pattern that goes nowhere has no long-term trend. Therefore the Up and Down Test rejects the Hypothesis of Randomness, but not a global test of runs for the entire series. In a long series, the Up and Down Test detects local trends. Reference: Gibbons & Chakraborti [Gibbons and Chakraborti, 2010] The Bartels Rank Test {X t } T 1 are continuous random variables with joint density f(x 1, X 2,..., X T ). Null Hypothesis: H 0 : The Hypothesis of Randomness Alt. Hypothesis: H a : The sequential variance of ranks is not equal to the usual variance of ranks. The test statistic is based on a parametric test originally suggested by John von Neumann, but substitutes ranks for observations. The Bartels statistic U is U = (R t R t+1 ) 2 / (R i (t + 1)/2) 2 where R t = rank(x t ), t = 1,..., T. It is known that the distribution of (U 2)/σ, where σ 2 = [4(T 2)(5T 2 2T 9)]/[5T (T + 1)(T 1) 2 ]. is asyptotically normal N(0, 1) under the Hypothesis of Randomness. If for example, the data are positively serially correlated, then U will be small and if negatively correlated, large compared to the average random arrangement. Simplified Explanation: The Bartels test is based on a test originally suggested by von Neumann, in which the ratio of a variance determinded by sequential differences is compared to the variance formed from deviations from the mean (ignoring a norming constant). In the von Neumann test, the asymptotic sampling distribution for a normal distribution can be calculated. The Bartels test substitutes ranks for the actual observations, and forms the ratio of the variance calculated from consecutive ranks R t R t+1 to that based on deviations from the mean R t (T + 1)/2. The reference below shows that the rank version has an asymptotically normal distribution, with mean and variance indicated above. In simulations run by Bartels against serially correlated alternatives, this test was more sensitive that the runs up and down test. Reference: Bartels [Bartels, 1982]. 14

15 4.2 Nonparametric Tests for Randomness: Trend Tests Another important deviation from randomness is a global trend. In elementary theory of time series, trends are treated simply and naïvely, as either random walks with drift or trend stationary processes. In either case there is a constant push up or down augmented by noise. But what if one wants to test for trend without restricted models of these types, as can occur when the drift is not constant or the process is nonlinear? This Section presents several nonparametric tests that can be used to detect such trends The Cox-Stuart Test Null Hypothesis: Alt. Hypothesis: Let f(x 1,..., x T ) be unknown density function for (X 1,..., X T ). H 0 : The Hypothesis of Randomness H a : There is either an excess number of positive (or negative) sequential values in the first half compared to the second. Data is split into two halves with the middle observation dropped if sample size is odd, giving equal numbers m in each half. Calling X t and Y t, t = 1, 2,..., m the respective observations in the two halves, the test statistic is t=m t=1 sgn(y t X t ). Under the null hypothesis, this statistic (known as a sign test) has a binomial distribution B(m, 1/2), i.e. with p = 0.5. When the value significantly exceeds m/2, there is evidence of an upward trend, and when it is significantly below, evidence of a downward trend. Simplified Explanation: The Cox-Stuart Test is a simple but not very powerful test. It compares the ranks of two halves of a sample using only the signs of pairwise equential ifferences. It s easy to calculate, but not appropriate if one wants to determine the degree of monotonicity in the data. Reference: Cox & Stuart [D. R. Cox, 1955] The Difference Sign Test Null Hypothesis: Let f(x 1,..., x T ) be unknown density function for(x 1,..., X T ). H 0 : The Hypothesis of Randomness Null Hypothesis: H 0 : (X 1,..., X T ) = D (X σ(1),..., X σ(t ) ), for all σ P T. Alt. Hypothesis: H aa : P [X t > X t 1 ] > 1/2 for all t or H ad : P [X t > X t 1 ] < 1/2 for all t. 15

16 This test counts the number of positive differences among sample data and performs a sign test. Under the null hypothesis, the number of positive differences X t X t 1 will be distributed as a binomial B(T 1, 1/2). Simplified Explanation: The test statistic is the number of positive differences which has under the null hypothesis a binomial distribution B(T 1, 1/2). To be effective, it needs a large sample. Reference: Moore & Wallis [Geoffrey H. Moore, 1943] The Mann Test for Trend The data have joint density f(x 1, x 2,..., x T ). Null Hypothesis: H 0 : The Hypothesis of Randomness. Alt. Hypothesis: H a : There is a monotone trend, P [X t2 > X t1 ] for t 2 > t 1. The original Mann test is S = t=t 1 t=1 s=t s=t+1 sgn(x s X t ). Under the null Hypothesis of Randomness, E[S] = 0 and the variance is p=g V ar(s) = [T (T 1)(2T + 5) t p (t p 1)(2t p + 5)]/18, where the number of groups of observations that are tied is g and the number of ties in each group p is t p. Asymptotically, Z = p=1 S V ar(s) has a N(0, 1) distribution. In small samples of T < 30 the test S 1 S > 0 V ar(s) Z = 0 S = 0 S+1 S < 0 V ar(s) which applies a continuity correction, is recommended. Simplified Explanation: When the Mann Test is positive and large, there is evidence of an upward trend; when highly negative, a downward trend. The two previous tests, the Cox-Stuart and the Difference Sign Test as well as this one all tests for monotone, not necessarily linear, trends. Of these three tests, the Mann is definitely the best. Moreover, as we see below in the Dietz- Killeen multivariate generalization of the Mann Test, short term correlations can be partially removed, allowing the dectection of an underlying trend not due to those correlations. Reference: Hirsh & Slack [Hirsh and Slack, 1984], Kendall [Kendall, 1975], Mann [Mann, 1945]. 16

17 4.2.4 The Dietz-Kileen Multivariate Test for Trend Null Hypothesis: Alt. Hypothesis: The p-variate time series X t = (X t1, X t2,..., X tp has unknown joint density f(x 1, x 2,..., x p ). H 0 : The Hypothesis of Randomness H a : There is a monotone trend in one or more of the p variables. The test statistic is complicated to describe; the interested reader is referred to the cited references. Under H 0 and full rank of a covariance matrix, the test statistic will be distributed as a χ 2 (p) distribution. Simplified Explanation: When the Test is positive and large, one rejects the hypothesis of no trend in any variable. Directionality is not measured some variables can have ascending trends, some descending and rejection will occur anyway. This test is useful in trading systems testing, as follows. One forms a multivariate series by taking every two or three trades and calling that vector an observation. Then perform the Dietz-Kileen procedure tests for monotone trends adjusted for serial or three-fold correlation. Under tests of this method by Hirsh and Slack [Hirsh and Slack, 1984], it proved quite effective if the serial correlation does not exceed 0.6. Thus the Dietz-Kileen test can address a common question about trading systems is serial correlation the cause of a trend or not? Reference: Dietz & Kileen [E. Jacquelin Dietz, 1981], [Hirsh et al., 1982], Mann [Mann, 1945]. 5 Nonparametric Tests for Comparing Distributions Because returns distributions have been shown to have systematic departures from normality, it is more appropriate to test i.i.d., white noise and random walk hypotheses using nonparametric or semiparametric methods. Below are several of the more common tests of this type. 5.1 A Comparison of Distributions Test: The two-sample test of Kolmogorov-Smirnov (KS test) One univariate random sample from continous distribution F 1 (x), the other from continuous distribution F 2 (x). Null Hypothesis: H 0 : F 1 (x) = F 2 (x) for all x R. Alt. Hypothesis: H a : F 1 (x) F 2 (x) > 0 for some x R. The test statistic uses the empirical distribution functions for the two samples. An empirical distribution ˆF (x) for a univariate sample {X i } i=1 i=n is defined 17

18 as ˆF (x) = # of X is in the sample that are x. (12) n The KS test statistic is just the largest absolute difference between the empirical distribution functions. KS = max x R ˆF 1 (x) ˆF 2 (x) (13) The strength of the KS test is that its asymptotic distribution does not depend on the hypothesized common distribution function. We omit discussion of the Cramer-von Mises [Anderson, 1962] and Anderson-Darling [Scholz and Stephens, 1987] tests, which are similar the KS test, but have power against different alternatives. Simplified Explanation: This test detects a worst case difference between two distribution functions. When sample sizes are large, it is quite sensitive to small differences in two distributions. Like the Kruskal-Wallis test and other tests that compare distributions, however, it is questionable for time series because the null hypothesis ignores the sample s original order. Reference: Hollander & Wolfe [Hollander and Wolfe, 1999]. 5.2 Subset Location Tests: Wilcoxon Rank-Sum (k = 2) and Kruskal-Wallis (k 2) Location Tests Assuming that there are k groups of approximately the same size, where the i th group has indices m i to m i, 0 = m 0 < m 1 < < m k = T For each i, j = 1, 2,..., k, define A ij = P [X i > X j ] P [X i = X j ] and for each i = 1, 2,..., k, let n i = m i m i 1 be the size of the i th group. Null Hypothesis: H 0 : For each i = 1, 2,..., k, j=k n j j=1 T A i,j = 0.5. Alt. Hypothesis: H a : For at least one i = 1, 2,..., k, j=k j=1 n j T A i,j 0.5. This rather convoluted statement of the null and alternative hypotheses can be reduced to the statement that the null hypothesis H 0 must be invariant under the conversion of returns to ranks. A weaker condition which ensures that this null hypothesis holds is 18

19 Let A i be the indices of group i. For each j A 1, let X j have the form X j = µ + e j, and for each i > 1 and each j A i, let X j = µ + i + e j. Assume that e l for l = 1, 2,..., T are a random sample from a continuous distribution. Null Hypothesis: H0 : 2 = 3 = = k = 0. Alt. Hypothesis: Ha : At least one i 0 (one-sided test is available for k = 2). The model assumes that each group s distribution is merely shifted from the others; the null hypothesis H 0 hypothesis is then true if all shifts are 0. Historically, the Wilcoxon test (k = 2) was proposed first and was generalized later to several groups. Thus 1. For k = 2, the test is known as the Wilcoxon Rank-Sum Test. 2. For k > 2, the test is known as the Kruskal-Wallis Test. The Wilcoxon and Kruskal-Wallis tests convert the combined samples to ranks and then applies a small or large-sample ANOVA to the groups of ranks. Simplified Explanation: The Kruskal-Wallis test is identical to the Wilcoxon Rank-Sum test when k = 2. Most users of these tests consider that they compare the medians of the distributions of the k groups and the null hypothesis will be rejected when there is evidence that not all of these medians are equal. The Kruskal-Wallis test, however, is not powerful against most differences that exist in returns data for two reasons: (1) it converts the data to ranks, which neglects their magnitudes, and (2) it ignores the time sequence of the data so that serial dependencies are ignored. Reference: Vargha & Delaney [Vargha and Delaney, 1998]; Hollander & Wolfe [Hollander and Wolfe, 1999]. NOTE: The following three tests, the Ansari-Bradley, Fligner-Kileen and Mood, all have the same null hypothesis for testing the scale of k groups, but have different test statistics. However, the default R-language implementations of the Ansari-Bradley and Mood tests supports k = 2 only, while the Fligner- Killeen test supports k 2. For simplicity, all null hypotheses are stated for two groups only. 19

20 5.3 The Ansari-Bradley Test of Equality of Variances Let F be an unknown continuous distribution function, let {X 1i } n1 1 be a random sample with d.f. F (x 1 m) and let {X 2i } n2 1 be an independent r.s. with d.f. F ((x 2 m)/s), where m is a common unknown location for the two distributions and where s is an unknown scale parameter. Null Hypothesis: H 0 : s = 1. Alt. Hypothesis: H a : s 1, or the distributions are not representable as described in the model. The test statistic ranks the combined N = n 1 + n 2 observations from the n 1 of X 1 and the n 2 of X 2 and assigns a score of a i = (N + 1)/2 i N (14) to the i th in the ranking. Thus the first and last have the smallest scores of 1, the next smallest and next largest, scores of 2 and so on. Designating the indices in the ranking corresponding to the X 1 sample as i 1, i 2,..., i n1, the test statistic is Z = j=n 1 j=1 a ij. (15) Under the null hypothesis, the distribution of Z is described in [Hollander and Wolfe, 1999]; for small samples, a formula is available and for large samples a normal approximation is available. It is advisable to calculate medians for each sample separately and to subtact them from their sample prior to performing the test. Note that this test, unlike the parametric F -test, does not require that means or variances of either sample exist. Note also that the R-language implementation of the Ansari-Bradley test ansari.test does not median-correct the data, and that correction should be performed prior to the test. Simplified Explanation: The Ansari-Bradley test relies on the fact that under the null distribution, all permutations of data are equally likely. On the other hand, if one sample has the same location but a different variance, then more of its scores will be near the extremes of the ranks, or more toward the middle of the ranks. In these cases, (15) will be either too large, or too small, respectively, compared to a typical case under the null hypothesis. Before applying this test, each sample s median should be subtracted from its observations. A word should be added contrasting this test with the Fligner-Kileen and Mood tests. The difference among these tests is the method of scoring; 7 in all other formal respects they are the same. Among the three scale tests, this one has the smallest weights for large observations, so that unlike the others, it is less sensitive to 7 Scores are the values assigned to the observations and then summed up to form the test statistic in the Ansari-Bradley, they are the a i of formula (14). 20

21 very large observations, but more sensitive to very small observations. As such, it should not perform as well on market data that has outliers. Like the Kruskal- Wallis test and other distributional comparison tests, it is questionable for time series because the null hypothesis ignores the sample s original order. Reference: Hollander & Wolfe [Hollander and Wolfe, 1999]. 5.4 The Fligner-Kileen Test of Scale Null Hypothesis: H 0 : s = 1. Alt. Hypothesis: Let F be an unknown continuous distribution function, random sample X 1 have distribution F (x 1 m) and independent random sample X 2 have distribution F ((x 2 m)/s), where m is a common unknown location for the distributions and s is an unknown scale parameter. H a : s 1, or the distributions are not representable as described in the model. Let N = n 1 +n 2 be the total number of observations, where n 1 is the number in the X 1 sample and n 2 is the number in the X 2 sample. The Fligner-Kileen test assigns scores a ij to observations x ij as follows: (1) observations in each group are centered separately using means or medians, (2) the centered data are combined into one sample and ranked from lowest to highest, yielding ranks r ij, 8 and (3) the score for the j th observation in the i th group, i = 1, 2 is assigned the score ( ) 1 a ij = Φ r ij (16) 2(N + 1) where Φ is the standard normal distribution and Φ 1 is its inverse, i.e. maps probabilites to quantiles. Scores are then assigned as in steps (2) and (3) above and the test statistic where χ 2 1 = n 1(Ā1 ā) + n 2 (Ā2 ā) V 2, (17) Ā i = 1 a ij, n i j ā = (n 1 A 1 + n 2 A 2 )/N, and i=2 V 2 i=1 j = (a ij ā) 2. N 1 As the notation indicates, χ 2 1 is distributed as a χ 2 distribution with 1 degree of freedom. Simplified Explanation: The Fligner-Kileen test relies on the fact that under the null distribution, all permutations of data are equally likely. If one sample 8 For the handling of ties, see the references below. 21

22 has the same location but a different variance, then more of its scores will be near the extremes of the ranks, or more toward the middle of the ranks. In either case the χ 2 1 statistic will be either too large or too small compared to typical ones under the null hypothesis. By default, the R-language implementation centers the data with medians. A word should be added contrasting this test with the Ansari-Bradley and Mood tests. The difference among these tests is the method of scoring; in all other formal respects they are the same. This test has the relatively largest weights for extreme observations, so that among the others it should be most sensitive to extreme observations; as such, it should perform better than the others on a sample with more outliers. Like the Kruskal-Wallis test and other distributional comparison tests, however, it is questionable for time series because the null hypothesis ignores the sample s original order. Reference: Conover, Johnson & Johnson [Conover et al., 1981]. 5.5 The Mood Test of Scale Null Hypothesis: H 0 : s = 1. Alt. Hypothesis: Let F be an unknown continuous distribution function, random sample X 1 have distribution F (x 1 m and independent random sample X 2 have distribution F ((x 2 m)/s), where m is a common unknown location for the distributions and s is an unknown scale parameter. H a : s 1, or the distributions are not representable as described in the model. The Mood test has the form (15) where r ij is the rank of the (ij) th observation, i = 1, 2, j = 1, 2,..., n i, and with scores ( a ij = r ij N + 1 ) 2 (18) 2 and ā i = j=n i j=1 a ij. In other words, it differs from the Ansari-Bradley test only in the form of its scores a ij. The critical values under the null hypothesis are given by a formula in small samples, and by a normal approximation in large samples. Simplified Explanation: The test statistic is based on the fact that under the null distribution, all permutations of data are equally likely. If one sample has the same location but a different variance, then more of its scores will be near the extremes of the ranks, or more toward the middle of the ranks. A word should be added contrasting this test with the Ansari-Bradley and Fligner- Kileen tests. The difference among these tests is the method of scoring; in all other formal respects they are the same. On large observations, this test has larger scores than Ansari-Bradley, but less than Fligner-Kileen, so that it will have intermediate sensitivity to outliers. By default, the R-language implementation does not center the data, so this should be done before analysis. Like the Kruskal-Wallis test and other distributional comparison tests, however, 22

23 it is questionable for time series because the null hypothesis ignores the sample s original order. Reference: Conover, Johnson & Jonhnson [Conover et al., 1981]. 6 Time Series Tests 6.1 General Tests (G.1) The Brock-Dechert-Schenkman (BDS) i.i.d. Test Null Hypothesis: Alt Hypothesis: Arbitrary data {X t } (of infinite extent above and below) H 0 : X i are i.i.d. from continuous distribution F (x). H a : X i are from a continuous distributions but not i.i.d.. This test is complicated to describe, but the basic idea is that an i.i.d. series X 1, X 2,..., X T has the property that P [ X t X s < ɛ] 2 = P [ X t X s < ɛ]p [ X t 1 X s 1 < ɛ] (19) for all s, t and ɛ > 0. Thus lagged pairs (X t, X t 1 ) must satisfy this relationship, and by extending the above relationship to sequences of length m, one can obtain analogous relationships among m-length sequences. The test itself uses a statistic which is an estimate of the correlation dimension of the series for an account of how the BDS is developed, see references on chaos or the original articles cited below. The BDS test is one of the few tests for i.i.d. robust against nonlinear and chaotic time series. It is valid for the standardized residuals of all ARIMA, ARCH and GARCH-family models provided the sample size is sufficient (> 500). Simplified Explanation: The BDS test is one of the few i.i.d. tests that is robust against nonlinear and chaotic time series. In general, the BDS test is appropriate for ARIMA and ARCH standardized residuals, but not for small samples from a GARCH model. Reference: Brock, Dechert & Scheinkman [Brock et al., 1996]; LeBaron [LeBaron, 1997], de Lima [de Lima, 1996]. (G.2) The KPSS Test of Stationarity 23

24 Null Hypothesis: Alt Hypothesis: X t = βt + Y t + U t, where U t is i.i.d white noise, Y t = Y t 1 + V t and X t is an observed series, β is a constant, V t is i.i.d. and V t N(0, σv). 2 H0 1 : σv 2 = 0 for trend stationarity and H0 2 : β = 0 for level stationarity. Ha 1 : σv 2 > 0 for stationarity (the series has a unit root.) Ha 2 : β 0 for level stationarity. Trend stationarity refers to a series with β 0, and level stationarity to one with β = 0. The KPSS null hypothesis of stationarity stipulates (1) that there is no random walk term Y t = Y t 1 +V t embedded in the observed series X t, and (2) that the time series has the same probability generating mechanism at all time periods. When a random walk term is present, the series will have a unit root, and therefore, will not be stationary. This test is a one-sided, right-tailed test, so rejection of occurs when the statistic is large. The level stationarity null hypothesis is satisfied when the long term trend of the series is zero. In this case, the test is two-sided, being positive when the series is increasing and negative when it is decreasing. Simplified Explanation: The KPSS test is typically used to detect the existence of an embedded random walk in the series, guaranteeing that ARMA models are not appropriate without a data transformation such as differencing. And if a series is not level stationary, then it is certainly not i.i.d.. Thus rejection of the stationary null hypothesis means that the series is not i.i.d.. The problem with this test is that there are other deviations from i.i.d. besides an embedded random walk, and this test may not be powerful against such alternatives. Reference: Kwiatkowski, Phillips and Schmidt [Kwiatkowski et al., 1991]. (G.3) The Terasvirta Neural Network Test For Neglected Nonlinearity This test is too complicated to described in detail here. The idea is to build a neural network with hidden processing units and to use these to test that the data have non-zero linearity. See the Terasvirta reference for details. Simplified Explanation: The Terasvirta tests the null hypothesis of i.i.d. against the alternative that there is nonlinear dependence of X t on X t 1, X t 2,.... If the null hypothesis is rejected, then there is evidence that the series not only fails to be i.i.d., but the relationship is not linear. Reference: Terasvirta, Lin & Granger [Tersvirta et al., 1993]. (G.4) The White Test for Neglected Nonlinearity 24

A nonparametric test for seasonal unit roots

A nonparametric test for seasonal unit roots Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna To be presented in Innsbruck November 7, 2007 Abstract We consider a nonparametric test for the

More information

Glossary. The ISI glossary of statistical terms provides definitions in a number of different languages:

Glossary. The ISI glossary of statistical terms provides definitions in a number of different languages: Glossary The ISI glossary of statistical terms provides definitions in a number of different languages: http://isi.cbs.nl/glossary/index.htm Adjusted r 2 Adjusted R squared measures the proportion of the

More information

BCT Lecture 3. Lukas Vacha.

BCT Lecture 3. Lukas Vacha. BCT Lecture 3 Lukas Vacha vachal@utia.cas.cz Stationarity and Unit Root Testing Why do we need to test for Non-Stationarity? The stationarity or otherwise of a series can strongly influence its behaviour

More information

Recall the Basics of Hypothesis Testing

Recall the Basics of Hypothesis Testing Recall the Basics of Hypothesis Testing The level of significance α, (size of test) is defined as the probability of X falling in w (rejecting H 0 ) when H 0 is true: P(X w H 0 ) = α. H 0 TRUE H 1 TRUE

More information

Unit 14: Nonparametric Statistical Methods

Unit 14: Nonparametric Statistical Methods Unit 14: Nonparametric Statistical Methods Statistics 571: Statistical Methods Ramón V. León 8/8/2003 Unit 14 - Stat 571 - Ramón V. León 1 Introductory Remarks Most methods studied so far have been based

More information

Multivariate Time Series: VAR(p) Processes and Models

Multivariate Time Series: VAR(p) Processes and Models Multivariate Time Series: VAR(p) Processes and Models A VAR(p) model, for p > 0 is X t = φ 0 + Φ 1 X t 1 + + Φ p X t p + A t, where X t, φ 0, and X t i are k-vectors, Φ 1,..., Φ p are k k matrices, with

More information

THE ROYAL STATISTICAL SOCIETY HIGHER CERTIFICATE

THE ROYAL STATISTICAL SOCIETY HIGHER CERTIFICATE THE ROYAL STATISTICAL SOCIETY 004 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE PAPER II STATISTICAL METHODS The Society provides these solutions to assist candidates preparing for the examinations in future

More information

Volatility. Gerald P. Dwyer. February Clemson University

Volatility. Gerald P. Dwyer. February Clemson University Volatility Gerald P. Dwyer Clemson University February 2016 Outline 1 Volatility Characteristics of Time Series Heteroskedasticity Simpler Estimation Strategies Exponentially Weighted Moving Average Use

More information

Statistical Procedures for Testing Homogeneity of Water Quality Parameters

Statistical Procedures for Testing Homogeneity of Water Quality Parameters Statistical Procedures for ing Homogeneity of Water Quality Parameters Xu-Feng Niu Professor of Statistics Department of Statistics Florida State University Tallahassee, FL 3306 May-September 004 1. Nonparametric

More information

Nonlinear Time Series

Nonlinear Time Series Nonlinear Time Series Recall that a linear time series {X t } is one that follows the relation, X t = µ + i=0 ψ i A t i, where {A t } is iid with mean 0 and finite variance. A linear time series is stationary

More information

If we want to analyze experimental or simulated data we might encounter the following tasks:

If we want to analyze experimental or simulated data we might encounter the following tasks: Chapter 1 Introduction If we want to analyze experimental or simulated data we might encounter the following tasks: Characterization of the source of the signal and diagnosis Studying dependencies Prediction

More information

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis Introduction to Time Series Analysis 1 Contents: I. Basics of Time Series Analysis... 4 I.1 Stationarity... 5 I.2 Autocorrelation Function... 9 I.3 Partial Autocorrelation Function (PACF)... 14 I.4 Transformation

More information

Nonparametric Statistics. Leah Wright, Tyler Ross, Taylor Brown

Nonparametric Statistics. Leah Wright, Tyler Ross, Taylor Brown Nonparametric Statistics Leah Wright, Tyler Ross, Taylor Brown Before we get to nonparametric statistics, what are parametric statistics? These statistics estimate and test population means, while holding

More information

Statistical Hypothesis Testing with SAS and R

Statistical Hypothesis Testing with SAS and R Statistical Hypothesis Testing with SAS and R Statistical Hypothesis Testing with SAS and R Dirk Taeger Institute for Prevention and Occupational Medicine of the German Social Accident Insurance, Institute

More information

AN EMPIRICAL LIKELIHOOD RATIO TEST FOR NORMALITY

AN EMPIRICAL LIKELIHOOD RATIO TEST FOR NORMALITY Econometrics Working Paper EWP0401 ISSN 1485-6441 Department of Economics AN EMPIRICAL LIKELIHOOD RATIO TEST FOR NORMALITY Lauren Bin Dong & David E. A. Giles Department of Economics, University of Victoria

More information

5.5.3 Statistical Innovative Trend Test Application Crossing Trend Analysis Methodology Rational Concept...

5.5.3 Statistical Innovative Trend Test Application Crossing Trend Analysis Methodology Rational Concept... Contents 1 Introduction.... 1 1.1 General... 1 1.2 Trend Definition and Analysis... 3 1.2.1 Conceptual and Visual Trends.... 4 1.2.2 Mathematical Trend.... 7 1.2.3 Statistical Trend.... 9 1.3 Trend in

More information

One-Sample Numerical Data

One-Sample Numerical Data One-Sample Numerical Data quantiles, boxplot, histogram, bootstrap confidence intervals, goodness-of-fit tests University of California, San Diego Instructor: Ery Arias-Castro http://math.ucsd.edu/~eariasca/teaching.html

More information

Topic 4 Unit Roots. Gerald P. Dwyer. February Clemson University

Topic 4 Unit Roots. Gerald P. Dwyer. February Clemson University Topic 4 Unit Roots Gerald P. Dwyer Clemson University February 2016 Outline 1 Unit Roots Introduction Trend and Difference Stationary Autocorrelations of Series That Have Deterministic or Stochastic Trends

More information

PSY 307 Statistics for the Behavioral Sciences. Chapter 20 Tests for Ranked Data, Choosing Statistical Tests

PSY 307 Statistics for the Behavioral Sciences. Chapter 20 Tests for Ranked Data, Choosing Statistical Tests PSY 307 Statistics for the Behavioral Sciences Chapter 20 Tests for Ranked Data, Choosing Statistical Tests What To Do with Non-normal Distributions Tranformations (pg 382): The shape of the distribution

More information

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing 1 In most statistics problems, we assume that the data have been generated from some unknown probability distribution. We desire

More information

Nonstationary Time Series:

Nonstationary Time Series: Nonstationary Time Series: Unit Roots Egon Zakrajšek Division of Monetary Affairs Federal Reserve Board Summer School in Financial Mathematics Faculty of Mathematics & Physics University of Ljubljana September

More information

Textbook Examples of. SPSS Procedure

Textbook Examples of. SPSS Procedure Textbook s of IBM SPSS Procedures Each SPSS procedure listed below has its own section in the textbook. These sections include a purpose statement that describes the statistical test, identification of

More information

Research Article A Nonparametric Two-Sample Wald Test of Equality of Variances

Research Article A Nonparametric Two-Sample Wald Test of Equality of Variances Advances in Decision Sciences Volume 211, Article ID 74858, 8 pages doi:1.1155/211/74858 Research Article A Nonparametric Two-Sample Wald Test of Equality of Variances David Allingham 1 andj.c.w.rayner

More information

Econ 423 Lecture Notes: Additional Topics in Time Series 1

Econ 423 Lecture Notes: Additional Topics in Time Series 1 Econ 423 Lecture Notes: Additional Topics in Time Series 1 John C. Chao April 25, 2017 1 These notes are based in large part on Chapter 16 of Stock and Watson (2011). They are for instructional purposes

More information

CHAPTER 17 CHI-SQUARE AND OTHER NONPARAMETRIC TESTS FROM: PAGANO, R. R. (2007)

CHAPTER 17 CHI-SQUARE AND OTHER NONPARAMETRIC TESTS FROM: PAGANO, R. R. (2007) FROM: PAGANO, R. R. (007) I. INTRODUCTION: DISTINCTION BETWEEN PARAMETRIC AND NON-PARAMETRIC TESTS Statistical inference tests are often classified as to whether they are parametric or nonparametric Parameter

More information

Lecture 5: Unit Roots, Cointegration and Error Correction Models The Spurious Regression Problem

Lecture 5: Unit Roots, Cointegration and Error Correction Models The Spurious Regression Problem Lecture 5: Unit Roots, Cointegration and Error Correction Models The Spurious Regression Problem Prof. Massimo Guidolin 20192 Financial Econometrics Winter/Spring 2018 Overview Stochastic vs. deterministic

More information

SAS/STAT 14.1 User s Guide. Introduction to Nonparametric Analysis

SAS/STAT 14.1 User s Guide. Introduction to Nonparametric Analysis SAS/STAT 14.1 User s Guide Introduction to Nonparametric Analysis This document is an individual chapter from SAS/STAT 14.1 User s Guide. The correct bibliographic citation for this manual is as follows:

More information

Financial Econometrics and Quantitative Risk Managenent Return Properties

Financial Econometrics and Quantitative Risk Managenent Return Properties Financial Econometrics and Quantitative Risk Managenent Return Properties Eric Zivot Updated: April 1, 2013 Lecture Outline Course introduction Return definitions Empirical properties of returns Reading

More information

Hypothesis testing:power, test statistic CMS:

Hypothesis testing:power, test statistic CMS: Hypothesis testing:power, test statistic The more sensitive the test, the better it can discriminate between the null and the alternative hypothesis, quantitatively, maximal power In order to achieve this

More information

Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8]

Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8] 1 Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8] Insights: Price movements in one market can spread easily and instantly to another market [economic globalization and internet

More information

Application of Variance Homogeneity Tests Under Violation of Normality Assumption

Application of Variance Homogeneity Tests Under Violation of Normality Assumption Application of Variance Homogeneity Tests Under Violation of Normality Assumption Alisa A. Gorbunova, Boris Yu. Lemeshko Novosibirsk State Technical University Novosibirsk, Russia e-mail: gorbunova.alisa@gmail.com

More information

* Tuesday 17 January :30-16:30 (2 hours) Recored on ESSE3 General introduction to the course.

* Tuesday 17 January :30-16:30 (2 hours) Recored on ESSE3 General introduction to the course. Name of the course Statistical methods and data analysis Audience The course is intended for students of the first or second year of the Graduate School in Materials Engineering. The aim of the course

More information

Analyzing Small Sample Experimental Data

Analyzing Small Sample Experimental Data Analyzing Small Sample Experimental Data Session 2: Non-parametric tests and estimators I Dominik Duell (University of Essex) July 15, 2017 Pick an appropriate (non-parametric) statistic 1. Intro to non-parametric

More information

A comparison study of the nonparametric tests based on the empirical distributions

A comparison study of the nonparametric tests based on the empirical distributions 통계연구 (2015), 제 20 권제 3 호, 1-12 A comparison study of the nonparametric tests based on the empirical distributions Hyo-Il Park 1) Abstract In this study, we propose a nonparametric test based on the empirical

More information

A Monte Carlo Simulation of the Robust Rank- Order Test Under Various Population Symmetry Conditions

A Monte Carlo Simulation of the Robust Rank- Order Test Under Various Population Symmetry Conditions Journal of Modern Applied Statistical Methods Volume 12 Issue 1 Article 7 5-1-2013 A Monte Carlo Simulation of the Robust Rank- Order Test Under Various Population Symmetry Conditions William T. Mickelson

More information

Nonlinear Characterization of Activity Dynamics in Online Collaboration Websites

Nonlinear Characterization of Activity Dynamics in Online Collaboration Websites Nonlinear Characterization of Activity Dynamics in Online Collaboration Websites Tiago Santos 1 Simon Walk 2 Denis Helic 3 1 Know-Center, Graz, Austria 2 Stanford University 3 Graz University of Technology

More information

Lecture 6a: Unit Root and ARIMA Models

Lecture 6a: Unit Root and ARIMA Models Lecture 6a: Unit Root and ARIMA Models 1 2 Big Picture A time series is non-stationary if it contains a unit root unit root nonstationary The reverse is not true. For example, y t = cos(t) + u t has no

More information

Introduction and Descriptive Statistics p. 1 Introduction to Statistics p. 3 Statistics, Science, and Observations p. 5 Populations and Samples p.

Introduction and Descriptive Statistics p. 1 Introduction to Statistics p. 3 Statistics, Science, and Observations p. 5 Populations and Samples p. Preface p. xi Introduction and Descriptive Statistics p. 1 Introduction to Statistics p. 3 Statistics, Science, and Observations p. 5 Populations and Samples p. 6 The Scientific Method and the Design of

More information

Solutions exercises of Chapter 7

Solutions exercises of Chapter 7 Solutions exercises of Chapter 7 Exercise 1 a. These are paired samples: each pair of half plates will have about the same level of corrosion, so the result of polishing by the two brands of polish are

More information

Modeling Hydrologic Chanae

Modeling Hydrologic Chanae Modeling Hydrologic Chanae Statistical Methods Richard H. McCuen Department of Civil and Environmental Engineering University of Maryland m LEWIS PUBLISHERS A CRC Press Company Boca Raton London New York

More information

Contents. Acknowledgments. xix

Contents. Acknowledgments. xix Table of Preface Acknowledgments page xv xix 1 Introduction 1 The Role of the Computer in Data Analysis 1 Statistics: Descriptive and Inferential 2 Variables and Constants 3 The Measurement of Variables

More information

Introduction to Nonparametric Analysis (Chapter)

Introduction to Nonparametric Analysis (Chapter) SAS/STAT 9.3 User s Guide Introduction to Nonparametric Analysis (Chapter) SAS Documentation This document is an individual chapter from SAS/STAT 9.3 User s Guide. The correct bibliographic citation for

More information

APPLIED TIME SERIES ECONOMETRICS

APPLIED TIME SERIES ECONOMETRICS APPLIED TIME SERIES ECONOMETRICS Edited by HELMUT LÜTKEPOHL European University Institute, Florence MARKUS KRÄTZIG Humboldt University, Berlin CAMBRIDGE UNIVERSITY PRESS Contents Preface Notation and Abbreviations

More information

Lecture Slides. Elementary Statistics. by Mario F. Triola. and the Triola Statistics Series

Lecture Slides. Elementary Statistics. by Mario F. Triola. and the Triola Statistics Series Lecture Slides Elementary Statistics Tenth Edition and the Triola Statistics Series by Mario F. Triola Slide 1 Chapter 13 Nonparametric Statistics 13-1 Overview 13-2 Sign Test 13-3 Wilcoxon Signed-Ranks

More information

Lecture Slides. Section 13-1 Overview. Elementary Statistics Tenth Edition. Chapter 13 Nonparametric Statistics. by Mario F.

Lecture Slides. Section 13-1 Overview. Elementary Statistics Tenth Edition. Chapter 13 Nonparametric Statistics. by Mario F. Lecture Slides Elementary Statistics Tenth Edition and the Triola Statistics Series by Mario F. Triola Slide 1 Chapter 13 Nonparametric Statistics 13-1 Overview 13-2 Sign Test 13-3 Wilcoxon Signed-Ranks

More information

Physics 509: Non-Parametric Statistics and Correlation Testing

Physics 509: Non-Parametric Statistics and Correlation Testing Physics 509: Non-Parametric Statistics and Correlation Testing Scott Oser Lecture #19 Physics 509 1 What is non-parametric statistics? Non-parametric statistics is the application of statistical tests

More information

Chapte The McGraw-Hill Companies, Inc. All rights reserved.

Chapte The McGraw-Hill Companies, Inc. All rights reserved. er15 Chapte Chi-Square Tests d Chi-Square Tests for -Fit Uniform Goodness- Poisson Goodness- Goodness- ECDF Tests (Optional) Contingency Tables A contingency table is a cross-tabulation of n paired observations

More information

Distribution-Free Procedures (Devore Chapter Fifteen)

Distribution-Free Procedures (Devore Chapter Fifteen) Distribution-Free Procedures (Devore Chapter Fifteen) MATH-5-01: Probability and Statistics II Spring 018 Contents 1 Nonparametric Hypothesis Tests 1 1.1 The Wilcoxon Rank Sum Test........... 1 1. Normal

More information

Testing for non-stationarity

Testing for non-stationarity 20 November, 2009 Overview The tests for investigating the non-stationary of a time series falls into four types: 1 Check the null that there is a unit root against stationarity. Within these, there are

More information

Empirical Power of Four Statistical Tests in One Way Layout

Empirical Power of Four Statistical Tests in One Way Layout International Mathematical Forum, Vol. 9, 2014, no. 28, 1347-1356 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/imf.2014.47128 Empirical Power of Four Statistical Tests in One Way Layout Lorenzo

More information

Asymptotic Statistics-VI. Changliang Zou

Asymptotic Statistics-VI. Changliang Zou Asymptotic Statistics-VI Changliang Zou Kolmogorov-Smirnov distance Example (Kolmogorov-Smirnov confidence intervals) We know given α (0, 1), there is a well-defined d = d α,n such that, for any continuous

More information

Testing for Normality

Testing for Normality Testing for Normality For each mean and standard deviation combination a theoretical normal distribution can be determined. This distribution is based on the proportions shown below. This theoretical normal

More information

Analysis of 2x2 Cross-Over Designs using T-Tests

Analysis of 2x2 Cross-Over Designs using T-Tests Chapter 234 Analysis of 2x2 Cross-Over Designs using T-Tests Introduction This procedure analyzes data from a two-treatment, two-period (2x2) cross-over design. The response is assumed to be a continuous

More information

A nonparametric two-sample wald test of equality of variances

A nonparametric two-sample wald test of equality of variances University of Wollongong Research Online Faculty of Informatics - Papers (Archive) Faculty of Engineering and Information Sciences 211 A nonparametric two-sample wald test of equality of variances David

More information

CHAPTER 2 SIMPLE LINEAR REGRESSION

CHAPTER 2 SIMPLE LINEAR REGRESSION CHAPTER 2 SIMPLE LINEAR REGRESSION 1 Examples: 1. Amherst, MA, annual mean temperatures, 1836 1997 2. Summer mean temperatures in Mount Airy (NC) and Charleston (SC), 1948 1996 Scatterplots outliers? influential

More information

NAG Library Chapter Introduction. G08 Nonparametric Statistics

NAG Library Chapter Introduction. G08 Nonparametric Statistics NAG Library Chapter Introduction G08 Nonparametric Statistics Contents 1 Scope of the Chapter.... 2 2 Background to the Problems... 2 2.1 Parametric and Nonparametric Hypothesis Testing... 2 2.2 Types

More information

Contents 1. Contents

Contents 1. Contents Contents 1 Contents 1 One-Sample Methods 3 1.1 Parametric Methods.................... 4 1.1.1 One-sample Z-test (see Chapter 0.3.1)...... 4 1.1.2 One-sample t-test................. 6 1.1.3 Large sample

More information

Contents Kruskal-Wallis Test Friedman s Two-way Analysis of Variance by Ranks... 47

Contents Kruskal-Wallis Test Friedman s Two-way Analysis of Variance by Ranks... 47 Contents 1 Non-parametric Tests 3 1.1 Introduction....................................... 3 1.2 Advantages of Non-parametric Tests......................... 4 1.3 Disadvantages of Non-parametric Tests........................

More information

E 4160 Autumn term Lecture 9: Deterministic trends vs integrated series; Spurious regression; Dickey-Fuller distribution and test

E 4160 Autumn term Lecture 9: Deterministic trends vs integrated series; Spurious regression; Dickey-Fuller distribution and test E 4160 Autumn term 2016. Lecture 9: Deterministic trends vs integrated series; Spurious regression; Dickey-Fuller distribution and test Ragnar Nymoen Department of Economics, University of Oslo 24 October

More information

Non-Parametric Statistics: When Normal Isn t Good Enough"

Non-Parametric Statistics: When Normal Isn t Good Enough Non-Parametric Statistics: When Normal Isn t Good Enough" Professor Ron Fricker" Naval Postgraduate School" Monterey, California" 1/28/13 1 A Bit About Me" Academic credentials" Ph.D. and M.A. in Statistics,

More information

Nonparametric statistic methods. Waraphon Phimpraphai DVM, PhD Department of Veterinary Public Health

Nonparametric statistic methods. Waraphon Phimpraphai DVM, PhD Department of Veterinary Public Health Nonparametric statistic methods Waraphon Phimpraphai DVM, PhD Department of Veterinary Public Health Measurement What are the 4 levels of measurement discussed? 1. Nominal or Classificatory Scale Gender,

More information

5 Introduction to the Theory of Order Statistics and Rank Statistics

5 Introduction to the Theory of Order Statistics and Rank Statistics 5 Introduction to the Theory of Order Statistics and Rank Statistics This section will contain a summary of important definitions and theorems that will be useful for understanding the theory of order

More information

For the full text of this licence, please go to:

For the full text of this licence, please go to: This item was submitted to Loughborough s Institutional Repository by the author and is made available under the following Creative Commons Licence conditions. For the full text of this licence, please

More information

G. S. Maddala Kajal Lahiri. WILEY A John Wiley and Sons, Ltd., Publication

G. S. Maddala Kajal Lahiri. WILEY A John Wiley and Sons, Ltd., Publication G. S. Maddala Kajal Lahiri WILEY A John Wiley and Sons, Ltd., Publication TEMT Foreword Preface to the Fourth Edition xvii xix Part I Introduction and the Linear Regression Model 1 CHAPTER 1 What is Econometrics?

More information

Non-parametric tests, part A:

Non-parametric tests, part A: Two types of statistical test: Non-parametric tests, part A: Parametric tests: Based on assumption that the data have certain characteristics or "parameters": Results are only valid if (a) the data are

More information

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY & Contents PREFACE xiii 1 1.1. 1.2. Difference Equations First-Order Difference Equations 1 /?th-order Difference

More information

E 4101/5101 Lecture 9: Non-stationarity

E 4101/5101 Lecture 9: Non-stationarity E 4101/5101 Lecture 9: Non-stationarity Ragnar Nymoen 30 March 2011 Introduction I Main references: Hamilton Ch 15,16 and 17. Davidson and MacKinnon Ch 14.3 and 14.4 Also read Ch 2.4 and Ch 2.5 in Davidson

More information

Small Sample Properties of Alternative Tests for Martingale Difference Hypothesis

Small Sample Properties of Alternative Tests for Martingale Difference Hypothesis Small Sample Properties of Alternative Tests for Martingale Difference Hypothesis Amélie Charles, Olivier Darné, Jae Kim To cite this version: Amélie Charles, Olivier Darné, Jae Kim. Small Sample Properties

More information

Testing for Normality

Testing for Normality Testing for Normality For each mean and standard deviation combination a theoretical normal distribution can be determined. This distribution is based on the proportions shown below. This theoretical normal

More information

Univariate, Nonstationary Processes

Univariate, Nonstationary Processes Univariate, Nonstationary Processes Jamie Monogan University of Georgia March 20, 2018 Jamie Monogan (UGA) Univariate, Nonstationary Processes March 20, 2018 1 / 14 Objectives By the end of this meeting,

More information

When is a copula constant? A test for changing relationships

When is a copula constant? A test for changing relationships When is a copula constant? A test for changing relationships Fabio Busetti and Andrew Harvey Bank of Italy and University of Cambridge November 2007 usetti and Harvey (Bank of Italy and University of Cambridge)

More information

ON SMALL SAMPLE PROPERTIES OF PERMUTATION TESTS: INDEPENDENCE BETWEEN TWO SAMPLES

ON SMALL SAMPLE PROPERTIES OF PERMUTATION TESTS: INDEPENDENCE BETWEEN TWO SAMPLES ON SMALL SAMPLE PROPERTIES OF PERMUTATION TESTS: INDEPENDENCE BETWEEN TWO SAMPLES Hisashi Tanizaki Graduate School of Economics, Kobe University, Kobe 657-8501, Japan e-mail: tanizaki@kobe-u.ac.jp Abstract:

More information

Nearest-Neighbor Forecasts Of U.S. Interest Rates

Nearest-Neighbor Forecasts Of U.S. Interest Rates 1 Nearest-Neighbor Forecasts Of U.S. Interest Rates John Barkoulas 1 Department of Economics University of Tennessee 534 Stokely Management Center Knoxville, TN 37996 Christopher F. Baum Department of

More information

Application of Homogeneity Tests: Problems and Solution

Application of Homogeneity Tests: Problems and Solution Application of Homogeneity Tests: Problems and Solution Boris Yu. Lemeshko (B), Irina V. Veretelnikova, Stanislav B. Lemeshko, and Alena Yu. Novikova Novosibirsk State Technical University, Novosibirsk,

More information

Research Article The Laplace Likelihood Ratio Test for Heteroscedasticity

Research Article The Laplace Likelihood Ratio Test for Heteroscedasticity International Mathematics and Mathematical Sciences Volume 2011, Article ID 249564, 7 pages doi:10.1155/2011/249564 Research Article The Laplace Likelihood Ratio Test for Heteroscedasticity J. Martin van

More information

Stochastic Processes

Stochastic Processes Stochastic Processes Stochastic Process Non Formal Definition: Non formal: A stochastic process (random process) is the opposite of a deterministic process such as one defined by a differential equation.

More information

Inferential Statistics

Inferential Statistics Inferential Statistics Eva Riccomagno, Maria Piera Rogantin DIMA Università di Genova riccomagno@dima.unige.it rogantin@dima.unige.it Part G Distribution free hypothesis tests 1. Classical and distribution-free

More information

TESTING FOR NORMALITY IN THE LINEAR REGRESSION MODEL: AN EMPIRICAL LIKELIHOOD RATIO TEST

TESTING FOR NORMALITY IN THE LINEAR REGRESSION MODEL: AN EMPIRICAL LIKELIHOOD RATIO TEST Econometrics Working Paper EWP0402 ISSN 1485-6441 Department of Economics TESTING FOR NORMALITY IN THE LINEAR REGRESSION MODEL: AN EMPIRICAL LIKELIHOOD RATIO TEST Lauren Bin Dong & David E. A. Giles Department

More information

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley Time Series Models and Inference James L. Powell Department of Economics University of California, Berkeley Overview In contrast to the classical linear regression model, in which the components of the

More information

Dr. Maddah ENMG 617 EM Statistics 10/12/12. Nonparametric Statistics (Chapter 16, Hines)

Dr. Maddah ENMG 617 EM Statistics 10/12/12. Nonparametric Statistics (Chapter 16, Hines) Dr. Maddah ENMG 617 EM Statistics 10/12/12 Nonparametric Statistics (Chapter 16, Hines) Introduction Most of the hypothesis testing presented so far assumes normally distributed data. These approaches

More information

H 2 : otherwise. that is simply the proportion of the sample points below level x. For any fixed point x the law of large numbers gives that

H 2 : otherwise. that is simply the proportion of the sample points below level x. For any fixed point x the law of large numbers gives that Lecture 28 28.1 Kolmogorov-Smirnov test. Suppose that we have an i.i.d. sample X 1,..., X n with some unknown distribution and we would like to test the hypothesis that is equal to a particular distribution

More information

Does k-th Moment Exist?

Does k-th Moment Exist? Does k-th Moment Exist? Hitomi, K. 1 and Y. Nishiyama 2 1 Kyoto Institute of Technology, Japan 2 Institute of Economic Research, Kyoto University, Japan Email: hitomi@kit.ac.jp Keywords: Existence of moments,

More information

Non-parametric Tests for Complete Data

Non-parametric Tests for Complete Data Non-parametric Tests for Complete Data Non-parametric Tests for Complete Data Vilijandas Bagdonavičius Julius Kruopis Mikhail S. Nikulin First published 2011 in Great Britain and the United States by

More information

Diagnostic Test for GARCH Models Based on Absolute Residual Autocorrelations

Diagnostic Test for GARCH Models Based on Absolute Residual Autocorrelations Diagnostic Test for GARCH Models Based on Absolute Residual Autocorrelations Farhat Iqbal Department of Statistics, University of Balochistan Quetta-Pakistan farhatiqb@gmail.com Abstract In this paper

More information

Multivariate Regression

Multivariate Regression Multivariate Regression The so-called supervised learning problem is the following: we want to approximate the random variable Y with an appropriate function of the random variables X 1,..., X p with the

More information

Non-parametric (Distribution-free) approaches p188 CN

Non-parametric (Distribution-free) approaches p188 CN Week 1: Introduction to some nonparametric and computer intensive (re-sampling) approaches: the sign test, Wilcoxon tests and multi-sample extensions, Spearman s rank correlation; the Bootstrap. (ch14

More information

Simulating Uniform- and Triangular- Based Double Power Method Distributions

Simulating Uniform- and Triangular- Based Double Power Method Distributions Journal of Statistical and Econometric Methods, vol.6, no.1, 2017, 1-44 ISSN: 1792-6602 (print), 1792-6939 (online) Scienpress Ltd, 2017 Simulating Uniform- and Triangular- Based Double Power Method Distributions

More information

Econometric Methods for Panel Data

Econometric Methods for Panel Data Based on the books by Baltagi: Econometric Analysis of Panel Data and by Hsiao: Analysis of Panel Data Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies

More information

Sequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process

Sequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process Applied Mathematical Sciences, Vol. 4, 2010, no. 62, 3083-3093 Sequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process Julia Bondarenko Helmut-Schmidt University Hamburg University

More information

MA Advanced Econometrics: Applying Least Squares to Time Series

MA Advanced Econometrics: Applying Least Squares to Time Series MA Advanced Econometrics: Applying Least Squares to Time Series Karl Whelan School of Economics, UCD February 15, 2011 Karl Whelan (UCD) Time Series February 15, 2011 1 / 24 Part I Time Series: Standard

More information

Rank-Based Methods. Lukas Meier

Rank-Based Methods. Lukas Meier Rank-Based Methods Lukas Meier 20.01.2014 Introduction Up to now we basically always used a parametric family, like the normal distribution N (µ, σ 2 ) for modeling random data. Based on observed data

More information

NON-PARAMETRIC STATISTICS * (http://www.statsoft.com)

NON-PARAMETRIC STATISTICS * (http://www.statsoft.com) NON-PARAMETRIC STATISTICS * (http://www.statsoft.com) 1. GENERAL PURPOSE 1.1 Brief review of the idea of significance testing To understand the idea of non-parametric statistics (the term non-parametric

More information

Module 9: Nonparametric Statistics Statistics (OA3102)

Module 9: Nonparametric Statistics Statistics (OA3102) Module 9: Nonparametric Statistics Statistics (OA3102) Professor Ron Fricker Naval Postgraduate School Monterey, California Reading assignment: WM&S chapter 15.1-15.6 Revision: 3-12 1 Goals for this Lecture

More information

unadjusted model for baseline cholesterol 22:31 Monday, April 19,

unadjusted model for baseline cholesterol 22:31 Monday, April 19, unadjusted model for baseline cholesterol 22:31 Monday, April 19, 2004 1 Class Level Information Class Levels Values TRETGRP 3 3 4 5 SEX 2 0 1 Number of observations 916 unadjusted model for baseline cholesterol

More information

arxiv:math/ v1 [math.pr] 9 Sep 2003

arxiv:math/ v1 [math.pr] 9 Sep 2003 arxiv:math/0309164v1 [math.pr] 9 Sep 003 A NEW TEST FOR THE MULTIVARIATE TWO-SAMPLE PROBLEM BASED ON THE CONCEPT OF MINIMUM ENERGY G. Zech and B. Aslan University of Siegen, Germany August 8, 018 Abstract

More information

ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications

ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications Yongmiao Hong Department of Economics & Department of Statistical Sciences Cornell University Spring 2019 Time and uncertainty

More information

ANALYSIS OF VARIANCE OF BALANCED DAIRY SCIENCE DATA USING SAS

ANALYSIS OF VARIANCE OF BALANCED DAIRY SCIENCE DATA USING SAS ANALYSIS OF VARIANCE OF BALANCED DAIRY SCIENCE DATA USING SAS Ravinder Malhotra and Vipul Sharma National Dairy Research Institute, Karnal-132001 The most common use of statistics in dairy science is testing

More information

Financial Econometrics

Financial Econometrics Financial Econometrics Nonlinear time series analysis Gerald P. Dwyer Trinity College, Dublin January 2016 Outline 1 Nonlinearity Does nonlinearity matter? Nonlinear models Tests for nonlinearity Forecasting

More information

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis. 401 Review Major topics of the course 1. Univariate analysis 2. Bivariate analysis 3. Simple linear regression 4. Linear algebra 5. Multiple regression analysis Major analysis methods 1. Graphical analysis

More information

Chapter 15: Nonparametric Statistics Section 15.1: An Overview of Nonparametric Statistics

Chapter 15: Nonparametric Statistics Section 15.1: An Overview of Nonparametric Statistics Section 15.1: An Overview of Nonparametric Statistics Understand Difference between Parametric and Nonparametric Statistical Procedures Parametric statistical procedures inferential procedures that rely

More information