False Alarm Probability based on bootstrap and extreme-value methods for periodogram peaks

Similar documents
arxiv: v1 [astro-ph.sr] 2 Dec 2015

Sharp statistical tools Statistics for extremes

Period. End of Story?

The analysis of indexed astronomical time series XI. The statistics of oversampled white noise periodograms

arxiv: v1 [astro-ph] 15 Jun 2007

arxiv: v1 [astro-ph.im] 16 Jun 2011

Additional Keplerian Signals in the HARPS data for Gliese 667C from a Bayesian re-analysis

arxiv:astro-ph/ v1 17 May 2006

Statistical Applications in the Astronomy Literature II Jogesh Babu. Center for Astrostatistics PennState University, USA

If we want to analyze experimental or simulated data we might encounter the following tasks:

UNIFORMLY MOST POWERFUL CYCLIC PERMUTATION INVARIANT DETECTION FOR DISCRETE-TIME SIGNALS

Degrees-of-freedom estimation in the Student-t noise model.

Modelação de valores extremos e sua importância na

A Correntropy based Periodogram for Light Curves & Semi-supervised classification of VVV periodic variables

The Spectral Density Estimation of Stationary Time Series with Missing Data

Financial Econometrics and Volatility Models Extreme Value Theory

Analysing the Hipparcos epoch photometry of λ Bootis stars

Extreme Value Analysis and Spatial Extremes

Variable and Periodic Signals in Astronomy

Modelling Multivariate Peaks-over-Thresholds using Generalized Pareto Distributions

Parameter estimation in epoch folding analysis

Variable and Periodic Signals in Astronomy

Regional Estimation from Spatially Dependent Data

Frequency estimation uncertainties of periodic signals

TIME SERIES ANALYSIS

Gaussian processes. Basic Properties VAG002-

On the Application of the Generalized Pareto Distribution for Statistical Extrapolation in the Assessment of Dynamic Stability in Irregular Waves

Time-dependent Behaviour of the Low Amplitude δ Scuti Star HD 52788

High Time Resolution Photometry of V458 Vul

Payer, Küchenhoff: Modelling extreme wind speeds in the context of risk analysis for high speed trains

Overview of Extreme Value Theory. Dr. Sawsan Hilal space

Journal of Environmental Statistics

Change Point Analysis of Extreme Values

Sequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process

Extremogram and ex-periodogram for heavy-tailed time series

Photometry of the δ Scuti star HD 40372

The Goodness-of-fit Test for Gumbel Distribution: A Comparative Study

2. SPECTRAL ANALYSIS APPLIED TO STOCHASTIC PROCESSES

TOTAL JITTER MEASUREMENT THROUGH THE EXTRAPOLATION OF JITTER HISTOGRAMS

Introduction to Algorithmic Trading Strategies Lecture 10

arxiv: v1 [astro-ph.sr] 6 Jul 2013

Investigation of an Automated Approach to Threshold Selection for Generalized Pareto

Bootstrap Confidence Intervals

Algorithm Independent Topics Lecture 6

Extremogram and Ex-Periodogram for heavy-tailed time series

1 Random walks and data

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80

Covariance function estimation in Gaussian process regression

Lecture 2 APPLICATION OF EXREME VALUE THEORY TO CLIMATE CHANGE. Rick Katz

Bootstrap & Confidence/Prediction intervals

Statistical techniques for data analysis in Cosmology

Bayesian Modelling of Extreme Rainfall Data

PAijpam.eu ADAPTIVE K-S TESTS FOR WHITE NOISE IN THE FREQUENCY DOMAIN Hossein Arsham University of Baltimore Baltimore, MD 21201, USA

Extreme Value Theory and Applications

Time Series Analysis. Eric Feigelson Penn State University. Summer School in Statistics for Astronomers XIII June 2017

Frequency Estimation of Rare Events by Adaptive Thresholding

The Spatial Variation of the Maximum Possible Pollutant Concentration from Steady Sources

Variable inspection plans for continuous populations with unknown short tail distributions

arxiv: v1 [astro-ph] 22 Dec 2008

Statistical distributions: Synopsis

REGULARIZED ESTIMATION OF LINE SPECTRA FROM IRREGULARLY SAMPLED ASTROPHYSICAL DATA. Sébastien Bourguignon, Hervé Carfantan, Loïc Jahan

Directionally Sensitive Multivariate Statistical Process Control Methods

Overview of Extreme Value Analysis (EVA)

Analysis of the AIC Statistic for Optimal Detection of Small Changes in Dynamic Systems

Frequency estimation by DFT interpolation: A comparison of methods

IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 57, NO. 3, MARCH X/$ IEEE

arxiv:astro-ph/ v1 6 Jun 2002

Zwiers FW and Kharin VV Changes in the extremes of the climate simulated by CCC GCM2 under CO 2 doubling. J. Climate 11:

HIERARCHICAL MODELS IN EXTREME VALUE THEORY

Time Series: Theory and Methods

Extreme Precipitation: An Application Modeling N-Year Return Levels at the Station Level

Assessing Dependence in Extreme Values

Power Functions for. Process Behavior Charts

On Information Maximization and Blind Signal Deconvolution

NONINFORMATIVE NONPARAMETRIC BAYESIAN ESTIMATION OF QUANTILES

Semi-parametric estimation of non-stationary Pickands functions

Monitoring Random Start Forward Searches for Multivariate Data

Kriging models with Gaussian processes - covariance function estimation and impact of spatial sampling

arxiv:cond-mat/ v1 [cond-mat.stat-mech] 25 Feb 2003

arxiv: v1 [astro-ph.im] 28 Mar 2017

Regression of Time Series

Richard L. Smith Department of Statistics and Operations Research University of North Carolina Chapel Hill, NC

n =10,220 observations. Smaller samples analyzed here to illustrate sample size effect.

Bayesian Inference for Clustered Extremes

ColourMatrix: White Paper

STAT 6350 Analysis of Lifetime Data. Probability Plotting

Statistical Inference

Design and Implementation of CUSUM Exceedance Control Charts for Unknown Location

A COMPLETE ASYMPTOTIC SERIES FOR THE AUTOCOVARIANCE FUNCTION OF A LONG MEMORY PROCESS. OFFER LIEBERMAN and PETER C. B. PHILLIPS

UNIVERSITÄT POTSDAM Institut für Mathematik

Probability Distributions

Spatial and temporal extremes of wildfire sizes in Portugal ( )

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

Computational statistics

Uncertainties in extreme surge level estimates from observational records

Statistical Applications in Genetics and Molecular Biology

A spatio-temporal model for extreme precipitation simulated by a climate model

The International Journal of Biostatistics

Physics 509: Bootstrap and Robust Parameter Estimation

EVA Tutorial #2 PEAKS OVER THRESHOLD APPROACH. Rick Katz

Transcription:

False Alarm Probability based on bootstrap and extreme-value methods for periodogram peaks M. Süveges ISDC Data Centre for Astrophysics, University of Geneva, Switzerland Abstract One of the most pertinent problems in spectral analysis of astronomical time series is the assessment of significance of periodogram peaks. The origin of many difficulties is the often occurring non-gaussianity, the unclear meaning of the number of independent frequencies and the manner how it should be taken into account. I propose to avoid these problems by a bootstrap of the original time series, which makes the procedure independent of the distributional characteristics, and by the application of extreme-value methods. Generalized extreme-value distributions are used as general limiting distributions for maxima; they depend only on the decay of the right tail of the underlying distribution, but not on its precise form. Moreover, there is no need to compute the complete oversampled frequency range for each bootstrap repetition: under the zero hypothesis of the observed time series being a pure noise, the maxima of only a part of the spectra of the repetitions are sufficient to estimate the False Alarm Probability for the observed original peak. The procedure is tested on simulations. Keywords: methods: data analysis methods: statistical stars: oscillations. 1 Introduction Assessment of the significance of peaks in the periodogram of variable objects is of utmost importance for asteroseismology, detection of extrasolar planetary systems, analysis of variable stars and a number of other fields of astrophysics. Periodicities in a signal sampled at equally spaced time intervals is most efficiently searched for by the Fourier transform, which gives the projections of the time series to a set of orthogonal basis sine functions with frequencies defined by the number of time points and the time span of the observation. The statistical behaviour of the resulting periodograms is well known (see, e.g., Brockwell and Davis 2006). In astrophysical applications, the situation is different, because it is most often impossible to evenly sample the light curve of the objects. Instead, we have scarce observations scattered irregularly over a long time span, sometimes as few as 25, and from this, we derive a periodogram over a dense grid of frequencies. The periodogram contains information on how well an oscillating signal at any grid frequency fits the observed data, and unfortunately, it comprises signs of many e-mail: Maria.Suveges@unige.ch 1

undesirable effects as well: aliases, parasite frequencies, and spectral leakage. These effects make the identification of the peak corresponding to a periodic oscillation very difficult. Assessment of the statistical significance of an eventual peak involves testing of the zero hypothesis H 0 of the observed time series being purely white noise against the alternative, H 1, stating that there is a periodic deterministic signal in it. We do this either by computing the probability of a peak of the observed height or higher under H 0 (the false alarm probability or FAP), or by computing levels corresponding to prescribed FAP values under H 0. This requires the knowledge of the distribution of the maximum of the periodogram, which is extremely difficult for several reasons. Theoretical derivations build on many restrictive assumptions, such as Gaussian distribution under H 0 and independence. Moreover, the distribution of the maximum, especially at high quantiles, is extremely sensitive to the tail of the underlying distribution of the periodogram and to a parameter called the effective number of independent frequencies in the investigated frequency range, both of which are extremely hard to estimate precisely. We propose a new procedure that avoids these drawbacks. It combines nonparametric bootstrap, which is able to reproduce any empirical distribution, with extreme-value models that provide generally valid, asymptotically well-motivated models for the tails of most continuous distributions. We give a brief summary of the problems in the assessment of significance for periodogram peaks and of extreme-value distributions in Section 2, then describe the procedure and give the motivation for each step in Section 3. Results on simulations are given in Section 5, and the improvements on other methods for significance assessment are summarized in Section 6. 2 Statistical methods 2.1 Frequency analysis There are many methods used in astronomy for frequency analysis; to name a few, the Deeming method (Deeming, 1975), PDM-Jurkevich (Stellingwerf, 1978; Dupuy and Hoffman, 1985), string length (Clarke, 2002), SuperSmoother (Friedman, 1984; Reimann, 1994), Lomb-Scargle and its extension, the generalized least squares method (GLS, Lomb 1976; Scargle 1982; Zechmeister and Kürster 2009) and the FastChi2 of Palmer (2009). We fit periodic functions taking frequencies over a grid to the observed light curve, and using some goodness-of-fit statistic, we look for the frequency which yields a significantly better fit than our zero hypothesis, a constant mean plus white noise. The collection of test statistic values as a function of the test frequency is termed somewhat loosely the periodogram. The frequency where it attains its maximum (or, depending on the chosen statistics, its minimum) is accepted as the most likely frequency of variability, if there is indeed variability. Statistically put, we ask how likely the found maximum or minimum value is under the zero hypothesis H 0, and compare this probability (the False Alarm Probability, abbreviated to FAP) to a specified level of confidence, say 0.01. For its calculation, we need to know the distribution of the test statistic under H 0, namely, the distribution of the maximum of a set of random variables. Minima can be transformed to maxima by simple monotone 2

transformations, so we do not consider them separately anymore. The distribution G of a set of M independent random variables Z 1,..., Z M with common distribution function F is G(z) = F (z) M. For periodogram peaks, variants of this formula are commonly used, with F based on theoretical arguments, depending on the test statistic used to estimate the periodogram. However, before application many things must be accounted for or modified. First, even when the marginal distribution of the periodogram is derived theoretically, it is only approximate due to an assumption of Gaussianity of the time series observations and to the asymptotic nature of the distribution of the test statistic. Gaussianity in the underlying time series is very often unjustified, especially in the hypothesis testing situation. We must allow for the possibility of non-gaussianity: if the observed time series is indeed noise, then it can be non-gaussian. Second, in order to find a critical quantile value corresponding to a FAP of say 0.01, we should compute 1 0.01 = G(z) = F (z) M, and since M is usually of the order of 100 (Horne and Baliunas, 1986; Frescura et al., 2008; Schwarzenberg-Czerny, 2012), we need to know very high quantiles of F very precisely. Fitted marginal models to obtain ˆF are usually the least precise in these high regions, because such extreme values are most often absent or very rare among the data. Thus, models extrapolate, reasonably or not, in this region. Third, not only we do not know the value of M to use in the formula, but in fact, the formula itself is not valid. The testing situation is very far from the testing of the maximum of a set of independent variables. For uneven sampling, there is no set of mutually independent test frequencies. In addition, we use the maximum of sometimes several hundreds of thousand periodogram values computed from often as few as 50 or 100 observations, which implies degeneracy in the joint distribution of the periodogram values, and makes the assumption of independence implausible. The formula is thus not motivated, and can be used only as an ad hoc approximation. M lacks interpretation, and must be estimated as a parameter (a recent proposition is in Schwarzenberg-Czerny 2012), which adds to the instability of the FAP estimates. 2.2 Extreme-value statistics Extreme-value methods (see e.g. Leadbetter et al. 1983; Coles 2001; Beirlant et al. 2004) are widely used to assess probabilities of rare (and often catastrophic) events, or estimate the time within which an extreme event of a certain size should be experienced once on average. fundamental theorem states that the distribution of the properly rescaled maximum tends to the generalized extreme-value family (GEV), of the form { ( G(z) = exp 1 + ξ z µ ) } 1/ξ, ξ R, µ R, σ > 0, (1) σ regardless of the underlying distribution of the variables: maxima of variables from practically all continuous distributions used in applications of statistics tend to limits in the GEV family. The distribution has three parameters. The shape parameter, ξ, determines the decay rate of Its 3

the tail probabilities. GEV distributions with negative ξ have a finite upper endpoint, so there is zero probability to observe maxima larger than a certain value, yielding the limiting family for maxima of uniform or beta distributed variables. Positive ξ indicates power law-like heavytailed behaviour, with high probabilities to find extremely large maxima, like in the Cauchy or Student s t distribution. Finally, ξ = 0 implies relatively quickly, exponentially decaying tail probabilities, and thus it is the limiting distribution of the exponential, χ 2, gamma, normal or lognormal distributions. The other two parameters, µ and σ are location and scale parameters. The GEV distributions are the limiting distributions for maxima of dependent sets of variables too, under two conditions: one is an asymptotic decrease of the strength of dependence at extreme levels, and the other is stationarity. The first of these conditions is not straightforward to translate for periodograms, as the limit must be precisely specified, and for any finite-length time series and a periodogram calculated on a fixed grid, the quality of the extreme-value approximation is not assured because of the long-range dependence in the periodograms. Diagnostic tests of the fits are therefore always necessary, for example by means of return level plots and quantile-quantile plots. Estimation of extreme-value behaviour in a time series is usually performed by dividing the sequence into long blocks say of size N (e.g. years for a sequence of daily temperature measurements spanning several decades). N must be sufficiently large to provide a good extremevalue approximation, because maxima of too small sets do not follow distributions close enough to the asymptotic validity region of extreme-value theory. Usually, under mild dependence in the random sequence, blocks of several hundred variables yield GEV models of acceptable quality (in the previous example, maxima of 365 days). We then select the maxima of each block (in the example, the several ten annual maximum temperatures), and find the parameters of the GEV by maximum likelihood. The estimated GEV distribution can then be used to obtain probabilities of very high levels or distributions of maxima of longer periods. The fact that the form of the GEV remains the same if we take longer blocks (it is max-stable in statistical terminology) implies that quantiles and probabilities for much longer periods can be calculated by a reasonable extrapolation based on the model for the shorter blocks. An important quantity is the return level; the return level z p associated with the return period 1/p is the level that is expected to be exceeded once in every 1/p blocks of the same length N (in the previous example, the z 1/20 return level is the temperature which is exceeded only once in every 20 years). This is simply the 1 p quantile of the GEV distribution, and can be obtained by inverting the formula (1): z p = G 1 (1 p) = µ σ [ 1 { log(1 p)} ξ]. (2) ξ Extreme-value methods have been used for periodograms also by Baluev (2008), who derives upper limits for the FAP based on upcrossings of χ 2, beta and F random processes. However, these methods are built on the fundamental assumption of Gaussianity under H 0, and their quality decreases in the case of strong aliasing and spectral leakage. 4

3 The proposed procedure We propose to avoid the pitfalls listed in Section 2.1 by combining nonparametric bootstrap and extreme-value methods. Suppose we have N irregularly spaced observations, and from these data, we derive the periodogram between 0 and f max on a grid of n test frequencies. Let moreover K denote the oversampling factor between the used frequency grid and the Fourier frequency system. The proposed procedure to assess the significance of a found peak is as follows: 1. Create R bootstrap repetitions of the original time series: preserving exactly the epochs of the observations, each observed magnitude value is replaced with another one, randomly chosen with equal probabilities from the original time series. One observation may be selected more than once. 2. For each bootstrap repetition, we select a subset of L frequencies from the frequency grid randomly with equal probabilities, with L such that L n and n/(kl) N/2. Around each of these L frequencies, we take a symmetric interval of K frequencies. Compute the periodogram at these KL frequencies with the same method as used for the original time series, and select the maximum from the KL values. This maximum is all we need to retain from one repetition. 3. Fit an extreme-value model G L (z) to the R maxima: estimate the parameters ˆξ, ˆσ and ˆµ by maximum likelihood. Use diagnostic plots: the return level plot and the quantile-quantile plot to check the quality of the extreme-value fit. 4. Estimation can be based on the high quantiles of G L (z). To estimate a periodogram level corresponding to FAP = 0.01, we should consider that this is the level that is exceeded only once in a hundred complete periodograms. Also, the complete periodogram contains n/(kl) times more frequencies than the subsets based on which G L (z) was estimated. Thus, the return level corresponding to this probability is calculated by substituting 1 p = KL/(100 n) into equation (2). The procedure is the result of a consideration of several important aspects. First, nonparametric bootstrap uses the empirical distribution function of the observations as sampling distribution, and does not force a pre-specified distribution to the data. The result is a white noise sequence with a distribution close to the observed one, which corresponds perfectly to H 0. The following period search on these noise sequences using part of the frequency grid, and the selection of the maximum of each, yields a sample of independent maxima of sub-periodograms of noise sequences with approximately the same marginal distribution as the original time series. Next, the choice of the size KL (equivalently, of L) of the frequency subsets is motivated by several principles and aims. The most important aim is of course to avoid having to compute the entire periodogram for each bootstrap repetition, as this is excessively time-consuming, so L n. Nevertheless, L must be sufficiently large to ensure that KL provides a good extremevalue approximation. The quality of the extreme-value models must therefore be checked for a 5

double reason: one is the long-range dependence that might invalidate the theory described in Section 2.2, and the other is to ascertain whether L is chosen sufficiently large. The long-range dependence also motivates the condition n/(kl) N/2, since this ensures that the number of subsets of size KL comprised in a complete periodogram of size n is less than the number of input observations, and thus avoids problems of degeneracy in the underlying joint distributions of the maxima of the sub-periodograms. The factor 1/2 is due to the fact that the periodogram loses half of the information in the observations: it holds information on the amplitudes of the basis functions, but not on their phases. We select the frequency subsets in a specific manner. Choosing the L frequencies randomly with uniform probability across the entire grid implies that frequencies separated both by short and long frequency intervals will occur in the set. Thus, we obtain a sample containing information on the long-range dependence structure. Taking a symmetric interval of K frequencies around each of the L random frequencies is equivalent to finding the local maximum around each random frequency within peaks due to spectral leakage, and accounts for the short-range dependence patterns. This strategy for the choice of the KL test frequencies, together with the selection of a new set for each bootstrap repetition, ensures that the subsets reflect the characteristics of the complete periodogram under H 0. 4 Simulation The performance of the above procedure was assessed on simulations. We present here a pure sine-wave model of the form g(t) = A i sin(f 0 t) with frequency F 0 = 3.379865 d 1 and with three different amplitudes of A 1 = 0.15, A 2 = 0.05 and A 3 = 0.025 mag. To this variability pattern, we added random noise generated from Gaussian distributions with time-varying variance, yielding the model Y i = g(t i ) + ɛ i, ɛ i N (0, σi 2 ), where the standard errors σ i were randomly selected from a gamma distribution such that they were on average 0.05 mag. The signal-to-noise ratios of the three time series were thus SNR i = A i / σ = 3, 1 and 0.5. Epochs of observations were chosen randomly on a time grid of 0.005 day (about 10 minutes) with a total span of 25 days; imitating random nighttime observations during 4 hours each night, we have randomly uniformly selected two different sequences of epochs, one consisting of N = 100 observations, the other, of N = 25. These time grid parameters lead to an upper detection limit f max = 100d 1, a Fourier frequency set of {0, 1/25, 2/25,..., 100}d 1 and a corresponding leakage peak width of 0.04 d 1. The oversampling factor used is K = 16, providing a test frequency grid {0, 0.0025, 0.005,..., 100} d 1. The generalized Lomb-Scargle (GLS) period search method was performed on the six simulated time series using this frequency grid, once without weighting, and once weighted the usual way, by the inverse variances normalized to have sum 1. 6

5 Results The periodograms of the six simulations estimated by the unweighted GLS method are shown in Figure 1. The results are very similar for weighted GLS. Applying the procedure described in Section 3, we created 1000 nonparametric bootstrap repetitions from each time series, and in order to test the stability of the GEV estimates against the size L of the test frequency set, we computed the maxima on subsets of size L = 50, 100, 200, 300, 400 and 500 for each repetition. GEV models were fitted to all six times six samples of 1000 maxima by maximum likelihood, yielding estimated distributions ĜL,i(z; ˆξ L,i, ˆσ L,i, ˆµ L,i ), and return levels (quantiles) for the complete periodogram corresponding to FAP = 0.05, 0.01 and 0.005 were calculated based on these distributions. The levels obtained for FAP = 0.01 are shown in Figure 1 as lines of different colours. The estimated quantiles agree well with the judgment of the human eye: for N = 100 with SNR = 3 and 1, the signal is clearly discernible above the noise level, and accordingly, the quantiles pass well below their peak. In the case of N = 100, SNR = 0.5, even though there is a group of peaks at the true frequency, we find other groups with comparable values. The signal, though it is there, is undetectable, and the quantile levels confirm this. With fewer observations (N = 25), the strongest signal with SNR = 3 is just discernible as a group of peaks at the right frequency, but it exceeds other peaks only slightly. The quantile estimates agree that this peak just attains the FAP = 0.01 level (or remains barely below); it is significant though at the level of 0.05. For the two cases with lower SNR, the signal is masked by the noise, and accordingly, the quantile levels pass well above the periodogram values. In addition, the estimated quantiles are remarkably stable with respect to the size L of the test frequency set. Inspection of an enlarged version of the quantile levels in Figure 2 for SNR = 0.5 for L = 100 and 500, plotted together with a 95% confidence interval based on bootstrap, confirms this: the large overlap of the confidence intervals indicate statistically indistinguishable estimates when using L = 100 or 500. In order to keep computational time as low as possible, it is also important to investigate the quality of the estimates with respect to the number of bootstrap repetitions. Using only R = 200, 400, 600 and 800 repetitions instead of 1000, the estimates were found to remain stable down to R = 400 in the case of N = 25 and to 200 in the case of N = 100, though as expected, with increasing variances. Return level and quantile-quantile plots indicated that the quality of the fits is good in the full range of the tried parameter combinations, and the use of the GEV as statistical model for maxima of periodograms is justified. It is possible that though the estimated high quantiles seem to give a reasonable idea about the significance of the peak, and they have very good stability properties, they yield biased estimates. Their quality was checked by simulations: 2000 white noise sequences with the same distributional characteristics were generated, and the full periodogram was calculated for each. Then the empirical probability of the maxima to exceed the quantile levels derived from the GEV fits was estimated by the proportion of exceedances among the maxima of 2000 full periodograms. The results using the non-weighted GLS method, L = 500 and R = 1000 are 7

Figure 1: The level corresponding to FAP = 0.01 for SNR = 3 (upper row), SNR = 1 (middle row) and SNR = 0.5 (bottom row), for time series of 100 (left) and 25 (right) simulated observations. The size of the used test frequency set for the estimation is indicated by the colours. The black spikes represent the periodogram of the original time series obtained by non-weighted GLS. summarized in Table 1. It appears that the GEV slightly overestimates the quantiles, but as this results in a conservative significance estimate for an observed peak, with a tendency rather to judge a dubious peak to be nonsignificant than significant, it implies a lower false alarm rate. 6 Discussion The paper deals with problems in the estimation of FAP in the frequency analysis of astronomical time series. These problems include the failure of theoretically derived distributions due to non-gaussianity in the observations, loss of the orthogonal Fourier frequency system due to SNR =3 SNR = 1 SNR = 0.5 FAP from GEV 0.01 0.005 0.01 0.005 0.01 0.005 Empirical FAP (N = 100) 0.01 0.005 0.005 0.003 0.009 0.004 Empirical FAP (N = 25) 0.008 0.005 0.008 0.004 0.006 0.003 Table 1: The proportion of cases in 2000 noise simulations 8

100 observations 25 observations Normalised χ 2 0.10 0.20 0.30 L = 100 L = 500 0.2 0.6 0 20 40 60 80 100 Frequency 0 20 40 60 80 100 Frequency Figure 2: The estimated 0.99 quantile for SNR = 0.5, enlarged and together with 95% confidence intervals based on bootstrap. the irregular time sampling, the lack of independence originating in the sparse irregular time sampling and to the tremendous oversampling in the periodogram, and the sensitivity of the commonly used formula to parameters that are exceedingly hard to well estimate. The procedure combines bootstrap and GEV model fitting, which helps to reduce many shortcomings of the commonly used procedures. The assumption of Gaussianity and the need for an excellent estimate of the tail of the marginal periodogram distribution is avoided by using nonparametric bootstrap of the original time series. The periodogram of the bootstrap repetitions is computed only at sets of randomly selected frequencies, which are chosen so that they carry information on both the long-range and the short-range dependence. The use of the GEV model, since it is a generally valid limiting distribution for maxima, ensures a reasonable assumption on the tail behaviour of the maxima. Moreover, because of its max-stability, the same model fitted to maxima of only subsets of periodogram can be used in a straightforward manner to estimate the high quantiles of the complete periodogram. The problematic need of the classical methods to estimate a non-interpretable parameter inducing strong instability of the estimates, the number of independent frequencies, is completely avoided. According to simulations, the procedure gives good estimates for the quantile levels (though slightly over-estimates them), and provides reasonable results for the significance of the peaks in the original periodogram, agreeing with its visual aspect and the knowledge about the signal in the simulations. It is stable both with respect of the size of the frequency subset (around 100 local maxima are necessary) and of the number of bootstrap repetitions (requiring a few hundred), and therefore needs only moderate computational time (several times the computation of a complete periodogram). In summary, the procedure performs very well in estimating the high quantiles of the distribution of the maximum of the periodogram for any underlying distribution, and is a very promising method to assess significances of signals with low SNR in astronomical time series analysis. References Baluev, R. V.: 2008, MNRAS 385, 1279 9

Beirlant, J., Goegebeur, Y., Segers, J., and Teugels, J.: 2004, Statistics of Extremes, John Wiley Brockwell, P. J. and Davis, R. A.: 2006, Time Series: Theory and Methods, Springer Science+Business Media, 2nd edition Clarke, D.: 2002, AAp 386, 763 Coles, S. G.: 2001, An Introduction to Statistical Modeling of Extreme Values, Springer-Verlag, London Deeming, T. J.: 1975, ApSS 36, 137 Dupuy, D. L. and Hoffman, G. A.: 1985, Photometry Communications 20, 1 International Amateur-Professional Photoelectric Frescura, F. A. M., Engelbrecht, C. A., and Frank, B. S.: 2008, MNRAS 388, 1693 Friedman, J. H.: 1984, A variable span smoother, Technical report, LCS Technical Report 5 Horne, J. H. and Baliunas, S. L.: 1986, ApJ 302, 757 Leadbetter, M. R., Lindgren, G., and Rootzén, H.: 1983, Extremes and Related Properties of Random Sequences and Processes, Springer-Verlag, New York Lomb, N. R.: 1976, ApSS 39, 447 Palmer, D. M.: 2009, ApJ 695, 496 Reimann, J. D.: 1994, Ph.D. thesis, Univ. of California, Berkeley Scargle, J. D.: 1982, ApJ 263, 835 Schwarzenberg-Czerny, A.: 2012, in IAU Symposium, Vol. 285 of IAU Symposium, pp 81 86 Stellingwerf, R. F.: 1978, ApJ 224, 953 Zechmeister, M. and Kürster, M.: 2009, AAp 496, 577 10