CHANGE-POINT DETECTION IN TEMPERATURE SERIES

Size: px
Start display at page:

Download "CHANGE-POINT DETECTION IN TEMPERATURE SERIES"

Transcription

1 CZECH TECHNICAL UNIVERSITY FACULTY OF CIVIL ENGINEERING CHANGE-POINT DETECTION IN TEMPERATURE SERIES PhD. Thesis MONIKA RENCOVÁ Supervisor: Prof. RNDr. Daniela Jarušková, CSc. Department of Mathematics Thákurova 7, Prague 9

2 Abstract This work is devoted to some change-point detection problems in temperature series. Results of this thesis are based on working with real data. The submitted work presents suggestions on how the change-point methods may be applied to detect changes in annual maximal, resp. minimal temperatures and to detect changes in occurrences of unusually hot, resp. cold days. Solving these practical examples we came across some theoretical problems, we tried to work out in this thesis. In the first problem we apply the change-point theory and we will be looking for a change in parameters in a large class of independent random variables with a GEV distribution not satisfying regularity conditions. In the second problem we will focus on dependent variables and show how the change-point theory might be extended from linear processes to strong-mixing sequences. i

3 Acknowledgements It is a real pleasure to express my thanks to Prof. RNDr. Daniela Jarušková, CSc., my thesis supervisor, who introduced me into a field of change-point detection, encouraged me and spent a lot of time in discussions on this topic. I would like to thank the Management and the Staff of Department of Mathematics for providing a good environment for my work on the thesis. However, first of all I thank to my family for supporting me during my work, especially, I would like to appreciate my husband Jirka for his patience. ii

4 Contents Preface Data 7 3 The change-point detection for the GEV distributions 7 3. The change-point detection for the Weibull distributions Main results for the Weibull distributions The change-point detection for the Fréchet distributions Results Conclusion The change-point detection for dependent data The change-point detection for strong-mixing sequences Application Conclusion Block permutation 8 5. Block permutation for strong-mixing sequences Application Comparison Results A Some useful theorems and inequalities A. The test statistic for a change-point detection A. The extreme value distributions A.3 Limit theorems A.4 Strong-mixing sequences A.5 Rank statistics

5 Chapter Preface This work is devoted to some change-point detection problems in temperature series. Results of this thesis are based on working with real data. The broadly accepted hypothesis of global warming stimulated an interest in studying long temperature series. Some scientists assume that changes do not necessarily occur in the mean of the series but rather in some other characteristics, e.g. appearance of some extreme events or increase of difference between summer and winter temperatures etc. This raises an interest in studying statistical properties of extremes of random sequences, see e.g. Embrechts et al. [], Leadbetter et al. []. Our paper presents suggestions on how the change-point methods may be applied to detect changes in annual maximal, resp. minimal temperatures and to detect changes in occurrences of unusually hot, resp. cold days. Solving these practical examples we came across some theoretical problems, we tried to work out in this thesis. The world is filled with changes. We encounter them in economics, medicine, meteorology, climatology etc. A change-point analysis is a statistical method allowing to decide whether an observed stochastic process follows one model or whether the model changes. In the case of a change, we might be interested in following problems: when a change was detected and how many changes have occurred. The change-point detection is formulated in terms of hypotheses testing. The null hypothesis claims that the series is stationary, usually it means that the parameters of the model do not change, while the alternative hypothesis claims that at an unknown time point the model changes. The decision rule for rejecting the null hypothesis is based on test statistics. The earliest change-point studies go back to the 95s, where they arose in the context of quality control. We observe an output of a manufacturing process and assume that a certain characteristic varies around a certain in-control constant a. Sometimes, for example due to a failure of the production device, this constant starts to vary around another out-of-control constant a a and we want to know if and when such a change occurred. Statistical procedures in change-point analysis can be divided into two categories: on-line and off-line procedures. The on-line approach, coming from the manufacturing process, is based on the idea that after each observation we apply a new test and hope to be warned that the change occurred. In this thesis we will work with the off-line analysis when we already have all the observations and we apply a test for

6 . Preface 3 the whole data to decide whether and when the change occurred. This historically first change-point problem can be formulated as follows. For simplicity we assume that the starting value a and variance σ are known and the observations are independent and distributed according to the normal distribution. Moreover we standardize the observations and obtain variables Y i, i =... n with a zero mean and unit variance at the beginning. Then change-point problem formulated by hypotheses testing is: H : Y i = e i, i =,..., n,. A : there exists k {,..., n } such that Y i = e i, i =,..., k, Y i = a + e i, i = k +,..., n, where a. Using likelihood ratio method we obtain a so-called maximum-type statistic { } n max Y k n i.. n k i=k+ However, in practice the distribution of this statistic is very complex, so that it can be computed only for small sample sizes. Therefore, for n large, the asymptotic behavior of the statistic. is of interest. The maximum-type statistic goes to infinity as n a.s., however we can approximate this statistic by a maximum of a standardized Wiener process satisfying = o p t log log n max k n n i= Y i W t sup k /n t The approximate critical values can be calculated from the asymptotic behavior of the probabilities under H: { } n P max Y i > x + b n exp { e x}, x R,.3 k n n k a n where i=k+ a n = log log n, b n = log log n + log log log n log π. This approximation was derived by Darling and Erdös [8] in 956. However, it often happens that the starting value a is unknown. In such case we test the following null hypothesis H against the alternative A: H : Y i = a + e i, i =,..., n,.4 A : there exists k {,..., n } such that Y i = a + e i, i =,..., k, Y i = a + δ + e i, i = k +,..., n.

7 . Preface 4 with δ. Again for i.i.d. random variables distributed according to the normal distribution N, σ with σ known we obtain the maximum-type statistic of a form { } n k max Y k n σ i Y n..5 k n k For n large we may approximate the statistic.5 by the maximum of a standardized Brownian bridges n k max Y k n σ k n k i Y n sup Bt = o p. t t log log n i= i= n t n Yao and Davis [8] proved similar approximation as in.3 for a sequence of independent normal variables. For x R it holds { n k } P max Yi Y n > x + b n exp { e x}. k n σ k n k a n i= For a quite extensive survey on change-point detection we refer to Csörgő Horváth [7]. They used the log-likelihood ratio for the general model working with a sequence of independent random vectors X, X,..., X n with distribution functions F x; θ,..., F x; θ n, respectively, where θ i Θ R d for i =,..., n are parameters of the distribution functions and are assumed to change at unknown time. The general problem tested in Csörgő Horváth [7] has a form: The asymptotics of a testing statistic H : ϕ = ϕ =... = ϕ n A : there exists k {,..., n } such that ϕ =... = ϕ k ϕ k+ =... = ϕ n. max logλ k k n is under null hypothesis again / lim P Alog n max logλ k t + D d log n n k n for all t R, where and where Γt is the gamma function defined Ax = log x D d x = log x + d/ log log x log Γd/, Γt = y t exp ydy. = exp e t

8 . Preface 5 In Chapter 4 of Csörgő Horváth [7] it is also shown that the limit results remain true for a large class of dependent observations. Many articles have been published on change-point detection, see e.g. Antoch et al. [], Antoch J. and Hušková M. [], Gombay E. and Horváth, L. []. For application in climatology we refer to publications of Jarušková [8] and Jandhyala [?]. In this thesis we will apply the change-point methods for two special cases. In the first problem we apply the change-point theory and we will be looking for a change in parameters in a large class of independent random variables with the GEV distribution. In the second problem we will focus on dependent variables and show how the change-point theory might be extended from linear processes to strong-mixing sequences. The material is divided into chapters, sections and paragraphs. The studied data are presented in the second chapter. We summarize the origin of the data sets and provide some statistical characteristics of the observations. The results of the third chapter were obtained while solving change-point detection in annual maximal and minimal temperatures. We are looking for a change in parameters in a large class of random variables with the GEV distribution. The general Theorem.3.. presented by Csörgő Horváth [7], confer Appendix Theorem A.., can not be applied here directly, as extremal distributions do not satisfy conditions on regularity. Since the density function hx; µ, ψ, ξ is defined on the set {x; + ξx µ/ψ > }, the classical regularity conditions for maximum likelihood estimators are not satisfied. The next problem is caused by the conditions C.4 and C.5 of Theorem A.., since they require the continuity of third derivatives. This can be weakened by Smith s theorem, see Appendix Theorem A.3., and we will show that for ξ > there exists a sequence µ n, ψ n, ξ n of solutions of the likelihood equations such that n µ n µ, ψ n ψ, ξ n ξ converge in distribution to a zero mean normal vector with a variance covariance matrix M M is a Fisher information matrix and hence ξ > is still a regular case. From here an idea comes that the assertion of Theorem A.. is still valid. The results on temperature series are presented at the end of the chapter. Many articles have been published on independent observations. Clearly, working with real temperature series, we can not expect that the condition of independency is fulfilled. This problem we encountered solving the second example with occurrences of unusually hot or cold days. While for the annual maximal and minimal temperatures from the first example we might assume that the data form i.i.d. sequence, for occurrences of unusually hot or cold days we have a strong correlation between the temperature values measured at subsequent days, the value of correlation coefficient is for all series very close to.8. In the fourth chapter we define the problem by a model working with data forming strongmixing processes. We generalize the theory for dependent data presented in Csörgő and Horváth [7], and show that the asymptotic distribution of the testing statistic T n t is valid not only for linear processes but for strong-mixing sequences as well. In the context, it is an important question how to estimate σ. We can replace σ with an estimator,

9 where the rate of convergency to σ must be at least o p log log n, which is fulfilled by: ψn σ = R + Ri,.6 where Rj = n n j i= i= Yi Y n Yi+j Y n, Yn = n j n Y j and ψn tends to infinity with a certain speed. The estimator σ is a simplified version of the Bartlett log window estimator σnl = ˆR L + i ˆRi. L For more information about this estimator we refer to Antoch et al. []. i= It is wide known that the rate of convergency to distributional asymptotics under the null hypothesis is very slow, therefore in the fifth chapter we propose a permutation principle for obtaining the corresponding critical values. We generalize the theory presented by Kirch [] from linear processes to strong-mixing processes. We show that the estimator of variance σ LK = KL [ L K Y Kl + k Y n ], l= k= where Y n = n j n Y j, satisfies necessary condition on the rate of convergency. Critical values obtained from our data using the permutation test are listed at the end of the chapter. In Appendix we summarize the known theory for extremal distributions, the theorems concerning change-point analysis for i.i.d. data, Smith s theorem and some results on strong-mixing sequences and rank statistics.

10 Chapter Data The data sets we have studied were taken from a CD-ROM that was a part of the book edited by Camuffo and Jones [6]. The book sums up results of EU research project IM- PROVE. One of the main goals of the project was to produce seven highly reliable daily series Brussels, Cadiz, Milan, Padua, St. Petersburg, Stockholm, Uppsala, extending over more than two centuries, by correcting errors and inhomogeneities caused by changes in measurement style etc. We add one more data set, which attracted our attention the most Prague temperature series obtained from Of course, statisticians enjoy analyzing such long natural series but the length of the series also brings problems. Temperature often started to be measured at famous universities which are now mostly situated in city centers with their heat island effect. The climatologists who analyzed the Milan and Stockholm series tried to remove this effect by comparing the series with the measurements taken in nearby observatories, while the authors of the other series were not able to do it. This is not the only reason why the properties of the studied series are difficult to compare. The other reason is that the way in which the daily averages were calculated differs from place to place. In spite of the effort of the climatologists participating in the project, the series are not complete. Table and show the periods of measurement. period missing data number of obser. Brussels none 4 Cadiz ; Milan none 36 Padua ; 855; 865; 97; 947 St. Petersburg ; ; ; 793; ; 8 85; 846 Stockholm 756 none 45 Uppsala 75 none 66 Prague none 3 Table. Periods of observations together with missing data for annual minima. 7

11 . Data 8 period missing data number of obser. Brussels none 4 Cadiz 87 85; 873; 964; 988; 76 99; Milan none 36 Padua ; 9 9; 947; 954; 4 994; 996 St. Petersburg ; 784; 787; ; ; 8 84 Stockholm 756 none 45 Uppsala 75 none 66 Prague none 3 Table. Periods of observations together with missing data for annual maxima. The following tables summarize basic descriptive statistics of the data. maxima x σ n sk Brussels Cadiz Milan Padua St. Petersburg Stockholm Uppsala Prague Table 3. Descriptive statistics for annual maximal temperatures. minima x σ n sk Brussels Cadiz Milan Padua St. Petersburg Stockholm Uppsala Prague Table 4. Descriptive statistics for annual minimal temperatures. Comparing Tables 3 and 4 we can see that skewness for minimal temperatures is negative and its absolute values are larger, while skewness for maximal temperatures is positive with smaller absolute values. It might suggest that maximal temperatures might be modelled by the normal distribution. For minimal temperatures, the three parameter

12 . Data 9 Weibull distribution fits better, see Rencová [3]. The following figures show the behavior of the series under study Figure. Annual minimal temperatures in C in Brussels Figure. Annual maximal temperatures in C in Brussels Figure 3. Annual minimal temperatures in C in Cadiz. Figure 4. Annual maximal temperatures in C in Cadiz Figure 5. Annual minimal temperatures in C in Milan Figure 6. Annual maximal temperatures in C in Milan.

13 . Data Figure 7. Annual minimal temperatures in C in Padua. Figure 8. Annual maximal temperatures in C in Padua Figure 9. Annual minimal temperatures in C in St. Petersburg Figure. Annual maximal temperatures in C in St. Petersburg Figure. Annual minimal temperatures in C in Stockholm Figure. Annual maximal temperatures in C in Stockholm.

14 . Data Figure 3. Annual minimal temperatures in C in Uppsala. Figure 4. Annual maximal temperatures in C in Uppsala Figure 5. Annual minimal temperatures in C in Prague. Figure 6. Annual maximal temperatures in C in Prague. Figures 7 3 describe the second problem - changes in occurrences of unusually hot, resp. cold days. We provide graphs of sums of exceedances over a chosen level h =.5 and under a chosen level c =.5 for standardized daily series of studied data sets, see Section 4..

15 . Data x 4 Figure 7. Sums of exceedances over a chosen level h =.5 in Brussels x 4 Figure 9. Sums of exceedances over a chosen level h =.5 in Cadiz x 4 Figure 8. Sums of exceedances under a chosen level c =.5 in Brussels x 4 Figure. Sums of exceedances under a chosen level c =.5 in Cadiz x 4 Figure. Sums of exceedances over a chosen level h =.5 in Milan x 4 Figure. Sums of exceedances under a chosen level c =.5 in Milan.

16 . Data x 4 Figure 3. Sums of exceedances over a chosen level h =.5 in Padua x 4 Figure 4. Sums of exceedances under a chosen level c =.5 in Padua x 4 Figure 5. Sums of exceedances over a chosen level h =.5 in St. Petersburg x 4 Figure 6. Sums of exceedances under a chosen level c =.5 in St. Petersburg x 4 Figure 7. Sums of exceedances over a chosen level h =.5 in Stockholm x 4 Figure 8. Sums of exceedances under a chosen level c =.5 in Stockholm.

17 . Data x 4 Figure 9. Sums of exceedances over a chosen level h =.5 in Uppsala x 4 Figure 3. Sums of exceedances under a chosen level c =.5 in Uppsala x 4 Figure 3. Sums of exceedances over a chosen level h =.5 in Prague x 4 Figure 3. Sums of exceedances under a chosen level c =.5 in Prague. Figure 33 shows that there is a strong correlation between the temperature values measured at subsequent days, its value is for all series very close to.8, see the upper stationary graph in Figure 33. The following graphs, from the top to the bottom, depict correlation coefficients between two days with lag equal to, 3 and 4, e.g. the first value in the upper graph is the value of the correlation coefficient between st January and nd January, the first value in the second graph from the top is the value of the correlation coefficient between st January and 3 rd January etc.

18 . Data 5.9 Milan Figure 33. The autocorrelation coefficients between subsequent days in Milan. We can notice that the autocorrelation coefficients for lag equal to oscillate about the value.8 during the whole year, while the autocorrelation coefficients for larger lags are smaller in summer than in winter, see Figure 33.

19 Problem Application of change-point detection for annual maxima and minima

20 Chapter 3 The change-point detection for the GEV distributions In the first part of the thesis we study annual maximal and minimal temperatures. Figures 6 and Tables 3, 4 show behavior of the series under study. Extremes of random sequences are modelled by the GEV distribution, for details confer Appendix, Section A.. Our goal is applying the general Csörgő and Horváth theory for detecting a sudden change Appendix, Section A. of the GEV distribution Hx; µ, ψ, ξ with a density function hx; µ, ψ, ξ = + ξ x µ { ξ exp + ξ x µ } ξ, 3. ψ ψ ψ provided + ξx µ/ψ >. Notice that the support of the density function hx; µ, ψ, ξ depends on the parameters µ, ψ, ξ. Suppose that X,..., X n are independent random variables, we are to test the null hypothesis H against the alternative A : H : X i GEV µ, ψ, ξ, i =,..., n, 3. A : there exists k {,..., n n } such that X i GEV µ, ψ, ξ, i =,..., k, X i GEV µ, ψ, ξ, i = k +,..., n, where the parameters µ, ψ, ξ before the change point are known while µ, ψ, ξ µ, ψ, ξ are unknown or to test the null hypothesis H against the alternative A : A : there exists k { n,..., n n } such that X i GEV µ, ψ, ξ, i =,..., k, 3.3 X i GEV µ, ψ, ξ, i = k +,..., n, where neither the parameters before nor after the change point are known and µ, ψ, ξ µ, ψ, ξ. The constant n may be any fixed integer larger than three. However, to obtain a good estimates of all three parameters we need to have enough observations. For testing the problem 3.3 we may use the twice log likelihood ratio max logλ k = [L k ϕ k + L kϕ k L n ϕ n ], k n 7

21 3. The change-point detection for the GEV distributions 8 while testing the problem 3. yields in a simplified version max k n logλ k = max k n [L k ϕ k L kϕ ]. for more details we refer to Appendix, Remark A... To find critical values we have to find distribution of the test statistics under H. As the exact distribution of max k n logλ k, resp. max k n logλ k, under H are very complex, the approximate critical values can be found using the limit behavior of max k n logλ k, resp. max k n logλ k, see Appendix, Csörgő Horváth theorem. The conditions of Theorem A.. are classical regularity conditions for the existence of the maximum likelihood estimator. Since the density function 3. is defined on the set {x; +ξx µ/ψ > }, the classical regularity conditions for maximum likelihood estimators are not satisfied. The next problem is caused by the conditions C.4 and C.5 of Theorem A.., since they require the continuity of third derivatives. This can be weakened by Smith s Theorem A.3. and we will show that for ξ > there exists a sequence µ n, ψ n, ξ n of solutions of the likelihood equations such that n µ n µ, ψ n ψ, ξ n ξ converge in distribution to a zero mean normal vector with a variance covariance matrix M M is a Fisher information matrix and hence ξ > is still a regular case. We will proceed in two steps. At first we show theory for the three parameter Weibull distributions using the results of Smith s theorem, see Appendix Theorem A.3., and then we will focus on the Fréchet distributions. For those purposes we can use the following reparameterization of the GEV distribution. For ξ <, substituting θ = µ ψ ξ, β = ξ ψ ξ, α = ξ we obtain hx; θ, α, β = αβ x + θ α exp{ β x + θ α } for x θ, = for x < θ. 3.4 It is the three parameter Weibull distribution W eibθ, α, β of a random variable x not the Weibull distribution as a limit distribution for maxima concentrated on, θ from the Fisher Tippet theorem, see Appendix Theorem A... ξ, α =, we obtain a reparameteriza- ξ For ξ >, substituting θ = µ ψ, β = ξ tion ξ ψ hx; θ, α, β = αβx θ α exp{ βx θ α } for x θ, = for x < θ, 3.5 which corresponds to the Fréchet distribution F réch θ, α, β.

22 3. The change-point detection for the GEV distributions 9 Remark 3... Results for the Gumbel distribution from Fisher Tippet theorem A.. corresponding to a case ξ = are obtained by ξ in The change-point detection for the Weibull distributions The general theory presented in Csörgő and Horváth [7] was applied by Jandhyala et al. [?] to develop a test for detecting a sudden change in the two parameter Weibull distribution. However, in the case we use as a model the three parameter Weibull distribution W eib θ, α, β with a density function fx; θ, α, β = x θ α αβ exp{ βx θ α } for x θ, 3.6 = for x < θ, it seems more natural to look for a change in all three parameters. Notice that the support of the density function f is given by the parameter θ. Suppose that X,..., X n are independent random variables, we are to test the null hypothesis H against the alternative A : H : X i W eib θ, α, β, i =,..., n, 3.7 A : there exists k {,..., n n } such that X i W eib θ, α, β, i =,..., k, X i W eib θ, α, β, i = k +,..., n, where the parameters θ, α, β before the change point are known while θ, α, β θ, α, β are unknown or to test the null hypothesis H against the alternative A : A : there exists k { n,..., n n } such that X i W eib θ, α, β, i =,..., k, 3.8 X i W eib θ, α, β, i = k +,..., n, where neither the parameters before nor after the change point are known and θ, α, β θ, α, β. The constant n may be any fixed integer larger than three. For testing the problem 3.8 we may use the twice log likelihood ratio max logλ k = [L k ϕ k + L kϕ k L n ϕ n ], k n while testing the problem 3.7 yields in a simplified version max k n logλ k = max k n [L k ϕ k L kϕ ],

23 3. The change-point detection for the GEV distributions for more details we refer to Appendix. As the exact distribution of max k n logλ k, resp. max k n logλ k, under H are very complex, the approximate critical values can be found using the limit behavior of test statistics max k n logλ k, resp. max k n logλ k, see Appendix, Csörgő Horváth Theorem A... Let θ, α, β be the true values of the parameters under H. We will assume that θ R, α > and β >. The assumptions C.4. and C.5. of Csörgő and Horváth are not satisfied and Theorem A.. cannot be applied directly to get a limit distribution of max k n logλ k, resp. max k n logλ k. On the other hand, Smith showed, confer Appendix Theorem A.3., that for α > there exists a sequence θ n, α n, β n of solutions of the likelihood equations such that n θ n θ, α n α, β n β converge in distribution to a zero mean normal vector with a variance covariance matrix M M is a Fisher information matrix and hence α > is still a regular case. From here an idea comes that the assertion of Csörgő and Horváth theorem A.. is still valid for α >. 3. Main results for the Weibull distributions Our main results concern the asymptotic distribution of the statistic max k n logλ k under H for testing a change in all three parameters, when the parameters before a change point are known while after it they are unknown as well as the statistic max k n logλ k under H for testing a change in all three parameters, when the parameters both before and after a change point are unknown. We start with the characteristics of the log likelihood function L k. The first and second derivatives of L k θ, α, β may be expressed as follows: L k θ = i= L k α = k L k β = L k θ = k [ ] α X i θ + αβx i θ α, 3.9 [ logx i θ + ] α βx i θ α logx i θ, 3. i= k [ ] β X i θ α, 3. k [ ] α X i θ αα βx i θ α, 3. [ ] X i θ + βx i θ α + αβx i θ α logx i θ, 3.3 i= i= L k k θ α = i=

24 3. The change-point detection for the GEV distributions L k k θ β = [ αxi θ α ], 3.4 L k α = i= k i= [ ] α βx i θ α log X i θ, 3.5 L k k α β = [ X i θ α logx i θ], 3.6 L k β = i= k i= For simplicity we denote the true value of parameter [ β ]. 3.7 ϕ = θ, α, β, a maximum likelihood estimator based on X,..., X k when it exists by It holds ϕ k = θ k, α k, β k. E log fxi ; ϕ =, 3.8 θ E log fxi ; ϕ =, α E log fxi ; ϕ =. β Let s denote a Fisher information matrix M on a parameter ϕ = θ, α, β with elements m θθ m θα m θβ M = m αθ m αα m αβ, m βθ m βα m ββ where m θθ = E{ θ logfx i; ϕ θ logfx i; ϕ } = E{ θ logfx i; ϕ }, m αα = E{ α logfx i; ϕ α logfx i; ϕ } = E{ α logfx i; ϕ }, m ββ = E{ β logfx i; ϕ β logfx i; ϕ } = E{ β logfx i; ϕ },

25 3. The change-point detection for the GEV distributions m θα = m αθ = E{ θ logfx i; ϕ α logfx i; ϕ } = E{ θ α logfx i; ϕ }, m θβ = m βθ = E{ θ logfx i; ϕ β logfx i; ϕ } = E{ θ β logfx i; ϕ }, m αβ = m βα = E{ α logfx i; ϕ β logfx i; ϕ } = E{ α β logfx i; ϕ }. 3.9 A maximum likelihood estimator ϕ k = θ k, α k, β k satisfies L k θ ϕ k =, L k α ϕ k =, L k β ϕ k =. 3. The existence of ϕ k = θ k, α k, β k is guaranteed for α > by Theorem A.3.. If we consider positive moments only, then E log fxi ; ϕ r < θ if and only if r < α and for the same r we also have E log fxi ; ϕ r <, α E log fxi ; ϕ r <. β For the second derivatives E log fxi ; ϕ θ s < if and only if s < α and for the same s is also E log fxi ; ϕ s <, θ α E log fxi ; ϕ s <, θ β E log fxi ; ϕ α s <, E log fxi ; ϕ s <, α β E log fxi ; ϕ β s <.

26 3. The change-point detection for the GEV distributions 3 M is a positive definite matrix. According to the Marcinkiewicz-Zygmund law, Appendix - Theorem A.3., for any τ such that < τ < /α lim k kτ k θ L kϕ + m θθ = a.s., lim k kτ k θ α L kϕ + m θα = a.s., lim k kτ k lim k kτ k lim k kτ k lim k kτ k θ β L kϕ + m θβ = a.s., α L kϕ + m αα = a.s., α β L kϕ + m αβ = a.s., β L kϕ + m ββ = a.s. 3. By the law of the iterated logarithm, confer Appendix Theorem A.3.3 lim sup k lim sup k lim sup k L θ kϕ = O k log log k L α kϕ = O k log log k L β kϕ = O k log log k a.s., a.s., a.s. 3. We start with several technical lemmas on the three parameter Weibull distribution W eib θ, α, β. Lemma 3... Let X i W eib θ, α, β, then a for θ R, β > and α > 3 E b for θ R, β > and < α 3 k i= Proof. See Feller []. X i θ <, E X i θ <, E X i θ < 3 E X i θ <, E X i θ <, X i θ 3 = o k 3/α log k 3/α+ a.s. for some >. 3.3

27 3. The change-point detection for the GEV distributions 4 Lemma 3... Let X, X,..., X n are i.i.d. random variables, X i W eib θ, α, β. Under H for any >, the minimum X k and the second minimum X k satisfy Proof. See Jarušková [8]. = o log k + a.s. 3.4 X k θ X k θ Further, we prove the following lemma. We denote for any δ k > } I δk = { θ R, α R, β R; θ θ < δ k, α α < δ k, β β < δ k. Lemma For any sequence {δ k } satisfying δ k k /α+δ for some δ > and for any τ such that < τ < /α lim k kτ sup I δk k θ L k ϕ θ L kϕ = a.s., 3.5 lim k kτ sup I δk k θ α L k ϕ θ α L kϕ = a.s., 3.6 lim k kτ sup I δk k θ β L k ϕ θ β L kϕ = a.s., 3.7 lim k kτ sup I δk k α L k ϕ α L kϕ = a.s., 3.8 lim k kτ sup I δk k α β L k ϕ α β L kϕ = a.s., 3.9 lim k kτ sup k β L k ϕ β L kϕ = a.s. 3.3 I δk Proof. The proof is divided into two parts: the first part, rather lengthy, corresponds to the condition < α 3 and the second part corresponds to the condition α > 3. Let s suppose < α 3. discontinuous at X i = θ. We consider only terms in second derivatives, which are We start with proving 3.5. Substituting 3. into 3.5 and concentrating on terms discontinuous at X i = θ we obtain a following assertion to be examined: k lim k kτ sup I δk k X i θ = a.s., < τ < /α. 3.3 X i θ i=

28 3. The change-point detection for the GEV distributions 5 The proof can be found in Jarušková [8], Lemma 3. Next we look into the assertion 3.6. Substituting 3.3 into 3.6 we get following assertions to be proved k lim k kτ sup I δk k X i θ = a.s. 3.3 X i θ and lim k kτ sup I δk k i= k α βx i θ bα logx i θ i= α β X i θ α logx i θ = a.s We start with 3.3. Similarly as in Smith [6] and Jarušková [8] we write k k X i θ = X i θ i= kx k θ kx k θ k + k i= X ki θ X ki θ First, let X has the three parameter Weibull distribution with the density function 3.6. Then Z = X θ β /α has the Weibull distribution with parameters θ =, β = and α. Random variable Y = has the density Z fy = /y α+ α exp{ /y α } for y, 3.35 = for y <, with a finite moment EY r < for any r < α. Then, according to Theorem A.3.4, we get / k /r X k θ a.s. for any r < α, i.e. the second term on the right side of 3.34 satisfies / k X k θ = ok + r a.s. for any r < α. Second, choose r satisfying /α < /r < /α + δ, then we can write the first term on the right side of 3.34 as follows: Recall that δ k = o k /α+δ then X k θ =. θ θ X k θ b X k θ sup b θ θ <δ k θ θ X k θ = ok/r /k /α+δ = o a.s. 3.36

29 3. The change-point detection for the GEV distributions 6 In the other words, for two first terms in 3.34 sup kx k θ = ok + r a.s kx k θ for any r < α. b θ θ <δ k Further, for the third term in 3.34, again as in Smith [6] and Jarušková [8], using the Taylor expansion, there exists θ that θ θ < θ θ and k k i= X ki θ X ki θ = θ θ k k X ki θ i= We have to distinguish between two cases: i θ ˆθ > and ii θ ˆθ <. We start with i. For θ satisfying ˆθ < θ < θ we have in 3.38 θ θ k k i= X ki θ θ θ k k i= X ki θ. Applying Lemma 3.. for the right side of the above inequality, we obtain sup I δk k for any < υ < α. k X i θ = ok /α+δ O = ok υ a.s. X i θ i= For the proof of ii we use a characteristic of the first minimum that X ki θ X ki X k for every i =,..., n. We obtain following inequalities θ θ k k i= X ki θ θ θ k = θ θ k k i= i= X ki X k k X ki θ X k θ = θ θ k k i= θ θ X k θ k X k θ X ki θ X k θ X ki θ k i= X ki θ.

30 3. The change-point detection for the GEV distributions 7 Applying Lemma 3.. and Lemma 3.. for the factors on the right side of the above inequality, we obtain similarly as in the case i sup I δk k for any < υ < α. k X i θ = ok /α+δ olog k + O X i θ i= = ok υ a.s. Since we assume < α 3, we have α α implying for any < τ < α the rate of convergency of the third term of 3.34 sup I δk k k X i θ = ok τ a.s. X i θ i= and with the result 3.37 we have the assertion of 3.3. Now we prove the assertion The Taylor expansion implies that there exists θ, α, β, such that θ θ < θ θ, α α < α α, β β < β β and k k i= α βx i θ bα logx i θ α β X i θ α logx i θ = k k βx i θ logx α i θ α α i= + α βx i θ log α X i θ α α + αx i θ logx α i θ β β + α α βx i θ logx α i θ θ θ + α βx i θ α θ θ. 3.39

31 3. The change-point detection for the GEV distributions 8 The first term on the right side of 3.39 can be rewritten k k i= = k = k βx i θ logx α i θ α α k i= k i= βx i θ α log X i θ X i θ Xi θ α α α X i θ X i θ βx i θ α + θ θ X i θ α logxi θ + βx i θ α + θ θ α log + θ θ X i θ X i θ From inequality θ θ X i θ < θ θ X k θ for every i =,.... and 3.36 we have θ θ sup θ θ b <δ k X i θ = o a.s. α α. 3.4 Then Similarly sup b θ θ <δ k + θ α θ = O a.s. 3.4 X i θ log + θ θ log + θ θ K θ θ X i θ X k θ X k θ for some K R a.s. and then sup log + θ θ = o a.s. 3.4 θ θ b <δ k X i θ We can suppose that for sufficiently large k, is α > since we have α > and then EX i θ α logx i θ <, EX i θ α <. Therefore the first term in 3.39 satisfies sup I δk k k i= βx i θ logx α i θ α α = ok α +δ.

32 3. The change-point detection for the GEV distributions 9 The second term on the right side of 3.39 can be rewritten k k i= = k = k α βx i θ log α X i θ α α k α βx i θ α Xi θ i= [ k i= α X i θ log X i θ X i θ X i θ α βx i θ α + θ θ X i θ α log X i θ + α βx i θ α + θ θ X i θ α logxi θ log + α βx i θ α + θ θ X i θ α log Using inequalities 3.4, 3.4 and a property we get sup I δk k + θ θ X i θ EX i θ α log X i θ < for α > k i= + θ θ X i θ ] α βx i θ log α X i θ α α = ok α +δ. α α α α Similarly we can rewrite the other terms on the right side of 3.39 for which, supposing again α >, it holds EX i θ α logx i θ <, EX i θ α logx i θ <, EX i θ α <, 3.44 therefore we have for all the terms in 3.39 sup I δk k k i= for any < υ < α. α βx i θ bα logx i θ α β X i θ α logx i θ = ok υ a.s. Since we assume < α 3, we have α α implying for any < τ < α the rate of convergency in 3.33 sup I δk k k i= α βx i θ bα logx i θ α β X i θ α logx i θ = ok τ a.s.

33 3. The change-point detection for the GEV distributions 3 Now we investigate the assertion 3.8. Substituting 3.5 into 3.8 we have to prove k lim k kτ sup I δk k β log X i θx i θ bα + β log X i θ X i θ α i= = a.s From the Taylor expansion we get that there exists θ, α, β, such that θ θ < θ θ, α α < α α, β β < β β and k βx i k θ bα log X i θ β X i θ α log X i θ i= [ = k βx i log θ α 3 X i k θ α α i= + X i log θ α X i θ β β + βx i θ logx α i θ θ θ ] + α βx i θ log α X i θ θ θ Similarly as in the second term of 3.39 we have k sup sup I δk k βx i θ bα log X i θ + β X i θ α log X i θ I δk i= = ok υ a.s. for any < υ <. Since for < α 3 it holds and then α α α k sup sup I δk k βx i θ bα log X i θ + β X i θ α log X i θ I δk i= for any < τ < α. = ok τ a.s. That were all the terms in 3.5, 3.6, 3.7, 3.8, 3.9, 3.3 discontinuous at X i = θ. The proof for the condition α > 3 is trivial, as according to Lemma 3.., all the terms of Taylor expansions have finite expectations. The next theorem gives the convergency of the proposed maximum likelihood estimators.

34 3. The change-point detection for the GEV distributions 3 Theorem There exists a sequence of real number {δ k }, such that δ k k/ log log k and δk k /α+δ for some δ >, and there exists a set A with P A =, such that for any ω A we can find k ω, such that for all k k there exists a local maximum of L k θ, α, β denoted by ϕ k = θ k, α k, β k satisfying θ L k θ k, α k, β k =, α L k θ k, α k, β k =, β L k θ k, α k, β k = and θ k θ δ k, α k α δ k, β k β δ k. Proof. Similarly as in Smith [6], for any sequence {δ k } satisfying assumptions of Theorem 3..4 we define for t R, x R, y R the function f k t, x, y = δ k k L kθ + δ k t, α + δ k x, β o + δ k y. The Taylor expansion for any t R, x R, y R satisfying t + x + y implies that there exist t θ <, x θ <, ỹ θ < such that f k t t, x, y = f k t,, + f k t t θ, x θ, ỹ θ t + f k t x t θ, x θ, ỹ θ x+ + f k t y t θ, x θ, ỹ θ y = = L k δ k k θ ϕ + t L k k θ ϕ θ + x L k k θ α ϕ θ + y L k k θ β ϕ θ = L k δ k k θ ϕ + t L k k θ ϕ θ + x L k k θ α ϕ θ + y L k k θ β ϕ θ + t L k k θ ϕ + x L k k θ α ϕ + y L k k θ β ϕ t L k k θ ϕ x L k k θ α ϕ y L k k θ β ϕ + t m θθ t m θθ + x m θα x m θα + y m θβ y m θβ, 3.47 where m θθ, m θα and m θβ are the elements of the Fisher information matrix M, see the definition 3.9. For t θ <, x θ <, ỹ θ < we denote ϕ θ = θ θ, α θ, β θ such as θ θ = θ + t θ δ k, α θ = α + x θ δ k, βθ = β + ỹ θ δ k satisfying θ θ θ < δ k, α θ α < δ k, β θ β < δ k

35 3. The change-point detection for the GEV distributions 3 and we denote ε k,θ t θ, x θ, ỹ θ = L k δ k k θ ϕ + t L k k θ ϕ θ L k θ ϕ + x L k k θ α ϕ θ L k θ α ϕ + y L k k θ β ϕ θ L k θ β ϕ L k + t k θ ϕ + m θθ L k + x k θ α ϕ + m θα L k + y k θ β ϕ + m θβ. We can then rewrite 3.47 in a form f k t t, x, y = t m θθ x m θα y m θβ + ε k,θ t θ, x θ, ỹ θ, For the term ε k,θ t θ, x θ, ỹ θ it holds ε k,θ t θ, x θ, ỹ θ L k δ k k θ ϕ + t sup L k I δk k θ ϕ θ L k k θ ϕ + x sup L k I δk k θ α ϕ θ L k k θ α ϕ + y sup L k I δk k θ β ϕ θ L k k θ β ϕ + t L k k θ ϕ + m θθ + x L k k θ α ϕ + m θα + y L k k θ β ϕ + m θβ From the law of the iterated logarithm 3. and the characteristics of the sequence δ k that log log k δ k k, we get L k δ k k θ ϕ. Then, combining 3.5, 3.6, 3.7 with 3., we get that also all the next terms in 3.48 tend to and so we obtain ε k,θ t θ, x θ, ỹ θ a.s.

36 3. The change-point detection for the GEV distributions 33 Similarly, f k L k t, x, y = x δ k k α ϕ + t L k k α θ ϕ α + x L k k α ϕ α + y L k k α β ϕ α = t m θα x m αα y m αβ + ε k,α t α, x α, ỹ α, where m θα, m αα and m αβ are the elements of the Fisher information matrix M, see the definition 3.9. For t α <, x α <, ỹ α < we denote ϕ α = θ α, α α, β α, where θ α = θ + t α δ k, α α = α + x α δ k, β α = β + ỹ α δ k satisfying θ α θ < δ k, α α α < δ k, β α β < δ k. For ε k,α t α, x α, ỹ α we obtain following inequalities. ε k,α t α, x α, ỹ α L k δ k k α ϕ + t sup L k I δk k θ α ϕ α L k k θ α ϕ + x sup L k I δk k α ϕ α L k k α ϕ + y sup L k I δk k α β ϕ α L k k α β ϕ + t L k k θ α ϕ + m θα + x L k k α ϕ + m αα + y L k k α β ϕ + m αβ From the law of the iterated logarithm 3. and the characteristics of the sequence δ k that log log k δ k k, we get L k δ k k α ϕ. Then, combining 3.8, 3.6, 3.9 with 3., we get that also all the next terms in 3.49 tend to and so we obtain ε k,α t α, x α, ỹ α a.s. Further, f k L k t, x, y = y δ k k β ϕ + t L k k β θ ϕ β + x L k k α β ϕ β + y L k k β ϕ β = t m θβ x m αβ y m ββ + ε k,β t β, x β, ỹ β,

37 3. The change-point detection for the GEV distributions 34 where m θβ, m αβ and m ββ are the elements of the Fisher information matrix M, see the definition 3.9. For t β <, x β <, ỹ β < we denote ϕ β = θ β, α β, β β, where θ β = θ + t β δ k, α β = α + x β δ k, β β = β + ỹ β δ k satisfying θ β θ < δ k, α β α < δ k, β β β < δ k. For ε k,β t β, x β, ỹ β we have following inequalities ε k,β t β, x β, ỹ β L k δ k k β ϕ + t sup L k I δk k β θ ϕ β L k k β θ ϕ + x sup L k I δk k α β ϕ β L k k α β ϕ + y sup L k I δk k β ϕ β L k k β ϕ + t L k k θ β ϕ + m θβ + x L k k α β ϕ + m αβ + y L k k β ϕ + m ββ. 3.5 From the law of the iterated logarithm 3. and the characteristics of the sequence δ k that log log k δ k k, we get L k δ k k β ϕ. Then, combining 3.7, 3.9, 3.3 with 3., we get that also all the next terms in 3.5 tend to and so we obtain ε k,β t β, x β, ỹ β a.s. Let t + x + y =. Then we have t f k t + x f k x + y f k y = t m θθ x m αα y m ββ x t m θα + y t m θβ + x y m αβ + t ε k,θ t θ, x θ, ỹ θ + x ε k,α t α, x α, ỹ α + y ε k,β t β, x β, ỹ β 3.5 Since we proved that ε k,θ t θ, x θ, ỹ θ a.s., ε k,α t α, x α, ỹ α a.s., ε k,β t β, x β, ỹ β a.s., the expression 3.5 is as k strictly negative by the assumed positivedefiniteness of M. Hence Lemma A.3.5 shows that f k has a local maximum in the range t + x + y < as k.

38 3. The change-point detection for the GEV distributions 35 Clearly, there exists a sequence { θ k, α k, β k } and a set A with P A = such that for any ω A there exists k ω such that for all k k, θ k, α k, β k is a local maximum of L k θ, α, β satisfying θ L k θ k, α k, β k =, α L k θ k, α k, β k =, β L k θ k, α k, β k = and θ k θ δ k, α k α δ k, β k β δ k. The next lemma shows that, in the limit, the proposed maximum likelihood estimators behave as if they were partial sums of random vectors. Lemma For any τ such that < τ < /α and for k it holds L θ kϕ lim k kτ L k log log k α kϕ L β kϕ θk θ k M α k α β k β = a.s. 3.5 Proof. Now, for instance we choose the sequence δ k = log log log k log log k/k} satisfying conditions from Lemma For maximum likelihood estimators ϕ k = θ k, α k, β k we have = θ L k ϕ k = = θ L kϕ k θ k θ m θθ k α k α m θα k β k β m θβ + θ L k ϕ k θ L kϕ θ k θ + θ L k ϕ k θ α L kϕ α k α + θ β L k ϕ k θ β L kϕ β k β + θ L kϕ + k m θθ θ k θ + θ α L kϕ + k m θα α k α + θ β L kϕ + k m θβ β k β 3.53

39 3. The change-point detection for the GEV distributions 36 for ϕ θ = θ θ, α θ, β θ satisfying θ k θ < θ k θ, α k α < α k α, β k β < β k β. Then k log log k θ L kϕ k θ k θ m θθ k α k α m θα k β k β m θβ k = log log k k θ L k ϕ k θ L kϕ θ k θ k log log k k θ L k ϕ k θ α L kϕ α k α k log log k k θ β L k ϕ k θ β L kϕ β k β k log log k k θ L kϕ + k m θθ θ k θ k log log k k θ α L kϕ + k m θα α k α k log log k k θ β L kϕ + k m θβ β k β Using the Marcinkiewicz-Zygmund law 3., characteristics 3.5, 3.6, 3.7 and the characteristics of the sequence { δ k } that log log k θ k θ log log log k, k log log k α k α log log log k, k log log k β k β log log log k k 3.55 for the right side of 3.54 we get lim k kτ k log log k θ L kϕ k m θθ θ k θ k m θα α k α k m θβ β k β = a.s for any τ satisfying < τ < /α, where the coefficients m θθ, m θα, m θβ are the elements of the first line of the matrix M. From the similar expressions for derivatives L α k ϕ k and L β k ϕ k we obtain lim k kτ k log log k α L kϕ km αθ θ k θ km αα α k α km αβ β k β = a.s. 3.57

40 3. The change-point detection for the GEV distributions 37 lim k kτ k log log k β L kϕ km βθ θ k θ km βα α k α km ββ β k β = a.s., 3.58 where coefficients m αθ, m αα, m αβ are the elements of the second line of the matrix M and where coefficients m βθ, m βα, m ββ are the elements of the third line of the matrix M. From 3.56, 3.57, 3.58 we get the assertion 3.5. The next corollary gives the rate of convergency of the proposed maximum likelihood estimators. Corollary The sequence of the proposed maximum likelihood estimators ϕ k = θ k, α k, β k from Theorem 3..4 satisfies k lim sup α k α = O a.s., k log log k k lim sup β k β = O a.s., k log log k lim sup k k log log k θ k θ = O a.s Proof. The proof is an easy consequence of 3.5 and the law of iterated logarithm 3.. Corollary For any τ such that < τ < /α it holds lim k kτ k θ L kϕ, α L kϕ, β L kϕ M L θ kϕ L α kϕ L β kϕ k θk θ, α k α, β θk θ k β M α k α = a.s. 3.6 β k β Proof. For a matrix P, such that P T P = M, we can write equation 3.5 as follows: L lim k kτ P T θ kϕ L log log k α kϕ θk θ kp α k α = a.s. 3.6 k L β kϕ β k β For k we have k θ L kϕ, α L kϕ, β L kϕ P + k θk θ, α k α, β k β P T = O log log k a.s. 3.6 and combining equations 3.6 and 3.6 we obtain the assertion of Corollary 3..7.

41 3. The change-point detection for the GEV distributions 38 Theorem For any τ such that < τ < /α and for k k τ L k ϕ k L k ϕ k θ L kϕ, α L kϕ, β L kϕ M L θ kϕ L α kϕ L β kϕ a.s. Proof. The Taylor expansion L k ϕ k L k ϕ = D cϕk ϕ L k ϕ k + D cϕ k ϕ L k ϕ, 3.63 where D cϕk ϕ L k ϕ k is the first differential at the point ϕ k in the direction ϕ k ϕ, D cϕ k ϕ L k ϕ is the second differential at the point ϕ in the direction ϕ k ϕ and θ θ < θ k θ, α α < α k α, β β < β k β. We can rewrite 3.63 L k ϕ k L k ϕ k θ L kϕ, α L kϕ, β L kϕ M L θ kϕ L α kϕ L β kϕ = D cϕ k ϕ L k ϕ D cϕ k ϕ L k ϕ + D cϕ k ϕ L k ϕ k θk θ, α k α, β θk θ k β M α k α β k β + k θk θ, α k α, β θk θ k β M α k α β k β k θ L kϕ, α L kϕ, For the differences on the right side of 3.64 we obtain: β L kϕ M L θ kϕ L α kϕ L β kϕ. 3.64

42 3. The change-point detection for the GEV distributions 39 - The difference of second differentials is Dcϕ k ϕ L k ϕ Dcϕ k ϕ L k ϕ = θ L k ϕ θ L kϕ θ k θ + α L k ϕ α L kϕ α k α + β L k ϕ β L kϕ β k β + θ α L k ϕ θ α L kϕ θ k θ α k α + θ β L k ϕ θ β L kϕ θ k θ β k β + α β L k ϕ α β L kϕ α k α β k β and its elements are ok τ according to Lemma 3..5 and Corollary Similarly for the difference Dcϕ k ϕ L k ϕ k θk θ, α k α, β θk θ k β M α k α β k β = θ L kϕ + k m θθ θ k θ + α L kϕ + k m αα α k α + β L kϕ + k m ββ β k β + θ α L kϕ + k m θα θ k θ α k α + θ β L kϕ + k m θβ θ k θ β k β + α β L kϕ + k m αβ α k α β k β and its elements are ok τ according to the Marcinkiewicz-Zygmund law 3. and Corollary 3..6.

43 3. The change-point detection for the GEV distributions 4 - The difference k θk θ, α k α, β θk θ k β M α k α β k β k θ L kϕ, α L kϕ, β L kϕ M L θ kϕ L α kϕ L β kϕ is ok τ according to the Corollary Summarizing these three results we obtain that the right side of 3.64 is ok τ a.s. We introduce Ax = log x and D d x = log x+d/ log log x log Γd/, similarly as in Theorem A... Theorem The asymptotic distribution of the maximum likelihood statistic for testing the problem 3.7 under H provided α > is given by / lim P Alog n max n k n logλ k t + D 3 logn = exp e t and for the maximum likelihood statistic for testing the problem 3.8 we have for all t R. / lim P Alog n max logλ k t + D 3 logn n k n = exp e t Proof. Using Theorem 3..8 we can similarly as in Csörgő and Horváth [7] prove that max Lk ϕ k L k ϕ k n max k n k θ L kϕ, α L kϕ, β L kϕ M L θ kϕ L α kϕ L β kϕ = o P log log n and the assertion of Theorem 3..9 is an easy consequence. Now, coming back to the GEV distribution and using parameters µ, ψ, ξ we can write the Theorem 3..9 as follows.

44 3. The change-point detection for the GEV distributions 4 Theorem 3... Provided < ξ <, the asymptotic distribution of the maximum likelihood statistic for testing the problem 3. under H is given by lim P / Alog n max n k n logλ k t + D 3 logn = exp e t and for the maximum likelihood statistic for testing the problem 3.3 we have for all t R. / lim P Alog n max logλ k t + D 3 logn n k n Proof. An easy consequence of the inequality α >. = exp e t 3.3 The change-point detection for the Fréchet distributions Now we concentrate on proving a similar theorem as Theorem 3.. for parameter ξ > corresponding to the Fréchet distribution F réch θ, α, β with the density function 3.5 hx; θ, α, β = αβx θ α exp{ βx θ α } for x θ, = for x < θ. Suppose that X,..., X n are independent random variables, we are to test the null hypothesis H against the alternative A : H : X i F réch θ, α, β, i =,..., n, 3.65 A : there exists k {,..., n n } such that X i F réch θ, α, β, i =,..., k, X i F réch θ, α, β, i = k +,..., n, where the parameters θ, α, β before the change point are known while θ, α, β θ, α, β are unknown or to test the null hypothesis H against the alternative A : A : there exists k { n,..., n n } such that X i F réch θ, α, β, i =,..., k, 3.66 X i F réch θ, α, β, i = k +,..., n, where neither the parameters before nor after the change point are known and θ, α, β θ, α, β. The constant n may be any fixed integer larger than three, α > is an unknown shape parameter, β > is an unknown scale parameter and θ R is an unknown location parameter.

45 3. The change-point detection for the GEV distributions 4 Our goal is to find the limit distribution of max k n logλ k for the problem 3.66, resp. max k n logλ k for the problem At first we show an important characteristic that the right tail of the density function hx; θ, α, β defined in 3.5 decreases faster than any power of x θ. Lemma For every m R Proof. It is an easy consequence of a limit lim x x θm hx; θ, α, β for x θ lim y y p = for every p R. ey The log likelihood of 3.5 is given by k L k θ, α, β = k log α + k log β + α log X i θ i= k β X i θ α i= First and second derivatives of L k θ, α, β are: L k θ = i= L k α = k L k β = L k θ = k [ ] α + X i θ αβx i θ α, [ logx i θ + α βx i θ α logx i θ i= i= i= L k k θ α = k [ ] β X i θ α, k [ ] α + X i θ αα + βx i θ α, [ X i θ βx i θ α αβx i θ α logx i θ i= L k k θ β = L k α = i= k i= [ αxi θ α ], [ ] α βx i θ α log X i θ, ], ],

46 3. The change-point detection for the GEV distributions 43 L k k α β = [ Xi θ α logx i θ ], L k β = i= k i= [ β ] It holds E log hxi ; ϕ =, 3.69 θ E log hxi ; ϕ =, α E log hxi ; ϕ =. β According to Lemma 3.3., for every s R it holds E log hxi ; ϕ θ s <, E log hxi ; ϕ s <, θ α E log hxi ; ϕ s <, θ β E log hxi ; ϕ α s <, E log hxi ; ϕ s <, α β E log hxi ; ϕ β s <. Let s denote a Fisher information matrix M on a parameter ϕ = θ, α, β with elements m θθ m θα m θβ M = m αθ m αα m αβ, m βθ m βα m ββ where m θθ = E{ θ loghx i; ϕ θ loghx i; ϕ } = E{ θ loghx i; ϕ }, m αα = E{ α loghx i; ϕ α loghx i; ϕ } = E{ α loghx i; ϕ },

MFM Practitioner Module: Quantitiative Risk Management. John Dodson. October 14, 2015

MFM Practitioner Module: Quantitiative Risk Management. John Dodson. October 14, 2015 MFM Practitioner Module: Quantitiative Risk Management October 14, 2015 The n-block maxima 1 is a random variable defined as M n max (X 1,..., X n ) for i.i.d. random variables X i with distribution function

More information

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Put your solution to each problem on a separate sheet of paper. Problem 1. (5166) Assume that two random samples {x i } and {y i } are independently

More information

The largest eigenvalues of the sample covariance matrix. in the heavy-tail case

The largest eigenvalues of the sample covariance matrix. in the heavy-tail case The largest eigenvalues of the sample covariance matrix 1 in the heavy-tail case Thomas Mikosch University of Copenhagen Joint work with Richard A. Davis (Columbia NY), Johannes Heiny (Aarhus University)

More information

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a

More information

Extreme Value Analysis and Spatial Extremes

Extreme Value Analysis and Spatial Extremes Extreme Value Analysis and Department of Statistics Purdue University 11/07/2013 Outline Motivation 1 Motivation 2 Extreme Value Theorem and 3 Bayesian Hierarchical Models Copula Models Max-stable Models

More information

An exponential family of distributions is a parametric statistical model having densities with respect to some positive measure λ of the form.

An exponential family of distributions is a parametric statistical model having densities with respect to some positive measure λ of the form. Stat 8112 Lecture Notes Asymptotics of Exponential Families Charles J. Geyer January 23, 2013 1 Exponential Families An exponential family of distributions is a parametric statistical model having densities

More information

CHANGE DETECTION IN TIME SERIES

CHANGE DETECTION IN TIME SERIES CHANGE DETECTION IN TIME SERIES Edit Gombay TIES - 2008 University of British Columbia, Kelowna June 8-13, 2008 Outline Introduction Results Examples References Introduction sunspot.year 0 50 100 150 1700

More information

EXTREMAL MODELS AND ENVIRONMENTAL APPLICATIONS. Rick Katz

EXTREMAL MODELS AND ENVIRONMENTAL APPLICATIONS. Rick Katz 1 EXTREMAL MODELS AND ENVIRONMENTAL APPLICATIONS Rick Katz Institute for Study of Society and Environment National Center for Atmospheric Research Boulder, CO USA email: rwk@ucar.edu Home page: www.isse.ucar.edu/hp_rick/

More information

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,

More information

For iid Y i the stronger conclusion holds; for our heuristics ignore differences between these notions.

For iid Y i the stronger conclusion holds; for our heuristics ignore differences between these notions. Large Sample Theory Study approximate behaviour of ˆθ by studying the function U. Notice U is sum of independent random variables. Theorem: If Y 1, Y 2,... are iid with mean µ then Yi n µ Called law of

More information

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3 Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest

More information

Homework 7: Solutions. P3.1 from Lehmann, Romano, Testing Statistical Hypotheses.

Homework 7: Solutions. P3.1 from Lehmann, Romano, Testing Statistical Hypotheses. Stat 300A Theory of Statistics Homework 7: Solutions Nikos Ignatiadis Due on November 28, 208 Solutions should be complete and concisely written. Please, use a separate sheet or set of sheets for each

More information

Change Point Analysis of Extreme Values

Change Point Analysis of Extreme Values Change Point Analysis of Extreme Values TIES 2008 p. 1/? Change Point Analysis of Extreme Values Goedele Dierckx Economische Hogeschool Sint Aloysius, Brussels, Belgium e-mail: goedele.dierckx@hubrussel.be

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

1 Directional Derivatives and Differentiability

1 Directional Derivatives and Differentiability Wednesday, January 18, 2012 1 Directional Derivatives and Differentiability Let E R N, let f : E R and let x 0 E. Given a direction v R N, let L be the line through x 0 in the direction v, that is, L :=

More information

Review. December 4 th, Review

Review. December 4 th, Review December 4 th, 2017 Att. Final exam: Course evaluation Friday, 12/14/2018, 10:30am 12:30pm Gore Hall 115 Overview Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 6: Statistics and Sampling Distributions Chapter

More information

1 General problem. 2 Terminalogy. Estimation. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ).

1 General problem. 2 Terminalogy. Estimation. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ). Estimation February 3, 206 Debdeep Pati General problem Model: {P θ : θ Θ}. Observe X P θ, θ Θ unknown. Estimate θ. (Pick a plausible distribution from family. ) Or estimate τ = τ(θ). Examples: θ = (µ,

More information

Computer Intensive Methods in Mathematical Statistics

Computer Intensive Methods in Mathematical Statistics Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of

More information

Statistical Inference

Statistical Inference Statistical Inference Robert L. Wolpert Institute of Statistics and Decision Sciences Duke University, Durham, NC, USA Week 12. Testing and Kullback-Leibler Divergence 1. Likelihood Ratios Let 1, 2, 2,...

More information

Statistics & Data Sciences: First Year Prelim Exam May 2018

Statistics & Data Sciences: First Year Prelim Exam May 2018 Statistics & Data Sciences: First Year Prelim Exam May 2018 Instructions: 1. Do not turn this page until instructed to do so. 2. Start each new question on a new sheet of paper. 3. This is a closed book

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth

More information

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function

More information

TESTING A HOMOGENEITY OF STOCHASTIC PROCESSES

TESTING A HOMOGENEITY OF STOCHASTIC PROCESSES K Y B E R N E I K A V O L U M E 4 3 0 0 7, N U M B E R 4, P A G E S 4 5 4 3 0 ESING A HOMOGENEIY OF SOCHASIC PROCESSES Jaromír Antoch and Daniela Jarušová he paper concentrates on modeling the data that

More information

INFORMATION CRITERION AND CHANGE POINT PROBLEM FOR REGULAR MODELS 1. Jianmin Pan

INFORMATION CRITERION AND CHANGE POINT PROBLEM FOR REGULAR MODELS 1. Jianmin Pan INFORMATION CRITERION AND CHANGE POINT PROBLEM FOR REGULAR MODELS 1 Jiahua Chen Department of Statistics & Actuarial Science University of Waterloo Waterloo, Ontario, Canada N2L 3G1 A. K. Gupta Department

More information

First Year Examination Department of Statistics, University of Florida

First Year Examination Department of Statistics, University of Florida First Year Examination Department of Statistics, University of Florida August 20, 2009, 8:00 am - 2:00 noon Instructions:. You have four hours to answer questions in this examination. 2. You must show

More information

Shape of the return probability density function and extreme value statistics

Shape of the return probability density function and extreme value statistics Shape of the return probability density function and extreme value statistics 13/09/03 Int. Workshop on Risk and Regulation, Budapest Overview I aim to elucidate a relation between one field of research

More information

STAT 830 Hypothesis Testing

STAT 830 Hypothesis Testing STAT 830 Hypothesis Testing Hypothesis testing is a statistical problem where you must choose, on the basis of data X, between two alternatives. We formalize this as the problem of choosing between two

More information

Richard L. Smith Department of Statistics and Operations Research University of North Carolina Chapel Hill, NC

Richard L. Smith Department of Statistics and Operations Research University of North Carolina Chapel Hill, NC EXTREME VALUE THEORY Richard L. Smith Department of Statistics and Operations Research University of North Carolina Chapel Hill, NC 27599-3260 rls@email.unc.edu AMS Committee on Probability and Statistics

More information

EXTREMAL QUANTILES OF MAXIMUMS FOR STATIONARY SEQUENCES WITH PSEUDO-STATIONARY TREND WITH APPLICATIONS IN ELECTRICITY CONSUMPTION ALEXANDR V.

EXTREMAL QUANTILES OF MAXIMUMS FOR STATIONARY SEQUENCES WITH PSEUDO-STATIONARY TREND WITH APPLICATIONS IN ELECTRICITY CONSUMPTION ALEXANDR V. MONTENEGRIN STATIONARY JOURNAL TREND WITH OF ECONOMICS, APPLICATIONS Vol. IN 9, ELECTRICITY No. 4 (December CONSUMPTION 2013), 53-63 53 EXTREMAL QUANTILES OF MAXIMUMS FOR STATIONARY SEQUENCES WITH PSEUDO-STATIONARY

More information

Chapter 3. Point Estimation. 3.1 Introduction

Chapter 3. Point Estimation. 3.1 Introduction Chapter 3 Point Estimation Let (Ω, A, P θ ), P θ P = {P θ θ Θ}be probability space, X 1, X 2,..., X n : (Ω, A) (IR k, B k ) random variables (X, B X ) sample space γ : Θ IR k measurable function, i.e.

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random

More information

Math 576: Quantitative Risk Management

Math 576: Quantitative Risk Management Math 576: Quantitative Risk Management Haijun Li lih@math.wsu.edu Department of Mathematics Washington State University Week 11 Haijun Li Math 576: Quantitative Risk Management Week 11 1 / 21 Outline 1

More information

Gibbs Sampling in Linear Models #2

Gibbs Sampling in Linear Models #2 Gibbs Sampling in Linear Models #2 Econ 690 Purdue University Outline 1 Linear Regression Model with a Changepoint Example with Temperature Data 2 The Seemingly Unrelated Regressions Model 3 Gibbs sampling

More information

Weighted Exponential Distribution and Process

Weighted Exponential Distribution and Process Weighted Exponential Distribution and Process Jilesh V Some generalizations of exponential distribution and related time series models Thesis. Department of Statistics, University of Calicut, 200 Chapter

More information

P n. This is called the law of large numbers but it comes in two forms: Strong and Weak.

P n. This is called the law of large numbers but it comes in two forms: Strong and Weak. Large Sample Theory Large Sample Theory is a name given to the search for approximations to the behaviour of statistical procedures which are derived by computing limits as the sample size, n, tends to

More information

Comprehensive Examination Quantitative Methods Spring, 2018

Comprehensive Examination Quantitative Methods Spring, 2018 Comprehensive Examination Quantitative Methods Spring, 2018 Instruction: This exam consists of three parts. You are required to answer all the questions in all the parts. 1 Grading policy: 1. Each part

More information

Ch. 5 Hypothesis Testing

Ch. 5 Hypothesis Testing Ch. 5 Hypothesis Testing The current framework of hypothesis testing is largely due to the work of Neyman and Pearson in the late 1920s, early 30s, complementing Fisher s work on estimation. As in estimation,

More information

δ -method and M-estimation

δ -method and M-estimation Econ 2110, fall 2016, Part IVb Asymptotic Theory: δ -method and M-estimation Maximilian Kasy Department of Economics, Harvard University 1 / 40 Example Suppose we estimate the average effect of class size

More information

X n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2)

X n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2) 14:17 11/16/2 TOPIC. Convergence in distribution and related notions. This section studies the notion of the so-called convergence in distribution of real random variables. This is the kind of convergence

More information

Statistical Inference

Statistical Inference Statistical Inference Classical and Bayesian Methods Revision Class for Midterm Exam AMS-UCSC Th Feb 9, 2012 Winter 2012. Session 1 (Revision Class) AMS-132/206 Th Feb 9, 2012 1 / 23 Topics Topics We will

More information

INFORMATION APPROACH FOR CHANGE POINT DETECTION OF WEIBULL MODELS WITH APPLICATIONS. Tao Jiang. A Thesis

INFORMATION APPROACH FOR CHANGE POINT DETECTION OF WEIBULL MODELS WITH APPLICATIONS. Tao Jiang. A Thesis INFORMATION APPROACH FOR CHANGE POINT DETECTION OF WEIBULL MODELS WITH APPLICATIONS Tao Jiang A Thesis Submitted to the Graduate College of Bowling Green State University in partial fulfillment of the

More information

1. Let A be a 2 2 nonzero real matrix. Which of the following is true?

1. Let A be a 2 2 nonzero real matrix. Which of the following is true? 1. Let A be a 2 2 nonzero real matrix. Which of the following is true? (A) A has a nonzero eigenvalue. (B) A 2 has at least one positive entry. (C) trace (A 2 ) is positive. (D) All entries of A 2 cannot

More information

Lecture 2 APPLICATION OF EXREME VALUE THEORY TO CLIMATE CHANGE. Rick Katz

Lecture 2 APPLICATION OF EXREME VALUE THEORY TO CLIMATE CHANGE. Rick Katz 1 Lecture 2 APPLICATION OF EXREME VALUE THEORY TO CLIMATE CHANGE Rick Katz Institute for Study of Society and Environment National Center for Atmospheric Research Boulder, CO USA email: rwk@ucar.edu Home

More information

Fractional integration and cointegration: The representation theory suggested by S. Johansen

Fractional integration and cointegration: The representation theory suggested by S. Johansen : The representation theory suggested by S. Johansen hhu@ssb.no people.ssb.no/hhu www.hungnes.net January 31, 2007 Univariate fractional process Consider the following univariate process d x t = ε t, (1)

More information

Introduction to Algorithmic Trading Strategies Lecture 10

Introduction to Algorithmic Trading Strategies Lecture 10 Introduction to Algorithmic Trading Strategies Lecture 10 Risk Management Haksun Li haksun.li@numericalmethod.com www.numericalmethod.com Outline Value at Risk (VaR) Extreme Value Theory (EVT) References

More information

The properties of L p -GMM estimators

The properties of L p -GMM estimators The properties of L p -GMM estimators Robert de Jong and Chirok Han Michigan State University February 2000 Abstract This paper considers Generalized Method of Moment-type estimators for which a criterion

More information

Review: General Approach to Hypothesis Testing. 1. Define the research question and formulate the appropriate null and alternative hypotheses.

Review: General Approach to Hypothesis Testing. 1. Define the research question and formulate the appropriate null and alternative hypotheses. 1 Review: Let X 1, X,..., X n denote n independent random variables sampled from some distribution might not be normal!) with mean µ) and standard deviation σ). Then X µ σ n In other words, X is approximately

More information

Discrete time processes

Discrete time processes Discrete time processes Predictions are difficult. Especially about the future Mark Twain. Florian Herzog 2013 Modeling observed data When we model observed (realized) data, we encounter usually the following

More information

GMM and SMM. 1. Hansen, L Large Sample Properties of Generalized Method of Moments Estimators, Econometrica, 50, p

GMM and SMM. 1. Hansen, L Large Sample Properties of Generalized Method of Moments Estimators, Econometrica, 50, p GMM and SMM Some useful references: 1. Hansen, L. 1982. Large Sample Properties of Generalized Method of Moments Estimators, Econometrica, 50, p. 1029-54. 2. Lee, B.S. and B. Ingram. 1991 Simulation estimation

More information

Mathematics Ph.D. Qualifying Examination Stat Probability, January 2018

Mathematics Ph.D. Qualifying Examination Stat Probability, January 2018 Mathematics Ph.D. Qualifying Examination Stat 52800 Probability, January 2018 NOTE: Answers all questions completely. Justify every step. Time allowed: 3 hours. 1. Let X 1,..., X n be a random sample from

More information

Statistics 3858 : Maximum Likelihood Estimators

Statistics 3858 : Maximum Likelihood Estimators Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,

More information

MATHEMATICAL ECONOMICS: OPTIMIZATION. Contents

MATHEMATICAL ECONOMICS: OPTIMIZATION. Contents MATHEMATICAL ECONOMICS: OPTIMIZATION JOÃO LOPES DIAS Contents 1. Introduction 2 1.1. Preliminaries 2 1.2. Optimal points and values 2 1.3. The optimization problems 3 1.4. Existence of optimal points 4

More information

Estimates for probabilities of independent events and infinite series

Estimates for probabilities of independent events and infinite series Estimates for probabilities of independent events and infinite series Jürgen Grahl and Shahar evo September 9, 06 arxiv:609.0894v [math.pr] 8 Sep 06 Abstract This paper deals with finite or infinite sequences

More information

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes Neyman-Pearson paradigm. Suppose that a researcher is interested in whether the new drug works. The process of determining whether the outcome of the experiment points to yes or no is called hypothesis

More information

Exercises and Answers to Chapter 1

Exercises and Answers to Chapter 1 Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean

More information

A note on unsatisfiable k-cnf formulas with few occurrences per variable

A note on unsatisfiable k-cnf formulas with few occurrences per variable A note on unsatisfiable k-cnf formulas with few occurrences per variable Shlomo Hoory Department of Computer Science University of British Columbia Vancouver, Canada shlomoh@cs.ubc.ca Stefan Szeider Department

More information

HIERARCHICAL MODELS IN EXTREME VALUE THEORY

HIERARCHICAL MODELS IN EXTREME VALUE THEORY HIERARCHICAL MODELS IN EXTREME VALUE THEORY Richard L. Smith Department of Statistics and Operations Research, University of North Carolina, Chapel Hill and Statistical and Applied Mathematical Sciences

More information

Questions and Answers on Unit Roots, Cointegration, VARs and VECMs

Questions and Answers on Unit Roots, Cointegration, VARs and VECMs Questions and Answers on Unit Roots, Cointegration, VARs and VECMs L. Magee Winter, 2012 1. Let ɛ t, t = 1,..., T be a series of independent draws from a N[0,1] distribution. Let w t, t = 1,..., T, be

More information

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics Jiti Gao Department of Statistics School of Mathematics and Statistics The University of Western Australia Crawley

More information

Static Problem Set 2 Solutions

Static Problem Set 2 Solutions Static Problem Set Solutions Jonathan Kreamer July, 0 Question (i) Let g, h be two concave functions. Is f = g + h a concave function? Prove it. Yes. Proof: Consider any two points x, x and α [0, ]. Let

More information

Classical Extreme Value Theory - An Introduction

Classical Extreme Value Theory - An Introduction Chapter 1 Classical Extreme Value Theory - An Introduction 1.1 Introduction Asymptotic theory of functions of random variables plays a very important role in modern statistics. The objective of the asymptotic

More information

EM Algorithm II. September 11, 2018

EM Algorithm II. September 11, 2018 EM Algorithm II September 11, 2018 Review EM 1/27 (Y obs, Y mis ) f (y obs, y mis θ), we observe Y obs but not Y mis Complete-data log likelihood: l C (θ Y obs, Y mis ) = log { f (Y obs, Y mis θ) Observed-data

More information

Extreme Value Theory and Applications

Extreme Value Theory and Applications Extreme Value Theory and Deauville - 04/10/2013 Extreme Value Theory and Introduction Asymptotic behavior of the Sum Extreme (from Latin exter, exterus, being on the outside) : Exceeding the ordinary,

More information

Abrupt change in mean avoiding variance estimation and block bootstrap. Barbora Peštová The Czech Academy of Sciences Institute of Computer Science

Abrupt change in mean avoiding variance estimation and block bootstrap. Barbora Peštová The Czech Academy of Sciences Institute of Computer Science Abrupt change in mean avoiding variance estimation and block bootstrap Barbora Peštová The Czech Academy of Sciences Institute of Computer Science Structural changes 2 24 Motivation To know whether a change

More information

On the number of ways of writing t as a product of factorials

On the number of ways of writing t as a product of factorials On the number of ways of writing t as a product of factorials Daniel M. Kane December 3, 005 Abstract Let N 0 denote the set of non-negative integers. In this paper we prove that lim sup n, m N 0 : n!m!

More information

Chapter 12: An introduction to Time Series Analysis. Chapter 12: An introduction to Time Series Analysis

Chapter 12: An introduction to Time Series Analysis. Chapter 12: An introduction to Time Series Analysis Chapter 12: An introduction to Time Series Analysis Introduction In this chapter, we will discuss forecasting with single-series (univariate) Box-Jenkins models. The common name of the models is Auto-Regressive

More information

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales.

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. Lecture 2 1 Martingales We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. 1.1 Doob s inequality We have the following maximal

More information

Lecture 2: Univariate Time Series

Lecture 2: Univariate Time Series Lecture 2: Univariate Time Series Analysis: Conditional and Unconditional Densities, Stationarity, ARMA Processes Prof. Massimo Guidolin 20192 Financial Econometrics Spring/Winter 2017 Overview Motivation:

More information

On the lines passing through two conjugates of a Salem number

On the lines passing through two conjugates of a Salem number Under consideration for publication in Math. Proc. Camb. Phil. Soc. 1 On the lines passing through two conjugates of a Salem number By ARTŪRAS DUBICKAS Department of Mathematics and Informatics, Vilnius

More information

Ph.D. Qualifying Exam Monday Tuesday, January 4 5, 2016

Ph.D. Qualifying Exam Monday Tuesday, January 4 5, 2016 Ph.D. Qualifying Exam Monday Tuesday, January 4 5, 2016 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Find the maximum likelihood estimate of θ where θ is a parameter

More information

Parameter Estimation

Parameter Estimation Parameter Estimation Consider a sample of observations on a random variable Y. his generates random variables: (y 1, y 2,, y ). A random sample is a sample (y 1, y 2,, y ) where the random variables y

More information

Reliable Inference in Conditions of Extreme Events. Adriana Cornea

Reliable Inference in Conditions of Extreme Events. Adriana Cornea Reliable Inference in Conditions of Extreme Events by Adriana Cornea University of Exeter Business School Department of Economics ExISta Early Career Event October 17, 2012 Outline of the talk Extreme

More information

Econ 101A Problem Set 1 Solution

Econ 101A Problem Set 1 Solution Econ 101A Problem Set 1 Solution Problem 1. Univariate unconstrained maximization. (10 points) Consider the following maximization problem: max x f(x; x 0)=exp( (x x 0 ) 2 ) 1. Write down the first order

More information

Lecture 06 01/31/ Proofs for emergence of giant component

Lecture 06 01/31/ Proofs for emergence of giant component M375T/M396C: Topics in Complex Networks Spring 2013 Lecture 06 01/31/13 Lecturer: Ravi Srinivasan Scribe: Tianran Geng 6.1 Proofs for emergence of giant component We now sketch the main ideas underlying

More information

Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama

Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama Instructions This exam has 7 pages in total, numbered 1 to 7. Make sure your exam has all the pages. This exam will be 2 hours

More information

Optimization. The value x is called a maximizer of f and is written argmax X f. g(λx + (1 λ)y) < λg(x) + (1 λ)g(y) 0 < λ < 1; x, y X.

Optimization. The value x is called a maximizer of f and is written argmax X f. g(λx + (1 λ)y) < λg(x) + (1 λ)g(y) 0 < λ < 1; x, y X. Optimization Background: Problem: given a function f(x) defined on X, find x such that f(x ) f(x) for all x X. The value x is called a maximizer of f and is written argmax X f. In general, argmax X f may

More information

STA205 Probability: Week 8 R. Wolpert

STA205 Probability: Week 8 R. Wolpert INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and

More information

If we want to analyze experimental or simulated data we might encounter the following tasks:

If we want to analyze experimental or simulated data we might encounter the following tasks: Chapter 1 Introduction If we want to analyze experimental or simulated data we might encounter the following tasks: Characterization of the source of the signal and diagnosis Studying dependencies Prediction

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)

More information

Stability of Stochastic Differential Equations

Stability of Stochastic Differential Equations Lyapunov stability theory for ODEs s Stability of Stochastic Differential Equations Part 1: Introduction Department of Mathematics and Statistics University of Strathclyde Glasgow, G1 1XH December 2010

More information

Asymptotic Statistics-III. Changliang Zou

Asymptotic Statistics-III. Changliang Zou Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (

More information

GARCH Models Estimation and Inference. Eduardo Rossi University of Pavia

GARCH Models Estimation and Inference. Eduardo Rossi University of Pavia GARCH Models Estimation and Inference Eduardo Rossi University of Pavia Likelihood function The procedure most often used in estimating θ 0 in ARCH models involves the maximization of a likelihood function

More information

INSTITUTE AND FACULTY OF ACTUARIES. Curriculum 2019 SPECIMEN SOLUTIONS

INSTITUTE AND FACULTY OF ACTUARIES. Curriculum 2019 SPECIMEN SOLUTIONS INSTITUTE AND FACULTY OF ACTUARIES Curriculum 09 SPECIMEN SOLUTIONS Subject CSA Risk Modelling and Survival Analysis Institute and Faculty of Actuaries Sample path A continuous time, discrete state process

More information

Correlation 1. December 4, HMS, 2017, v1.1

Correlation 1. December 4, HMS, 2017, v1.1 Correlation 1 December 4, 2017 1 HMS, 2017, v1.1 Chapter References Diez: Chapter 7 Navidi, Chapter 7 I don t expect you to learn the proofs what will follow. Chapter References 2 Correlation The sample

More information

Submitted to the Brazilian Journal of Probability and Statistics

Submitted to the Brazilian Journal of Probability and Statistics Submitted to the Brazilian Journal of Probability and Statistics Multivariate normal approximation of the maximum likelihood estimator via the delta method Andreas Anastasiou a and Robert E. Gaunt b a

More information

Bootstrap - theoretical problems

Bootstrap - theoretical problems Date: January 23th 2006 Bootstrap - theoretical problems This is a new version of the problems. There is an added subproblem in problem 4, problem 6 is completely rewritten, the assumptions in problem

More information

On the Estimation and Application of Max-Stable Processes

On the Estimation and Application of Max-Stable Processes On the Estimation and Application of Max-Stable Processes Zhengjun Zhang Department of Statistics University of Wisconsin Madison, WI 53706, USA Co-author: Richard Smith EVA 2009, Fort Collins, CO Z. Zhang

More information

Course 4 Solutions November 2001 Exams

Course 4 Solutions November 2001 Exams Course 4 Solutions November 001 Exams November, 001 Society of Actuaries Question #1 From the Yule-Walker equations: ρ φ + ρφ 1 1 1. 1 1+ ρ ρφ φ Substituting the given quantities yields: 0.53 φ + 0.53φ

More information

Applying the Benjamini Hochberg procedure to a set of generalized p-values

Applying the Benjamini Hochberg procedure to a set of generalized p-values U.U.D.M. Report 20:22 Applying the Benjamini Hochberg procedure to a set of generalized p-values Fredrik Jonsson Department of Mathematics Uppsala University Applying the Benjamini Hochberg procedure

More information

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1, Economics 520 Lecture Note 9: Hypothesis Testing via the Neyman-Pearson Lemma CB 8., 8.3.-8.3.3 Uniformly Most Powerful Tests and the Neyman-Pearson Lemma Let s return to the hypothesis testing problem

More information

Doubly Indexed Infinite Series

Doubly Indexed Infinite Series The Islamic University of Gaza Deanery of Higher studies Faculty of Science Department of Mathematics Doubly Indexed Infinite Series Presented By Ahed Khaleel Abu ALees Supervisor Professor Eissa D. Habil

More information

STAT 512 sp 2018 Summary Sheet

STAT 512 sp 2018 Summary Sheet STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}

More information

ECONOMETRICS II, FALL Testing for Unit Roots.

ECONOMETRICS II, FALL Testing for Unit Roots. ECONOMETRICS II, FALL 216 Testing for Unit Roots. In the statistical literature it has long been known that unit root processes behave differently from stable processes. For example in the scalar AR(1)

More information

A Conditional Approach to Modeling Multivariate Extremes

A Conditional Approach to Modeling Multivariate Extremes A Approach to ing Multivariate Extremes By Heffernan & Tawn Department of Statistics Purdue University s April 30, 2014 Outline s s Multivariate Extremes s A central aim of multivariate extremes is trying

More information

Minimum and maximum values *

Minimum and maximum values * OpenStax-CNX module: m17417 1 Minimum and maximum values * Sunil Kumar Singh This work is produced by OpenStax-CNX and licensed under the Creative Commons Attribution License 2.0 In general context, a

More information

Minimax lower bounds I

Minimax lower bounds I Minimax lower bounds I Kyoung Hee Kim Sungshin University 1 Preliminaries 2 General strategy 3 Le Cam, 1973 4 Assouad, 1983 5 Appendix Setting Family of probability measures {P θ : θ Θ} on a sigma field

More information

STAT 830 Hypothesis Testing

STAT 830 Hypothesis Testing STAT 830 Hypothesis Testing Richard Lockhart Simon Fraser University STAT 830 Fall 2018 Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 1 / 30 Purposes of These

More information

MAS223 Statistical Inference and Modelling Exercises

MAS223 Statistical Inference and Modelling Exercises MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,

More information

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics

Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics Wooldridge, Introductory Econometrics, 4th ed. Appendix C: Fundamentals of mathematical statistics A short review of the principles of mathematical statistics (or, what you should have learned in EC 151).

More information

STAT 461/561- Assignments, Year 2015

STAT 461/561- Assignments, Year 2015 STAT 461/561- Assignments, Year 2015 This is the second set of assignment problems. When you hand in any problem, include the problem itself and its number. pdf are welcome. If so, use large fonts and

More information