Sequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process

Similar documents
Hypothesis testing (cont d)

ECE531 Screencast 9.2: N-P Detection with an Infinite Number of Possible Observations

Lecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics. 1 Executive summary

ECE531 Lecture 6: Detection of Discrete-Time Signals with Random Parameters

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

Sequential Detection. Changes: an overview. George V. Moustakides

Testing Simple Hypotheses R.L. Wolpert Institute of Statistics and Decision Sciences Duke University, Box Durham, NC 27708, USA

Intelligent Systems Statistical Machine Learning

Lecture 8: Information Theory and Statistics

2. What are the tradeoffs among different measures of error (e.g. probability of false alarm, probability of miss, etc.)?

Detection theory. H 0 : x[n] = w[n]

Introduction to Bayesian Statistics

Intelligent Systems Statistical Machine Learning

Lecture 7 Introduction to Statistical Decision Theory

LM threshold unit root tests

Institute of Actuaries of India

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Partitioning the Parameter Space. Topic 18 Composite Hypotheses

Optimal Sequential Procedures with Bayes Decision Rules

Constructing Ensembles of Pseudo-Experiments

Module 2. Random Processes. Version 2, ECE IIT, Kharagpur

Testing Restrictions and Comparing Models

Early Detection of a Change in Poisson Rate After Accounting For Population Size Effects

2 Mathematical Model, Sequential Probability Ratio Test, Distortions

Optimal Sequential Tests for Two Simple Hypotheses

Statistical Methods for Particle Physics (I)

Chapter 2. Binary and M-ary Hypothesis Testing 2.1 Introduction (Levy 2.1)

ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing

More Empirical Process Theory

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80

Detection and Estimation Chapter 1. Hypothesis Testing

If there exists a threshold k 0 such that. then we can take k = k 0 γ =0 and achieve a test of size α. c 2004 by Mark R. Bell,

ECE531 Lecture 13: Sequential Detection of Discrete-Time Signals

Lecture 8: Information Theory and Statistics

Detection and Estimation Theory

Physics 403. Segev BenZvi. Classical Hypothesis Testing: The Likelihood Ratio Test. Department of Physics and Astronomy University of Rochester

DETECTION theory deals primarily with techniques for

Testing Statistical Hypotheses

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.

On detection of unit roots generalizing the classic Dickey-Fuller approach

Lecture Testing Hypotheses: The Neyman-Pearson Paradigm

Two-stage Adaptive Randomization for Delayed Response in Clinical Trials

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided

ECE531 Lecture 4b: Composite Hypothesis Testing

Hypothesis Testing. Part I. James J. Heckman University of Chicago. Econ 312 This draft, April 20, 2006

Detection theory 101 ELEC-E5410 Signal Processing for Communications

OPTIMAL SEQUENTIAL PROCEDURES WITH BAYES DECISION RULES

Statistical Inference

UNIFORMLY MOST POWERFUL CYCLIC PERMUTATION INVARIANT DETECTION FOR DISCRETE-TIME SIGNALS

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

g(.) 1/ N 1/ N Decision Decision Device u u u u CP

Mathematical Statistics

Decentralized Sequential Hypothesis Testing. Change Detection

STAT 461/561- Assignments, Year 2015

Parametric Techniques Lecture 3

Ch. 5 Hypothesis Testing

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing

STAT 830 Hypothesis Testing

4 Hypothesis testing. 4.1 Types of hypothesis and types of error 4 HYPOTHESIS TESTING 49

INFORMATION PROCESSING ABILITY OF BINARY DETECTORS AND BLOCK DECODERS. Michael A. Lexa and Don H. Johnson

Hypothesis Testing - Frequentist

SYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions

Chapter 2 Signal Processing at Receivers: Detection Theory

Parametric Techniques

Analysis of the AIC Statistic for Optimal Detection of Small Changes in Dynamic Systems

Topic 10: Hypothesis Testing

Empirical Market Microstructure Analysis (EMMA)

Some General Types of Tests

Uncertainty. Jayakrishnan Unnikrishnan. CSL June PhD Defense ECE Department

Approximation of Average Run Length of Moving Sum Algorithms Using Multivariate Probabilities

Solving Classification Problems By Knowledge Sets

Statistical Methods for Particle Physics Lecture 4: discovery, exclusion limits

Quantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing

Optimal SPRT and CUSUM Procedures using Compressed Limit Gauges

Research Article Degenerate-Generalized Likelihood Ratio Test for One-Sided Composite Hypotheses

SEQUENTIAL TESTS FOR COMPOSITE HYPOTHESES

Practice Problems Section Problems

Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution.

Topic 17: Simple Hypotheses


Spring 2012 Math 541B Exam 1

Lecture 8 Inequality Testing and Moment Inequality Models

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

PATTERN RECOGNITION AND MACHINE LEARNING

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

Least Favorable Distributions for Robust Quickest Change Detection

A CUSUM approach for online change-point detection on curve sequences

Statistical Methods for Particle Physics Lecture 1: parameter estimation, statistical tests

Parameter Estimation, Sampling Distributions & Hypothesis Testing

Topic 3: Hypothesis Testing

14.30 Introduction to Statistical Methods in Economics Spring 2009

Simultaneous and sequential detection of multiple interacting change points

1/24/2008. Review of Statistical Inference. C.1 A Sample of Data. C.2 An Econometric Model. C.4 Estimating the Population Variance and Other Moments

Recall the Basics of Hypothesis Testing

Statistics for the LHC Lecture 1: Introduction

Performance Evaluation and Comparison

Statistical Inference

Physics 403. Segev BenZvi. Credible Intervals, Confidence Intervals, and Limits. Department of Physics and Astronomy University of Rochester

Statistical inference

Testing Statistical Hypotheses

Transcription:

Applied Mathematical Sciences, Vol. 4, 2010, no. 62, 3083-3093 Sequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process Julia Bondarenko Helmut-Schmidt University Hamburg University of the Federal Armed Forces Hamburg Holstenhofweg 85, 22043 Hamburg, Germany bonda@hsu-hh.de Abstract Sequential Probability Ratio Test SPRT has been widely used to detect process anomalies. The purpose of this paper is to present a practical procedure on the basis of SPRT, which recognizes if the mean of a latent Gaussian process Y t significantly deviates from presumed value, via an observable signal resulting from Y t. An equation for estimating the unknown nuisance parameter of the process is obtained, values of likelihood ratio statistic and thresholds are determined. The empirical analysis of the procedure performance is conducted. Keywords: Sequential Probability Ratio Test; Maximum Likelihood Estimate; Nuisance Parameter; Average Sample Number 1 Introduction A. Wald 11] has developed his famous sequential probability ratio test SPRT - a sequential extension of the Neyman-Pearson Lemma, which implies that likelihood ratio test gives the best result in fixed size samples. Let Y 1, Y 2... be an infinite sequence of the random variables with values in some data space Y, and let µ be a sigma-finite measure on Y. Then Y 1, Y 2,., Y n Y n and µ n = µ n product measure is sigma-finite on Y n. Let fy 1, Y 2,., Y n, n = 1, 2,..., denote the density function of Y 1, Y 2,., Y n with respect to µ n on Y n. In the case of parameterized family of probability density functions, we can also write fy 1, Y 2,., Y n ; θ. This includes the continuous case where µ n is the Lebesque measure and fy 1, Y 2,., Y n ; θ is a Lebesque density, or joint probability density function - PDF, the discrete case where µ n is the counting measure and fy 1, Y 2,., Y n ; θ is a probability function - PF, as well as more complex models. Let us assume first that we want to test two simple hypotheses, H 0 : θ = θ 0 vs H 1 : θ = θ 1, θ 0 θ 1. The corresponding densities are fy 1, Y 2,., Y n ; θ 0 =

3084 J. Bondarenko f 0 Y 1, Y 2,., Y n and fy 1, Y 2,., Y n ; θ 1 = f 1 Y 1, Y 2,., Y n. We consider the first n 1 random variables observations Y 1, Y 2,., Y n 1. When the new observation Y n arrives, the SPRT decides whether to reject the alternative hypothesis, reject the null hypothesis, or continue sampling. The test involves the likelihood ratio Λ n = f 1Y 1, Y 2,., Y n f 0 Y 1, Y 2,., Y n = LY 1, Y 2,..., Y n ; H 1 LY 1, Y 2,..., Y n ; H 0, 1 where n = 1, 2,..., LY 1, Y 2,..., Y n ; H j is the likelihood value under hypothesis j, j = {0; 1}. The decision rule defined by SPRT splits the sample space Y n into three subsets called acceptance, rejection and continuation/indifference regions. In fixed size samples, accordingly to the Neyman-Pearson Lemma, the likelihood ratio statistic is compared to certain threshold value, which determines accuracy. By analogy, we select two thresholds A and B which are in general depending on n, 0 < B < A <, and draw a comparison between them and our likelihood ratio at each consecutive observation. If a threshold is reached or crossed, a decision regarding the null or alternative hypothesis is reached as well, that is: if Λ n B - terminate and reject H 1 assuming that either H 0 or H 1 is true, it can be also interpreted as terminate and accept H 0 - acceptance region; if Λ n A - terminate and reject H 0 rejection region; if B < Λ n < A - continue sampling since our hope for a reduced sample size i is not justified and we need more observations for decision continuation/indifference region. Constants A and B are related to the error rates α probability of eventually rejecting H 0 when it is true, α = P H 1 H 0 ] = P 0 H 1 ] - error of Type I, or the false alarm probability and β probability of eventually accepting H 0 when it is false, β = P H 0 H 1 ] = P 1 H 0 ] - error of Type II, or missed alarm probability. The probabilities α and β reflect the costs of making these two types of errors. Note, that the Neyman-Pearson Lemma provides the test with maximal power 1 β, where the above-mentioned threshold value is chosen by setting the false alarm probability α. Provided that the procedure terminates with probability 1, Wald has also shown, that α and β and the bounds A and B satisfy the following inequalities 11], p. 43; 8]; 2]: B β, A 1 β, see Appendix 2 for the proof. In 1 α α general, these inequalities need not to hold with equalities because by termination Λ n might fall not exactly in B or A. But usually, under large n, the value of Λ n is only slightly smaller than B or larger than A. Therefore, Wald s approximation replaces the inequalities by equalities, for obtaining in practice the thresholds for desired α = α and β = β, see, for example, 8], p. 11: B = β, A = 1 β. Note, that a special case of SPRT is the open-ended 1 α α test, when the test statistic Λ n excludes the lower bound B. Going back to the above-mentioned Neyman-Pearson Lemma, let s remind that to provide the most powerful test for a fixed samples one needs to fix a

Sequential procedure 3085 false alarm probability test size α, that is, the constraints here are a sample size and α. The Wald-Wolfowitz Theorem 12] cited below states, that the SPRT in testing simple H 0 vs simple H 1 hypotheses based on the i.i.d. observations, which operates with infinite sample, can present even better result. Really, additionally to the desired error probabilities α and β, playing a role of constraints in this case, it gives also the smaller average sample number ASN E n] under both H 0 and H 1. We remind briefly, that ASN is the average number of observations required for decision making on step n under assumption that H 0 or H 1 is true. Hence, by implementing the SPRT to fixed sample the sampling procedure may be terminated even before the entire sample is considered. Theorem 1 Wald and Wolfowitz. Among all tests fixed sample or sequential for which P H 1 H 0 ] α and P H 0 H 1 ] β and for which E n H j ] <, j = {0; 1}, H j are simple, the SPRT with error probabilities α and β minimizes both E n H 0 ] and E n H 1 ], under i.i.d.-assumption. For the case of non-i.i.d. observations the optimality of the SPRT in the sense of the Wald-Wolfowitz Theorem, is in general not ensured, although the asymptotic optimality under small error probabilities and some mild constraint stability of the process {log Λ n } is fulfilled. Note, that by the Wald and Wolfowitz Theorem, the optimality of SPRT implies that the expected sample size needed is smallest only when θ = θ 0 or θ = θ 1. But for values θ between θ 0 and θ 1, the SPRT may perform quite unsatisfactory: the expected sample sizes tend to be large and the decision about hypothesis acceptance is delayed, hence, optimality doesn t hold here and SPRT may require a sample size even larger than a non-sequential test. It gave rise to consider the following optimization problem: one tests H 0 : θ = θ 0 vs H 1 : θ = θ 1 such that the expected sample size for θ = θ, θ 0 < θ < θ 1, is minimized subject to error probabilities at two other points, θ 0 and θ 1, that is, P H 1 H 0 ] α and P H 0 H 1 ] β. This problem was solved by Kiefer and Weiss 4], and also extended in later works. In most realistic situations the hypotheses to be tested are composite, and the simple hypotheses are only a particular case of them. The composite hypotheses can be written as H 0 : θ < θ 0 vs H 1 : θ > θ 0 θ > θ 1, θ 0 < θ 1, 2 with Type I and Type II error probabilities not exceeding α and β and indifference zone i.e., zone separating zero and alternative hypotheses θ 0, θ 1. For fixed sample size, the composite hypotheses may be converted to simple ones under monotone likelihood ratio property. Unfortunately, it is not true in

3086 J. Bondarenko the sequential case 1], p. 54. There are numerous approaches to the testing of such kind of hypotheses, but an obvious optimality criterion doesn t exist here. For example, Wald s proposal was to transform the problem to testing between two simple hypotheses K 0 : θ = θ 0 vs K 1 : θ = θ 1 by means the method of weight functions - some sort of prior distribution 11]. The sequential testing solution for composite hypotheses based upon Bayesian strategy is considered also in 6]. A detailed review of methods and approaches for sequential testing of composite hypotheses is given in 5], 7]. In this work, we consider a sequential test of composite hypotheses about parameter of weakly stationary gaussian latent process. The outline of the paper is as follows. In Section 2, we introduce our testing problem for parameter of interest, present a nuisance parameter eliminating procedure and derive the test thresholds. Section 3 shows the empirical results of the problem solution and concludes with a few remarks. 2 Sequential Testing Procedure 2.1 Problem Formulation Let Y 1, Y 2,.. be some underlying unobservable hidden time series realization of Y t supposed to be independent normally distributed random variables, Y t N µ, σ 2, for each t = 1, 2,..., where µ > 0 and σ 2 are unknown. Assume that if Y t > θ, where θ = const, then the signal is observed. Without losing generality, we may take θ = 0. That is, we shall concerned then with the nonnegative random variables X t = max0, Y t, i = 1, 2,.... 3 The purpose is to detect as quickly as possible if the value of parameter of interest µ differs from a known value µ 0, in the presence of nuisance parameter σ 2. We can formulate the following, analogous to 2, hypotheses: H 0 : µ < µ 0 vs H 1 : µ > µ 0, 4 with false and missed alarm probabilities α and β, correspondingly. Similar problem statement can arise in medical, socio-economic and physical fields, for example, with application to financial decision-making in particular, option pricing problem, accidents monitoring, disease control, seismic signal processing, etc. In order to construct a likelihood ratio 1 corresponding to the testing problem 4, we need, in the first instance, to derive the distribution of the random variable X, under all values µ. The distribution of X can be easily obtained as follows. Let s replace the zero value in 3 for every t with a random variable Z, normally distributed with parameters 0 and τ 2, where

Sequential procedure 3087 τ 0, Z N 0, τ 2, and independent of Y. Then the distribution and density functions of the random variable X t = max Z t, Y t are, correspondingly, F X x; µ, σ 2, τ 2 = P Z < x, Y < x = F Z x; 0, τ 2 F Y x; µ, σ 2, 5 f X x; µ, σ 2, τ 2 = F Z x; 0, τ 2 f Y x; µ, σ 2 + F Y x; µ, σ 2 f Z x; 0, τ 2. Notation 2 Another possibility would be to assign a uniform distribution to the variable Z, i.e., Z U τ, τ, where τ 0. 2.2 Nuisance Parameter Estimation Elimination of the nuisance parameter σ 2 still remains an open question. The algorithm that produces approximate maximum likelihood estimator for parameter σ 2 is similar to the Stein two-stage procedure 9]. A training set of data first M observations X 1,..., X M has been used to construct the estimate σ 2. Notation 3 Alternatively, one can consider purely sequential procedure for the problem solution, based on the Adaptive Sequential Probability Ratio Test ASPRT see 10]. ASPRT employs estimates of the unknown nuisance parameters parameter σ 2 in our case, which involve only the data up to the previous time point. First, we specify the maximum loglikelihood equations: { l µ X 1,..., X M µ, σ 2, τ 2 = 0, l σ X 1,..., X M µ, σ 2, τ 2 = 0, where l X 1,..., X M µ, σ 2, τ 2 = ln L X 1,..., X M µ, σ 2, τ 2 = M ln F 1 X i f 2 X i + F 2 X i f 1 X i, F 1 x = F Z x; 0, τ 2, f 1 x = f Z x; 0, τ 2, F 2 x = F Y x; µ, σ 2, f 2 x = f Y x; µ, σ 2. Let η be a parameter, η = {µ, σ}, then we denote S = l η X 1,..., X M µ, σ 2, τ 2 = M F 1 X i f 2 η X i+ F 2 η X i f 1 X i F 1 X i f 2 X i +F 2 X i f 1 X i. Passing to the limit under τ 0, one can distinguish two cases: 1 X i 3τ and 2 X i > 3τ. The sum S is partitioned into two smaller sums: the first one, S 1, involves only K summands with X i 3τ, the second one, S 2, M K summands with X i > 3τ, S = S 1 +S 2. In first case we can see that f 1 X i and F 1 X i 1, and therefore 2 6

3088 J. Bondarenko S 1 = K K F 2 η X i F 2 X i F 1 X i f 2 η X i+ F 2 η X i f 1 X i F 1 X i f 2 X i +F 2 X i f 1 X i = K F 2 η 0 F 2 0. = K f 1 X i f 1 X i F 1 X i f 1 X i f 2 η X i+ F 2 ] η X i F 1 X i f 1 X i f 2X i +F 2 X i In the second case, f 1 X i 0 and F 1 X i 1, and consequently, S 2 = M K M K f 2 η X i f 2 X i. F 1 X i f 2 η X i+ F 2 η X i f 1 X i F 1 X i f 2 X i +F 2 X i f 1 X i = M K In respect that f 2 X µ i = X i µ f σ 2 X i, f 2 X σ i = 1σ + X i µ 2 f σ 3 2 X i, F 2 0 = 1 0 µ 2πσ 3 s µ exp s µ2 ds, 2σ 2 F 2 σ 0 = 1 σ F 2 0 + 1 2πσ 4 0 s µ 2 exp s µ2 2σ 2 ] ] f F 1 X i 2 η X i+ F 2 η X i f 1X i F 1 X i ] F 1 X i f 2 X i +F 2 X i f 1X i F 1 X i ds, the system of equations 6 is transformed in the following one: K exp µ2 2σ 2 2π 1 Φ σ = 1 µ σ K µ exp µ2 2σ 2 2π σ 1 Φ σ = M K 1 µ where Φ x = 1 2π exp X i >0 X i M K µ σ, x2 2 7 is a standard normal distribution function. Multiplying the first equation in 7 by µ σ we obtain that µ = σ 2 X i >0 X2 i + 2 σ 2 µ X i >0 X i M K σ 2 µ 2, Xi 2 M K σ2 X i >0 X i X i >0 and resolve it with respect to σ. and subtracting it from the second one,. We substitute then µ into first equation Notation 4 Method of moments is another way to the solution of parameter estimation problem. 2.3 Sequential Likelihood Ratio Test Statistic and Boundaries For practical purposes, we can replace the composite hypotheses by simple ones see 11], p. 71. Let s choose two constants ɛ 1 and ɛ 2, ɛ 1 > 0, ɛ 2 > 0, and consider two simple hypotheses

Sequential procedure 3089 H 0 : µ = µ 0 ɛ 1 vs H 1 : µ = µ 0 + ɛ 2. 8 If ɛ 1 and ɛ 2 are very small, the test of hypotheses 8, the error probabilities will be approximately equal to error probabilities of original problem 4, α and β.the new problem formulation provides us with indifference zone, unlike the original problem. This means, that for the values of parameter µ inside the interval µ 0 ɛ 1, µ 0 + ɛ 2, i.e. sufficiently close to µ 0, the decision is difficult to make. The existence of indifference zone introduces now an undesirable imperspicuity for the parameter values inside it, but this drawback can be adjusted partially. On the basis of 8, the following statistic can be employed for sequential testing of hypotheses H 0 and H 1 : l X M+1,..., X n µ 0 + ɛ 2, σ 2, τ 2 Λ n = l X M+1,..., X n µ 0 ɛ 1, σ = 9 2, τ 2 n ln F 1 X t f 2 X t ; µ 0 + ɛ 2 + F 2 X t ; µ 0 + ɛ 2 f 1 X t F 1 X t f 2 X t ; µ 0 ɛ 1 + F 2 X t ; µ 0 ɛ 1 f 1 X t, t=m+1 where l is a log-likelihood function, F 1 x F Z x; 0, τ 2, f 1 x f Z x; 0, τ 2, F 2 x; s F Y x; s, σ 2, f 2 x; s f Y x; s, σ 2 and σ 2 is an estimate of unknown variance. We can also write 9 in the following form Λ n = Λ n 1 + C n, 10 where C n = ln F 1X t f 2 X t;µ 0 +ɛ 2 +F 2 X t;µ 0 +ɛ 2 f 1 X t F 1 X t f 2 X t;µ 0 ɛ 1 +F 2 X t;µ 0 ɛ 1 f 1 X t. The test involves comparing the value Λ n to thresholds b = log β 1 α and a = log 1 β α, where α and β are the error probabilities of Type I and Type II, respectively. If Λ n b, we stop the test and accept H 0. If Λ n a, we stop and declare H 1. Otherwise, no decision is made and we keep on testing. As before, we distinguish two cases: 1 X n 3τ and 2 X n > C. In second case we have that f 1 X n 0 and F 1 X n 1, and therefore, from 10, Λ n Λ n 1 + ln f 2 X n ; µ 0 + ɛ 2 f 2 X n ; µ 0 ɛ 1, 11 that corresponds to the normal case. More detailed, C n = 1 σ2 ɛ 2 + ɛ 1 X n µ 0 1 2 σ ɛ2 2 2 ɛ 2 1. The first case gives us f 1 X n = 0 and F 1 X n = 0 1 under τ 0. 2 One obtains here C n = ln f 10 f 1 0 F 1 0 f 2 0;µ 0 +ɛ 2 f 1 0 +F 2 0;µ 0 +ɛ 2 F 1 X n f 2 0;µ 0 ɛ 1 f 1 0 +F 2 0;µ 0 ɛ 1 ] ] ln F 20;µ 0 +ɛ 2, thus, F 2 0;µ 0 ɛ 1

3090 J. Bondarenko Λ n Λ n 1 + ln F 2 0; µ 0 + ɛ 2 F 2 0; µ 0 ɛ 1. 12 Finally, we can write our likelihood ratio statistic in a slightly different form: { Λ n = Λ n 1 + 1 σ ɛ 2 1 + ɛ 2 X n, if X n > 3τ, 13 0, if X n 3τ. with upper and lower bounds { U n = U n 1 + 1 2 σ 2 ɛ2 2 ɛ 2 1 + µ 0 σ 2 ɛ 2 + ɛ 1 and L n = L n 1 + 1 2 σ 2 ɛ2 2 ɛ 2 1 + µ 0 σ 2 ɛ 2 + ɛ 1, if X n > 3τ, { U n = U n 1 ln F 20;µ 0 +ɛ 2 F 2 0;µ 0 ɛ 1 and L n = L n 1 ln F 20;µ 0 +ɛ 2 F 2 0;µ 0 ɛ 1, if X n 3τ, where Λ M = 0, U M = a, L M = b. 3 Empirical Results and Future Research Direction 14 15 Here we illustrate a SPRT-procedure performance comparison for different values of the mean parameter µ, standard deviation parameter σ and indifference zone width. The boundaries defined by ɛ 1 and ɛ 2 in 8 are equal to µ 0 ɛ and µ 0 + ɛ, where ɛ = ɛ 1 = ɛ 2, µ 0 = 0.5. Error probabilities have been set to α = β = 0.05. We assume also that a set of training data of size M = 150 for parameter σ 2 estimation is available. Each set of parameters µ, σ, ɛ implies simulation with 2000 replications. The reported values of power and ASN are presented in Figures 1-4. In general, for fixed values α and β, an increase in ɛ-value entails better results. The relatively small value of ɛ produces high detection probability and short ASN only under small variance σ. For large variance values, a certain hesitation between indifference and weak preference is illustrative. Note, that as the difference between µ and µ 0 increases, performance of the procedure becomes not sensitive to the variance values: power converges to 1 and ASN go quickly down. Notation 5 Average Sample Number can be also determined analytically as following see, for instance, 3]: En H 1 = EΛ n H 1 EC t H 1, 16 where C t is the loglikelihood ratio component corresponding to the tth observation. From the previous section,

Sequential procedure 3091 Figure 1: Empirical Power and Average Sample Length, µ = 0.55, µ0 = 0.5, α = β = 0.05 Figure 2: Empirical Power and Average Sample Length, µ = 0.6, µ0 = 0.5, α = β = 0.05 Figure 3: Empirical Power and Average Sample Length, µ = 1.0, µ0 = 0.5, α = β = 0.05

3092 J. Bondarenko Figure 4: Empirical Power and Average Sample Length, µ = 2.0, µ 0 = 0.5, α = β = 0.05 { C t = 1 σ ɛ 2 2 + ɛ 1 X t µ 0 1 2 σ ɛ2 2 2 ɛ 2 1, if X t > 3τ, ln F 20;µ 0 +ɛ 2, if X F 2 0;µ 0 ɛ 1 t 3τ. Computing an expectation of the random value C t under alternative hypothesis H 1 and keeping in mind that the numerator in 16 may be approximated as EΛ n H 1 = 1 β ln 1 β + β ln β, α 1 α we can obtain the theoretical results and compare them with empirical evidence. The ASN under null hypothesis H 0 are evaluated analogously. Theoretical analysis of ASN proposed above is a subject tackled in future research. Moreover, the results shown here are also preliminary steps towards a complexification of the considered sequential problem: particularly, we are interested in solving change detection problems from option pricing when the underlying unobservable process Y t is a geometric Brownian motion process. References 1] B. Eisenberg, B. K. Ghosh, The sequential probability ratio test, In: Handbook of Sequential Analysis, Eds. B. K. Ghosh and P. K. Sen, Marcel Dekker Inc., New York, 1991, 47-65. 2] N. L. Johnson, Sequential analysis: a survey, Journal of the Royal Statistical Society Series A, 124 1991, 372-411. 3] D. Johnson, Sequential Hypothesis Testing, http://cnx.org/content/m11242/latest/, 2003. 4] J. C. Kiefer, L. Weiss, Some properties of generalized sequential probability ratio tests, The Annals of Mathematical Statistics, 28 1957, 57-74. 5] T. L. Lai, On optimal stopping in sequential hypothesis testing, Statistica Sinica, 7 1997, 33-51.

Sequential procedure 3093 6] T. L. Lai, Nearly optimal sequential tests of composite hypotheses, Annals of Statistics, 16 1998, 856-886. 7] T. L. Lai, Sequential anlysis: some classical problems and new challeges, Statistica Sinica, 11 2001, 303-408. 8] D. Siegmund, Sequential Analysis: Test and Confidence Intervals, Springer-Verlag, New York, 1985. 9] C. Stein, A two-sample test for a linear hypothesis whose power is independent of the variance, The Annals of Mathematical Statistics, 16 1945, 243-258. 10] A. Tartakovsky, An efficient adaptive sequential procedure for detecting targets, IEEE Aerospace Conference, Big Sky, Montana, 9-16 March 2002. 11] A. Wald, Sequential Analysis, John Wiley& Sons, New York-London- Sydney, 1947. 12] A. Wald, J. Wolfowitz, Optimum character of the sequential probability ratio test, The Annals of Mathematical Statistics, 19 1948, 326-339. Received: April, 2010