Optional Stopping Theorem Let X be a martingale and T be a stopping time such

Similar documents
Plan Martingales cont d. 0. Questions for Exam 2. More Examples 3. Overview of Results. Reading: study Next Time: first exam

Point Process Control

Poisson Processes. Stochastic Processes. Feb UC3M

Manual for SOA Exam MLC.

Lectures 16-17: Poisson Approximation. Using Lemma (2.4.3) with θ = 1 and then Lemma (2.4.4), which is valid when max m p n,m 1/2, we have

Poisson Processes. Particles arriving over time at a particle detector. Several ways to describe most common model.

Exponential Distribution and Poisson Process

Chapter 2. Poisson Processes. Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan

STAT 380 Markov Chains

The exponential distribution and the Poisson process

Continuous-time Markov Chains

Modern Discrete Probability Branching processes

EDRP lecture 7. Poisson process. Pawe J. Szab owski

1 Delayed Renewal Processes: Exploiting Laplace Transforms

errors every 1 hour unless he falls asleep, in which case he just reports the total errors

Filtrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition

Figure 10.1: Recording when the event E occurs

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

Chapter 2. Poisson Processes

Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided

Chapter 1. Poisson processes. 1.1 Definitions

A Large-Sample Approach to Controlling the False Discovery Rate

Exercise Exercise Homework #6 Solutions Thursday 6 April 2006

Data analysis and stochastic modeling

Solution: The process is a compound Poisson Process with E[N (t)] = λt/p by Wald's equation.

1 Poisson processes, and Compound (batch) Poisson processes

1 Some basic renewal theory: The Renewal Reward Theorem

Lecture 4a: Continuous-Time Markov Chain Models

Math 6810 (Probability) Fall Lecture notes

H 2 : otherwise. that is simply the proportion of the sample points below level x. For any fixed point x the law of large numbers gives that

False Discovery Rate

Stochastic Modelling Unit 1: Markov chain models

µ X (A) = P ( X 1 (A) )

Some General Types of Tests

Central Limit Theorem ( 5.3)

3. Poisson Processes (12/09/12, see Adult and Baby Ross)

Alpha-Investing. Sequential Control of Expected False Discoveries

(b) What is the variance of the time until the second customer arrives, starting empty, assuming that we measure time in minutes?

Renewal theory and its applications

Stochastic integration. P.J.C. Spreij

Optimization and Simulation

Modelling the risk process

The story of the film so far... Mathematics for Informatics 4a. Continuous-time Markov processes. Counting processes

Looking at the Other Side of Bonferroni

CDA5530: Performance Models of Computers and Networks. Chapter 3: Review of Practical

1 Stationary point processes

Jump Processes. Richard F. Bass

Selected Exercises on Expectations and Some Probability Inequalities

Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama

Theory of Computation

BEST TESTS. Abstract. We will discuss the Neymann-Pearson theorem and certain best test where the power function is optimized.

STAT 461/561- Assignments, Year 2015

Since D has an exponential distribution, E[D] = 0.09 years. Since {A(t) : t 0} is a Poisson process with rate λ = 10, 000, A(0.

Probability Midterm Exam 2:15-3:30 pm Thursday, 21 October 1999

STAT Sample Problem: General Asymptotic Results

ISyE 6644 Fall 2014 Test #2 Solutions (revised 11/7/16)

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539

µ n 1 (v )z n P (v, )

Q = (c) Assuming that Ricoh has been working continuously for 7 days, what is the probability that it will remain working at least 8 more days?

The goodness-of-fit test Having discussed how to make comparisons between two proportions, we now consider comparisons of multiple proportions.

1. Stochastic Processes and filtrations

Lecture 7 April 16, 2018

CS 237: Probability in Computing

Dynkin (λ-) and π-systems; monotone classes of sets, and of functions with some examples of application (mainly of a probabilistic flavor)

MATH 56A: STOCHASTIC PROCESSES CHAPTER 6

Spring 2016 Network Science. Solution of Quiz I

Markov chains. 1 Discrete time Markov chains. c A. J. Ganesh, University of Bristol, 2015

Solutions to Homework Discrete Stochastic Processes MIT, Spring 2011

Specific Differences. Lukas Meier, Seminar für Statistik

UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017)

SOLUTIONS IEOR 3106: Second Midterm Exam, Chapters 5-6, November 8, 2012

Chapter 5. Chapter 5 sections

Massachusetts Institute of Technology

Tyler Hofmeister. University of Calgary Mathematical and Computational Finance Laboratory

Markov processes and queueing networks

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Arrivals and waiting times

Summer School in Statistics for Astronomers & Physicists June 5-10, 2005

1 Generating functions

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

Multiple Testing. Hoang Tran. Department of Statistics, Florida State University

S1. Proofs of the Key Properties of CIATA-Ph

STAT331 Lebesgue-Stieltjes Integrals, Martingales, Counting Processes

Logic, Sets, and Proofs

Linear maps. Matthew Macauley. Department of Mathematical Sciences Clemson University Math 8530, Spring 2017

Massachusetts Institute of Technology

Statistical Inference. Hypothesis Testing

E[X n ]= dn dt n M X(t). ). What is the mgf? Solution. Found this the other day in the Kernel matching exercise: 1 M X (t) =

Journal Club: Higher Criticism

IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Thursday, October 4 Renewal Theory: Renewal Reward Processes

Summary: the confidence interval for the mean (σ 2 known) with gaussian assumption

n E(X t T n = lim X s Tn = X s

Asymptotics of random sums of heavy-tailed negatively dependent random variables with applications

An essay on the general theory of stochastic processes

6. Bernoulli Trials and the Poisson Process

Doing Cosmology with Balls and Envelopes

Problem Sheet 1. You may assume that both F and F are σ-fields. (a) Show that F F is not a σ-field. (b) Let X : Ω R be defined by 1 if n = 1

ELEMENTS OF PROBABILITY THEORY

Distribution-Free Procedures (Devore Chapter Fifteen)

Transcription:

Plan Counting, Renewal, and Point Processes 0. Finish FDR Example 1. The Basic Renewal Process 2. The Poisson Process Revisited 3. Variants and Extensions 4. Point Processes Reading: G&S: 7.1 7.3, 7.10 Next Time: Poisson and Point Processes and Applications Homework 7 due Thursday Left-over Question 1. Does E Z < implie that E Z 1{ Z > c} 0 as c? Theorem 2. Optional Sampling Let X be a submartingale with respect to (F n ). 1. If T is a stopping time for which P{T m} = 1 for some m <, then EX + T < and X 0 E(X T F 0 ). (Hence, EX 0 EX T.) 2. If T 0 T 1 are stopping times with P{T i m i } = 1 for m i <, then Y n = X Tn is a submartingale as well. Corollary 3. that Optional Stopping Theorem Let X be a martingale and T be a stopping time such 1. P{T < } = 1. 2. E X T < 3. EX n 1{T > n} 0 as n. Then, EX T = EX 0. Corollary 4. Suppose that X = (X n ) n 0 is a uniformly integrable martingale and T is a stopping time with P{T < } = 1, then EX T = EX 0. Corollary 5. If X is a martingale and T is a stopping time with ET < and assume that X T n is a uniformly integrable sequence. Then EX 0 = EX T. 1 11 Apr 2006

Application 6. False Discovery Rates Suppose we have m hypothesis tests and corresponding p-values P 1,..., P m. Let P (0) 0 and P (1) < P (m) are the p-values in increasing order. Then, Benjamini and Hocbherg (1995) used the following threshold to determine which null hypotheses would be rejected. { P = max P (i) : P (i) α i } (1) m They showed that if we reject all null hypothesis for which P i P, we will control the false discovery rate at level α. That is, falsely rejected nulls E α. (2) rejected nulls Let s make this more precise. Let H 0 and H 1 be a partition of {1,..., m}. H 0 represents the indices of true null hypotheses; H 1 represents the indices of true alternatives (though which alternative is not specified by the index alone). Let m 0 = #(H 0 ). For any p-value threshold, 0 t 1, define the following: V (t) = 1{P i t} (3) i H 0 R(t) = 1{P i t}. (4) i H 0 H 1 V (t) is the number of false rejects at threshold t; R(t) is the total number of rejections. Define the False Discovery Proportion at threshold t by FDP(t) = V (t) max(r(t), 1). (5) Then, for a threshold T, fixed or random, the False Discovery Rate (FDR) is given by FDR = E FDP(T ). (6) Benjamini and Hochberg (1995) showed that a threshold equivalent to { } mt T α = sup t [0, 1]: R(t) α (7) satisfies FDR α. Let s give a martingale proof in three steps. But in this case we need to do two things: take the threshold as our time index, which is thus a continuous index, and reverse time. Both are technical variations on the discrete-time martingales we ve seen, but the basic ideas are the same. Assume that the p-values for which the null is true are independent. Claim 1. The process t V (t)/t is a reversed-time martingale with respect to the filtration F t = σ (1{P i s}, s [t, 1], i {1,..., m}). (8) Claim 2. T α is a stopping time (with time reversed) with respect to (F t ). Claim 3. The optional stopping theorem implies that E FDP(T α ) = (m 0 /m)α. 2 11 Apr 2006

Proof of Claim 1. What do we need to show? What do we know (or need to assume)? What are the first steps? Proof of Claim 2. What do we need to show? (It s not exactly the same as in the discrete case but what s the natural analogue?) What do we know (or need to assume)? How does mt/r(t) change with t? 3 11 Apr 2006

Proof of Claim 3. Thinking about the process mt/r(t), we see that Hence, Note that the stopped process is bounded by m/α. Hmmmm... The optional stopping theorem gives R(T α ) = m α T α. (9) FDP(T α ) = α V (T α ). (10) m T α and thus FDR Tα = α m EV (T α) T α = m 0 m α. E V (T α) = E V (1) = m 0, (11) T α 1 Sweet. Definition 7. A counting process N = (N t ) t 0 satisfies 1. N t is non-negative integer-valued. 2. N t is non-decreasing. The increment N t N s for s t represents the number of events that have occurred in the interval (s, t]. Figure 8. 5 4 3 2 1 N t X 1 = S 1 S 2 S 3 S 4 S 5 }{{}}{{}}{{}}{{} X 2 X 3 X 4 X 5 The Basic Renewal Process 9. Let X 1, X 2,... be a sequence of iid, non-negative random variables with common cdf F. Assume, to avoid odd boundary cases, that F (0) = 0. Let µ = EX 1. We will interpret X i as the time between the i 1st and ith event or renewal. Write S 0 = 0 and S n = n i=1 X i. Then, S n is the time of the nth event. Finally, define N t = sup{n Z : S n t}. (12) This counting process is called a renewal process. t 4 11 Apr 2006

Question 10. Can an infinite number of events occur in a finite time interval? Why or why not? A Representation 11. Consider the following claim: N t = 1{S k t}. (13) k=1 Is this true? Why? Definition 12. we have Let m(t) = EN t. This is called the renewal function. By the above representation, m(t) = where F k is the cdf of S k. P{S k t} = F k (t), (14) k=1 Question 13. What is the distribution of S k in terms of the distribution of S k 1? (Start with k = 1.) k=1 Derivation 14. The Distribution of N t. We have the following logical relation: From this, we have N t n S n t. (15) P{N t = n} = P{N t n} P{N t n + 1} (16) = P{S n t} P{S n+1 t} (17) = F n (t) F n+1 (t). (18) 5 11 Apr 2006

Notation 15. For our purposes here, we will use for the convolution operator given by (G H)(t) = t 0 G(t s) dh(s), (19) whenever the integral exists, for real-valued functions G, H on [0, ). You can show that this a commutative and associative operator. Equation 16. A function u is a solution to a renewal-type equation if it satisfies for bounded G on [0, ). In particular, the renewal function satisfies u = G + u F, (20) m = F + m F. (21) There are a couple ways to see this. First, we can plug in k F k into the equation. Check. Second, we can condition on the first arrival. The function E(N t X 1 near s) takes the following form: Why? It follows that E(N t X 1 near s) = { 1 + m(t s) if s t 0 if s > t. (22) m(t) = EN t (23) = EE(N t X 1 ) (24) = = 0 t 0 E(N t X 1 near s) ds (25) (1 + m(t s)) df (s) (26) = F (t) + (m F )(t). (27) More generally, a solution to the renewal-type equation (20) is given by u = G + G m. (28) To see this, convolve both sides of the equation with F : hence, u F = G F + G m F (29) = G F + G (m F ) (30) = G m (31) = u G, (32) u = G + u F. (33) We can also show that if G is bounded on compact sets, then u is a unique solution and is also bounded on compact sets. 6 11 Apr 2006

Definition 17. A real-valued random variable is said to be lattice if it takes values in a set {ka: k Z}. The span (or period) of the lattice is the maximal such a. Limit Theorems 18. Renewal processes behave asymptotically as you might expect. (ha ha) As t, the following hold N t 1. 1 almost surely. t µ 2. m(t) 1 t µ. m(t + h) m(t) 3. 1 h µ, for all h if X 1 is not lattice and for all h that are integer multiples of the span if X 1 is lattice. In addition, N t /t, properly centered and scaled, is asymptotically Normal. See the book (section 10.2) for proofs. Process 19. The Poisson Process is a renewal process with interarrival distribution F equal to an Exponential λ. More precisely, this is called a (homogeneous) Poisson process with rate λ. Recall that from this setup, we showed that the Poisson process has two other important properties. First, independent increments. The random variables N t N s and N v N u are independent whenever (s, t] and (u, v] are disjoint. (The same goes for multiple such intervals.) Second, stationary increments, the distribution of N t N s depends only on t s. We can thus characterize this process in two other equivalent ways. Alternate Definition 1 20. Let N = (N t ) t 0 be a process with the following properties: 1. N 0 = 0 2. For 0 s < t, N t N s has a Poisson λ(t s ) distribution. 3. For 0 s 1 < t 1 < s 2 < t 2 < < s m < t m, the random variables N ti N si are independent. Then, N is a Poisson process with rate λ. Alternate Definition 2 21. Let N = (N t ) t 0 be a process with the following properties: 1. N 0 = 0 2. The process has stationary and independent increments. 3. P{N h = 1} = λh + o(h). 4. P{N h 2} = o(h). Then, N is a Poisson process with rate λ. Theorem 22. The two alternate definitions are equivalent. Basic idea of the proof: Poisson approximation to the binomial. Note that from these alternative definitions, we can derive the interarrival distribution by P{X 1 > t} = P{N t = 0} = e λt. (34) Useful Fact 23.... which you derived in homework. Given N t = n, the set of n arrival times {S 1,..., S n } have the same distribution as the order statistics of n independent Uniform 0, t random variables. 7 11 Apr 2006

Process 24. By generalizing the alternate definitions, we get a whole new class of interesting processes. Let N = (N t ) t 0 be a process with the following properties: 1. N 0 = 0 2. The process has independent increments. 3. P{N t+h N h 2} = o(h). 4. P{N t+h N h = 1} = λ(t)h + o(h). Then, N is called an inhomogeneous Poisson process with rate function λ(t). We can show that N t+s N t has a Poisson distribution with mean m(t + s) m(t), where m(t) = t 0 λ(s) ds. This leads us to the idea of Point Processes which we ll take up next time. 8 11 Apr 2006