Runs of Identical Outcomes in a Sequence of Bernoulli Trials

Size: px
Start display at page:

Download "Runs of Identical Outcomes in a Sequence of Bernoulli Trials"

Transcription

1 Western Kentucky University TopSCHOLAR Masters Theses & Specialist Projects Graduate School Spring 2018 Runs of Identical Outcomes in a Sequence of Bernoulli Trials Matthew Riggle Western Kentucky University, matthew.riggle262@topper.wku.edu Follow this and additional works at: Part of the Applied Mathematics Commons, and the Discrete Mathematics and Combinatorics Commons Recommended Citation Riggle, Matthew, "Runs of Identical Outcomes in a Sequence of Bernoulli Trials" (2018). Masters Theses & Specialist Projects. Paper This Thesis is brought to you for free and open access by TopSCHOLAR. It has been accepted for inclusion in Masters Theses & Specialist Projects by an authorized administrator of TopSCHOLAR. For more information, please contact topscholar@wku.edu.

2 RUNS OF IDENTICAL OUTCOMES IN A SEQUENCE OF BERNOULLI TRIALS A Thesis Presented to The Faculty of the Department of Mathematics Western Kentucky University Bowling Green, Kentucky In Partial Fulfillment Of the Requirements for the Degree Master of Science By Matthew Riggle May 2018

3

4 Dedicated to Mom and Dad and Michael

5 ACKNOWLEDGMENTS There are many individuals who contributed to the success of this project. First and foremost, I would like to thank my advisor, Dr. David Neal. He has provided a wealth of guidance and insight throughout the course of the project, and it is an honor to have written the last master s thesis that he will advise. (Happy retirement!) Next, I would like to thank the other readers of my committee, Dr. Melanie Autin and Dr. Ngoc Nguyen. Both of them have been supportive during the project, and were great professors for the many statistics classes I was able to take during my time at WKU. In addition to those named above, I would also like to thank the many other mathematics professors who have taught and advised me during my undergraduate and graduate career. I have enjoyed working with all of you, and I am grateful for all of the challenges and opportunities that helped me develop as a mathematician. Lastly, I would like to thank my friends and family who have supported me during my academic journey. In particular, Kathleen Bell deserves special recognition for doing countless homework assignments together, sharing in the experience of being calculus TAs, being a great office mate, and finally for joining me for several Thesis Thursday sessions. Also, I would like to thank Emily Keith, who came all the way from Illinois to attend my defense. iv

6 CONTENTS List of Tables vii Abstract viii Chapter 1. Introduction Chapter 2. Expectation for Obtaining a Run of Length The Direct Counting Approach for E[T 2,2 ] The Direct Counting Approach for E[T 2 ] Conditional Average for E[T 2,2 ] Numerical Examples Chapter 3. Markov Chains and Recursive Sequences Markov Chain Solutions A Larger Markov Chain Example Pseudo-Fibonacci Sequences Chapter 4. Expectation for Obtaining a Run of Length n Deriving E[T n ] Recursively Deriving E[T n,m ] Recursively Relationships Between E[T n ] and E[T n,m ] Conditional Expectations Applications of Conditional Runs Chapter 5. Simulations with Mathematica Chapter 6. Generalizations and Further Study Obtaining a Run with a Non-Constant Probability Obtaining a Run in a Finite Set v

7 6.3. Obtaining a Run with a Categorical Distribution Obtaining a Run from Multiple Sequences Appendix A Code for Theorem Appendix B Code for Theorem Appendix C Code for Theorems 4.4.1/ Appendix D Code for Conjecture References vi

8 LIST OF TABLES Table 3.1 Selected Pseudo-Fibonacci Sequences Table 4.1 Probabilities for a Fair Run Game Table 5.1 Simulation Results vii

9 RUNS OF IDENTICAL OUTCOMES IN A SEQUENCE OF BERNOULLI TRIALS Matthew Riggle May Pages Directed by: David Neal, Melanie Autin, and Ngoc Nguyen Department of Mathematics Western Kentucky University The Bernoulli distribution is a basic, well-studied distribution in probability. In this thesis, we will consider repeated Bernoulli trials in order to study runs of identical outcomes. More formally, for t N, we let X t Bernoulli(p), where p is the probability of success, q 1 p is the probability of failure, and all X t are independent. Then X t gives the outcome of the t th trial, which is 1 for success or 0 for failure. For n, m N, we define T n to be the number of trials needed to first observe n consecutive successes (where the n th success occurs on trial X Tn ). Likewise, we define T n,m to be the number of trials needed to first observe either n consecutive successes or m consecutive failures. We shall primarily focus our attention on calculating E[T n ] and E[T n,m ]. Starting with the simple cases of E[T 2 ] and E[T 2,2 ], we will use a variety of techniques, such as counting arguments and Markov chains, in order to derive the expectations. When possible, we shall also provide closed-form expressions for the probability mass function, cumulative distribution function, variance, and other values of interest. Eventually we will work our way to general formulas for E[T n ] and E[T n,m ]. We will also derive formulas for conditional averages, and discuss how famous results from probability such as Wald s Identity apply to our problem. Numerical examples will also be given in order to supplement the discussion and clarify the results. viii

10 CHAPTER 1 INTRODUCTION Throughout, we let X t Bernoulli(p), where p is the probability of success, q 1 p is the probability of failure, and all X t are independent. X t gives the outcome of the t th trial, which is 1 for success and 0 for failure. For integers n, m 2, we define T n to be the number of trials needed to first observe n consecutive successes (where the n th success occurs on trial X Tn ). Likewise, we define the stopping time T n,m to be the number of trials needed to first observe either n consecutive successes or m consecutive failures. What are E[T n ] and E[T n,m ]? Motivation for the main problem can be found in a variety of applications. In particular, this problem is closely related to the concept of the hot-hand fallacy, which was discussed in a famous psychology paper by Gilovich, Vallone, and Tversky [2]. The idea is that basketball fans have a mental picture of a player who makes several consecutive shots as being on fire, when in reality most such streaks can be ascribed to random chance. The hot-hand fallacy is closely related to the gambler s fallacy [8], which falsely assumes that even for independent events, long streaks of identical outcomes are less likely than probability would dictate. While we are largely interested in the average number of trials needed to observe a run of a given length, other work has been done on related problems. Several papers by Schilling discuss the average length of the longest run in a given finite number of Bernoulli trials [9], [10], [11]. These papers provide examples of some of the techniques we will use for our problem, including counting arguments and recursive methods. Another application of studying runs is related to the concept of optimal stopping, which seeks to identify when to take a certain action so as to maximize 1

11 (or minimize) some value. An example of a gambling application will be presented in this project, in which the player will quit after some number of consecutive wins or losses. We will investigate the probability of coming out ahead and the associated conditional winnings. Other research related to optimal stopping and the Bernoulli distribution can be found in [6] and [12]. Finally, while not the main focus of this project, it is possible to closely examine the probability mass function for the case T 2 in order to develop recursive Fibonacci-like sequences. It turns out that using p 1/2 returns precisely the Fibonacci sequence, and varying the probability parameter will provide a different recursive sequence. Using a modified version of Binet s Formula for the Fibonacci numbers allows us to define such sequences as a side result. Similar recursive sequences are examined in [1]. The two trivial cases, where p 0 or p 1, can be handled with little effort. If p 1, we will obtain a success on every trial, so it takes n trials to obtain a run of n successes. Thus E[T n ] n and E[T n,m ] n. If instead p 0, we will obtain a failure on every trial, so it takes m trials to obtain a run of m failures and E[T n,m ] m. However, we will never obtain a run of n successes, so E[T n ]. We will thus focus on the case where 0 < p < 1. We begin in Chapter 2 by studying the case where we wish to obtain a run of length 2, in order to illustrate the techniques used to derive formulas for the regular and conditional expectations. In Chapter 3, we use Markov chains as an alternate approach to the problem, which allows use to highlight some nice results related to the Fibonacci sequence. In Chapter 4, we generalize the problem for runs of length n or m where n, m 2 and compare the results with those obtained in Chapter 2. Chapter 5 describes simulation programs used to verify the results obtained in previous chapters. Finally, we present further generalizations and avenues for future study in Chapter 6. 2

12 CHAPTER 2 EXPECTATION FOR OBTAINING A RUN OF LENGTH 2 In this chapter, we utilize a variety of approaches to calculate E[T 2 ] and E[T 2,2 ], beginning with a direct counting approach. Surprisingly, this approach turns out to be more straightforward for the calculation of E[T 2,2 ], so we begin there The Direct Counting Approach for E[T 2,2 ] We begin by deriving the probability mass function f 2,2 (t) P (T 2,2 t), defined for t 2. We note that we must end either by obtaining a run of 2 successes (which occurs with probability p 2 ) or by obtaining a run of 2 failures (which occurs with probability q 2 ). Also, since we cannot obtain a run of length 2 before the observed run, all previous trials must alternate between success and failure. This means that for any given value of t, there are two valid options, depending on the parity of t. If t is even, then the only valid options are or , where 1 indicates a success and 0 indicates a failure. If t is odd, then the two valid options are or Thus, the outcome of the starting trial will depend on whether t is even or odd. We may then find the probability for any particular value of t, depending on whether it is even or odd. If t is even, then there will be either t successes and t 2 1 failures, or vice versa. Also, each valid sequence of trials has only one possible ordering. The corresponding probabilities are p t 2 +1 q t 2 1 and p t 2 1 q t 2 +1, so the total probability is p t 2 +1 q t p t 2 1 q t 2 +1, which can be rewritten as (pq) t 2 1 (p 2 + q 2 ). 3

13 If instead t is odd, then there will be either t+1 2 successes and t 1 2 failures, or vice versa. The corresponding probabilities are p t+1 2 q t 1 2 and p t 1 2 q t+1 2, so the total probability is p t+1 2 q t p t 1 2 q t+1 2, which can be simplified to (pq) t 1 2. Thus the probability mass function f 2,2 (t) for T 2,2 is as follows: p t 2 1 q t 2 1 (p 2 + q 2 ), if t is even f 2,2 (t) (pq) t 1 2, if t is odd. (2.1) Summing f 2,2 (t) over all possible values of t allows us to prove the following result. Theorem For 0 p 1, we will obtain a run of length 2 almost surely. Proof. For p 0 and p 1, it is guaranteed that the run will occur in the first two trials. For 0 < p < 1, we let even t 2k and odd t 2k + 1, and consider the piecewise sum of f 2,2 (t): P (T 2,2 < ) (pq) k 1 (p 2 + q 2 ) + k1 (pq) k (p 2 + q 2 ) + p2 + q 2 1 pq + pq 1 pq p2 + q 2 + pq 1 pq p2 + q(q + p) 1 p(1 p) p2 + q 1 p + p 2 p2 + q q + p 2 1. (pq) k k1 (pq) k+1 Since the sum of f 2,2 (t) over all values of t is equal to 1, we will obtain a run of 4

14 two successes or two failures almost surely. Theorem holds even though there are two sequences which never give a run of 2; i.e., those which forever alternate between success and failure. It is not difficult to show that these two sequences occur with probability 0. However, just because T 2,2 is finite-valued, it is not obvious that E[T 2,2 ] is finite. As a corollary to our work thus far, it is now convenient to utilize the probability mass function to develop a formula for the cumulative distribution function F 2,2 (t) P (T 2,2 t). We note that for even t 2k, the even sum will contain an extra term. For odd t 2k + 1, we have F 2,2 (t) k (pq) i 1 (p 2 + q 2 ) + i1 k (pq) i i1 (p2 + q 2 )(1 (pq) k ) pq (pq)k pq 1 pq (p2 + q 2 )(1 (pq) k ) pq (pq)k pq 1 pq (p2 + q 2 )(1 (pq) k ) + pq (pq) k+1 1 pq (p2 + q 2 )(1 (pq) k ) + pq(1 (pq) k ) 1 pq (p2 + q 2 + pq)(1 (pq) k ) 1 (1 q)q (p(p + q) + q2 )(1 (pq) k ) 1 q + q 2 (p + q2 )(1 (pq) k ) p + q 2 1 (pq) k, and for even t 2k, we have F 2,2 (t) k k 1 (pq) i 1 (p 2 + q 2 ) + (pq) i i1 i1 5

15 (p2 + q 2 )(1 (pq) k ) pq (pq)k + 1 pq 1 pq (p2 + q 2 )(1 (pq) k ) + pq (pq) k 1 pq (p2 + q 2 )(1 (pq) k ) + pq 1 + (1 (pq) k ) 1 pq (p2 + q 2 + 1)(1 (pq) k ) + pq 1 1 pq (p2 + q 2 + 1)(1 (pq) k ) 1 1 pq (p2 + (1 p) 2 + 1)(1 (pq) k ) 1 1 p(1 p) (p p + p 2 + 1)(1 (pq) k ) 1 1 p + p 2 (2p2 2p + 2)(1 (pq) k ) 1 p + p 2 1 2(1 (pq) k ) 1 1 2(pq) k. Then we reach the following corollary. Corollary The cumulative mass function F 2,2 (t), which gives the probability of obtaining a run of two identical trials within t trials, is given by 1 2(pq) t 2, if t is even F 2,2 (t) 1 (pq) t 1 2, if t is odd. Returning to the probability mass function allows us to calculate E[T 2,2 ]. Theorem Let T 2,2 be the number of trials required to obtain either a run of two successes or a run of two failures. Then E[T 2,2 ] 2 + pq 1 pq. Proof. We use two sums, one for even t and one for odd t. For even t 2k, we 6

16 have (2k)(pq) k 1 (p 2 + q 2 ), k1 and for odd t 2k + 1, we have (2k + 1)(pq) k. k1 Combining the even and odd sums, we obtain: E[T 2,2 ] (2k)(pq) k 1 (p 2 + q 2 ) + (2k + 1)(pq) k. k1 k1 Applying the formulas rk 1 1 r and that p + q 1 allows us to simplify the sum: E[T 2,2 ] (2k)(pq) k 1 (p 2 + q 2 ) + k1 2(p 2 + q 2 ) k1 krk 1 (2k + 1)(pq) k k1 k(pq) k 1 + 2pq k1 k(pq) k 1 + pq k1 2(p2 + q 2 ) (1 pq) 2 + 2pq (1 pq) 2 + pq (1 pq) 2p2 + 2q 2 + 2pq + pq(1 pq) (1 pq) 2 2p2 + 2q 2 + 3pq p 2 q 2 (1 pq) 2 (2p2 + 2pq) + (2q 2 + 2pq) pq p 2 q 2 (1 pq) 2 (2p2 + 2pq) + (2q 2 + 2pq) pq p 2 q 2 (1 pq) 2 2p(p + q) + 2q(p + q) pq p2 q 2 (1 pq) 2 1 and the fact (1 r) 2 (pq) k 7

17 2 pq p2 q 2 (1 pq) 2 (2 + pq)(1 pq) (1 pq) pq 1 pq. We note that the formula is also valid for p 0 or p 1; in either case, we obtain E[T 2,2 ] The Direct Counting Approach for E[T 2 ] We now consider the probability mass function f 2 (t) P (T t), defined for t 2, for first obtaining a run of two successes ending on the t th trial. The first few values of f 2 (t) are straightforward to calculate: t 2 corresponds to a run of two successes, so f 2 (2) p 2. Likewise, for t 3, we must obtain successes on trials 2 and 3, and a failure on trial 1 (or else we would have achieved the run at t 2). Thus f 2 (3) p 2 q. It is clear that for t > 3, the last three trials must be failure, success, and success, in that order, but there are more possibilities for the first t 3 trials (hereafter referred to as the pre-run ). The only restriction is that we cannot obtain two consecutive successes. Note that the number of successes in the pre-run cannot exceed t 2 2 run of two successes earlier than the t th trial., since otherwise we would have a We now consider the choice of successes within the pre-run. If the prerun contains no successes, this can only occur in one way (we will write this as ) for reasons that will soon be apparent). If the pre-run contains exactly ( t success, there are ( t 3 1 ) ways this may occur. If the pre-run contains more than 1 success, then each success we place prior to the last will necessarily be preceded by a failure, which removes 1 position from consideration. Thus a prerun that contains exactly 2 successes can occur in ( ) t 4 2 ways, and a pre-run that 8

18 contains exactly k successes can occur in ( ) t 2 k k ways. Finally, note that a prerun of length t 3 with k successes occurs with probability p k q t 3 k, so the entire observed sequence occurs with probability p k+2 q t 2 k. Thus, the probability mass function can be written as f 2 (t) t 2 2 ( t 2 k which leads to the following expected value k ) p k+2 q t 2 k, defined for integers t 2, (2.2) E[T 2 ] t2 t 2 2 ( t 2 k t k ) p k+2 q t 2 k. (2.3) As can be seen, the formula in (2.3) is quite unwieldy, so we now pursue a different method of enumeration. Instead of developing a formula based on the number of trials t, we will consider the number of failures. We know that the sequence must end with two successes, but before that, there may be any number of failures k {0, 1, 2,...}. Between any two failures or before the first failure, there can be at most one success. Thus there are i {0, 1, 2,..., k} successes, in addition to the two successes in the run. Thus the probability of our sequence containing exactly k failures is k i0 ( ) k p 2+i q k p 2 q k i Using this probability allows us to prove the following result. k i0 ( ) k p i p 2 q k (p + 1) k. (2.4) i Theorem For 0 < p 1, we will obtain a run of two successes almost surely. Proof. For p 1, it is guaranteed that the run will occur in the first two trials. 9

19 For 0 < p < 1, we have P (T 2 < ) p 2 q k (p + 1) k p 2 [q(p + 1)] k p 2 1 q(p + 1) p 2 1 pq q p2 p pq p 2 p(1 q) 1. As the total probability is equal to 1, we will obtain the run almost surely. In this case, there are infinitely many sequences which never give a run of 2 successes; however, the union of all such sequences has probability 0. We can also use the probability function in (2.4) to obtain a formula for E[T 2 ]. Theorem Let T 2 be the number of trials required to obtain a run of 2 successes. Then E[T 2 ] 1 + p p 2. Proof. We sum on the number of failures k, for k 0. Since we might have a success before each failure, there are i successes before the run, for some 0 i k. Together with the run of 2 successes, there is a total of 2 + k + i trials. Then E[T 2 ] k i0 ( ) k p 2+i q k (2 + k + i) i 10

20 k ( ) k p 2 q k p i (2 + k + i) i i0 k ( ) k p 2 q ((2 k + k) p i + i i0 k i0 ( ) ) k p i i i p 2 q ( k (2 + k)(p + 1) k + kp(p + 1) k 1) ( p 2 q k (2 + k)(p + 1) k + p 2 (2 q k (p + 1) k + ) q k kp(p + 1) k 1 q k (p + 1) k k + pq ) kq k 1 (p + 1) k 1 ( ) p q(p + 1) + (p + 1)q (1 q(p + 1)) + pq 2 (1 q(p + 1)) 2 p2 (2 2q(p + 1) + (p + 1)q + pq) (1 q(p + 1)) 2 p2 (2 q(p + 1) + pq) (1 qp q) 2 p2 (2 qp q + pq) (p pq) 2 p2 (2 q) (p(1 q)) 2 2 q p p p 2. The formula also works for p 1, as E[T 2 ] Conditional Average for E[T 2,2 ] We may also be interested in obtaining a run of n successes before we obtain a run of m failures. In order to derive the conditional expectation of the number of trials needed in this case, we will need the probability of first obtaining a run of successes. As before, we first seek a derivation for the case where n 2 and m 2. 11

21 Theorem Let P (S2) be the probability of obtaining a run of 2 successes before obtaining a run of 2 failures. Then P (S2) p(1 q2 ) 1 pq. Proof. Suppose we obtain the run of 2 successes first and that it takes t trials. Then the sequence of trials must end with two successes and be preceded by alternating successes and failures. If t is even, we will have two successes (occurring with probability p 2 ), plus an equal number of successes and failures in the preceding t 2 trials. If instead t is odd, we will have two successes (occurring with probability p 2 ), and there will be one additional success in the preceding t 2 trials. Again, by summing on the number of failures k, we obtain P (S2) p 2 (pq) k + qp 2 p2 1 pq + qp2 1 pq p2 + qp 2 1 pq p2 (1 + q) 1 pq p(1 q)(1 + q) 1 pq p(1 q2 ) 1 pq. (pq) k By symmetry, we also obtain the analogous result. Corollary Let P (F 2) be the probability of obtaining a run of 2 failures before obtaining a run of 2 successes. Then P (F 2) q(1 p2 ) 1 pq. Now, we shall derive a formula for E[T 2,2 S2]. 12

22 Theorem Let S2 be the event of obtaining a run of 2 successes before 2 + 3q pq2 obtaining a run of 2 failures. Then E[T 2,2 S2] (1 + q)(1 pq). Proof. We begin with two sums, corresponding to an even number of trials and an odd number of trials, respectively. We multiply by the number of trials in order to calculate the conditional expectation, and we divide by the probability of the given event P (S2): ( E[T 2,2 S2] 1 pq p 2 p(1 q 2 ) ( 1 pq 2p 2 p 2 (1 + q) ( 2pq (pq) k (2k + 2) + qp 2 (pq) k k + 2p 2 (pq) k (2k + 3) (pq) k + 2qp 2 1 pq 1 + q (1 pq) pq + 2q2 p (1 pq) + 3q 2 1 pq 1 pq ( ) 2pq + 2(1 pq) + 2q 2 p + 3q(1 pq) 1 + q (1 pq) 2 1 pq ( ) 2 + 3q q 2 p 1 + q (1 pq) q pq2 (1 + q)(1 pq). ) ) (pq) k k + 3qp 2 (pq) k ) Again, we obtain the analogous result by symmetry. Corollary Let F 2 be the event of obtaining a run of 2 failures before 2 + 3p qp2 obtaining a run of 2 successes. Then E[T 2,2 F 2] (1 + p)(1 pq) Numerical Examples We now present an example utilizing the formulas derived in the previous sections. Example Let p 0.4 be the probability of success on each trial. 13

23 By Theorem 2.2.2, the average number of trials needed to obtain a run of two successes is E[T 2 ] 1 + p p By Theorem 2.1.3, the average number of trials needed to obtain a run of two successes or a run of two failures is E[T 2,2 ] 2 + pq 1 pq 2 + (0.4)(0.6) 1 (.4)(.6) By Theorem 2.3.1, the probability of obtaining a run of two successes before a run of two failures is P (S2) p(1 q2 ) 1 pq (0.4)( ) 1 (.4)(.6) , and, similarly, by Corollary 2.3.2, the probability of obtaining a run of two failures before a run of two successes is P (F 2) q(1 p2 ) 1 pq (0.6)( ) 1 (.4)(.6) Finally, by Theorem 2.3.3, the average number of trials needed to obtain a run of two successes, given that we obtain such a run before a run of two failures, is E[T 2,2 S2] 2 + 3q pq2 (1 + q)(1 pq) 2 + 3(.6) (.4)(.6)2 (1 +.6)(1 (.4)(.6))

24 Likewise, by Corollary 2.3.4, the average number of trials needed to obtain a run of two failures, given that we obtain such a run before a run of two successes, is E[T 2,2 F 2] 2 + 3p qp2 (1 + p)(1 pq) 2 + 3(.4) (.6)(.4)2 (1 +.4)(1 (.4)(.6)) The previous two results demonstrate that T 2,2 depends on which run occurs first. The next example will treat the repeated Bernoulli trials as corresponding to individual steps in a one-dimensional random walk. A success corresponds to an upward step, and a failure corresponds to a downward step. More formally, if upward steps are a units each, and downward steps are b units each, then we can define the height H t as H t t (a1 {Xi 1} b1 {Xi 0}) i1 where 1 {Xt1} is an indicator variable that returns 1 for a success and 0 for a failure, and 1 {Xt0} is an indicator variable that returns 0 for a success and 1 for a failure. Then H t represents the height at time t (i.e., after t trials). A very useful result for studying random walks is Wald s Identity. Next, we will present a proof of Wald s Identity from Harper and Ross [3]. Theorem Wald s Identity: Suppose that H t is the height of a random walk defined by H t t i1 (a1 {X i 1} b1 {Xi 0}). If X t are i.i.d., T is a stopping time, and E[X 1 ] and E[T ] are both finite, then E[H T ] E[X 1 ] E[T ]. 15

25 Proof. As X t and the indicator 1 {T n} are independent, we have E[H T ] E[X n 1 {T n} ] n1 E[X n ] E[1 {T n} ] n1 E[X 1 ] P (T n) n1 E[X 1 ] n P (T n) n1 E[X 1 ] E[T ]. We will use Wald s Identity in the next example. Example Let p 0.4 be the probability of success on each trial, and let each upward step be a units and each downward step be b units. We would like to compute E[H T2 ] and E[H T2,2 ]. First, we observe that E[X 1 ] ap bq. Then, by applying Wald s Identity, we find E[H T2 ] E[X 1 ] E[T 2 ] ( ) 1 + p (ap bq) p 2 (0.4a 0.6b)(8.75) 3.5a 5.25b, 16

26 and E[H T2,2 ] E[X 1 ] E[T 2,2 ] ( ) 2 + pq (ap bq) 1 pq ( ) 56 (0.4a 0.6b) a b

27 CHAPTER 3 MARKOV CHAINS AND RECURSIVE SEQUENCES Markov Chain Solutions The formula for the probability mass function f 2 (t) derived with the direct counting approach, (2.2), is difficult to use directly. For this reason, we attempt to develop a simpler formula by using a Markov chain. In this situation, we have three different states, corresponding to having a current run of either 0, 1, or 2 successes. Thus we have a 1 3 initial state matrix B and a 3 3 transition matrix A: ( ) B p p 0 A 1 p 0 p The entries in B correspond to the initial probabilities; since we always begin with a run of 0 successes, there is a 1 in the first entry and 0 elsewhere. Each entry in A corresponds to the probability of moving from its row state to its column state. Note that the third row represents the absorbing state, since once we obtain a run of 2 successes, we will remain in that state indefinitely. Then the 1 3 matrix B A t represents the probabilities of being in each state after t trials. Because computing A t is quite difficult, we seek a diagonalization of A using standard techniques (see [7]). The eigenvalues of A can be found by solving its characteristic equation: (1 p λ)( λ)(1 λ) (1 p)(p)(1 λ) 0 18

28 (1 λ)[λ 2 (1 p)λ p(1 p)] 0 Solving for λ, we obtain: λ 1 or λ (1 p) ± (1 p)(1 + 3p). 2 Then the three eigenvalues of A are λ 1 1, λ 2 (1 p) + (1 p)(1 + 3p) 2, and λ 3 (1 p) (1 p)(1 + 3p). 2 We can also find the corresponding matrix of eigenvectors P, which is 1 (1 p)+ (1 p)(1+3p) (1 p) (1 p)(1+3p) 2(1 p) 2(1 p) P , where the i th column of P represents the eigenvector corresponding to λ i. Note that P is invertible, and P 1 1 p (1 p)(1 + 3p) (1 p)+ (1 p)(1+3p) 2(1 p) (1 p)+ (1 p)(1+3p) 2(1 p) (1 p)(1+3p) 1 p (1 p) (1 p)(1+3p) 2(1 p) (1 p) (1 p)(1+3p) 2(1 p). Then we have the diagonalization A P DP 1, where D is simply the diagonal matrix containing the eigenvalues on the main diagonal: D 0 (1 p)+ (1 p)(1+3p) (1 p) (1 p)(1+3p) 2.

29 This allows us to easily calculate A t P D t P 1. Note that the matrix ( ) B A t P (0) P (1) P (2) represents the final probabilities of being in each state after t trials. In particular, P (2) is the probability of obtaining a run of 2 successes in t or fewer trials, which is the cumulative mass function F 2 (t). Thus, when we perform the matrix multiplication A t P D t P 1, we can focus on this entry of the final state matrix, which is as follows: [ 1 p (1 p)(1 + 3p) P (2) (1 p)(1 + 3p) 1 p ( (1 p) + ) 2 ( (1 p)(1 + 3p) (1 p) + ) t (1 p)(1 + 3p) 2(1 p) 2 + ( (1 p) (1 p)(1 + 3p) 2(1 p) ) 2 ( (1 p) (1 p)(1 + 3p) 2 ) t. The above result simplifies to yield F 2 (t) 1 [ (1 p) + ] t+2 [ (1 p)(1 + 3p) (1 p) ] t+2 (1 p)(1 + 3p) 2 t+2 (1 p), (1 p)(1 + 3p) (3.1) from which we can derive f 2 (t) F 2 (t) F 2 (t 1). A remarkable simplification is possible in the case where p 1/2; in this case, we obtain [(1 12 ) + (1 12 )( ) ] t+2 [(1 12 ) (1 12 )( ) ] t+2 F 2 (t) 1 2 t+2 (1 1 2 ) (1 1 2 )( ) 20

30 ( ) t+2 ( ( ) t ( ) t+2 ( We note that t ) t+2 ) t ϕ, where ϕ is the golden ratio, and ϕ. This allows us to apply Binet s Formula, which states F n ϕn (1 ϕ) n ϕ (1 ϕ), where F n is the n th Fibonacci number. Thus we obtain the following theorem. Theorem The cumulative distribution function F 2 (t) for the probability of obtaining a run of 2 successes within t trials is given by F 2 (t) 1 F t+2 2 t, where F t is the t th Fibonacci number. Using Theorem 3.1.1, we can also obtain the pdf f 2 (t): f 2 (t) F 2 (t) F 2 (t 1) 1 F t+2 2 t 2F t+1 F t+2 2 t ( 1 F t+1 2 t 1 2F t+1 (F t+1 + F t ) 2 t F t+1 F t 2 t. ) We thus obtain the following result. Theorem The probability mass function f 2 (t) for the probability of obtaining a run of 2 successes ending on the t th trial is given by f 2 (t) F t 1 2 t, where F t is the t th Fibonacci number. We can use the probability distribution function in Theorem to provide an alternate proof that a run of 2 successes must have finite expectation in 21

31 the case where p 1/2. Theorem Suppose p 1/2. Then T 2 is finite almost surely, and E[T 2 ] is finite. Proof. To prove the first part of the theorem, we will show that the following sum is equal to 1: f 2 (t) t2 F t 1 2 t t2 ϕ t 1 (1 ϕ) t 1 2 t [ϕ (1 ϕ)] t2 1 ( ϕ t 1 2[ϕ (1 ϕ)] t2 (1 ϕ)t 1 2t 1 2 t 1 [ 1 (ϕ ) ( ) ] t t (1 ϕ) 2[ϕ (1 ϕ)] 2 2 t1 [ 1 ( ϕ ) t ( ) ] t (1 ϕ) 4ϕ t1 t1 ( ) ϕ 2 1 ϕ ϕ ( ) ϕ 1 1+ϕ 4ϕ ( 1 2 4ϕ 2 2 ϕ 2 ) 1 + ϕ ( ) ϕ 4 + 2ϕ 4ϕ 2 (2 ϕ)(1 + ϕ) ϕ ϕ ( 1+ 5) )

32 For the second part of the theorem, we will use the ratio test on the series E[T 2 ] t2 F t 1 2 t t, obtaining the following limit: ( ) ( ) ( ) Ft 2 t t + 1 lim t F t 1 2 t+1 t ( Ft lim t F t 1 ) ( ) 1 ϕ 2 2 < 1. Hence E[T 2 ] is finite A Larger Markov Chain Example We may apply the same Markov chain technique for longer runs where n > 2. In these cases, however, it will not be as straightforward to solve the characteristic equation and obtain a diagonalization of the transition matrix. We can, however, obtain numerical approximations for given values of p and t. Example Let p What is the probability of obtaining a run of 4 successes within 20 trials? Within 50 trials? What is the probability of first obtaining a run of 4 successes on the 10 th trial? To set up the Markov chain, we note that our initial state matrix will be a 1 5 matrix (allowing for runs of length 0, 1, 2, 3, and 4), and the transition matrix will be a 5 5 matrix. As before, the initial state matrix will contain a single 1 followed by all zeros. All but the last row of the transition matrix will contain 1 p in the first entry, p on row i, column i + 1, and zeros elsewhere. The last row, representing the absorbing state, will contain all zeros followed by a single 1 in the last entry. Thus, ( ) B

33 and A Then, to find the probability of obtaining a run of 4 successes within 20 trials, we simply need to compute ( ) B A 20 P (0) P (1) P (2) P (3) P (4), which yields the following (the values do not sum to 1 due to rounding): ( ) B A Therefore, from the last entry in the matrix, there is about a 79.53% chance of obtaining a run of 4 successes within 20 trials. Similarly, we can compute ( ) B A to find that there is about a 98.60% chance of obtaining a run of 4 successes within 50 trials. Essentially, we are computing specific values of the cumulative distribution function. Thus, in order to answer the last question, we need to compute two of these cdf values and subtract them. f 4 (10) F 4 (10) F 4 (9) (B A 10 ) 5 (B A 9 ) 5 24

34 Thus, there is about a 4.64% chance of first completing a run of 4 successes on the tenth trial Pseudo-Fibonacci sequences We can derive Binet s Formula for the n th Fibonacci number by using a generating function. We let G(x) F 0 + F 1 x + F 2 x 2 + F 3 x be the generating function, with coefficient F n representing the n th Fibonacci number. We also have the initial conditions F 0 0, F 1 1, and F n F n 1 + F n 2. Then we have G(x) F 0 + F 1 x + F 2 x 2 + F 3 x xg(x) F 0 x + F 1 x 2 + F 2 x 3 + F 3 x x 2 G(x) F 0 x 2 + F 1 x 3 + F 2 x 4 + F 3 x Subtracting the last two lines from the first line gives (1 x x 2 )G(x) F 0 + (F 1 F 0 )x, or G(x) x 1 x x 2. Thus we have solved for the generating function G(x). Note that the denominator can be factored (using ϕ and ϕ); hence, 2 2 G(x) x (1 ϕx)(1 (1 ϕ)x). 25

35 The right side of the equation can be rewritten using the partial fraction expansion G(x) A 1 ϕx + B 1 (1 ϕ)x. Through some straightforward algebra, we obtain A 1 5 and B 1 5. Note that 5 ϕ (1 ϕ). Thus we have G(x) ( ) 1 1 ϕ (1 ϕ) 1 ϕx (1 ϕ)x Note that we can rewrite the above using power series as follows: G(x) ( 1 (ϕx) n ϕ (1 ϕ) n0 ) [(1 ϕ)x] n. Then the coefficient F n for the x n term of G(x) is ϕn (1 ϕ) n, which is exactly ϕ (1 ϕ) Binet s Formula. F 2 (t) 1 We now consider the more general formula for F 2 (t) from (3.1): [ (1 p) + ] t+2 [ (1 p)(1 + 3p) (1 p) ] t+2 (1 p)(1 + 3p) 2 t+2 (1 p). (1 p)(1 + 3p) n0 We let α (1 p) + (1 p)(1 + 3p) and β (1 p) (1 p)(1 + 3p), and note that α β 2 (1 p)(1 + 3p), which is present in the denominator of [ (1 p) + ] n [ (1 p)(1 + 3p) (1 p) ] n (1 p)(1 + 3p) 2. (1 p)(1 + 3p) Then for n 0 we obtain 0, and for n 1 we obtain 1, matching the initial values of the Fibonacci sequence. However, we will need a different recursive relation than the one for the Fibonacci sequence. 26

36 For the Fibonacci generating function, we used the polynomial 1 x x 2 in order to eliminate all but a finite number of terms to give (1 x x 2 )G(x) F 0 + (F 1 F 0 )x. Now, the Fibonacci recursive relation may be written as F n F n 1 F n 2 0, which reveals the key insight. The coefficients of the polynomial and the recursion match (constant with F n, x with F n 1, and x 2 with F n 2 ). Thus, we seek a polynomial that factors into (1 αx)(1 βx). This turns out to be 4p(1 p) 2(1 p)x x 2. Then we can define a new recursive sequence, P n, of pseudo-fibonacci numbers with conditions P 0 0, P 1 1 P n 2(1 p)p n 1 + 4p(1 p)p n 2. Then by following the same argument as in the derivation of Binet s Formula, we obtain P n [ (1 p) + ] n [ (1 p)(1 + 3p) (1 p) ] n (1 p)(1 + 3p) 2 (1 p)(1 + 3p). (3.2) For p 1/2, (3.2) gives Binet s Formula for F n. Thus we have defined recursive sequences {P n } corresponding to probability values, which we can use to rewrite F 2 (t) P (T 2 t), as follows: F 2 (t) 1 P t+2 2 t+1 (1 p) 1 2(1 p)p t+1 + 4p(1 p)p t 2 t+1 (1 p) 27

37 1 P t+1 + 2pP t 2 t. It may be of interest to examine the sequences {P n }. Table 3.1 presents the beginning of the sequences for selected values of p. Table 3.1: Selected Pseudo-Fibonacci Sequences p First 7 terms of {P n } 0 0, 1, 2, 4, 8, 16, 32, , 1, 1.6, 3.2, 6.144, , , , 1, 1.4, 2.8, 5.096, , , , 1, 1, 2, 3, 5, 8, , 1, 0.5, 1, 0.875, , 1.25,... ϕ/2 0, 1, , , , , , , 1, 0.2, 0.4, 0.152, , , , 1, 0, 0, 0, 0, 0,... Note that some sequences are strictly increasing, while others oscillate. Some properties of the family {P n } may be observed. For instance, the only values which yield sequences of only integers are p 0, p 1/2, and p 1. Using p 0 gives powers of 2, and of course p 1/2 returns the familiar Fibonacci sequence. Examining the end behavior of the sequences also provides some interesting properties. For 0 p 1/2, {P n } is a monotone increasing sequence. For 1/2 < p < ϕ/2, {P n } oscillates a finite number of times before eventually increasing to infinity. For p ϕ/2, {P n } oscillates infinitely often, but the oscillations become smaller and smaller, and the sequence converges to the limit ϕ 1. For p > ϕ/2, {P n } eventually becomes decreasing and converges to 0. 28

38 Variations on the Fibonacci sequence are not a new concept. For instance, the generalized Fibonacci sequence modifies the starting values while preserving the coefficients in the recursive relation. In contrast, the family of pseudo- Fibonacci sequences defined here modified the coefficients in the recursive relation while preserving the starting values of 0 and 1. More results and applications related to generalized Fibonacci numbers can be found in Koshy [5]. 29

39 CHAPTER 4 EXPECTATION FOR OBTAINING A RUN OF LENGTH N Again, we define the stopping time T n to be the number of trials needed to first observe n consecutive successes, and we define the stopping time T n,m to be the number of trials needed to first observe either n consecutive successes or m consecutive failures. Because the counting arguments presented in Chapter 2 would seemingly be quite complicated for n, m > 2, we seek a more general method to calculate E[T n ] and E[T n,m ] Deriving E[T n ] Recursively Before we begin our procedure to derive a formula for E[T n ], we must first show that it is finite for any non-zero choice of p and any choice of n. Theorem For 0 < p 1 and fixed n, E[T n ] is finite. Proof. We have already proven the result for the trivial case of p 1, so suppose 0 < p < 1. Consider the trials in blocks of n; i.e., trials X 1 through X n, trials X n+1 through X 2n, and so on. Each block contains the run we desire with probability p n. The blocks correspond to a geometric distribution with parameter p n, so the expected number of blocks required to observe a run is 1/p n, which corresponds to n/p n trials. This expectation, which is finite, in fact provides an upper bound for E[T n ], since the first run might occur across two different blocks. Thus we can safely conclude that E[T n ] must be finite for any choice of n. As an aside, note that the case n 1 is equivalent to the geometric distribution with parameter p, which has expected value 1/p. We have already derived a formula for n 2 in Theorem 2.2.2, but we would like a general formula. We now develop such a formula by recursively calculating the weighted average. 30

40 Theorem Let T n be the number of trials required to obtain a run of n consecutive successes. Then for 0 < p < 1, E[T n ] 1 pn p n (1 p). Proof. Suppose we obtain a run of n successes immediately. This situation occurs with probability p n and requires n trials. If instead we do not obtain a run of n successes, then we must have first obtained a run of k successes followed by a failure, for some k {0, 1,..., n 1}. This situation occurs with probability p k (1 p). We have already observed k + 1 trials, and since the failure resets the run, we will require E[T n ] additional trials, for a total of k +1+E[T n ] trials. Thus we obtain the following: n 1 E[T n ] p n n + p k (1 p)(k E[T n ]) n 1 n 1 p n n + p k (1 p)e[t n ] + p k (1 p)(k + 1) n 1 n 1 p n n + (1 p n )E[T n ] + p k (k + 1) p k+1 (k + 1). Because E[T n ] is finite, we can subtract it from the right-hand side. Solving for E[T n ], we obtain n 1 p n E[T n ] p n n + p k (k + 1) n p k k k1 n 1 n 1 p n n p k (k + 1) p k k p n n n n 1 k1 p k 1 pn 1 p, p k k1 31 k1

41 which gives the result. Note that if we use n 1 in Theorem 4.1.2, we obtain the expectation E[T 1 ] 1, which is consistent with the geometric distribution. Furthermore, for p n 2, we obtain E[T 1 ] 1 p2 p 2 (1 p) 1 + p, which agrees with Theorem p 2 Thus we have solved for E[T n ] in all cases:, p 0 1 p E[T n ] n p n (1 p), 0 < p < 1 n, p 1. (4.1) We now turn our attention to E[T n,m ] Deriving E[T n,m ] Recursively For 0 < p < 1, it is not difficult to see that E[T n,m ] must be finite by using Theorem 4.1.1, as it is clear that E[T n,m, ] E[T n ]. Corollary For 0 p 1 and fixed n and m, E[T n,m ] is finite. Since Corollary guarantees E[T n,m ] must be finite for any parameter p, we now seek its derivation. The resulting formula is given as the following theorem. Theorem Let T n,m be the number of trials required to obtain either a run of n successes or a run of m failures. Then E[T n,m ] is given by m, p 0 (1 p E[T n,m ] n )(1 q m ) p n q + pq m p n q, 0 < p < 1 m n, p 1. 32

42 Proof. In order to facilitate the creation of formulas in this proof, we define two additional variables. We let A 1 be the expected number of additional trials needed when starting with a single success, and we let A 2 be the expected number of additional trials needed when starting with a single failure. Both A 1 and A 2 must be finite, as both values are less than or equal to T n,m. Now, consider the first trial, which results in success with probability p and failure with probability q. We may then write E[T n,m ] 1 + pa 1 + qa 2. (4.2) Suppose we have a current run of one success. If we obtain successes in the next n 1 trials, we are finished. This situation occurs with probability p n 1. If instead we do not obtain a run of n 1 successes, then we must have first obtained a run of k successes followed by a failure, for some k {0, 1,..., n 2}. This situation occurs with probability p k q. We have then observed k + 1 trials, and since we now have a current run of one failure, we will require A 2 additional trials, for a total of k A 2 trials. Thus we obtain the following: n 2 A 1 p n 1 (n 1) + p k q(k A 2 ) n 2 n 2 p n 1 (n 1) + p k (1 p)a 2 + p k (1 p)(k + 1) n 2 n 2 p n 1 (n 1) + (1 p n 1 )A 2 + p k (k + 1) p k+1 (k + 1). We proceed by moving the A 2 term to the left side: n 2 n 1 A 1 A 2 (1 p n 1 ) p n 1 (n 1) + p k (k + 1) p k k k1 n 2 n 2 p n 1 (n 1) p k (k + 1) p k k p n 1 (n 1) 33 k1 k1

43 The result is as follows: n k1 n 2 p k. p k A 1 A 2 (1 p n 1 ) 1 pn 1 1 p. (4.3) If we apply the same steps with success and failure reversed, we also obtain A 2 A 1 (1 q m 1 ) 1 qm 1. (4.4) 1 q Thus, (4.3) and (4.4) give a system of two equations in the two variables A 1 and A 2. Solving the system gives the following: A 1 (1 pn 1 )(1 q m ) p n q + pq m p n q m, A 2 (1 qm 1 )(1 p n ) p n q + pq m p n q m. (4.5) For (4.2), we need pa 1 and pa 2. First, we have pa 1 p(1 pn 1 )(1 q m ) p n q + pq m p n q m p pn pq m + p n q m p n q + pq m p n q m p pn + (p n q p n q) pq m + p n q m p n q + pq m p n q m p pn + p n q p n q + pq m p n q 1 m p p n+1 p n q + pq m p n q 1. m 34

44 Similarly, we have pa 2 q q m+1 p n q + pq m p n q m 1. Substituting these expressions into (4.2) and simplifying, we obtain E[T n,m ] 1 + pa 1 + qa 2 p p n+1 p n q + pq m p n q m + q q m+1 p n q + pq m p n q m 1 p pn+1 + q q m+1 p n q pq m + p n q m p n q + pq m p n q m 1 pn (p + q) q m (q + p) + p n q m p n q + pq m p n q m 1 pn q m + p n q m p n q + pq m p n q m (1 pn )(1 q m ) p n q + pq m p n q m, which is valid for 0 < p < 1. It should be noted that the formula in Theorem confirms the formula for n, m 2 derived in Chapter 2. Making the substitutions n 2 and m 2 and simplifying yields E[T 2,2 ] (1 p2 )(1 q 2 ) p 2 q + pq 2 p 2 q 2 1 p2 q 2 + p 2 q 2 pq(p + q) p 2 q 2 p + q p2 q 2 + p 2 q 2 pq p 2 q 2 p(1 p) + q(1 q) + p2 q 2 pq(1 pq) 2pq + p2 q 2 pq(1 pq) 35

45 2 + pq 1 pq, which agrees with Theorem Relationships Between E[T n ] and E[T n,m ] For 0 < p < 1, we have obtained the following expectations E[T n ] 1 pn p n (1 p), E[T n,m] (1 pn )(1 q m ) p n q + pq m p n q m. (4.6) Also, suppose that we wish to obtain a run of m consecutive failures, where the probability of failure is q. We denote the number of trials required by T m. Then, by symmetry, E[T m] is given by E[T m] 1 qm q m (1 q). We then have the following identity. Theorem For 0 < p < 1, 1 E[T n ] + 1 E[Tm] 1 E[T n,m ]. Proof. Simplifying the left side of the identity gives 1 E[T n ] + 1 E[Tm] pn (1 p) + qm (1 q) 1 p n 1 q m pn q(1 q m ) + q m p(1 p n ) (1 p n )(1 q m ) pn q p n q m+1 + q m p q m p n+1 (1 p n )(1 q m ) pn q + pq m p n q m (1 p n )(1 q m ) 1 E[T n,m ]. 36

46 In a sense, we can think of E[T n ] and E[Tm] as two components of E[T n,m ]. Theorem then says that the reciprocal of E[T n,m ] is equal to the sum of the reciprocals of its two components. In the special case where p q 1/2, we can relate E[T n ] and E[T n,n ]. Theorem For p q 1/2, E[T n ] 2E[T n,n ]. Proof. For p 1/2, we have the following: ) n E[T n ] 1 ( 1 2 ) n (1 1 ( 1 2 ) n 2 ) 1 ( 1 2 ( 1 n+1 ( 2) ( 1 1 ) n ) 2 2 ) n ( 1 2 2(2 n 1); [ ( 1 1 ) n ] [ ( n ] E[T n,n ] ( 2) 1 ) n+1 ( ) n+1 ( 2 1 2n 2) [ 1 ( 1 2) n ] 2 ( 1 ) n ( 2 1 2n 2) [ ( 1 1 n ] 2 ( 2) 1 ) n [ ( n ] 2) 1 ( 1 2 ( 1 n 2) ) n 2 n 1 E[T n]. 2 37

47 4.4 - Conditional Expectations Suppose we are now interested in the event that we obtain a run of n successes before we obtain a run of m failures, which we denote R nbm. We shall derive P (R nbm ) by considering cases of how smaller runs may occur before a run of n successes. Theorem Let R nbm be the event that we obtain a run of n successes before p n 1 (1 q m ) a run of m failures. Then P (R nbm ) p n 1 + q m 1 p n 1 q. m 1 Analogously, if Rnbm C is the event that we obtain a run of m failures before a run of n successes, then P (Rnbm C ) q m 1 (1 p n ) p n 1 + q m 1 p n 1 q m 1 Proof. For R nbm, we have two cases. In Case I, we start with a success, which occurs with probability p. We might immediately obtain all remaining n 1 successes, or our potential run of successes might be interrupted. An interruption will have two parts: some number of successes followed by a failure (part 1), then some number of failures followed by a success (part 2). Note that in part 1, we must obtain a failure within n 1 trials, and in part 2, we must obtain a success within m 1 trials (or else we will obtain a run of m failures, which is not allowed). After the interruption, we have one success, which returns to the original Case I scenario. To summarize, we have one success, followed by some number of interruptions (possibly 0), followed by the remaining n 1 successes. Also note that we can use the geometric distribution for each interruption; i.e., the probability of obtaining a failure within n 1 trials is 1 p n 1, and the probability of obtaining a success within m 1 trials is 1 q m 1. Thus, by summing on the number of 38

48 possible interruptions k, the probability of Case I is P 1 p n [(1 p n 1 )(1 q m 1 )] k. In Case II, we start with a failure. We must then obtain some number of failures followed by a success, which returns us to the beginning of case I. We then obtain some number of interruptions followed by the final n 1 successes. In this case, we will have the probability of the failure, the probability of the n 1 successes, and a single geometric probability outside the summation for the interruptions. The probability in this case is P 2 qp n 1 (1 q m 1 ) [(1 p n 1 )(1 q m 1 )] k. The total probability, P (R nbm ), is simply the sum of P 1 and P 2 : P (R nbm ) P 1 + P 2 p n [(1 p n 1 )(1 q m 1 )] k + qp n 1 (1 q m 1 ) [(1 p n 1 )(1 q m 1 )] k [p n + qp n 1 (1 q m 1 )] [(1 p n 1 )(1 q m 1 )] k pn + qp n 1 (1 q m 1 ) 1 (1 p n 1 )(1 q m 1 ) pn 1 (p + q(1 q m 1 )) p n 1 + q m 1 p n 1 q m 1 p n 1 (1 q m ) p n 1 + q m 1 p n 1 q. m 1 We can define Rnbm C analogously, and we can find P (RC nbm ) similarly. Thus, 39

49 we have P (R nbm ) p n 1 (1 q m ) p n 1 + q m 1 p n 1 q m 1, and P (R C nbm) q m 1 (1 p n ) p n 1 + q m 1 p n 1 q m 1. The derivation of the conditional probabilities leads to the following result. Theorem Let T n,m R nbm be the number of trials required to obtain a run of n successes given that we obtain such a run before a run of m failures. Then E[T n,m R nbm ] n + (1 pn )(q q m (m+q mq))+(1 q m )(1 q m 1 )(p p n (n+p np)) (1 q m )(p n q+pq m p n q m ). Also, if T n,m R C nbm is the number of trials required to obtain a run of m failures given that we obtain such a run before a run of n successes, then E[T n,m R C nbm ] m + (1 qm )(p p n (n+p np))+(1 p n )(1 p n 1 )(q q m (m+q mq)) (1 p n )(p n q+pq m p n q m ). Proof. As before, we consider the average in each of the two cases. To facilitate this, let us define conditional averages for the geometrically distributed interruptions. For X Geometric(p), we define e 1 E[X X m 1], and likewise, for Y Geometric(q), we define e 2 E[Y Y n 1]. Then for Case I we have A 1 p n [n + k(e 1 + e 2 )][(1 p n 1 )(1 q m 1 )] k, (4.7) 40

50 and for Case II we have A 2 qp n 1 (1 q m 1 ) We then obtain [n + e 1 + k(e 1 + e 2 )][(1 p n 1 )(1 q m 1 )] k. (4.8) We will now derive e 1 and e 2. Lemma Suppose X Geometric(p). Then E[T n,m R nbm ] A 1 + A 2 P (R nbm ). (4.9) E[X X m 1] 1 qm 1 (m+q mq). p(1 q m 1 ) m 1 k1 kp (k) Proof. We have E[X X m 1], and since X Geometric(p), P (X m 1) P (k) pq k 1 and P (X m 1) 1 q m 1. Thus we obtain E[X X m 1] m 1 k1 kpqk 1 1 q m 1 p[1 qm 1 (m + q mq)] p 2 (1 q m 1 ) 1 qm 1 (m + q mq) p(1 q m 1 ) Then, by Lemma 4.4.3, we have e 1 1 qm 1 (m + q mq) p(1 q m 1 ) 1 p n 1 (n + p np). Now consider A q(1 p n 1 1 from (4.7): ) and e 2 A 1 p n [n + k(e 1 + e 2 )][(1 p n 1 )(1 q m 1 )] k [ p n n[(1 p n 1 )(1 q m 1 )] k + ( ) 1 q m 1 (m + q mq) k [(1 p n 1 )(1 q m 1 )] k p(1 q m 1 ) 41

51 ( ) ] 1 p n 1 (n + p np) + k [(1 p n 1 )(1 q m 1 )] k q(1 p n 1 ) np n [(1 p n 1 )(1 q m 1 )] k + [ p n 1 (1 p n 1 )(1 q m 1 (m + q mq)) ] + pn q (1 qm 1 )(1 p n 1 (n + p np)) k[(1 p n 1 )(1 q m 1 )] k 1 np n 1 (1 p n 1 )(1 q m 1 ) + pn 1 q(1 p n 1 )(1 q m 1 (m + q mq)) q[1 (1 p n 1 )(1 q m 1 )] 2 + pn (1 q m 1 (1 p n 1 (n + p np)) q[1 (1 p n 1 )(1 q m 1 )] ( ) 2 1 (np n q(p n 1 + q m 1 p n 1 q m 1 ) q[p n 1 + q m 1 p n 1 q m 1 ] 2 +p n 1 q(1 p n 1 )(1 q m 1 (m + q mq)) +p n (1 q m 1 )(1 p n 1 (n + p np)) ). Consider A 2 from (4.8). A 2 qp n 1 (1 q m 1 ) [n + e 1 + k(e 1 + e 2 )][(1 p n 1 )(1 q m 1 )] k [ qp n 1 (1 q m 1 ) n[(1 p n 1 )(1 q m 1 )] k ( 1 q m 1 (m + q mq) + (k + 1) p(1 q m 1 ) ( 1 p n 1 (n + p np) + k q(1 p n 1 ) nqp n 1 (1 q m 1 ) 1 (1 p n 1 )(1 q m 1 ) + qp n 2 (1 q m 1 (m + q mq)) ) [(1 p n 1 )(1 q m 1 )] k ) [(1 p n 1 )(1 q m 1 )] k ] (k + 1)[(1 p n 1 )(1 q m 1 )] k + pn 1 (1 q m 1 )(1 p n 1 (n + p np)) 1 p n 1 42 k[(1 p n 1 )(1 q m 1 )] k

Boundary Problems for One and Two Dimensional Random Walks

Boundary Problems for One and Two Dimensional Random Walks Western Kentucky University TopSCHOLAR Masters Theses & Specialist Projects Graduate School 5-2015 Boundary Problems for One and Two Dimensional Random Walks Miky Wright Western Kentucky University, miky.wright768@topper.wku.edu

More information

Gambler s Ruin with Catastrophes and Windfalls

Gambler s Ruin with Catastrophes and Windfalls Journal of Grace Scientific Publishing Statistical Theory and Practice Volume 2, No2, June 2008 Gambler s Ruin with Catastrophes and Windfalls B unter, Department of Mathematics, University of California,

More information

The Geometric Random Walk: More Applications to Gambling

The Geometric Random Walk: More Applications to Gambling MATH 540 The Geometric Random Walk: More Applications to Gambling Dr. Neal, Spring 2008 We now shall study properties of a random walk process with only upward or downward steps that is stopped after the

More information

1 Gambler s Ruin Problem

1 Gambler s Ruin Problem 1 Gambler s Ruin Problem Consider a gambler who starts with an initial fortune of $1 and then on each successive gamble either wins $1 or loses $1 independent of the past with probabilities p and q = 1

More information

Chapter 11 - Sequences and Series

Chapter 11 - Sequences and Series Calculus and Analytic Geometry II Chapter - Sequences and Series. Sequences Definition. A sequence is a list of numbers written in a definite order, We call a n the general term of the sequence. {a, a

More information

THE N-VALUE GAME OVER Z AND R

THE N-VALUE GAME OVER Z AND R THE N-VALUE GAME OVER Z AND R YIDA GAO, MATT REDMOND, ZACH STEWARD Abstract. The n-value game is an easily described mathematical diversion with deep underpinnings in dynamical systems analysis. We examine

More information

Introduction to Techniques for Counting

Introduction to Techniques for Counting Introduction to Techniques for Counting A generating function is a device somewhat similar to a bag. Instead of carrying many little objects detachedly, which could be embarrassing, we put them all in

More information

Random walk on a polygon

Random walk on a polygon IMS Lecture Notes Monograph Series Recent Developments in Nonparametric Inference and Probability Vol. 50 (2006) 4 c Institute of Mathematical Statistics, 2006 DOI: 0.24/0749270600000058 Random walk on

More information

Examining the accuracy of the normal approximation to the poisson random variable

Examining the accuracy of the normal approximation to the poisson random variable Eastern Michigan University DigitalCommons@EMU Master's Theses and Doctoral Dissertations Master's Theses, and Doctoral Dissertations, and Graduate Capstone Projects 2009 Examining the accuracy of the

More information

Lecture 9 Classification of States

Lecture 9 Classification of States Lecture 9: Classification of States of 27 Course: M32K Intro to Stochastic Processes Term: Fall 204 Instructor: Gordan Zitkovic Lecture 9 Classification of States There will be a lot of definitions and

More information

We are going to discuss what it means for a sequence to converge in three stages: First, we define what it means for a sequence to converge to zero

We are going to discuss what it means for a sequence to converge in three stages: First, we define what it means for a sequence to converge to zero Chapter Limits of Sequences Calculus Student: lim s n = 0 means the s n are getting closer and closer to zero but never gets there. Instructor: ARGHHHHH! Exercise. Think of a better response for the instructor.

More information

THESIS. Presented in Partial Fulfillment of the Requirements for the Degree Master of Science in the Graduate School of The Ohio State University

THESIS. Presented in Partial Fulfillment of the Requirements for the Degree Master of Science in the Graduate School of The Ohio State University The Hasse-Minkowski Theorem in Two and Three Variables THESIS Presented in Partial Fulfillment of the Requirements for the Degree Master of Science in the Graduate School of The Ohio State University By

More information

General Boundary Problem: System of Equations Solution

General Boundary Problem: System of Equations Solution MATH 54 General Boundary Problem: System of Equations Solution Consider again the scenario of a random walk X that starts at positive height j, and on each independent step, moves upward a units with probability

More information

The Kth M-ary Partition Function

The Kth M-ary Partition Function Indiana University of Pennsylvania Knowledge Repository @ IUP Theses and Dissertations (All) Fall 12-2016 The Kth M-ary Partition Function Laura E. Rucci Follow this and additional works at: http://knowledge.library.iup.edu/etd

More information

THE DYNAMICS OF SUCCESSIVE DIFFERENCES OVER Z AND R

THE DYNAMICS OF SUCCESSIVE DIFFERENCES OVER Z AND R THE DYNAMICS OF SUCCESSIVE DIFFERENCES OVER Z AND R YIDA GAO, MATT REDMOND, ZACH STEWARD Abstract. The n-value game is a dynamical system defined by a method of iterated differences. In this paper, we

More information

1 Presessional Probability

1 Presessional Probability 1 Presessional Probability Probability theory is essential for the development of mathematical models in finance, because of the randomness nature of price fluctuations in the markets. This presessional

More information

Equality of P-partition Generating Functions

Equality of P-partition Generating Functions Bucknell University Bucknell Digital Commons Honors Theses Student Theses 2011 Equality of P-partition Generating Functions Ryan Ward Bucknell University Follow this and additional works at: https://digitalcommons.bucknell.edu/honors_theses

More information

At the start of the term, we saw the following formula for computing the sum of the first n integers:

At the start of the term, we saw the following formula for computing the sum of the first n integers: Chapter 11 Induction This chapter covers mathematical induction. 11.1 Introduction to induction At the start of the term, we saw the following formula for computing the sum of the first n integers: Claim

More information

Lecture 4: Probability and Discrete Random Variables

Lecture 4: Probability and Discrete Random Variables Error Correcting Codes: Combinatorics, Algorithms and Applications (Fall 2007) Lecture 4: Probability and Discrete Random Variables Wednesday, January 21, 2009 Lecturer: Atri Rudra Scribe: Anonymous 1

More information

RANDOM WALKS AND THE PROBABILITY OF RETURNING HOME

RANDOM WALKS AND THE PROBABILITY OF RETURNING HOME RANDOM WALKS AND THE PROBABILITY OF RETURNING HOME ELIZABETH G. OMBRELLARO Abstract. This paper is expository in nature. It intuitively explains, using a geometrical and measure theory perspective, why

More information

1 Normal Distribution.

1 Normal Distribution. Normal Distribution.. Introduction A Bernoulli trial is simple random experiment that ends in success or failure. A Bernoulli trial can be used to make a new random experiment by repeating the Bernoulli

More information

4 Branching Processes

4 Branching Processes 4 Branching Processes Organise by generations: Discrete time. If P(no offspring) 0 there is a probability that the process will die out. Let X = number of offspring of an individual p(x) = P(X = x) = offspring

More information

The Boundary Problem: Markov Chain Solution

The Boundary Problem: Markov Chain Solution MATH 529 The Boundary Problem: Markov Chain Solution Consider a random walk X that starts at positive height j, and on each independent step, moves upward a units with probability p, moves downward b units

More information

4. Linear transformations as a vector space 17

4. Linear transformations as a vector space 17 4 Linear transformations as a vector space 17 d) 1 2 0 0 1 2 0 0 1 0 0 0 1 2 3 4 32 Let a linear transformation in R 2 be the reflection in the line = x 2 Find its matrix 33 For each linear transformation

More information

Class 26: review for final exam 18.05, Spring 2014

Class 26: review for final exam 18.05, Spring 2014 Probability Class 26: review for final eam 8.05, Spring 204 Counting Sets Inclusion-eclusion principle Rule of product (multiplication rule) Permutation and combinations Basics Outcome, sample space, event

More information

= W z1 + W z2 and W z1 z 2

= W z1 + W z2 and W z1 z 2 Math 44 Fall 06 homework page Math 44 Fall 06 Darij Grinberg: homework set 8 due: Wed, 4 Dec 06 [Thanks to Hannah Brand for parts of the solutions] Exercise Recall that we defined the multiplication of

More information

Ranking Score Vectors of Tournaments

Ranking Score Vectors of Tournaments Utah State University DigitalCommons@USU All Graduate Plan B and other Reports Graduate Studies 5-2011 Ranking Score Vectors of Tournaments Sebrina Ruth Cropper Utah State University Follow this and additional

More information

UC Berkeley Department of Electrical Engineering and Computer Science. EE 126: Probablity and Random Processes. Solutions 5 Spring 2006

UC Berkeley Department of Electrical Engineering and Computer Science. EE 126: Probablity and Random Processes. Solutions 5 Spring 2006 Review problems UC Berkeley Department of Electrical Engineering and Computer Science EE 6: Probablity and Random Processes Solutions 5 Spring 006 Problem 5. On any given day your golf score is any integer

More information

(x 1 +x 2 )(x 1 x 2 )+(x 2 +x 3 )(x 2 x 3 )+(x 3 +x 1 )(x 3 x 1 ).

(x 1 +x 2 )(x 1 x 2 )+(x 2 +x 3 )(x 2 x 3 )+(x 3 +x 1 )(x 3 x 1 ). CMPSCI611: Verifying Polynomial Identities Lecture 13 Here is a problem that has a polynomial-time randomized solution, but so far no poly-time deterministic solution. Let F be any field and let Q(x 1,...,

More information

DR.RUPNATHJI( DR.RUPAK NATH )

DR.RUPNATHJI( DR.RUPAK NATH ) Contents 1 Sets 1 2 The Real Numbers 9 3 Sequences 29 4 Series 59 5 Functions 81 6 Power Series 105 7 The elementary functions 111 Chapter 1 Sets It is very convenient to introduce some notation and terminology

More information

Probability and Distributions

Probability and Distributions Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated

More information

The Geometric Distribution

The Geometric Distribution MATH 382 The Geometric Distribution Dr. Neal, WKU Suppose we have a fixed probability p of having a success on any single attempt, where p > 0. We continue to make independent attempts until we succeed.

More information

arxiv: v1 [math.co] 21 Sep 2015

arxiv: v1 [math.co] 21 Sep 2015 Chocolate Numbers arxiv:1509.06093v1 [math.co] 21 Sep 2015 Caleb Ji, Tanya Khovanova, Robin Park, Angela Song September 22, 2015 Abstract In this paper, we consider a game played on a rectangular m n gridded

More information

Introduction to Probability

Introduction to Probability LECTURE NOTES Course 6.041-6.431 M.I.T. FALL 2000 Introduction to Probability Dimitri P. Bertsekas and John N. Tsitsiklis Professors of Electrical Engineering and Computer Science Massachusetts Institute

More information

Assignment 16 Assigned Weds Oct 11

Assignment 16 Assigned Weds Oct 11 Assignment 6 Assigned Weds Oct Section 8, Problem 3 a, a 3, a 3 5, a 4 7 Section 8, Problem 4 a, a 3, a 3, a 4 3 Section 8, Problem 9 a, a, a 3, a 4 4, a 5 8, a 6 6, a 7 3, a 8 64, a 9 8, a 0 56 Section

More information

MATHEMATICS 23a/E-23a, Fall 2015 Linear Algebra and Real Analysis I Module #1, Week 4 (Eigenvectors and Eigenvalues)

MATHEMATICS 23a/E-23a, Fall 2015 Linear Algebra and Real Analysis I Module #1, Week 4 (Eigenvectors and Eigenvalues) MATHEMATICS 23a/E-23a, Fall 205 Linear Algebra and Real Analysis I Module #, Week 4 (Eigenvectors and Eigenvalues) Author: Paul Bamberg R scripts by Paul Bamberg Last modified: June 8, 205 by Paul Bamberg

More information

ALL TEXTS BELONG TO OWNERS. Candidate code: glt090 TAKEN FROM

ALL TEXTS BELONG TO OWNERS. Candidate code: glt090 TAKEN FROM How are Generating Functions used in finding the closed form of sequences involving recurrence relations and in the analysis of probability distributions? Mathematics Extended Essay Word count: 3865 Abstract

More information

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes Lecture Notes 7 Random Processes Definition IID Processes Bernoulli Process Binomial Counting Process Interarrival Time Process Markov Processes Markov Chains Classification of States Steady State Probabilities

More information

5 CORRELATION. Expectation of the Binomial Distribution I The Binomial distribution can be defined as: P(X = r) = p r q n r where p + q = 1 and 0 r n

5 CORRELATION. Expectation of the Binomial Distribution I The Binomial distribution can be defined as: P(X = r) = p r q n r where p + q = 1 and 0 r n 5 CORRELATION The covariance of two random variables gives some measure of their independence. A second way of assessing the measure of independence will be discussed shortly but first the expectation

More information

Pistol-Dueling and the Thue-Morse Sequence

Pistol-Dueling and the Thue-Morse Sequence Pistol-Dueling and the Thue-Morse Sequence Tarkan Al-Kazily Spring 2017 Contents 1 Introduction 1 2 The Pistol Duel Rules 2 3 Enter the Thue-Morse Sequence 3 4 The Fairest Sequence of Them All? 7 5 Conclusions

More information

CS 246 Review of Proof Techniques and Probability 01/14/19

CS 246 Review of Proof Techniques and Probability 01/14/19 Note: This document has been adapted from a similar review session for CS224W (Autumn 2018). It was originally compiled by Jessica Su, with minor edits by Jayadev Bhaskaran. 1 Proof techniques Here we

More information

1.1.1 Algebraic Operations

1.1.1 Algebraic Operations 1.1.1 Algebraic Operations We need to learn how our basic algebraic operations interact. When confronted with many operations, we follow the order of operations: Parentheses Exponentials Multiplication

More information

ENUMERATION BY KERNEL POSITIONS

ENUMERATION BY KERNEL POSITIONS ENUMERATION BY KERNEL POSITIONS Abstract. We introduce a class of two-player games on posets with a rank function, in which each move of the winning strategy is unique. This allows to enumerate the kernel

More information

TEMPERATURE THEORY AND THE THERMOSTATIC STRATEGY

TEMPERATURE THEORY AND THE THERMOSTATIC STRATEGY TEMPERATURE THEORY AND THE THERMOSTATIC STRATEGY KAREN YE Abstract. In this paper, we differentiate between cold games, which are easier to analyze and play, and hot games, much more difficult in terms

More information

Review of Probability Theory

Review of Probability Theory Review of Probability Theory Arian Maleki and Tom Do Stanford University Probability theory is the study of uncertainty Through this class, we will be relying on concepts from probability theory for deriving

More information

Errata List for the Second Printing

Errata List for the Second Printing Errata List for the Second Printing The Second Revised Edition has had two printings, and it also exists on the web. The second printing can be distinguished from the first printing by the fact that on

More information

Generating Function Notes , Fall 2005, Prof. Peter Shor

Generating Function Notes , Fall 2005, Prof. Peter Shor Counting Change Generating Function Notes 80, Fall 00, Prof Peter Shor In this lecture, I m going to talk about generating functions We ve already seen an example of generating functions Recall when we

More information

EXAMPLES OF PROOFS BY INDUCTION

EXAMPLES OF PROOFS BY INDUCTION EXAMPLES OF PROOFS BY INDUCTION KEITH CONRAD 1. Introduction In this handout we illustrate proofs by induction from several areas of mathematics: linear algebra, polynomial algebra, and calculus. Becoming

More information

Introduction to Information Entropy Adapted from Papoulis (1991)

Introduction to Information Entropy Adapted from Papoulis (1991) Introduction to Information Entropy Adapted from Papoulis (1991) Federico Lombardo Papoulis, A., Probability, Random Variables and Stochastic Processes, 3rd edition, McGraw ill, 1991. 1 1. INTRODUCTION

More information

AUTOMORPHISM GROUPS AND SPECTRA OF CIRCULANT GRAPHS

AUTOMORPHISM GROUPS AND SPECTRA OF CIRCULANT GRAPHS AUTOMORPHISM GROUPS AND SPECTRA OF CIRCULANT GRAPHS MAX GOLDBERG Abstract. We explore ways to concisely describe circulant graphs, highly symmetric graphs with properties that are easier to generalize

More information

Roberto s Notes on Linear Algebra Chapter 10: Eigenvalues and diagonalization Section 3. Diagonal matrices

Roberto s Notes on Linear Algebra Chapter 10: Eigenvalues and diagonalization Section 3. Diagonal matrices Roberto s Notes on Linear Algebra Chapter 10: Eigenvalues and diagonalization Section 3 Diagonal matrices What you need to know already: Basic definition, properties and operations of matrix. What you

More information

Homework set 2 - Solutions

Homework set 2 - Solutions Homework set 2 - Solutions Math 495 Renato Feres Simulating a Markov chain in R Generating sample sequences of a finite state Markov chain. The following is a simple program for generating sample sequences

More information

Math 110 Linear Algebra Midterm 2 Review October 28, 2017

Math 110 Linear Algebra Midterm 2 Review October 28, 2017 Math 11 Linear Algebra Midterm Review October 8, 17 Material Material covered on the midterm includes: All lectures from Thursday, Sept. 1st to Tuesday, Oct. 4th Homeworks 9 to 17 Quizzes 5 to 9 Sections

More information

1 Proof techniques. CS 224W Linear Algebra, Probability, and Proof Techniques

1 Proof techniques. CS 224W Linear Algebra, Probability, and Proof Techniques 1 Proof techniques Here we will learn to prove universal mathematical statements, like the square of any odd number is odd. It s easy enough to show that this is true in specific cases for example, 3 2

More information

Toppling Conjectures

Toppling Conjectures Games of No Chance 4 MSRI Publications Volume 63, 2015 Toppling Conjectures ALEX FINK, RICHARD NOWAKOWSKI, AARON SIEGEL AND DAVID WOLFE Positions of the game of TOPPLING DOMINOES exhibit many familiar

More information

Inference for Stochastic Processes

Inference for Stochastic Processes Inference for Stochastic Processes Robert L. Wolpert Revised: June 19, 005 Introduction A stochastic process is a family {X t } of real-valued random variables, all defined on the same probability space

More information

Lecture 4 - Random walk, ruin problems and random processes

Lecture 4 - Random walk, ruin problems and random processes Lecture 4 - Random walk, ruin problems and random processes Jan Bouda FI MU April 19, 2009 Jan Bouda (FI MU) Lecture 4 - Random walk, ruin problems and random processesapril 19, 2009 1 / 30 Part I Random

More information

Introduction to Statistical Data Analysis Lecture 3: Probability Distributions

Introduction to Statistical Data Analysis Lecture 3: Probability Distributions Introduction to Statistical Data Analysis Lecture 3: Probability Distributions James V. Lambers Department of Mathematics The University of Southern Mississippi James V. Lambers Statistical Data Analysis

More information

Discrete Mathematics and Probability Theory Fall 2013 Vazirani Note 1

Discrete Mathematics and Probability Theory Fall 2013 Vazirani Note 1 CS 70 Discrete Mathematics and Probability Theory Fall 013 Vazirani Note 1 Induction Induction is a basic, powerful and widely used proof technique. It is one of the most common techniques for analyzing

More information

Diagonalization. MATH 322, Linear Algebra I. J. Robert Buchanan. Spring Department of Mathematics

Diagonalization. MATH 322, Linear Algebra I. J. Robert Buchanan. Spring Department of Mathematics Diagonalization MATH 322, Linear Algebra I J. Robert Buchanan Department of Mathematics Spring 2015 Motivation Today we consider two fundamental questions: Given an n n matrix A, does there exist a basis

More information

Compositions, Bijections, and Enumerations

Compositions, Bijections, and Enumerations Georgia Southern University Digital Commons@Georgia Southern Electronic Theses & Dissertations COGS- Jack N. Averitt College of Graduate Studies Fall 2012 Compositions, Bijections, and Enumerations Charles

More information

Chapter 5. Chapter 5 sections

Chapter 5. Chapter 5 sections 1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions

More information

COMBINATORIAL GAMES AND SURREAL NUMBERS

COMBINATORIAL GAMES AND SURREAL NUMBERS COMBINATORIAL GAMES AND SURREAL NUMBERS MICHAEL CRONIN Abstract. We begin by introducing the fundamental concepts behind combinatorial game theory, followed by developing operations and properties of games.

More information

INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING

INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING ERIC SHANG Abstract. This paper provides an introduction to Markov chains and their basic classifications and interesting properties. After establishing

More information

Generalized eigenvector - Wikipedia, the free encyclopedia

Generalized eigenvector - Wikipedia, the free encyclopedia 1 of 30 18/03/2013 20:00 Generalized eigenvector From Wikipedia, the free encyclopedia In linear algebra, for a matrix A, there may not always exist a full set of linearly independent eigenvectors that

More information

Further Mathematical Methods (Linear Algebra) 2002

Further Mathematical Methods (Linear Algebra) 2002 Further Mathematical Methods (Linear Algebra) 2002 Solutions For Problem Sheet 4 In this Problem Sheet, we revised how to find the eigenvalues and eigenvectors of a matrix and the circumstances under which

More information

Chapter 8: An Introduction to Probability and Statistics

Chapter 8: An Introduction to Probability and Statistics Course S3, 200 07 Chapter 8: An Introduction to Probability and Statistics This material is covered in the book: Erwin Kreyszig, Advanced Engineering Mathematics (9th edition) Chapter 24 (not including

More information

One Pile Nim with Arbitrary Move Function

One Pile Nim with Arbitrary Move Function One Pile Nim with Arbitrary Move Function by Arthur Holshouser and Harold Reiter Arthur Holshouser 3600 Bullard St. Charlotte, NC, USA, 28208 Harold Reiter Department of Mathematics UNC Charlotte Charlotte,

More information

A Collection of Dice Problems

A Collection of Dice Problems with solutions and useful appendices (a work continually in progress) version July 5, 208 doctormatt at madandmoonly dot com www.matthewconroy.com Thanks A number of people have sent corrections, comments

More information

Ahlswede Khachatrian Theorems: Weighted, Infinite, and Hamming

Ahlswede Khachatrian Theorems: Weighted, Infinite, and Hamming Ahlswede Khachatrian Theorems: Weighted, Infinite, and Hamming Yuval Filmus April 4, 2017 Abstract The seminal complete intersection theorem of Ahlswede and Khachatrian gives the maximum cardinality of

More information

CHAPTER 8: EXPLORING R

CHAPTER 8: EXPLORING R CHAPTER 8: EXPLORING R LECTURE NOTES FOR MATH 378 (CSUSM, SPRING 2009). WAYNE AITKEN In the previous chapter we discussed the need for a complete ordered field. The field Q is not complete, so we constructed

More information

Grade 8 Chapter 7: Rational and Irrational Numbers

Grade 8 Chapter 7: Rational and Irrational Numbers Grade 8 Chapter 7: Rational and Irrational Numbers In this chapter we first review the real line model for numbers, as discussed in Chapter 2 of seventh grade, by recalling how the integers and then the

More information

1 Continued Fractions

1 Continued Fractions Continued Fractions To start off the course, we consider a generalization of the Euclidean Algorithm which has ancient historical roots and yet still has relevance and applications today.. Continued Fraction

More information

Section 4.1: Sequences and Series

Section 4.1: Sequences and Series Section 4.1: Sequences and Series In this section, we shall introduce the idea of sequences and series as a necessary tool to develop the proof technique called mathematical induction. Most of the material

More information

16. . Proceeding similarly, we get a 2 = 52 1 = , a 3 = 53 1 = and a 4 = 54 1 = 125

16. . Proceeding similarly, we get a 2 = 52 1 = , a 3 = 53 1 = and a 4 = 54 1 = 125 . Sequences When we first introduced a function as a special type of relation in Section.3, we did not put any restrictions on the domain of the function. All we said was that the set of x-coordinates

More information

Number Theory: Niven Numbers, Factorial Triangle, and Erdos' Conjecture

Number Theory: Niven Numbers, Factorial Triangle, and Erdos' Conjecture Sacred Heart University DigitalCommons@SHU Mathematics Undergraduate Publications Mathematics -2018 Number Theory: Niven Numbers, Factorial Triangle, and Erdos' Conjecture Sarah Riccio Sacred Heart University,

More information

Section 11.1: Sequences

Section 11.1: Sequences Section 11.1: Sequences In this section, we shall study something of which is conceptually simple mathematically, but has far reaching results in so many different areas of mathematics - sequences. 1.

More information

Econ Slides from Lecture 7

Econ Slides from Lecture 7 Econ 205 Sobel Econ 205 - Slides from Lecture 7 Joel Sobel August 31, 2010 Linear Algebra: Main Theory A linear combination of a collection of vectors {x 1,..., x k } is a vector of the form k λ ix i for

More information

Calculus (Real Analysis I)

Calculus (Real Analysis I) Calculus (Real Analysis I) (MAT122β) Department of Mathematics University of Ruhuna A.W.L. Pubudu Thilan Department of Mathematics University of Ruhuna Calculus (Real Analysis I)(MAT122β) 1/172 Chapter

More information

4 Sums of Independent Random Variables

4 Sums of Independent Random Variables 4 Sums of Independent Random Variables Standing Assumptions: Assume throughout this section that (,F,P) is a fixed probability space and that X 1, X 2, X 3,... are independent real-valued random variables

More information

Lectures on Probability and Statistical Models

Lectures on Probability and Statistical Models Lectures on Probability and Statistical Models Phil Pollett Professor of Mathematics The University of Queensland c These materials can be used for any educational purpose provided they are are not altered

More information

q-series Michael Gri for the partition function, he developed the basic idea of the q-exponential. From

q-series Michael Gri for the partition function, he developed the basic idea of the q-exponential. From q-series Michael Gri th History and q-integers The idea of q-series has existed since at least Euler. In constructing the generating function for the partition function, he developed the basic idea of

More information

Exchangeable Bernoulli Random Variables And Bayes Postulate

Exchangeable Bernoulli Random Variables And Bayes Postulate Exchangeable Bernoulli Random Variables And Bayes Postulate Moulinath Banerjee and Thomas Richardson September 30, 20 Abstract We discuss the implications of Bayes postulate in the setting of exchangeable

More information

SIMPLE RANDOM WALKS: IMPROBABILITY OF PROFITABLE STOPPING

SIMPLE RANDOM WALKS: IMPROBABILITY OF PROFITABLE STOPPING SIMPLE RANDOM WALKS: IMPROBABILITY OF PROFITABLE STOPPING EMILY GENTLES Abstract. This paper introduces the basics of the simple random walk with a flair for the statistical approach. Applications in biology

More information

1 Basic Combinatorics

1 Basic Combinatorics 1 Basic Combinatorics 1.1 Sets and sequences Sets. A set is an unordered collection of distinct objects. The objects are called elements of the set. We use braces to denote a set, for example, the set

More information

arxiv: v1 [math.ho] 28 Jul 2017

arxiv: v1 [math.ho] 28 Jul 2017 Generalized Fibonacci Sequences and Binet-Fibonacci Curves arxiv:1707.09151v1 [math.ho] 8 Jul 017 Merve Özvatan and Oktay K. Pashaev Department of Mathematics Izmir Institute of Technology Izmir, 35430,

More information

Settling Time Reducibility Orderings

Settling Time Reducibility Orderings Settling Time Reducibility Orderings by Clinton Loo A thesis presented to the University of Waterloo in fulfillment of the thesis requirement for the degree of Master of Mathematics in Pure Mathematics

More information

ON THE POSSIBLE QUANTITIES OF FIBONACCI NUMBERS THAT OCCUR IN SOME TYPES OF INTERVALS

ON THE POSSIBLE QUANTITIES OF FIBONACCI NUMBERS THAT OCCUR IN SOME TYPES OF INTERVALS Acta Math. Univ. Comenianae Vol. LXXXVII, 2 (2018), pp. 291 299 291 ON THE POSSIBLE QUANTITIES OF FIBONACCI NUMBERS THAT OCCUR IN SOME TYPES OF INTERVALS B. FARHI Abstract. In this paper, we show that

More information

Bounded Tiling-Harmonic Functions on the Integer Lattice

Bounded Tiling-Harmonic Functions on the Integer Lattice Bounded Tiling-Harmonic Functions on the Integer Lattice Jacob Klegar Choate Rosemary Hall mentored by Prof. Sergiy Merenkov, CCNY-CUNY as part of the MIT PRIMES-USA research program January 24, 16 Abstract

More information

1.3 Convergence of Regular Markov Chains

1.3 Convergence of Regular Markov Chains Markov Chains and Random Walks on Graphs 3 Applying the same argument to A T, which has the same λ 0 as A, yields the row sum bounds Corollary 0 Let P 0 be the transition matrix of a regular Markov chain

More information

A = 3 1. We conclude that the algebraic multiplicity of the eigenvalues are both one, that is,

A = 3 1. We conclude that the algebraic multiplicity of the eigenvalues are both one, that is, 65 Diagonalizable Matrices It is useful to introduce few more concepts, that are common in the literature Definition 65 The characteristic polynomial of an n n matrix A is the function p(λ) det(a λi) Example

More information

REALIZING TOURNAMENTS AS MODELS FOR K-MAJORITY VOTING

REALIZING TOURNAMENTS AS MODELS FOR K-MAJORITY VOTING California State University, San Bernardino CSUSB ScholarWorks Electronic Theses, Projects, and Dissertations Office of Graduate Studies 6-016 REALIZING TOURNAMENTS AS MODELS FOR K-MAJORITY VOTING Gina

More information

Incompatibility Paradoxes

Incompatibility Paradoxes Chapter 22 Incompatibility Paradoxes 22.1 Simultaneous Values There is never any difficulty in supposing that a classical mechanical system possesses, at a particular instant of time, precise values of

More information

Notes on generating functions in automata theory

Notes on generating functions in automata theory Notes on generating functions in automata theory Benjamin Steinberg December 5, 2009 Contents Introduction: Calculus can count 2 Formal power series 5 3 Rational power series 9 3. Rational power series

More information

Sequences and infinite series

Sequences and infinite series Sequences and infinite series D. DeTurck University of Pennsylvania March 29, 208 D. DeTurck Math 04 002 208A: Sequence and series / 54 Sequences The lists of numbers you generate using a numerical method

More information

Chapter 4: An Introduction to Probability and Statistics

Chapter 4: An Introduction to Probability and Statistics Chapter 4: An Introduction to Probability and Statistics 4. Probability The simplest kinds of probabilities to understand are reflected in everyday ideas like these: (i) if you toss a coin, the probability

More information

On Projective Planes

On Projective Planes C-UPPSATS 2002:02 TFM, Mid Sweden University 851 70 Sundsvall Tel: 060-14 86 00 On Projective Planes 1 2 7 4 3 6 5 The Fano plane, the smallest projective plane. By Johan Kåhrström ii iii Abstract It was

More information

Sequences and the Binomial Theorem

Sequences and the Binomial Theorem Chapter 9 Sequences and the Binomial Theorem 9. Sequences When we first introduced a function as a special type of relation in Section.3, we did not put any restrictions on the domain of the function.

More information

Random variables and expectation

Random variables and expectation Department of Mathematics Ma 3/103 KC Border Introduction to Probability and Statistics Winter 2017 Lecture 5: Random variables and expectation Relevant textbook passages: Pitman [5]: Sections 3.1 3.2

More information

Math Circle: Recursion and Induction

Math Circle: Recursion and Induction Math Circle: Recursion and Induction Prof. Wickerhauser 1 Recursion What can we compute, using only simple formulas and rules that everyone can understand? 1. Let us use N to denote the set of counting

More information