Modeling Random Experiments

Similar documents
Random Experiments; Probability Spaces; Random Variables; Independence

Random variables and expectation

Random variables and expectation

the time it takes until a radioactive substance undergoes a decay

Lecture 8: Probability

Introduction to Probability. Ariel Yadin. Lecture 1. We begin with an example [this is known as Bertrand s paradox]. *** Nov.

Lecture 4: Probability and Discrete Random Variables

Module 1. Probability

ECE353: Probability and Random Processes. Lecture 2 - Set Theory

CIS 2033 Lecture 5, Fall

Axioms of Probability

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

1. When applied to an affected person, the test comes up positive in 90% of cases, and negative in 10% (these are called false negatives ).

Basic counting techniques. Periklis A. Papakonstantinou Rutgers Business School

Section 2: Classes of Sets

Measure and integration

LECTURE 1. 1 Introduction. 1.1 Sample spaces and events

Basic Probability. Introduction

Sample Spaces, Random Variables

POISSON PROCESSES 1. THE LAW OF SMALL NUMBERS

Discrete Mathematics and Probability Theory Fall 2014 Anant Sahai Note 15. Random Variables: Distributions, Independence, and Expectations

STA Module 4 Probability Concepts. Rev.F08 1

1 Normal Distribution.

God doesn t play dice. - Albert Einstein

Introducing the Normal Distribution

Introducing the Normal Distribution

Lecture 4: Probability, Proof Techniques, Method of Induction Lecturer: Lale Özkahya

Probability and the Second Law of Thermodynamics

Lecture Slides. Elementary Statistics Tenth Edition. by Mario F. Triola. and the Triola Statistics Series. Slide 1

1 The Basic Counting Principles

Lecture Slides. Elementary Statistics Eleventh Edition. by Mario F. Triola. and the Triola Statistics Series 4.1-1

Chapter 4: An Introduction to Probability and Statistics

Probability: Terminology and Examples Class 2, Jeremy Orloff and Jonathan Bloom

Lecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019

Randomized Algorithms

Fundamentals of Probability CE 311S

Probability Notes (A) , Fall 2010

PROBABILITY THEORY 1. Basics

Department of Mathematics

Probability Theory Review

Statistics 1 - Lecture Notes Chapter 1

Probability 1 (MATH 11300) lecture slides

STA111 - Lecture 1 Welcome to STA111! 1 What is the difference between Probability and Statistics?

Chapter 8: An Introduction to Probability and Statistics

Probability. Lecture Notes. Adolfo J. Rumbos

Statistical Theory 1

Stochastic Processes

Building Infinite Processes from Finite-Dimensional Distributions

Notes 1 Autumn Sample space, events. S is the number of elements in the set S.)

Probability Year 9. Terminology

Set theory background for probability

Discrete Mathematics and Probability Theory Fall 2013 Vazirani Note 12. Random Variables: Distribution and Expectation

MATH2206 Prob Stat/20.Jan Weekly Review 1-2

2 Measure Theory. 2.1 Measures

POL502: Foundations. Kosuke Imai Department of Politics, Princeton University. October 10, 2005

Notes Week 2 Chapter 3 Probability WEEK 2 page 1

Undergraduate Probability I. Class Notes for Engineering Students

Some Background Material

Chapter 1 Review of Equations and Inequalities

1.1. MEASURES AND INTEGRALS

Probability Year 10. Terminology

Lecture Notes 1 Basic Probability. Elements of Probability. Conditional probability. Sequential Calculation of Probability

Law of Trichotomy and Boundary Equations

Probability (Devore Chapter Two)

Econ 325: Introduction to Empirical Economics

Expectation is a positive linear operator

Properties of Probability

1 What is the area model for multiplication?

Probability and distributions. Francesco Corona

Probability: Sets, Sample Spaces, Events

Important Concepts Read Chapter 2. Experiments. Phenomena. Probability Models. Unpredictable in detail. Examples

CS 361: Probability & Statistics

Discrete Mathematics: Lectures 6 and 7 Sets, Relations, Functions and Counting Instructor: Arijit Bishnu Date: August 4 and 6, 2009

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 16. Random Variables: Distribution and Expectation

Chapter 14. From Randomness to Probability. Copyright 2012, 2008, 2005 Pearson Education, Inc.

MAT Mathematics in Today's World

Topic 5 Basics of Probability

Lecture 1. Chapter 1. (Part I) Material Covered in This Lecture: Chapter 1, Chapter 2 ( ). 1. What is Statistics?

Stats Probability Theory

Why should you care?? Intellectual curiosity. Gambling. Mathematically the same as the ESP decision problem we discussed in Week 4.

CHAPTER 1: Functions

Probability theory basics

UMASS AMHERST MATH 300 SP 05, F. HAJIR HOMEWORK 8: (EQUIVALENCE) RELATIONS AND PARTITIONS

Probability Experiments, Trials, Outcomes, Sample Spaces Example 1 Example 2

1 Basic Combinatorics

P (A) = P (B) = P (C) = P (D) =

RVs and their probability distributions

Probability COMP 245 STATISTICS. Dr N A Heard. 1 Sample Spaces and Events Sample Spaces Events Combinations of Events...

Problems from Probability and Statistical Inference (9th ed.) by Hogg, Tanis and Zimmerman.

18.175: Lecture 2 Extension theorems, random variables, distributions

P(A) = Definitions. Overview. P - denotes a probability. A, B, and C - denote specific events. P (A) - Chapter 3 Probability

Continuum Probability and Sets of Measure Zero

Discrete Mathematics and Probability Theory Fall 2012 Vazirani Note 14. Random Variables: Distribution and Expectation

CMPSCI 240: Reasoning about Uncertainty

5 Set Operations, Functions, and Counting

Elementary Discrete Probability

STAT 7032 Probability Spring Wlodek Bryc

2. Two binary operations (addition, denoted + and multiplication, denoted

Discrete Mathematics for CS Spring 2007 Luca Trevisan Lecture 27

2) There should be uncertainty as to which outcome will occur before the procedure takes place.

Transcription:

Department of Mathematics Ma 3/103 KC Border Introduction to Probability and Statistics Winter 2018 Lecture 2: Modeling Random Experiments Relevant textbook passages: Pitman [4]: Sections 1.3 1.4., pp. 26 46. Larsen Marx [3]: Sections 2.2 2.5, pp. 18 66. 2.1 Axioms for probability measures Recall from last time that a random experiment is an experiment that may be conducted under seemingly identical conditions, yet give different results. Coin tossing is everyone s go-to example of a random experiment. The way we model random experiments is through the use of probabilities. We start with the sample space S, the set of possible outcomes of the experiment, and consider events, which are subsets E of the sample sample space. (We let E denote the collection of events.) A probability measure P or probability distribution attaches to each event E a number between 0 and 1 (inclusive) so as to obey the following axioms of probability: Normalization: P ( ) = 0; and P (S) = 1. Nonnegativity: For each event E, we have P (E) 0. Additivity: If EF =, then P ( F ) = P (E) + P (F ). Note that while the domain of P is technically E, the set of events, that is P : E [0, 1], we may also refer to P as a probability (measure) on S, the set of samples. 2.1.1 Remark To reduce the visual clutter created by layers of delimiters in our notation, we may omit the some of them simply write something like P (f(s) = 1) or P {s S : f(s) = 1} instead of P ( {s S : f(s) = 1} ) and we may write P (s) instead of P ( {s} ). You will come to appreciate this. 2.1.1 Elementary Probability Identities There are some elementary properties of probability measures that follow in a straightforward way from the axioms. Here are a few together with their proofs (derivations, if you prefer). 1. P (E c ) = 1 P (E) Larsen Marx [3]: 2.3 Proof : Since E and E c are disjoint, and E E c = S, additivity gives us P (E) + P (E c ) = P (S). Normalization gives us P (S) = 1, and the desired expression is just a rearrangement. 2. If F E, then P (E \ F ) = P (E) P (F ) Proof : Recall the definition of E\F as the set of points in E that are not in F, that is EF c. Thus E = F (E \ F ) where F and E \ F are disjoint. By additivity then P (E) = P (F ) + P (E \ F ), and the desired expression is just a rearrangement. KC Border v. 2018.02.05::11.21

KC Border Random Experiments and Probability Spaces 2 2 3. Monotonicity: If F E, then P (F ) P (E) Proof : We have just shown that P (E) = P (F )+P (E\F ), and nonnegativity implies P (E\F ) 0. Therefore P (E) P (F ). 4. Finite additivity: If E 1,..., E n is a finite sequence of pairwise disjoint events, that is, i j = E i E j =, then ( n ) n P E i = P (E i ). Proof : We shall prove this by induction on n. Let P(n) stand for the above equality for n pairwise disjoint sets. Then P(2) is just Additivity. Assume P(n 1). Write n E i = n 1 E i E n }{{} =B Then BE n =, so P ( n E i ) = P (B E n ) = P (B) + P (E n ) by Additivity n 1 = P (E i ) + P (E n ) by P(n 1) = n P (E i ). The next result is obvious, but it can be derived from our axions. 2.1.2 Boole s Inequality Even if events E 1,..., E n are not pairwise disjoint, we may still conclude that ( n ) n P E i P (E i ). Before I demonstrate how to prove Boole s Inequality, let me describe a trick for disjunctifying a sequence of sets. 2.1.3 Lemma Let E 1,..., E n,... be a sequence (finite or infinite) of events. The there is a sequence A 1,..., A n,... of events such that: For each n, A n E n. The A i s are pairwise disjoint. That is, if i j, then A i A j =. For each n, The last fact implies that n A i = n E i. A i = E i. v. 2018.02.05::11.21 KC Border

KC Border Random Experiments and Probability Spaces 2 3 Proof : Define and for n > 1 define A 1 = E 1, A n = E n \ (E 1 E n 1 ). I leave it to you to prove that that the sequence A 1,..., A n,... has the desired properties. See Figure 2.1 for an example with three events. E 1 A 1 A 2 A 3 E 2 E 3 Figure 2.1. The E i events are the discs, and the three colored regions represent the events A 1 = E 1, A 2 = E 2 \ E 1, and A 3 = E 3 \ (E 1 E 2 ). Proof of Boole s Inequality: Let A i be a sequence of pairwise disjoint events satisfying the conclusion of Lemma 2.1.3. That is, each A i E i and n A i = n E i. Then P ( n ) ( n ) E i = P A i = n P (A i ) n P (E i ), where the second equality follows from the fact that the A i s are pairwise disjoint and finite additivity, and the inequality follows from P (A i ) P (E i ) (monotonicity) for each i. 2.1.2 Digression on terminology: Odds I find it exasperating that even generally linguistically reliable sources, such as The New York Times, confuse probabilities and odds. Pitman [4]: pp. 6 8 2.1.4 Definition The odds against the event E is the ratio P (E c ) P (E). The odds in favor of the event E is the ratio P (E) P (E c ). KC Border v. 2018.02.05::11.21

KC Border Random Experiments and Probability Spaces 2 4 That is, odds are a ratio of probabilities, not a probability. 1 It is customary to say something like the odds against the event E are P (E c )/P (E) to one, or as a ratio of integers. (That is, we typically say 3 to 2 instead of 1 1/2 to 1. ) Unfortunately, you often run across statements such as, the odds are one in ten that event E will happen, when the author probably means, the probability that E will happen is onetenth, so that the odds against E happening are nine to one. This is to be distinguished from the payoff odds. The payoff odds are the ratio of the amount won to the amount wagered for a simple bet. For instance, in roulette, if you bet $1 on the number 2 and 2 comes up, you get a payoff of $35, so the payoff odds are 35 to 1. But (assuming that all numbers on a roulette wheel are equally likely) the odds against 2 are 37 to 1 since a roulette wheel has the numbers 0 and 00 in addition to the numbers 1 through 36. 2 3 4 (Pitman [4, p. 7] describes the outcomes and bets for a roulette wheel.) 2.2 Additivity and the Inclusion Exclusion Principle Pitman [4]: p. 22 The Inclusion Exclusion Principle describes the full power of additivity of probability measures when applied to unions of not necessarily pairwise disjoint sets. Early on, we expect small children to understand the relation between sets and their cardinality If Alex has three apples and Blair has two apples, then how many apples do they have together? The implicit assumption is that the two sets of apples are disjoint (since they belong to different children), then the measure (count) of the union is the sum of the counts. But what if Alex and Blair own some of their apples in common? 2.2.1 Proposition (Inclusion Exclusion Principle, I) Even if AB, P (A B) = P (A) + P (B) P (AB). Proof : Now A B = (AB c ) (AB) (A c B). The three sets being unioned on the right-hand side are pairwise disjoint: Therefore, by finite additivity, Now also by additivity, (AB c )(AB) = (AB)(A c B) = (AB c )(A c B) =. P (A B) = P (AB c ) + P (AB) + P (A c B). P (A) = P (AB c ) + P (AB) P (B) = P (BA c ) + P (AB). 1 You may hove noticed that I used the singular verb is in the definition above and the plural verb are in the sentences preceding and following this note. The word odds is both plural and singular. When in doubt, you can always say the odds ratio is... Note to self: Check the OED. 2 Actually, there are (at least) two kinds of roulette wheels. In Las Vegas, roulette wheels have 0 and 00, but in Monte Carlo, the 00 is missing. 3 The term roulette wheel is a pleonasm, since roulette is French for little wheel. 4 The word pleonasm is one of my favorites. Look it up. v. 2018.02.05::11.21 KC Border

KC Border Random Experiments and Probability Spaces 2 5 So, adding and regrouping, P (A) + P (B) = P (AB c ) + P (AB) + P (BA c ) +P (AB) }{{} = P (A B) + P (AB). This implies P (A B) = P (A) + P (B) P (AB). Additionally, P (A B C) = P (A) + P (B) + P (C) P (AB) P (AC) P (BC) + P (ABC). To see this refer to Figure 2.2. The events A, B, and C are represented by the three circles. p 1 A p 6 p 7 p 4 p 3 C p 5 B p 2 Figure 2.2. Inclusion-Exclusion for three sets. The probability of each shaded region is designated by p i, i = 1,..., 7. Observe that P (A B C) = p 1 + p 2 + p 3 + p 4 + p 5 + p 6 + p 7 P (A) = p 1 + p 4 + p 6 + p 7 P (B) = p 2 + p 4 + p 5 + p 7 P (C) = p 3 + p 5 + p 6 + p 7 P (AB) = p 4 + p 7 P (AC) = p 6 + p 7 P (BC) = p 5 + p 7 P (ABC) = p 7. KC Border v. 2018.02.05::11.21

KC Border Random Experiments and Probability Spaces 2 6 Thus So P (A) + P (B) + P (C) = p 1 + p 2 + p 3 + 2p 4 + 2p 5 + 2p 6 + 3p 7 P (AB) + P (AC) + P (BC) = p 4 + p 5 + p 6 + 3p 7. [ P (A) + P (B) + P (C) ] [ P (AB) + P (AC) + P (BC) ] + P (ABC) = p 1 + p 2 + p 3 + p 4 + p 5 + p 6 + p 7 = P (A B C). The general version of the Inclusion Exclusion Principle may be found in Pitman [4], Exercise 1.3.12, p. 31. 2.2.2 Proposition (General Inclusion Exclusion Principle) ( n ) P E i = P (E i ) i i<j P (E i E j ) +. i<j<k P (E i E j E k ) + ( 1) n+1 P (E 1 E 2... E n ). (Recall that intersection is denoted by placing sets next to each other. Note that the sign preceding a sum with the intersection of m sets is ( 1) m+1. The reason for summing over increasing indices is to avoid double counting.) Note that if the sets are pairwise disjoint, the intersections above are all empty and so have probability zero, and this reduces to finite additivity. While it is possible to prove this result now using induction, I will put off a proof until we learn about the expectation of random variables, which will make the proof much easier. 2.3 Countable additivity Most probabilists assume further that E is a σ-algebra or σ-field, which requires in addition that 3. If E 1, E 2,... belong to E, then E i and E i belong to E. Note that if S is finite and E is an algebra then, it is automatically a σ-algebra. Why? Since there are only finitely many subsets of a finite set, any infinite sequence of sets has only finite may distinct sets, the rest are all copies of something else. So any infinite intersection or union is actually the same as some finite intersection or union. Most probabilists also require the following stronger property, called countable additivity: v. 2018.02.05::11.21 KC Border

KC Border Random Experiments and Probability Spaces 2 7 ( ) Countable additivity P E i = P (E i) provided E i E j = for i j. 2.3.1 Remark If the sample space S is finite, it has only finitely many subsets, so the only way an infinite sequence E 1, E 2,... can be pairwise disjoint, is if all but finitely many of the events E i are equal to the empty set. In this case, since P ( ) = 0, the infinite series P (E i) reduces to a finite sum. In other words, for a finite sample space, finite additivity guarantees countable additivity. (Cf. Section 2.1.1, item 4.) You need to take an advanced analysis course to understand that for infinite sample spaces, there can be probability measures that are additive, but not countably additive. So don t worry too much about it. The next results may seem theoretical and of no practical relevance, but they are crucial to understanding the properties of cumulative distribution functions. A sequence E 1,..., E n... of events is decreasing, written E n, if E 1 E 2 E n. A sequence E 1,..., E n... of events is increasing, written E n, if E 1 E 2 E n. Add the proof. 2.3.2 Proposition (Continuity and countable additivity) If P is an additive probability, then ( ) 1. P is countably additive if and only if E n implies P E n = lim n P (E n ). n ( ) 2. P is countably additive if and only if E n implies P E n = lim n P (E n ). n 2.4 Random variables and random vectors For some reason, your textbooks postpone the definition of random variables, even though they are fundamental concepts. A random variable is a numerical measurement of the outcome of a random experiment. 2.4.1 Example (Some random variables) Here are some examples of random variables. The random experiment is to roll two dice, so the sample space is the set of ordered pairs of integers from 1 through 6. The sum of these to numbers is a random variable. The pair itself is a random vector. The difference of the numbers is another random variable. The experiment is to roll two dice repeatedly until boxcars (a sum of twelve) appear. The number of rolls is a random variable, which may take on the value if boxcars never appear. (This is an idealization of course.) The random experiment takes a sample of blood and smears it on a microscope slide with a little rectangle marked on it. The number of platelets lying in the rectangle is a random variable. The experiment is to record all the earthquakes in Southern California. The number of magnitude 5+ earthquakes in a year is a random variable. KC Border v. 2018.02.05::11.21

KC Border Random Experiments and Probability Spaces 2 8 A letter is drawn at random from the alphabet. If we assign a number to each letter, then that number is a random variable. But unlike the cases above it doe not make sense to take the results of two such experiments and average them. What is the average of a and b? In such cases, where the result of the experiment is categorical and not inherently numeric, it may make more sense to take the outcome to be a random vector, indexed by the categories. This interpretation is often used in communication theory by electrical engineers, e.g., Robert Gray [2] or Thomas Cover and Joy Thomas [1]. An experiment by Rutherford, Chadwick, and Ellis counted the number of α-particles emitted by a radioactive sample for consecutive 7.5-second time intervals. Each count is a random variable. Being numerical, we can add random variables, take ratios, etc., to get new random variables. But to understand how these are related we have to go back to our formal model of random experiments as probability spaces, and define random variables in terms of a probability space. 2.4.2 Definition A random variable on a probability space (S, E, P ) is an (extended) a real-valued function on S which has the property that for every interval I R the inverse image of I is an event. b A random vector is simply a finite-dimensional vector (ordered list) of random variables. a The extended real numbers include two additional symbols, and. We ll have more to say about them later. b Note that when the collection E of events consists of all subsets of S, then the requirement that inverse images of intervals be events is automatically satisfied. So a random variable is not a variable in the usual sense of the word variable in mathematics. A random variable is simply an (extended) real-valued function. Traditionally, probabilists and statisticians use upper-case Latin letters near the end of the alphabet to denote random variables. This has confused generations of students, who have trouble thinking of random variables as functions. For the sake of tradition, and so that you get used to it, we follow suit. So a random variable X is a function X : S R such that for each interval I, {s S : X(s) I} E. We shall adopt the following notational convention, which I refer to as statistician s notation, that (X I) means {s S : X(s) I}. Likewise (X t) means {s S : X(s) t}, etc. If E belongs to E, then its indicator function 1 E, defined by { 0 s / E 1 E (s) = 1 s E, is a random variable. v. 2018.02.05::11.21 KC Border

KC Border Random Experiments and Probability Spaces 2 9 A random variable X, is a mapping from the sample space S to the real numbers, that is, X maps each point s S to a real number X(s). The function X is different from its value X(s) at the point s, which is simply a real number. The value X(s) is frequently referred to as a realization of the random variable X. A realization is just the value that X takes on for some outcome s in the sample space. 2.4.1 Distribution of a random variable A random variable X on the probability space (S, E, P ) induces a probability measure or distribution on the real line as follows. Given an interval I, we define P X (I) = P ({s S : X(s) I}). Similarly, a random m-vector X = (X 1,..., X m ) induces a probability measure or distribution on R m by P X (I 1 I m ) = P ({s S : X i (s) I i, i = 1,..., m}). We shall discuss this further in Section 5.3. 2.5 Independent events 2.5.1 Definition Events E and F are (stochastically) independent if P (EF ) = P (E) P (F ). Larsen Marx [3]: 2.5 Pitman [4]: 1.4 Whether or not two events are independent is determined by the probability measure, but note that for ane event F, we have that F and S are independent and also that F and are independent. 2.5.2 Lemma If E and F are independent, then E and F c are independent; and E c and F c are independent; and E c and F are independent. Proof : It suffices to prove that if E and F are independent, then E c and F are independent. The other conclusions follow by symmetry. So write so by additivity F = (EF ) (E c F ), P (F ) = P (EF ) + P (E c F ) = P (E)P (F ) + P (E c F ), where the second equality follows from the independence of E and F. Now solve for P (E c F ) to get P (E c F ) = ( 1 P (E) ) P (F ) = P (E c )P (F ). But this is just the definition of independence of E c and F. KC Border v. 2018.02.05::11.21

KC Border Random Experiments and Probability Spaces 2 10 2.6 Repeated experiments and product spaces One of the chief uses of the theory of probability is to understand long-run frequencies of outcomes of experiments. If S is the sample space of a random experiment and E is its set of events, and we repeat the experiment, we have a compound experiment or joint experiment whose sample space is the Cartesian product S 2 = S S, the set of ordered pairs of outcomes. Similarly, the sample space for n repetitions of the experiments is S n = S S. The }{{} n copies of S individual experiments in such a sequence are called trials. The set of events for the compound experiment should certainly include sets of the form set of the form E = E 1 E n, where each E i is an event in E, the common set of events for a single experiment is an event for the repeated experiment. Such a set is called a rectangle. (Geometrical rectangles are products of intervals.) But there are non-rectangular event that the outcomes are the same in each trial, {(s 1,..., s n ) : s 1 = s 2 = = s n }. We want the set of events in the compound experiment to be a algebra. The smallest algebra that includes all the rectangles is called the product algebra and is denoted E n. For example consider rolling a die, where every subset of {1,..., 6} of is an event in E. The sample space for the repeated experiment is the set of ordered pairs of the numbers one through six. The event both tosses give the same result is not a rectangular event, but it is the union of finitely many rectangles: = 6 ( ) {i} {i}, and so is an event in the product algebra. So is the complementary event, the rolls are different. 2.7 Independent repeated experiments Add a picture, and some examples. Our mathematical model of a random experiment is a probability space (S, E, P ). And of the repeated experiment is (S 2, E 2,?). The question mark is there because we need to decide the probabilities of events in a repeated experiment. To do this in a simple way we shall consider the case where the experiments are independent. That is, the outcome of the first experiment provides no information about the outcome of the second experiment. Consider a compound event E 1 E 2, which means that the outcome of the first experiment was in E 1 E and the outcome of the second experiment was in E 2 E. The event E 1 in the first experiment is really the event E 1 S in the compound experiment. That is, it is the set of all ordered pairs where the first coordinate belongs to E 1. Similarly the event E 2 in the second experiment corresponds to the event S E 2 in the compound experiment. Now observe that (E 1 S) (S E 2 ) = E 1 E 2. Since the experiments are independent the probability of the intersection (E 1 S)(S E 2 ) should be the probability of (E 1 S) times the probability of (S E 2 ). But these probabilities are just P (E 1 ) and P (E 2 ) respectively. Thus for independently repeated experiments and rectangular events, Prob(E 1 E 2 ) = P (E 1 ) P (E 2 ). This is enough to pin down the probability of all the events in the product algebra E 2, and the resulting probability measure is called the product probability, and may be denoted by P P, or P 2, or by really abusing notation, simply P again. The point to remember is that independent experiments give rise to products of probabilities. How do we know when two experiments are independent? We rely on our knowledge of physics or biology or whatever to tell us that the outcome of one experiment yields no information on v. 2018.02.05::11.21 KC Border

KC Border Random Experiments and Probability Spaces 2 11 the outcome of the other. It s built into our modeling decision. I am no expert, but my understanding is that quantum entanglement implies that experiments that our intuition tells are independent are not really independent. 5 But that is an exceptional case. For coin tossing, die rolling, roulette spinning, etc., independence is probably a good modeling choice. As an example of experiments that are not independent, consider testing potentially fatal drugs on lab rats, with the same set of rats. If a rat dies in the first experiment, it diminishes the probability he survives in the second. 2.8 A digression on infinity Now consider an experiment with a stopping rule. For example, consider the experiment, toss a coin until Heads occurs, then stop. What is the natural sample space, and set of events for this experiment. You might think the simplest sample space for this experiment is the set of sequences of finite length. For a given natural number n, the space of all sequences of length n is the n-fold Cartesian product S n = S S, so the grand sample space is }{{} n copies S = n=1 S n. This sample space is infinite, but at least is a nice infinity it is countably infinite. The event H n that the first Head occurs on the n th toss belongs E n, and so it should be an event in the larger experiment. Now consider the event H = (a Head eventually occurs). The event H is the infinite union n=1h n. Is this union an event as we have defined things? No, it is not. One way to see this is to ask, what is the complement of H? It would be the event that no Head occurs, so we would have to toss forever. But the infinite sequence of all Tails (while admittedly a probability zero occurrence) does not appear in our sample space. Another way to say this is that E n is not a σ-algebra. So if we want the set of events to include H, n=1 we need to do something drastic. One possibility is never to consider H to be an event. After all, how could we ever observe such an event happening? In other words, we could say that ensuring that the set of events is a σ-algebra instead of merely an algebra is not worth the trouble. (Your textbooks simply ignore this difficulty, and they are still full of useful results.) On the other hand, we might really care about the probability of the event H. If we want to do that, we have to agree that the real sample space is actually the set of all infinite sequences of outcomes of the original experiment, that is, the infinite Cartesian product Ω = S. Hopefully, you remember from Ma 1a that S is a countable set while Ω is uncountable if S has at least two elements. (Think of binary expansions of real numbers in the unit interval.) This makes Ω a more difficult object to work with. I ll have more to say about this when we discuss the Strong and Weak Law of Large Numbers. Each approach has its advantages and disadvantages. I should discuss some of these issues in an appendix for the mathematically inclined, and shall when I can find the time. Fortunately, there are still plenty of interesting things we can say without having to worry about making H an event. Bibliography [1] T. M. Cover and J. A. Thomas. 2006. Elements of information theory, 2d. ed. Hoboken, New Jersey: Wiley Interscience. 5 I asked David Politzer if this was a fair statement, and he gave his blessing. KC Border v. 2018.02.05::11.21

KC Border Random Experiments and Probability Spaces 2 12 [2] R. M. Gray. 1988. Probability, random processes, and ergodic properties. New York: Springer Verlag. [3] R. J. Larsen and M. L. Marx. 2012. An introduction to mathematical statistics and its applications, fifth ed. Boston: Prentice Hall. [4] J. Pitman. 1993. Probability. Springer Texts in Statistics. New York, Berlin, and Heidelberg: Springer. v. 2018.02.05::11.21 KC Border