On a horizontal number line, list all possible values of x; in this case: Chapter 3. Section 3.1: The probability histogram

Size: px
Start display at page:

Download "On a horizontal number line, list all possible values of x; in this case: Chapter 3. Section 3.1: The probability histogram"

Transcription

1 Chapter 3 Section 3.1: The probability histogram Below is the probability histogram for the IS (p. 80 of text): On a horizontal number line, list all possible values of x; in this case: 0.90, 0.70, 0.50, 0.30, 0.10, 0.10, 0.30, 0.50, 0.70, Draw rectangles: Each x serves as the center of the base of its rectangle The base of each rectangle equals δ (this means the rectangles touch, but do not overlap). The height of the rectangle centered at x is: P(X = x)/δ How is it determined? Below is the verification of two heights given in our picture: P(X = 0.10)/0.20 = /0.20 = P(X = 0.30)/0.20 = /0.20 = Why this strange definition of height? B/c we want to consider areas. The area of the triangle centered at x is, of course, its base times its height: δ[p(x = x)/δ] = P(X = x). Here is the main thing to remember: In a probability histogram, the area of a rectangle equals the probability of its center value. Once we know this, we can see the symmetry in the sampling distribution for the IS. We are now going to add another adjective (remember actual?). The sampling distribution of Chapter 2 will be called the exact sampling distribution and it yields the exact P-value. We say this b/c in Chapter 3 we will learn two aways to approximate a sampling distribution: computer simulation and fancy math Section 3.2: Computer simulation In the text, I talk about a Colloquium Study (CQS). Its data are below: Treat. S F Total Total It can be shown that there are 40,116,600 possible assignments, and that 12,932,920 of these give x = 0 (a = 4). Thus, P(X = 0) =12,932,920/40,116,600= The idea of the computer simulation approximation is quite simple: Perhaps we can obtain a good approximation to any probability by looking at some, not all of the assignments. + 84

2 For example, I looked at 10,000 assignments for the CQS and discovered that 3267 of them gave a = 4 and x = 0. Thus, the relative frequency (RF) of occurrence of 0 is , which is very close to its probability, But, I am ahead of myself. Once we decide to look at only some of the assignments, two questions arise. 1. How many should we look at? The answer is called the number of runs of the computer simulation. As we shall see, 10,000 is a good choice for the number of runs. 2. Which ones should we look at? Well, to avoid bias, we select assignments at random. If you want to see more details on this, read pages 84 and 85 in the text. But you don t need to understand this Below are the results of my computer simulation study with 10,000 runs for the CQS: x RF Probability 4/ / / / / / / / Total First, note that the RFs and probabilities are close. (Remember Section 2.4) We can use the RFs to approximate the P- value. The ingredients: the actual x = 3/7, the alternative was > Thus, the exact P-value is P(X 3/7) = = We can approximate this by RF (X 3/7) = = The picture below is taken from p. 87 of the text Runs 2/7 1/7 4/ Runs Results of a simulation experiment with 10,000 runs for the Ballerina study: Rel. Freq. Rel. Freq. Rel. Freq. x of x of x of x /7 0 3/7 10,000 Runs Recall, the P-value is P(X 0.24) = We approximate this with RF(X 0.24) = /7 1/7 2/

3 Results of a simulation experiment with 10,000 runs for the Crohn s study: Rel. Freq. Rel. Freq. Rel. Freq. x of x of x of x Recall, the P-value = P(X 0.27) = , which we approximate by RF(X 0.27) = Section 3.3: Center and Spread Page 89 of the text shows probability histograms for four studies in the text. In each picture, if you sum the areas of the teal colored rectangles, you get the P-value. There are two facts (for all FTs) that are revealed in these pictures. Each picture has one central peak. The peak can be one or two rectangles wide, but never three. As you move away from the central peak in either direction, the rectangles become shorter and shorter. It is useful to have a concept of a left to right center for such a picture Clearly, if the picture is symmetric, then its center is 0. To include all pictures, we define the center of the PH to be its center of gravity. It can be shown that for every FT, the center of gravity is 0. In general, the center of gravity, or mean, of a PH is denoted by the Greek letter mu: µ. For FT, µ = 0. (In Chapter 5, we will have pictures for which µ 0.) Thus, all FTs are similar in that they all have µ = 0. But the pictures on page 89 look very different. This is b/c they have different amounts of spread. For the four pictures, IS has the most spread, then CQS, then Soccer and finally CCD has the least spread. (This is an obvious visual assessment.) + 91 We need more than a visual assessment of spread. We need a number that summarizes the spread in a PH. The number is the standard deviation of the PH, denoted by the Greek letter sigma: σ. There is a simple formula for calculating σ: m1 m σ = 2 n 1 n 2 (n 1). Below are the standard deviations for the four pictures on page 89. (See text for details.) Study: IS CQS Soccer CCD σ: Note that more spread corresponds to a larger σ. Why do we want to measure spread? Be patient please. + 92

4 Recall that X is the test statistic for FT. It is also called a random variable. Let X be any random variable, with mean µ and standard deviation σ. Define the standardized version of X to be Z, where Z = X µ. σ Transforming X to Z is called standardizing X. The observed value of Z is denoted by z and is computed by: z = (x µ)/σ. For FT, b/c µ = 0: Z = X/σ and z = x/σ. Data X x Z z + 93 In Chapter 2, we learned about the sampling distribution for X. In a similar way, Z has a sampling distribution. Assuming we have the sampling distribution for X, it is very easy to get the sampling distribution for Z. We will illustrate with the IS. Recall for the IS: σ = and the possible values of x are: 0.90, ,0.90. Thus, z = x/ and we get all possible values of z by taking every possible value of x and dividing it by Namely, 0.90/ = 3.94, 0.70/ = / = The probabilities for the z s are automatic. For example, the event that z = 3.94 is the same event as x = 0.90, so they have the same probability. To summarize, given the sampling distribution for X, it is easy to get the sampling distribution for Z. You don t need to worry about reproducing these details; this result is simply to motivate what happens next. B/c Z has a sampling distribution, we can draw its probability histogram and I have done so on page 94 of the text. (Ignore the teal color and smooth curve on p. 94.) Compare the pictures for the IS on pages 89 and 94 of the text: These pictures have the same shape and both are centered at 0. They differ in their spreads I repeated the above procedure (standardizing) for the other three probability histograms on p. 89. The details will not be given. The pictures I get for the three Z s are on pages 95 and 96. As with the IS, if you compare each picture for X on page 89 with its picture for Z, you will find that the two pictures have the same shape and both are centered at 0, but they have different spreads. Let s go back to the picture for Z for the IS on page 94. B/c Z is equivalent to X, we may use X or Z to find the P-value. To make this precise: P(X 0.30) = P(Z 1.31) and P(X 0.30) = P(Z 1.31). + 96

5 Thus, the area of the teal-shaded rectangles on page 94 is the P-value for the IS. Similarly, the area of the teal-shaded rectangles is the P-value for the three pictures on pages 95 and 96. Now, look at the four pictures on pages They look very similar. Why? B/c: the mean of each picture is 0 and the standard deviation of each picture is 1. This is what standardizing does to a picture: It creates a new picture which has the same shape as the old picture. In addition, the new picture is centered at 0 and is scaled to have a standard deviation of 1. Working with X, different data sets give very different pictures (see page 89), but working with Z, different data sets give very similar pictures. But why do we desire similar pictures? This is where the smooth curve enters the argument. B/c the Z pictures are similar, we can use one curve to as an approximation to each of them. A B C 1.31 Exact Area = B + C Approximate Area = A + B + 97 If A = C, then the approx. is perfect The smooth curve is called the standard normal curve (snc). The approximation method motivated above is useful only if it is easy to find areas under the snc Two facts: The snc is symmetric around 0 and its total area is 1. Suppose that I want to find the area under the snc to the right of, say, z = Good news: The table in the front of the book is designed to answer this question. First, take the z and divide it into two pieces: 1.3 and Then go to the table as shown below. z

6 For another example, suppose we want the area under the snc to the right of z = Break z into 1.5 and z Suppose we want the area under the snc to the right of z = The difficulty is: What to do with the extra digits? In my class, just round-off the z to 1.47 and proceed as above. Finally, suppose we want the area under the snc to the right of z = The difficulty is that the table goes no higher than But read the 3.5 row in the table (does it remind you of the movie The Shining? ). Fact: For any z > 3.59, the area under the snc to the right of z is What is the area under the snc to the left of z = 1.27? Thinking like the MITK, use symmetry to realize that the area to the left of z = 1.27 equals the area to the right of z = +1.27, and we know how to find this latter area (it is ). Fact: The area under the snc to the left of any z equals the area under the snc to the right of z. Now we are ready to use the snc to obtain an approximate P-value for FT. Consider again the CCD. Recall that the exact P-value is and, symbolically, is P(X ). Also, recall that for the CCD, σ = Thus, P(X ) = P(X/σ /0.1193) = P(Z 2.27) Look at the picture on p. 96 of the text. This P-value is the area of the rectangle centered at 2.27 plus the areas of all the rectangles to the right of this rectangle. (Only one of these rectangles can be seen on p. 96). It can be shown (more on this later) that the rectangle centered at 2.27 has for its endpoints: 2.04 and 2.51 (these look funny b/c of round-off error). The snc approximation to this P-value should be the area under the snc to the right of 2.04, which is This is a pretty good approximation, almost as good as we obtained with computer simulation (that was ). BUT now I must tell you something strange about the rest of the world, outside this classroom. Every other book I have seen says, It s too much work to find the 2.04 above, let s just calculate the area to the right of z = 2.27! I am NOT making this up! The area to the right of 2.27 is , which is a horrible approximation of To summarize, every other book I have seen, after spending all of Chapter 2 convincing you that the P-value is important, gets to Chapter 3 and says, It is too much trouble to approximate it accurately. For now, let s focus on what all the lazy people do. Whereas the method gives a lousy answer, it is very easy to use. 1. Calculate z = x/σ. 2. For the alternative >: The approximate P-value is the area under the snc to the right of z

7 For the alternative < the exact P-value is P(X x) = P(Z z), where z = x/σ. Thus, the lazy approximation is the area under the snc to the left of z, which is equal to the area under the snc to the right of z. Finally, for the alternative, we know that the P-value has two pieces to it. The approximation is: Twice the area under the snc to the right of z. This three-part rule (one for each alternative) is presented on p. 102 of the text. Its main advantage is that we calculate the same z regardless of alternative and then look up either z, z or z in the table. Remembering to double the area we find for the alternative. Its main disadvantage is that it often gives horribly inaccurate approximations Of course, nobody else calls this the lazy method. They call it the method. Mostly, they don t acknowledge the existence of an exact P-value, thus avoiding the issue that their approximation is bad. If anything, they call the lazy method the approximation without the continuity correction and my better method the approximation with the continuity correction. There is an additional bit of perversity going on: the approximation with the continuity correction given in all books that give it is actually a very bad correction and gives answers barely different from the without. This helps them discount people like me who want to improve on their answers: by presenting an ineffective improvement, they send the message that improvements are not needed! In their defense it is true that for very large studies the continuity correction does not change the answer much Thus, one can reasonably take the following approach: For small studies use the website to get the exact P-value and for larger studies if you use the snc, it does not matter much whether you use the continuity correction. (BTW, I am convinced that after some threshold for being a large study the website that we use actually uses an snc approximation.) When I have managed to challenge, in person, somebody who teaches the lazy method they always say, Oh, we only use the snc when the study is too large to get the exact P-value. But they taught the lazy method before there were computers readily available to obtain the exact answer. Now I am going to, perhaps, surprise you. On our homework and especially on the midterm, I encourage you to use the lazy method. (Discuss.) If you submit project 1 (A or B) you will need to use the approximation with the continuity correction. Its steps are outlined on page 3 of the Course Notes on the course webpage. (Discuss projects briefly.) Consider BSS. Recall that x = 0.24, the alternative is <, the exact P-value is Also, σ = 38(12) 25(25)(49) = Following page 3 of the Course Notes, g = δ/2 = 50/[2(25)(25)] = Thus, x 2 = x + g = = 0.20, which standardized gives z 2 = 0.20/ = We look up z 2 = 1.64 and find that the approximate P-value is W/o the c/c, z = 0.24/ = 1.97, giving an approximate P-value of

8 Let s look at CCD again. As I mentioned earlier in lecture, b/c of the nasty numbers for n s we need to be more precise in our computations. In particular, ˆp 1 = 22/37 = , ˆp 2 = 11/34 = , x = = , g = δ/2 = 71/[2(37)(34)] = Thus, x 1 = x g = = and z 1 = / = 2.04 Finally, consider IS as an example of. Recall that δ = 0.20, so g = First, we compare x to g. If x g then the exact P-value is 1 and no approximation is needed. In this case, x = 0.30, so we must continue. Next, x 3 = x g = = As stated earlier, σ = , making z 3 = 0.20/ = The area to the right of 0.88 is ; doubling this, we get as the approximation to the exact W/o the c/c, z = 0.30/ = 1.31 giving an approximation of 2(0.0951) = One somewhat positive comment about the lazy method. If all the margins are really large, w/ and w/o the c/c give about the same answers. For example, take n 1 = n 2 = 5000, m 1 = 4000 and x = One can verify σ = and g = With the c/c, z = / = 2.03 and w/o it, z = 0.02/ = But this is a huge study! Chapter 5 We spent Chapters 2 and 3 examining the Skeptic s argument. The Skeptic makes no attempt to (formally) extend conclusions beyond the subjects in the study. For example, we concluded that cyclosporine was superior to placebo for the 71 people in the study. We concluded that Julie was better spinning right than left for the 50 trials in her study. Typically, for better or worse, researchers want to extend their conclusions to a larger situation. There are many techniques for such extensions, but central to every one is the notion of a population. For our purposes, there are two types of populations: finite and infinite. And the word infinite here means it is not what we call a finite population. (?) A finite population is a well-defined collection of individuals. Examples: All persons who will vote in this year s presidential election; all persons eligible to vote in this year s presidential election; all persons in this room; all persons enrolled for one or more credits this semester at UW- Madison We need a way to think about a finite population. Imagine a box of cards, called the population box. Each member of the population has a card in the box. On the member s card are the values of one or more variables, or features, of the member. (Same features for all members.) For simplicity, we begin with one dichotomous feature per card. As in Chapter 1, the possible values of the feature are labeled success and failure. A 1 on a card denotes that that member is a success for the feature, and a 0 denotes a failure. Thus, every card has a 1 or a 0 on it. Statisticians paraphrase Snoopy who once famously said: I love mankind, it s people I can t stand! That is, statisticians are interested in the box in totality, not in any particular member s card

9 For a given population box: Let s denote the number of cards in the box marked 1. Let f denote the number of cards in the box marked 0. Let N = s+f denote the total number of cards in the box. Let p = s/n denote the proportion of cards in the box marked 1. Let q = f/n denote the proportion of cards in the box marked These five numbers, s, f, N, p and q tell us what is in the box. And, of course, knowledge of two of these numbers, N and p, allow us to determine the others. As a result, I will describe a box as Box(N; p). For example, Box(10;0.60) is a box with N = 10 cards, of which a proportion p = 0.60 of them are successes. In practice, a researcher does not know p and often does not even know N. There are two ways for a researcher to learn about what is in the population box. First, is a census; this means one examines every card in the box. Discuss. Second, is a survey; this means one examines only some of the cards in the box. Discuss. We will focus on surveys. The cards actually examined in a survey comprise the sample A sample is called representative if it looks like the box. Every (honest) researcher wants a representative sample, but, alas, there is no way to guarantee getting one. Let s consider this idea of representative again. Suppose that two researchers, A and B, each select samples of size n = 5 from Box(N;0.60). Below are their samples: A: 1, 1, 0, 1, 1 B: 0, 1, 1, 0, 1. Which sample is representative? B s b/c its ˆp = p. Discuss Bill Clinton s cabinet. We cannot guarantee a representative sample, so we advocate selecting what is called a random sample. A random sample (much like randomization) is a process and does not guarantee a sample that is representative, or necessarily even close to representative Its great virture is that a random sample allows us to calculate the probability that we will obtain a sample that is close to representative. As stated earlier, in practice, a researcher does not know p and often does not even know N. But, for now, let s assume that we know both of these numbers. Consider the chance mechanism of selecting n cards at random from Box(N; p). Imagine that we select the cards one-at-atime. But once we think of selecting the cards oneby-one, two ways of sampling come to mind: Without replacement (smart), and With replacement (dumb). Why do I label these smart and dumb? Discuss

10 Two probability histograms for X, the number of successes in a sample of size n = 10 from Box(1000;0.60) Solid [Dashed] rectangles are for a random sample with [without] replacement As above, but from Box(20;0.60) Note that in the first of these pictures, probabilities are the same for both methods of sampling, but for the second picture the probabilities are quite different. Also, note that the probabilities are better for the smarter method of sampling. Finally, note that for the dumb way of sampling, the probabilities do not depend on the value of N. The key is the ratio, n/n of sample size to population size. In the first example, this ratio is 10/1000 = 0.01 and in the second example, 10/20 = The general guideline is: If n/n 0.05, then probabilities calculated w/replacement are approximately equal to probabilities calculated w/o/replacement Why does any of this matter? Well, it turns out that it is much easier, both computationally and theoretically, to work with the probabilities for the dumb way of sampling. Thus, regardless of how you select a random sample, provided n/n 0.05, it is valid to calculate probabilities the easy way. Extended enrichment example: It is sometimes better to sample the dumb way. Recall the CCD study. There are N = possible assignments of subjects to treatments. Let each possible assignment be a pop. member and as a group they form our (very very large) finite population. Define an assignment to be a success if it would yield x (Remember, the actual x = 0.27.) Thus, the p for this population box equals , the P-value for the FT. Our computer simulation experiment selected a sample of n =10,000 assignments in the dumb way. But it would be very difficult to to write a computer program to select assignments the smart way (and it would require lots of memory and would run slowly). And, as our above result shows, the smart way and dumb way of sampling give the same answers. We will now investigate computing probabilities for the dumb way of sampling (with replacement). Recall, that we plan to (probabilities are always about the future) select n cards at random with replacement from the population box. Define X 1 to be the number on the first card selected (0 or 1); X 2 to be the number on the second card selected;...; and X n to be the number on the nth (last) card selected

11 Let s begin by computing probabilities for X 1. Clearly, P(X 1 = 1) = s/n = p and P(X 1 = 0) = f/n = q. We can present these equations in a table: Value Probability 0 q 1 p This table presents the sampling distribution for X 1. Upon reflection, this table is the sampling distribution for X 2, X 3,..., and X n. We summarize by saying that X 1, X 2, X 3,..., and X n are identically distributed. So, we can calculate probabilities for any individual card Next, we consider two cards simultaneously. For example, P(X 1 = 1 and X 2 = 0) = P(X 1 = 1, X 2 = 0). The result is the multiplication rule for probabilities. In this example, the multiplication rule says: P(X 1 = 1, X 2 = 0) = P(X 1 = 1)P(X 2 = 0) = pq. In words, we replace the word and by the operation of multiplying. (Similarly, recall, the addition rule replaced or by adding.) My argument for justifying the multiplication rule is a little tricky, so I will give it for a specific example, p = For this p, P(X 1 = 1, X 2 = 0) = pq = 0.60(0.40) = A brute force argument is given on pages of the text. The argument below appeals to the long-run interpretation of probability given in Section Consider the chance mechanism of selecting two cards at random w/replacement from the box. Now, imagine operating this chance mechanism a large number of times. We know that in the long-run approximately 60% of the operations will give a first card of 1. And of those operations that first give a 1, approximately 40% will give a second card of 0. Thus, in the long-run, 40% of 60% of the operations will give a 1 followed by a 0.. Next, remember that a percent of a percent is computed by converting to decimals and multiplying: 0.40(0.60) = qp = The multiplication rule can be extended in two directions. First, it is true for any two cards, not just the first two. Thus, for example, P(X 3 = 1, X 7 = 1) = pp = p 2. Second, it is true for more than two cards, for example, P(X 3 = 1, X 6 = 0, X 7 = 1) = pqp = p 2 q Do you recognize 4!? This is read 4-factorial (you don t need to shout) and it is calculated by: 4! = 4(3)(2)(1) = 24. Similarly, 3! = 3(2)(1) = 6 and 5! = 5(4)(3)(2)(1) = 120. These guys get big fast; for example, 50! = By special definition, 0! = 1. Now, define X to be X 1 + X X n. Literally, X is the sum of the numbers on the n cards selected. But b/c each card has a 1 or a 0,, X can be interpreted as the total number of successes in the sample. The variable X is very important in scientific applications. Thus, we would like to know its sampling distribution. Fortunately, there is a simple (?) formula for it, given on page 159 of the text: + 124

12 P(X = x) = n! x!(n x)! px q n x, for x = 0,1,..., n. This is a pretty amazing formula. It works for any choice of n and any value of p. It is called the Binomial sampling distribution with parameters n and p, written Bin(n, p). In this class, you need to be able to evaluate this formula with a hand calculator for n 6. For example, suppose we select n = 5 cards at random with replacement from a box with p = What is the probability we will get a representative sample? First, we realize that ˆp = x/5 will equal p = 0.60 if, and only if, x = 3. Thus, we want to calculate P(X = 3): P(X = 3) = 5! 3!(5 3)! (0.60)3 (0.40) 5 3 = (0.216)(0.16) = (2) For another example, for the same n and box, let s calculate P(X = 5). First, note that we can calculate this by using the multiplication rule. The event (X = 5) means that every card is a 1. Thus, P(X 1 = 1, X 2 = 1, X 3 = 1, X 4 = 1, X 5 = 1) = p 5 = (0.60) 5 = Using the binomial formula we get: P(X = 5) = 5! 5!(5 5)! (0.60)5 (0.40) 5 5 = 120 (0.0778)(1) = (1) Note that we need the definition 0! = 1 so that the formula works There are many statistical software packages that will compute binomial probabilities for us. For example, in the text on page 161 I give the computer-generated probabilities for Bin(25,0.50). Once we have binomial probabilities, we can draw probability histograms for binomials. Several such pictures are given on pages of the text. These pictures illustrate the following facts about the binomial. *δ = 1; thus, the height of each rectangle equals the probability of its center. *Just like the pictures for FT, there is one central peak one or two but never more rectangles wide and the probabilities steadily decrease as you move away from the peak. *The ph is symmetric, if and only if, p = *For p 0.50, the ph is roughly symmetric provided that both np and nq are large. People disagree on what large means; most say 5 or 10 or 15. Yes, we can use a computer to calculate binomial probabilities, but a computer should not be seen as a panacea. For example, if I try to use my software package for Bin(130,0.50) I get an error message. Unless a program is written with extreme care, the accurate calculation of, say, n! for large n is difficult. Thus, we might want to find a way to approximate binomial probabilities. Given the similarities of the binomial ph to the ph for FT, it is appealing to use the snc as an approximation. To this end, note that it can be shown that for Bin(n, p), µ = np and σ = npq. Thus, for example, Bin(100,0.50) has: µ = 100(0.50) = 50 and σ = 100(0.50)(0.50) = 25 = 5. As a result, it is easy to standardize X: Z = X np npq

13 On pages of the text are probability histograms for three Z s, each with an snc for comparison. Visually, it is clear that the snc can give good approximations for Z, and hence for X. The details will not be given and you are not responsible for them. This approximation will be used in Chapter 6. Section 5.3: Bernoulli Trials Consider the following experiments: Julie spins in circles to the right; Clyde shoots free throws; and Bob repeatedly tosses a coin. In each experiment, a person is conducting a sequence of trials. Consider the following question: Suppose that on Monday, Clyde attempts 100 free throws and achieves 77 successes. Clyde plans to attempt 200 free throws on Tuesday What is the probability that he will make 150 or more free throws on Tuesday? In order to answer this and similar questions, we need a mathematical model for the process that generates the results of the trials, whatever that means. We begin with a simple sequence of trials: repeated tosses of a fair coin. Think about this for a minute. What does it mean to you when you read, Repeated tosses of a fair coin? In particular, I want us to write down mathematical assumptions that describe this notion. There are three assumptions needed Each toss results in one of two outcomes: a heads or a tails. 2. The probability of heads is 0.50 for every toss. 3. The tosses exhibit no memory. We have studied the chance mechanism of selecting cards at random, with replacement, from a population box. For this CM, we learned two very useful techniques: the multiplication rule for a particular sequence of outcomes and the binomial sampling distribution for the total number of successes. Is there any relationship between the assumptions above and selecting cards from a box? Well, imagine a box with two cards, one card marked 1 for heads and the other marked 0 for tails. Suppose that we select cards at random with replacement from this box. I claim that this selection of cards from this box satisfies the three assumptions given above. Discuss. Thus, we can perform the following computations: If I toss a fair coin four times, the probability I get all heads is: P(H, H, H, H) = (0.50) 4 = If I toss a fair coin eight times, the probability that I get a total of exactly six heads is: 8! 6!2! (0.50)8 = Next, we generalize the above assumptions

14 Suppose that we have a sequence of trials. If they satisfy the following three assumptions, then we say that we have Bernoulli Trials (BT). 1. Each trial results in one of two outcomes: a success or a failure. 2. The probability of success equals p for every trial. 3. The trials exhibit no memory. we can calculate the following probabilities. If Katie shoots three free throws, the probability she makes all three is: P(S, S, S) = (0.85) 3 = If Katie shoots ten free throws, the probability she makes a total of exactly nine is: 10! 9!1! (0.85)9 (0.15) = To summarize if we are told that we have BT and we are told the value of p, then we can calculate probabilities about the outcomes of the trials. As argued above for a coin, BT are mathematically equivalent to selecting cards at random with replacement from Box(N; p). For example, Katie is a very good free throw shooter. On the assumption that Katie s free throws are BT with p = 0.85, What are the difficulties with this? Well, education should be more than learning to obey authority figures! I acknowledge this above by prefacing my computation by saying, On the assumption... are BT, but we can (sometimes) do more If we have previous outcomes of the trials, we can use these data to investigate (not determine!) whether the assumptions of BT seem reasonable. I offer two ways to investigate. One way is designed to explore the second assumption (constancy of success probability) and the other focuses on the third assumption (lack of memory). To be precise, suppose we have observed the following results of n = 10 trials: We investigate constancy by creating the following table: Half S F Total ˆp Total We need to be careful. We want to know whether p remains constant, but we never get to see p. We can see the ˆp s. For the table above, ˆp 1 = ˆp 2 = B/c the ˆp s do not change from the first half to the second half, there is no evidence that p has changed. Of course, this argument is not conclusive. We could make it more formal, but that is not my goal. I encourage you to think of this as an informal hypothesis test for which the null hypothesis is that p remains constant. Let s look at the trials again: Do you see memory? Well the last five trials yield exactly the same outcomes as the first five, so this looks like memory. But it could be simply the result of chance

15 But now you need to be a scientist. Does this 5 step memory make any sense to you? Discuss. Often, what makes sense is 1 step memory; i.e. memory in which a current trial is influenced by the outcome of the trial immediately before it. We take the 10 trials and form 9 overlapping pairs. (It is too difficult to do this w/my word processor. I will illustrate it on the board.) We create the following table. Note that in this table we are counting pairs not trials. Previous Current Trial Trial S F Total ˆp Total B/c the two ˆp s are equal, there is no evidence of 1 step memory. Discuss. I want to show you an easier and faster way to create the above memory table. First, we know that the number of pairs will be (n 1), which is nine for these data. Also, we know from previously comparing halves that there is a total of four successes and six failures. We put these numbers in the margins, yielding the partial table below. Previous Current Trial Trial S F Total ˆp Total A difficulty with this table, of course, is that = 10, not 9. But look at the first trial in the sequence, a success The first trial appears in only one pair b/c it has no trial before it. In other words, the first trial never gets to be current in the memory table. Thus, even though there are four successes in the sequence, only three of them are current. Thus, we change the column total for S to 3. Similarly, the last trial, also an S, never gets to be previous. Thus, we subtract one from the row total for S to get the following partial table, which has the correct margins. Previous Current Trial Trial S F Total ˆp Total Now, we simply determine by counting one of the entries in the table, and obtain the others by subtracting. For example, the pair SS (11) occurs exactly once in the sequence Here is another example with n = 20: These data yield the following tables. Discuss. Half S F Total ˆp Total Previous Current Trial Trial S F Total ˆp Total Over the years, I have had trouble convincing my students that BT exist in the world. (Math majors are extremely willing to believe assumptions; nonmath majors are not.) I decided to use the video game Tetris to convince my students that BTs do exist. Sadly, as we will see, I failed miserably

16 I collected these data circa 1990 using a Nintendo system played on my television. A trial is a shape of blocks falling from the top of the screen. First difficultly: There are 7, not 2 shapes Play Solution: Define log to be an S; all others are F s. I played 8 games and observed n = 1872 trials. Instead of dividing the trials into halves, I created the following table: Play S F Total ˆp First Second Third Fourth Fifth Sixth Seventh Eighth Total Looking at this table and its plot, I was feeling pretty good. There is some evidence that p is not constant, but the evidence seems weak to me. (A formal HT of the null hypothesis that p is constant versus the alternative that it can change every game see Chapter 11 for details gives a very large P-value, ) But then I created the memory table: Current Shape Previous Shape S F Total ˆp Success Failure Total Note: The grand total for this table is (n 8) b/c we lose one pair for each play. (I believe that it does not make sense to look for memory by pairing the last trial of one game with the first trial of the next game.) It is obvious (?) that there is memory. (The P-value is ) One of my favorite projects: Describe the petturtle study. Section 5.4: Some Practical Considerations The big result of Chapter 5 is the multiplication rule, from which we obtain the Binomial and many other results not discussed in this course. For a finite population, we get the MR if we sample at random w/replacement and for an infinite population if we have BTs B/c humans, and I would say especially mathematicians, like answers, there is a tremendous pressure to assume/pretend/deceive that we have the MR. In my experience, it is very common for researchers to claim that they have a random sample from a finite population, even when they clearly do not. Examples: Pretty much every survey you ever read about. My Wisconsin DOT study. (Discuss.) Birthdays. Below is a very famous table in probability theory

17 n Prob. n Prob. n Prob n Prob. n Prob Probabilities must be computed before the CM is operated. But despite my feeling this way, many people try to calculate probabilities after the CM has been operated. In my experience, the huge majority of such computations lead to gibberish. There are three major mistakes that people make. I call them: ignoring failed attempts; focusing on a winner; and inappropriate use of ELC. Multiple lotto winner. Said to beat-the-odds of one in 16 trillion. Discuss Discuss batting order example. Yahtzee. Assuming that the five dice are balanced (ELC) and act independently (no memory from die to die) then the probability of getting a Yahtzee on a single throw of the five dice is: It was reported on TV that the probability of this happening was 1 in Let s put that number in perspective. There are currently 30 teams in MLB, playing 162 games per year for a total of 2430 games per year. The above event is, thus, expected to happen once every 54,190,080 years. I really doubt that MLB will survive that long! (1/6) 4 = 1/1296. Thus, if you pick up five dice and throw a Yahtzee, that is pretty great. But suppose you toss the dice 10 times per minute for two hours and get a Yahtzee; what then? Suppose that 12,960 people each throw five dice at once. Do you think there will be any Yahtzees? How many?

Section 3.1: The probability histogram. Below is the probability histogram for the IS (p. 80 of text):

Section 3.1: The probability histogram. Below is the probability histogram for the IS (p. 80 of text): Chapter 3 Section 3.1: The probability histogram Below is the probability histogram for the IS (p. 80 of text): 1.5 1.5755 1.0 0.7500 0.5 0.90 0.30 0.30 0.90 How is it determined? + 81 On a horizontal

More information

2.1 Independent, Identically Distributed Trials

2.1 Independent, Identically Distributed Trials Chapter 2 Trials 2.1 Independent, Identically Distributed Trials In Chapter 1 we considered the operation of a CM. Many, but not all, CMs can be operated more than once. For example, a coin can be tossed

More information

Chapter 3. Estimation of p. 3.1 Point and Interval Estimates of p

Chapter 3. Estimation of p. 3.1 Point and Interval Estimates of p Chapter 3 Estimation of p 3.1 Point and Interval Estimates of p Suppose that we have Bernoulli Trials (BT). So far, in every example I have told you the (numerical) value of p. In science, usually the

More information

1 Normal Distribution.

1 Normal Distribution. Normal Distribution.. Introduction A Bernoulli trial is simple random experiment that ends in success or failure. A Bernoulli trial can be used to make a new random experiment by repeating the Bernoulli

More information

Chapter 8: An Introduction to Probability and Statistics

Chapter 8: An Introduction to Probability and Statistics Course S3, 200 07 Chapter 8: An Introduction to Probability and Statistics This material is covered in the book: Erwin Kreyszig, Advanced Engineering Mathematics (9th edition) Chapter 24 (not including

More information

P (A) = P (B) = P (C) = P (D) =

P (A) = P (B) = P (C) = P (D) = STAT 145 CHAPTER 12 - PROBABILITY - STUDENT VERSION The probability of a random event, is the proportion of times the event will occur in a large number of repititions. For example, when flipping a coin,

More information

Chapter 4: An Introduction to Probability and Statistics

Chapter 4: An Introduction to Probability and Statistics Chapter 4: An Introduction to Probability and Statistics 4. Probability The simplest kinds of probabilities to understand are reflected in everyday ideas like these: (i) if you toss a coin, the probability

More information

CS 124 Math Review Section January 29, 2018

CS 124 Math Review Section January 29, 2018 CS 124 Math Review Section CS 124 is more math intensive than most of the introductory courses in the department. You re going to need to be able to do two things: 1. Perform some clever calculations to

More information

The Celsius temperature scale is based on the freezing point and the boiling point of water. 12 degrees Celsius below zero would be written as

The Celsius temperature scale is based on the freezing point and the boiling point of water. 12 degrees Celsius below zero would be written as Prealgebra, Chapter 2 - Integers, Introductory Algebra 2.1 Integers In the real world, numbers are used to represent real things, such as the height of a building, the cost of a car, the temperature of

More information

Senior Math Circles November 19, 2008 Probability II

Senior Math Circles November 19, 2008 Probability II University of Waterloo Faculty of Mathematics Centre for Education in Mathematics and Computing Senior Math Circles November 9, 2008 Probability II Probability Counting There are many situations where

More information

Chapter 1 Review of Equations and Inequalities

Chapter 1 Review of Equations and Inequalities Chapter 1 Review of Equations and Inequalities Part I Review of Basic Equations Recall that an equation is an expression with an equal sign in the middle. Also recall that, if a question asks you to solve

More information

Please bring the task to your first physics lesson and hand it to the teacher.

Please bring the task to your first physics lesson and hand it to the teacher. Pre-enrolment task for 2014 entry Physics Why do I need to complete a pre-enrolment task? This bridging pack serves a number of purposes. It gives you practice in some of the important skills you will

More information

MA 1125 Lecture 15 - The Standard Normal Distribution. Friday, October 6, Objectives: Introduce the standard normal distribution and table.

MA 1125 Lecture 15 - The Standard Normal Distribution. Friday, October 6, Objectives: Introduce the standard normal distribution and table. MA 1125 Lecture 15 - The Standard Normal Distribution Friday, October 6, 2017. Objectives: Introduce the standard normal distribution and table. 1. The Standard Normal Distribution We ve been looking at

More information

Discrete Mathematics and Probability Theory Fall 2014 Anant Sahai Note 15. Random Variables: Distributions, Independence, and Expectations

Discrete Mathematics and Probability Theory Fall 2014 Anant Sahai Note 15. Random Variables: Distributions, Independence, and Expectations EECS 70 Discrete Mathematics and Probability Theory Fall 204 Anant Sahai Note 5 Random Variables: Distributions, Independence, and Expectations In the last note, we saw how useful it is to have a way of

More information

Discrete Mathematics and Probability Theory Fall 2013 Vazirani Note 12. Random Variables: Distribution and Expectation

Discrete Mathematics and Probability Theory Fall 2013 Vazirani Note 12. Random Variables: Distribution and Expectation CS 70 Discrete Mathematics and Probability Theory Fall 203 Vazirani Note 2 Random Variables: Distribution and Expectation We will now return once again to the question of how many heads in a typical sequence

More information

Notes 1 Autumn Sample space, events. S is the number of elements in the set S.)

Notes 1 Autumn Sample space, events. S is the number of elements in the set S.) MAS 108 Probability I Notes 1 Autumn 2005 Sample space, events The general setting is: We perform an experiment which can have a number of different outcomes. The sample space is the set of all possible

More information

Chapter 18. Sampling Distribution Models. Copyright 2010, 2007, 2004 Pearson Education, Inc.

Chapter 18. Sampling Distribution Models. Copyright 2010, 2007, 2004 Pearson Education, Inc. Chapter 18 Sampling Distribution Models Copyright 2010, 2007, 2004 Pearson Education, Inc. Normal Model When we talk about one data value and the Normal model we used the notation: N(μ, σ) Copyright 2010,

More information

Module 8 Probability

Module 8 Probability Module 8 Probability Probability is an important part of modern mathematics and modern life, since so many things involve randomness. The ClassWiz is helpful for calculating probabilities, especially those

More information

UNIT NUMBER PROBABILITY 6 (Statistics for the binomial distribution) A.J.Hobson

UNIT NUMBER PROBABILITY 6 (Statistics for the binomial distribution) A.J.Hobson JUST THE MATHS UNIT NUMBER 19.6 PROBABILITY 6 (Statistics for the binomial distribution) by A.J.Hobson 19.6.1 Construction of histograms 19.6.2 Mean and standard deviation of a binomial distribution 19.6.3

More information

Chapter 18. Sampling Distribution Models /51

Chapter 18. Sampling Distribution Models /51 Chapter 18 Sampling Distribution Models 1 /51 Homework p432 2, 4, 6, 8, 10, 16, 17, 20, 30, 36, 41 2 /51 3 /51 Objective Students calculate values of central 4 /51 The Central Limit Theorem for Sample

More information

Discrete Mathematics and Probability Theory Fall 2012 Vazirani Note 14. Random Variables: Distribution and Expectation

Discrete Mathematics and Probability Theory Fall 2012 Vazirani Note 14. Random Variables: Distribution and Expectation CS 70 Discrete Mathematics and Probability Theory Fall 202 Vazirani Note 4 Random Variables: Distribution and Expectation Random Variables Question: The homeworks of 20 students are collected in, randomly

More information

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n =

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n = Hypothesis testing I I. What is hypothesis testing? [Note we re temporarily bouncing around in the book a lot! Things will settle down again in a week or so] - Exactly what it says. We develop a hypothesis,

More information

Module 03 Lecture 14 Inferential Statistics ANOVA and TOI

Module 03 Lecture 14 Inferential Statistics ANOVA and TOI Introduction of Data Analytics Prof. Nandan Sudarsanam and Prof. B Ravindran Department of Management Studies and Department of Computer Science and Engineering Indian Institute of Technology, Madras Module

More information

10.1. Randomness and Probability. Investigation: Flip a Coin EXAMPLE A CONDENSED LESSON

10.1. Randomness and Probability. Investigation: Flip a Coin EXAMPLE A CONDENSED LESSON CONDENSED LESSON 10.1 Randomness and Probability In this lesson you will simulate random processes find experimental probabilities based on the results of a large number of trials calculate theoretical

More information

Steve Smith Tuition: Maths Notes

Steve Smith Tuition: Maths Notes Maths Notes : Discrete Random Variables Version. Steve Smith Tuition: Maths Notes e iπ + = 0 a + b = c z n+ = z n + c V E + F = Discrete Random Variables Contents Intro The Distribution of Probabilities

More information

MITOCW ocw f99-lec30_300k

MITOCW ocw f99-lec30_300k MITOCW ocw-18.06-f99-lec30_300k OK, this is the lecture on linear transformations. Actually, linear algebra courses used to begin with this lecture, so you could say I'm beginning this course again by

More information

MITOCW ocw f99-lec05_300k

MITOCW ocw f99-lec05_300k MITOCW ocw-18.06-f99-lec05_300k This is lecture five in linear algebra. And, it will complete this chapter of the book. So the last section of this chapter is two point seven that talks about permutations,

More information

Probability Year 9. Terminology

Probability Year 9. Terminology Probability Year 9 Terminology Probability measures the chance something happens. Formally, we say it measures how likely is the outcome of an event. We write P(result) as a shorthand. An event is some

More information

Descriptive Statistics (And a little bit on rounding and significant digits)

Descriptive Statistics (And a little bit on rounding and significant digits) Descriptive Statistics (And a little bit on rounding and significant digits) Now that we know what our data look like, we d like to be able to describe it numerically. In other words, how can we represent

More information

Introduction to Algebra: The First Week

Introduction to Algebra: The First Week Introduction to Algebra: The First Week Background: According to the thermostat on the wall, the temperature in the classroom right now is 72 degrees Fahrenheit. I want to write to my friend in Europe,

More information

Counting Out πr 2. Teacher Lab Discussion. Overview. Picture, Data Table, and Graph. Part I Middle Counting Length/Area Out πrinvestigation

Counting Out πr 2. Teacher Lab Discussion. Overview. Picture, Data Table, and Graph. Part I Middle Counting Length/Area Out πrinvestigation 5 6 7 Middle Counting Length/rea Out πrinvestigation, page 1 of 7 Counting Out πr Teacher Lab Discussion Figure 1 Overview In this experiment we study the relationship between the radius of a circle and

More information

1.1 The Language of Mathematics Expressions versus Sentences

1.1 The Language of Mathematics Expressions versus Sentences The Language of Mathematics Expressions versus Sentences a hypothetical situation the importance of language Study Strategies for Students of Mathematics characteristics of the language of mathematics

More information

Notes Week 2 Chapter 3 Probability WEEK 2 page 1

Notes Week 2 Chapter 3 Probability WEEK 2 page 1 Notes Week 2 Chapter 3 Probability WEEK 2 page 1 The sample space of an experiment, sometimes denoted S or in probability theory, is the set that consists of all possible elementary outcomes of that experiment

More information

Probability Experiments, Trials, Outcomes, Sample Spaces Example 1 Example 2

Probability Experiments, Trials, Outcomes, Sample Spaces Example 1 Example 2 Probability Probability is the study of uncertain events or outcomes. Games of chance that involve rolling dice or dealing cards are one obvious area of application. However, probability models underlie

More information

Sampling Distribution Models. Chapter 17

Sampling Distribution Models. Chapter 17 Sampling Distribution Models Chapter 17 Objectives: 1. Sampling Distribution Model 2. Sampling Variability (sampling error) 3. Sampling Distribution Model for a Proportion 4. Central Limit Theorem 5. Sampling

More information

P (E) = P (A 1 )P (A 2 )... P (A n ).

P (E) = P (A 1 )P (A 2 )... P (A n ). Lecture 9: Conditional probability II: breaking complex events into smaller events, methods to solve probability problems, Bayes rule, law of total probability, Bayes theorem Discrete Structures II (Summer

More information

Note: Please use the actual date you accessed this material in your citation.

Note: Please use the actual date you accessed this material in your citation. MIT OpenCourseWare http://ocw.mit.edu 18.06 Linear Algebra, Spring 2005 Please use the following citation format: Gilbert Strang, 18.06 Linear Algebra, Spring 2005. (Massachusetts Institute of Technology:

More information

MITOCW ocw f99-lec16_300k

MITOCW ocw f99-lec16_300k MITOCW ocw-18.06-f99-lec16_300k OK. Here's lecture sixteen and if you remember I ended up the last lecture with this formula for what I called a projection matrix. And maybe I could just recap for a minute

More information

Last few slides from last time

Last few slides from last time Last few slides from last time Example 3: What is the probability that p will fall in a certain range, given p? Flip a coin 50 times. If the coin is fair (p=0.5), what is the probability of getting an

More information

Probability (Devore Chapter Two)

Probability (Devore Chapter Two) Probability (Devore Chapter Two) 1016-345-01: Probability and Statistics for Engineers Fall 2012 Contents 0 Administrata 2 0.1 Outline....................................... 3 1 Axiomatic Probability 3

More information

Sequences and Series

Sequences and Series Sequences and Series What do you think of when you read the title of our next unit? In case your answers are leading us off track, let's review the following IB problems. 1 November 2013 HL 2 3 November

More information

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 16. Random Variables: Distribution and Expectation

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 16. Random Variables: Distribution and Expectation CS 70 Discrete Mathematics and Probability Theory Spring 206 Rao and Walrand Note 6 Random Variables: Distribution and Expectation Example: Coin Flips Recall our setup of a probabilistic experiment as

More information

Probability, For the Enthusiastic Beginner (Exercises, Version 1, September 2016) David Morin,

Probability, For the Enthusiastic Beginner (Exercises, Version 1, September 2016) David Morin, Chapter 8 Exercises Probability, For the Enthusiastic Beginner (Exercises, Version 1, September 2016) David Morin, morin@physics.harvard.edu 8.1 Chapter 1 Section 1.2: Permutations 1. Assigning seats *

More information

Probability Year 10. Terminology

Probability Year 10. Terminology Probability Year 10 Terminology Probability measures the chance something happens. Formally, we say it measures how likely is the outcome of an event. We write P(result) as a shorthand. An event is some

More information

CIS 2033 Lecture 5, Fall

CIS 2033 Lecture 5, Fall CIS 2033 Lecture 5, Fall 2016 1 Instructor: David Dobor September 13, 2016 1 Supplemental reading from Dekking s textbook: Chapter2, 3. We mentioned at the beginning of this class that calculus was a prerequisite

More information

ACCESS TO SCIENCE, ENGINEERING AND AGRICULTURE: MATHEMATICS 2 MATH00040 SEMESTER / Probability

ACCESS TO SCIENCE, ENGINEERING AND AGRICULTURE: MATHEMATICS 2 MATH00040 SEMESTER / Probability ACCESS TO SCIENCE, ENGINEERING AND AGRICULTURE: MATHEMATICS 2 MATH00040 SEMESTER 2 2017/2018 DR. ANTHONY BROWN 5.1. Introduction to Probability. 5. Probability You are probably familiar with the elementary

More information

Grade 8 Chapter 7: Rational and Irrational Numbers

Grade 8 Chapter 7: Rational and Irrational Numbers Grade 8 Chapter 7: Rational and Irrational Numbers In this chapter we first review the real line model for numbers, as discussed in Chapter 2 of seventh grade, by recalling how the integers and then the

More information

Outline. Probability. Math 143. Department of Mathematics and Statistics Calvin College. Spring 2010

Outline. Probability. Math 143. Department of Mathematics and Statistics Calvin College. Spring 2010 Outline Math 143 Department of Mathematics and Statistics Calvin College Spring 2010 Outline Outline 1 Review Basics Random Variables Mean, Variance and Standard Deviation of Random Variables 2 More Review

More information

Section 5.2. Section Carol performs 100 dichotomous trials and obtains the following data. Extra Exercises; Statistics 301; Professor Wardrop

Section 5.2. Section Carol performs 100 dichotomous trials and obtains the following data. Extra Exercises; Statistics 301; Professor Wardrop Section 5.2 Extra Exercises; Statistics 301; Professor Wardrop 1. Four cards are selected at random with replacement from a dichotomous box with p = 0.46. Calculate the probability of each of the following

More information

MITOCW ocw f99-lec01_300k

MITOCW ocw f99-lec01_300k MITOCW ocw-18.06-f99-lec01_300k Hi. This is the first lecture in MIT's course 18.06, linear algebra, and I'm Gilbert Strang. The text for the course is this book, Introduction to Linear Algebra. And the

More information

1.1 Administrative Stuff

1.1 Administrative Stuff 601.433 / 601.633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Introduction, Karatsuba/Strassen Date: 9/4/18 1.1 Administrative Stuff Welcome to Algorithms! In this class you will learn the

More information

Discrete Mathematics for CS Spring 2006 Vazirani Lecture 22

Discrete Mathematics for CS Spring 2006 Vazirani Lecture 22 CS 70 Discrete Mathematics for CS Spring 2006 Vazirani Lecture 22 Random Variables and Expectation Question: The homeworks of 20 students are collected in, randomly shuffled and returned to the students.

More information

Relationships Between Quantities

Relationships Between Quantities Algebra 1 Relationships Between Quantities Relationships Between Quantities Everyone loves math until there are letters (known as variables) in problems!! Do students complain about reading when they come

More information

Experiment 2 Random Error and Basic Statistics

Experiment 2 Random Error and Basic Statistics PHY191 Experiment 2: Random Error and Basic Statistics 7/12/2011 Page 1 Experiment 2 Random Error and Basic Statistics Homework 2: turn in the second week of the experiment. This is a difficult homework

More information

Binomial Distribution. Collin Phillips

Binomial Distribution. Collin Phillips Mathematics Learning Centre Binomial Distribution Collin Phillips c 00 University of Sydney Thanks To Darren Graham and Cathy Kennedy for turning my scribble into a book and to Jackie Nicholas and Sue

More information

Hi, my name is Dr. Ann Weaver of Argosy University. This WebEx is about something in statistics called z-

Hi, my name is Dr. Ann Weaver of Argosy University. This WebEx is about something in statistics called z- Hi, my name is Dr. Ann Weaver of Argosy University. This WebEx is about something in statistics called z- Scores. I have two purposes for this WebEx, one, I just want to show you how to use z-scores in

More information

5.3 Conditional Probability and Independence

5.3 Conditional Probability and Independence 28 CHAPTER 5. PROBABILITY 5. Conditional Probability and Independence 5.. Conditional Probability Two cubical dice each have a triangle painted on one side, a circle painted on two sides and a square painted

More information

STAT/SOC/CSSS 221 Statistical Concepts and Methods for the Social Sciences. Random Variables

STAT/SOC/CSSS 221 Statistical Concepts and Methods for the Social Sciences. Random Variables STAT/SOC/CSSS 221 Statistical Concepts and Methods for the Social Sciences Random Variables Christopher Adolph Department of Political Science and Center for Statistics and the Social Sciences University

More information

(c) P(BC c ) = One point was lost for multiplying. If, however, you made this same mistake in (b) and (c) you lost the point only once.

(c) P(BC c ) = One point was lost for multiplying. If, however, you made this same mistake in (b) and (c) you lost the point only once. Solutions to First Midterm Exam, Stat 371, Fall 2010 There are two, three or four versions of each problem. The problems on your exam comprise a mix of versions. As a result, when you examine the solutions

More information

CISC 1100/1400 Structures of Comp. Sci./Discrete Structures Chapter 7 Probability. Outline. Terminology and background. Arthur G.

CISC 1100/1400 Structures of Comp. Sci./Discrete Structures Chapter 7 Probability. Outline. Terminology and background. Arthur G. CISC 1100/1400 Structures of Comp. Sci./Discrete Structures Chapter 7 Probability Arthur G. Werschulz Fordham University Department of Computer and Information Sciences Copyright Arthur G. Werschulz, 2017.

More information

Looking Ahead to Chapter 10

Looking Ahead to Chapter 10 Looking Ahead to Chapter Focus In Chapter, you will learn about polynomials, including how to add, subtract, multiply, and divide polynomials. You will also learn about polynomial and rational functions.

More information

CS 361: Probability & Statistics

CS 361: Probability & Statistics February 19, 2018 CS 361: Probability & Statistics Random variables Markov s inequality This theorem says that for any random variable X and any value a, we have A random variable is unlikely to have an

More information

Introduction to Algorithms

Introduction to Algorithms Lecture 1 Introduction to Algorithms 1.1 Overview The purpose of this lecture is to give a brief overview of the topic of Algorithms and the kind of thinking it involves: why we focus on the subjects that

More information

STA Why Sampling? Module 6 The Sampling Distributions. Module Objectives

STA Why Sampling? Module 6 The Sampling Distributions. Module Objectives STA 2023 Module 6 The Sampling Distributions Module Objectives In this module, we will learn the following: 1. Define sampling error and explain the need for sampling distributions. 2. Recognize that sampling

More information

Note: Please use the actual date you accessed this material in your citation.

Note: Please use the actual date you accessed this material in your citation. MIT OpenCourseWare http://ocw.mit.edu 18.02 Multivariable Calculus, Fall 2007 Please use the following citation format: Denis Auroux. 18.02 Multivariable Calculus, Fall 2007. (Massachusetts Institute of

More information

Experiment 2 Random Error and Basic Statistics

Experiment 2 Random Error and Basic Statistics PHY9 Experiment 2: Random Error and Basic Statistics 8/5/2006 Page Experiment 2 Random Error and Basic Statistics Homework 2: Turn in at start of experiment. Readings: Taylor chapter 4: introduction, sections

More information

Basic Probability. Introduction

Basic Probability. Introduction Basic Probability Introduction The world is an uncertain place. Making predictions about something as seemingly mundane as tomorrow s weather, for example, is actually quite a difficult task. Even with

More information

Statistics 1L03 - Midterm #2 Review

Statistics 1L03 - Midterm #2 Review Statistics 1L03 - Midterm # Review Atinder Bharaj Made with L A TEX October, 01 Introduction As many of you will soon find out, I will not be holding the next midterm review. To make it a bit easier on

More information

MITOCW ocw f99-lec17_300k

MITOCW ocw f99-lec17_300k MITOCW ocw-18.06-f99-lec17_300k OK, here's the last lecture in the chapter on orthogonality. So we met orthogonal vectors, two vectors, we met orthogonal subspaces, like the row space and null space. Now

More information

7.1 What is it and why should we care?

7.1 What is it and why should we care? Chapter 7 Probability In this section, we go over some simple concepts from probability theory. We integrate these with ideas from formal language theory in the next chapter. 7.1 What is it and why should

More information

Problems from Probability and Statistical Inference (9th ed.) by Hogg, Tanis and Zimmerman.

Problems from Probability and Statistical Inference (9th ed.) by Hogg, Tanis and Zimmerman. Math 224 Fall 2017 Homework 1 Drew Armstrong Problems from Probability and Statistical Inference (9th ed.) by Hogg, Tanis and Zimmerman. Section 1.1, Exercises 4,5,6,7,9,12. Solutions to Book Problems.

More information

Special Theory of Relativity Prof. Shiva Prasad Department of Physics Indian Institute of Technology, Bombay. Lecture - 15 Momentum Energy Four Vector

Special Theory of Relativity Prof. Shiva Prasad Department of Physics Indian Institute of Technology, Bombay. Lecture - 15 Momentum Energy Four Vector Special Theory of Relativity Prof. Shiva Prasad Department of Physics Indian Institute of Technology, Bombay Lecture - 15 Momentum Energy Four Vector We had started discussing the concept of four vectors.

More information

Appendix A. Review of Basic Mathematical Operations. 22Introduction

Appendix A. Review of Basic Mathematical Operations. 22Introduction Appendix A Review of Basic Mathematical Operations I never did very well in math I could never seem to persuade the teacher that I hadn t meant my answers literally. Introduction Calvin Trillin Many of

More information

Dealing with the assumption of independence between samples - introducing the paired design.

Dealing with the assumption of independence between samples - introducing the paired design. Dealing with the assumption of independence between samples - introducing the paired design. a) Suppose you deliberately collect one sample and measure something. Then you collect another sample in such

More information

MITOCW ocw-18_02-f07-lec02_220k

MITOCW ocw-18_02-f07-lec02_220k MITOCW ocw-18_02-f07-lec02_220k The following content is provided under a Creative Commons license. Your support will help MIT OpenCourseWare continue to offer high quality educational resources for free.

More information

Grades 7 & 8, Math Circles 10/11/12 October, Series & Polygonal Numbers

Grades 7 & 8, Math Circles 10/11/12 October, Series & Polygonal Numbers Faculty of Mathematics Waterloo, Ontario N2L G Centre for Education in Mathematics and Computing Introduction Grades 7 & 8, Math Circles 0//2 October, 207 Series & Polygonal Numbers Mathematicians are

More information

Guide to Proofs on Sets

Guide to Proofs on Sets CS103 Winter 2019 Guide to Proofs on Sets Cynthia Lee Keith Schwarz I would argue that if you have a single guiding principle for how to mathematically reason about sets, it would be this one: All sets

More information

DIFFERENTIAL EQUATIONS

DIFFERENTIAL EQUATIONS DIFFERENTIAL EQUATIONS Basic Concepts Paul Dawkins Table of Contents Preface... Basic Concepts... 1 Introduction... 1 Definitions... Direction Fields... 8 Final Thoughts...19 007 Paul Dawkins i http://tutorial.math.lamar.edu/terms.aspx

More information

Manipulating Radicals

Manipulating Radicals Lesson 40 Mathematics Assessment Project Formative Assessment Lesson Materials Manipulating Radicals MARS Shell Center University of Nottingham & UC Berkeley Alpha Version Please Note: These materials

More information

Special Theory Of Relativity Prof. Shiva Prasad Department of Physics Indian Institute of Technology, Bombay

Special Theory Of Relativity Prof. Shiva Prasad Department of Physics Indian Institute of Technology, Bombay Special Theory Of Relativity Prof. Shiva Prasad Department of Physics Indian Institute of Technology, Bombay Lecture - 6 Length Contraction and Time Dilation (Refer Slide Time: 00:29) In our last lecture,

More information

Conditional Probability, Independence and Bayes Theorem Class 3, Jeremy Orloff and Jonathan Bloom

Conditional Probability, Independence and Bayes Theorem Class 3, Jeremy Orloff and Jonathan Bloom Conditional Probability, Independence and Bayes Theorem Class 3, 18.05 Jeremy Orloff and Jonathan Bloom 1 Learning Goals 1. Know the definitions of conditional probability and independence of events. 2.

More information

Lecture 4: Constructing the Integers, Rationals and Reals

Lecture 4: Constructing the Integers, Rationals and Reals Math/CS 20: Intro. to Math Professor: Padraic Bartlett Lecture 4: Constructing the Integers, Rationals and Reals Week 5 UCSB 204 The Integers Normally, using the natural numbers, you can easily define

More information

MITOCW ocw f99-lec09_300k

MITOCW ocw f99-lec09_300k MITOCW ocw-18.06-f99-lec09_300k OK, this is linear algebra lecture nine. And this is a key lecture, this is where we get these ideas of linear independence, when a bunch of vectors are independent -- or

More information

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.] Math 43 Review Notes [Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty Dot Product If v (v, v, v 3 and w (w, w, w 3, then the

More information

Math 3361-Modern Algebra Lecture 08 9/26/ Cardinality

Math 3361-Modern Algebra Lecture 08 9/26/ Cardinality Math 336-Modern Algebra Lecture 08 9/26/4. Cardinality I started talking about cardinality last time, and you did some stuff with it in the Homework, so let s continue. I said that two sets have the same

More information

MAT Mathematics in Today's World

MAT Mathematics in Today's World MAT 1000 Mathematics in Today's World Last Time We discussed the four rules that govern probabilities: 1. Probabilities are numbers between 0 and 1 2. The probability an event does not occur is 1 minus

More information

Chapter 14. Simple Linear Regression Preliminary Remarks The Simple Linear Regression Model

Chapter 14. Simple Linear Regression Preliminary Remarks The Simple Linear Regression Model Chapter 14 Simple Linear Regression 14.1 Preliminary Remarks We have only three lectures to introduce the ideas of regression. Worse, they are the last three lectures when you have many other things academic

More information

STAT 285 Fall Assignment 1 Solutions

STAT 285 Fall Assignment 1 Solutions STAT 285 Fall 2014 Assignment 1 Solutions 1. An environmental agency sets a standard of 200 ppb for the concentration of cadmium in a lake. The concentration of cadmium in one lake is measured 17 times.

More information

79 Wyner Math Academy I Spring 2016

79 Wyner Math Academy I Spring 2016 79 Wyner Math Academy I Spring 2016 CHAPTER NINE: HYPOTHESIS TESTING Review May 11 Test May 17 Research requires an understanding of underlying mathematical distributions as well as of the research methods

More information

Basics of Proofs. 1 The Basics. 2 Proof Strategies. 2.1 Understand What s Going On

Basics of Proofs. 1 The Basics. 2 Proof Strategies. 2.1 Understand What s Going On Basics of Proofs The Putnam is a proof based exam and will expect you to write proofs in your solutions Similarly, Math 96 will also require you to write proofs in your homework solutions If you ve seen

More information

Rules for Means and Variances; Prediction

Rules for Means and Variances; Prediction Chapter 7 Rules for Means and Variances; Prediction 7.1 Rules for Means and Variances The material in this section is very technical and algebraic. And dry. But it is useful for understanding many of the

More information

Sequences and infinite series

Sequences and infinite series Sequences and infinite series D. DeTurck University of Pennsylvania March 29, 208 D. DeTurck Math 04 002 208A: Sequence and series / 54 Sequences The lists of numbers you generate using a numerical method

More information

Lesson 1 WHAT S A GROUP?

Lesson 1 WHAT S A GROUP? WHAT S A GROUP? What is a group? Well, that s a question we can begin to answer in several different ways. First and foremost, we can say that groups are composed of cycles and that s why groups are so

More information

Mathematics for Health and Physical Sciences

Mathematics for Health and Physical Sciences 1 Mathematics for Health and Physical Sciences Collection edited by: Wendy Lightheart Content authors: Wendy Lightheart, OpenStax, Wade Ellis, Denny Burzynski, Jan Clayton, and John Redden Online:

More information

Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institution of Technology, Kharagpur

Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institution of Technology, Kharagpur Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institution of Technology, Kharagpur Lecture No. # 36 Sampling Distribution and Parameter Estimation

More information

Creating and Exploring Circles

Creating and Exploring Circles Creating and Exploring Circles 1. Close your compass, take a plain sheet of paper and use the compass point to make a tiny hole (point) in what you consider to be the very centre of the paper. The centre

More information

Roberto s Notes on Linear Algebra Chapter 11: Vector spaces Section 1. Vector space axioms

Roberto s Notes on Linear Algebra Chapter 11: Vector spaces Section 1. Vector space axioms Roberto s Notes on Linear Algebra Chapter 11: Vector spaces Section 1 Vector space axioms What you need to know already: How Euclidean vectors work. What linear combinations are and why they are important.

More information

2. Probability. Chris Piech and Mehran Sahami. Oct 2017

2. Probability. Chris Piech and Mehran Sahami. Oct 2017 2. Probability Chris Piech and Mehran Sahami Oct 2017 1 Introduction It is that time in the quarter (it is still week one) when we get to talk about probability. Again we are going to build up from first

More information

Chapter 1. Probability. 1.1 The Standard Normal Curve.

Chapter 1. Probability. 1.1 The Standard Normal Curve. 0 Chapter 1 Probability 1.1 The Standard Normal Curve. In this section we will learn how to use an approximating device that we will need throughout the semester. The standard normal curve, henceforth

More information

1 What is the area model for multiplication?

1 What is the area model for multiplication? for multiplication represents a lovely way to view the distribution property the real number exhibit. This property is the link between addition and multiplication. 1 1 What is the area model for multiplication?

More information