Expected Value and Variance
|
|
- Gerard Jenkins
- 5 years ago
- Views:
Transcription
1 Expected Value and Variance This handout, reorganizes the material in. of the text. It also adds to it slightly (Theorem on p.) and presents two results from chapter of the text, because they fit in here and will simplify the presentation of some of the material in chapters and. In addition, the handout contains a motivating discussion that leads up to the definition of expected value. Expected Value.. Motivating Discussion and Definition. Suppose that I play the following game with anyone who is interested in playing: I will charge a player an up-front fee F (to be determined). The player then flips a fair coin three times, and I pay him/her $.00 for each head that (s)he gets and $2.00 for each tail that (s)he gets. Below, I show the random variable Y that logs my payout as a function defined on a sample space as well as the the distribution of Y. Sample Space: Distribution: Sample point s: p(s) : Y(s) : HHH T HH HT H HHT HT T T HT T T H T T T 6 Value y: 6 P(Y = y): The question I want to consider is this: what would be an equitable value for F, the up-front fee? If I overcharge, I will be swindling people, but if I undercharge, I will eventually lose all my savings. In order to be fair, I should take in from each player the amount that, on average, I must pay out to each player: F should be the average payout per game. () The number referred to in box () is the theoretical average of the random variable Y ; it is called the expected value of Y and is denoted EY or µ Y. In order to discover the correct formula for EY, let us see what an µ is the Greek lowercase letter mu; I am certain that it was chosen because m is the first letter in the word mean.
2 estimate of it looks like. Suppose we estimate the average payout per game by repeating the game a huge number of times say 0 9 times, to be definite and then calculating the arithmetic average of the payouts: arithmetic average = y + y y , (2) where the numbers (y, y 2,...) are the amounts of the individual payouts. We now need to simplify (2). replaced by (): The first step is to group like terms together in the numerator, so that (2) is (++ +) + (++ +) + (++ +) + ( ) arithmetic average = 0 9 () (++ +) (++ +) (++ +) ( ) = ( ) ( ) ( ) ( 0 9 ) # of s # of s # of s # of 6 s arithmetic average = () Now, if you look carefully at (), you can discern the theoretical quantity that the arithmetic average is estimating. Since 0 9 is such a big number, we can be sure that ( ) # of s 0 9 P (Y = ), ( ) # of s 0 9 P (Y = ), ( ) # of s 0 9 P (Y = ), ( ) # of 6 s 0 9 and P (Y = 6). Making these approximate substitutions into the right side of () gives us the approximate equation arithmetic average P (Y = ) + P (Y = ) + P (Y = ) + 6P (Y = 6). () We have identified the theoretical average value of this particular random variable: it is the right side of (). The corresponding formula works for any discrete random variable. 2 Definition Let Y be a discrete random variable with distribution given by The expected value of Y (EY or µ Y ) is given by P (Y = y i ) = p i, (i ). EY = µ Y := y P (Y = y ) + y 2 P (Y = y 2 ) + y P (Y = y ) + = i y i P (Y = y i ). (6).2 Basic Properties of Expected Value. There are quite a few of these; together, they make expected value a very user-friendly quantity. The first of these, which is not in the text, simplifies the proofs of some of the others; it shows that E Y, which is defined from the distribution of Y, can also be viewed as a sum over the sample space on which Y is defined. 2 When the random variable has infinitely many values, the sum in (6) can diverge. It is strongly suggested that at this point you compute the expected value of the random variable on p. both from the definition and from formula (7). 2
3 Theorem E Y = Y (s)p(s). (7) Proof. For each i, let E i be the event Y = y i so that P (Y = y i ) = P (E i ). Observe that the events {E, E 2,...} partition S. Starting from the definition, we get: E Y = i y i P (Y = y i ) = y i P (E i ) i (by the definition of P (E i ) ) = ( ) y i p(s) i s E i (because y i = Y (s) for s E i ) = ( ) Y (s) p(s) i s E i (distribute ) = Y (s)p(s) i s E i (because the events {E, E 2,...} partition S ) = Y (s)p(s). Theorem 2 Let X, Y and be random variables defined on the same sample space, and let c and d be constants. a: If X is constant that is, if P (X = x 0 ) = for some x 0 then E X = x 0. b: E cy + d = ce Y + d (or, equivalently: µ cy +d = cµ Y + d.) c: E Y + = E Y + E (or, equivalently: µ Y + = µ Y + µ Y2.) Proof of a. E X = x 0 P (X = x 0 ) = x 0 = x 0. Proof of b. E cy + d = i (cy i + d)p (Y = y i ) = i ( cyi P (Y = y i ) + dp (Y = y i ) ) = i cy i P (Y = y i ) + i dp (Y = y i )) = c y i P (Y = y i ) + d P (Y = y i )) i i = ce Y + d () = ce Y + d. Proof of c. This can be proved more easily using (7) than it can by using the definition of expected value. E Y + = ) (Y + (s)p(s) (definition of Y + ) = ( ) Y (s) + (s) p(s) (distribute ) = Y (s)p(s) + (s)p(s)
4 (group into two sums ) = Y (s)p(s) + (s)p(s) (equation (7), used twice ) = E Y + E. Exercise Show that E Y µ Y = 0.. Independence and Expected Values. The concept of independent random variables is built upon the idea of independent events. Definition 2 Let X and Y be two random variables defined on the same sample space. X and Y are independent iff for every value x i in the distribution of X and every y j in the distribution of Y, the events X = x i and Y = y j are independent events; that is, ) P (X = x i Y = yj = P (X = x i )P (Y = y j ). Many desirable things happen when random variables are independent; in this section, I will focus on one particular one. Theorem If X and Y are independent random variables, then E XY = E X E Y (if X and Y are independent) () Proof. The simplest proof I can find uses both the definition and Theorem. For each value x j of X, let F j = X = x j, and for each value y i of Y, let E i = Y = y i. Note that the events {(F j E i )} partition S. E X E Y = x j P (X = x j ) y i P (Y = y i ) j i = x j P (F j ) y i P (E i ) j i (multiply out ) = x j y i P (F j )P (E i ) j i (because X and Y are independent ) = x j y i P (F j E i ) j i (definition of P (F j E i ) ) = x j y i p(s) j i s F j E i (multiply out ) = x j y i p(s) j i s F j E i (correct for s (F j E i ) ) = X(s)Y (s)p(s) j i s F j E i (because the events{(f j E i )} partition S ) = X(s)Y (s)p(s) (formula (7) ) = E XY. Because they are independent, they are necessarily defined on the same sample space.
5 2 Variance and Standard Deviation. 2. Motivating Discussion and Definition. Just as the expected value µ Y of a random variable Y is analogous to the mean y of a sample (y, y 2,..., y n ), there is a number called the variance of Y that is analogous to the variance s 2 of the sample y, y 2,..., y n. The considerations that lead to the definition of s 2 apply in this case as well:. The goal is to find to measure the spread of the distribution of Y ; that is, to find a number that is small for a tight, compact distribution and large for a widely spread-out distribution. 2. The expected (signed) distance of a random value of Y from the mean µ Y does not work, because E Y µ Y = 0, always (Exercise ).. The expected (unsigned) distance of Y from µ Y, namely E Y µ Y, does work, but it is not particularly easy to work with.. The expected squared distance of Y from µ Y the variance also works, just as #2 does; but unlike #2, the variance has a number of nice mathematical properties, and (probably because of them) it is by far the most commonly used measure of dispersion.. The variance has one undesirable property, namely that its units are the square of the units of Y ; this can be fixed by taking the square root of the variance, to get the so-called standard deviation of Y. Definition Let Y be a discrete random variable.. The variance of Y, denoted V(Y) or σy, 2 is given by := E (Y µ Y ) 2. σ 2 Y 2. The standard deviation of Y, denoted σ Y, is the square root of the variance of Y : σ Y := σ 2 Y = E (Y µ Y ) Basic Properties of the Variance and Standard Deviation. While there are parallels between the properties of variance and those of expected value, there are also differences. Notice in particular that part d of Theorem requires independence, whereas part c of Theorem 2 (p.) does not. Theorem Let X, Y and be random variables defined on the same sample space, and let c and d be constants. a: σ 2 Y = E Y 2 µ 2 Y. b: If X is constant that is, if P (X = x 0 ) = for some x 0 then σ 2 X = 0.6 c: V (cy + d) = c 2 V (Y ) (or, equivalently: σ 2 cy +d = c 2 σ 2 Y.) d: If Y and are independent, V (Y + ) = V (Y )+V ( ) (or, equivalently: σ 2 Y + = σ 2 Y + σ 2.) with the exception of the (n ) in the denominator in the definition of s 2. The reason for that will become clear when you reach Chapter. 6 The converse of b is also true (Exercise ).
6 Proof of a. (by Theorem 2 part c ) = E (by Theorem 2 parts a and b ) = E (because E Y = µ Y ) = E = E σ 2 Y = E (Y µ Y ) 2 = E 2Y µ Y + µ 2 Y + E 2µ Y E 2µ 2 Y + µ 2 Y µ 2 Y. 2Y µ Y + E µ 2 Y Y + µ 2 Y Proof of b. Since P (X = x 0 ) =, it is also true that P (X 2 = x 2 0) =, so that by part a of Theorem 2, E X 2 = x 2 0. Then, by part a of Theorem, Proof of c. σx 2 = E X 2 (µ X ) 2 = x 2 0 (µ X ) 2 = x 2 0 (x 0 ) 2 = 0. V (cy + d) = ( ) 2 E cy + d µ cy +d (by Theorem 2b ) = ( ) 2 E cy + d (cµ Y + d) ( ) 2 (arithmetic ) = E c(y µ Y ) (arithmetic ) = E c 2 (Y µ Y ) 2 (by Theorem 2b ) = c 2 E (Y µ Y ) 2 = c 2 V (Y ). Proof of d. V (Y + ) = ) 2 E (Y + µ Y+Y2 (by Theorem 2c ) = ( ) 2 E Y + (µ Y + µ Y2 ) ( ) 2 (arithmetic ) = E (Y µ Y ) + ( µ Y2 ) (arithmetic ) = E ((Y µ Y ) 2 + 2(Y µ Y )( µ Y2 ) + ( µ Y2 ) 2 ( (by Theorem 2 b,c ) = E (Y µ Y ) 2 +2E (Y µ Y )( µ Y2 ) +E ( µ Y2 ) 2 ( (by Theorem (we have independence) ) = E (Y µ Y ) 2 +2E Y µ Y E µ Y2 +E ( µ Y2 ) 2 ( (by Exercise on p. ) = E (Y µ Y ) 2 +2(0)(0)+E ( µ Y2 ) 2 = V (Y ) + V ( ). Exercise 2 For each part of Theorem, there is a corresponding fact about standard deviations. Write these facts down. Exercise Show that if σ 2 Y = 0, then Y is a constant. 6
Chapter 35 out of 37 from Discrete Mathematics for Neophytes: Number Theory, Probability, Algorithms, and Other Stuff by J. M. Cargal.
35 Mixed Chains In this chapter we learn how to analyze Markov chains that consists of transient and absorbing states. Later we will see that this analysis extends easily to chains with (nonabsorbing)
More information1 Random variables and distributions
Random variables and distributions In this chapter we consider real valued functions, called random variables, defined on the sample space. X : S R X The set of possible values of X is denoted by the set
More informationIntroduction to Probability
Introduction to Probability Salvatore Pace September 2, 208 Introduction In a frequentist interpretation of probability, a probability measure P (A) says that if I do something N times, I should see event
More informationDiscrete Random Variables
Discrete Random Variables We have a probability space (S, Pr). A random variable is a function X : S V (X ) for some set V (X ). In this discussion, we must have V (X ) is the real numbers X induces a
More informationLecture 6 - Random Variables and Parameterized Sample Spaces
Lecture 6 - Random Variables and Parameterized Sample Spaces 6.042 - February 25, 2003 We ve used probablity to model a variety of experiments, games, and tests. Throughout, we have tried to compute probabilities
More informationExpected Value 7/7/2006
Expected Value 7/7/2006 Definition Let X be a numerically-valued discrete random variable with sample space Ω and distribution function m(x). The expected value E(X) is defined by E(X) = x Ω x m(x), provided
More informationP [(E and F )] P [F ]
CONDITIONAL PROBABILITY AND INDEPENDENCE WORKSHEET MTH 1210 This worksheet supplements our textbook material on the concepts of conditional probability and independence. The exercises at the end of each
More informationV. RANDOM VARIABLES, PROBABILITY DISTRIBUTIONS, EXPECTED VALUE
V. RANDOM VARIABLES, PROBABILITY DISTRIBUTIONS, EXPECTED VALUE A game of chance featured at an amusement park is played as follows: You pay $ to play. A penny a nickel are flipped. You win $ if either
More informationWhat is a random variable
OKAN UNIVERSITY FACULTY OF ENGINEERING AND ARCHITECTURE MATH 256 Probability and Random Processes 04 Random Variables Fall 20 Yrd. Doç. Dr. Didem Kivanc Tureli didemk@ieee.org didem.kivanc@okan.edu.tr
More informationWhy should you care?? Intellectual curiosity. Gambling. Mathematically the same as the ESP decision problem we discussed in Week 4.
I. Probability basics (Sections 4.1 and 4.2) Flip a fair (probability of HEADS is 1/2) coin ten times. What is the probability of getting exactly 5 HEADS? What is the probability of getting exactly 10
More informationChapter 8: An Introduction to Probability and Statistics
Course S3, 200 07 Chapter 8: An Introduction to Probability and Statistics This material is covered in the book: Erwin Kreyszig, Advanced Engineering Mathematics (9th edition) Chapter 24 (not including
More informationTOPIC 12: RANDOM VARIABLES AND THEIR DISTRIBUTIONS
TOPIC : RANDOM VARIABLES AND THEIR DISTRIBUTIONS In the last section we compared the length of the longest run in the data for various players to our expectations for the longest run in data generated
More informationProbability Theory and Simulation Methods
Feb 28th, 2018 Lecture 10: Random variables Countdown to midterm (March 21st): 28 days Week 1 Chapter 1: Axioms of probability Week 2 Chapter 3: Conditional probability and independence Week 4 Chapters
More informationExam 1 Review With Solutions Instructor: Brian Powers
Exam Review With Solutions Instructor: Brian Powers STAT 8, Spr5 Chapter. In how many ways can 5 different trees be planted in a row? 5P 5 = 5! =. ( How many subsets of S = {,,,..., } contain elements?
More informationLecture 8: Continuous random variables, expectation and variance
Lecture 8: Continuous random variables, expectation and variance Lejla Batina Institute for Computing and Information Sciences Digital Security Version: autumn 2013 Lejla Batina Version: autumn 2013 Wiskunde
More informationDept. of Linguistics, Indiana University Fall 2015
L645 Dept. of Linguistics, Indiana University Fall 2015 1 / 34 To start out the course, we need to know something about statistics and This is only an introduction; for a fuller understanding, you would
More informationFirst Digit Tally Marks Final Count
Benford Test () Imagine that you are a forensic accountant, presented with the two data sets on this sheet of paper (front and back). Which of the two sets should be investigated further? Why? () () ()
More informationA birthday in St. Petersburg
Mathematical Assoc of America College Mathematics Journal 45: July 29, 207 2:27 pm StPetersburgtex page A birthday in St Petersburg Bio to be included if accepted for publication Consider the following
More informationSolution to HW 12. Since B and B 2 form a partition, we have P (A) = P (A B 1 )P (B 1 ) + P (A B 2 )P (B 2 ). Using P (A) = 21.
Solution to HW 12 (1) (10 pts) Sec 12.3 Problem A screening test for a disease shows a positive result in 92% of all cases when the disease is actually present and in 7% of all cases when it is not. Assume
More informationWhat is Probability? Probability. Sample Spaces and Events. Simple Event
What is Probability? Probability Peter Lo Probability is the numerical measure of likelihood that the event will occur. Simple Event Joint Event Compound Event Lies between 0 & 1 Sum of events is 1 1.5
More informationSociology 6Z03 Topic 10: Probability (Part I)
Sociology 6Z03 Topic 10: Probability (Part I) John Fox McMaster University Fall 2014 John Fox (McMaster University) Soc 6Z03: Probability I Fall 2014 1 / 29 Outline: Probability (Part I) Introduction Probability
More informationChapter 14. From Randomness to Probability. Copyright 2012, 2008, 2005 Pearson Education, Inc.
Chapter 14 From Randomness to Probability Copyright 2012, 2008, 2005 Pearson Education, Inc. Dealing with Random Phenomena A random phenomenon is a situation in which we know what outcomes could happen,
More informationNotes 1 Autumn Sample space, events. S is the number of elements in the set S.)
MAS 108 Probability I Notes 1 Autumn 2005 Sample space, events The general setting is: We perform an experiment which can have a number of different outcomes. The sample space is the set of all possible
More informationStatistics 251: Statistical Methods
Statistics 251: Statistical Methods Probability Module 3 2018 file:///volumes/users/r/renaes/documents/classes/lectures/251301/renae/markdown/master%20versions/module3.html#1 1/33 Terminology probability:
More informationMAS275 Probability Modelling Exercises
MAS75 Probability Modelling Exercises Note: these questions are intended to be of variable difficulty. In particular: Questions or part questions labelled (*) are intended to be a bit more challenging.
More information5 CORRELATION. Expectation of the Binomial Distribution I The Binomial distribution can be defined as: P(X = r) = p r q n r where p + q = 1 and 0 r n
5 CORRELATION The covariance of two random variables gives some measure of their independence. A second way of assessing the measure of independence will be discussed shortly but first the expectation
More informationThe topics in this section concern with the first course objective.
1.1 Systems & Probability The topics in this section concern with the first course objective. A system is one of the most fundamental concepts and one of the most useful and powerful tools in STEM (science,
More information4/17/2012. NE ( ) # of ways an event can happen NS ( ) # of events in the sample space
I. Vocabulary: A. Outcomes: the things that can happen in a probability experiment B. Sample Space (S): all possible outcomes C. Event (E): one outcome D. Probability of an Event (P(E)): the likelihood
More informationCS 361: Probability & Statistics
February 19, 2018 CS 361: Probability & Statistics Random variables Markov s inequality This theorem says that for any random variable X and any value a, we have A random variable is unlikely to have an
More informationProbability: Sets, Sample Spaces, Events
Probability: Sets, Sample Spaces, Events Engineering Statistics Section 2.1 Josh Engwer TTU 01 February 2016 Josh Engwer (TTU) Probability: Sets, Sample Spaces, Events 01 February 2016 1 / 29 The Need
More information18.05 Exam 1. Table of normal probabilities: The last page of the exam contains a table of standard normal cdf values.
Name 18.05 Exam 1 No books or calculators. You may have one 4 6 notecard with any information you like on it. 6 problems, 8 pages Use the back side of each page if you need more space. Simplifying expressions:
More informationDiscrete Structures for Computer Science
Discrete Structures for Computer Science William Garrison bill@cs.pitt.edu 6311 Sennott Square Lecture #24: Probability Theory Based on materials developed by Dr. Adam Lee Not all events are equally likely
More informationChapter 2.5 Random Variables and Probability The Modern View (cont.)
Chapter 2.5 Random Variables and Probability The Modern View (cont.) I. Statistical Independence A crucially important idea in probability and statistics is the concept of statistical independence. Suppose
More information2011 Pearson Education, Inc
Statistics for Business and Economics Chapter 3 Probability Contents 1. Events, Sample Spaces, and Probability 2. Unions and Intersections 3. Complementary Events 4. The Additive Rule and Mutually Exclusive
More informationPropagation of Probabilistic Uncertainty: The Simplest Case (A Brief Pedagogical Introduction)
University of Texas at El Paso DigitalCommons@UTEP Departmental Technical Reports (CS) Department of Computer Science 11-2017 Propagation of Probabilistic Uncertainty: The Simplest Case (A Brief Pedagogical
More informationProblems from Probability and Statistical Inference (9th ed.) by Hogg, Tanis and Zimmerman.
Math 224 Fall 2017 Homework 1 Drew Armstrong Problems from Probability and Statistical Inference (9th ed.) by Hogg, Tanis and Zimmerman. Section 1.1, Exercises 4,5,6,7,9,12. Solutions to Book Problems.
More informationChapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables
Chapter 2 Some Basic Probability Concepts 2.1 Experiments, Outcomes and Random Variables A random variable is a variable whose value is unknown until it is observed. The value of a random variable results
More informationNotes slides from before lecture. CSE 21, Winter 2017, Section A00. Lecture 16 Notes. Class URL:
Notes slides from before lecture CSE 21, Winter 2017, Section A00 Lecture 16 Notes Class URL: http://vlsicad.ucsd.edu/courses/cse21-w17/ Notes slides from before lecture Notes March 8 (1) This week: Days
More informationNotice how similar the answers are in i,ii,iii, and iv. Go back and modify your answers so that all the parts look almost identical.
RANDOM VARIABLES MATH CIRCLE (ADVANCED) 3/3/2013 0) a) Suppose you flip a fair coin 3 times. i) What is the probability you get 0 heads? ii) 1 head? iii) 2 heads? iv) 3 heads? b) Suppose you are dealt
More informationRandom variables (discrete)
Random variables (discrete) Saad Mneimneh 1 Introducing random variables A random variable is a mapping from the sample space to the real line. We usually denote the random variable by X, and a value that
More informationLecture Lecture 5
Lecture 4 --- Lecture 5 A. Basic Concepts (4.1-4.2) 1. Experiment: A process of observing a phenomenon that has variation in its outcome. Examples: (E1). Rolling a die, (E2). Drawing a card form a shuffled
More informationCSC Discrete Math I, Spring Discrete Probability
CSC 125 - Discrete Math I, Spring 2017 Discrete Probability Probability of an Event Pierre-Simon Laplace s classical theory of probability: Definition of terms: An experiment is a procedure that yields
More informationProbability and random variables
Probability and random variables Events A simple event is the outcome of an experiment. For example, the experiment of tossing a coin twice has four possible outcomes: HH, HT, TH, TT. A compound event
More informationOutline. Probability. Math 143. Department of Mathematics and Statistics Calvin College. Spring 2010
Outline Math 143 Department of Mathematics and Statistics Calvin College Spring 2010 Outline Outline 1 Review Basics Random Variables Mean, Variance and Standard Deviation of Random Variables 2 More Review
More informationConditional Probability, Independence and Bayes Theorem Class 3, Jeremy Orloff and Jonathan Bloom
Conditional Probability, Independence and Bayes Theorem Class 3, 18.05 Jeremy Orloff and Jonathan Bloom 1 Learning Goals 1. Know the definitions of conditional probability and independence of events. 2.
More informationMODULE 2 RANDOM VARIABLE AND ITS DISTRIBUTION LECTURES DISTRIBUTION FUNCTION AND ITS PROPERTIES
MODULE 2 RANDOM VARIABLE AND ITS DISTRIBUTION LECTURES 7-11 Topics 2.1 RANDOM VARIABLE 2.2 INDUCED PROBABILITY MEASURE 2.3 DISTRIBUTION FUNCTION AND ITS PROPERTIES 2.4 TYPES OF RANDOM VARIABLES: DISCRETE,
More informationArkansas Tech University MATH 3513: Applied Statistics I Dr. Marcel B. Finan
2.4 Random Variables Arkansas Tech University MATH 3513: Applied Statistics I Dr. Marcel B. Finan By definition, a random variable X is a function with domain the sample space and range a subset of the
More informationPart (A): Review of Probability [Statistics I revision]
Part (A): Review of Probability [Statistics I revision] 1 Definition of Probability 1.1 Experiment An experiment is any procedure whose outcome is uncertain ffl toss a coin ffl throw a die ffl buy a lottery
More informationEECS 126 Probability and Random Processes University of California, Berkeley: Spring 2015 Abhay Parekh February 17, 2015.
EECS 126 Probability and Random Processes University of California, Berkeley: Spring 2015 Abhay Parekh February 17, 2015 Midterm Exam Last name First name SID Rules. You have 80 mins (5:10pm - 6:30pm)
More informationEcon 371 Problem Set #1 Answer Sheet
Econ 371 Problem Set #1 Answer Sheet 2.1 In this question, you are asked to consider the random variable Y, which denotes the number of heads that occur when two coins are tossed. a. The first part of
More informationIAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES
IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES VARIABLE Studying the behavior of random variables, and more importantly functions of random variables is essential for both the
More informationrandom variables T T T T H T H H
random variables T T T T H T H H random variables A random variable X assigns a real number to each outcome in a probability space. Ex. Let H be the number of Heads when 20 coins are tossed Let T be the
More informationSTAT 285 Fall Assignment 1 Solutions
STAT 285 Fall 2014 Assignment 1 Solutions 1. An environmental agency sets a standard of 200 ppb for the concentration of cadmium in a lake. The concentration of cadmium in one lake is measured 17 times.
More informationRandom Variables. Statistics 110. Summer Copyright c 2006 by Mark E. Irwin
Random Variables Statistics 110 Summer 2006 Copyright c 2006 by Mark E. Irwin Random Variables A Random Variable (RV) is a response of a random phenomenon which is numeric. Examples: 1. Roll a die twice
More informationExam #1 March 9, 2016
Mathematics 1372/D552, Spring 2016 Instructor: Suman Ganguli Exam #1 March 9, 2016 Name: 1. (20 points) The following data set lists the high temperatures (in degrees Fahrenheit) observed in Central Park
More informationDiscrete Random Variable
Discrete Random Variable Outcome of a random experiment need not to be a number. We are generally interested in some measurement or numerical attribute of the outcome, rather than the outcome itself. n
More informationChapter 4: An Introduction to Probability and Statistics
Chapter 4: An Introduction to Probability and Statistics 4. Probability The simplest kinds of probabilities to understand are reflected in everyday ideas like these: (i) if you toss a coin, the probability
More informationLecture 6: The Pigeonhole Principle and Probability Spaces
Lecture 6: The Pigeonhole Principle and Probability Spaces Anup Rao January 17, 2018 We discuss the pigeonhole principle and probability spaces. Pigeonhole Principle The pigeonhole principle is an extremely
More informationCS4705. Probability Review and Naïve Bayes. Slides from Dragomir Radev
CS4705 Probability Review and Naïve Bayes Slides from Dragomir Radev Classification using a Generative Approach Previously on NLP discriminative models P C D here is a line with all the social media posts
More informationProbability and random variables. Sept 2018
Probability and random variables Sept 2018 2 The sample space Consider an experiment with an uncertain outcome. The set of all possible outcomes is called the sample space. Example: I toss a coin twice,
More information2. Variance and Covariance: We will now derive some classic properties of variance and covariance. Assume real-valued random variables X and Y.
CS450 Final Review Problems Fall 08 Solutions or worked answers provided Problems -6 are based on the midterm review Identical problems are marked recap] Please consult previous recitations and textbook
More informationClass 8 Review Problems 18.05, Spring 2014
1 Counting and Probability Class 8 Review Problems 18.05, Spring 2014 1. (a) How many ways can you arrange the letters in the word STATISTICS? (e.g. SSSTTTIIAC counts as one arrangement.) (b) If all arrangements
More informationToday. Statistical Learning. Coin Flip. Coin Flip. Experiment 1: Heads. Experiment 1: Heads. Which coin will I use? Which coin will I use?
Today Statistical Learning Parameter Estimation: Maximum Likelihood (ML) Maximum A Posteriori (MAP) Bayesian Continuous case Learning Parameters for a Bayesian Network Naive Bayes Maximum Likelihood estimates
More informationMassachusetts Institute of Technology Department of Electrical Engineering & Computer Science 6.041/6.431: Probabilistic Systems Analysis
6.04/6.43: Probabilistic Systems Analysis Question : Multiple choice questions. CLEARLY circle the best answer for each question below. Each question is worth 4 points each, with no partial credit given.
More informationManipulating Equations
Manipulating Equations Now that you know how to set up an equation, the next thing you need to do is solve for the value that the question asks for. Above all, the most important thing to remember when
More informationDiscrete Mathematics and Probability Theory Fall 2012 Vazirani Note 14. Random Variables: Distribution and Expectation
CS 70 Discrete Mathematics and Probability Theory Fall 202 Vazirani Note 4 Random Variables: Distribution and Expectation Random Variables Question: The homeworks of 20 students are collected in, randomly
More informationRandom Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R
In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample
More informationBusiness Statistics. Lecture 3: Random Variables and the Normal Distribution
Business Statistics Lecture 3: Random Variables and the Normal Distribution 1 Goals for this Lecture A little bit of probability Random variables The normal distribution 2 Probability vs. Statistics Probability:
More informationTo find the median, find the 40 th quartile and the 70 th quartile (which are easily found at y=1 and y=2, respectively). Then we interpolate:
Joel Anderson ST 37-002 Lecture Summary for 2/5/20 Homework 0 First, the definition of a probability mass function p(x) and a cumulative distribution function F(x) is reviewed: Graphically, the drawings
More informationTotal. Name: Student ID: CSE 21A. Midterm #2. February 28, 2013
Name: Student ID: CSE 21A Midterm #2 February 28, 2013 There are 6 problems. The number of points a problem is worth is shown next to the problem. Show your work (even on multiple choice questions)! Also,
More informationMonty Hall Puzzle. Draw a tree diagram of possible choices (a possibility tree ) One for each strategy switch or no-switch
Monty Hall Puzzle Example: You are asked to select one of the three doors to open. There is a large prize behind one of the doors and if you select that door, you win the prize. After you select a door,
More informationMATH 151, FINAL EXAM Winter Quarter, 21 March, 2014
Time: 3 hours, 8:3-11:3 Instructions: MATH 151, FINAL EXAM Winter Quarter, 21 March, 214 (1) Write your name in blue-book provided and sign that you agree to abide by the honor code. (2) The exam consists
More informationSTEP Support Programme. Statistics STEP Questions: Solutions
STEP Support Programme Statistics STEP Questions: Solutions 200 S Q2 Preparation (i) (a) The sum of the probabilities is, so we have k + 2k + 3k + 4k k 0. (b) P(X 3) P(X 3) + P(X 4) 7 0. (c) E(X) 0 ( +
More informationOutline. 1. Define likelihood 2. Interpretations of likelihoods 3. Likelihood plots 4. Maximum likelihood 5. Likelihood ratio benchmarks
Outline 1. Define likelihood 2. Interpretations of likelihoods 3. Likelihood plots 4. Maximum likelihood 5. Likelihood ratio benchmarks Likelihood A common and fruitful approach to statistics is to assume
More informationIntroduction to Probability, Fall 2009
Introduction to Probability, Fall 2009 Math 30530 Review questions for exam 1 solutions 1. Let A, B and C be events. Some of the following statements are always true, and some are not. For those that are
More informationk P (X = k)
Math 224 Spring 208 Homework Drew Armstrong. Suppose that a fair coin is flipped 6 times in sequence and let X be the number of heads that show up. Draw Pascal s triangle down to the sixth row (recall
More informationCS 124 Math Review Section January 29, 2018
CS 124 Math Review Section CS 124 is more math intensive than most of the introductory courses in the department. You re going to need to be able to do two things: 1. Perform some clever calculations to
More information1 INFO Sep 05
Events A 1,...A n are said to be mutually independent if for all subsets S {1,..., n}, p( i S A i ) = p(a i ). (For example, flip a coin N times, then the events {A i = i th flip is heads} are mutually
More information8 Laws of large numbers
8 Laws of large numbers 8.1 Introduction We first start with the idea of standardizing a random variable. Let X be a random variable with mean µ and variance σ 2. Then Z = (X µ)/σ will be a random variable
More information324 Stat Lecture Notes (1) Probability
324 Stat Lecture Notes 1 robability Chapter 2 of the book pg 35-71 1 Definitions: Sample Space: Is the set of all possible outcomes of a statistical experiment, which is denoted by the symbol S Notes:
More informationMATH Solutions to Probability Exercises
MATH 5 9 MATH 5 9 Problem. Suppose we flip a fair coin once and observe either T for tails or H for heads. Let X denote the random variable that equals when we observe tails and equals when we observe
More information1 4 3 A Scalar Multiplication
1 Matrices A matrix is a rectangular array of variables or constants in horizontal rows and vertical columns, usually enclosed in brackets. In a matrix, the numbers or data are organized so that each position
More informationMATH 19B FINAL EXAM PROBABILITY REVIEW PROBLEMS SPRING, 2010
MATH 9B FINAL EXAM PROBABILITY REVIEW PROBLEMS SPRING, 00 This handout is meant to provide a collection of exercises that use the material from the probability and statistics portion of the course The
More informationRecitation 6. Randomization. 6.1 Announcements. RandomLab has been released, and is due Monday, October 2. It s worth 100 points.
Recitation 6 Randomization 6.1 Announcements RandomLab has been released, and is due Monday, October 2. It s worth 100 points. FingerLab will be released after Exam I, which is going to be on Wednesday,
More informationCS206 Review Sheet 3 October 24, 2018
CS206 Review Sheet 3 October 24, 2018 After ourintense focusoncounting, wecontinue withthestudyofsomemoreofthebasic notions from Probability (though counting will remain in our thoughts). An important
More informationProbability: Terminology and Examples Class 2, Jeremy Orloff and Jonathan Bloom
1 Learning Goals Probability: Terminology and Examples Class 2, 18.05 Jeremy Orloff and Jonathan Bloom 1. Know the definitions of sample space, event and probability function. 2. Be able to organize a
More informationSimultaneous Equations Solve for x and y (What are the values of x and y): Summation What is the value of the following given x = j + 1. x i.
1 Algebra Simultaneous Equations Solve for x and y (What are the values of x and y): x + 2y = 6 x - y = 3 Summation What is the value of the following given x = j + 1. Summation Calculate the following:
More informationProbability (Devore Chapter Two)
Probability (Devore Chapter Two) 1016-345-01: Probability and Statistics for Engineers Fall 2012 Contents 0 Administrata 2 0.1 Outline....................................... 3 1 Axiomatic Probability 3
More informationMath 180B Homework 4 Solutions
Math 80B Homework 4 Solutions Note: We will make repeated use of the following result. Lemma. Let (X n ) be a time-homogeneous Markov chain with countable state space S, let A S, and let T = inf { n 0
More informationp. 4-1 Random Variables
Random Variables A Motivating Example Experiment: Sample k students without replacement from the population of all n students (labeled as 1, 2,, n, respectively) in our class. = {all combinations} = {{i
More informationExpectation, inequalities and laws of large numbers
Chapter 3 Expectation, inequalities and laws of large numbers 3. Expectation and Variance Indicator random variable Let us suppose that the event A partitions the sample space S, i.e. A A S. The indicator
More informationCombinatorial Analysis
Chapter 1 Combinatorial Analysis STAT 302, Department of Statistics, UBC 1 A starting example: coin tossing Consider the following random experiment: tossing a fair coin twice There are four possible outcomes,
More informationEdexcel past paper questions
Edexcel past paper questions Statistics 1 Discrete Random Variables Past examination questions Discrete Random variables Page 1 Discrete random variables Discrete Random variables Page 2 Discrete Random
More informationHomework 4 Solution, due July 23
Homework 4 Solution, due July 23 Random Variables Problem 1. Let X be the random number on a die: from 1 to. (i) What is the distribution of X? (ii) Calculate EX. (iii) Calculate EX 2. (iv) Calculate Var
More informationLecture 3 Probability Basics
Lecture 3 Probability Basics Thais Paiva STA 111 - Summer 2013 Term II July 3, 2013 Lecture Plan 1 Definitions of probability 2 Rules of probability 3 Conditional probability What is Probability? Probability
More informationName: Practice Final Exam May 8, 2012
Math 00 Finite Math Practice Final Exam May 8, 0 Name: Be sure that you have all 7 pages of the test. The exam lasts for hours. The Honor Code is in effect for this examination, including keeping your
More informationDynamic Programming Lecture #4
Dynamic Programming Lecture #4 Outline: Probability Review Probability space Conditional probability Total probability Bayes rule Independent events Conditional independence Mutual independence Probability
More informationIntroduction to discrete probability. The rules Sample space (finite except for one example)
Algorithms lecture notes 1 Introduction to discrete probability The rules Sample space (finite except for one example) say Ω. P (Ω) = 1, P ( ) = 0. If the items in the sample space are {x 1,..., x n }
More information1 Normal Distribution.
Normal Distribution.. Introduction A Bernoulli trial is simple random experiment that ends in success or failure. A Bernoulli trial can be used to make a new random experiment by repeating the Bernoulli
More informationNotes 6 Autumn Example (One die: part 1) One fair six-sided die is thrown. X is the number showing.
MAS 08 Probability I Notes Autumn 005 Random variables A probability space is a sample space S together with a probability function P which satisfies Kolmogorov s aioms. The Holy Roman Empire was, in the
More information