CMPSCI 240: Reasoning Under Uncertainty
|
|
- Homer Evans
- 5 years ago
- Views:
Transcription
1 CMPSCI 240: Reasoning Under Uncertainty Lecture 8 Prof. Hanna Wallach wallach@cs.umass.edu February 16, 2012
2 Reminders Check the course website: ~wallach/courses/s12/cmpsci240/ Second homework is due TOMORROW
3 Recap
4 Last Time: Expectation PMF: p X (x) = P(X =x) = P({X =x})
5 Last Time: Expectation PMF: p X (x) = P(X =x) = P({X =x}) Expectation: E[X ] = x x p X (x)
6 Last Time: Expectation PMF: p X (x) = P(X =x) = P({X =x}) Expectation: E[X ] = x x p X (x) If Y = g(x ) then E[Y ] = x g(x) p X (x)
7 Last Time: Expectation PMF: p X (x) = P(X =x) = P({X =x}) Expectation: E[X ] = x x p X (x) If Y = g(x ) then E[Y ] = x g(x) p X (x) If Y = a X + b then E[Y ] = a E[X ] + b
8 Last Time: Variance Variance: var(x ) = E[(X E[X ]) 2 ] = E[X 2 ] E[X ] 2
9 Last Time: Variance Variance: var(x ) = E[(X E[X ]) 2 ] = E[X 2 ] E[X ] 2 If Y = a X + b then var(y ) = a 2 var(x )
10 Last Time: Variance Variance: var(x ) = E[(X E[X ]) 2 ] = E[X 2 ] E[X ] 2 If Y = a X + b then var(y ) = a 2 var(x ) Standard deviation: σ X = var(x )
11 Last Time: Variance Variance: var(x ) = E[(X E[X ]) 2 ] = E[X 2 ] E[X ] 2 If Y = a X + b then var(y ) = a 2 var(x ) Standard deviation: σ X = var(x ) If Y = a X + b then σ Y = a σ X
12 Last Time: Multiple Random Variables Joint PMF: p X,Y (x, y) = P({X =x} {Y =y})
13 Last Time: Multiple Random Variables Joint PMF: p X,Y (x, y) = P({X =x} {Y =y}) p X (x) = y p X,Y (x, y) and p Y (y) = x p X,Y (x, y)
14 Last Time: Multiple Random Variables Joint PMF: p X,Y (x, y) = P({X =x} {Y =y}) p X (x) = y p X,Y (x, y) and p Y (y) = x p X,Y (x, y) If Z = g(x, Y ) then E[Z] = x y g(x, y) p X,Y (x, y)
15 Last Time: Multiple Random Variables Joint PMF: p X,Y (x, y) = P({X =x} {Y =y}) p X (x) = y p X,Y (x, y) and p Y (y) = x p X,Y (x, y) If Z = g(x, Y ) then E[Z] = x y g(x, y) p X,Y (x, y) If Z = ax + by + c then E[Z] = a E[X ] + b E[Y ] + c
16 Multiple Random Variables (Cont.)
17 Three or More Random Variables The joint PMF of X, Y, and Z is denoted p X,Y,Z where p X,Y,Z (x, y, z) = P({X =x} {Y =y} {Z =z})
18 Three or More Random Variables The joint PMF of X, Y, and Z is denoted p X,Y,Z where p X,Y,Z (x, y, z) = P({X =x} {Y =y} {Z =z}) The expected value of g(x, Y, Z) is E[g(X, Y, Z)] = g(x, y, z) P X,Y,Z (x, y, z) x y z
19 Three or More Random Variables The joint PMF of X, Y, and Z is denoted p X,Y,Z where p X,Y,Z (x, y, z) = P({X =x} {Y =y} {Z =z}) The expected value of g(x, Y, Z) is E[g(X, Y, Z)] = x g(x, y, z) P X,Y,Z (x, y, z) y z E[aX + by + cz + d] = a E[X ] + b E[Y ] + c E[Z] + d
20 Linearity of Expectation In general, the expectation of a sum of random variables is equal to the sum of their expectations, i.e., E[X X n ] = n E[X i ] i=1
21 Linearity of Expectation In general, the expectation of a sum of random variables is equal to the sum of their expectations, i.e., E[X X n ] = n E[X i ] i=1 If X i has the same PMF as X j they are identically distributed: E[X i ] = x x p Xi (x) = x x p Xj (x) = E[X j ]
22 Examples of Three or More Random Variables e.g., an office party decides to do a gift exchange for their annual holiday party. Everyone in the office purchases a gift and places it in a box. Each person is then given a random gift drawn from the box. What is the expected number of people who get back the gift that they purchased?
23 Expectations of Standard Random Variables e.g., if X is a binomial random variable, what is E[X ]?
24 Conditioning
25 Conditioning Conditional PMF of X given Y : denoted by p X Y, where p X Y (x y) = P(X =x Y =y) = P({X =x} {Y =y})
26 Conditioning Conditional PMF of X given Y : denoted by p X Y, where p X Y (x y) = P(X =x Y =y) = P({X =x} {Y =y}) Compute p X Y using the definition of conditional probability: P(A B) = P(A B) P(B) = p X Y (x y) = p X,Y (x, y) p Y (y)
27 Conditioning and the Tabular Representation y 1 y 2 y 3 y 4 x x x Compute conditional PMF for Y given X = x by renormalizing the values in the row for x and conditional PMF for X given Y =y by renormalizing the values in the column for y
28 Conditioning and the Tabular Representation y 1 y 2 y 3 y 4 x x x e.g., to compute p X Y (x 1 y 2 ): p X Y (x 1 y 2 ) = p X,Y (x 1, y 2 ) 3 i=1 p X,Y (x i, y 2 ) = p X,Y (x 1, y 2 ) p Y (y 2 )
29 Calculating Joint and Marginal PMFs Apply the multiplication rule using the underlying events {X =x} {Y =y}, {X =x}, and {Y =y} to give p X,Y (x, y) = p X (x) p Y X (y x) = p Y (y) p X Y (x y)
30 Calculating Joint and Marginal PMFs Apply the multiplication rule using the underlying events {X =x} {Y =y}, {X =x}, and {Y =y} to give p X,Y (x, y) = p X (x) p Y X (y x) = p Y (y) p X Y (x y) Divide-and-conquer approach to calculating marginal PMFs: p X (x) = y p X,Y (x, y) = y p Y (y) p X Y (x y)
31 Examples of Conditioning e.g., if my computer has a virus (which is true with probability 0.2) my antivirus program will detect this virus with probability 0.7. If my computer does not have a virus, my antivirus program will mistakenly detect a virus with probability 0.1. Let X and Y be binary random variables that are equal to 1 if my computer has a virus and the program detects a virus, respectively, and 0 otherwise. What is p X,Y?
32 Conditioning on Events We can condition random variables on events too
33 Conditioning on Events We can condition random variables on events too If X is associated with the same experiment as event A then p X A (k) = P(X =k A) = P({X =k} A)
34 Conditioning on Events We can condition random variables on events too If X is associated with the same experiment as event A then p X A (k) = P(X =k A) = P({X =k} A) e.g., X = sum of two rolls, A = first roll is odd
35 Conditioning on Events We can condition random variables on events too If X is associated with the same experiment as event A then p X A (k) = P(X =k A) = P({X =k} A) e.g., X = sum of two rolls, A = first roll is odd Nothing fancy here this is just notation when in doubt, think about the underlying events and outcomes
36 Examples of Conditioning on Events e.g., a student will take a test repeatedly up to a maximum of n times, each time with a probability p of passing, independent of previous attempts. What is the PMF of the number of attempts, given that the student passes the test?
37 Conditional Expectation Conditional version of expectation: E[X Y =y] = x x p X Y (x y)
38 Conditional Expectation Conditional version of expectation: E[X Y =y] = x x p X Y (x y) For any function g( ) of X, what is E[g(X ) Y =y]?
39 Conditional Expectation Conditional version of expectation: E[X Y =y] = x x p X Y (x y) For any function g( ) of X, what is E[g(X ) Y =y]? E[g(X ) Y =y] = x g(x) p X Y (x y)
40 Conditional Expectation Conditional version of expectation: E[X Y =y] = x x p X Y (x y) For any function g( ) of X, what is E[g(X ) Y =y]? E[g(X ) Y =y] = x g(x) p X Y (x y) Expectations can be conditioned on events too
41 Conditional Variance and Standard Deviation Conditional variance of X given Y =y: var(x Y =y) = E[X 2 Y =y] E[X Y =y] 2
42 Conditional Variance and Standard Deviation Conditional variance of X given Y =y: var(x Y =y) = E[X 2 Y =y] E[X Y =y] 2 What is the conditional standard deviation of X given Y =y?
43 Conditional Variance and Standard Deviation Conditional variance of X given Y =y: var(x Y =y) = E[X 2 Y =y] E[X Y =y] 2 What is the conditional standard deviation of X given Y =y? σ X Y=y = var(x Y =y)
44 Conditional Variance and Standard Deviation Conditional variance of X given Y =y: var(x Y =y) = E[X 2 Y =y] E[X Y =y] 2 What is the conditional standard deviation of X given Y =y? σ X Y=y = var(x Y =y) Both can be conditioned on events too
45 Total Expectation Theorem Total expectation theorem: for any random variables X and Y E[X ] = y p Y (y) E[X Y =y]
46 Total Expectation Theorem Total expectation theorem: for any random variables X and Y E[X ] = y p Y (y) E[X Y =y] For any disjoint events A 1,..., A n that partition Ω: E[X ] = n P(A i ) E[X A i ] i=1
47 Expectations of Standard Random Variables e.g., if X is a geometric random variable, what is E[X ]?
48 Independence
49 Independence of Events Two events A and B are independent if and only if P(A B) = P(A) P(B)
50 Independence of Events Two events A and B are independent if and only if P(A B) = P(A) P(B) If P(B) > 0 this is equivalent to P(A B) P(B) = P(A B) = P(A)
51 Independence of Events Two events A and B are independent if and only if P(A B) = P(A) P(B) If P(B) > 0 this is equivalent to P(A B) P(B) = P(A B) = P(A) If P(A) > 0 this is equivalent to P(B A) = P(B)
52 Independence of Random Variables X and Y are independent if and only if for all x and y p X,Y (x, y) = p X (x) p Y (y)
53 Independence of Random Variables X and Y are independent if and only if for all x and y p X,Y (x, y) = p X (x) p Y (y) Equivalently if for all x and y such that p Y (y) > 0 p X,Y (x, y) p Y (y) = p X Y (x y) = p X (x)
54 Independence of Random Variables X and Y are independent if and only if for all x and y p X,Y (x, y) = p X (x) p Y (y) Equivalently if for all x and y such that p Y (y) > 0 p X,Y (x, y) p Y (y) = p X Y (x y) = p X (x) Or if p Y X (y x) = p Y (y) for all x and y such that p X (x) > 0
55 Independence and the Tabular Representation y 1 y 2 y 3 y 4 x x x Are X and Y independent? How can we tell?
56 Expectation and Independence If X and Y are independent then E[XY ] = x y p X,Y (x, y) x y = x y p X (x) p Y (y) = E[X ] E[Y ] x y
57 Expectation and Independence If X and Y are independent then E[XY ] = x y p X,Y (x, y) x y = x y p X (x) p Y (y) = E[X ] E[Y ] x y This is only true if X and Y are independent
58 Independence and Three or More Random Variables X, Y, and Z are independent if and only if for all x, y, and z p X,Y,Z (x, y, z) = p X (x) p Y (y) p Z (z)
59 Independence and Three or More Random Variables X, Y, and Z are independent if and only if for all x, y, and z p X,Y,Z (x, y, z) = p X (x) p Y (y) p Z (z) If random variables X, Y, and Z are independent then so are random variables of the form f (X ), g(y ), and h(z)
60 Variance and Independence If X 1, X 2,..., X n are independent then var(x 1 + X X n ) = n var(x i ) i=1
61 Variance and Independence If X 1, X 2,..., X n are independent then var(x 1 + X X n ) = n var(x i ) i=1 If X 1, X 2,..., X n are independent and identically distributed, then var(x 1 ) = var(x 2 ) =... = var(x n ) and var(x 1 + X X n ) = n var(x 1 )
62 For Next Time Check the course website: ~wallach/courses/s12/cmpsci240/ Second homework is due TOMORROW
CMPSCI 240: Reasoning Under Uncertainty
CMPSCI 240: Reasoning Under Uncertainty Lecture 7 Prof. Hanna Wallach wallach@cs.umass.edu February 14, 2012 Reminders Check the course website: http://www.cs.umass.edu/ ~wallach/courses/s12/cmpsci240/
More informationSDS 321: Introduction to Probability and Statistics
SDS 321: Introduction to Probability and Statistics Lecture 13: Expectation and Variance and joint distributions Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin
More informationCMPSCI 240: Reasoning Under Uncertainty
CMPSCI 240: Reasoning Under Uncertainty Lecture 5 Prof. Hanna Wallach wallach@cs.umass.edu February 7, 2012 Reminders Pick up a copy of B&T Check the course website: http://www.cs.umass.edu/ ~wallach/courses/s12/cmpsci240/
More informationCMPSCI 240: Reasoning Under Uncertainty
CMPSCI 240: Reasoning Under Uncertainty Lecture 24 Prof. Hanna Wallach wallach@cs.umass.edu April 24, 2012 Reminders Check the course website: http://www.cs.umass.edu/ ~wallach/courses/s12/cmpsci240/ Eighth
More informationCMPSCI 240: Reasoning Under Uncertainty
CMPSCI 240: Reasoning Under Uncertainty Lecture 2 Prof. Hanna Wallach wallach@cs.umass.edu January 26, 2012 Reminders Pick up a copy of B&T Check the course website: http://www.cs.umass.edu/ ~wallach/courses/s12/cmpsci240/
More information3 Multiple Discrete Random Variables
3 Multiple Discrete Random Variables 3.1 Joint densities Suppose we have a probability space (Ω, F,P) and now we have two discrete random variables X and Y on it. They have probability mass functions f
More informationMATHEMATICS 154, SPRING 2009 PROBABILITY THEORY Outline #11 (Tail-Sum Theorem, Conditional distribution and expectation)
MATHEMATICS 154, SPRING 2009 PROBABILITY THEORY Outline #11 (Tail-Sum Theorem, Conditional distribution and expectation) Last modified: March 7, 2009 Reference: PRP, Sections 3.6 and 3.7. 1. Tail-Sum Theorem
More informationRecitation 2: Probability
Recitation 2: Probability Colin White, Kenny Marino January 23, 2018 Outline Facts about sets Definitions and facts about probability Random Variables and Joint Distributions Characteristics of distributions
More informationNotes 12 Autumn 2005
MAS 08 Probability I Notes Autumn 005 Conditional random variables Remember that the conditional probability of event A given event B is P(A B) P(A B)/P(B). Suppose that X is a discrete random variable.
More informationSDS 321: Introduction to Probability and Statistics
SDS 321: Introduction to Probability and Statistics Lecture 17: Continuous random variables: conditional PDF Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin
More informationRandom Variables. Saravanan Vijayakumaran Department of Electrical Engineering Indian Institute of Technology Bombay
1 / 13 Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay August 8, 2013 2 / 13 Random Variable Definition A real-valued
More informationRandom Variables. Lecture 6: E(X ), Var(X ), & Cov(X, Y ) Random Variables - Vocabulary. Random Variables, cont.
Lecture 6: E(X ), Var(X ), & Cov(X, Y ) Sta230/Mth230 Colin Rundel February 5, 2014 We have been using them for a while now in a variety of forms but it is good to explicitly define what we mean Random
More informationStatistics for Economists Lectures 6 & 7. Asrat Temesgen Stockholm University
Statistics for Economists Lectures 6 & 7 Asrat Temesgen Stockholm University 1 Chapter 4- Bivariate Distributions 41 Distributions of two random variables Definition 41-1: Let X and Y be two random variables
More informationSDS 321: Introduction to Probability and Statistics
SDS 321: Introduction to Probability and Statistics Lecture 10: Expectation and Variance Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin www.cs.cmu.edu/ psarkar/teaching
More informationSTAT 430/510 Probability Lecture 7: Random Variable and Expectation
STAT 430/510 Probability Lecture 7: Random Variable and Expectation Pengyuan (Penelope) Wang June 2, 2011 Review Properties of Probability Conditional Probability The Law of Total Probability Bayes Formula
More informationDiscrete Random Variable
Discrete Random Variable Outcome of a random experiment need not to be a number. We are generally interested in some measurement or numerical attribute of the outcome, rather than the outcome itself. n
More informationExam 1. Problem 1: True or false
Exam 1 Problem 1: True or false We are told that events A and B are conditionally independent, given a third event C, and that P(B C) > 0. For each one of the following statements, decide whether the statement
More informationLecture 11. Probability Theory: an Overveiw
Math 408 - Mathematical Statistics Lecture 11. Probability Theory: an Overveiw February 11, 2013 Konstantin Zuev (USC) Math 408, Lecture 11 February 11, 2013 1 / 24 The starting point in developing the
More informationBivariate Distributions
STAT/MATH 395 A - PROBABILITY II UW Winter Quarter 17 Néhémy Lim Bivariate Distributions 1 Distributions of Two Random Variables Definition 1.1. Let X and Y be two rrvs on probability space (Ω, A, P).
More informationSTAT 430/510: Lecture 16
STAT 430/510: Lecture 16 James Piette June 24, 2010 Updates HW4 is up on my website. It is due next Mon. (June 28th). Starting today back at section 6.7 and will begin Ch. 7. Joint Distribution of Functions
More informationIntroduction to Stochastic Processes
Stat251/551 (Spring 2017) Stochastic Processes Lecture: 1 Introduction to Stochastic Processes Lecturer: Sahand Negahban Scribe: Sahand Negahban 1 Organization Issues We will use canvas as the course webpage.
More informationSDS 321: Introduction to Probability and Statistics
SDS 321: Introduction to Probability and Statistics Lecture 14: Continuous random variables Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin www.cs.cmu.edu/
More informationBivariate distributions
Bivariate distributions 3 th October 017 lecture based on Hogg Tanis Zimmerman: Probability and Statistical Inference (9th ed.) Bivariate Distributions of the Discrete Type The Correlation Coefficient
More informationSTAT2201. Analysis of Engineering & Scientific Data. Unit 3
STAT2201 Analysis of Engineering & Scientific Data Unit 3 Slava Vaisman The University of Queensland School of Mathematics and Physics What we learned in Unit 2 (1) We defined a sample space of a random
More informationMA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems
MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Review of Basic Probability The fundamentals, random variables, probability distributions Probability mass/density functions
More informationMidterm Exam 1 Solution
EECS 126 Probability and Random Processes University of California, Berkeley: Fall 2015 Kannan Ramchandran September 22, 2015 Midterm Exam 1 Solution Last name First name SID Name of student on your left:
More informationProbability Review. Chao Lan
Probability Review Chao Lan Let s start with a single random variable Random Experiment A random experiment has three elements 1. sample space Ω: set of all possible outcomes e.g.,ω={1,2,3,4,5,6} 2. event
More informationAnalysis of Engineering and Scientific Data. Semester
Analysis of Engineering and Scientific Data Semester 1 2019 Sabrina Streipert s.streipert@uq.edu.au Example: Draw a random number from the interval of real numbers [1, 3]. Let X represent the number. Each
More informationMAS113 Introduction to Probability and Statistics. Proofs of theorems
MAS113 Introduction to Probability and Statistics Proofs of theorems Theorem 1 De Morgan s Laws) See MAS110 Theorem 2 M1 By definition, B and A \ B are disjoint, and their union is A So, because m is a
More informationProbabilistic Systems Analysis Spring 2018 Lecture 6. Random Variables: Probability Mass Function and Expectation
EE 178 Probabilistic Systems Analysis Spring 2018 Lecture 6 Random Variables: Probability Mass Function and Expectation Probability Mass Function When we introduce the basic probability model in Note 1,
More informationLecture 2: Repetition of probability theory and statistics
Algorithms for Uncertainty Quantification SS8, IN2345 Tobias Neckel Scientific Computing in Computer Science TUM Lecture 2: Repetition of probability theory and statistics Concept of Building Block: Prerequisites:
More informationContinuous Random Variables
1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables
More informationLecture 13. Poisson Distribution. Text: A Course in Probability by Weiss 5.5. STAT 225 Introduction to Probability Models February 16, 2014
Lecture 13 Text: A Course in Probability by Weiss 5.5 STAT 225 Introduction to Probability Models February 16, 2014 Whitney Huang Purdue University 13.1 Agenda 1 2 3 13.2 Review So far, we have seen discrete
More informationClass 8 Review Problems solutions, 18.05, Spring 2014
Class 8 Review Problems solutions, 8.5, Spring 4 Counting and Probability. (a) Create an arrangement in stages and count the number of possibilities at each stage: ( ) Stage : Choose three of the slots
More informationRandom Variables and Their Distributions
Chapter 3 Random Variables and Their Distributions A random variable (r.v.) is a function that assigns one and only one numerical value to each simple event in an experiment. We will denote r.vs by capital
More information1 Random Variable: Topics
Note: Handouts DO NOT replace the book. In most cases, they only provide a guideline on topics and an intuitive feel. 1 Random Variable: Topics Chap 2, 2.1-2.4 and Chap 3, 3.1-3.3 What is a random variable?
More informationMathematics 426 Robert Gross Homework 9 Answers
Mathematics 4 Robert Gross Homework 9 Answers. Suppose that X is a normal random variable with mean µ and standard deviation σ. Suppose that PX > 9 PX
More informationDiscrete Probability Refresher
ECE 1502 Information Theory Discrete Probability Refresher F. R. Kschischang Dept. of Electrical and Computer Engineering University of Toronto January 13, 1999 revised January 11, 2006 Probability theory
More informationJoint Distribution of Two or More Random Variables
Joint Distribution of Two or More Random Variables Sometimes more than one measurement in the form of random variable is taken on each member of the sample space. In cases like this there will be a few
More informationGOV 2001/ 1002/ Stat E-200 Section 1 Probability Review
GOV 2001/ 1002/ Stat E-200 Section 1 Probability Review Solé Prillaman Harvard University January 28, 2015 1 / 54 LOGISTICS Course Website: j.mp/g2001 lecture notes, videos, announcements Canvas: problem
More informationPCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities
PCMI 207 - Introduction to Random Matrix Theory Handout #2 06.27.207 REVIEW OF PROBABILITY THEORY Chapter - Events and Their Probabilities.. Events as Sets Definition (σ-field). A collection F of subsets
More informationGov Multiple Random Variables
Gov 2000-4. Multiple Random Variables Matthew Blackwell September 29, 2015 Where are we? Where are we going? We described a formal way to talk about uncertain outcomes, probability. We ve talked about
More informationMore on Distribution Function
More on Distribution Function The distribution of a random variable X can be determined directly from its cumulative distribution function F X. Theorem: Let X be any random variable, with cumulative distribution
More informationLecture 10: Bayes' Theorem, Expected Value and Variance Lecturer: Lale Özkahya
BBM 205 Discrete Mathematics Hacettepe University http://web.cs.hacettepe.edu.tr/ bbm205 Lecture 10: Bayes' Theorem, Expected Value and Variance Lecturer: Lale Özkahya Resources: Kenneth Rosen, Discrete
More informationLecture 16. Lectures 1-15 Review
18.440: Lecture 16 Lectures 1-15 Review Scott Sheffield MIT 1 Outline Counting tricks and basic principles of probability Discrete random variables 2 Outline Counting tricks and basic principles of probability
More informationAlgorithms for Uncertainty Quantification
Algorithms for Uncertainty Quantification Tobias Neckel, Ionuț-Gabriel Farcaș Lehrstuhl Informatik V Summer Semester 2017 Lecture 2: Repetition of probability theory and statistics Example: coin flip Example
More informationCMPSCI 240: Reasoning about Uncertainty
CMPSCI 240: Reasoning about Uncertainty Lecture 2: Sets and Events Andrew McGregor University of Massachusetts Last Compiled: January 27, 2017 Outline 1 Recap 2 Experiments and Events 3 Probabilistic Models
More informationIntroduction to Probability and Stocastic Processes - Part I
Introduction to Probability and Stocastic Processes - Part I Lecture 2 Henrik Vie Christensen vie@control.auc.dk Department of Control Engineering Institute of Electronic Systems Aalborg University Denmark
More informationMath 3215 Intro. Probability & Statistics Summer 14. Homework 5: Due 7/3/14
Math 325 Intro. Probability & Statistics Summer Homework 5: Due 7/3/. Let X and Y be continuous random variables with joint/marginal p.d.f. s f(x, y) 2, x y, f (x) 2( x), x, f 2 (y) 2y, y. Find the conditional
More informationProbability and Distributions
Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated
More informationLecture 10. Variance and standard deviation
18.440: Lecture 10 Variance and standard deviation Scott Sheffield MIT 1 Outline Defining variance Examples Properties Decomposition trick 2 Outline Defining variance Examples Properties Decomposition
More informationTopic 3: The Expectation of a Random Variable
Topic 3: The Expectation of a Random Variable Course 003, 2017 Page 0 Expectation of a discrete random variable Definition (Expectation of a discrete r.v.): The expected value (also called the expectation
More informationMidterm #1. Lecture 10: Joint Distributions and the Law of Large Numbers. Joint Distributions - Example, cont. Joint Distributions - Example
Midterm #1 Midterm 1 Lecture 10: and the Law of Large Numbers Statistics 104 Colin Rundel February 0, 01 Exam will be passed back at the end of class Exam was hard, on the whole the class did well: Mean:
More informationRandom Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R
In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample
More informationLectures 22-23: Conditional Expectations
Lectures 22-23: Conditional Expectations 1.) Definitions Let X be an integrable random variable defined on a probability space (Ω, F 0, P ) and let F be a sub-σ-algebra of F 0. Then the conditional expectation
More informationChapter 4. Chapter 4 sections
Chapter 4 sections 4.1 Expectation 4.2 Properties of Expectations 4.3 Variance 4.4 Moments 4.5 The Mean and the Median 4.6 Covariance and Correlation 4.7 Conditional Expectation SKIP: 4.8 Utility Expectation
More informationSTAT Chapter 5 Continuous Distributions
STAT 270 - Chapter 5 Continuous Distributions June 27, 2012 Shirin Golchi () STAT270 June 27, 2012 1 / 59 Continuous rv s Definition: X is a continuous rv if it takes values in an interval, i.e., range
More informationData Mining Techniques. Lecture 3: Probability
Data Mining Techniques CS 6220 - Section 3 - Fall 2016 Lecture 3: Probability Jan-Willem van de Meent (credit: Zhao, CS 229, Bishop) Project Vote 1. Freeform: Develop your own project proposals 30% of
More informationPROBABILITY: CONDITIONING, BAYES THEOREM [DEVORE 2.4]
PROBABILITY: CONDITIONING, BAYES THEOREM [DEVORE 2.4] CONDITIONAL PROBABILITY: Let events E, F be events in the sample space Ω of an experiment. Then: The conditional probability of F given E, P(F E),
More informationProbability. Paul Schrimpf. January 23, Definitions 2. 2 Properties 3
Probability Paul Schrimpf January 23, 2018 Contents 1 Definitions 2 2 Properties 3 3 Random variables 4 3.1 Discrete........................................... 4 3.2 Continuous.........................................
More informationConditional distributions (discrete case)
Conditional distributions (discrete case) The basic idea behind conditional distributions is simple: Suppose (XY) is a jointly-distributed random vector with a discrete joint distribution. Then we can
More informationCHAPTER 4 MATHEMATICAL EXPECTATION. 4.1 Mean of a Random Variable
CHAPTER 4 MATHEMATICAL EXPECTATION 4.1 Mean of a Random Variable The expected value, or mathematical expectation E(X) of a random variable X is the long-run average value of X that would emerge after a
More informationENGG2430A-Homework 2
ENGG3A-Homework Due on Feb 9th,. Independence vs correlation a For each of the following cases, compute the marginal pmfs from the joint pmfs. Explain whether the random variables X and Y are independent,
More informationReview of Basic Probability Theory
Review of Basic Probability Theory James H. Steiger Department of Psychology and Human Development Vanderbilt University James H. Steiger (Vanderbilt University) 1 / 35 Review of Basic Probability Theory
More informationSTAT 414: Introduction to Probability Theory
STAT 414: Introduction to Probability Theory Spring 2016; Homework Assignments Latest updated on April 29, 2016 HW1 (Due on Jan. 21) Chapter 1 Problems 1, 8, 9, 10, 11, 18, 19, 26, 28, 30 Theoretical Exercises
More informationPractice Problem - Skewness of Bernoulli Random Variable. Lecture 7: Joint Distributions and the Law of Large Numbers. Joint Distributions - Example
A little more E(X Practice Problem - Skewness of Bernoulli Random Variable Lecture 7: and the Law of Large Numbers Sta30/Mth30 Colin Rundel February 7, 014 Let X Bern(p We have shown that E(X = p Var(X
More informationMAS113 Introduction to Probability and Statistics. Proofs of theorems
MAS113 Introduction to Probability and Statistics Proofs of theorems Theorem 1 De Morgan s Laws) See MAS110 Theorem 2 M1 By definition, B and A \ B are disjoint, and their union is A So, because m is a
More informationUCSD CSE 21, Spring 2014 [Section B00] Mathematics for Algorithm and System Analysis
UCSD CSE 21, Spring 2014 [Section B00] Mathematics for Algorithm and System Analysis Lecture 10 Class URL: http://vlsicad.ucsd.edu/courses/cse21-s14/ Lecture 10 Notes Midterm Good job overall! = 81; =
More informationa zoo of (discrete) random variables
a zoo of (discrete) random variables 42 uniform random variable Takes each possible value, say {1..n} with equal probability. Say random variable uniform on S Recall envelopes problem on homework... Randomization
More informationProbability. Paul Schrimpf. January 23, UBC Economics 326. Probability. Paul Schrimpf. Definitions. Properties. Random variables.
Probability UBC Economics 326 January 23, 2018 1 2 3 Wooldridge (2013) appendix B Stock and Watson (2009) chapter 2 Linton (2017) chapters 1-5 Abbring (2001) sections 2.1-2.3 Diez, Barr, and Cetinkaya-Rundel
More informationPart IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationThe mean, variance and covariance. (Chs 3.4.1, 3.4.2)
4 The mean, variance and covariance (Chs 3.4.1, 3.4.2) Mean (Expected Value) of X Consider a university having 15,000 students and let X equal the number of courses for which a randomly selected student
More informationDept. of Linguistics, Indiana University Fall 2015
L645 Dept. of Linguistics, Indiana University Fall 2015 1 / 34 To start out the course, we need to know something about statistics and This is only an introduction; for a fuller understanding, you would
More informationP (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n
JOINT DENSITIES - RANDOM VECTORS - REVIEW Joint densities describe probability distributions of a random vector X: an n-dimensional vector of random variables, ie, X = (X 1,, X n ), where all X is are
More informationProbability Review. Yutian Li. January 18, Stanford University. Yutian Li (Stanford University) Probability Review January 18, / 27
Probability Review Yutian Li Stanford University January 18, 2018 Yutian Li (Stanford University) Probability Review January 18, 2018 1 / 27 Outline 1 Elements of probability 2 Random variables 3 Multiple
More informationLecture 14. Text: A Course in Probability by Weiss 5.6. STAT 225 Introduction to Probability Models February 23, Whitney Huang Purdue University
Lecture 14 Text: A Course in Probability by Weiss 5.6 STAT 225 Introduction to Probability Models February 23, 2014 Whitney Huang Purdue University 14.1 Agenda 14.2 Review So far, we have covered Bernoulli
More informationAppendix A : Introduction to Probability and stochastic processes
A-1 Mathematical methods in communication July 5th, 2009 Appendix A : Introduction to Probability and stochastic processes Lecturer: Haim Permuter Scribe: Shai Shapira and Uri Livnat The probability of
More informationEE514A Information Theory I Fall 2013
EE514A Information Theory I Fall 2013 K. Mohan, Prof. J. Bilmes University of Washington, Seattle Department of Electrical Engineering Fall Quarter, 2013 http://j.ee.washington.edu/~bilmes/classes/ee514a_fall_2013/
More informationProbability and Statistics Notes
Probability and Statistics Notes Chapter Five Jesse Crawford Department of Mathematics Tarleton State University Spring 2011 (Tarleton State University) Chapter Five Notes Spring 2011 1 / 37 Outline 1
More informationNotice how similar the answers are in i,ii,iii, and iv. Go back and modify your answers so that all the parts look almost identical.
RANDOM VARIABLES MATH CIRCLE (ADVANCED) 3/3/2013 0) a) Suppose you flip a fair coin 3 times. i) What is the probability you get 0 heads? ii) 1 head? iii) 2 heads? iv) 3 heads? b) Suppose you are dealt
More information18.440: Lecture 26 Conditional expectation
18.440: Lecture 26 Conditional expectation Scott Sheffield MIT 1 Outline Conditional probability distributions Conditional expectation Interpretation and examples 2 Outline Conditional probability distributions
More information1 Basic continuous random variable problems
Name M362K Final Here are problems concerning material from Chapters 5 and 6. To review the other chapters, look over previous practice sheets for the two exams, previous quizzes, previous homeworks and
More informationBrief Review of Probability
Brief Review of Probability Nuno Vasconcelos (Ken Kreutz-Delgado) ECE Department, UCSD Probability Probability theory is a mathematical language to deal with processes or experiments that are non-deterministic
More informationSTOR : Lecture 17. Properties of Expectation - II Limit Theorems
STOR 435.001: Lecture 17 Properties of Expectation - II Limit Theorems Jan Hannig UNC Chapel Hill 1 / 14 Properties of expectation Recall: For two random variables X and Y, conditional distribution of
More informationExpectation of Random Variables
1 / 19 Expectation of Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 13, 2015 2 / 19 Expectation of Discrete
More informationRandom Variables and Expectations
Inside ECOOMICS Random Variables Introduction to Econometrics Random Variables and Expectations A random variable has an outcome that is determined by an experiment and takes on a numerical value. A procedure
More information3. Probability and Statistics
FE661 - Statistical Methods for Financial Engineering 3. Probability and Statistics Jitkomut Songsiri definitions, probability measures conditional expectations correlation and covariance some important
More informationDiscrete Random Variables
CPSC 53 Systems Modeling and Simulation Discrete Random Variables Dr. Anirban Mahanti Department of Computer Science University of Calgary mahanti@cpsc.ucalgary.ca Random Variables A random variable is
More informationSOR201 Solutions to Examples 3
page 0SOR0(00) SOR0 lutions to Examples (a) An outcome is an unordered sample {i,, i n }, a subset of {,, N}, where the i j s are all different The random variable X can take the values n, n +,, N If X
More informationSTAT 430/510: Lecture 10
STAT 430/510: Lecture 10 James Piette June 9, 2010 Updates HW2 is due today! Pick up your HW1 s up in stat dept. There is a box located right when you enter that is labeled "Stat 430 HW1". It ll be out
More informationCSE 312: Foundations of Computing II Random Variables, Linearity of Expectation 4 Solutions
CSE 31: Foundations of Computing II Random Variables, Linearity of Expectation Solutions Review of Main Concepts (a Random Variable (rv: A numeric function X : Ω R of the outcome. (b Range/Support: The
More informationReview of probability
Review of probability Computer Sciences 760 Spring 2014 http://pages.cs.wisc.edu/~dpage/cs760/ Goals for the lecture you should understand the following concepts definition of probability random variables
More informationCovariance. Lecture 20: Covariance / Correlation & General Bivariate Normal. Covariance, cont. Properties of Covariance
Covariance Lecture 0: Covariance / Correlation & General Bivariate Normal Sta30 / Mth 30 We have previously discussed Covariance in relation to the variance of the sum of two random variables Review Lecture
More information18.440: Lecture 28 Lectures Review
18.440: Lecture 28 Lectures 17-27 Review Scott Sheffield MIT 1 Outline Continuous random variables Problems motivated by coin tossing Random variable properties 2 Outline Continuous random variables Problems
More informationCME 106: Review Probability theory
: Probability theory Sven Schmit April 3, 2015 1 Overview In the first half of the course, we covered topics from probability theory. The difference between statistics and probability theory is the following:
More informationJoint Probability Distributions and Random Samples (Devore Chapter Five)
Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete
More informationMAT 271E Probability and Statistics
MAT 71E Probability and Statistics Spring 013 Instructor : Class Meets : Office Hours : Textbook : Supp. Text : İlker Bayram EEB 1103 ibayram@itu.edu.tr 13.30 1.30, Wednesday EEB 5303 10.00 1.00, Wednesday
More informationBasic Probability. Introduction
Basic Probability Introduction The world is an uncertain place. Making predictions about something as seemingly mundane as tomorrow s weather, for example, is actually quite a difficult task. Even with
More informationHomework 4 Solution, due July 23
Homework 4 Solution, due July 23 Random Variables Problem 1. Let X be the random number on a die: from 1 to. (i) What is the distribution of X? (ii) Calculate EX. (iii) Calculate EX 2. (iv) Calculate Var
More informationECE353: Probability and Random Processes. Lecture 7 -Continuous Random Variable
ECE353: Probability and Random Processes Lecture 7 -Continuous Random Variable Xiao Fu School of Electrical Engineering and Computer Science Oregon State University E-mail: xiao.fu@oregonstate.edu Continuous
More information