CS145: Probability & Computing
|
|
- Madeline Cunningham
- 5 years ago
- Views:
Transcription
1 CS45: Probability & Computing Lecture 5: Concentration Inequalities, Law of Large Numbers, Central Limit Theorem Instructor: Eli Upfal Brown University Computer Science Figure credits: Bertsekas & Tsitsiklis, Introduction to Probability, 2008 Pitman, Probability, 999
2 CS45: Lecture 5 Outline Ø Markov and Chebyshev Inequalities Ø Law of Large Numbers Ø Central Limit Theorem
3 E[X] = Theorem Markov s Inequality [Markov Inequality] For any non-negative random variable, and for all a > 0, Pr(X a) apple E[X ] a. Z 0 xf X (x) dx Z a xf X (x) dx Z a af X (x) dx = ap (X a) a f X (x)
4 Characterizing Random Variables Suppose we care about some summary of random variable X: Y = g(x) Ø Expected Value: Value of a Z typical sample E[Y ]=E[g(X)] = g(x)f X (x) dx Ø Variance: Expected squared distance from mean Var[Y ]=E[(g(X) E[g(X)]) 2 ] x Ø Cumulative distribution: A complete characterization of g(x) F Y (y) =P (Y apple y) =P (g(x) apple y) Often tractable to compute, & leads to bounds on extreme events! f Y (y) = df Y (y) dy If g(x) is complex or high-dimensional, may be hard to compute!
5 Theorem Chebyshev s Inequality For any random variable X,andanya > 0, Proof. Pr( X E[X ] a) apple Var[X ] a 2. By Markov inequality Pr( X E[X ] a) =Pr((X E[X ]) 2 a 2 ) Pr((X E[X ]) 2 a 2 ) apple E[(X E[X ])2 ] a 2 = Var[X ] a 2
6 Theorem Chebyshev s Inequality For any random variable X,andanya > 0, Pr( X E[X ] a) apple Var[X ] a 2. Ø Another way of parameterizing Chebyshev s inequality: µ = E[X], = p Var[X] P ( X µ k ) apple k 2 Ø Chebyshev bound is vacuous (above one) for events less than one standard deviation from the mean. But this could be likely! k 2
7 Chebyshev s Inequality Ø Another way of parameterizing Chebyshev s inequality: E(X) µ = E[X], = p Var[X] P ( X µ k ) apple k 2 Ø Chebyshev bound is vacuous (above one) for events less than one standard deviation from the mean. But this could be likely! k 2
8 . Markov vs. Chebyshev s Inequalities
9 CS45: Lecture 5 Outline Ø Markov and Chebyshev Inequalities Ø Law of Large Numbers Ø Central Limit Theorem
10 Convergence in Probability Convergence in Probability Let Y,Y 2,... be a sequence of random variables (not necessarily independent), and let a be a real number. We say that the sequence Y n converges to a in probability, if for every ϵ > 0, we have lim n P( Y n a ϵ ) =0. (almost all) of the PMF/PDF of Y n eventually gets concentrated (arbitrarily) close to a Convergence of a Deterministic Sequence Let a,a 2,... be a sequence of real numbers, and let a be another real number. We say that the sequence a n converges to a, or lim n a n = a, if for every ϵ > 0 there exists some n 0 such that a n a ϵ, for all n n 0. a n eventually gets and stays (arbitrarily) close to a
11 Convergence in Probability Convergence in Probability Let Y,Y 2,... be a sequence of random variables (not necessarily independent), and let a be a real number. We say that the sequence Y n converges to a in probability, if for every ϵ > 0, we have lim n P( Y n a ϵ ) =0. Example: X n is a sequence of independent uniform variables on [0, ] and Y n =min{x,...,x n } Ø We expect that Y n converges to zero. To verify: P ( Y n 0 ϵ ) = P(X ϵ,...,x n ϵ) = P(X ϵ) P(X n ϵ) = ( ϵ) n.
12 The (Weak) Law of Large Numbers X,X 2,... i.i.d. finite mean µ and variance 2 E[M n ]= Var[M n ]= E[X ]+ + E[X n ] = nµ n n = µ, M = n var(x + + X n ) = var(x )+ +var(x n ) n 2 n 2 X + + X n n = nσ2 n 2 Var( M 2 n) P( M n µ ) = 2 2 n = σ2 n. Ø Chebyshev s inequality bounds distance between the true mean and the empirical or sample mean: sample mean or empirical mean Ø The empirical mean converges to the true mean in probability lim n! P ( M n µ ) =0 Ø True even if variance not finite, but proof more challenging.
13 Why is it a Weak Law of Large Numbers? Convergence in Probability Let Y,Y 2,... be a sequence of random variables (not necessarily independent), and let a be a real number. We say that the sequence Y n converges to a in probability, if for every ϵ > 0, we have lim n P( Y n a ϵ ) =0. Example: - /n pmf of Y n /n 0 n For every > 0, lim n! P ( Y n 0 ) =0. But even though Y n converges in probability, occasionally it takes on very large values: E[Y n ] = for all n.
14 The Strong Law of Large Numbers The Strong Law of Large Numbers (SLLN) Let X,X 2,... be a sequence of independent identically distributed random variables with mean µ. Then, the sequence of sample means M n =(X + + X n )/n converges to µ, with probability, in the sense that ( P lim n X + + X n n ) = µ =. Ø This stronger (but more technically challenging) notion of convergence rules out cases like the previous example Ø For many practical scenarios, both forms of convergence hold, but convergence in probability is easier to show Ø We focus exclusively on the weak law in this course
15 CS45: Lecture 5 Outline Ø Markov and Chebyshev Inequalities Ø Law of Large Numbers Ø Central Limit Theorem
16 What is the shape of the distribution of Z n for large n? Scaling of the Sample Mean Ø Sequence of independent, identically distributed random variables: X,X 2,...,X n E[X i ]=µ Var[X i ]= 2 < Ø The variance of their sum increases with n: S n = P n i= X i E[S n ]=nµ Var[S n ]=n 2 Ø Law of Large Numbers: variance of the empirical mean decreases with n: M n = n S n E[M n ]=µ Var[M n ]= 2 n Ø Standardized sum: transform so mean and variance constant for all n p Var[Sn ] Z n = S n E[S n ] = S n nµ p n E[Z n ]=0 Var[Z n ]=
17 Central Limit Theorem (CLT) Standardized S n = X + + X n : S n E[S n ] S n Z n = = zero mean S n ne[x] n Theorem: For every c: P(Z n c) P(Z c) unit variance Let Z be a standard normal r.v. (zero mean, unit variance) f Z (z) = p 2 e z2 2 P(Z c) is the standard normal CDF, (c), available from the normal tables
18 Central Limit Theorem (CLT) accurate computational shortcut justification of Usefulness normal models What exactly does it say? universal; only Usefulness means, variances matter universal; accurate CDF of Z computational only n converges to means, variances shortcut normal CDF matter accurate justification notwhat a statement computational exactly of normal about does models convergence shortcut it say? of justification CDF PDFs of or Z PMFs n of converges normal models to normal CDF not a statement about convergence of PDFs What or PMFs exactly does it say? CDF ofnormal Z n converges approximation to normal CDF Treat not azstatement n as if normal about convergence of PDFs also treat Normal PMFs S n as approximation if normal Treat Z n as if normal also treat S n as if normal Normal approximation Can Treat wez n use as it if normal when n is moderate? Yes, alsobut treat nos n nice as if theorems normal to this e ect Can Symmetry we use helps it when a lot n is moderate? Yes, but no nice theorems to this e ect Symmetry helps a lot Can we use it when n is moderate? Theorem: For every c: P(Z n c) P(Z c) P(Z c) is the standard normal CDF, (c), available from the normal tables
19 CLT: Uniform Random Variables n=l o n=2... o C "iii c () Cl Joint density of (X + Y, W) o 2 I t t ",, " n=3 o 2 3 n=4 o n=5 S n = P n i= X i Den ity of X + Y f Xi (x i )=if0apple x i apple, 0 otherwise. o
20 .0 CLT: Exponential Random Variables f Xi (x) = e x,x 0. E[X Var[X i ]= i ]= 2 n= n=2 n=3 S n = P n i= X i n=0 0.0 o time in multiples of /,\
21 CLT: Intuitions for Proof Ø Theorem: A sum of independent Gaussians is Gaussian f Xi (x i )= p e 2 S = P n i i= X i 2 2 i f S (s) = p 2 2 e 2( s µ ) 2 µ = nx µ i, 2 = i= nx i= x µi 2 Ø A family of distributions is stable if a linear combination of independent random variables stays in the same family Ø Theorem: The only (non-degenerate) stable family of random variables with finite variance is the Gaussian. Ø One proof: Take Taylor expansion of log-pdf (or its Fourier transform), show that all terms of order three or higher approach zero as n grows, giving a quadratic (Gaussian) limit 2 i
22 CLT: Binomial Distribution Fix p, where 0 <p< X i : Bernoulli(p) S n = X + + X n : Binomial(n, p) mean np, variance np( p) CDF of q S n np standard normal np( p) k l k l De Moivre Laplace Approximation to the Binomial If S n is a binomial random variable with parameters n and p, n is large, and k, l are nonnegative integers, then P(k S n l) Φ ( l + 2 np ) ( k 2 Φ np ). np( p) np( p) P(S n 2) = P(S n < 22), because S n is integer Compromise: consider P(S n 2.5)
23 CLT: Binomial Distribution When the /2 correction is used, CLT can also approximate the binomial p.m.f. (not just the binomial CDF) P(S n = 9) = P(8.5 S n 9.5) n = 36,p= S n S n Z n 0.5 P(S n = 9) P(0.7 Z 0.5) = P(Z 0.5) P(Z 0.7) = = 0.24 k l k l De Moivre Laplace Approximation to the Binomial If S n is a binomial random variable with parameters n and p, n is large, and k, l are nonnegative integers, then P(k S n l) Φ ( l + 2 np ) ( k 2 Φ np ). np( p) np( p) P(S n 2) = P(S n < 22), because S n is integer Compromise: consider P(S n 2.5) Exact answer: =
24 Pollster s Problem: Chebyshev f: fraction of population that... M n =(X + + X n )/n fraction of yes in our sample ith (randomly selected) person polled: Goal: 95% confidence of % error, if yes, X i = 0, if no. P( M n f.0).05 Use Chebyshev s inequality: P( M n M 2 n f.0) (0.0) 2 = 2 x n(0.0) 2 4n(0.0) 2 For any binary variable, Var(X i ) apple 2 2 If n = 50, 000, then P( M n f.0).05 (conservative)
25 M n =(X + + X n )/n < Suppose we want: : f: fraction of population that... ith (randomly selected) person polled:, if yes, X i = 0, if no. Event of interest: M n f.0 X + + X n n Pollster s Problem: CLT P( Mn f.0).05 Event of interest: M n f.0 X + + X n n nf nf. 0 X + + X n nf.0 n n. 0 X + + X n nf.0 n n M n =(X + + X n )/n fraction of yes in our sample Goal: 95% confidence of % error P( M n f.0).05 For any binary variable, Var(X i ) apple 2 2 Std(X i ) apple 2 P( M n f.0) P( Z.0 n/ ) P( Z.02 n) p n = = 00,n= 0, 000
Introduction to Probability
LECTURE NOTES Course 6.041-6.431 M.I.T. FALL 2000 Introduction to Probability Dimitri P. Bertsekas and John N. Tsitsiklis Professors of Electrical Engineering and Computer Science Massachusetts Institute
More informationLecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN
Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and
More informationLecture 7: Chapter 7. Sums of Random Variables and Long-Term Averages
Lecture 7: Chapter 7. Sums of Random Variables and Long-Term Averages ELEC206 Probability and Random Processes, Fall 2014 Gil-Jin Jang gjang@knu.ac.kr School of EE, KNU page 1 / 15 Chapter 7. Sums of Random
More informationCS145: Probability & Computing Lecture 11: Derived Distributions, Functions of Random Variables
CS145: Probability & Computing Lecture 11: Derived Distributions, Functions of Random Variables Instructor: Erik Sudderth Brown University Computer Science March 5, 2015 Homework Submissions Electronic
More informationCOMPSCI 240: Reasoning Under Uncertainty
COMPSCI 240: Reasoning Under Uncertainty Andrew Lan and Nic Herndon University of Massachusetts at Amherst Spring 2019 Lecture 20: Central limit theorem & The strong law of large numbers Markov and Chebyshev
More information6.1 Moment Generating and Characteristic Functions
Chapter 6 Limit Theorems The power statistics can mostly be seen when there is a large collection of data points and we are interested in understanding the macro state of the system, e.g., the average,
More informationQuick Tour of Basic Probability Theory and Linear Algebra
Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra CS224w: Social and Information Network Analysis Fall 2011 Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra Outline Definitions
More informationSDS 321: Introduction to Probability and Statistics
SDS 321: Introduction to Probability and Statistics Lecture 14: Continuous random variables Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin www.cs.cmu.edu/
More informationChapter 4. Chapter 4 sections
Chapter 4 sections 4.1 Expectation 4.2 Properties of Expectations 4.3 Variance 4.4 Moments 4.5 The Mean and the Median 4.6 Covariance and Correlation 4.7 Conditional Expectation SKIP: 4.8 Utility Expectation
More informationTwelfth Problem Assignment
EECS 401 Not Graded PROBLEM 1 Let X 1, X 2,... be a sequence of independent random variables that are uniformly distributed between 0 and 1. Consider a sequence defined by (a) Y n = max(x 1, X 2,..., X
More informationProbability Review. Yutian Li. January 18, Stanford University. Yutian Li (Stanford University) Probability Review January 18, / 27
Probability Review Yutian Li Stanford University January 18, 2018 Yutian Li (Stanford University) Probability Review January 18, 2018 1 / 27 Outline 1 Elements of probability 2 Random variables 3 Multiple
More informationMAS113 Introduction to Probability and Statistics
MAS113 Introduction to Probability and Statistics School of Mathematics and Statistics, University of Sheffield 2018 19 Identically distributed Suppose we have n random variables X 1, X 2,..., X n. Identically
More informationSTAT Chapter 5 Continuous Distributions
STAT 270 - Chapter 5 Continuous Distributions June 27, 2012 Shirin Golchi () STAT270 June 27, 2012 1 / 59 Continuous rv s Definition: X is a continuous rv if it takes values in an interval, i.e., range
More informationCS145: Probability & Computing
CS45: Probability & Computing Lecture 0: Continuous Bayes Rule, Joint and Marginal Probability Densities Instructor: Eli Upfal Brown University Computer Science Figure credits: Bertsekas & Tsitsiklis,
More informationNorthwestern University Department of Electrical Engineering and Computer Science
Northwestern University Department of Electrical Engineering and Computer Science EECS 454: Modeling and Analysis of Communication Networks Spring 2008 Probability Review As discussed in Lecture 1, probability
More informationLimiting Distributions
Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the
More informationLECTURE 18: Inequalities, convergence, and the Weak Law of Large Numbers. Inequalities
LECTURE 18: Inequalities, convergence, and the Weak Law of Large Numbers Inequalities bound P(X > a) based on limited information about a distribution Markov inequality (based on the mean) Chebyshev inequality
More informationA Brief Analysis of Central Limit Theorem. SIAM Chapter Florida State University
1 / 36 A Brief Analysis of Central Limit Theorem Omid Khanmohamadi (okhanmoh@math.fsu.edu) Diego Hernán Díaz Martínez (ddiazmar@math.fsu.edu) Tony Wills (twills@math.fsu.edu) Kouadio David Yao (kyao@math.fsu.edu)
More informationLimiting Distributions
We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the two fundamental results
More informationFunctions of Several Random Variables (Ch. 5.5)
(Ch. 5.5) Iowa State University Mar 7, 2013 Iowa State University Mar 7, 2013 1 / 37 Outline Iowa State University Mar 7, 2013 2 / 37 several random variables We often consider functions of random variables
More informationProving the central limit theorem
SOR3012: Stochastic Processes Proving the central limit theorem Gareth Tribello March 3, 2019 1 Purpose In the lectures and exercises we have learnt about the law of large numbers and the central limit
More informationContinuous Random Variables and Continuous Distributions
Continuous Random Variables and Continuous Distributions Continuous Random Variables and Continuous Distributions Expectation & Variance of Continuous Random Variables ( 5.2) The Uniform Random Variable
More informationAnalysis of Engineering and Scientific Data. Semester
Analysis of Engineering and Scientific Data Semester 1 2019 Sabrina Streipert s.streipert@uq.edu.au Example: Draw a random number from the interval of real numbers [1, 3]. Let X represent the number. Each
More informationProbability Models. 4. What is the definition of the expectation of a discrete random variable?
1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions
More informationMathematical Statistics 1 Math A 6330
Mathematical Statistics 1 Math A 6330 Chapter 3 Common Families of Distributions Mohamed I. Riffi Department of Mathematics Islamic University of Gaza September 28, 2015 Outline 1 Subjects of Lecture 04
More informationChapter 3: Random Variables 1
Chapter 3: Random Variables 1 Yunghsiang S. Han Graduate Institute of Communication Engineering, National Taipei University Taiwan E-mail: yshan@mail.ntpu.edu.tw 1 Modified from the lecture notes by Prof.
More informationWeek 9 The Central Limit Theorem and Estimation Concepts
Week 9 and Estimation Concepts Week 9 and Estimation Concepts Week 9 Objectives 1 The Law of Large Numbers and the concept of consistency of averages are introduced. The condition of existence of the population
More informationSUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416)
SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416) D. ARAPURA This is a summary of the essential material covered so far. The final will be cumulative. I ve also included some review problems
More information8 Laws of large numbers
8 Laws of large numbers 8.1 Introduction We first start with the idea of standardizing a random variable. Let X be a random variable with mean µ and variance σ 2. Then Z = (X µ)/σ will be a random variable
More informationContinuous Expectation and Variance, the Law of Large Numbers, and the Central Limit Theorem Spring 2014
Continuous Expectation and Variance, the Law of Large Numbers, and the Central Limit Theorem 18.5 Spring 214.5.4.3.2.1-4 -3-2 -1 1 2 3 4 January 1, 217 1 / 31 Expected value Expected value: measure of
More informationChapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued
Chapter 3 sections Chapter 3 - continued 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions
More informationLecture 11. Probability Theory: an Overveiw
Math 408 - Mathematical Statistics Lecture 11. Probability Theory: an Overveiw February 11, 2013 Konstantin Zuev (USC) Math 408, Lecture 11 February 11, 2013 1 / 24 The starting point in developing the
More informationMath 151. Rumbos Spring Solutions to Review Problems for Exam 3
Math 151. Rumbos Spring 2014 1 Solutions to Review Problems for Exam 3 1. Suppose that a book with n pages contains on average λ misprints per page. What is the probability that there will be at least
More informationMidterm #1. Lecture 10: Joint Distributions and the Law of Large Numbers. Joint Distributions - Example, cont. Joint Distributions - Example
Midterm #1 Midterm 1 Lecture 10: and the Law of Large Numbers Statistics 104 Colin Rundel February 0, 01 Exam will be passed back at the end of class Exam was hard, on the whole the class did well: Mean:
More informationPractice Problem - Skewness of Bernoulli Random Variable. Lecture 7: Joint Distributions and the Law of Large Numbers. Joint Distributions - Example
A little more E(X Practice Problem - Skewness of Bernoulli Random Variable Lecture 7: and the Law of Large Numbers Sta30/Mth30 Colin Rundel February 7, 014 Let X Bern(p We have shown that E(X = p Var(X
More informationECE302 Exam 2 Version A April 21, You must show ALL of your work for full credit. Please leave fractions as fractions, but simplify them, etc.
ECE32 Exam 2 Version A April 21, 214 1 Name: Solution Score: /1 This exam is closed-book. You must show ALL of your work for full credit. Please read the questions carefully. Please check your answers
More information1 Random Variable: Topics
Note: Handouts DO NOT replace the book. In most cases, they only provide a guideline on topics and an intuitive feel. 1 Random Variable: Topics Chap 2, 2.1-2.4 and Chap 3, 3.1-3.3 What is a random variable?
More informationChapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University
Chapter 3, 4 Random Variables ENCS6161 - Probability and Stochastic Processes Concordia University ENCS6161 p.1/47 The Notion of a Random Variable A random variable X is a function that assigns a real
More information4. Distributions of Functions of Random Variables
4. Distributions of Functions of Random Variables Setup: Consider as given the joint distribution of X 1,..., X n (i.e. consider as given f X1,...,X n and F X1,...,X n ) Consider k functions g 1 : R n
More informationA Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.
A Probability Primer A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes. Are you holding all the cards?? Random Events A random event, E,
More informationn px p x (1 p) n x. p x n(n 1)... (n x + 1) x!
Lectures 3-4 jacques@ucsd.edu 7. Classical discrete distributions D. The Poisson Distribution. If a coin with heads probability p is flipped independently n times, then the number of heads is Bin(n, p)
More informationBasic concepts of probability theory
Basic concepts of probability theory Random variable discrete/continuous random variable Transform Z transform, Laplace transform Distribution Geometric, mixed-geometric, Binomial, Poisson, exponential,
More informationChapter 3: Random Variables 1
Chapter 3: Random Variables 1 Yunghsiang S. Han Graduate Institute of Communication Engineering, National Taipei University Taiwan E-mail: yshan@mail.ntpu.edu.tw 1 Modified from the lecture notes by Prof.
More informationBasic concepts of probability theory
Basic concepts of probability theory Random variable discrete/continuous random variable Transform Z transform, Laplace transform Distribution Geometric, mixed-geometric, Binomial, Poisson, exponential,
More informationMathematical Statistics 1 Math A 6330
Mathematical Statistics 1 Math A 6330 Chapter 2 Transformations and Expectations Mohamed I. Riffi Department of Mathematics Islamic University of Gaza September 14, 2015 Outline 1 Distributions of Functions
More informationECE 302 Division 2 Exam 2 Solutions, 11/4/2009.
NAME: ECE 32 Division 2 Exam 2 Solutions, /4/29. You will be required to show your student ID during the exam. This is a closed-book exam. A formula sheet is provided. No calculators are allowed. Total
More informationProbability and Distributions
Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated
More informationExpectation of Random Variables
1 / 19 Expectation of Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 13, 2015 2 / 19 Expectation of Discrete
More informationBasic concepts of probability theory
Basic concepts of probability theory Random variable discrete/continuous random variable Transform Z transform, Laplace transform Distribution Geometric, mixed-geometric, Binomial, Poisson, exponential,
More information18.440: Lecture 19 Normal random variables
18.440 Lecture 19 18.440: Lecture 19 Normal random variables Scott Sheffield MIT Outline Tossing coins Normal random variables Special case of central limit theorem Outline Tossing coins Normal random
More informationFundamental Tools - Probability Theory II
Fundamental Tools - Probability Theory II MSc Financial Mathematics The University of Warwick September 29, 2015 MSc Financial Mathematics Fundamental Tools - Probability Theory II 1 / 22 Measurable random
More informationChapter 6. Convergence. Probability Theory. Four different convergence concepts. Four different convergence concepts. Convergence in probability
Probability Theory Chapter 6 Convergence Four different convergence concepts Let X 1, X 2, be a sequence of (usually dependent) random variables Definition 1.1. X n converges almost surely (a.s.), or with
More informationp. 6-1 Continuous Random Variables p. 6-2
Continuous Random Variables Recall: For discrete random variables, only a finite or countably infinite number of possible values with positive probability (>). Often, there is interest in random variables
More informationRandom Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline.
Random Variables Amappingthattransformstheeventstotherealline. Example 1. Toss a fair coin. Define a random variable X where X is 1 if head appears and X is if tail appears. P (X =)=1/2 P (X =1)=1/2 Example
More informationReview of Probability Theory
Review of Probability Theory Arian Maleki and Tom Do Stanford University Probability theory is the study of uncertainty Through this class, we will be relying on concepts from probability theory for deriving
More informationECE353: Probability and Random Processes. Lecture 7 -Continuous Random Variable
ECE353: Probability and Random Processes Lecture 7 -Continuous Random Variable Xiao Fu School of Electrical Engineering and Computer Science Oregon State University E-mail: xiao.fu@oregonstate.edu Continuous
More informationCSE 312 Final Review: Section AA
CSE 312 TAs December 8, 2011 General Information General Information Comprehensive Midterm General Information Comprehensive Midterm Heavily weighted toward material after the midterm Pre-Midterm Material
More informationChap 2.1 : Random Variables
Chap 2.1 : Random Variables Let Ω be sample space of a probability model, and X a function that maps every ξ Ω, toa unique point x R, the set of real numbers. Since the outcome ξ is not certain, so is
More informationContinuous Random Variables
1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables
More informationTom Salisbury
MATH 2030 3.00MW Elementary Probability Course Notes Part V: Independence of Random Variables, Law of Large Numbers, Central Limit Theorem, Poisson distribution Geometric & Exponential distributions Tom
More informationSTAT2201. Analysis of Engineering & Scientific Data. Unit 3
STAT2201 Analysis of Engineering & Scientific Data Unit 3 Slava Vaisman The University of Queensland School of Mathematics and Physics What we learned in Unit 2 (1) We defined a sample space of a random
More informationChapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued
Chapter 3 sections 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions 3.6 Conditional
More information1 Review of Probability and Distributions
Random variables. A numerically valued function X of an outcome ω from a sample space Ω X : Ω R : ω X(ω) is called a random variable (r.v.), and usually determined by an experiment. We conventionally denote
More information3 Multiple Discrete Random Variables
3 Multiple Discrete Random Variables 3.1 Joint densities Suppose we have a probability space (Ω, F,P) and now we have two discrete random variables X and Y on it. They have probability mass functions f
More informationCDA6530: Performance Models of Computers and Networks. Chapter 2: Review of Practical Random
CDA6530: Performance Models of Computers and Networks Chapter 2: Review of Practical Random Variables Definition Random variable (RV)X (R.V.) X: A function on sample space X: S R Cumulative distribution
More informationthe law of large numbers & the CLT
the law of large numbers & the CLT Probability/Density 0.000 0.005 0.010 0.015 0.020 n = 4 0.0 0.2 0.4 0.6 0.8 1.0 x-bar 1 sums of random variables If X,Y are independent, what is the distribution of Z
More informationMidterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley
Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley Midterm 2: Format 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten
More informationDisjointness and Additivity
Midterm 2: Format Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten
More informationDeterministic. Deterministic data are those can be described by an explicit mathematical relationship
Random data Deterministic Deterministic data are those can be described by an explicit mathematical relationship Deterministic x(t) =X cos r! k m t Non deterministic There is no way to predict an exact
More informationFormulas for probability theory and linear models SF2941
Formulas for probability theory and linear models SF2941 These pages + Appendix 2 of Gut) are permitted as assistance at the exam. 11 maj 2008 Selected formulae of probability Bivariate probability Transforms
More informationProbability reminders
CS246 Winter 204 Mining Massive Data Sets Probability reminders Sammy El Ghazzal selghazz@stanfordedu Disclaimer These notes may contain typos, mistakes or confusing points Please contact the author so
More information7 Random samples and sampling distributions
7 Random samples and sampling distributions 7.1 Introduction - random samples We will use the term experiment in a very general way to refer to some process, procedure or natural phenomena that produces
More informationJoint Probability Distributions and Random Samples (Devore Chapter Five)
Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete
More information6 The normal distribution, the central limit theorem and random samples
6 The normal distribution, the central limit theorem and random samples 6.1 The normal distribution We mentioned the normal (or Gaussian) distribution in Chapter 4. It has density f X (x) = 1 σ 1 2π e
More informationContinuous Random Variables
Continuous Random Variables Recall: For discrete random variables, only a finite or countably infinite number of possible values with positive probability. Often, there is interest in random variables
More informationExample continued. Math 425 Intro to Probability Lecture 37. Example continued. Example
continued : Coin tossing Math 425 Intro to Probability Lecture 37 Kenneth Harris kaharri@umich.edu Department of Mathematics University of Michigan April 8, 2009 Consider a Bernoulli trials process with
More informationCDA5530: Performance Models of Computers and Networks. Chapter 2: Review of Practical Random Variables
CDA5530: Performance Models of Computers and Networks Chapter 2: Review of Practical Random Variables Definition Random variable (R.V.) X: A function on sample space X: S R Cumulative distribution function
More information3 Operations on One Random Variable - Expectation
3 Operations on One Random Variable - Expectation 3.0 INTRODUCTION operations on a random variable Most of these operations are based on a single concept expectation. Even a probability of an event can
More informationChapter 5. Chapter 5 sections
1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More information2 Continuous Random Variables and their Distributions
Name: Discussion-5 1 Introduction - Continuous random variables have a range in the form of Interval on the real number line. Union of non-overlapping intervals on real line. - We also know that for any
More informationcontinuous random variables
continuous random variables continuous random variables Discrete random variable: takes values in a finite or countable set, e.g. X {1,2,..., 6} with equal probability X is positive integer i with probability
More informationP (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n
JOINT DENSITIES - RANDOM VECTORS - REVIEW Joint densities describe probability distributions of a random vector X: an n-dimensional vector of random variables, ie, X = (X 1,, X n ), where all X is are
More informationLecture 10: Normal RV. Lisa Yan July 18, 2018
Lecture 10: Normal RV Lisa Yan July 18, 2018 Announcements Midterm next Tuesday Practice midterm, solutions out on website SCPD students: fill out Google form by today Covers up to and including Friday
More information3. Probability and Statistics
FE661 - Statistical Methods for Financial Engineering 3. Probability and Statistics Jitkomut Songsiri definitions, probability measures conditional expectations correlation and covariance some important
More informationCDA6530: Performance Models of Computers and Networks. Chapter 2: Review of Practical Random Variables
CDA6530: Performance Models of Computers and Networks Chapter 2: Review of Practical Random Variables Two Classes of R.V. Discrete R.V. Bernoulli Binomial Geometric Poisson Continuous R.V. Uniform Exponential,
More informationECE 4400:693 - Information Theory
ECE 4400:693 - Information Theory Dr. Nghi Tran Lecture 8: Differential Entropy Dr. Nghi Tran (ECE-University of Akron) ECE 4400:693 Lecture 1 / 43 Outline 1 Review: Entropy of discrete RVs 2 Differential
More informationBandits, Experts, and Games
Bandits, Experts, and Games CMSC 858G Fall 2016 University of Maryland Intro to Probability* Alex Slivkins Microsoft Research NYC * Many of the slides adopted from Ron Jin and Mohammad Hajiaghayi Outline
More informationSTAT 430/510: Lecture 16
STAT 430/510: Lecture 16 James Piette June 24, 2010 Updates HW4 is up on my website. It is due next Mon. (June 28th). Starting today back at section 6.7 and will begin Ch. 7. Joint Distribution of Functions
More informationHT Introduction. P(X i = x i ) = e λ λ x i
MODS STATISTICS Introduction. HT 2012 Simon Myers, Department of Statistics (and The Wellcome Trust Centre for Human Genetics) myers@stats.ox.ac.uk We will be concerned with the mathematical framework
More informationn! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2
Order statistics Ex. 4. (*. Let independent variables X,..., X n have U(0, distribution. Show that for every x (0,, we have P ( X ( < x and P ( X (n > x as n. Ex. 4.2 (**. By using induction or otherwise,
More informationSection 27. The Central Limit Theorem. Po-Ning Chen, Professor. Institute of Communications Engineering. National Chiao Tung University
Section 27 The Central Limit Theorem Po-Ning Chen, Professor Institute of Communications Engineering National Chiao Tung University Hsin Chu, Taiwan 3000, R.O.C. Identically distributed summands 27- Central
More informationTopic 7: Convergence of Random Variables
Topic 7: Convergence of Ranom Variables Course 003, 2016 Page 0 The Inference Problem So far, our starting point has been a given probability space (S, F, P). We now look at how to generate information
More informationLecture 18: Central Limit Theorem. Lisa Yan August 6, 2018
Lecture 18: Central Limit Theorem Lisa Yan August 6, 2018 Announcements PS5 due today Pain poll PS6 out today Due next Monday 8/13 (1:30pm) (will not be accepted after Wed 8/15) Programming part: Java,
More informationF X (x) = P [X x] = x f X (t)dt. 42 Lebesgue-a.e, to be exact 43 More specifically, if g = f Lebesgue-a.e., then g is also a pdf for X.
10.2 Properties of PDF and CDF for Continuous Random Variables 10.18. The pdf f X is determined only almost everywhere 42. That is, given a pdf f for a random variable X, if we construct a function g by
More informationLecture 2: Review of Basic Probability Theory
ECE 830 Fall 2010 Statistical Signal Processing instructor: R. Nowak, scribe: R. Nowak Lecture 2: Review of Basic Probability Theory Probabilistic models will be used throughout the course to represent
More information1 Review of Probability
1 Review of Probability Random variables are denoted by X, Y, Z, etc. The cumulative distribution function (c.d.f.) of a random variable X is denoted by F (x) = P (X x), < x
More information4 Branching Processes
4 Branching Processes Organise by generations: Discrete time. If P(no offspring) 0 there is a probability that the process will die out. Let X = number of offspring of an individual p(x) = P(X = x) = offspring
More informationCSE 312, 2017 Winter, W.L. Ruzzo. 7. continuous random variables
CSE 312, 2017 Winter, W.L. Ruzzo 7. continuous random variables The new bit continuous random variables Discrete random variable: values in a finite or countable set, e.g. X {1,2,..., 6} with equal probability
More informationEE4601 Communication Systems
EE4601 Communication Systems Week 2 Review of Probability, Important Distributions 0 c 2011, Georgia Institute of Technology (lect2 1) Conditional Probability Consider a sample space that consists of two
More informationMore on Distribution Function
More on Distribution Function The distribution of a random variable X can be determined directly from its cumulative distribution function F X. Theorem: Let X be any random variable, with cumulative distribution
More information