Twelfth Problem Assignment

Size: px
Start display at page:

Download "Twelfth Problem Assignment"

Transcription

1 EECS 401 Not Graded PROBLEM 1 Let X 1, X 2,... be a sequence of independent random variables that are uniformly distributed between 0 and 1. Consider a sequence defined by (a) Y n = max(x 1, X 2,..., X n ) (b) Y n is the median of the values of X 1, X 2,..., X 2n+1. Determine whether or not the sequence {Y n } converges in probability in each case. (a) Y n 1 in probability. P( Y n 1 > ε) = P(Y n < 1 ε) = (b) Y n 1 2 in probability. n P(X i < 1 ε) = (1 ε) n 0(as n ). P( Y n 0.5 > ε) P(Y n > ε) + P(Y n < 0.5 ε) Each of these terms converges to zero. Consider P(Y n > ε). For the event {Y n > 0.5+ε} to occur, we must have at least n+1 of the random variables X 1, X 2,... X 2n+1 to have a value of ε or larger. Let Z i be a Bernoulli random variable which is equal to 1 if and only if X i 0.5+ε. Then, the event {Y n 0.5+ε} is the same as the event (Z 1 + Z 2n+1 )/(2n + 1) 0.5. Note that P(Z i = 1) = 0.5 ε. By the weak law of large numbers, the sequence (Z 1 + Z 2n+1 )/(2n + 1) converges to 0.5 ε. Therefore (Z1 + Z 2n+1 ) P ε (2n + 1) converges to zero, which implies P(Y n ε) also converges to zero. By a similar argument it can be shown that P Yn 0.5 ε converges to zero, and therefore Y n converges to Not Graded

2 s PROBLEM 2 (a) Show the one sided Chebyshev s inequality: If X is a random variable with mean 0 and finite variance σ 2, the for any a > 0 For any b 0, P(X a) σ2 σ 2 + a 2 P(X a) = P(X + b a + b) E[(X + b)2 ] (a + b) 2 Treat the RHS as a function of b and minimize with respect to b. The minimum is attained at b = E[X2 ] = Var(X). Substituting back, we get a a 2 Var(X) Var(X) + a P(X a) 2 = a + Var(X) a (b) Generalize to the case when E[X] = µ. In particular, show that P(X µ + a) P(X µ a) σ2 σ 2 + a 2 σ2 σ 2 + a 2 Var(X) a 2 + Var(X) Take X to be X µ in part (a), and the first result follows. Take X to be µ X and the second result follows. PROBLEM 3 Many people believe that the daily change of price of a company s stock on the stock market is a random variable with mean 0 and variance σ 2. That is, if Y n represents the price of the stock on the nth day, then Y n = Y n 1 + X n n 1 where X 1, X 2,... are independent and identically distributed random variables with mean 0 and variance σ 2. Suppose that the stock s price today is 100. If σ 2 = 1, what can you say about the probability that the stock s price will exceed 105 after 10 days? Given that Y 0 = 100, we want to find P(Y 10 > 105). Note that Y 10 = Y 0 + X 1 + X X 10. Therefore, E[Y 10 ] = E[Y 0 ] = 100 and Var(Y 10 ) = Var(X 1 ) + 2 Not Graded

3 s Var(X 2 ) + + Var(X 10 ) = 10. Now, using the Chebyshev s inequality P(Y 10 > 105) = P(Y > 5) < P( Y > 5) Var(Y 10) 5 2 = = 0.4 We can get a tighter bound using the one sided Chebyshev s inequality, P(Y 10 > 105) = P(Y 10 > ) Var(Y 10) Var(Y 10 ) = = PROBLEM 4 Consider a Gaussian random variable X, with mean m and variance σ 2 = (m/2) 2. (a) Determine the probability that an experimental value of X is negative. Pr(X < 0) = Pr X m m/2 < 2 = Pr (Z < 2) = Φ ( 2) = 1 Φ (2) = (b) Determine the probability that the sum of four independent experimental values of X is negative. Let Y = X 1 + X 2 + X 3 + X 4. Then, E[Y] = 4m and Var(Y) = m 2. Thus Y 4m Pr(Y < 0) = Pr m < 4 = Φ( 4) = 1 Φ(4) 0 (c) Two independent experimental values of X (X 1, X 2 ) are obtained. Determine the PDF and mean and variance for: (i) ax 1 + bx 2 (ii) ax 1 bx 2 (iii) X 1 X 2 a, b > 0 (i) ax 1 + bx 2. Sum of two Gaussian r.v.s is also a Gaussian r.v. with mean (a + b)m and variance (a 2 + b 2 )m 2 /4. (ii) ax 1 bx 2. Sum of two Gaussian r.v.s is also a Gaussian r.v. with mean (a b)m and variance (a 2 + b 2 )m 2 /4. (iii) X 1 X 2. Let Y = X 1 X 2, then Y is a Gaussian r.v. with mean 0 and variance m 2 /2. Thus the density of Z = Y is Thus, 4 f Z (z) = m /m2 e z2, z > 0 2π 3 Not Graded

4 s and E[Z] = Var(Z) = 0 0 4z m 2π 4z 2 m 2π /m2 2 e z2 dz = m π /m2 2m e z2 dz = m π Γ 1 2 (d) Determine the probability that the product of four independent experimental values of X is negative. Let Y = X 1 X 2 X 3 X 4. Let p be the probability that X i < 0. X m p = Pr(X i < 0) = Pr m/2 < 2 = Φ( 2) = 1 Φ(2) = Y will be negative if either one of X i s is negative or three X i s are negative, that is 4 4 Pr(Y < 0) = p(1 p) 3 + p 3 (1 p) = PROBLEM 5 Consider a Poisson process with mean arrival rate λ = 1 and let X n be the number of arrivals between 0 and n. Does X n /n converge in probability? We know that X n is a Poisson random variable with parameter λn = n. Therefore, E[X n ] = n = var(x n ), which means that E[X n /n] = 1 and var(x n /n) = 1/n. By using the Chebyshev inequality, we have the following for any ɛ > 0: ( ) X n P n 1 ɛ var(x n) ɛ 2 = 1/n ɛ 2 Therefore, taking the limit as n : ( ) lim P X n n n 1 ɛ 1/n lim n ɛ 2 = 0 Since a probability is always greater than or equal to 0, then the above inequality implies that: ( ) lim P X n n n 1 ɛ = 0 Thus, we have shown that X n /n converges in probability to 1. 4 Not Graded

5 s PROBLEM 6 Suppose that the number of units produced daily at factory A is a random variable with mean 20 and standard deviation 3 and the number produced at factory B is a random variable with mean 18 and standard deviation 6. Assuming independence, derive an upper bound for the probability that more units are produced today at factory B than factory A. Let X be the number of units produced by factory A and let Y be the number of units produced by factory B. Then, we have that E[Y X] = E[Y] E[X] = = 2. Also, since Y and X are independent, then: var(y X) = var(y) + var( X) = var(y) + var(x) = = 45. We can use this information to apply the one-sided Chebyshev s inequality and obtain the following upperbound: P (Y X 0) = P (Y X ( 2) 2) var(y X)var(Y x) = Thus,an upperbound for the probability that factory B produces at least as many units as factory B is given by 45/49. PROBLEM 7 Consider the number of 3s which result from 600 tosses of a fair six-sided die. (a) Determine the probability that there are exactly 100 3s, using a form of Stirling s approximation for n! which is very accurate for these values, Pr( 100 3s) = n! 2πe n n n+0.5 ( ) ( 1 6 ) = 600! 100! 500! πe πe πe = 1 2π = Not Graded

6 s (b) Use the Poisson approximation to the binomial PMF to obtain the probability that there are exactly 100 3s. Using the Poisson approximation Pr(100 3s) = ( ) 100 e(600 1/6) 6 100! e ! = = π = e 100 2πe (c) Repeat part (b), using the central limit theorem intelligently. Let X i denote the indicator function of the event that the outcome of the ith toss is a 3. Then X i is a Bernoulli random variable with parameter 1/6. Let K = 100 X i. Then E[K] = 100/6 and Var(K) = 500/36 = 125/9. Thus, Pr(99 K 101) Φ Φ = 2Φ(0.268) 1 = /125 9/125 (d) Use the Chebyshev inequality to find a lower bound on the probability that the number of 3s is between 97 and 103 inclusive, between 90 and 110 inclusive, and between 60 and 140 inclusive. From Chebyshev s inequality, Pr( K 100 a) 1 Var(K) a 2 a Pr (e) Repeat part (d), using the central limit theorem and employing the DeMoivre-Laplace result when it appears relevant. Compare your answers with those obtained above, and comment. 6 Not Graded

7 s Using DeMoivre-Laplace transform, Pr(100 a < K < a) = 2Φ ( (a +.5) 9/125 ) 1 a Pr When we are close to the mean of the sum, central limit theorem gives a better approximation. When we are away from the mean, Chebyshev s inequality gives a better approximation. PROBLEM 8 Student scores on exams are given by a certain instructor have mean 74 and standard deviation 14. This instructor is about to give two exams, one to a class of size 25 and the other to a class of size 64. n Let X i denote the score of student i. LetM n = 1 n X i. Then E[M n ] = 74 and Var(M n ) = 196/n. (a) Approximate the probability that the average test score in the class of size 25 exceeds 80. Pr(M ) 196/25 196/ = (b) Repeat part (a) for the class size of 64. Pr(M ) 196/64 196/ = (c) Approximate the probability that the average test score in the larger class exceeds that of the other class by over 2.2 points. Let M = M 64 M 25. Then E[M] = 0 and Var(M) = Var(M 64 ) + Var(M 25 ) = 196(1/25 + 1/64) Thus 7 Not Graded

8 s Pr(M 2.2) Var(M) Var(M) = (d) Approximate the probability that the average test score in the smaller class exceeds that of the other class by over 2.2 points. Pr(M 2.2) Var(M) Var(M) = PROBLEM 9 The energy of any individual particle in a certain system is an independent random variable with probability density function { f E (e) = 2e 2e e 0 0 e < 0 The total energy in the system is the sum of the energies of the individual particles. Energy of each particle is an exponential random variable with mean 1 2 and variance 1 4. If there are n particles in the system, the total energy K has mean n/2 and variance n/4 (a) If there are 1,600 particles in the system, determine the probability that there are between 780 and 840 energy units in the system. In this case E[K] = 800 and Var(K) = 400. Using the central limit theorem Pr(780 K 840) Φ Φ = Φ(2) Φ( 1) = (b) What is the largest number of particles the system may contain if the probability that its total energy is less than 440 units must be at least Suppose the number of particles in the system is n. Using Markov inequality Pr(K < 440) = 1 Pr(K 440) 1 n Solving for n we get n Hence the largest number of particles that the system may contain is 855. (c) Each particle will escape from the system if its energy exceeds (ln 3)/2 units. If the system originally contained 4,800 particles, what is the probability that at least 1,700 particles will escape? 8 Not Graded

9 s Let Y i be the indicator of the event that particle i escapes. Then Pr(Y i = 1) = ln 3/2 2e 2x dx = 1 3 Let L = 4800 Y i, then L denotes the number of particles that have escaped. E[L] = 4800/3 = 1600 and Var(L) = 4800(1/3)(2/3) = 3200/3 = Using the one sided Chebyshev s inequality, we have Pr(L ) = 8 83 = (d) If there are 10 particles in the system, determine an exact expression for the PDF for the total energy in the system. The total energy of all particles in the system is the sum of 10 exponential random variables. Thus, the PDF is an Erlang distribution. If T denotes the total energy, then 2 10 x 9 e 2x x 0 f T (x) = 9! 0 x < 0 PROBLEM 10 A set of 200 people, consisting of 100 men and 100 women, randomly divided into 100 pairs of 2 each. Give an upper bound to the probability that at most 30 of these pairs will consist of a man and a woman. Number the 100 men from 1 to 100. Let X i be the indicator of the event that man i is paired with a woman, that is { 1 if man i is paired with a woman X i = 0 otherwise Let X denote the number of man-woman pairs. Then X = 100 Also Therefore, E[X i ] = Pr(X i = 1) = X i. Then, E[X i X i ] = Pr(X i = 1, X j = 1) = Pr(X i = 1) Pr(X j = 1 X i = 1) = Not Graded

10 s and 100 E[X] = E[X i ] = Var(X) = Var(X i ) + 2 Cov(X i, X j ) = (a) Use Chebyshev s inequality i<j ( ) [ ] Pr(X 30) Pr( X ) (20.25) (b) Use one sided Chebyshev s inequality Pr(X 30) = Pr(X ) (20.25) (c) Use the Central Limit Theorem Pr(X 30) Φ( = Φ( 4.04) Notice that we are away from the mean of the process, so central limit theorem gives a bad approximation. PROBLEM 11 The number of rabbits in the generation i is N i. Variable N i+1, the number of rabbits in generation i + 1, depends on the random effects of light, heat, water, food, and predator population, as well as on the number N i. The relation is 0.2 j = 1 P Ni+1 (jn i ) = 0.3 j = j = 3 10 Not Graded

11 s This states, for instance, that with probability 0.5 there will be three times as many rabbits in the generation i + 1 as there were in generation i. The rabbit population in generation 1 was 2. Find an approximation to the number of rabbits in generation 12. (Hint: To use the central limit theorem, you must find an expression involving the sum of random variables. Let N 12 denote the population after 12 generations. Then By weak law of large numbers N 12 = 2J 2 J 3 J 12 log N 12 = log 2 + log J 2 + log J 12 E[log N 12 ] = log E[log J] log N 12 log E[log J] = N 12 = 8290 (in probability) (in probability) Therefore there are approximately 8290 rabbits in generation 12 (assuming that rabbits do not die) 11 Not Graded

Eleventh Problem Assignment

Eleventh Problem Assignment EECS April, 27 PROBLEM (2 points) The outcomes of successive flips of a particular coin are dependent and are found to be described fully by the conditional probabilities P(H n+ H n ) = P(T n+ T n ) =

More information

Math Spring Practice for the final Exam.

Math Spring Practice for the final Exam. Math 4 - Spring 8 - Practice for the final Exam.. Let X, Y, Z be three independnet random variables uniformly distributed on [, ]. Let W := X + Y. Compute P(W t) for t. Honors: Compute the CDF function

More information

CS145: Probability & Computing

CS145: Probability & Computing CS45: Probability & Computing Lecture 5: Concentration Inequalities, Law of Large Numbers, Central Limit Theorem Instructor: Eli Upfal Brown University Computer Science Figure credits: Bertsekas & Tsitsiklis,

More information

Math 151. Rumbos Spring Solutions to Review Problems for Exam 3

Math 151. Rumbos Spring Solutions to Review Problems for Exam 3 Math 151. Rumbos Spring 2014 1 Solutions to Review Problems for Exam 3 1. Suppose that a book with n pages contains on average λ misprints per page. What is the probability that there will be at least

More information

Lecture 7: Chapter 7. Sums of Random Variables and Long-Term Averages

Lecture 7: Chapter 7. Sums of Random Variables and Long-Term Averages Lecture 7: Chapter 7. Sums of Random Variables and Long-Term Averages ELEC206 Probability and Random Processes, Fall 2014 Gil-Jin Jang gjang@knu.ac.kr School of EE, KNU page 1 / 15 Chapter 7. Sums of Random

More information

ECE 302 Division 2 Exam 2 Solutions, 11/4/2009.

ECE 302 Division 2 Exam 2 Solutions, 11/4/2009. NAME: ECE 32 Division 2 Exam 2 Solutions, /4/29. You will be required to show your student ID during the exam. This is a closed-book exam. A formula sheet is provided. No calculators are allowed. Total

More information

STAT Chapter 5 Continuous Distributions

STAT Chapter 5 Continuous Distributions STAT 270 - Chapter 5 Continuous Distributions June 27, 2012 Shirin Golchi () STAT270 June 27, 2012 1 / 59 Continuous rv s Definition: X is a continuous rv if it takes values in an interval, i.e., range

More information

Exercises in Probability Theory Paul Jung MA 485/585-1C Fall 2015 based on material of Nikolai Chernov

Exercises in Probability Theory Paul Jung MA 485/585-1C Fall 2015 based on material of Nikolai Chernov Exercises in Probability Theory Paul Jung MA 485/585-1C Fall 2015 based on material of Nikolai Chernov Many of the exercises are taken from two books: R. Durrett, The Essentials of Probability, Duxbury

More information

Practice Questions for Final

Practice Questions for Final Math 39 Practice Questions for Final June. 8th 4 Name : 8. Continuous Probability Models You should know Continuous Random Variables Discrete Probability Distributions Expected Value of Discrete Random

More information

COMPSCI 240: Reasoning Under Uncertainty

COMPSCI 240: Reasoning Under Uncertainty COMPSCI 240: Reasoning Under Uncertainty Andrew Lan and Nic Herndon University of Massachusetts at Amherst Spring 2019 Lecture 20: Central limit theorem & The strong law of large numbers Markov and Chebyshev

More information

Example continued. Math 425 Intro to Probability Lecture 37. Example continued. Example

Example continued. Math 425 Intro to Probability Lecture 37. Example continued. Example continued : Coin tossing Math 425 Intro to Probability Lecture 37 Kenneth Harris kaharri@umich.edu Department of Mathematics University of Michigan April 8, 2009 Consider a Bernoulli trials process with

More information

Chapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University

Chapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University Chapter 3, 4 Random Variables ENCS6161 - Probability and Stochastic Processes Concordia University ENCS6161 p.1/47 The Notion of a Random Variable A random variable X is a function that assigns a real

More information

Introduction to Probability

Introduction to Probability LECTURE NOTES Course 6.041-6.431 M.I.T. FALL 2000 Introduction to Probability Dimitri P. Bertsekas and John N. Tsitsiklis Professors of Electrical Engineering and Computer Science Massachusetts Institute

More information

n(1 p i ) n 1 p i = 1 3 i=1 E(X i p = p i )P(p = p i ) = 1 3 p i = n 3 (p 1 + p 2 + p 3 ). p i i=1 P(X i = 1 p = p i )P(p = p i ) = p1+p2+p3

n(1 p i ) n 1 p i = 1 3 i=1 E(X i p = p i )P(p = p i ) = 1 3 p i = n 3 (p 1 + p 2 + p 3 ). p i i=1 P(X i = 1 p = p i )P(p = p i ) = p1+p2+p3 Introduction to Probability Due:August 8th, 211 Solutions of Final Exam Solve all the problems 1. (15 points) You have three coins, showing Head with probabilities p 1, p 2 and p 3. You perform two different

More information

Midterm Exam 1 Solution

Midterm Exam 1 Solution EECS 126 Probability and Random Processes University of California, Berkeley: Fall 2015 Kannan Ramchandran September 22, 2015 Midterm Exam 1 Solution Last name First name SID Name of student on your left:

More information

Math 151. Rumbos Fall Solutions to Review Problems for Exam 2. Pr(X = 1) = ) = Pr(X = 2) = Pr(X = 3) = p X. (k) =

Math 151. Rumbos Fall Solutions to Review Problems for Exam 2. Pr(X = 1) = ) = Pr(X = 2) = Pr(X = 3) = p X. (k) = Math 5. Rumbos Fall 07 Solutions to Review Problems for Exam. A bowl contains 5 chips of the same size and shape. Two chips are red and the other three are blue. Draw three chips from the bowl at random,

More information

STAT 414: Introduction to Probability Theory

STAT 414: Introduction to Probability Theory STAT 414: Introduction to Probability Theory Spring 2016; Homework Assignments Latest updated on April 29, 2016 HW1 (Due on Jan. 21) Chapter 1 Problems 1, 8, 9, 10, 11, 18, 19, 26, 28, 30 Theoretical Exercises

More information

(Practice Version) Midterm Exam 2

(Practice Version) Midterm Exam 2 EECS 126 Probability and Random Processes University of California, Berkeley: Fall 2014 Kannan Ramchandran November 7, 2014 (Practice Version) Midterm Exam 2 Last name First name SID Rules. DO NOT open

More information

Brief Review of Probability

Brief Review of Probability Maura Department of Economics and Finance Università Tor Vergata Outline 1 Distribution Functions Quantiles and Modes of a Distribution 2 Example 3 Example 4 Distributions Outline Distribution Functions

More information

MATH Solutions to Probability Exercises

MATH Solutions to Probability Exercises MATH 5 9 MATH 5 9 Problem. Suppose we flip a fair coin once and observe either T for tails or H for heads. Let X denote the random variable that equals when we observe tails and equals when we observe

More information

Expectation. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Expectation. DS GA 1002 Statistical and Mathematical Models.   Carlos Fernandez-Granda Expectation DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Aim Describe random variables with a few numbers: mean, variance,

More information

Analysis of Engineering and Scientific Data. Semester

Analysis of Engineering and Scientific Data. Semester Analysis of Engineering and Scientific Data Semester 1 2019 Sabrina Streipert s.streipert@uq.edu.au Example: Draw a random number from the interval of real numbers [1, 3]. Let X represent the number. Each

More information

X = X X n, + X 2

X = X X n, + X 2 CS 70 Discrete Mathematics for CS Fall 2003 Wagner Lecture 22 Variance Question: At each time step, I flip a fair coin. If it comes up Heads, I walk one step to the right; if it comes up Tails, I walk

More information

Random Variables and Their Distributions

Random Variables and Their Distributions Chapter 3 Random Variables and Their Distributions A random variable (r.v.) is a function that assigns one and only one numerical value to each simple event in an experiment. We will denote r.vs by capital

More information

Section 9.1. Expected Values of Sums

Section 9.1. Expected Values of Sums Section 9.1 Expected Values of Sums Theorem 9.1 For any set of random variables X 1,..., X n, the sum W n = X 1 + + X n has expected value E [W n ] = E [X 1 ] + E [X 2 ] + + E [X n ]. Proof: Theorem 9.1

More information

STAT 418: Probability and Stochastic Processes

STAT 418: Probability and Stochastic Processes STAT 418: Probability and Stochastic Processes Spring 2016; Homework Assignments Latest updated on April 29, 2016 HW1 (Due on Jan. 21) Chapter 1 Problems 1, 8, 9, 10, 11, 18, 19, 26, 28, 30 Theoretical

More information

Expectation. DS GA 1002 Probability and Statistics for Data Science. Carlos Fernandez-Granda

Expectation. DS GA 1002 Probability and Statistics for Data Science.   Carlos Fernandez-Granda Expectation DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Aim Describe random variables with a few numbers: mean,

More information

Essentials on the Analysis of Randomized Algorithms

Essentials on the Analysis of Randomized Algorithms Essentials on the Analysis of Randomized Algorithms Dimitris Diochnos Feb 0, 2009 Abstract These notes were written with Monte Carlo algorithms primarily in mind. Topics covered are basic (discrete) random

More information

1. Let X be a random variable with probability density function. 1 x < f(x) = 0 otherwise

1. Let X be a random variable with probability density function. 1 x < f(x) = 0 otherwise Name M36K Final. Let X be a random variable with probability density function { /x x < f(x = 0 otherwise Compute the following. You can leave your answers in integral form. (a ( points Find F X (t = P

More information

ECE302 Spring 2015 HW10 Solutions May 3,

ECE302 Spring 2015 HW10 Solutions May 3, ECE32 Spring 25 HW Solutions May 3, 25 Solutions to HW Note: Most of these solutions were generated by R. D. Yates and D. J. Goodman, the authors of our textbook. I have added comments in italics where

More information

DO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO

DO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO QUESTION BOOKLET EE 26 Spring 2006 Final Exam Wednesday, May 7, 8am am DO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO You have 80 minutes to complete the final. The final consists of five

More information

University of Illinois ECE 313: Final Exam Fall 2014

University of Illinois ECE 313: Final Exam Fall 2014 University of Illinois ECE 313: Final Exam Fall 2014 Monday, December 15, 2014, 7:00 p.m. 10:00 p.m. Sect. B, names A-O, 1013 ECE, names P-Z, 1015 ECE; Section C, names A-L, 1015 ECE; all others 112 Gregory

More information

1 Exercises for lecture 1

1 Exercises for lecture 1 1 Exercises for lecture 1 Exercise 1 a) Show that if F is symmetric with respect to µ, and E( X )

More information

EECS 70 Discrete Mathematics and Probability Theory Fall 2015 Walrand/Rao Final

EECS 70 Discrete Mathematics and Probability Theory Fall 2015 Walrand/Rao Final EECS 70 Discrete Mathematics and Probability Theory Fall 2015 Walrand/Rao Final PRINT Your Name:, (last) SIGN Your Name: (first) PRINT Your Student ID: CIRCLE your exam room: 220 Hearst 230 Hearst 237

More information

Notes for Math 324, Part 19

Notes for Math 324, Part 19 48 Notes for Math 324, Part 9 Chapter 9 Multivariate distributions, covariance Often, we need to consider several random variables at the same time. We have a sample space S and r.v. s X, Y,..., which

More information

STAT/MA 416 Answers Homework 6 November 15, 2007 Solutions by Mark Daniel Ward PROBLEMS

STAT/MA 416 Answers Homework 6 November 15, 2007 Solutions by Mark Daniel Ward PROBLEMS STAT/MA 4 Answers Homework November 5, 27 Solutions by Mark Daniel Ward PROBLEMS Chapter Problems 2a. The mass p, corresponds to neither of the first two balls being white, so p, 8 7 4/39. The mass p,

More information

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

Math 447. Introduction to Probability and Statistics I. Fall 1998.

Math 447. Introduction to Probability and Statistics I. Fall 1998. Math 447. Introduction to Probability and Statistics I. Fall 1998. Schedule: M. W. F.: 08:00-09:30 am. SW 323 Textbook: Introduction to Mathematical Statistics by R. V. Hogg and A. T. Craig, 1995, Fifth

More information

Probability Models. 4. What is the definition of the expectation of a discrete random variable?

Probability Models. 4. What is the definition of the expectation of a discrete random variable? 1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions

More information

Course: ESO-209 Home Work: 1 Instructor: Debasis Kundu

Course: ESO-209 Home Work: 1 Instructor: Debasis Kundu Home Work: 1 1. Describe the sample space when a coin is tossed (a) once, (b) three times, (c) n times, (d) an infinite number of times. 2. A coin is tossed until for the first time the same result appear

More information

Random Variables and Expectations

Random Variables and Expectations Inside ECOOMICS Random Variables Introduction to Econometrics Random Variables and Expectations A random variable has an outcome that is determined by an experiment and takes on a numerical value. A procedure

More information

IE 230 Probability & Statistics in Engineering I. Closed book and notes. 60 minutes.

IE 230 Probability & Statistics in Engineering I. Closed book and notes. 60 minutes. Closed book and notes. 60 minutes. A summary table of some univariate continuous distributions is provided. Four Pages. In this version of the Key, I try to be more complete than necessary to receive full

More information

FINAL EXAM: Monday 8-10am

FINAL EXAM: Monday 8-10am ECE 30: Probabilistic Methods in Electrical and Computer Engineering Fall 016 Instructor: Prof. A. R. Reibman FINAL EXAM: Monday 8-10am Fall 016, TTh 3-4:15pm (December 1, 016) This is a closed book exam.

More information

Question Points Score Total: 76

Question Points Score Total: 76 Math 447 Test 2 March 17, Spring 216 No books, no notes, only SOA-approved calculators. true/false or fill-in-the-blank question. You must show work, unless the question is a Name: Question Points Score

More information

Tail Inequalities. The Chernoff bound works for random variables that are a sum of indicator variables with the same distribution (Bernoulli trials).

Tail Inequalities. The Chernoff bound works for random variables that are a sum of indicator variables with the same distribution (Bernoulli trials). Tail Inequalities William Hunt Lane Department of Computer Science and Electrical Engineering, West Virginia University, Morgantown, WV William.Hunt@mail.wvu.edu Introduction In this chapter, we are interested

More information

ECE 302 Division 1 MWF 10:30-11:20 (Prof. Pollak) Final Exam Solutions, 5/3/2004. Please read the instructions carefully before proceeding.

ECE 302 Division 1 MWF 10:30-11:20 (Prof. Pollak) Final Exam Solutions, 5/3/2004. Please read the instructions carefully before proceeding. NAME: ECE 302 Division MWF 0:30-:20 (Prof. Pollak) Final Exam Solutions, 5/3/2004. Please read the instructions carefully before proceeding. If you are not in Prof. Pollak s section, you may not take this

More information

Continuous RVs. 1. Suppose a random variable X has the following probability density function: π, zero otherwise. f ( x ) = sin x, 0 < x < 2

Continuous RVs. 1. Suppose a random variable X has the following probability density function: π, zero otherwise. f ( x ) = sin x, 0 < x < 2 STAT 4 Exam I Continuous RVs Fall 7 Practice. Suppose a random variable X has the following probability density function: f ( x ) = sin x, < x < π, zero otherwise. a) Find P ( X < 4 π ). b) Find µ = E

More information

Homework 10 (due December 2, 2009)

Homework 10 (due December 2, 2009) Homework (due December, 9) Problem. Let X and Y be independent binomial random variables with parameters (n, p) and (n, p) respectively. Prove that X + Y is a binomial random variable with parameters (n

More information

Midterm #1. Lecture 10: Joint Distributions and the Law of Large Numbers. Joint Distributions - Example, cont. Joint Distributions - Example

Midterm #1. Lecture 10: Joint Distributions and the Law of Large Numbers. Joint Distributions - Example, cont. Joint Distributions - Example Midterm #1 Midterm 1 Lecture 10: and the Law of Large Numbers Statistics 104 Colin Rundel February 0, 01 Exam will be passed back at the end of class Exam was hard, on the whole the class did well: Mean:

More information

Discrete Mathematics for CS Spring 2007 Luca Trevisan Lecture 20

Discrete Mathematics for CS Spring 2007 Luca Trevisan Lecture 20 CS 70 Discrete Mathematics for CS Spring 2007 Luca Trevisan Lecture 20 Today we shall discuss a measure of how close a random variable tends to be to its expectation. But first we need to see how to compute

More information

CS 1538: Introduction to Simulation Homework 1

CS 1538: Introduction to Simulation Homework 1 CS 1538: Introduction to Simulation Homework 1 1. A fair six-sided die is rolled three times. Let X be a random variable that represents the number of unique outcomes in the three tosses. For example,

More information

Chapter 1: Revie of Calculus and Probability

Chapter 1: Revie of Calculus and Probability Chapter 1: Revie of Calculus and Probability Refer to Text Book: Operations Research: Applications and Algorithms By Wayne L. Winston,Ch. 12 Operations Research: An Introduction By Hamdi Taha, Ch. 12 OR441-Dr.Khalid

More information

Lecture 13. Poisson Distribution. Text: A Course in Probability by Weiss 5.5. STAT 225 Introduction to Probability Models February 16, 2014

Lecture 13. Poisson Distribution. Text: A Course in Probability by Weiss 5.5. STAT 225 Introduction to Probability Models February 16, 2014 Lecture 13 Text: A Course in Probability by Weiss 5.5 STAT 225 Introduction to Probability Models February 16, 2014 Whitney Huang Purdue University 13.1 Agenda 1 2 3 13.2 Review So far, we have seen discrete

More information

Continuous distributions

Continuous distributions CHAPTER 7 Continuous distributions 7.. Introduction A r.v. X is said to have a continuous distribution if there exists a nonnegative function f such that P(a X b) = ˆ b a f(x)dx for every a and b. distribution.)

More information

CS280, Spring 2004: Final

CS280, Spring 2004: Final CS280, Spring 2004: Final 1. [4 points] Which of the following relations on {0, 1, 2, 3} is an equivalence relation. (If it is, explain why. If it isn t, explain why not.) Just saying Yes or No with no

More information

Brief Review of Probability

Brief Review of Probability Brief Review of Probability Nuno Vasconcelos (Ken Kreutz-Delgado) ECE Department, UCSD Probability Probability theory is a mathematical language to deal with processes or experiments that are non-deterministic

More information

1 Basic continuous random variable problems

1 Basic continuous random variable problems Name M362K Final Here are problems concerning material from Chapters 5 and 6. To review the other chapters, look over previous practice sheets for the two exams, previous quizzes, previous homeworks and

More information

Topic 3: The Expectation of a Random Variable

Topic 3: The Expectation of a Random Variable Topic 3: The Expectation of a Random Variable Course 003, 2017 Page 0 Expectation of a discrete random variable Definition (Expectation of a discrete r.v.): The expected value (also called the expectation

More information

n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2

n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2 Order statistics Ex. 4. (*. Let independent variables X,..., X n have U(0, distribution. Show that for every x (0,, we have P ( X ( < x and P ( X (n > x as n. Ex. 4.2 (**. By using induction or otherwise,

More information

Three hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER.

Three hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER. Three hours To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER EXTREME VALUES AND FINANCIAL RISK Examiner: Answer QUESTION 1, QUESTION

More information

Final Solutions Fri, June 8

Final Solutions Fri, June 8 EE178: Probabilistic Systems Analysis, Spring 2018 Final Solutions Fri, June 8 1. Small problems (62 points) (a) (8 points) Let X 1, X 2,..., X n be independent random variables uniformly distributed on

More information

Probability reminders

Probability reminders CS246 Winter 204 Mining Massive Data Sets Probability reminders Sammy El Ghazzal selghazz@stanfordedu Disclaimer These notes may contain typos, mistakes or confusing points Please contact the author so

More information

MATH 407 FINAL EXAM May 6, 2011 Prof. Alexander

MATH 407 FINAL EXAM May 6, 2011 Prof. Alexander MATH 407 FINAL EXAM May 6, 2011 Prof. Alexander Problem Points Score 1 22 2 18 Last Name: First Name: USC ID: Signature: 3 20 4 21 5 27 6 18 7 25 8 28 Total 175 Points total 179 but 175 is maximum. This

More information

Statistics 100A Homework 5 Solutions

Statistics 100A Homework 5 Solutions Chapter 5 Statistics 1A Homework 5 Solutions Ryan Rosario 1. Let X be a random variable with probability density function a What is the value of c? fx { c1 x 1 < x < 1 otherwise We know that for fx to

More information

Midterm Exam 2 (Solutions)

Midterm Exam 2 (Solutions) EECS 6 Probability and Random Processes University of California, Berkeley: Spring 07 Kannan Ramchandran March, 07 Midterm Exam (Solutions) Last name First name SID Name of student on your left: Name of

More information

1 Basic continuous random variable problems

1 Basic continuous random variable problems Name M362K Final Here are problems concerning material from Chapters 5 and 6. To review the other chapters, look over previous practice sheets for the two exams, previous quizzes, previous homeworks and

More information

IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES

IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES VARIABLE Studying the behavior of random variables, and more importantly functions of random variables is essential for both the

More information

Normal Distribution. Distribution function and Graphical Representation - pdf - identifying the mean and variance

Normal Distribution. Distribution function and Graphical Representation - pdf - identifying the mean and variance Distribution function and Graphical Representation - pdf - identifying the mean and variance f ( x ) 1 ( ) x e Distribution function and Graphical Representation - pdf - identifying the mean and variance

More information

6.041/6.431 Fall 2010 Final Exam Wednesday, December 15, 9:00AM - 12:00noon.

6.041/6.431 Fall 2010 Final Exam Wednesday, December 15, 9:00AM - 12:00noon. 6.041/6.431 Fall 2010 Final Exam Wednesday, December 15, 9:00AM - 12:00noon. DO NOT TURN THIS PAGE OVER UNTIL YOU ARE TOLD TO DO SO Name: Recitation Instructor: TA: Question Score Out of 1.1 4 1.2 4 1.3

More information

6.1 Moment Generating and Characteristic Functions

6.1 Moment Generating and Characteristic Functions Chapter 6 Limit Theorems The power statistics can mostly be seen when there is a large collection of data points and we are interested in understanding the macro state of the system, e.g., the average,

More information

Closed book and notes. 60 minutes. Cover page and four pages of exam. No calculators.

Closed book and notes. 60 minutes. Cover page and four pages of exam. No calculators. IE 230 Seat # Closed book and notes. 60 minutes. Cover page and four pages of exam. No calculators. Score Exam #3a, Spring 2002 Schmeiser Closed book and notes. 60 minutes. 1. True or false. (for each,

More information

MATH/STAT 3360, Probability Sample Final Examination Model Solutions

MATH/STAT 3360, Probability Sample Final Examination Model Solutions MATH/STAT 3360, Probability Sample Final Examination Model Solutions This Sample examination has more questions than the actual final, in order to cover a wider range of questions. Estimated times are

More information

Basic concepts of probability theory

Basic concepts of probability theory Basic concepts of probability theory Random variable discrete/continuous random variable Transform Z transform, Laplace transform Distribution Geometric, mixed-geometric, Binomial, Poisson, exponential,

More information

EECS 126 Probability and Random Processes University of California, Berkeley: Fall 2014 Kannan Ramchandran November 13, 2014.

EECS 126 Probability and Random Processes University of California, Berkeley: Fall 2014 Kannan Ramchandran November 13, 2014. EECS 126 Probability and Random Processes University of California, Berkeley: Fall 2014 Kannan Ramchandran November 13, 2014 Midterm Exam 2 Last name First name SID Rules. DO NOT open the exam until instructed

More information

Chapter 3: Random Variables 1

Chapter 3: Random Variables 1 Chapter 3: Random Variables 1 Yunghsiang S. Han Graduate Institute of Communication Engineering, National Taipei University Taiwan E-mail: yshan@mail.ntpu.edu.tw 1 Modified from the lecture notes by Prof.

More information

n px p x (1 p) n x. p x n(n 1)... (n x + 1) x!

n px p x (1 p) n x. p x n(n 1)... (n x + 1) x! Lectures 3-4 jacques@ucsd.edu 7. Classical discrete distributions D. The Poisson Distribution. If a coin with heads probability p is flipped independently n times, then the number of heads is Bin(n, p)

More information

Practice Problem - Skewness of Bernoulli Random Variable. Lecture 7: Joint Distributions and the Law of Large Numbers. Joint Distributions - Example

Practice Problem - Skewness of Bernoulli Random Variable. Lecture 7: Joint Distributions and the Law of Large Numbers. Joint Distributions - Example A little more E(X Practice Problem - Skewness of Bernoulli Random Variable Lecture 7: and the Law of Large Numbers Sta30/Mth30 Colin Rundel February 7, 014 Let X Bern(p We have shown that E(X = p Var(X

More information

Review of Probability. CS1538: Introduction to Simulations

Review of Probability. CS1538: Introduction to Simulations Review of Probability CS1538: Introduction to Simulations Probability and Statistics in Simulation Why do we need probability and statistics in simulation? Needed to validate the simulation model Needed

More information

ECE 302, Final 3:20-5:20pm Mon. May 1, WTHR 160 or WTHR 172.

ECE 302, Final 3:20-5:20pm Mon. May 1, WTHR 160 or WTHR 172. ECE 302, Final 3:20-5:20pm Mon. May 1, WTHR 160 or WTHR 172. 1. Enter your name, student ID number, e-mail address, and signature in the space provided on this page, NOW! 2. This is a closed book exam.

More information

Solution: By Markov inequality: P (X > 100) 0.8. By Chebyshev s inequality: P (X > 100) P ( X 80 > 20) 100/20 2 = The second estimate is better.

Solution: By Markov inequality: P (X > 100) 0.8. By Chebyshev s inequality: P (X > 100) P ( X 80 > 20) 100/20 2 = The second estimate is better. MA 485-1E, Probability (Dr Chernov) Final Exam Wed, Dec 12, 2001 Student s name Be sure to show all your work. Each problem is 4 points. Full credit will be given for 9 problems (36 points). You are welcome

More information

IE 230 Probability & Statistics in Engineering I. Closed book and notes. 120 minutes.

IE 230 Probability & Statistics in Engineering I. Closed book and notes. 120 minutes. Closed book and notes. 10 minutes. Two summary tables from the concise notes are attached: Discrete distributions and continuous distributions. Eight Pages. Score _ Final Exam, Fall 1999 Cover Sheet, Page

More information

ECE353: Probability and Random Processes. Lecture 7 -Continuous Random Variable

ECE353: Probability and Random Processes. Lecture 7 -Continuous Random Variable ECE353: Probability and Random Processes Lecture 7 -Continuous Random Variable Xiao Fu School of Electrical Engineering and Computer Science Oregon State University E-mail: xiao.fu@oregonstate.edu Continuous

More information

CDA6530: Performance Models of Computers and Networks. Chapter 2: Review of Practical Random Variables

CDA6530: Performance Models of Computers and Networks. Chapter 2: Review of Practical Random Variables CDA6530: Performance Models of Computers and Networks Chapter 2: Review of Practical Random Variables Two Classes of R.V. Discrete R.V. Bernoulli Binomial Geometric Poisson Continuous R.V. Uniform Exponential,

More information

Expectation of Random Variables

Expectation of Random Variables 1 / 19 Expectation of Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 13, 2015 2 / 19 Expectation of Discrete

More information

Chapter 5 continued. Chapter 5 sections

Chapter 5 continued. Chapter 5 sections Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions

More information

UC Berkeley Department of Electrical Engineering and Computer Sciences. EECS 126: Probability and Random Processes

UC Berkeley Department of Electrical Engineering and Computer Sciences. EECS 126: Probability and Random Processes UC Berkeley Department of Electrical Engineering and Computer Sciences EECS 6: Probability and Random Processes Problem Set 3 Spring 9 Self-Graded Scores Due: February 8, 9 Submit your self-graded scores

More information

ECON Fundamentals of Probability

ECON Fundamentals of Probability ECON 351 - Fundamentals of Probability Maggie Jones 1 / 32 Random Variables A random variable is one that takes on numerical values, i.e. numerical summary of a random outcome e.g., prices, total GDP,

More information

Bandits, Experts, and Games

Bandits, Experts, and Games Bandits, Experts, and Games CMSC 858G Fall 2016 University of Maryland Intro to Probability* Alex Slivkins Microsoft Research NYC * Many of the slides adopted from Ron Jin and Mohammad Hajiaghayi Outline

More information

14.30 Introduction to Statistical Methods in Economics Spring 2009

14.30 Introduction to Statistical Methods in Economics Spring 2009 MIT OpenCourseWare http://ocw.mit.edu 14.30 Introduction to Statistical Methods in Economics Spring 2009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

Chapter 3: Random Variables 1

Chapter 3: Random Variables 1 Chapter 3: Random Variables 1 Yunghsiang S. Han Graduate Institute of Communication Engineering, National Taipei University Taiwan E-mail: yshan@mail.ntpu.edu.tw 1 Modified from the lecture notes by Prof.

More information

Lecture 14. Text: A Course in Probability by Weiss 5.6. STAT 225 Introduction to Probability Models February 23, Whitney Huang Purdue University

Lecture 14. Text: A Course in Probability by Weiss 5.6. STAT 225 Introduction to Probability Models February 23, Whitney Huang Purdue University Lecture 14 Text: A Course in Probability by Weiss 5.6 STAT 225 Introduction to Probability Models February 23, 2014 Whitney Huang Purdue University 14.1 Agenda 14.2 Review So far, we have covered Bernoulli

More information

MATH/STAT 3360, Probability

MATH/STAT 3360, Probability MATH/STAT 3360, Probability Sample Final Examination This Sample examination has more questions than the actual final, in order to cover a wider range of questions. Estimated times are provided after each

More information

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable Distributions of Functions of Random Variables 5.1 Functions of One Random Variable 5.2 Transformations of Two Random Variables 5.3 Several Random Variables 5.4 The Moment-Generating Function Technique

More information

Relationship between probability set function and random variable - 2 -

Relationship between probability set function and random variable - 2 - 2.0 Random Variables A rat is selected at random from a cage and its sex is determined. The set of possible outcomes is female and male. Thus outcome space is S = {female, male} = {F, M}. If we let X be

More information

EECS 126 Probability and Random Processes University of California, Berkeley: Spring 2015 Abhay Parekh February 17, 2015.

EECS 126 Probability and Random Processes University of California, Berkeley: Spring 2015 Abhay Parekh February 17, 2015. EECS 126 Probability and Random Processes University of California, Berkeley: Spring 2015 Abhay Parekh February 17, 2015 Midterm Exam Last name First name SID Rules. You have 80 mins (5:10pm - 6:30pm)

More information

Exercises with solutions (Set D)

Exercises with solutions (Set D) Exercises with solutions Set D. A fair die is rolled at the same time as a fair coin is tossed. Let A be the number on the upper surface of the die and let B describe the outcome of the coin toss, where

More information

Limiting Distributions

Limiting Distributions Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the

More information

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

STAT2201. Analysis of Engineering & Scientific Data. Unit 3 STAT2201 Analysis of Engineering & Scientific Data Unit 3 Slava Vaisman The University of Queensland School of Mathematics and Physics What we learned in Unit 2 (1) We defined a sample space of a random

More information

A random variable is said to have a beta distribution with parameters (a, b) ifits probability density function is equal to

A random variable is said to have a beta distribution with parameters (a, b) ifits probability density function is equal to 224 Chapter 5 Continuous Random Variables A random variable is said to have a beta distribution with parameters (a, b) ifits probability density function is equal to 1 B(a, b) xa 1 (1 x) b 1 x 1 and is

More information

2. Suppose (X, Y ) is a pair of random variables uniformly distributed over the triangle with vertices (0, 0), (2, 0), (2, 1).

2. Suppose (X, Y ) is a pair of random variables uniformly distributed over the triangle with vertices (0, 0), (2, 0), (2, 1). Name M362K Final Exam Instructions: Show all of your work. You do not have to simplify your answers. No calculators allowed. There is a table of formulae on the last page. 1. Suppose X 1,..., X 1 are independent

More information