Probability Review. Gonzalo Mateos
|
|
- Gervais Long
- 5 years ago
- Views:
Transcription
1 Probability Review Gonzalo Mateos Dept. of ECE and Goergen Institute for Data Science University of Rochester September 11, 2018 Introduction to Random Processes Probability Review 1
2 Sigma-algebras and probability spaces Sigma-algebras and probability spaces Conditional probability, total probability, Bayes rule Independence Random variables Discrete random variables Continuous random variables Expected values Joint probability distributions Joint expectations Introduction to Random Processes Probability Review 2
3 Probability An event is something that happens A random event has an uncertain outcome The probability of an event measures how likely it is to occur Example I ve written a student s name in a piece of paper. Who is she/he? Event: Student x s name is written in the paper Probability: P(x) measures how likely it is that x s name was written Probability is a measurement tool Mathematical language for quantifying uncertainty Introduction to Random Processes Probability Review 3
4 Sigma-algebra Given a sample space or universe S Ex: All students in the class S = {x1, x 2,..., x N } (x n denote names) Def: An outcome is an element or point in S, e.g., x 3 Def: An event E is a subset of S Ex: {x1}, student with name x 1 Ex: Also {x1, x 4}, students with names x 1 and x 4 Outcome x 3 and event {x 3 } are different, the latter is a set Def: A sigma-algebra F is a collection of events E S such that (i) The empty set belongs to F: F (ii) Closed under complement: If E F, then E c F (iii) Closed under countable unions: If E 1, E 2,... F, then i=1e i F F is a set of sets Introduction to Random Processes Probability Review 4
5 Examples of sigma-algebras Example No student and all students, i.e., F 0 := {, S} Example Empty set, women, men, everyone, i.e., F 1 := {, Women, Men, S} Example F 2 including the empty set plus All events (sets) with one student {x 1 },..., {x N } plus All events with two students {x 1, x 2 }, {x 1, x 3 },..., {x 1, x N }, {x 2, x 3 },..., {x 2, x N },... {x N 1, x N } plus All events with three, four,..., N students F 2 is known as the power set of S, denoted 2 S Introduction to Random Processes Probability Review 5
6 Axioms of probability Define a function P(E) from a sigma-algebra F to the real numbers P(E) qualifies as a probability if A1) Non-negativity: P(E) 0 A2) Probability of universe: P(S) = 1 A3) Additivity: Given sequence of disjoint events E 1, E 2,... ( ) P E i = i=1 P (E i ) i=1 Disjoint (mutually exclusive) events means E i E j =, i j Union of countably infinite many disjoint events Triplet (S, F, P( )) is called a probability space Introduction to Random Processes Probability Review 6
7 Consequences of the axioms Implications of the axioms A1)-A3) Impossible event: P( ) = 0 Monotonicity: E 1 E 2 P(E 1 ) P(E 2 ) Range: 0 P(E) 1 Complement: P(E c ) = 1 P(E) Finite disjoint union: For disjoint events E 1,..., E N ( N ) N P E i = P (E i ) i=1 i=1 Inclusion-exclusion: For any events E 1 and E 2 P(E 1 E 2 ) = P(E 1 ) + P(E 2 ) P(E 1 E 2 ) Introduction to Random Processes Probability Review 7
8 Probability example Let s construct a probability space for our running example Universe of all students in the class S = {x 1, x 2,..., x N } Sigma-algebra with all combinations of students, i.e., F = 2 S Suppose names are equiprobable P({x n}) = 1/N for all n Have to specify probability for all E F Define P(E) = E S Q: Is this function a probability? A1): P(E) = E S 0 A2): P(S) = = 1 S S ( N ) A3): P i=1 E N i=1 E i Ni=1 E i = = i = N S S i=1 P(E i) The P( ) just defined is called uniform probability distribution Introduction to Random Processes Probability Review 8
9 Conditional probability, total probability, Bayes rule Sigma-algebras and probability spaces Conditional probability, total probability, Bayes rule Independence Random variables Discrete random variables Continuous random variables Expected values Joint probability distributions Joint expectations Introduction to Random Processes Probability Review 9
10 Conditional probability Consider events E and F, and suppose we know F occurred Q: What does this information imply about the probability of E? Def: Conditional probability of E given F is (need P(F ) > 0) P(E F ) = P(E F ) P(F ) In general P(E F ) P(F E) Renormalize probabilities to the set F Discard a piece of S F E 2 F E 2 May discard a piece of E as well E 1 S For given F with P(F ) > 0, P( F ) satisfies the axioms of probability Introduction to Random Processes Probability Review 10
11 Conditional probability example The name I wrote is male. What is the probability of name x n? Assume male names are F = {x 1,..., x M } P(F ) = M N If name x n is male, x n F and we have for event E = {x n } P(E F ) = P({x n }) = 1 N Conditional probability is as you would expect P(E F ) = P(E F ) P(F ) = 1/N M/N = 1 M If name is female x n / F, then P(E F ) = P( ) = 0 As you would expect, then P(E F ) = 0 Introduction to Random Processes Probability Review 11
12 Law of total probability Consider event E and events F and F c F and F c form a partition of the space S (F F c = S, F F c = ) Because F F c = S cover space S, can write the set E as E = E S = E [F F c ] = [E F ] [E F c ] Because F F c = are disjoint, so is [E F ] [E F c ] = P(E) = P([E F ] [E F c ]) = P(E F ) + P(E F c ) Use definition of conditional probability P(E) = P(E F )P(F ) + P(E F c )P(F c ) Translate conditional information P(E F ) and P(E F c ) Into unconditional information P(E) Introduction to Random Processes Probability Review 12
13 Law of total probability (continued) In general, consider (possibly infinite) partition F i, i = 1, 2,... of S Sets are disjoint F i F j = for i j Sets cover the space i=1 F i = S F 1 E F 1 E F 3 E F2 F 3 F 2 As before, because i=1 F i = S cover the space, can write set E as [ ] E = E S = E F i = [E F i ] i=1 i=1 Because F i F j = are disjoint, so is [E F i ] [E F j ] =. Thus ( ) P(E) = P [E F i ] = P(E F i ) = i=1 i=1 P(E Fi )P(F i ) i=1 Introduction to Random Processes Probability Review 13
14 Total probability example Consider a probability class in some university Seniors get an A with probability (w.p.) 0.9, juniors w.p. 0.8 An exchange student is a senior w.p. 0.7, and a junior w.p. 0.3 Q: What is the probability of the exchange student scoring an A? Let A = exchange student gets an A, S denote senior, and J junior Use the law of total probability P(A) = P(A S)P(S) + P(A J)P(J) = = 0.87 Introduction to Random Processes Probability Review 14
15 Bayes rule From the definition of conditional probability P(E F )P(F ) = P(E F ) Likewise, for F conditioned on E we have P(F E)P(E) = P(F E) Quantities above are equal, giving Bayes rule P(E F ) = P(F E)P(E) P(F ) Bayes rule allows time reversion. If F (future) comes after E (past), P(E F ), probability of past (E) having seen the future (F ) P(F E), probability of future (F ) having seen past (E) Models often describe future past. Interest is often in past future Introduction to Random Processes Probability Review 15
16 Bayes rule example Consider the following partition of my E 1 = spam w.p. P(E 1) = 0.7 E 2 = low priority w.p. P(E 2) = 0.2 E 3 = high priority w.p. P(E 3) = 0.1 Let F = an contains the word free From experience know P(F E1) = 0.9, P(F E2) = P(F E3) = 0.01 I got an containing free. What is the probability that it is spam? Apply Bayes rule P(E 1 F ) = P(F E 1)P(E 1) P(F ) = P(F E 1)P(E 1) 3 i=1 P(F E i )P(E i ) = Law of total probability very useful when applying Bayes rule Introduction to Random Processes Probability Review 16
17 Independence Sigma-algebras and probability spaces Conditional probability, total probability, Bayes rule Independence Random variables Discrete random variables Continuous random variables Expected values Joint probability distributions Joint expectations Introduction to Random Processes Probability Review 17
18 Independence Def: Events E and F are independent if P(E F ) = P(E)P(F ) Events that are not independent are dependent According to definition of conditional probability P(E F ) = P(E F ) P(F ) = P(E)P(F ) P(F ) = P(E) Intuitive, knowing F does not alter our perception of E F bears no information about E The symmetric is also true P(F E) = P(F ) Whether E and F are independent relies strongly on P( ) Avoid confusing with disjoint events, meaning E F = Q: Can disjoint events with P(E) > 0, P(F ) > 0 be independent? No Introduction to Random Processes Probability Review 18
19 Independence example Wrote one name, asked a friend to write another (possibly the same) Probability space (S, F, P( )) for this experiment S is the set of all pairs of names [x n (1), x n (2)], S = N 2 Sigma-algebra is (cartesian product) power set F = 2 S Define P(E) = E S as the uniform probability distribution Consider the events E 1 = I wrote x 1 and E 2 = My friend wrote x 2 Q: Are they independent? Yes, since P(E 1 E 2 ) = P ( {(x 1, x 2 )} ) = {(x 1, x 2 )} S = 1 N 2 = P(E 1)P(E 2 ) Dependent events: E 1 = I wrote x 1 and E 3 = Both names are male Introduction to Random Processes Probability Review 19
20 Independence for more than two events Def: Events E i, i = 1, 2,... are called mutually independent if ( ) P E i = P(E i ) i I i I for every finite subset I of at least two integers Ex: Events E 1, E 2, and E 3 are mutually independent if all the following hold P(E 1 E 2 E 3) = P(E 1)P(E 2)P(E 3) P(E 1 E 2) = P(E 1)P(E 2) P(E 1 E 3) = P(E 1)P(E 3) P(E 2 E 3) = P(E 2)P(E 3) If P(E i E j ) = P(E i )P(E j ) for all (i, j), the E i are pairwise independent Mutual independence pairwise independence. Not the other way Introduction to Random Processes Probability Review 20
21 Random variables Sigma-algebras and probability spaces Conditional probability, total probability, Bayes rule Independence Random variables Discrete random variables Continuous random variables Expected values Joint probability distributions Joint expectations Introduction to Random Processes Probability Review 21
22 Random variable (RV) definition Def: RV X (s) is a function that assigns a value to an outcome s S Think of RVs as measurements associated with an experiment Example Throw a ball inside a 1m 1m square. Interested in ball position Uncertain outcome is the place s [0, 1] 2 where the ball falls Random variables are X (s) and Y (s) position coordinates RV probabilities inferred from probabilities of underlying outcomes P(X (s) = x) = P({s S : X (s) = x}) P(X (s) (, x]) = P({s S : X (s) (, x]}) X (s) is the random variable and x a particular value of X (s) Introduction to Random Processes Probability Review 22
23 Example 1 Throw coin for head (H) or tails (T ). Coin is fair P(H) = 1/2, P(T ) = 1/2. Pay $1 for H, charge $1 for T. Earnings? Possible outcomes are H and T To measure earnings define RV X with values X (H) = 1, X (T ) = 1 Probabilities of the RV are P(X = 1) = P(H) = 1/2, P(X = 1) = P(T ) = 1/2 Also have P(X = x) = 0 for all other x ±1 Introduction to Random Processes Probability Review 23
24 Example 2 Throw 2 coins. Pay $1 for each H, charge $1 for each T. Earnings? Now the possible outcomes are HH, HT, TH, and TT To measure earnings define RV Y with values Y (HH) = 2, Y (HT ) = 0, Y (TH) = 0, Y (TT ) = 2 Probabilities of the RV are P(Y = 2) = P(HH) = 1/4, P(Y = 0) = P(HT ) + P(TH) = 1/2, P(Y = 2) = P(TT ) = 1/4 Introduction to Random Processes Probability Review 24
25 About Examples 1 and 2 RVs are easier to manipulate than events Let s 1 {H, T } be outcome of coin 1 and s 2 {H, T } of coin 2 Can relate Y and X s as Y (s 1, s 2 ) = X 1 (s 1 ) + X 2 (s 2 ) Throw N coins. Earnings? Enumeration becomes cumbersome Alternatively, let s n {H, T } be outcome of n-th toss and define Y (s 1, s 2,..., s N ) = N X n (s n ) n=1 Will usually abuse notation and write Y = N n=1 X n Introduction to Random Processes Probability Review 25
26 Example 3 Throw a coin until landing heads for the first time. P(H) = p Number of throws until the first head? Outcomes are H, TH, TTH, TTTH,... Note that S = Stop tossing after first H (thus THT not a possible outcome) Let N be a RV counting the number of throws N = n if we land T in the first n 1 throws and H in the n-th P(N = 1) = P(H) P(N = 2) = P(TH) = p = (1 p)p. P(N = n) = P(TT }{{... T } H) = (1 p) n 1 p n 1 tails Introduction to Random Processes Probability Review 26
27 Example 3 (continued) From A2) we should have P(S) = n=1 P(N =n) = 1 Holds because n=1 (1 p)n 1 is a geometric series (1 p) n 1 = 1 + (1 p) + (1 p) = 1 (1 p) = 1 p n=1 Plug the sum of the geometric series in the expression for P(S) P(N = n) = p (1 p) n 1 = p 1 p = 1 n=1 n=1 Introduction to Random Processes Probability Review 27
28 Indicator function The indicator function of an event is a random variable Let s S be an outcome, and E S be an event { 1, if s E I {E}(s) = 0, if s / E Example Indicates that outcome s belongs to set E, by taking value 1 Number of throws N until first H. Interested on N exceeding N 0 Event is {N : N > N 0 }. Possible outcomes are N = 1, 2,... Denote indicator function as I N0 = I {N : N > N 0 } Probability P(I N0 = 1) = P(N > N 0 ) = (1 p) N0 For N to exceed N 0 need N 0 consecutive tails Doesn t matter what happens afterwards Introduction to Random Processes Probability Review 28
29 Discrete random variables Sigma-algebras and probability spaces Conditional probability, total probability, Bayes rule Independence Random variables Discrete random variables Continuous random variables Expected values Joint probability distributions Joint expectations Introduction to Random Processes Probability Review 29
30 Probability mass and cumulative distribution functions Discrete RV takes on, at most, a countable number of values Probability mass function (pmf) p X (x) = P(X = x) If RV is clear from context, just write px (x) = p(x) If X supported in {x 1, x 2,...}, pmf satisfies (i) p(x i ) > 0 for i = 1, 2,... (ii) p(x) = 0 for all other x x i (iii) i=1 p(x i) = 1 Pmf for throw to first heads (p = 0.3) Cumulative distribution function (cdf) F X (x) = P(X x) = p(x i ) i:x i x Staircase function with jumps at x i Cdf for throw to first heads (p = 0.3) Introduction to Random Processes Probability Review 30
31 Bernoulli A trial/experiment/bet can succeed w.p. p or fail w.p. q := 1 p Ex: coin throws, any indication of an event Bernoulli X can be 0 or 1. Pmf is p(x) = p x q 1 x Cdf is 0, x < 0 F (x) = q, 0 x < 1 1, x 1 pmf (p = 0.4) cdf (p = 0.4) Introduction to Random Processes Probability Review 31
32 Geometric Count number of Bernoulli trials needed to register first success Trials succeed w.p. p and are independent Number of trials X until success is geometric with parameter p Pmf is p(x) = p(1 p) x 1 One success after x 1 failures, trials are independent Cdf is F (x) = 1 (1 p) x Recall P (X > x) = (1 p) x ; or just sum the geometric series pmf (p = 0.3) cdf (p = 0.3) Introduction to Random Processes Probability Review 32
33 Binomial Count number of successes X in n Bernoulli trials Trials succeed w.p. p and are independent Number of successes X is binomial with parameters (n, p). Pmf is ( ) n p(x) = p x (1 p) n x n! = x (n x)!x! px (1 p) n x X = x for x successes (p x ) and n x failures ((1 p) n x ). ( n x) ways of drawing x successes and n x failures pmf (n = 9, p = 0.4) cdf (n = 9, p = 0.4) Introduction to Random Processes Probability Review 33
34 Binomial (continued) Let Y i, i = 1,... n be Bernoulli RVs with parameter p Y i associated with independent events Can write binomial X with parameters (n, p) as X = Example Consider binomials Y and Z with parameters (n Y, p) and (n Z, p) Q: Probability distribution of X = Y + Z? Write Y = n Y i=1 Y i and Z = n Z i=1 Z i, thus X = n Y i=1 Y i + n Z Z i i=1 X is binomial with parameter (n Y + n Z, p) n i=1 Y i Introduction to Random Processes Probability Review 34
35 Poisson Counts of rare events (radioactive decay, packet arrivals, accidents) Usually modeled as Poisson with parameter λ and pmf λ λx p(x) = e x! Q: Is this a properly defined pmf? Yes Taylor s expansion of e x = 1 + x + x 2 / x i /i! Then P(S) = p(i) = e λ λ i i! = e λ e λ = 1 i=0 i=0 pmf (λ = 4) cdf (λ = 4) Introduction to Random Processes Probability Review 35
36 Poisson approximation of binomial X is binomial with parameters (n, p) Let n while maintaining a constant product np = λ If we just let n number of successes diverges. Boring Compare with Poisson distribution with parameter λ λ = 5, n = 6, 8, 10, 15, 20, Introduction to Random Processes Probability Review 36
37 Poisson and binomial (continued) This is, in fact, the motivation for the definition of a Poisson RV Substituting p = λ/n in the pmf of a binomial RV p n(x) = = n! (n x)!x! ( ) x ( λ 1 λ ) n x n n n(n 1)... (n x + 1) λ x (1 λ/n) n n x x! (1 λ/n) x Used factorials defs., (1 λ/n) n x = (1 λ/n)n (1 λ/n) x, and reordered terms In the limit, red term is lim n (1 λ/n) n = e λ Black and blue terms converge to 1. From both observations lim p n(x) = 1 λx e λ λx = e λ n x! 1 x! Limit is the pmf of a Poisson RV Introduction to Random Processes Probability Review 37
38 Closing remarks Binomial distribution is motivated by counting successes The Poisson is an approximation for large number of trials n Poisson distribution is more tractable (compare pmfs) Sometimes called law of rare events Individual events (successes) happen with small probability p = λ/n Aggregate event (number of successes), though, need not be rare Notice that all four RVs seen so far are related to coin tosses Introduction to Random Processes Probability Review 38
39 Where did the probability space go? Random variables are mappings X (s) : S R Example The underlying probability space often disappears This is for notational convenience, but it s still there Let s construct a probability space for a Bernoulli RV Let S = [0, 1], F the Borel sigma-field and P ([a, b]) = b a, a b Fix a parameter p [0, 1] and define { 1, s p, X (s) = 0, s > p. P (X = 1) = P (s p) = P ([0, p]) = p and P (X = 0) = 1 p Can do a similar construction for all distributions consider so far Introduction to Random Processes Probability Review 39
40 Continuous random variables Sigma-algebras and probability spaces Conditional probability, total probability, Bayes rule Independence Random variables Discrete random variables Continuous random variables Expected values Joint probability distributions Joint expectations Introduction to Random Processes Probability Review 40
41 Continuous RVs, probability density function Possible values for continuous RV X form a dense subset X R Uncountably infinite number of possible values Probability density function (pdf) f X (x) 0 is such that for any subset X R (Normal pdf to the right) P(X X ) = f X (x)dx Will have P(X = x) = 0 for all x X Cdf defined as before and related to the pdf (Normal cdf to the right) X F X (x) = P(X x) = x f X (u) du P(X ) = F X ( ) = lim x F X (x) = Introduction to Random Processes Probability Review 41
42 More on cdfs and pdfs When the set X = [a, b] is an interval of R P (X [a, b]) = P (X b) P (X a) = F X (b) F X (a) In terms of the pdf it can be written as P (X [a, b]) = b For small interval [x 0, x 0 + δx], in particular P (X [x 0, x 0 + δx]) = a x0 +δx x 0 f X (x) dx f X (x) dx f X (x 0)δx Probability is the area under the pdf (thus density ) Another relationship between pdf and cdf is F X (x) = f X (x) x Fundamental theorem of calculus ( derivative inverse of integral ) Introduction to Random Processes Probability Review 42
43 Uniform Model problems with equal probability of landing on an interval [a, b] Pdf of uniform RV is f (x) = 0 outside the interval [a, b] and f (x) = 1 b a, for a x b Cdf is F (x) = (x a)/(b a) in the interval [a, b] (0 before, 1 after) Prob. of interval [α, β] [a, b] is β f (x)dx = (β α)/(b a) α Depends on interval s width β α only, not on its position pdf (a = 1, b = 1) cdf (a = 1, b = 1) Introduction to Random Processes Probability Review 43
44 Exponential Model duration of phone calls, lifetime of electronic components Pdf of exponential RV is f (x) = { λe λx, x 0 0, x < 0 As parameter λ increases, height increases and width decreases Cdf obtained by integrating pdf F (x) = x f (u) du = x 0 x λe λu du = e λu = 1 e λx 0 1 pdf (λ = 1) cdf (λ = 1) Introduction to Random Processes Probability Review 44
45 Normal / Gaussian Model randomness arising from large number of random effects Pdf of normal RV is f (x) = 1 2πσ e (x µ)2 /2σ 2 µ is the mean (center), σ 2 is the variance (width) 0.68 prob. between µ ± σ, prob. in µ ± 3σ Standard normal RV has µ = 0 and σ 2 = 1 Cdf F (x) cannot be expressed in terms of elementary functions 0.4 pdf (µ = 0, σ 2 = 1) cdf (µ = 0, σ 2 = 1) Introduction to Random Processes Probability Review 45
46 Expected values Sigma-algebras and probability spaces Conditional probability, total probability, Bayes rule Independence Random variables Discrete random variables Continuous random variables Expected values Joint probability distributions Joint expectations Introduction to Random Processes Probability Review 46
47 Expected values We are asked to summarize information about a RV in a single value What should this value be? If we are allowed a description with a few values What should they be? Expected (mean) values are convenient answers to these questions Beware: Expectations are condensed descriptions They overlook some aspects of the random phenomenon Whole story told by the probability distribution (cdf) Introduction to Random Processes Probability Review 47
48 Definition for discrete RVs Discrete RV X taking on values x i, i = 1, 2,... with pmf p(x) Def: The expected value of the discrete RV X is E [X ] := x i p(x i ) = xp(x) i=1 x:p(x)>0 Weighted average of possible values x i. Probabilities are weights Common average if RV takes values x i, i = 1,..., N equiprobably E [X ] = N x i p(x i ) = i=1 N i=1 x i 1 N = 1 N N i=1 x i Introduction to Random Processes Probability Review 48
49 Expected value of Bernoulli and geometric RVs Ex: For a Bernoulli RV p(x) = p x q 1 x, for x {0, 1} E [X ] = 1 p + 0 q = p Ex: For a geometric RV p(x) = p(1 p) x 1 = pq x 1, for x 1 Note that q x / q = xq x 1 and that derivatives are linear operators E [X ] = xpq x 1 = p x=1 x=1 q x q = p ( ) q x q x=1 Sum inside derivative is geometric. Sums to q/(1 q), thus E [X ] = p ( ) q p = q 1 q (1 q) 2 = 1 p Time to first success is inverse of success probability. Reasonable Introduction to Random Processes Probability Review 49
50 Expected value of Poisson RV Ex: For a Poisson RV p(x) = e λ (λ x /x!), for x 0 First summand in definition is 0, pull λ out, and use x x! = 1 (x 1)! E [X ] = x=0 λ λx xe x! = λe λ x=1 λ x 1 (x 1)! Sum is Taylor s expansion of e λ = 1 + λ + λ 2 /2! λ x /x! E [X ] = λe λ e λ = λ Poisson is limit of binomial for large number of trials n, with λ = np Counts number of successes in n trials that succeed w.p. p Expected number of successes is λ = np Number of trials probability of individual success. Reasonable Introduction to Random Processes Probability Review 50
51 Definition for continuous RVs Continuous RV X taking values on R with pdf f (x) Def: The expected value of the continuous RV X is E [X ] := xf (x) dx Compare with E [X ] := x:p(x)>0 xp(x) in the discrete RV case Note that the integral or sum are assumed to be well defined Otherwise we say the expectation does not exist Introduction to Random Processes Probability Review 51
52 Expected value of normal RV Ex: For a normal RV add and subtract µ, separate integrals E [X ] = 1 2πσ = 1 2πσ = µ 1 2πσ xe (x µ) 2 2σ 2 dx (x + µ µ)e (x µ) 2 2σ 2 e (x µ) 2 2σ 2 dx + 1 dx 2πσ First integral is 1 because it integrates a pdf in all R (x µ)e (x µ) 2 2σ 2 Second integral is 0 by symmetry. Both observations yield E [X ] = µ The mean of a RV with a symmetric pdf is the point of symmetry dx Introduction to Random Processes Probability Review 52
53 Expected value of uniform and exponential RVs Ex: For a uniform RV f (x) = 1/(b a), for a x b E [X ] = xf (x) dx = b a x b a dx = b2 a 2 (a + b) = 2(b a) 2 Makes sense, since pdf is symmetric around midpoint (a + b)/2 Ex: For an exponential RV (non symmetric) integrate by parts E [X ] = 0 xλe λx dx = xe λx + 0 = xe λx 0 0 e λx λ e λx dx 0 = 1 λ Introduction to Random Processes Probability Review 53
54 Expected value of a function of a RV Consider a function g(x ) of a RV X. Expected value of g(x )? g(x) is also a RV, then it also has a pmf p g(x ) ( g(x) ) E [g(x )] = g(x):p g(x ) (g(x))>0 g(x)p g(x ) ( g(x) ) Requires calculating the pmf of g(x ). There is a simpler way Theorem Consider a function g(x ) of a discrete RV X with pmf p X (x). Then E [g(x )] = g(x i )p X (x i ) i=1 Weighted average of functional values. No need to find pmf of g(x ) Same can be proved for a continuous RV E [g(x )] = g(x)f X (x) dx Introduction to Random Processes Probability Review 54
55 Expected value of a linear transformation Consider a linear function (actually affine) g(x ) = ax + b E [ax + b] = (ax i + b)p X (x i ) i=1 = ax i p X (x i ) + bp X (x i ) i=1 i=1 = a x i p X (x i ) + b p X (x i ) i=1 i=1 = ae [X ] + b1 Can interchange expectation with additive/multiplicative constants E [ax + b] = ae [X ] + b Again, the same holds for a continuous RV Introduction to Random Processes Probability Review 55
56 Expected value of an indicator function Let X be a RV and X be a set I {X X } = { 1, if x X 0, if x / X Expected value of I {X X } in the discrete case E [I {X X }] = x:p X (x)>0 I {x X }p X (x) = x X Likewise in the continuous case E [I {X X }] = I {x X }f X (x)dx = x X p X (x) = P (X X ) f X (x)dx = P (X X ) Expected value of indicator RV = Probability of indicated event Recall E [X ] = p for Bernoulli RV (it indicates success ) Introduction to Random Processes Probability Review 56
57 Moments, central moments and variance Def: The n-th moment (n 0) of a RV is E [X n ] = xi n p(x i ) Def: The n-th central moment corrects for the mean, that is [ (X ) ] n ( E E [X ] = xi E [X ] ) n p(xi ) i=1 i=1 0-th order moment is E [ X 0] = 1; 1-st moment is the mean E [X ] 2-nd central moment is the variance. Measures width of the pmf [ (X ) ] 2 var [X ] = E E [X ] = E [ X 2] E 2 [X ] Ex: For affine functions var [ax + b] = a 2 var [X ] Introduction to Random Processes Probability Review 57
58 Variance of Bernoulli and Poisson RVs Ex: For a Bernoulli RV X with parameter p, E [X ] = E [ X 2] = p var [X ] = E [ X 2] E 2 [X ] = p p 2 = p(1 p) Ex: For Poisson RV Y with parameter λ, second moment is E [ Y 2] = = y=0 y=1 y 2 λ λy e y! = y=1 y e λ λ y (y 1)! (y 1) e λ λ y (y 1)! + e λ λ y (y 1)! = e λ λ 2 y=2 y=1 λ y 2 (y 2)! + e λ λ y=1 = e λ λ 2 e λ + e λ λe λ = λ 2 + λ var [Y ] = E [ Y 2] E 2 [Y ] = λ 2 + λ λ 2 = λ λ y 1 (y 1)! Introduction to Random Processes Probability Review 58
59 Joint probability distributions Sigma-algebras and probability spaces Conditional probability, total probability, Bayes rule Independence Random variables Discrete random variables Continuous random variables Expected values Joint probability distributions Joint expectations Introduction to Random Processes Probability Review 59
60 Joint cdf Want to study problems with more than one RV. Say, e.g., X and Y Probability distributions of X and Y are not sufficient Joint probability distribution (cdf) of (X, Y ) defined as F XY (x, y) = P (X x, Y y) If X, Y clear from context omit subindex to write F XY (x, y) = F (x, y) Can recover F X (x) by considering all possible values of Y F X (x) = P (X x) = P (X x, Y ) = F XY (x, ) F X (x) and F Y (y) = F XY (, y) are called marginal cdfs Introduction to Random Processes Probability Review 60
61 Joint pmf Consider discrete RVs X and Y X takes values in X := {x 1, x 2,...} and Y in Y := {y 1, y 2,...} Joint pmf of (X, Y ) defined as p XY (x, y) = P (X = x, Y = y) Possible values (x, y) are elements of the Cartesian product X Y (x1, y 1), (x 1, y 2),..., (x 2, y 1), (x 2, y 2),..., (x 3, y 1), (x 3, y 2),... Marginal pmf p X (x) obtained by summing over all values of Y p X (x) = P (X = x) = y Y P (X = x, Y = y) = y Y p XY (x, y) Likewise p Y (y) = x Xp XY (x, y). Marginalize by summing Introduction to Random Processes Probability Review 61
62 Joint pdf Consider continuous RVs X, Y. Arbitrary set A R 2 Joint pdf is a function f XY (x, y) : R 2 R + such that P ((X, Y ) A) = f XY (x, y) dxdy Marginalization. There are two ways of writing P (X X ) P (X X ) = P (X X, Y R) = A + X X Definition of f X (x) P (X X ) = X X f X (x) dx f XY (x, y) dydx Introduction to Random Processes Probability Review 62
63 Joint pdf Consider continuous RVs X, Y. Arbitrary set A R 2 Joint pdf is a function f XY (x, y) : R 2 R + such that P ((X, Y ) A) = f XY (x, y) dxdy Marginalization. There are two ways of writing P (X X ) P (X X ) = P (X X, Y R) = A + X X f XY (x, y) dydx Definition of f X (x) P (X X ) = X X f X (x) dx Lipstick on a pig (same thing written differently is still same thing) f X (x) = + f XY (x, y) dy, f Y (y) = + f XY (x, y) dx Introduction to Random Processes Probability Review 63
64 Example Consider two Bernoulli RVs B 1, B 2, with the same parameter p Define X = B 1 and Y = B 1 + B 2 The pmf of X is p X (0) = 1 p, p X (1) = p Likewise, the pmf of Y is p Y (0) = (1 p) 2, p Y (1) = 2p(1 p), p Y (2) = p 2 The joint pmf of X and Y is p XY (0, 0) = (1 p) 2, p XY (0, 1) = p(1 p), p XY (0, 2) = 0 p XY (1, 0) = 0, p XY (1, 1) = p(1 p), p XY (1, 2) = p 2 Introduction to Random Processes Probability Review 64
65 Random vectors For convenience often arrange RVs in a vector Prob. distribution of vector is joint distribution of its entries Consider, e.g., two RVs X and Y. Random vector is X = [X, Y ] T If X and Y are discrete, vector variable X is discrete with pmf ( p X (x) = p X [x, y] T ) = p XY (x, y) If X, Y continuous, X continuous with pdf ( f X (x) = f X [x, y] T ) = f XY (x, y) ( Vector cdf is F X (x) = F ) X [x, y] T = F XY (x, y) In general, can define n-dimensional RVs X := [X 1, X 2,..., X n ] T Just notation, definitions carry over from the n = 2 case Introduction to Random Processes Probability Review 65
66 Joint expectations Sigma-algebras and probability spaces Conditional probability, total probability, Bayes rule Independence Random variables Discrete random variables Continuous random variables Expected values Joint probability distributions Joint expectations Introduction to Random Processes Probability Review 66
67 Joint expectations RVs X and Y and function g(x, Y ). Function g(x, Y ) also a RV Expected value of g(x, Y ) when X and Y discrete can be written as E [g(x, Y )] = g(x, y)p XY (x, y) When X and Y are continuous E [g(x, Y )] = x,y:p XY (x,y)>0 g(x, y)f XY (x, y) dxdy Can have more than two RVs and use vector notation Ex: Linear transformation of a vector RV X R n : g(x) = a T X E [ a T X ] = a T xf X (x) dx R n Introduction to Random Processes Probability Review 67
68 Expected value of a sum of random variables Expected value of the sum of two continuous RVs E [X + Y ] = = (x + y)f XY (x, y) dxdy x f XY (x, y) dxdy + y f XY (x, y) dxdy Remove x (y) from innermost integral in first (second) summand E [X + Y ] = x f XY (x, y) dy dx + = xf X (x) dx + = E [X ] + E [Y ] Used marginal expressions yf Y (y) dy Expectation summation E [ i X ] i = i E [X i] y f XY (x, y) dx dy Introduction to Random Processes Probability Review 68
69 Expected value is a linear operator Combining with earlier result E [ax + b] = ae [X ] + b proves that E [a x X + a y Y + b] = a x E [X ] + a y E [Y ] + b Better yet, using vector notation (with a R n, X R n, b a scalar) E [ a T X + b ] = a T E [X] + b Also, if A is an m n matrix with rows a T 1,..., at m and b R m a vector with elements b 1,..., b m, we can write E [AX + b]= E [ a T 1 X + b 1 ] E [ a T 2 X + b 2 ]. E [ a T mx + b m ] = a T 1 E [X] + b 1 a T 2 E [X] + b 2. a T me [X] + b m = AE [X] + b Expected value operator can be interchanged with linear operations Introduction to Random Processes Probability Review 69
70 Independence of RVs Events E and F are independent if P (E F ) = P (E) P (F ) Def: RVs X and Y are independent if events X x and Y y are independent for all x and y, i.e. P (X x, Y y) = P (X x) P (Y y) By definition, equivalent to F XY (x, y) = F X (x)f Y (y) For discrete RVs equivalent to analogous relation between pmfs p XY (x, y) = p X (x)p Y (y) For continuous RVs the analogous is true for pdfs f XY (x, y) = f X (x)f Y (y) Independence Joint distribution factorizes into product of marginals Introduction to Random Processes Probability Review 70
71 Sum of independent Poisson RVs Independent Poisson RVs X and Y with parameters λ x and λ y Q: Probability distribution of the sum RV Z := X + Y? Z = n only if X = k, Y = n k for some 0 k n (use independence, Poisson pmf, rearrange terms, binomial theorem) p Z (n) = = = n P (X = k, Y = n k) = k=0 n k=0 e λx λk x e (λx +λy ) n! λ n k y k! e λy (n k)! (λ x + λ y ) n = n P (X = k) P (Y = n k) k=0 e (λx +λy ) n! n k=0 n! (n k)!k! λk x λ n k y Z is Poisson with parameter λ z := λ x + λ y Sum of independent Poissons is Poisson (parameters added) Introduction to Random Processes Probability Review 71
72 Expected value of a binomial RV Binomial RVs count number of successes in n Bernoulli trials Ex: Let X i, i = 1,... n be n independent Bernoulli RVs n n Can write binomial X = X i E [X ] = E [X i ] = np i=1 i=1 Expected nr. successes = nr. trials prob. individual success Same interpretation that we observed for Poisson RVs n Ex: Dependent Bernoulli trials. Y = X i, but X i are not independent Expected nr. successes is still E [Y ] = np Linearity of expectation does not require independence Y is not binomial distributed i=1 Introduction to Random Processes Probability Review 72
73 Expected value of a product of independent RVs Theorem For independent RVs X and Y, and arbitrary functions g(x ) and h(y ): E [g(x )h(y )] = E [g(x )] E [h(y )] The expected value of the product is the product of the expected values Can show that g(x ) and h(y ) are also independent. Intuitive Ex: Special case when g(x ) = X and h(y ) = Y yields E [XY ] = E [X ] E [Y ] Expectation and product can be interchanged if RVs are independent Different from interchange with linear operations (always possible) Introduction to Random Processes Probability Review 73
74 Expected value of a product of independent RVs Proof. Suppose X and Y continuous RVs. Use definition of independence E [g(x )h(y )] = = g(x)h(y)f XY (x, y) dxdy g(x)h(y)f X (x)f Y (y) dxdy Integrand is product of a function of x and a function of y E [g(x )h(y )] = g(x)f X (x) dx = E [g(x )] E [h(y )] h(y)f Y (y) dy Introduction to Random Processes Probability Review 74
75 Variance of a sum of independent RVs Let X n, n = 1,... N be independent with E [X n ] = µ n, var [X n ] = σ 2 n Q: Variance of sum X := N n=1 X n? Notice that mean of X is E [X ] = N n=1 µ n. Then ( N ) 2 ( N N ) 2 var [X ] = E X n µ n = E (X n µ n ) n=1 n=1 n=1 Expand square and interchange summation and expectation var [X ] = N N n=1 m=1 [ ] E (X n µ n )(X m µ m ) Introduction to Random Processes Probability Review 75
76 Variance of a sum of independent RVs (continued) Separate terms in sum. Then use independence and E(X n µ n ) = 0 var [X ] = = N N n=1,n m m=1 N n=1,n m m=1 [ ] E (X n µ n )(X m µ m ) + N E(X n µ n )E(X m µ m ) + N n=1 E [(X n µ n ) 2] N σn 2 = If RVs are independent Variance of sum is sum of variances Slightly more general result holds for independent X i, i = 1,..., n [ ] var (a i X i + b i ) = ai 2 var [X i ] i i n=1 N n=1 σ 2 n Introduction to Random Processes Probability Review 76
77 Variance of binomial RV and sample mean Ex: Let X i, i = 1,... n be independent Bernoulli RVs Recall E [X i ] = p and var [X i ] = p(1 p) Write binomial X with parameters (n, p) as: X = Variance of binomial then var [X ] = n i=1 X i n var [X i ] = np(1 p) Ex: Let Y i, i = 1,... n be independent RVs and E [Y i ] = µ, var [Y i ] = σ 2 Sample mean is Ȳ = 1 n Y i. What about E [ Ȳ ] and var [ Ȳ ]? n i=1 Expected value E [ Ȳ ] = 1 n Variance var [ Ȳ ] = 1 n 2 n i=1 i=1 n E [Y i ] = µ i=1 var [Y i ] = σ2 n (used independence) Introduction to Random Processes Probability Review 77
78 Covariance Def: The covariance of X and Y is (generalizes variance to pairs of RVs) cov(x, Y ) = E [(X E [X ])(Y E [Y ])] = E [XY ] E [X ] E [Y ] If cov(x, Y ) = 0 variables X and Y are said to be uncorrelated If X, Y independent then E [XY ] = E [X ] E [Y ] and cov(x, Y ) = 0 Independence implies uncorrelated RVs Opposite is not true, may have cov(x, Y ) = 0 for dependent X, Y Ex: X uniform in [ a, a] and Y = X 2 But uncorrelatedness implies independence if X, Y are normal If cov(x, Y ) > 0 then X and Y tend to move in the same direction Positive correlation If cov(x, Y ) < 0 then X and Y tend to move in opposite directions Negative correlation Introduction to Random Processes Probability Review 78
79 Covariance example Let X be a zero-mean random signal and Z zero-mean noise Signal X and noise Z are independent Consider received signals Y 1 = X + Z and Y 2 = X + Z (I) Y 1 and X are positively correlated (X, Y 1 move in same direction) cov(x, Y 1 ) = E [XY 1 ] E [X ] E [Y 1 ] = E [X (X + Z)] E [X ] E [X + Z] Second term is 0 (E [X ] = 0). For first term independence of X, Z E [X (X + Z)] = E [ X 2] + E [X ] E [Z] = E [ X 2] Combining observations cov(x, Y 1 ) = E [ X 2] > 0 Introduction to Random Processes Probability Review 79
80 Covariance example (continued) (II) Y 2 and X are negatively correlated (X, Y 2 move opposite direction) Same computations cov(x, Y 2 ) = E [ X 2] < 0 (III) Can also compute correlation between Y 1 and Y 2 cov(y 1, Y 2 ) = E [(X + Z)( X + Z)] E [(X + Z)] E [( X + Z)] = E [ X 2] + E [ Z 2] Negative correlation if E [ X 2] > E [ Z 2] (small noise) Positive correlation if E [ X 2] < E [ Z 2] (large noise) Correlation between X and Y 1 or X and Y 2 comes from causality Correlation between Y 1 and Y 2 does not. Latent variables X and Z Correlation does not imply causation Plausible, indeed commonly used, model of a communication channel Introduction to Random Processes Probability Review 80
81 Glossary Sample space Outcome and event Sigma-algebra Countable union Axioms of probability Probability space Conditional probability Law of total probability Bayes rule Independent events Random variable (RV) Discrete RV Bernoulli, binomial, Poisson Continuous RV Uniform, Normal, exponential Indicator RV Pmf, pdf and cdf Law of rare events Expected value Variance and standard deviation Joint probability distribution Marginal distribution Random vector Independent RVs Covariance Uncorrelated RVs Introduction to Random Processes Probability Review 81
Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable
Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed
More informationLecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable
Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed
More informationReview of Probability Theory
Review of Probability Theory Arian Maleki and Tom Do Stanford University Probability theory is the study of uncertainty Through this class, we will be relying on concepts from probability theory for deriving
More informationRecitation 2: Probability
Recitation 2: Probability Colin White, Kenny Marino January 23, 2018 Outline Facts about sets Definitions and facts about probability Random Variables and Joint Distributions Characteristics of distributions
More informationNorthwestern University Department of Electrical Engineering and Computer Science
Northwestern University Department of Electrical Engineering and Computer Science EECS 454: Modeling and Analysis of Communication Networks Spring 2008 Probability Review As discussed in Lecture 1, probability
More information1.1 Review of Probability Theory
1.1 Review of Probability Theory Angela Peace Biomathemtics II MATH 5355 Spring 2017 Lecture notes follow: Allen, Linda JS. An introduction to stochastic processes with applications to biology. CRC Press,
More informationWhy study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables
ECE 6010 Lecture 1 Introduction; Review of Random Variables Readings from G&S: Chapter 1. Section 2.1, Section 2.3, Section 2.4, Section 3.1, Section 3.2, Section 3.5, Section 4.1, Section 4.2, Section
More information1 Random Variable: Topics
Note: Handouts DO NOT replace the book. In most cases, they only provide a guideline on topics and an intuitive feel. 1 Random Variable: Topics Chap 2, 2.1-2.4 and Chap 3, 3.1-3.3 What is a random variable?
More informationRandom variables. DS GA 1002 Probability and Statistics for Data Science.
Random variables DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Motivation Random variables model numerical quantities
More informationMA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems
MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Review of Basic Probability The fundamentals, random variables, probability distributions Probability mass/density functions
More informationLectures on Elementary Probability. William G. Faris
Lectures on Elementary Probability William G. Faris February 22, 2002 2 Contents 1 Combinatorics 5 1.1 Factorials and binomial coefficients................. 5 1.2 Sampling with replacement.....................
More informationSTAT2201. Analysis of Engineering & Scientific Data. Unit 3
STAT2201 Analysis of Engineering & Scientific Data Unit 3 Slava Vaisman The University of Queensland School of Mathematics and Physics What we learned in Unit 2 (1) We defined a sample space of a random
More informationSingle Maths B: Introduction to Probability
Single Maths B: Introduction to Probability Overview Lecturer Email Office Homework Webpage Dr Jonathan Cumming j.a.cumming@durham.ac.uk CM233 None! http://maths.dur.ac.uk/stats/people/jac/singleb/ 1 Introduction
More informationLecture 4: Probability and Discrete Random Variables
Error Correcting Codes: Combinatorics, Algorithms and Applications (Fall 2007) Lecture 4: Probability and Discrete Random Variables Wednesday, January 21, 2009 Lecturer: Atri Rudra Scribe: Anonymous 1
More information1 Presessional Probability
1 Presessional Probability Probability theory is essential for the development of mathematical models in finance, because of the randomness nature of price fluctuations in the markets. This presessional
More informationDiscrete Random Variables
CPSC 53 Systems Modeling and Simulation Discrete Random Variables Dr. Anirban Mahanti Department of Computer Science University of Calgary mahanti@cpsc.ucalgary.ca Random Variables A random variable is
More information1: PROBABILITY REVIEW
1: PROBABILITY REVIEW Marek Rutkowski School of Mathematics and Statistics University of Sydney Semester 2, 2016 M. Rutkowski (USydney) Slides 1: Probability Review 1 / 56 Outline We will review the following
More informationChapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University
Chapter 3, 4 Random Variables ENCS6161 - Probability and Stochastic Processes Concordia University ENCS6161 p.1/47 The Notion of a Random Variable A random variable X is a function that assigns a real
More information18.440: Lecture 28 Lectures Review
18.440: Lecture 28 Lectures 18-27 Review Scott Sheffield MIT Outline Outline It s the coins, stupid Much of what we have done in this course can be motivated by the i.i.d. sequence X i where each X i is
More informationPart IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationRandom Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R
In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample
More informationChapter 2 Random Variables
Stochastic Processes Chapter 2 Random Variables Prof. Jernan Juang Dept. of Engineering Science National Cheng Kung University Prof. Chun-Hung Liu Dept. of Electrical and Computer Eng. National Chiao Tung
More informationDiscrete Random Variable
Discrete Random Variable Outcome of a random experiment need not to be a number. We are generally interested in some measurement or numerical attribute of the outcome, rather than the outcome itself. n
More informationLecture 4: Probability, Proof Techniques, Method of Induction Lecturer: Lale Özkahya
BBM 205 Discrete Mathematics Hacettepe University http://web.cs.hacettepe.edu.tr/ bbm205 Lecture 4: Probability, Proof Techniques, Method of Induction Lecturer: Lale Özkahya Resources: Kenneth Rosen, Discrete
More informationProbability review. September 11, Stoch. Systems Analysis Introduction 1
Probability review Alejandro Ribeiro Dept. of Electrical and Systems Engineering University of Pennsylvania aribeiro@seas.upenn.edu http://www.seas.upenn.edu/users/~aribeiro/ September 11, 2015 Stoch.
More informationProbability and Distributions
Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated
More informationMultivariate distributions
CHAPTER Multivariate distributions.. Introduction We want to discuss collections of random variables (X, X,..., X n ), which are known as random vectors. In the discrete case, we can define the density
More informationProbability, Random Processes and Inference
INSTITUTO POLITÉCNICO NACIONAL CENTRO DE INVESTIGACION EN COMPUTACION Laboratorio de Ciberseguridad Probability, Random Processes and Inference Dr. Ponciano Jorge Escamilla Ambrosio pescamilla@cic.ipn.mx
More informationChapter 2: Random Variables
ECE54: Stochastic Signals and Systems Fall 28 Lecture 2 - September 3, 28 Dr. Salim El Rouayheb Scribe: Peiwen Tian, Lu Liu, Ghadir Ayache Chapter 2: Random Variables Example. Tossing a fair coin twice:
More informationPerhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.
Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage
More informationLecture 2: Repetition of probability theory and statistics
Algorithms for Uncertainty Quantification SS8, IN2345 Tobias Neckel Scientific Computing in Computer Science TUM Lecture 2: Repetition of probability theory and statistics Concept of Building Block: Prerequisites:
More informationSummary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016
8. For any two events E and F, P (E) = P (E F ) + P (E F c ). Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 Sample space. A sample space consists of a underlying
More informationDeep Learning for Computer Vision
Deep Learning for Computer Vision Lecture 3: Probability, Bayes Theorem, and Bayes Classification Peter Belhumeur Computer Science Columbia University Probability Should you play this game? Game: A fair
More informationRelationship between probability set function and random variable - 2 -
2.0 Random Variables A rat is selected at random from a cage and its sex is determined. The set of possible outcomes is female and male. Thus outcome space is S = {female, male} = {F, M}. If we let X be
More informationFundamental Tools - Probability Theory II
Fundamental Tools - Probability Theory II MSc Financial Mathematics The University of Warwick September 29, 2015 MSc Financial Mathematics Fundamental Tools - Probability Theory II 1 / 22 Measurable random
More informationWe introduce methods that are useful in:
Instructor: Shengyu Zhang Content Derived Distributions Covariance and Correlation Conditional Expectation and Variance Revisited Transforms Sum of a Random Number of Independent Random Variables more
More informationProbability. Lecture Notes. Adolfo J. Rumbos
Probability Lecture Notes Adolfo J. Rumbos October 20, 204 2 Contents Introduction 5. An example from statistical inference................ 5 2 Probability Spaces 9 2. Sample Spaces and σ fields.....................
More informationProbability Theory. Introduction to Probability Theory. Principles of Counting Examples. Principles of Counting. Probability spaces.
Probability Theory To start out the course, we need to know something about statistics and probability Introduction to Probability Theory L645 Advanced NLP Autumn 2009 This is only an introduction; for
More informationBasics on Probability. Jingrui He 09/11/2007
Basics on Probability Jingrui He 09/11/2007 Coin Flips You flip a coin Head with probability 0.5 You flip 100 coins How many heads would you expect Coin Flips cont. You flip a coin Head with probability
More informationChapter 1. Sets and probability. 1.3 Probability space
Random processes - Chapter 1. Sets and probability 1 Random processes Chapter 1. Sets and probability 1.3 Probability space 1.3 Probability space Random processes - Chapter 1. Sets and probability 2 Probability
More informationMore on Distribution Function
More on Distribution Function The distribution of a random variable X can be determined directly from its cumulative distribution function F X. Theorem: Let X be any random variable, with cumulative distribution
More informationGeneral Random Variables
1/65 Chia-Ping Chen Professor Department of Computer Science and Engineering National Sun Yat-sen University Probability A general random variable is discrete, continuous, or mixed. A discrete random variable
More informationLecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019
Lecture 10: Probability distributions DANIEL WELLER TUESDAY, FEBRUARY 19, 2019 Agenda What is probability? (again) Describing probabilities (distributions) Understanding probabilities (expectation) Partial
More informationAlgorithms for Uncertainty Quantification
Algorithms for Uncertainty Quantification Tobias Neckel, Ionuț-Gabriel Farcaș Lehrstuhl Informatik V Summer Semester 2017 Lecture 2: Repetition of probability theory and statistics Example: coin flip Example
More informationChapter 3: Random Variables 1
Chapter 3: Random Variables 1 Yunghsiang S. Han Graduate Institute of Communication Engineering, National Taipei University Taiwan E-mail: yshan@mail.ntpu.edu.tw 1 Modified from the lecture notes by Prof.
More informationStatistical Inference
Statistical Inference Lecture 1: Probability Theory MING GAO DASE @ ECNU (for course related communications) mgao@dase.ecnu.edu.cn Sep. 11, 2018 Outline Introduction Set Theory Basics of Probability Theory
More informationDept. of Linguistics, Indiana University Fall 2015
L645 Dept. of Linguistics, Indiana University Fall 2015 1 / 34 To start out the course, we need to know something about statistics and This is only an introduction; for a fuller understanding, you would
More informationBivariate distributions
Bivariate distributions 3 th October 017 lecture based on Hogg Tanis Zimmerman: Probability and Statistical Inference (9th ed.) Bivariate Distributions of the Discrete Type The Correlation Coefficient
More informationELEG 3143 Probability & Stochastic Process Ch. 2 Discrete Random Variables
Department of Electrical Engineering University of Arkansas ELEG 3143 Probability & Stochastic Process Ch. 2 Discrete Random Variables Dr. Jingxian Wu wuj@uark.edu OUTLINE 2 Random Variable Discrete Random
More informationLecture Notes 2 Random Variables. Discrete Random Variables: Probability mass function (pmf)
Lecture Notes 2 Random Variables Definition Discrete Random Variables: Probability mass function (pmf) Continuous Random Variables: Probability density function (pdf) Mean and Variance Cumulative Distribution
More informationChapter 4. Continuous Random Variables 4.1 PDF
Chapter 4 Continuous Random Variables In this chapter we study continuous random variables. The linkage between continuous and discrete random variables is the cumulative distribution (CDF) which we will
More informationProbability: Why do we care? Lecture 2: Probability and Distributions. Classical Definition. What is Probability?
Probability: Why do we care? Lecture 2: Probability and Distributions Sandy Eckel seckel@jhsph.edu 22 April 2008 Probability helps us by: Allowing us to translate scientific questions into mathematical
More informationSample Spaces, Random Variables
Sample Spaces, Random Variables Moulinath Banerjee University of Michigan August 3, 22 Probabilities In talking about probabilities, the fundamental object is Ω, the sample space. (elements) in Ω are denoted
More informationRandom Variables. P(x) = P[X(e)] = P(e). (1)
Random Variables Random variable (discrete or continuous) is used to derive the output statistical properties of a system whose input is a random variable or random in nature. Definition Consider an experiment
More informationGaussian, Markov and stationary processes
Gaussian, Markov and stationary processes Gonzalo Mateos Dept. of ECE and Goergen Institute for Data Science University of Rochester gmateosb@ece.rochester.edu http://www.ece.rochester.edu/~gmateosb/ November
More informationDS-GA 1002 Lecture notes 2 Fall Random variables
DS-GA 12 Lecture notes 2 Fall 216 1 Introduction Random variables Random variables are a fundamental tool in probabilistic modeling. They allow us to model numerical quantities that are uncertain: the
More information18.440: Lecture 28 Lectures Review
18.440: Lecture 28 Lectures 17-27 Review Scott Sheffield MIT 1 Outline Continuous random variables Problems motivated by coin tossing Random variable properties 2 Outline Continuous random variables Problems
More informationMath Review Sheet, Fall 2008
1 Descriptive Statistics Math 3070-5 Review Sheet, Fall 2008 First we need to know about the relationship among Population Samples Objects The distribution of the population can be given in one of the
More informationPCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities
PCMI 207 - Introduction to Random Matrix Theory Handout #2 06.27.207 REVIEW OF PROBABILITY THEORY Chapter - Events and Their Probabilities.. Events as Sets Definition (σ-field). A collection F of subsets
More informationExpectation. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda
Expectation DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Aim Describe random variables with a few numbers: mean, variance,
More informationOrigins of Probability Theory
1 16.584: INTRODUCTION Theory and Tools of Probability required to analyze and design systems subject to uncertain outcomes/unpredictability/randomness. Such systems more generally referred to as Experiments.
More informationFinal Exam # 3. Sta 230: Probability. December 16, 2012
Final Exam # 3 Sta 230: Probability December 16, 2012 This is a closed-book exam so do not refer to your notes, the text, or any other books (please put them on the floor). You may use the extra sheets
More informationA Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.
A Probability Primer A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes. Are you holding all the cards?? Random Events A random event, E,
More informationThings to remember when learning probability distributions:
SPECIAL DISTRIBUTIONS Some distributions are special because they are useful They include: Poisson, exponential, Normal (Gaussian), Gamma, geometric, negative binomial, Binomial and hypergeometric distributions
More information1 Review of Probability
1 Review of Probability Random variables are denoted by X, Y, Z, etc. The cumulative distribution function (c.d.f.) of a random variable X is denoted by F (x) = P (X x), < x
More informationp. 4-1 Random Variables
Random Variables A Motivating Example Experiment: Sample k students without replacement from the population of all n students (labeled as 1, 2,, n, respectively) in our class. = {all combinations} = {{i
More informationChapter 2. Probability
2-1 Chapter 2 Probability 2-2 Section 2.1: Basic Ideas Definition: An experiment is a process that results in an outcome that cannot be predicted in advance with certainty. Examples: rolling a die tossing
More informationSUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416)
SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416) D. ARAPURA This is a summary of the essential material covered so far. The final will be cumulative. I ve also included some review problems
More informationLecture 2: Probability and Distributions
Lecture 2: Probability and Distributions Ani Manichaikul amanicha@jhsph.edu 17 April 2007 1 / 65 Probability: Why do we care? Probability helps us by: Allowing us to translate scientific questions info
More informationLecture 16. Lectures 1-15 Review
18.440: Lecture 16 Lectures 1-15 Review Scott Sheffield MIT 1 Outline Counting tricks and basic principles of probability Discrete random variables 2 Outline Counting tricks and basic principles of probability
More informationBrief Review of Probability
Brief Review of Probability Nuno Vasconcelos (Ken Kreutz-Delgado) ECE Department, UCSD Probability Probability theory is a mathematical language to deal with processes or experiments that are non-deterministic
More informationSTAT 712 MATHEMATICAL STATISTICS I
STAT 72 MATHEMATICAL STATISTICS I Fall 207 Lecture Notes Joshua M. Tebbs Department of Statistics University of South Carolina c by Joshua M. Tebbs TABLE OF CONTENTS Contents Probability Theory. Set Theory......................................2
More informationChapter 1 Statistical Reasoning Why statistics? Section 1.1 Basics of Probability Theory
Chapter 1 Statistical Reasoning Why statistics? Uncertainty of nature (weather, earth movement, etc. ) Uncertainty in observation/sampling/measurement Variability of human operation/error imperfection
More informationSDS 321: Introduction to Probability and Statistics
SDS 321: Introduction to Probability and Statistics Lecture 14: Continuous random variables Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin www.cs.cmu.edu/
More informationCME 106: Review Probability theory
: Probability theory Sven Schmit April 3, 2015 1 Overview In the first half of the course, we covered topics from probability theory. The difference between statistics and probability theory is the following:
More informationECE 302 Division 2 Exam 2 Solutions, 11/4/2009.
NAME: ECE 32 Division 2 Exam 2 Solutions, /4/29. You will be required to show your student ID during the exam. This is a closed-book exam. A formula sheet is provided. No calculators are allowed. Total
More informationExpectation. DS GA 1002 Probability and Statistics for Data Science. Carlos Fernandez-Granda
Expectation DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Aim Describe random variables with a few numbers: mean,
More informationModule 1. Probability
Module 1 Probability 1. Introduction In our daily life we come across many processes whose nature cannot be predicted in advance. Such processes are referred to as random processes. The only way to derive
More informationMidterm Exam 1 Solution
EECS 126 Probability and Random Processes University of California, Berkeley: Fall 2015 Kannan Ramchandran September 22, 2015 Midterm Exam 1 Solution Last name First name SID Name of student on your left:
More informationReview of Probabilities and Basic Statistics
Alex Smola Barnabas Poczos TA: Ina Fiterau 4 th year PhD student MLD Review of Probabilities and Basic Statistics 10-701 Recitations 1/25/2013 Recitation 1: Statistics Intro 1 Overview Introduction to
More informationStatistics for scientists and engineers
Statistics for scientists and engineers February 0, 006 Contents Introduction. Motivation - why study statistics?................................... Examples..................................................3
More informationM378K In-Class Assignment #1
The following problems are a review of M6K. M7K In-Class Assignment # Problem.. Complete the definition of mutual exclusivity of events below: Events A, B Ω are said to be mutually exclusive if A B =.
More informationPoisson approximations
Chapter 9 Poisson approximations 9.1 Overview The Binn, p) can be thought of as the distribution of a sum of independent indicator random variables X 1 + + X n, with {X i = 1} denoting a head on the ith
More informationIntroduction to Information Entropy Adapted from Papoulis (1991)
Introduction to Information Entropy Adapted from Papoulis (1991) Federico Lombardo Papoulis, A., Probability, Random Variables and Stochastic Processes, 3rd edition, McGraw ill, 1991. 1 1. INTRODUCTION
More informationMASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 8 10/1/2008 CONTINUOUS RANDOM VARIABLES
MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 8 10/1/2008 CONTINUOUS RANDOM VARIABLES Contents 1. Continuous random variables 2. Examples 3. Expected values 4. Joint distributions
More informationLecture 2: Review of Basic Probability Theory
ECE 830 Fall 2010 Statistical Signal Processing instructor: R. Nowak, scribe: R. Nowak Lecture 2: Review of Basic Probability Theory Probabilistic models will be used throughout the course to represent
More informationLecture Notes 2 Random Variables. Random Variable
Lecture Notes 2 Random Variables Definition Discrete Random Variables: Probability mass function (pmf) Continuous Random Variables: Probability density function (pdf) Mean and Variance Cumulative Distribution
More informationMATH 151, FINAL EXAM Winter Quarter, 21 March, 2014
Time: 3 hours, 8:3-11:3 Instructions: MATH 151, FINAL EXAM Winter Quarter, 21 March, 214 (1) Write your name in blue-book provided and sign that you agree to abide by the honor code. (2) The exam consists
More informationProbability Theory for Machine Learning. Chris Cremer September 2015
Probability Theory for Machine Learning Chris Cremer September 2015 Outline Motivation Probability Definitions and Rules Probability Distributions MLE for Gaussian Parameter Estimation MLE and Least Squares
More informationIAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES
IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES VARIABLE Studying the behavior of random variables, and more importantly functions of random variables is essential for both the
More information3. Probability and Statistics
FE661 - Statistical Methods for Financial Engineering 3. Probability and Statistics Jitkomut Songsiri definitions, probability measures conditional expectations correlation and covariance some important
More information1 INFO Sep 05
Events A 1,...A n are said to be mutually independent if for all subsets S {1,..., n}, p( i S A i ) = p(a i ). (For example, flip a coin N times, then the events {A i = i th flip is heads} are mutually
More informationChapter 3: Random Variables 1
Chapter 3: Random Variables 1 Yunghsiang S. Han Graduate Institute of Communication Engineering, National Taipei University Taiwan E-mail: yshan@mail.ntpu.edu.tw 1 Modified from the lecture notes by Prof.
More informationLecture 1: August 28
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random
More information2.1 Elementary probability; random sampling
Chapter 2 Probability Theory Chapter 2 outlines the probability theory necessary to understand this text. It is meant as a refresher for students who need review and as a reference for concepts and theorems
More informationTom Salisbury
MATH 2030 3.00MW Elementary Probability Course Notes Part V: Independence of Random Variables, Law of Large Numbers, Central Limit Theorem, Poisson distribution Geometric & Exponential distributions Tom
More informationSet Theory Digression
1 Introduction to Probability 1.1 Basic Rules of Probability Set Theory Digression A set is defined as any collection of objects, which are called points or elements. The biggest possible collection of
More information1 Probability and Random Variables
1 Probability and Random Variables The models that you have seen thus far are deterministic models. For any time t, there is a unique solution X(t). On the other hand, stochastic models will result in
More informationProbability Theory Review
Cogsci 118A: Natural Computation I Lecture 2 (01/07/10) Lecturer: Angela Yu Probability Theory Review Scribe: Joseph Schilz Lecture Summary 1. Set theory: terms and operators In this section, we provide
More informationChapter 5. Chapter 5 sections
1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More information