Chapter 5. Chapter 5 sections

Size: px
Start display at page:

Download "Chapter 5. Chapter 5 sections"

Transcription

1 1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions Continuous univariate distributions: 5.6 Normal distributions 5.7 Gamma distributions Just skim 5.8 Beta distributions Multivariate distributions Just skim 5.9 Multinomial distributions 5.10 Bivariate normal distributions

2 2 / Introduction Families of distributions How: Why: Parameter and Parameter space pf /pdf and cdf - new notation: f (x parameters ) Mean, variance and the m.g.f. ψ(t) Features, connections to other distributions, approximation Reasoning behind a distribution Natural justification for certain experiments A model for the uncertainty in an experiment All models are wrong, but some are useful George Box

3 3 / 43 Bernoulli distributions 5.2 Bernoulli and Binomial distributions Def: Bernoulli distributions Bernoulli(p) A r.v. X has the Bernoulli distribution with parameter p if P(X = 1) = p and P(X = 0) = 1 p. The pf of X is { p f (x p) = x (1 p) 1 x for x = 0, 1 0 otherwise Parameter space: p [0, 1] In an experiment with only two possible outcomes, success and failure, let X = number successes. Then X Bernoulli(p) where p is the probability of success. E(X) = p, Var(X) = p(1 p) and ψ(t) = E(e tx ) = pe t + (1 p) 0 for x < 0 The cdf is F (x p) = 1 p for 0 x < 1 1 for x 1

4 4 / 43 Binomial distributions 5.2 Bernoulli and Binomial distributions Def: Binomial distributions Binomial(n, p) A r.v. X has the Binomial distribution with parameters n and p if X has the pf { ( n ) f (x n, p) = x p x (1 p) n x for x = 0, 1,..., n 0 otherwise Parameter space: n is a positive integer and p [0, 1] If X is the number of successes in n independent tries where prob. of success is p each time, then X Binomial(n, p) Theorem If X 1, X 2,..., X n form n Bernoulli trials with parameter p (i.e. are i.i.d. Bernoulli(p)) then X = X X n Binomial(n, p)

5 5.2 Bernoulli and Binomial distributions Binomial distributions Let X Binomial(n, p) E(X) = np, Var(X) = np(1 p) To find the m.g.f. of X write X = X X n where X i s are i.i.d. Bernoulli(p). Then ψ i (t) = pe t + 1 p and we get ψ(t) = n ψ i (t) = i=1 n ( pe t + 1 p ) = (pe t + 1 p) n i=1 cdf: F(x n, p) = x t=0 ( n t ) p t (1 p) n t = yikes! Theorem If X i Binomial(n i, p), i = 1,..., k and the X i s are independent, then X = X X k Binomial( k i=1 n i, p) 5 / 43

6 6 / Bernoulli and Binomial distributions Example: Blood testing (Example 5.2.7) The setup: 1000 people need to be tested for a disease that affects 0.2% of all people. The test is guaranteed to detect the disease if it is present in a blood sample. Task: Find all the people that have the disease. Strategy: Test 1000 samples What s the expected number of people that have the disease? Any assumptions you need to make? Strategy (611): Divide the people into 10 groups of 100. For each group take a portion of each of the 100 blood samples and combine into one sample. Then test the combined blood samples (10 tests).

7 7 / Bernoulli and Binomial distributions Example: Blood testing (Example 5.2.7) continued Strategy (611): If all of these tests are negative then none of the 1000 people have the disease. Total number of tests needed: 10 If one of these tests are positive then we test each of the 100 people in that group. Total number of tests needed: If all of the 10 tests are positive we end up having to do 1010 tests Is this strategy better? What is the expected number of tests needed? When does this strategy lose?

8 8 / Bernoulli and Binomial distributions Example: Blood testing (Example 5.2.7) continued Let Y i = 1 if test for group i is positive and Y i = 0 otherwise Let Y = Y Y 10 = the number of groups where every individual has to be tested. Total number of tests needed: T = Y. Let Z i = number of people in group i that have the disease, i = 1,..., 10. Then Z i Binomial(100, 0.002) Then Y i is a Bernoulli(p) r.v. where p = P(Y i = 1) = P(Z i > 0) = 1 P(Z i = 0) ( ) 100 = ( ) 100 = = Then Y Binomial(10, 0.181) ET = E(10+100Y ) = E(Y ) = ( ) = 191

9 9 / Bernoulli and Binomial distributions Example: Blood testing (Example 5.2.7) continued When does this strategy (611) lose? Worst case scenario P(T 1000) = P(Y 9.9) = P(Y = 10) = ( ) Question: can we go further - a 611-A strategy Any further improvement?

10 Hypergeometric distributions 5.3 Hypergeometric distributions Def: Hypergeometric distributions A random variable X has the Hypergeometric distribution with parameters N, M and n if it has the pf ( N )( M ) x n x f (x N, M, n) = ) ( N+M n Parameter space: N, M and n are nonnegative integers with n N + M Reasoning: Say we have a finite population with N items of type I and M items of type II. Let X be the number of items of type I when we take n samples without replacement from that population Then X has the hypergeometric distribution 10 / 43

11 11 / Hypergeometric distributions Hypergeometric distributions Binomial: Sampling with replacement (effectively infinite population) Hypergeometric: Sample without replacement from a finite population You can also think of the Hypergeometric distribution as a sum of dependent Bernoulli trials Limiting situation: Theorem 5.3.4: If the samples size n is much smaller than the total population N + M then the Hypergeometric distribution with parameters N, M and n will be nearly the same as the Binomial distribution with parameters n and p = N N + M

12 12 / 43 Poisson distributions 5.4 Poisson distributions Def: Poisson distributions Poisson(λ) A random variable X has the Poisson distribution with mean λ if it has the pf { e λ λ x f (x λ) = x! for x = 0, 1, otherwise Parameter space: λ > 0 Show that f (x λ) is a pf E(X) = λ Var(X) = λ ψ(t) = e λ(et 1) The cdf: F(x λ) = x k=0 e λ λ k k! = yikes.

13 13 / Poisson distributions Why Poisson? The Poisson distribution is useful for modeling uncertainty in counts / arrivals Examples: How many calls arrive at a switch board in one hour? How many busses pass while you wait at the bus stop for 10 min? How many bird nests are there in a certain area? Under certain conditions (Poisson postulates) the Poisson distribution can be shown to be the distribution of the number of arrivals (Poisson process). However, the Poisson distribution is often used as a model for uncertainty of counts in other types of experiments. The Poisson distribution can also be used as an approximation to the Binomial(n, p) distribution when n is large and p is small.

14 Poisson Postulates 5.4 Poisson distributions For t 0, let X t be a random variable with possible values in N 0 (Think: X t = number of arrivals from time 0 to time t) (i) Start with no arrivals: X 0 = 0 (ii) Arrivals in disjoint time periods are ind.: X s and X t X s ind. if s < t (iii) Number of arrivals depends only on period length: X s and X t+s X t are identically distributed (iv) Arrival probability is proportional to period length, if length is small: P(X t = 1) lim = λ t 0 t (v) No simultaneous arrivals: lim t 0 P(X t >1) t = 0 If (i) - (v) hold then for any integer n λt (λt)n P(X t = n) = e that is, X t Poisson(λt) n! Can be defined in terms of spatial areas too. 14 / 43

15 15 / Poisson distributions Properties of the Poisson Distributions Useful recursive property: P(X = x) = λ x P(X = x 1) for x 1 Theorem 5.4.4: Sum of Poissons is a Poisson If X 1,..., X k are independent r.v. and X i Poisson(λ i ) for all i, then ( k ) X X k Poisson λ i Theorem 5.4.5: Approximation to Binomial Let X n Binomial(n, p n ), where 0 < p n < 1 for all n and {p n } n=1 is a sequence so that lim n np n = λ. Then for all x = 0, 1, 2,... i=1 lim f λ λx X n n (x n, p n ) = e x! = f Poisson (x λ)

16 16 / Poisson distributions Example: Poisson as approximation to Binomial Recall the disease testing example. We had 1000 X = X i Binomial(1000, 0.002) and i=1 Y Binomial(100, 0.181)

17 17 / 43 Geometric distributions 5.5 Negative Binomial distributions Def: Geometric distributions Geometric(p) A random variable X has the Geometric distribution with parameter p if it has the pf { p(1 p) x for x = 0, 1, 2... f (x r, p) = 0 otherwise Parameter space: 0 < p < 1 Say we have an infinite sequence of Bernoulli trials with parameter p X = number of failures before the first success. Then X Geometric(p)

18 18 / 43 Negative Binomial distributions 5.5 Negative Binomial distributions Def: Negative Binomial distributions NegBinomial(r, p) A random variable X has the Negative Binomial distribution with parameters r and p if it has the pf { ( r+x 1 ) f (x r, p) = x p r (1 p) x for x = 0, 1, otherwise Parameter space: 0 < p < 1 and r positive integer. Say we have an infinite sequence of Bernoulli trials with parameter p X = number of failures before the r th success. Then X NegBinomial(r, p) Geometric(p) = NegBinomial(1, p) Theorem 5.5.2: If X 1,..., X r are i.i.d. Geometric(p) then X = X X r NegBinomial(r, p)

19 19 / Negative Binomial distributions sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions Continuous univariate distributions: 5.6 Normal distributions 5.7 Gamma distributions Just skim 5.8 Beta distributions Multivariate distributions Just skim 5.9 Multinomial distributions 5.10 Bivariate normal distributions

20 20 / 43 Gamma distributions 5.7 Gamma distributions The Gamma function: Γ(α) = 0 x α 1 e x dx Γ(1) = 1 and Γ(0.5) = π Γ(α) = (α 1)Γ(α 1) if α > 1 Def: Gamma distributions Gamma(α, β) A continuous r.v. X has the gamma distribution with parameters α and β if it has the pdf { β α f (x α, β) = Γ(α) x α 1 e βx for x > 0 0 otherwise Parameter space: α > 0 and β > 0 Gamma(1, β) is the same as the exponential distribution with parameter β, Expo(β)

21 21 / Gamma distributions Properties of the gamma distributions ψ(t) = ( β β t ) α, for t < β. E(X) = α β and E(X) = α β 2 If X 1,..., X k are independent Γ(α i, β) r.v. then ( k ) X X k Gamma α i, β i=1

22 22 / Gamma distributions Properties of the gamma distributions Theorem 5.7.9: Exponential distribution is memoryless Let X Expo(β) and let t > 0. Then for any h > 0 P(X t + h X t) = P(X h) Theorem : Times between arrivals in a Poisson process Let Z k be the time until the k th arrival in a Poisson process with rate β. Let Y 1 = Z 1 and Y k = Z k Z k 1 for k 2. Then Y 1, Y 2, Y 3,... are i.i.d. with the exponential distribution with parameter β.

23 23 / 43 Beta distributions 5.8 Beta distributions Def: Beta distributions Beta(α, β) A continuous r.v. X has the beta distribution with parameters α and β if it has the pdf { Γ(α+β) f (x α, β) = Γ(α)Γ(β) x α 1 (1 x) β 1 for 0 < x < 1 0 otherwise Parameter space: α > 0 and β > 0 Beta(1, 1) = Uniform(0, 1) Used to model a r.v.that takes values between 0 and 1. The Beta distributions are often used as prior distributions for probability parameters, e.g. the p in the Binomial distribution.

24 24 / 43 Beta distributions 5.8 Beta distributions

25 25 / Beta distributions sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions Continuous univariate distributions: 5.6 Normal distributions 5.7 Gamma distributions Just skim 5.8 Beta distributions Multivariate distributions Just skim 5.9 Multinomial distributions 5.10 Bivariate normal distributions

26 26 / Normal distributions Why Normal? Works well in practice. Many physical experiments have distributions that are approximately normal Central Limit Theorem: Sum of many i.i.d. random variables are approximately normally distributed Mathematically convenient especially the multivariate normal distribution. Can explicitly obtain the distribution of many functions of a normally distributed random variable have. Marginal and conditional distributions of a multivariate normal are also normal (multivariate or univariate). Developed by Gauss and then Laplace in the early 1800s Also known at the Gaussian distributions Gauss Laplace

27 27 / 43 Normal distributions 5.6 Normal distributions Def: Normal distributions N(µ, σ 2 ) A continuous r.v. X has the normal distribution with mean µ and variance σ 2 if it has the pdf f (x µ, σ 2 ) = 1 ) (x µ)2 exp ( 2π σ 2σ 2, < x < Parameter space: µ R and σ 2 > 0 Show: ψ(t) = exp ( µt σ2 t 2) E(X) = µ Var(X) = σ 2

28 28 / 43 The Bell curve 5.6 Normal distributions

29 29 / 43 Standard normal 5.6 Normal distributions Standard normal distribution: N(0, 1) The normal distribution with µ = 0 and σ 2 = 1 is called the standard normal distribution and the pdf and cdf are denoted as φ(x) and Φ(x) The cdf for a normal distribution cannot be expressed in closed form and is evaluated using numerical approximations. Φ(x) is tabulated in the back of the book. Many calculators and programs such as R, Matlab, Excel etc. can calculate Φ(x). Φ( x) = 1 Φ(x) Φ 1 (p) = Φ 1 (1 p)

30 30 / Normal distributions Properties of the normal distributions Theorem 5.6.4: Linear transformation of a normal is still normal If X N(µ, σ 2 ) and Y = ax + b where a and b are constants and a 0 then Y N(aµ + b, a 2 σ 2 ) Let F be the cdf of X, where X N(µ, σ 2 ). Then ( ) x µ F(x) = Φ σ and F 1 (p) = µ + σφ 1 (p)

31 31 / 43 Example: Measured Voltage 5.6 Normal distributions Suppose the measured voltage, X, in a certain electric circuit has the normal distribution with mean 120 and standard deviation 2 1 What is the probability that the measured voltage is between 118 and 122? 2 Below what value will 95% of the measurements be?

32 32 / Normal distributions Properties of the normal distributions Theorem 5.6.7: Linear combination of ind. normals is a normal Let X 1,..., X k be independent r.v. and X i N(µ i, σi 2 ) for i = 1,..., k. Then ( ) X X k N µ µ k, σ σ2 k Also, if a 1,..., a k and b are constants where at least one a i is not zero: In particular: a 1 X a k X k + b N ( b + k a i µ i, i=1 k i=1 a 2 i σ2 i The sample mean: X n = 1 n n i=1 X i If X 1,..., X n are a random sample from a N(µ, σ 2 ), what is the distribution of the sample mean? )

33 33 / Normal distributions Example: Measured voltage continued Suppose the measured voltage, X, in a certain electric circuit has the normal distribution with mean 120 and standard deviation 2. If three independent measurements of the voltage are made, what is the probability that the sample mean X 3 will lie between 118 and 120? Find x that satisfies P( X x) = 0.95

34 34 / 43 Area under the curve 5.6 Normal distributions

35 35 / Normal distributions Lognormal distributions Def: Lognormal distributions If log(x) N(µ, σ 2 ) then we say that X has the Lognormal distribution with parameters µ and σ 2. The support of the lognormal distribution is (0, ). Often used to model time before failure. Example: Let X and Y be independent random variables such that log(x) N(1.6, 4.5) and log(y ) N(3, 6). What is the distribution of the product XY?

36 36 / 43 Bivariate normal distributions 5.10 Bivariate normal distributions Def: Bivariate normal Two continuous r.v. X 1 and X 2 have the bivariate normal distribution with means µ 1 and µ 2, variances σ1 2 and σ2 2 and correlation ρ if they have the joint pdf 1 f (x 1, x 2 ) = 2π(1 ρ 2 ) 1/2 σ 1 σ 2 ( [ ( ) ( ) 1 (x1 µ 1 ) 2 x1 µ 1 x2 µ 2 exp 2ρ + (x ]) 2 µ 2 ) 2 2(1 ρ 2 ) σ 2 1 σ 1 σ 2 σ 2 2 (1) Parameter space: µ i R, σ 2 i > 0 for i = 1, 2 and 1 ρ 1

37 37 / 43 Bivariate normal pdf 5.10 Bivariate normal distributions Bivariate normal pdf with different ρ: Contours:

38 5.10 Bivariate normal distributions Bivariate normal as linear combination Theorem : Bivariate normal from two ind. standard normals Let Z 1 N(0, 1) and Z 2 N(0, 1) be independent. Let µ i R, σi 2 > 0 for i = 1, 2 and 1 ρ 1 and let X 1 = σ 1 Z 1 + µ 1 X 2 = σ 2 (ρz ρ 2 Z 2 ) + µ 2 (2) Then the joint distribution of X 1 and X 2 is bivariate normal with parameters µ 1, µ 2, σ 2 1, σ2 2 and ρ Theorem (part 1) the other way Let X 1 and X 2 have the pdf in (1). Then there exist independent standard normal r.v. Z 1 and Z 2 so that (2) holds. 38 / 43

39 39 / 43 Properties of a bivariate normal 5.10 Bivariate normal distributions Theorem (part 2) Let X 1 and X 2 have the pdf in (1). Then the marginal distributions are X 1 N(µ 1, σ 2 1 ) and X 2 N(µ 2, σ 2 2 ) And the correlation between X 1 and X 2 is ρ Theorem : The conditional is normal Let X 1 and X 2 have the pdf in (1). Then the conditional distribution of X 2 given that X 1 = x 1 is (univariate) normal with E(X 2 X 1 = x 1 ) = µ 2 + ρσ 2 (x 1 µ 1 ) σ 1 and Var(X 2 X 1 = x 1 ) = (1 ρ 2 )σ 2 2

40 5.10 Bivariate normal distributions Properties of a bivariate normal Theorem : Uncorrelated Independent Let X 1 and X 2 have the bivariate normal distribution. Then X 1 and X 2 are independent if and only if they are uncorrelated. Only holds for the multivariate normal distribution One of the very convenient properties of the normal distribution Theorem : Linear combinations are normal Let X 1 and X 2 have the pdf in (1) and let a 1, a 2 and b be constants. Then Y = a 1 X 1 + a 2 X 2 + b is normally distributed with E(Y ) = a 1 µ 1 + a 2 µ 2 + b and Var(Y ) = a 2 1 σ2 1 + a2 2 σ a 1a 2 ρσ 1 σ 2 This extends what we already had for independent normals 40 / 43

41 41 / 43 Example 5.10 Bivariate normal distributions Let X 1 and X 2 have the bivariate normal distribution with means µ 1 = 3, µ 2 = 5, variances σ1 2 = 4, σ2 2 = 9 and correlation ρ = 0.6. a) Find the distribution of X 2 2X 1 b) What is expected value of X 2, given that we observed X 1 = 2? c) What is the probability that X 1 > X 2?

42 42 / Bivariate normal distributions Multivariate normal Matrix notation The pdf of an n-dimensional normal distribution, X N(µ, Σ): { 1 f (x) = (2π) n/2 exp 1 } Σ 1/2 2 (x µ) Σ 1 (x µ) where µ 1 x 1 µ 2 µ =., x = x 2. µ n x n σ 2 1 σ 1,2 σ 1,3 σ 1,n σ 2,1 σ2 2 σ 2,3 σ 2,n and Σ = σ 3,1 σ 3,2 σ3 2 σ 3,n σ n,1 σ n,2 σ n,3 σn 2 µ is the mean vector and Σ is called the variance-covariance matrix.

43 43 / Bivariate normal distributions Multivariate normal Matrix notation Same things hold for multivariate normal distribution as the bivariate. Let X N(µ, Σ) Linear combinations of X are normal AX + b is (multivariate) normal for fixed matrix A and vector b The marginal distribution of X i is normal with mean µ i and variance σ 2 i The off-diagonal elements of Σ are the covariances between individual elements of X, i.e. Cov(X i, X j ) = σ i,j. The joint marginal distributions are also normal where the mean and covariance matrix are found by picking the corresponding elements from µ and rows and columns from Σ. The conditional distributions are also normal (multivariate or univariate)

Chapter 5 continued. Chapter 5 sections

Chapter 5 continued. Chapter 5 sections Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions

More information

Things to remember when learning probability distributions:

Things to remember when learning probability distributions: SPECIAL DISTRIBUTIONS Some distributions are special because they are useful They include: Poisson, exponential, Normal (Gaussian), Gamma, geometric, negative binomial, Binomial and hypergeometric distributions

More information

15 Discrete Distributions

15 Discrete Distributions Lecture Note 6 Special Distributions (Discrete and Continuous) MIT 4.30 Spring 006 Herman Bennett 5 Discrete Distributions We have already seen the binomial distribution and the uniform distribution. 5.

More information

Stat 5101 Notes: Brand Name Distributions

Stat 5101 Notes: Brand Name Distributions Stat 5101 Notes: Brand Name Distributions Charles J. Geyer February 14, 2003 1 Discrete Uniform Distribution DiscreteUniform(n). Discrete. Rationale Equally likely outcomes. The interval 1, 2,..., n of

More information

Statistics STAT:5100 (22S:193), Fall Sample Final Exam B

Statistics STAT:5100 (22S:193), Fall Sample Final Exam B Statistics STAT:5 (22S:93), Fall 25 Sample Final Exam B Please write your answers in the exam books provided.. Let X, Y, and Y 2 be independent random variables with X N(µ X, σ 2 X ) and Y i N(µ Y, σ 2

More information

Random Variables and Their Distributions

Random Variables and Their Distributions Chapter 3 Random Variables and Their Distributions A random variable (r.v.) is a function that assigns one and only one numerical value to each simple event in an experiment. We will denote r.vs by capital

More information

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued Chapter 3 sections Chapter 3 - continued 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions

More information

Lecture 1: August 28

Lecture 1: August 28 36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random

More information

STAT Chapter 5 Continuous Distributions

STAT Chapter 5 Continuous Distributions STAT 270 - Chapter 5 Continuous Distributions June 27, 2012 Shirin Golchi () STAT270 June 27, 2012 1 / 59 Continuous rv s Definition: X is a continuous rv if it takes values in an interval, i.e., range

More information

Mathematical Statistics 1 Math A 6330

Mathematical Statistics 1 Math A 6330 Mathematical Statistics 1 Math A 6330 Chapter 3 Common Families of Distributions Mohamed I. Riffi Department of Mathematics Islamic University of Gaza September 28, 2015 Outline 1 Subjects of Lecture 04

More information

Probability and Distributions

Probability and Distributions Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated

More information

Stat 5101 Notes: Brand Name Distributions

Stat 5101 Notes: Brand Name Distributions Stat 5101 Notes: Brand Name Distributions Charles J. Geyer September 5, 2012 Contents 1 Discrete Uniform Distribution 2 2 General Discrete Uniform Distribution 2 3 Uniform Distribution 3 4 General Uniform

More information

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued Chapter 3 sections 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions 3.6 Conditional

More information

Probability Distributions Columns (a) through (d)

Probability Distributions Columns (a) through (d) Discrete Probability Distributions Columns (a) through (d) Probability Mass Distribution Description Notes Notation or Density Function --------------------(PMF or PDF)-------------------- (a) (b) (c)

More information

1 Presessional Probability

1 Presessional Probability 1 Presessional Probability Probability theory is essential for the development of mathematical models in finance, because of the randomness nature of price fluctuations in the markets. This presessional

More information

Lecture 2: Repetition of probability theory and statistics

Lecture 2: Repetition of probability theory and statistics Algorithms for Uncertainty Quantification SS8, IN2345 Tobias Neckel Scientific Computing in Computer Science TUM Lecture 2: Repetition of probability theory and statistics Concept of Building Block: Prerequisites:

More information

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Review of Basic Probability The fundamentals, random variables, probability distributions Probability mass/density functions

More information

Continuous Random Variables and Continuous Distributions

Continuous Random Variables and Continuous Distributions Continuous Random Variables and Continuous Distributions Continuous Random Variables and Continuous Distributions Expectation & Variance of Continuous Random Variables ( 5.2) The Uniform Random Variable

More information

Contents 1. Contents

Contents 1. Contents Contents 1 Contents 6 Distributions of Functions of Random Variables 2 6.1 Transformation of Discrete r.v.s............. 3 6.2 Method of Distribution Functions............. 6 6.3 Method of Transformations................

More information

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

Chapter 2. Discrete Distributions

Chapter 2. Discrete Distributions Chapter. Discrete Distributions Objectives ˆ Basic Concepts & Epectations ˆ Binomial, Poisson, Geometric, Negative Binomial, and Hypergeometric Distributions ˆ Introduction to the Maimum Likelihood Estimation

More information

Chapter 4. Chapter 4 sections

Chapter 4. Chapter 4 sections Chapter 4 sections 4.1 Expectation 4.2 Properties of Expectations 4.3 Variance 4.4 Moments 4.5 The Mean and the Median 4.6 Covariance and Correlation 4.7 Conditional Expectation SKIP: 4.8 Utility Expectation

More information

More on Distribution Function

More on Distribution Function More on Distribution Function The distribution of a random variable X can be determined directly from its cumulative distribution function F X. Theorem: Let X be any random variable, with cumulative distribution

More information

1.1 Review of Probability Theory

1.1 Review of Probability Theory 1.1 Review of Probability Theory Angela Peace Biomathemtics II MATH 5355 Spring 2017 Lecture notes follow: Allen, Linda JS. An introduction to stochastic processes with applications to biology. CRC Press,

More information

Formulas for probability theory and linear models SF2941

Formulas for probability theory and linear models SF2941 Formulas for probability theory and linear models SF2941 These pages + Appendix 2 of Gut) are permitted as assistance at the exam. 11 maj 2008 Selected formulae of probability Bivariate probability Transforms

More information

Continuous Probability Spaces

Continuous Probability Spaces Continuous Probability Spaces Ω is not countable. Outcomes can be any real number or part of an interval of R, e.g. heights, weights and lifetimes. Can not assign probabilities to each outcome and add

More information

Some Special Distributions (Hogg Chapter Three)

Some Special Distributions (Hogg Chapter Three) Some Special Distributions Hogg Chapter Three STAT 45-: Mathematical Statistics I Fall Semester 5 Contents Binomial & Related Distributions. The Binomial Distribution............... Some advice on calculating

More information

Actuarial Science Exam 1/P

Actuarial Science Exam 1/P Actuarial Science Exam /P Ville A. Satopää December 5, 2009 Contents Review of Algebra and Calculus 2 2 Basic Probability Concepts 3 3 Conditional Probability and Independence 4 4 Combinatorial Principles,

More information

1 Review of Probability and Distributions

1 Review of Probability and Distributions Random variables. A numerically valued function X of an outcome ω from a sample space Ω X : Ω R : ω X(ω) is called a random variable (r.v.), and usually determined by an experiment. We conventionally denote

More information

ACM 116: Lectures 3 4

ACM 116: Lectures 3 4 1 ACM 116: Lectures 3 4 Joint distributions The multivariate normal distribution Conditional distributions Independent random variables Conditional distributions and Monte Carlo: Rejection sampling Variance

More information

Probability Theory. Patrick Lam

Probability Theory. Patrick Lam Probability Theory Patrick Lam Outline Probability Random Variables Simulation Important Distributions Discrete Distributions Continuous Distributions Most Basic Definition of Probability: number of successes

More information

Brief Review of Probability

Brief Review of Probability Maura Department of Economics and Finance Università Tor Vergata Outline 1 Distribution Functions Quantiles and Modes of a Distribution 2 Example 3 Example 4 Distributions Outline Distribution Functions

More information

APM 504: Probability Notes. Jay Taylor Spring Jay Taylor (ASU) APM 504 Fall / 65

APM 504: Probability Notes. Jay Taylor Spring Jay Taylor (ASU) APM 504 Fall / 65 APM 504: Probability Notes Jay Taylor Spring 2015 Jay Taylor (ASU) APM 504 Fall 2013 1 / 65 Outline Outline 1 Probability and Uncertainty 2 Random Variables Discrete Distributions Continuous Distributions

More information

Stat410 Probability and Statistics II (F16)

Stat410 Probability and Statistics II (F16) Stat4 Probability and Statistics II (F6 Exponential, Poisson and Gamma Suppose on average every /λ hours, a Stochastic train arrives at the Random station. Further we assume the waiting time between two

More information

Covariance. Lecture 20: Covariance / Correlation & General Bivariate Normal. Covariance, cont. Properties of Covariance

Covariance. Lecture 20: Covariance / Correlation & General Bivariate Normal. Covariance, cont. Properties of Covariance Covariance Lecture 0: Covariance / Correlation & General Bivariate Normal Sta30 / Mth 30 We have previously discussed Covariance in relation to the variance of the sum of two random variables Review Lecture

More information

Tom Salisbury

Tom Salisbury MATH 2030 3.00MW Elementary Probability Course Notes Part V: Independence of Random Variables, Law of Large Numbers, Central Limit Theorem, Poisson distribution Geometric & Exponential distributions Tom

More information

Some Special Distributions (Hogg Chapter Three)

Some Special Distributions (Hogg Chapter Three) Some Special Distributions Hogg Chapter Three STAT 45-: Mathematical Statistics I Fall Semester 3 Contents Binomial & Related Distributions The Binomial Distribution Some advice on calculating n 3 k Properties

More information

Lecture 5: Moment generating functions

Lecture 5: Moment generating functions Lecture 5: Moment generating functions Definition 2.3.6. The moment generating function (mgf) of a random variable X is { x e tx f M X (t) = E(e tx X (x) if X has a pmf ) = etx f X (x)dx if X has a pdf

More information

2. Suppose (X, Y ) is a pair of random variables uniformly distributed over the triangle with vertices (0, 0), (2, 0), (2, 1).

2. Suppose (X, Y ) is a pair of random variables uniformly distributed over the triangle with vertices (0, 0), (2, 0), (2, 1). Name M362K Final Exam Instructions: Show all of your work. You do not have to simplify your answers. No calculators allowed. There is a table of formulae on the last page. 1. Suppose X 1,..., X 1 are independent

More information

Lecture 13. Poisson Distribution. Text: A Course in Probability by Weiss 5.5. STAT 225 Introduction to Probability Models February 16, 2014

Lecture 13. Poisson Distribution. Text: A Course in Probability by Weiss 5.5. STAT 225 Introduction to Probability Models February 16, 2014 Lecture 13 Text: A Course in Probability by Weiss 5.5 STAT 225 Introduction to Probability Models February 16, 2014 Whitney Huang Purdue University 13.1 Agenda 1 2 3 13.2 Review So far, we have seen discrete

More information

Common probability distributionsi Math 217 Probability and Statistics Prof. D. Joyce, Fall 2014

Common probability distributionsi Math 217 Probability and Statistics Prof. D. Joyce, Fall 2014 Introduction. ommon probability distributionsi Math 7 Probability and Statistics Prof. D. Joyce, Fall 04 I summarize here some of the more common distributions used in probability and statistics. Some

More information

SDS 321: Introduction to Probability and Statistics

SDS 321: Introduction to Probability and Statistics SDS 321: Introduction to Probability and Statistics Lecture 14: Continuous random variables Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin www.cs.cmu.edu/

More information

Common ontinuous random variables

Common ontinuous random variables Common ontinuous random variables CE 311S Earlier, we saw a number of distribution families Binomial Negative binomial Hypergeometric Poisson These were useful because they represented common situations:

More information

t x 1 e t dt, and simplify the answer when possible (for example, when r is a positive even number). In particular, confirm that EX 4 = 3.

t x 1 e t dt, and simplify the answer when possible (for example, when r is a positive even number). In particular, confirm that EX 4 = 3. Mathematical Statistics: Homewor problems General guideline. While woring outside the classroom, use any help you want, including people, computer algebra systems, Internet, and solution manuals, but mae

More information

Chapter 7: Special Distributions

Chapter 7: Special Distributions This chater first resents some imortant distributions, and then develos the largesamle distribution theory which is crucial in estimation and statistical inference Discrete distributions The Bernoulli

More information

Continuous Distributions

Continuous Distributions A normal distribution and other density functions involving exponential forms play the most important role in probability and statistics. They are related in a certain way, as summarized in a diagram later

More information

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ).

x. Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ 2 ). .8.6 µ =, σ = 1 µ = 1, σ = 1 / µ =, σ =.. 3 1 1 3 x Figure 1: Examples of univariate Gaussian pdfs N (x; µ, σ ). The Gaussian distribution Probably the most-important distribution in all of statistics

More information

1 Solution to Problem 2.1

1 Solution to Problem 2.1 Solution to Problem 2. I incorrectly worked this exercise instead of 2.2, so I decided to include the solution anyway. a) We have X Y /3, which is a - function. It maps the interval, ) where X lives) onto

More information

Conditional distributions (discrete case)

Conditional distributions (discrete case) Conditional distributions (discrete case) The basic idea behind conditional distributions is simple: Suppose (XY) is a jointly-distributed random vector with a discrete joint distribution. Then we can

More information

Lecture 6: Special probability distributions. Summarizing probability distributions. Let X be a random variable with probability distribution

Lecture 6: Special probability distributions. Summarizing probability distributions. Let X be a random variable with probability distribution Econ 514: Probability and Statistics Lecture 6: Special probability distributions Summarizing probability distributions Let X be a random variable with probability distribution P X. We consider two types

More information

Review 1: STAT Mark Carpenter, Ph.D. Professor of Statistics Department of Mathematics and Statistics. August 25, 2015

Review 1: STAT Mark Carpenter, Ph.D. Professor of Statistics Department of Mathematics and Statistics. August 25, 2015 Review : STAT 36 Mark Carpenter, Ph.D. Professor of Statistics Department of Mathematics and Statistics August 25, 25 Support of a Random Variable The support of a random variable, which is usually denoted

More information

2 Random Variable Generation

2 Random Variable Generation 2 Random Variable Generation Most Monte Carlo computations require, as a starting point, a sequence of i.i.d. random variables with given marginal distribution. We describe here some of the basic methods

More information

Advanced topics from statistics

Advanced topics from statistics Advanced topics from statistics Anders Ringgaard Kristensen Advanced Herd Management Slide 1 Outline Covariance and correlation Random vectors and multivariate distributions The multinomial distribution

More information

(y 1, y 2 ) = 12 y3 1e y 1 y 2 /2, y 1 > 0, y 2 > 0 0, otherwise.

(y 1, y 2 ) = 12 y3 1e y 1 y 2 /2, y 1 > 0, y 2 > 0 0, otherwise. 54 We are given the marginal pdfs of Y and Y You should note that Y gamma(4, Y exponential( E(Y = 4, V (Y = 4, E(Y =, and V (Y = 4 (a With U = Y Y, we have E(U = E(Y Y = E(Y E(Y = 4 = (b Because Y and

More information

Quick Tour of Basic Probability Theory and Linear Algebra

Quick Tour of Basic Probability Theory and Linear Algebra Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra CS224w: Social and Information Network Analysis Fall 2011 Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra Outline Definitions

More information

Probability Models. 4. What is the definition of the expectation of a discrete random variable?

Probability Models. 4. What is the definition of the expectation of a discrete random variable? 1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions

More information

Continuous Random Variables

Continuous Random Variables Continuous Random Variables Recall: For discrete random variables, only a finite or countably infinite number of possible values with positive probability. Often, there is interest in random variables

More information

It can be shown that if X 1 ;X 2 ;:::;X n are independent r.v. s with

It can be shown that if X 1 ;X 2 ;:::;X n are independent r.v. s with Example: Alternative calculation of mean and variance of binomial distribution A r.v. X has the Bernoulli distribution if it takes the values 1 ( success ) or 0 ( failure ) with probabilities p and (1

More information

Multivariate Random Variable

Multivariate Random Variable Multivariate Random Variable Author: Author: Andrés Hincapié and Linyi Cao This Version: August 7, 2016 Multivariate Random Variable 3 Now we consider models with more than one r.v. These are called multivariate

More information

Probability. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. August 2014

Probability. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. August 2014 Probability Machine Learning and Pattern Recognition Chris Williams School of Informatics, University of Edinburgh August 2014 (All of the slides in this course have been adapted from previous versions

More information

Joint p.d.f. and Independent Random Variables

Joint p.d.f. and Independent Random Variables 1 Joint p.d.f. and Independent Random Variables Let X and Y be two discrete r.v. s and let R be the corresponding space of X and Y. The joint p.d.f. of X = x and Y = y, denoted by f(x, y) = P(X = x, Y

More information

LIST OF FORMULAS FOR STK1100 AND STK1110

LIST OF FORMULAS FOR STK1100 AND STK1110 LIST OF FORMULAS FOR STK1100 AND STK1110 (Version of 11. November 2015) 1. Probability Let A, B, A 1, A 2,..., B 1, B 2,... be events, that is, subsets of a sample space Ω. a) Axioms: A probability function

More information

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models Fatih Cavdur fatihcavdur@uludag.edu.tr March 20, 2012 Introduction Introduction The world of the model-builder

More information

Continuous Random Variables

Continuous Random Variables 1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables

More information

Chapter 5 Joint Probability Distributions

Chapter 5 Joint Probability Distributions Applied Statistics and Probability for Engineers Sixth Edition Douglas C. Montgomery George C. Runger Chapter 5 Joint Probability Distributions 5 Joint Probability Distributions CHAPTER OUTLINE 5-1 Two

More information

THE QUEEN S UNIVERSITY OF BELFAST

THE QUEEN S UNIVERSITY OF BELFAST THE QUEEN S UNIVERSITY OF BELFAST 0SOR20 Level 2 Examination Statistics and Operational Research 20 Probability and Distribution Theory Wednesday 4 August 2002 2.30 pm 5.30 pm Examiners { Professor R M

More information

Bivariate distributions

Bivariate distributions Bivariate distributions 3 th October 017 lecture based on Hogg Tanis Zimmerman: Probability and Statistical Inference (9th ed.) Bivariate Distributions of the Discrete Type The Correlation Coefficient

More information

3. Probability and Statistics

3. Probability and Statistics FE661 - Statistical Methods for Financial Engineering 3. Probability and Statistics Jitkomut Songsiri definitions, probability measures conditional expectations correlation and covariance some important

More information

Topic 2: Review of Probability Theory

Topic 2: Review of Probability Theory CS 8850: Advanced Machine Learning Fall 2017 Topic 2: Review of Probability Theory Instructor: Daniel L. Pimentel-Alarcón c Copyright 2017 2.1 Why Probability? Many (if not all) applications of machine

More information

Review for the previous lecture

Review for the previous lecture Lecture 1 and 13 on BST 631: Statistical Theory I Kui Zhang, 09/8/006 Review for the previous lecture Definition: Several discrete distributions, including discrete uniform, hypergeometric, Bernoulli,

More information

PCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities

PCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities PCMI 207 - Introduction to Random Matrix Theory Handout #2 06.27.207 REVIEW OF PROBABILITY THEORY Chapter - Events and Their Probabilities.. Events as Sets Definition (σ-field). A collection F of subsets

More information

Test Problems for Probability Theory ,

Test Problems for Probability Theory , 1 Test Problems for Probability Theory 01-06-16, 010-1-14 1. Write down the following probability density functions and compute their moment generating functions. (a) Binomial distribution with mean 30

More information

Slides 8: Statistical Models in Simulation

Slides 8: Statistical Models in Simulation Slides 8: Statistical Models in Simulation Purpose and Overview The world the model-builder sees is probabilistic rather than deterministic: Some statistical model might well describe the variations. An

More information

Order Statistics. The order statistics of a set of random variables X 1, X 2,, X n are the same random variables arranged in increasing order.

Order Statistics. The order statistics of a set of random variables X 1, X 2,, X n are the same random variables arranged in increasing order. Order Statistics The order statistics of a set of random variables 1, 2,, n are the same random variables arranged in increasing order. Denote by (1) = smallest of 1, 2,, n (2) = 2 nd smallest of 1, 2,,

More information

MAS223 Statistical Inference and Modelling Exercises

MAS223 Statistical Inference and Modelling Exercises MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,

More information

Stat 100a, Introduction to Probability.

Stat 100a, Introduction to Probability. Stat 100a, Introduction to Probability. Outline for the day: 1. Geometric random variables. 2. Negative binomial random variables. 3. Moment generating functions. 4. Poisson random variables. 5. Continuous

More information

b. ( ) ( ) ( ) ( ) ( ) 5. Independence: Two events (A & B) are independent if one of the conditions listed below is satisfied; ( ) ( ) ( )

b. ( ) ( ) ( ) ( ) ( ) 5. Independence: Two events (A & B) are independent if one of the conditions listed below is satisfied; ( ) ( ) ( ) 1. Set a. b. 2. Definitions a. Random Experiment: An experiment that can result in different outcomes, even though it is performed under the same conditions and in the same manner. b. Sample Space: This

More information

Chapters 3.2 Discrete distributions

Chapters 3.2 Discrete distributions Chapters 3.2 Discrete distributions In this section we study several discrete distributions and their properties. Here are a few, classified by their support S X. There are of course many, many more. For

More information

Discrete Distributions

Discrete Distributions Chapter 2 Discrete Distributions 2.1 Random Variables of the Discrete Type An outcome space S is difficult to study if the elements of S are not numbers. However, we can associate each element/outcome

More information

Chapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University

Chapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University Chapter 3, 4 Random Variables ENCS6161 - Probability and Stochastic Processes Concordia University ENCS6161 p.1/47 The Notion of a Random Variable A random variable X is a function that assigns a real

More information

Definition: A random variable X is a real valued function that maps a sample space S into the space of real numbers R. X : S R

Definition: A random variable X is a real valued function that maps a sample space S into the space of real numbers R. X : S R Random Variables Definition: A random variable X is a real valued function that maps a sample space S into the space of real numbers R. X : S R As such, a random variable summarizes the outcome of an experiment

More information

Lecture 14: Multivariate mgf s and chf s

Lecture 14: Multivariate mgf s and chf s Lecture 14: Multivariate mgf s and chf s Multivariate mgf and chf For an n-dimensional random vector X, its mgf is defined as M X (t) = E(e t X ), t R n and its chf is defined as φ X (t) = E(e ıt X ),

More information

Algorithms for Uncertainty Quantification

Algorithms for Uncertainty Quantification Algorithms for Uncertainty Quantification Tobias Neckel, Ionuț-Gabriel Farcaș Lehrstuhl Informatik V Summer Semester 2017 Lecture 2: Repetition of probability theory and statistics Example: coin flip Example

More information

MATH : EXAM 2 INFO/LOGISTICS/ADVICE

MATH : EXAM 2 INFO/LOGISTICS/ADVICE MATH 3342-004: EXAM 2 INFO/LOGISTICS/ADVICE INFO: WHEN: Friday (03/11) at 10:00am DURATION: 50 mins PROBLEM COUNT: Appropriate for a 50-min exam BONUS COUNT: At least one TOPICS CANDIDATE FOR THE EXAM:

More information

Multiple Random Variables

Multiple Random Variables Multiple Random Variables This Version: July 30, 2015 Multiple Random Variables 2 Now we consider models with more than one r.v. These are called multivariate models For instance: height and weight An

More information

STAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)

STAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3) STAT/MATH 395 A - PROBABILITY II UW Winter Quarter 07 Néhémy Lim Moment functions Moments of a random variable Definition.. Let X be a rrv on probability space (Ω, A, P). For a given r N, E[X r ], if it

More information

1 Review of Probability

1 Review of Probability 1 Review of Probability Random variables are denoted by X, Y, Z, etc. The cumulative distribution function (c.d.f.) of a random variable X is denoted by F (x) = P (X x), < x

More information

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

Review of Probabilities and Basic Statistics

Review of Probabilities and Basic Statistics Alex Smola Barnabas Poczos TA: Ina Fiterau 4 th year PhD student MLD Review of Probabilities and Basic Statistics 10-701 Recitations 1/25/2013 Recitation 1: Statistics Intro 1 Overview Introduction to

More information

Chapter Learning Objectives. Probability Distributions and Probability Density Functions. Continuous Random Variables

Chapter Learning Objectives. Probability Distributions and Probability Density Functions. Continuous Random Variables Chapter 4: Continuous Random Variables and Probability s 4-1 Continuous Random Variables 4-2 Probability s and Probability Density Functions 4-3 Cumulative Functions 4-4 Mean and Variance of a Continuous

More information

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

STAT2201. Analysis of Engineering & Scientific Data. Unit 3 STAT2201 Analysis of Engineering & Scientific Data Unit 3 Slava Vaisman The University of Queensland School of Mathematics and Physics What we learned in Unit 2 (1) We defined a sample space of a random

More information

18.440: Lecture 28 Lectures Review

18.440: Lecture 28 Lectures Review 18.440: Lecture 28 Lectures 18-27 Review Scott Sheffield MIT Outline Outline It s the coins, stupid Much of what we have done in this course can be motivated by the i.i.d. sequence X i where each X i is

More information

Guidelines for Solving Probability Problems

Guidelines for Solving Probability Problems Guidelines for Solving Probability Problems CS 1538: Introduction to Simulation 1 Steps for Problem Solving Suggested steps for approaching a problem: 1. Identify the distribution What distribution does

More information

Class 26: review for final exam 18.05, Spring 2014

Class 26: review for final exam 18.05, Spring 2014 Probability Class 26: review for final eam 8.05, Spring 204 Counting Sets Inclusion-eclusion principle Rule of product (multiplication rule) Permutation and combinations Basics Outcome, sample space, event

More information

1 Random Variable: Topics

1 Random Variable: Topics Note: Handouts DO NOT replace the book. In most cases, they only provide a guideline on topics and an intuitive feel. 1 Random Variable: Topics Chap 2, 2.1-2.4 and Chap 3, 3.1-3.3 What is a random variable?

More information

Chapter 4 Multiple Random Variables

Chapter 4 Multiple Random Variables Review for the previous lecture Theorems and Examples: How to obtain the pmf (pdf) of U = g ( X Y 1 ) and V = g ( X Y) Chapter 4 Multiple Random Variables Chapter 43 Bivariate Transformations Continuous

More information

BMIR Lecture Series on Probability and Statistics Fall, 2015 Uniform Distribution

BMIR Lecture Series on Probability and Statistics Fall, 2015 Uniform Distribution Lecture #5 BMIR Lecture Series on Probability and Statistics Fall, 2015 Department of Biomedical Engineering and Environmental Sciences National Tsing Hua University s 5.1 Definition ( ) A continuous random

More information

18 Bivariate normal distribution I

18 Bivariate normal distribution I 8 Bivariate normal distribution I 8 Example Imagine firing arrows at a target Hopefully they will fall close to the target centre As we fire more arrows we find a high density near the centre and fewer

More information

Chapter 3 Common Families of Distributions

Chapter 3 Common Families of Distributions Lecture 9 on BST 631: Statistical Theory I Kui Zhang, 9/3/8 and 9/5/8 Review for the previous lecture Definition: Several commonly used discrete distributions, including discrete uniform, hypergeometric,

More information

Moments. Raw moment: February 25, 2014 Normalized / Standardized moment:

Moments. Raw moment: February 25, 2014 Normalized / Standardized moment: Moments Lecture 10: Central Limit Theorem and CDFs Sta230 / Mth 230 Colin Rundel Raw moment: Central moment: µ n = EX n ) µ n = E[X µ) 2 ] February 25, 2014 Normalized / Standardized moment: µ n σ n Sta230

More information