Sampling Distributions
|
|
- Coleen Lester
- 5 years ago
- Views:
Transcription
1 Sampling Distributions Mathematics 47: Lecture 9 Dan Sloughter Furman University March 16, 2006 Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
2 Definition We call the probability distribution of a statistic T a sampling distribution. Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
3 Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
4 Let X 1, X 2,..., X n be a random sample from a continuous distribution with probability density function f and (cumulative) distribution function F. Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
5 Let X 1, X 2,..., X n be a random sample from a continuous distribution with probability density function f and (cumulative) distribution function F. Let Y = X (n). Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
6 Let X 1, X 2,..., X n be a random sample from a continuous distribution with probability density function f and (cumulative) distribution function F. Let Y = X (n). If G is the distribution function of Y, then G(y) = P(Y y) = P(X 1 y, X 2 y,..., X n y) n = P(X i y) = (F (y)) n. i=1 Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
7 Let X 1, X 2,..., X n be a random sample from a continuous distribution with probability density function f and (cumulative) distribution function F. Let Y = X (n). If G is the distribution function of Y, then G(y) = P(Y y) = P(X 1 y, X 2 y,..., X n y) n = P(X i y) = (F (y)) n. i=1 Hence, if g is the probability density function of Y, then g(y) = d dy G(y) = n(f (y))n 1 f (y) = nf (y)(f (y)) n 1. Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
8 (cont d) Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
9 (cont d) Now let W = X (1). Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
10 (cont d) Now let W = X (1). If H is the the (cumulative) distribution function of W, then H(w) = P(W w) = 1 P(W > w) = 1 P(X 1 > w, X 2 > w,..., X n > w) n = 1 P(X i > w) = 1 (1 F (w)) n. i=1 Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
11 (cont d) Now let W = X (1). If H is the the (cumulative) distribution function of W, then H(w) = P(W w) = 1 P(W > w) = 1 P(X 1 > w, X 2 > w,..., X n > w) n = 1 P(X i > w) = 1 (1 F (w)) n. i=1 Hence, if h is the probability density function of W, then h(w) = d dy H(w) = n(1 F (w))n 1 ( f (w)) = nf (w)(1 F (w)) n 1. Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
12 Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
13 Suppose X 1, X 2,..., X n is a random sample from a uniform distribution on (0, θ). Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
14 Suppose X 1, X 2,..., X n is a random sample from a uniform distribution on (0, θ). Let Y = X (n) and W = X (1). Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
15 Suppose X 1, X 2,..., X n is a random sample from a uniform distribution on (0, θ). Let Y = X (n) and W = X (1). Recall: the uniform density on (0, θ) is 1, if 0 < x < θ, f (x) = θ 0, otherwise, and the distribution function is 0, if x < 0, x F (x) =, if 0 x < θ, θ 1, if x θ. Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
16 (cont d) Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
17 (cont d) Hence the probability density function for Y is n ( y ) n 1, if 0 < y < θ, g(y) = θ θ 0, otherwise, { n = θ n y n 1, if 0 < y < θ, 0, otherwise. Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
18 (cont d) Hence the probability density function for Y is n ( y ) n 1, if 0 < y < θ, g(y) = θ θ 0, otherwise, { n = θ n y n 1, if 0 < y < θ, 0, otherwise. And the probability density function for W is n ( 1 w ) n 1, if 0 < w < θ, h(w) = θ θ 0, otherwise. Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
19 Sample means Theorem If X 1, X 2,..., X n is a random sample from a distribution with mean µ and standard deviation σ, then E[ X ] = µ and var[ X ] = σ2 n. Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
20 Sample means Theorem If X 1, X 2,..., X n is a random sample from a distribution with mean µ and standard deviation σ, then E[ X ] = µ and var[ X ] = σ2 n. Proof. We have E[ X ] = E [ 1 n ] n X i = 1 n i=1 n E[X i ] = 1 n (nµ) = µ i=1 and var[ X ] = var [ 1 n ] n X i = 1 n n 2 var[x i ] = 1 n 2 (nσ2 ) = σ2 n. i=1 Dan Sloughter (Furman University) Sampling Distributions March 16, / 10 i=1
21 Sample means (cont d) If we let µ Y denote the mean of a random variable Y and σ 2 Y denote the variance of Y, then the above theorem says µ X = µ and or, equivalently, σ2 σ = 2 X n, σ X = σ n. Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
22 Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
23 Suppose X 1, X 2,..., X n is a random sample from a Bernoulli distribution with probability of success p. Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
24 Suppose X 1, X 2,..., X n is a random sample from a Bernoulli distribution with probability of success p. Let ˆp = X, the sample proportion of successes. Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
25 Suppose X 1, X 2,..., X n is a random sample from a Bernoulli distribution with probability of success p. Let ˆp = X, the sample proportion of successes. Then and µˆp = E[X 1 ] = p σˆp = σ n X 1 = p(1 p). n Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
26 Suppose X 1, X 2,..., X n is a random sample from a Bernoulli distribution with probability of success p. Let ˆp = X, the sample proportion of successes. Then and µˆp = E[X 1 ] = p σˆp = σ n X 1 = p(1 p). n Note: for any 0 p 1, 0 p(1 p) 1 4. Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
27 Suppose X 1, X 2,..., X n is a random sample from a Bernoulli distribution with probability of success p. Let ˆp = X, the sample proportion of successes. Then and µˆp = E[X 1 ] = p σˆp = σ n X 1 = p(1 p). n Note: for any 0 p 1, 0 p(1 p) 1 4. Hence, for any value of p, σˆp 1 2 n. Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
28 (cont d) Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
29 (cont d) For example, for a sample of size n = 2000, σˆp Dan Sloughter (Furman University) Sampling Distributions March 16, / 10
Pivotal Quantities. Mathematics 47: Lecture 16. Dan Sloughter. Furman University. March 30, 2006
Pivotal Quantities Mathematics 47: Lecture 16 Dan Sloughter Furman University March 30, 2006 Dan Sloughter (Furman University) Pivotal Quantities March 30, 2006 1 / 10 Pivotal quantities Definition Suppose
More informationNonparametric Tests. Mathematics 47: Lecture 25. Dan Sloughter. Furman University. April 20, 2006
Nonparametric Tests Mathematics 47: Lecture 25 Dan Sloughter Furman University April 20, 2006 Dan Sloughter (Furman University) Nonparametric Tests April 20, 2006 1 / 14 The sign test Suppose X 1, X 2,...,
More informationMathematics 13: Lecture 4
Mathematics 13: Lecture Planes Dan Sloughter Furman University January 10, 2008 Dan Sloughter (Furman University) Mathematics 13: Lecture January 10, 2008 1 / 10 Planes in R n Suppose v and w are nonzero
More informationMAS113 Introduction to Probability and Statistics
MAS113 Introduction to Probability and Statistics School of Mathematics and Statistics, University of Sheffield 2018 19 Identically distributed Suppose we have n random variables X 1, X 2,..., X n. Identically
More informationMathematics 22: Lecture 7
Mathematics 22: Lecture 7 Separation of Variables Dan Sloughter Furman University January 15, 2008 Dan Sloughter (Furman University) Mathematics 22: Lecture 7 January 15, 2008 1 / 8 Separable equations
More informationAntiderivatives. Mathematics 11: Lecture 30. Dan Sloughter. Furman University. November 7, 2007
Antiderivatives Mathematics 11: Lecture 30 Dan Sloughter Furman University November 7, 2007 Dan Sloughter (Furman University) Antiderivatives November 7, 2007 1 / 9 Definition Recall: Suppose F and f are
More informationMathematics 22: Lecture 5
Mathematics 22: Lecture 5 Autonomous Equations Dan Sloughter Furman University January 11, 2008 Dan Sloughter (Furman University) Mathematics 22: Lecture 5 January 11, 2008 1 / 11 Solving the logistics
More informationPractice Problem - Skewness of Bernoulli Random Variable. Lecture 7: Joint Distributions and the Law of Large Numbers. Joint Distributions - Example
A little more E(X Practice Problem - Skewness of Bernoulli Random Variable Lecture 7: and the Law of Large Numbers Sta30/Mth30 Colin Rundel February 7, 014 Let X Bern(p We have shown that E(X = p Var(X
More informationSome Trigonometric Limits
Some Trigonometric Limits Mathematics 11: Lecture 7 Dan Sloughter Furman University September 20, 2007 Dan Sloughter (Furman University) Some Trigonometric Limits September 20, 2007 1 / 14 The squeeze
More informationCalculus: Area. Mathematics 15: Lecture 22. Dan Sloughter. Furman University. November 12, 2006
Calculus: Area Mathematics 15: Lecture 22 Dan Sloughter Furman University November 12, 2006 Dan Sloughter (Furman University) Calculus: Area November 12, 2006 1 / 7 Area Note: formulas for the areas of
More informationGoodness of Fit Tests: Homogeneity
Goodness of Fit Tests: Homogeneity Mathematics 47: Lecture 35 Dan Sloughter Furman University May 11, 2006 Dan Sloughter (Furman University) Goodness of Fit Tests: Homogeneity May 11, 2006 1 / 13 Testing
More informationLimiting Distributions
Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the
More informationMathematics 22: Lecture 19
Mathematics 22: Lecture 19 Legendre s Equation Dan Sloughter Furman University February 5, 2008 Dan Sloughter (Furman University) Mathematics 22: Lecture 19 February 5, 2008 1 / 11 Example: Legendre s
More informationChange of Variables: Indefinite Integrals
Change of Variables: Indefinite Integrals Mathematics 11: Lecture 39 Dan Sloughter Furman University November 29, 2007 Dan Sloughter (Furman University) Change of Variables: Indefinite Integrals November
More informationMathematics 22: Lecture 10
Mathematics 22: Lecture 10 Euler s Method Dan Sloughter Furman University January 22, 2008 Dan Sloughter (Furman University) Mathematics 22: Lecture 10 January 22, 2008 1 / 14 Euler s method Consider the
More informationLecture 13. Poisson Distribution. Text: A Course in Probability by Weiss 5.5. STAT 225 Introduction to Probability Models February 16, 2014
Lecture 13 Text: A Course in Probability by Weiss 5.5 STAT 225 Introduction to Probability Models February 16, 2014 Whitney Huang Purdue University 13.1 Agenda 1 2 3 13.2 Review So far, we have seen discrete
More informationMathematics 22: Lecture 12
Mathematics 22: Lecture 12 Second-order Linear Equations Dan Sloughter Furman University January 28, 2008 Dan Sloughter (Furman University) Mathematics 22: Lecture 12 January 28, 2008 1 / 14 Definition
More informationHomework for 1/13 Due 1/22
Name: ID: Homework for 1/13 Due 1/22 1. [ 5-23] An irregularly shaped object of unknown area A is located in the unit square 0 x 1, 0 y 1. Consider a random point distributed uniformly over the square;
More informationMathematics 22: Lecture 4
Mathematics 22: Lecture 4 Population Models Dan Sloughter Furman University January 10, 2008 Dan Sloughter (Furman University) Mathematics 22: Lecture 4 January 10, 2008 1 / 6 Malthusian growth model Let
More informationProving the central limit theorem
SOR3012: Stochastic Processes Proving the central limit theorem Gareth Tribello March 3, 2019 1 Purpose In the lectures and exercises we have learnt about the law of large numbers and the central limit
More informationMathematics 22: Lecture 11
Mathematics 22: Lecture 11 Runge-Kutta Dan Sloughter Furman University January 25, 2008 Dan Sloughter (Furman University) Mathematics 22: Lecture 11 January 25, 2008 1 / 11 Order of approximations One
More informationExample. Mathematics 255: Lecture 17. Example. Example (cont d) Consider the equation. d 2 y dt 2 + dy
Mathematics 255: Lecture 17 Undetermined Coefficients Dan Sloughter Furman University October 10, 2008 6y = 5e 4t. so the general solution of 0 = r 2 + r 6 = (r + 3)(r 2), 6y = 0 y(t) = c 1 e 3t + c 2
More informationThe Chain Rule. Mathematics 11: Lecture 18. Dan Sloughter. Furman University. October 10, 2007
The Chain Rule Mathematics 11: Lecture 18 Dan Sloughter Furman University October 10, 2007 Dan Sloughter (Furman University) The Chain Rule October 10, 2007 1 / 15 Example Suppose that a pebble is dropped
More informationMathematics 13: Lecture 10
Mathematics 13: Lecture 10 Matrices Dan Sloughter Furman University January 25, 2008 Dan Sloughter (Furman University) Mathematics 13: Lecture 10 January 25, 2008 1 / 19 Matrices Recall: A matrix is a
More informationChapter 4. Chapter 4 sections
Chapter 4 sections 4.1 Expectation 4.2 Properties of Expectations 4.3 Variance 4.4 Moments 4.5 The Mean and the Median 4.6 Covariance and Correlation 4.7 Conditional Expectation SKIP: 4.8 Utility Expectation
More informationThe Geometry. Mathematics 15: Lecture 20. Dan Sloughter. Furman University. November 6, 2006
The Geometry Mathematics 15: Lecture 20 Dan Sloughter Furman University November 6, 2006 Dan Sloughter (Furman University) The Geometry November 6, 2006 1 / 18 René Descartes 1596-1650 Dan Sloughter (Furman
More information3 Conditional Expectation
3 Conditional Expectation 3.1 The Discrete case Recall that for any two events E and F, the conditional probability of E given F is defined, whenever P (F ) > 0, by P (E F ) P (E)P (F ). P (F ) Example.
More informationMidterm #1. Lecture 10: Joint Distributions and the Law of Large Numbers. Joint Distributions - Example, cont. Joint Distributions - Example
Midterm #1 Midterm 1 Lecture 10: and the Law of Large Numbers Statistics 104 Colin Rundel February 0, 01 Exam will be passed back at the end of class Exam was hard, on the whole the class did well: Mean:
More informationProbability Models. 4. What is the definition of the expectation of a discrete random variable?
1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions
More informationLecture 7: Chapter 7. Sums of Random Variables and Long-Term Averages
Lecture 7: Chapter 7. Sums of Random Variables and Long-Term Averages ELEC206 Probability and Random Processes, Fall 2014 Gil-Jin Jang gjang@knu.ac.kr School of EE, KNU page 1 / 15 Chapter 7. Sums of Random
More information6.041/6.431 Fall 2010 Quiz 2 Solutions
6.04/6.43: Probabilistic Systems Analysis (Fall 200) 6.04/6.43 Fall 200 Quiz 2 Solutions Problem. (80 points) In this problem: (i) X is a (continuous) uniform random variable on [0, 4]. (ii) Y is an exponential
More informationSTAT 430/510: Lecture 16
STAT 430/510: Lecture 16 James Piette June 24, 2010 Updates HW4 is up on my website. It is due next Mon. (June 28th). Starting today back at section 6.7 and will begin Ch. 7. Joint Distribution of Functions
More informationE X A M. Probability Theory and Stochastic Processes Date: December 13, 2016 Duration: 4 hours. Number of pages incl.
E X A M Course code: Course name: Number of pages incl. front page: 6 MA430-G Probability Theory and Stochastic Processes Date: December 13, 2016 Duration: 4 hours Resources allowed: Notes: Pocket calculator,
More informationLecture 14. Text: A Course in Probability by Weiss 5.6. STAT 225 Introduction to Probability Models February 23, Whitney Huang Purdue University
Lecture 14 Text: A Course in Probability by Weiss 5.6 STAT 225 Introduction to Probability Models February 23, 2014 Whitney Huang Purdue University 14.1 Agenda 14.2 Review So far, we have covered Bernoulli
More informationLecture 1: August 28
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random
More informationRandom Variables. Lecture 6: E(X ), Var(X ), & Cov(X, Y ) Random Variables - Vocabulary. Random Variables, cont.
Lecture 6: E(X ), Var(X ), & Cov(X, Y ) Sta230/Mth230 Colin Rundel February 5, 2014 We have been using them for a while now in a variety of forms but it is good to explicitly define what we mean Random
More informationBNAD 276 Lecture 5 Discrete Probability Distributions Exercises 1 11
1 / 15 BNAD 276 Lecture 5 Discrete Probability Distributions 1 11 Phuong Ho May 14, 2017 Exercise 1 Suppose we have the probability distribution for the random variable X as follows. X f (x) 20.20 25.15
More informationLecture 4: Law of Large Number and Central Limit Theorem
ECE 645: Estimation Theory Sring 2015 Instructor: Prof. Stanley H. Chan Lecture 4: Law of Large Number and Central Limit Theorem (LaTeX reared by Jing Li) March 31, 2015 This lecture note is based on ECE
More informationLimiting Distributions
We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the two fundamental results
More informationProbability and Statistics Notes
Probability and Statistics Notes Chapter Five Jesse Crawford Department of Mathematics Tarleton State University Spring 2011 (Tarleton State University) Chapter Five Notes Spring 2011 1 / 37 Outline 1
More informationChapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued
Chapter 3 sections Chapter 3 - continued 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions
More informationSDS 321: Introduction to Probability and Statistics
SDS 321: Introduction to Probability and Statistics Lecture 14: Continuous random variables Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin www.cs.cmu.edu/
More informationProbability and Distributions
Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated
More informationProperties of Random Variables
Properties of Random Variables 1 Definitions A discrete random variable is defined by a probability distribution that lists each possible outcome and the probability of obtaining that outcome If the random
More informationLecture 19: Properties of Expectation
Lecture 19: Properties of Expectation Dan Sloughter Furman University Mathematics 37 February 11, 4 19.1 The unconscious statistician, revisited The following is a generalization of the law of the unconscious
More informationDistributions of Functions of Random Variables. 5.1 Functions of One Random Variable
Distributions of Functions of Random Variables 5.1 Functions of One Random Variable 5.2 Transformations of Two Random Variables 5.3 Several Random Variables 5.4 The Moment-Generating Function Technique
More information1 Solution to Problem 2.1
Solution to Problem 2. I incorrectly worked this exercise instead of 2.2, so I decided to include the solution anyway. a) We have X Y /3, which is a - function. It maps the interval, ) where X lives) onto
More informationInverse Normal Distribution and Sampling Distributions
Inverse Normal Distribution and Sampling Distributions Section 4.3 & 4.4 Cathy Poliak, Ph.D. cathy@math.uh.edu Department of Mathematics University of Houston Lecture 11-2311 Cathy Poliak, Ph.D. cathy@math.uh.edu
More informationSome Assorted Formulae. Some confidence intervals: σ n. x ± z α/2. x ± t n 1;α/2 n. ˆp(1 ˆp) ˆp ± z α/2 n. χ 2 n 1;1 α/2. n 1;α/2
STA 248 H1S MIDTERM TEST February 26, 2008 SURNAME: SOLUTIONS GIVEN NAME: STUDENT NUMBER: INSTRUCTIONS: Time: 1 hour and 50 minutes Aids allowed: calculator Tables of the standard normal, t and chi-square
More informationForecasting. This optimal forecast is referred to as the Minimum Mean Square Error Forecast. This optimal forecast is unbiased because
Forecasting 1. Optimal Forecast Criterion - Minimum Mean Square Error Forecast We have now considered how to determine which ARIMA model we should fit to our data, we have also examined how to estimate
More informationSTAT 430/510: Lecture 10
STAT 430/510: Lecture 10 James Piette June 9, 2010 Updates HW2 is due today! Pick up your HW1 s up in stat dept. There is a box located right when you enter that is labeled "Stat 430 HW1". It ll be out
More information6 The normal distribution, the central limit theorem and random samples
6 The normal distribution, the central limit theorem and random samples 6.1 The normal distribution We mentioned the normal (or Gaussian) distribution in Chapter 4. It has density f X (x) = 1 σ 1 2π e
More informationProblem Y is an exponential random variable with parameter λ = 0.2. Given the event A = {Y < 2},
ECE32 Spring 25 HW Solutions April 6, 25 Solutions to HW Note: Most of these solutions were generated by R. D. Yates and D. J. Goodman, the authors of our textbook. I have added comments in italics where
More informationLecture Notes 2 Random Variables. Discrete Random Variables: Probability mass function (pmf)
Lecture Notes 2 Random Variables Definition Discrete Random Variables: Probability mass function (pmf) Continuous Random Variables: Probability density function (pdf) Mean and Variance Cumulative Distribution
More informationBMIR Lecture Series on Probability and Statistics Fall, 2015 Uniform Distribution
Lecture #5 BMIR Lecture Series on Probability and Statistics Fall, 2015 Department of Biomedical Engineering and Environmental Sciences National Tsing Hua University s 5.1 Definition ( ) A continuous random
More informationLecture 8 Sampling Theory
Lecture 8 Sampling Theory Thais Paiva STA 111 - Summer 2013 Term II July 11, 2013 1 / 25 Thais Paiva STA 111 - Summer 2013 Term II Lecture 8, 07/11/2013 Lecture Plan 1 Sampling Distributions 2 Law of Large
More informationChapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued
Chapter 3 sections 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions 3.6 Conditional
More informationMath 341: Probability Eighteenth Lecture (11/12/09)
Math 341: Probability Eighteenth Lecture (11/12/09) Steven J Miller Williams College Steven.J.Miller@williams.edu http://www.williams.edu/go/math/sjmiller/ public html/341/ Bronfman Science Center Williams
More informationTopic 7: Convergence of Random Variables
Topic 7: Convergence of Ranom Variables Course 003, 2016 Page 0 The Inference Problem So far, our starting point has been a given probability space (S, F, P). We now look at how to generate information
More informationConditional distributions (discrete case)
Conditional distributions (discrete case) The basic idea behind conditional distributions is simple: Suppose (XY) is a jointly-distributed random vector with a discrete joint distribution. Then we can
More informationMTMS Mathematical Statistics
MTMS.01.099 Mathematical Statistics Lecture 12. Hypothesis testing. Power function. Approximation of Normal distribution and application to Binomial distribution Tõnu Kollo Fall 2016 Hypothesis Testing
More informationSTA 111: Probability & Statistical Inference
STA 111: Probability & Statistical Inference Lecture Four Expectation and Continuous Random Variables Instructor: Olanrewaju Michael Akande Department of Statistical Science, Duke University Instructor:
More informationThe Geometric Random Walk: More Applications to Gambling
MATH 540 The Geometric Random Walk: More Applications to Gambling Dr. Neal, Spring 2008 We now shall study properties of a random walk process with only upward or downward steps that is stopped after the
More information4. Distributions of Functions of Random Variables
4. Distributions of Functions of Random Variables Setup: Consider as given the joint distribution of X 1,..., X n (i.e. consider as given f X1,...,X n and F X1,...,X n ) Consider k functions g 1 : R n
More informationLecture 5: Moment Generating Functions
Lecture 5: Moment Generating Functions IB Paper 7: Probability and Statistics Carl Edward Rasmussen Department of Engineering, University of Cambridge February 28th, 2018 Rasmussen (CUED) Lecture 5: Moment
More informationSTAT 430/510 Probability Lecture 16, 17: Compute by Conditioning
STAT 430/510 Probability Lecture 16, 17: Compute by Conditioning Pengyuan (Penelope) Wang Lecture 16-17 June 21, 2011 Computing Probability by Conditioning A is an arbitrary event If Y is a discrete random
More informationContinuous Random Variables
1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables
More informationExample continued. Math 425 Intro to Probability Lecture 37. Example continued. Example
continued : Coin tossing Math 425 Intro to Probability Lecture 37 Kenneth Harris kaharri@umich.edu Department of Mathematics University of Michigan April 8, 2009 Consider a Bernoulli trials process with
More informationEcon 325: Introduction to Empirical Economics
Econ 325: Introduction to Empirical Economics Lecture 6 Sampling and Sampling Distributions Ch. 6-1 Populations and Samples A Population is the set of all items or individuals of interest Examples: All
More informationMath438 Actuarial Probability
Math438 Actuarial Probability Jinguo Lian Department of Math and Stats Jan. 22, 2016 Continuous Random Variables-Part I: Definition A random variable X is continuous if its set of possible values is an
More informationLecture 4: Proofs for Expectation, Variance, and Covariance Formula
Lecture 4: Proofs for Expectation, Variance, and Covariance Formula by Hiro Kasahara Vancouver School of Economics University of British Columbia Hiro Kasahara (UBC) Econ 325 1 / 28 Discrete Random Variables:
More informationMoments. Raw moment: February 25, 2014 Normalized / Standardized moment:
Moments Lecture 10: Central Limit Theorem and CDFs Sta230 / Mth 230 Colin Rundel Raw moment: Central moment: µ n = EX n ) µ n = E[X µ) 2 ] February 25, 2014 Normalized / Standardized moment: µ n σ n Sta230
More informationIE 303 Discrete-Event Simulation
IE 303 Discrete-Event Simulation 1 L E C T U R E 5 : P R O B A B I L I T Y R E V I E W Review of the Last Lecture Random Variables Probability Density (Mass) Functions Cumulative Density Function Discrete
More informationCS145: Probability & Computing
CS45: Probability & Computing Lecture 5: Concentration Inequalities, Law of Large Numbers, Central Limit Theorem Instructor: Eli Upfal Brown University Computer Science Figure credits: Bertsekas & Tsitsiklis,
More informationLimit Theorems. STATISTICS Lecture no Department of Econometrics FEM UO Brno office 69a, tel
STATISTICS Lecture no. 6 Department of Econometrics FEM UO Brno office 69a, tel. 973 442029 email:jiri.neubauer@unob.cz 3. 11. 2009 If we repeat some experiment independently we can create using given
More informationMath 3215 Intro. Probability & Statistics Summer 14. Homework 5: Due 7/3/14
Math 325 Intro. Probability & Statistics Summer Homework 5: Due 7/3/. Let X and Y be continuous random variables with joint/marginal p.d.f. s f(x, y) 2, x y, f (x) 2( x), x, f 2 (y) 2y, y. Find the conditional
More informationSystem Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models
System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models Fatih Cavdur fatihcavdur@uludag.edu.tr March 20, 2012 Introduction Introduction The world of the model-builder
More information3 Multiple Discrete Random Variables
3 Multiple Discrete Random Variables 3.1 Joint densities Suppose we have a probability space (Ω, F,P) and now we have two discrete random variables X and Y on it. They have probability mass functions f
More informationSystem Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models
System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models Fatih Cavdur fatihcavdur@uludag.edu.tr March 29, 2014 Introduction Introduction The world of the model-builder
More informationExam 2. Problem 1: Independent normal random variables
Exam 2 Problem 1: Independent normal random variables Let U, V, and W be independent standard normal random variables (that is, independent normal random variables, each with mean 0 and variance 1), and
More informationAPPM/MATH 4/5520 Solutions to Exam I Review Problems. f X 1,X 2. 2e x 1 x 2. = x 2
APPM/MATH 4/5520 Solutions to Exam I Review Problems. (a) f X (x ) f X,X 2 (x,x 2 )dx 2 x 2e x x 2 dx 2 2e 2x x was below x 2, but when marginalizing out x 2, we ran it over all values from 0 to and so
More informationSTA 260: Statistics and Probability II
Al Nosedal. University of Toronto. Winter 2017 1 Chapter 7. Sampling Distributions and the Central Limit Theorem If you can t explain it simply, you don t understand it well enough Albert Einstein. Theorem
More informationRecitation 2: Probability
Recitation 2: Probability Colin White, Kenny Marino January 23, 2018 Outline Facts about sets Definitions and facts about probability Random Variables and Joint Distributions Characteristics of distributions
More informationSTA 256: Statistics and Probability I
Al Nosedal. University of Toronto. Fall 2017 My momma always said: Life was like a box of chocolates. You never know what you re gonna get. Forrest Gump. There are situations where one might be interested
More informationBivariate distributions
Bivariate distributions 3 th October 017 lecture based on Hogg Tanis Zimmerman: Probability and Statistical Inference (9th ed.) Bivariate Distributions of the Discrete Type The Correlation Coefficient
More informationDennis Bricker Dept of Mechanical & Industrial Engineering The University of Iowa
Dennis Bricker Dept of Mechanical & Industrial Engineering The University of Iowa dennis-bricker@uiowa.edu Probability Theory Results page 1 D.Bricker, U. of Iowa, 2002 Probability of simultaneous occurrence
More informationHT Introduction. P(X i = x i ) = e λ λ x i
MODS STATISTICS Introduction. HT 2012 Simon Myers, Department of Statistics (and The Wellcome Trust Centre for Human Genetics) myers@stats.ox.ac.uk We will be concerned with the mathematical framework
More informationM(t) = 1 t. (1 t), 6 M (0) = 20 P (95. X i 110) i=1
Math 66/566 - Midterm Solutions NOTE: These solutions are for both the 66 and 566 exam. The problems are the same until questions and 5. 1. The moment generating function of a random variable X is M(t)
More informationMathematics 426 Robert Gross Homework 9 Answers
Mathematics 4 Robert Gross Homework 9 Answers. Suppose that X is a normal random variable with mean µ and standard deviation σ. Suppose that PX > 9 PX
More informationChapter 6: Large Random Samples Sections
Chapter 6: Large Random Samples Sections 6.1: Introduction 6.2: The Law of Large Numbers Skip p. 356-358 Skip p. 366-368 Skip 6.4: The correction for continuity Remember: The Midterm is October 25th in
More informationTheory of Statistics.
Theory of Statistics. Homework V February 5, 00. MT 8.7.c When σ is known, ˆµ = X is an unbiased estimator for µ. If you can show that its variance attains the Cramer-Rao lower bound, then no other unbiased
More informationMore on Distribution Function
More on Distribution Function The distribution of a random variable X can be determined directly from its cumulative distribution function F X. Theorem: Let X be any random variable, with cumulative distribution
More informationCOS Lecture 16 Autonomous Robot Navigation
COS 495 - Lecture 16 Autonomous Robot Navigation Instructor: Chris Clark Semester: Fall 011 1 Figures courtesy of Siegwart & Nourbakhsh Control Structure Prior Knowledge Operator Commands Localization
More informationLecture 26 Section 8.4. Wed, Oct 14, 2009
PDFs n = Lecture 26 Section 8.4 Hampden-Sydney College Wed, Oct 14, 2009 Outline PDFs n = 1 2 PDFs n = 3 4 5 6 Outline PDFs n = 1 2 PDFs n = 3 4 5 6 PDFs n = Exercise 8.12, page 528. Suppose that 60% of
More informationVariance reduction. Michel Bierlaire. Transport and Mobility Laboratory. Variance reduction p. 1/18
Variance reduction p. 1/18 Variance reduction Michel Bierlaire michel.bierlaire@epfl.ch Transport and Mobility Laboratory Variance reduction p. 2/18 Example Use simulation to compute I = 1 0 e x dx We
More informationECE302 Spring 2015 HW10 Solutions May 3,
ECE32 Spring 25 HW Solutions May 3, 25 Solutions to HW Note: Most of these solutions were generated by R. D. Yates and D. J. Goodman, the authors of our textbook. I have added comments in italics where
More informationThis paper is not to be removed from the Examination Halls
~~ST104B ZA d0 This paper is not to be removed from the Examination Halls UNIVERSITY OF LONDON ST104B ZB BSc degrees and Diplomas for Graduates in Economics, Management, Finance and the Social Sciences,
More informationProbability Distributions
Probability Distributions Series of events Previously we have been discussing the probabilities associated with a single event: Observing a 1 on a single roll of a die Observing a K with a single card
More informationMathematical statistics
October 18 th, 2018 Lecture 16: Midterm review Countdown to mid-term exam: 7 days Week 1 Chapter 1: Probability review Week 2 Week 4 Week 7 Chapter 6: Statistics Chapter 7: Point Estimation Chapter 8:
More information