STAT 801: Mathematical Statistics. Distribution Theory

Size: px
Start display at page:

Download "STAT 801: Mathematical Statistics. Distribution Theory"

Transcription

1 STAT 81: Mathematical Statistics Distribution Theory Basic Problem: Start with assumptions about f or CDF of random vector X (X 1,..., X p ). Define Y g(x 1,..., X p ) to be some function of X (usually some statistic of interest). How can we compute the distribution or CDF or density of Y? Univariate Techniques Method 1: compute the CDF by integration and differentiate to find f Y. Example: U Uniform[, 1] and Y logu. so Y has standard exponential distribution. Example: Z N(, 1), i.e. and Y Z 2. Then F Y (y) P(Y y) P( log U y) P(log U y) P(U e y ) { 1 e y y > y. f Z (z) 1 2π e z2 /2 F Y (y) P(Z 2 y) { y < P( y Z y) y. Now differentiate to get Then P( y Z y) F Z ( y) F Z ( y) f Y (y) d dy F Z( y) [ y < d dy FZ ( y) F Z ( y) ] y > undefined y. f Z ( y) d y dy 1 ( exp ( ) y) 2 1 /2 2π 2 y 1/ πy e y/2. (Similar formula for other derivative.) Thus 1 2πy e y/2 y > f Y (y) y < undefined y. We will find indicator notation useful: 1(y > ) { 1 y > y which we use to write f Y (y) 1 2πy e y/2 1(y > ) 1

2 (changing definition unimportantly at y ). Notice: I never evaluated F Y before differentiating it. In fact F Y and F Z are integrals I can t do but I can differentiate then anyway. Remember fundamental theorem of calculus: d dx x at any x where f is continuous. Summary: for Y g(x) with X and Y each real valued a f(y)dy f(x) P(Y y) P(g(X) y) P(X g 1 (, y]). Take d/dy to compute the density f Y (y) d f X (x)dx. dy {x:g(x) y} Often can differentiate without doing integral. Method 2: Change of variables. Assume g is one to one. I do: g is increasing and differentiable. Interpretation of density (based on density F ): and f Y (y) P(y Y y + δy) lim δy δy lim δy F Y (y + δy) F Y (y) δy P(x X x + δx) f X (x) lim. δx δx Now assume y g(x). Define δy by y + δy g(x + δx). Then Get Take limit to get or P(y Y g(x + δx)) P(x X x + δx). P(y Y y + δy)) δy P(x X x + δx)/δx {g(x + δx) y}/δx. f Y (y) f X (x)/g (x) f Y (g(x))g (x) f X (x). Alternative view: Each probability is integral of a density. The first is the integral of the density of Y over the small interval from y g(x) to y g(x + δx). The interval is narrow so f Y is nearly constant and Since g has a derivative the difference P(y Y g(x + δx)) f Y (y)(g(x + δx) g(x)). g(x + δx) g(x) δxg (x) and we get P(y Y g(x + δx)) f Y (y)g (x)δx. 2

3 Same idea applied to P(x X x + δx) gives P(x X x + δx) f X (x)δx so that or, cancelling the δx in the limit If you remember y g(x) then you get f Y (y)g (x)δx f X (x)δx f Y (y)g (x) f X (x). f X (x) f Y (g(x))g (x). Or solve y g(x) to get x in terms of y, that is, x g 1 (y) and then f Y (y) f X (g 1 (y))/g (g 1 (y)). This is just the change of variables formula for doing integrals. Remark: For g decreasing g < but Then the interval (g(x), g(x + δx)) is really (g(x + δx), g(x)) so that g(x) g(x + δx) g (x)δx. In both cases this amounts to the formula Mnemonic: Example: X Weibull(shape α, scale β) or f X (x) α β f X (x) f Y (g(x)) g (x). f Y (y)dy f X (x)dx. ( ) α 1 x exp { (x/β) α }1(x > ). β Let Y log X or g(x) log(x). Solve y log x: x exp(y) or g 1 (y) e y. Then g (x) 1/x and 1/g (g 1 (y)) 1/(1/e y ) e y. Hence ( ) e y α 1 exp { (e y /β) α }1(e y > )e y. f Y (y) α β For any y, e y > so indicator 1. So β f Y (y) α β α exp {αy eαy /β α }. Define φ log β and θ 1/α; then, f Y (y) 1 θ exp { y φ θ { }} y φ exp. θ Extreme Value density with location parameter φ and scale parameter θ. (Note: several distributions are called Extreme Value.) Marginalization Simplest multivariate problem: (or in general Y is any X j ). X (X 1,..., X p ), Y X 1 3

4 Theorem 1 If X has density f(x 1,..., x p ) and q < p then Y (X 1,..., X q ) has density f Y (x 1,...,x q ) f(x 1,..., x p )dx q+1...dx p. f X1,...,X q is the marginal density of X 1,...,X q and f X the joint density of X but they are both just densities. Marginal just to distinguish from the joint density of X. Example The function f(x 1, x 2 ) Kx 1 x 2 1(x 1 >, x 2 >, x 1 + x 2 < 1) is a density provided The integral is P(X R 2 ) so K 24. The marginal density of x 1 is f X1 (x 1 ) K x1 24x 1 x 2 f(x 1, x 2 )dx 1 dx 2 1. x 1 x 2 dx 1 dx 2 K 1 x 1 (1 x 1 ) 2 dx 1 /2 K(1/2 2/3 + 1/4)/2 K/24 1(x 1 >, x 2 >, x 1 + x 2 < 1)dx 2 1 x1 x 1 x 2 1( < x 1 < 1)dx 2 12x 1 (1 x 1 ) 2 1( < x 1 < 1). This is a Beta(2, 3) density. General problem has Y (Y 1,..., Y q ) with Y i g i (X 1,...,X p ). Case 1: q > p. Y won t have density for smooth g. Y will have a singular or discrete distribution. Problem rarely of real interest. (But, e.g., residuals have singular distribution.) Case 2: q p. We use a change of variables formula which generalizes the one derived above for the case p q 1. (See below.) Case 3: q < p. Pad out Y add on p q more variables (carefully chosen) say Y q+1,..., Y p. Find functions g q+1,..., g p. Define for q < i p, Y i g i (X 1,..., X p ) and Z (Y 1,...,Y p ). Choose g i so that we can use change of variables on g (g 1,..., g p ) to compute f Z. Find f Y by integration: f Y (y 1,..., y q ) Change of Variables f Z (y 1,...,y q, z q+1,..., z p )dz q+1... dz p Suppose Y g(x) R p with X R p having density f X. Assume g is a one to one ( injective ) map, i.e., g(x 1 ) g(x 2 ) if and only if x 1 x 2. Find f Y : Step 1: Solve for x in terms of y: x g 1 (y). Step 2: Use basic equation: f Y (y)dy f X (x)dx 4

5 and rewrite it in the form Interpretation of derivative dx dy when p > 1: which is the so called Jacobian. Equivalent formula inverts the matrix: This notation means f Y (y) f X (g 1 (y)) dx dy. dx dy dy dx det ( ) det xi y j f Y (y) f X(g 1 (y)) dy dx y 1 x 1 y 1 x 2 y p x 1. y p x 2 but with x replaced by the corresponding value of y, that is, replace x by g 1 (y). Example: The density f X (x 1, x 2 ) 1 { 2π exp x2 1 + } x2 2 2 is the standard bivariate normal density. Let Y (Y 1, Y 2 ) where Y 1 X X2 2 and Y 2 < 2π is angle from the positive x axis to the ray from the origin to the point (X 1, X 2 ). I.e., Y is X in polar co-ordinates. Solve for x in terms of y: so that It follows that X 1 Y 1 cos(y 2 ) X 2 Y 1 sin(y 2 ) y 1 x p y p x p g(x 1, x 2 ) (g 1 (x 1, x 2 ), g 2 (x 1, x 2 )) ( x x2 2, argument(x 1, x 2 )) g 1 (y 1, y 2 ) (g1 1 (y 1, y 2 ), g2 1 (y 1, y 2 )) (y 1 cos(y 2 ), y 1 sin(y 2 )) ( dx dy det cos(y2 ) y 1 sin(y 2 ) sin(y 2 ) y 1 cos(y 2 ) y 1. f Y (y 1, y 2 ) 1 { } 2π exp y2 1 y 1 1( y 1 < )1( y 2 < 2π). 2 Next: marginal densities of Y 1, Y 2? Factor f Y as f Y (y 1, y 2 ) h 1 (y 1 )h 2 (y 2 ) where h 1 (y 1 ) y 1 e y2 1 /2 1( y 1 < ) ) and h 2 (y 2 ) 1( y 2 < 2π)/(2π). 5

6 Then f Y1 (y 1 ) h 1 (y 1 ) h 1 (y 1 )h 2 (y 2 )dy 2 h 2 (y 2 )dy 2 so marginal density of Y 1 is a multiple of h 1. Multiplier makes f Y1 1 but in this case so that h 2 (y 2 )dy 2 2π (2π) 1 dy 2 1 f Y1 (y 1 ) y 1 e y2 1 /2 1( y 1 < ). (Special case of Weibull or Rayleigh distribution.) Similarly f Y2 (y 2 ) 1( y 2 < 2π)/(2π) which is the Uniform((, 2π) density. Exercise: W Y1 2 /2 has standard exponential distribution. Recall: by definition U Y1 2 has a χ 2 distribution on 2 degrees of freedom. Exercise: find χ 2 2 density. Note: We show below factorization of density is equivalent to independence. 6

STAT 450: Statistical Theory. Distribution Theory. Reading in Casella and Berger: Ch 2 Sec 1, Ch 4 Sec 1, Ch 4 Sec 6.

STAT 450: Statistical Theory. Distribution Theory. Reading in Casella and Berger: Ch 2 Sec 1, Ch 4 Sec 1, Ch 4 Sec 6. STAT 45: Statistical Theory Distribution Theory Reading in Casella and Berger: Ch 2 Sec 1, Ch 4 Sec 1, Ch 4 Sec 6. Basic Problem: Start with assumptions about f or CDF of random vector X (X 1,..., X p

More information

STAT 450: Statistical Theory. Distribution Theory. Reading in Casella and Berger: Ch 2 Sec 1, Ch 4 Sec 1, Ch 4 Sec 6.

STAT 450: Statistical Theory. Distribution Theory. Reading in Casella and Berger: Ch 2 Sec 1, Ch 4 Sec 1, Ch 4 Sec 6. STAT 450: Statistical Theory Distribution Theory Reading in Casella and Berger: Ch 2 Sec 1, Ch 4 Sec 1, Ch 4 Sec 6. Example: Why does t-statistic have t distribution? Ingredients: Sample X 1,...,X n from

More information

2 Functions of random variables

2 Functions of random variables 2 Functions of random variables A basic statistical model for sample data is a collection of random variables X 1,..., X n. The data are summarised in terms of certain sample statistics, calculated as

More information

Change Of Variable Theorem: Multiple Dimensions

Change Of Variable Theorem: Multiple Dimensions Change Of Variable Theorem: Multiple Dimensions Moulinath Banerjee University of Michigan August 30, 01 Let (X, Y ) be a two-dimensional continuous random vector. Thus P (X = x, Y = y) = 0 for all (x,

More information

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued Chapter 3 sections Chapter 3 - continued 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions

More information

MTH739U/P: Topics in Scientific Computing Autumn 2016 Week 6

MTH739U/P: Topics in Scientific Computing Autumn 2016 Week 6 MTH739U/P: Topics in Scientific Computing Autumn 16 Week 6 4.5 Generic algorithms for non-uniform variates We have seen that sampling from a uniform distribution in [, 1] is a relatively straightforward

More information

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued Chapter 3 sections 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions 3.6 Conditional

More information

MAS223 Statistical Inference and Modelling Exercises

MAS223 Statistical Inference and Modelling Exercises MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,

More information

Lecture 3: Random variables, distributions, and transformations

Lecture 3: Random variables, distributions, and transformations Lecture 3: Random variables, distributions, and transformations Definition 1.4.1. A random variable X is a function from S into a subset of R such that for any Borel set B R {X B} = {ω S : X(ω) B} is an

More information

Chapter 4 Multiple Random Variables

Chapter 4 Multiple Random Variables Review for the previous lecture Theorems and Examples: How to obtain the pmf (pdf) of U = g ( X Y 1 ) and V = g ( X Y) Chapter 4 Multiple Random Variables Chapter 43 Bivariate Transformations Continuous

More information

Math 3215 Intro. Probability & Statistics Summer 14. Homework 5: Due 7/3/14

Math 3215 Intro. Probability & Statistics Summer 14. Homework 5: Due 7/3/14 Math 325 Intro. Probability & Statistics Summer Homework 5: Due 7/3/. Let X and Y be continuous random variables with joint/marginal p.d.f. s f(x, y) 2, x y, f (x) 2( x), x, f 2 (y) 2y, y. Find the conditional

More information

Tangent Planes, Linear Approximations and Differentiability

Tangent Planes, Linear Approximations and Differentiability Jim Lambers MAT 80 Spring Semester 009-10 Lecture 5 Notes These notes correspond to Section 114 in Stewart and Section 3 in Marsden and Tromba Tangent Planes, Linear Approximations and Differentiability

More information

Multivariate distributions

Multivariate distributions CHAPTER Multivariate distributions.. Introduction We want to discuss collections of random variables (X, X,..., X n ), which are known as random vectors. In the discrete case, we can define the density

More information

STAT 450. Moment Generating Functions

STAT 450. Moment Generating Functions STAT 450 Moment Generating Functions There are many uses of generating functions in mathematics. We often study the properties of a sequence a n of numbers by creating the function a n s n n0 In statistics

More information

p. 6-1 Continuous Random Variables p. 6-2

p. 6-1 Continuous Random Variables p. 6-2 Continuous Random Variables Recall: For discrete random variables, only a finite or countably infinite number of possible values with positive probability (>). Often, there is interest in random variables

More information

Transformation of Probability Densities

Transformation of Probability Densities Transformation of Probability Densities This Wikibook shows how to transform the probability density of a continuous random variable in both the one-dimensional and multidimensional case. In other words,

More information

Mathematical Statistics 1 Math A 6330

Mathematical Statistics 1 Math A 6330 Mathematical Statistics 1 Math A 6330 Chapter 2 Transformations and Expectations Mohamed I. Riffi Department of Mathematics Islamic University of Gaza September 14, 2015 Outline 1 Distributions of Functions

More information

JUST THE MATHS UNIT NUMBER INTEGRATION APPLICATIONS 13 (Second moments of a volume (A)) A.J.Hobson

JUST THE MATHS UNIT NUMBER INTEGRATION APPLICATIONS 13 (Second moments of a volume (A)) A.J.Hobson JUST THE MATHS UNIT NUMBER 13.13 INTEGRATION APPLICATIONS 13 (Second moments of a volume (A)) by A.J.Hobson 13.13.1 Introduction 13.13. The second moment of a volume of revolution about the y-axis 13.13.3

More information

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable Distributions of Functions of Random Variables 5.1 Functions of One Random Variable 5.2 Transformations of Two Random Variables 5.3 Several Random Variables 5.4 The Moment-Generating Function Technique

More information

Definition: If y = f(x), then. f(x + x) f(x) y = f (x) = lim. Rules and formulas: 1. If f(x) = C (a constant function), then f (x) = 0.

Definition: If y = f(x), then. f(x + x) f(x) y = f (x) = lim. Rules and formulas: 1. If f(x) = C (a constant function), then f (x) = 0. Definition: If y = f(x), then Rules and formulas: y = f (x) = lim x 0 f(x + x) f(x). x 1. If f(x) = C (a constant function), then f (x) = 0. 2. If f(x) = x k (a power function), then f (x) = kx k 1. 3.

More information

Chapter 2: Differentiation

Chapter 2: Differentiation Chapter 2: Differentiation Winter 2016 Department of Mathematics Hong Kong Baptist University 1 / 75 2.1 Tangent Lines and Their Slopes This section deals with the problem of finding a straight line L

More information

4 Pairs of Random Variables

4 Pairs of Random Variables B.Sc./Cert./M.Sc. Qualif. - Statistical Theory 4 Pairs of Random Variables 4.1 Introduction In this section, we consider a pair of r.v. s X, Y on (Ω, F, P), i.e. X, Y : Ω R. More precisely, we define a

More information

Continuous Random Variables

Continuous Random Variables Continuous Random Variables Recall: For discrete random variables, only a finite or countably infinite number of possible values with positive probability. Often, there is interest in random variables

More information

4. CONTINUOUS RANDOM VARIABLES

4. CONTINUOUS RANDOM VARIABLES IA Probability Lent Term 4 CONTINUOUS RANDOM VARIABLES 4 Introduction Up to now we have restricted consideration to sample spaces Ω which are finite, or countable; we will now relax that assumption We

More information

Chapter 2: Differentiation

Chapter 2: Differentiation Chapter 2: Differentiation Spring 2018 Department of Mathematics Hong Kong Baptist University 1 / 82 2.1 Tangent Lines and Their Slopes This section deals with the problem of finding a straight line L

More information

Multiple Random Variables

Multiple Random Variables Multiple Random Variables Joint Probability Density Let X and Y be two random variables. Their joint distribution function is F ( XY x, y) P X x Y y. F XY ( ) 1, < x

More information

Random Variables and Their Distributions

Random Variables and Their Distributions Chapter 3 Random Variables and Their Distributions A random variable (r.v.) is a function that assigns one and only one numerical value to each simple event in an experiment. We will denote r.vs by capital

More information

ECONOMICS 207 SPRING 2006 LABORATORY EXERCISE 5 KEY. 8 = 10(5x 2) = 9(3x + 8), x 50x 20 = 27x x = 92 x = 4. 8x 2 22x + 15 = 0 (2x 3)(4x 5) = 0

ECONOMICS 207 SPRING 2006 LABORATORY EXERCISE 5 KEY. 8 = 10(5x 2) = 9(3x + 8), x 50x 20 = 27x x = 92 x = 4. 8x 2 22x + 15 = 0 (2x 3)(4x 5) = 0 ECONOMICS 07 SPRING 006 LABORATORY EXERCISE 5 KEY Problem. Solve the following equations for x. a 5x 3x + 8 = 9 0 5x 3x + 8 9 8 = 0(5x ) = 9(3x + 8), x 0 3 50x 0 = 7x + 7 3x = 9 x = 4 b 8x x + 5 = 0 8x

More information

The Multivariate Normal Distribution 1

The Multivariate Normal Distribution 1 The Multivariate Normal Distribution 1 STA 302 Fall 2017 1 See last slide for copyright information. 1 / 40 Overview 1 Moment-generating Functions 2 Definition 3 Properties 4 χ 2 and t distributions 2

More information

1 Solution to Problem 2.1

1 Solution to Problem 2.1 Solution to Problem 2. I incorrectly worked this exercise instead of 2.2, so I decided to include the solution anyway. a) We have X Y /3, which is a - function. It maps the interval, ) where X lives) onto

More information

Joint Probability Distributions and Random Samples (Devore Chapter Five)

Joint Probability Distributions and Random Samples (Devore Chapter Five) Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete

More information

Lecture 1: August 28

Lecture 1: August 28 36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random

More information

Lecture 13 - Wednesday April 29th

Lecture 13 - Wednesday April 29th Lecture 13 - Wednesday April 29th jacques@ucsdedu Key words: Systems of equations, Implicit differentiation Know how to do implicit differentiation, how to use implicit and inverse function theorems 131

More information

Multivariable Calculus and Matrix Algebra-Summer 2017

Multivariable Calculus and Matrix Algebra-Summer 2017 Multivariable Calculus and Matrix Algebra-Summer 017 Homework 4 Solutions Note that the solutions below are for the latest version of the problems posted. For those of you who worked on an earlier version

More information

conditional cdf, conditional pdf, total probability theorem?

conditional cdf, conditional pdf, total probability theorem? 6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random

More information

Sec. 14.3: Partial Derivatives. All of the following are ways of representing the derivative. y dx

Sec. 14.3: Partial Derivatives. All of the following are ways of representing the derivative. y dx Math 2204 Multivariable Calc Chapter 14: Partial Derivatives I. Review from math 1225 A. First Derivative Sec. 14.3: Partial Derivatives 1. Def n : The derivative of the function f with respect to the

More information

Lecture 3. Probability - Part 2. Luigi Freda. ALCOR Lab DIAG University of Rome La Sapienza. October 19, 2016

Lecture 3. Probability - Part 2. Luigi Freda. ALCOR Lab DIAG University of Rome La Sapienza. October 19, 2016 Lecture 3 Probability - Part 2 Luigi Freda ALCOR Lab DIAG University of Rome La Sapienza October 19, 2016 Luigi Freda ( La Sapienza University) Lecture 3 October 19, 2016 1 / 46 Outline 1 Common Continuous

More information

Complex Differentials and the Stokes, Goursat and Cauchy Theorems

Complex Differentials and the Stokes, Goursat and Cauchy Theorems Complex Differentials and the Stokes, Goursat and Cauchy Theorems Benjamin McKay June 21, 2001 1 Stokes theorem Theorem 1 (Stokes) f(x, y) dx + g(x, y) dy = U ( g y f ) dx dy x where U is a region of the

More information

Continuous Random Variables

Continuous Random Variables 1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables

More information

STAT Chapter 5 Continuous Distributions

STAT Chapter 5 Continuous Distributions STAT 270 - Chapter 5 Continuous Distributions June 27, 2012 Shirin Golchi () STAT270 June 27, 2012 1 / 59 Continuous rv s Definition: X is a continuous rv if it takes values in an interval, i.e., range

More information

2.2 Separable Equations

2.2 Separable Equations 2.2 Separable Equations Definition A first-order differential equation that can be written in the form Is said to be separable. Note: the variables of a separable equation can be written as Examples Solve

More information

Lecture 4. Continuous Random Variables and Transformations of Random Variables

Lecture 4. Continuous Random Variables and Transformations of Random Variables Math 408 - Mathematical Statistics Lecture 4. Continuous Random Variables and Transformations of Random Variables January 25, 2013 Konstantin Zuev (USC) Math 408, Lecture 4 January 25, 2013 1 / 13 Agenda

More information

Chapter 4. Continuous Random Variables 4.1 PDF

Chapter 4. Continuous Random Variables 4.1 PDF Chapter 4 Continuous Random Variables In this chapter we study continuous random variables. The linkage between continuous and discrete random variables is the cumulative distribution (CDF) which we will

More information

Exercises and Answers to Chapter 1

Exercises and Answers to Chapter 1 Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean

More information

IEOR 4703: Homework 2 Solutions

IEOR 4703: Homework 2 Solutions IEOR 4703: Homework 2 Solutions Exercises for which no programming is required Let U be uniformly distributed on the interval (0, 1); P (U x) = x, x (0, 1). We assume that your computer can sequentially

More information

Lecture 3. David Aldous. 31 August David Aldous Lecture 3

Lecture 3. David Aldous. 31 August David Aldous Lecture 3 Lecture 3 David Aldous 31 August 2015 This size-bias effect occurs in other contexts, such as class size. If a small Department offers two courses, with enrollments 90 and 10, then average class (faculty

More information

Partial Derivatives Formulas. KristaKingMath.com

Partial Derivatives Formulas. KristaKingMath.com Partial Derivatives Formulas KristaKingMath.com Domain and range of a multivariable function A function f of two variables is a rule that assigns to each ordered pair of real numbers (x, y) in a set D

More information

Introduction to Probability Theory

Introduction to Probability Theory Introduction to Probability Theory Ping Yu Department of Economics University of Hong Kong Ping Yu (HKU) Probability 1 / 39 Foundations 1 Foundations 2 Random Variables 3 Expectation 4 Multivariate Random

More information

The linear model is the most fundamental of all serious statistical models encompassing:

The linear model is the most fundamental of all serious statistical models encompassing: Linear Regression Models: A Bayesian perspective Ingredients of a linear model include an n 1 response vector y = (y 1,..., y n ) T and an n p design matrix (e.g. including regressors) X = [x 1,..., x

More information

3 Algebraic Methods. we can differentiate both sides implicitly to obtain a differential equation involving x and y:

3 Algebraic Methods. we can differentiate both sides implicitly to obtain a differential equation involving x and y: 3 Algebraic Methods b The first appearance of the equation E Mc 2 in Einstein s handwritten notes. So far, the only general class of differential equations that we know how to solve are directly integrable

More information

19 : Slice Sampling and HMC

19 : Slice Sampling and HMC 10-708: Probabilistic Graphical Models 10-708, Spring 2018 19 : Slice Sampling and HMC Lecturer: Kayhan Batmanghelich Scribes: Boxiang Lyu 1 MCMC (Auxiliary Variables Methods) In inference, we are often

More information

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows. Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage

More information

This exam is closed book and closed notes. (You will have access to a copy of the Table of Common Distributions given in the back of the text.

This exam is closed book and closed notes. (You will have access to a copy of the Table of Common Distributions given in the back of the text. TEST #3 STA 5326 December 4, 214 Name: Please read the following directions. DO NOT TURN THE PAGE UNTIL INSTRUCTED TO DO SO Directions This exam is closed book and closed notes. (You will have access to

More information

10550 PRACTICE FINAL EXAM SOLUTIONS. x 2 4. x 2 x 2 5x +6 = lim x +2. x 2 x 3 = 4 1 = 4.

10550 PRACTICE FINAL EXAM SOLUTIONS. x 2 4. x 2 x 2 5x +6 = lim x +2. x 2 x 3 = 4 1 = 4. 55 PRACTICE FINAL EXAM SOLUTIONS. First notice that x 2 4 x 2x + 2 x 2 5x +6 x 2x. This function is undefined at x 2. Since, in the it as x 2, we only care about what happens near x 2 an for x less than

More information

Sampling Distributions

Sampling Distributions Sampling Distributions In statistics, a random sample is a collection of independent and identically distributed (iid) random variables, and a sampling distribution is the distribution of a function of

More information

Lecture 25: Review. Statistics 104. April 23, Colin Rundel

Lecture 25: Review. Statistics 104. April 23, Colin Rundel Lecture 25: Review Statistics 104 Colin Rundel April 23, 2012 Joint CDF F (x, y) = P [X x, Y y] = P [(X, Y ) lies south-west of the point (x, y)] Y (x,y) X Statistics 104 (Colin Rundel) Lecture 25 April

More information

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

Order Statistics and Distributions

Order Statistics and Distributions Order Statistics and Distributions 1 Some Preliminary Comments and Ideas In this section we consider a random sample X 1, X 2,..., X n common continuous distribution function F and probability density

More information

This does not cover everything on the final. Look at the posted practice problems for other topics.

This does not cover everything on the final. Look at the posted practice problems for other topics. Class 7: Review Problems for Final Exam 8.5 Spring 7 This does not cover everything on the final. Look at the posted practice problems for other topics. To save time in class: set up, but do not carry

More information

Math 312 Lecture Notes Linearization

Math 312 Lecture Notes Linearization Math 3 Lecture Notes Linearization Warren Weckesser Department of Mathematics Colgate University 3 March 005 These notes discuss linearization, in which a linear system is used to approximate the behavior

More information

Chapter 2 Continuous Distributions

Chapter 2 Continuous Distributions Chapter Continuous Distributions Continuous random variables For a continuous random variable X the probability distribution is described by the probability density function f(x), which has the following

More information

Chapter 7. Functions of Random Variables

Chapter 7. Functions of Random Variables Chapter 7. Functions of Random Variables Sections 7.2 -- 7.4: Functions of Discrete Random Variables, Method of Distribution Functions and Method of Transformations in One Dimension Jiaping Wang Department

More information

CALCULUS. Berkant Ustaoğlu CRYPTOLOUNGE.NET

CALCULUS. Berkant Ustaoğlu CRYPTOLOUNGE.NET CALCULUS Berkant Ustaoğlu CRYPTOLOUNGE.NET Secant 1 Definition Let f be defined over an interval I containing u. If x u and x I then f (x) f (u) Q = x u is the difference quotient. Also if h 0, such that

More information

Ordinary Differential Equations

Ordinary Differential Equations Ordinary Differential Equations (MA102 Mathematics II) Shyamashree Upadhyay IIT Guwahati Shyamashree Upadhyay ( IIT Guwahati ) Ordinary Differential Equations 1 / 25 First order ODE s We will now discuss

More information

Multiple Random Variables

Multiple Random Variables Multiple Random Variables This Version: July 30, 2015 Multiple Random Variables 2 Now we consider models with more than one r.v. These are called multivariate models For instance: height and weight An

More information

UNIT 3 INTEGRATION 3.0 INTRODUCTION 3.1 OBJECTIVES. Structure

UNIT 3 INTEGRATION 3.0 INTRODUCTION 3.1 OBJECTIVES. Structure Calculus UNIT 3 INTEGRATION Structure 3.0 Introduction 3.1 Objectives 3.2 Basic Integration Rules 3.3 Integration by Substitution 3.4 Integration of Rational Functions 3.5 Integration by Parts 3.6 Answers

More information

9.07 Introduction to Probability and Statistics for Brain and Cognitive Sciences Emery N. Brown

9.07 Introduction to Probability and Statistics for Brain and Cognitive Sciences Emery N. Brown 9.07 Introduction to Probability and Statistics for Brain and Cognitive Sciences Emery N. Brown I. Objectives Lecture 5: Conditional Distributions and Functions of Jointly Distributed Random Variables

More information

4. Distributions of Functions of Random Variables

4. Distributions of Functions of Random Variables 4. Distributions of Functions of Random Variables Setup: Consider as given the joint distribution of X 1,..., X n (i.e. consider as given f X1,...,X n and F X1,...,X n ) Consider k functions g 1 : R n

More information

Probability and Distributions

Probability and Distributions Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated

More information

BMIR Lecture Series on Probability and Statistics Fall 2015 Discrete RVs

BMIR Lecture Series on Probability and Statistics Fall 2015 Discrete RVs Lecture #7 BMIR Lecture Series on Probability and Statistics Fall 2015 Department of Biomedical Engineering and Environmental Sciences National Tsing Hua University 7.1 Function of Single Variable Theorem

More information

Chapter 5. Random Variables (Continuous Case) 5.1 Basic definitions

Chapter 5. Random Variables (Continuous Case) 5.1 Basic definitions Chapter 5 andom Variables (Continuous Case) So far, we have purposely limited our consideration to random variables whose ranges are countable, or discrete. The reason for that is that distributions on

More information

Introduction to Normal Distribution

Introduction to Normal Distribution Introduction to Normal Distribution Nathaniel E. Helwig Assistant Professor of Psychology and Statistics University of Minnesota (Twin Cities) Updated 17-Jan-2017 Nathaniel E. Helwig (U of Minnesota) Introduction

More information

JUST THE MATHS UNIT NUMBER NUMERICAL MATHEMATICS 6 (Numerical solution) of (ordinary differential equations (A)) A.J.Hobson

JUST THE MATHS UNIT NUMBER NUMERICAL MATHEMATICS 6 (Numerical solution) of (ordinary differential equations (A)) A.J.Hobson JUST THE MATHS UNIT NUMBER 17.6 NUMERICAL MATHEMATICS 6 (Numerical solution) of (ordinary differential equations (A)) by A.J.Hobson 17.6.1 Euler s unmodified method 17.6.2 Euler s modified method 17.6.3

More information

(c) The first thing to do for this problem is to create a parametric curve for C. One choice would be. (cos(t), sin(t)) with 0 t 2π

(c) The first thing to do for this problem is to create a parametric curve for C. One choice would be. (cos(t), sin(t)) with 0 t 2π 1. Let g(x, y) = (y, x) ompute gds for a circle with radius 1 centered at the origin using the line integral. (Hint: use polar coordinates for your parametrization). (a) Write out f((t)) so that f is a

More information

Introduction to Probabilistic Graphical Models: Exercises

Introduction to Probabilistic Graphical Models: Exercises Introduction to Probabilistic Graphical Models: Exercises Cédric Archambeau Xerox Research Centre Europe cedric.archambeau@xrce.xerox.com Pascal Bootcamp Marseille, France, July 2010 Exercise 1: basics

More information

Math Refresher Course

Math Refresher Course Math Refresher Course Columbia University Department of Political Science Fall 2007 Day 2 Prepared by Jessamyn Blau 6 Calculus CONT D 6.9 Antiderivatives and Integration Integration is the reverse of differentiation.

More information

MATH 411 NOTES (UNDER CONSTRUCTION)

MATH 411 NOTES (UNDER CONSTRUCTION) MATH 411 NOTES (NDE CONSTCTION 1. Notes on compact sets. This is similar to ideas you learned in Math 410, except open sets had not yet been defined. Definition 1.1. K n is compact if for every covering

More information

MAS223 Statistical Modelling and Inference Examples

MAS223 Statistical Modelling and Inference Examples Chapter MAS3 Statistical Modelling and Inference Examples Example : Sample spaces and random variables. Let S be the sample space for the experiment of tossing two coins; i.e. Define the random variables

More information

Probability Models. 4. What is the definition of the expectation of a discrete random variable?

Probability Models. 4. What is the definition of the expectation of a discrete random variable? 1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions

More information

PCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities

PCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities PCMI 207 - Introduction to Random Matrix Theory Handout #2 06.27.207 REVIEW OF PROBABILITY THEORY Chapter - Events and Their Probabilities.. Events as Sets Definition (σ-field). A collection F of subsets

More information

Generation from simple discrete distributions

Generation from simple discrete distributions S-38.3148 Simulation of data networks / Generation of random variables 1(18) Generation from simple discrete distributions Note! This is just a more clear and readable version of the same slide that was

More information

Ch3. Generating Random Variates with Non-Uniform Distributions

Ch3. Generating Random Variates with Non-Uniform Distributions ST4231, Semester I, 2003-2004 Ch3. Generating Random Variates with Non-Uniform Distributions This chapter mainly focuses on methods for generating non-uniform random numbers based on the built-in standard

More information

5.6 Logarithmic and Exponential Equations

5.6 Logarithmic and Exponential Equations SECTION 5.6 Logarithmic and Exponential Equations 305 5.6 Logarithmic and Exponential Equations PREPARING FOR THIS SECTION Before getting started, review the following: Solving Equations Using a Graphing

More information

UCSD ECE 153 Handout #20 Prof. Young-Han Kim Thursday, April 24, Solutions to Homework Set #3 (Prepared by TA Fatemeh Arbabjolfaei)

UCSD ECE 153 Handout #20 Prof. Young-Han Kim Thursday, April 24, Solutions to Homework Set #3 (Prepared by TA Fatemeh Arbabjolfaei) UCSD ECE 53 Handout #0 Prof. Young-Han Kim Thursday, April 4, 04 Solutions to Homework Set #3 (Prepared by TA Fatemeh Arbabjolfaei). Time until the n-th arrival. Let the random variable N(t) be the number

More information

Physics 403 Probability Distributions II: More Properties of PDFs and PMFs

Physics 403 Probability Distributions II: More Properties of PDFs and PMFs Physics 403 Probability Distributions II: More Properties of PDFs and PMFs Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Last Time: Common Probability Distributions

More information

Stat 5101 Lecture Slides: Deck 8 Dirichlet Distribution. Charles J. Geyer School of Statistics University of Minnesota

Stat 5101 Lecture Slides: Deck 8 Dirichlet Distribution. Charles J. Geyer School of Statistics University of Minnesota Stat 5101 Lecture Slides: Deck 8 Dirichlet Distribution Charles J. Geyer School of Statistics University of Minnesota 1 The Dirichlet Distribution The Dirichlet Distribution is to the beta distribution

More information

S6880 #7. Generate Non-uniform Random Number #1

S6880 #7. Generate Non-uniform Random Number #1 S6880 #7 Generate Non-uniform Random Number #1 Outline 1 Inversion Method Inversion Method Examples Application to Discrete Distributions Using Inversion Method 2 Composition Method Composition Method

More information

1 Joint and marginal distributions

1 Joint and marginal distributions DECEMBER 7, 204 LECTURE 2 JOINT (BIVARIATE) DISTRIBUTIONS, MARGINAL DISTRIBUTIONS, INDEPENDENCE So far we have considered one random variable at a time. However, in economics we are typically interested

More information

MATH34032 Mid-term Test 10.00am 10.50am, 26th March 2010 Answer all six question [20% of the total mark for this course]

MATH34032 Mid-term Test 10.00am 10.50am, 26th March 2010 Answer all six question [20% of the total mark for this course] MATH3432: Green s Functions, Integral Equations and the Calculus of Variations 1 MATH3432 Mid-term Test 1.am 1.5am, 26th March 21 Answer all six question [2% of the total mark for this course] Qu.1 (a)

More information

SDS 321: Introduction to Probability and Statistics

SDS 321: Introduction to Probability and Statistics SDS 321: Introduction to Probability and Statistics Lecture 17: Continuous random variables: conditional PDF Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin

More information

When Are Two Random Variables Independent?

When Are Two Random Variables Independent? When Are Two Random Variables Independent? 1 Introduction. Almost all of the mathematics of inferential statistics and sampling theory is based on the behavior of mutually independent random variables,

More information

STA 256: Statistics and Probability I

STA 256: Statistics and Probability I Al Nosedal. University of Toronto. Fall 2017 My momma always said: Life was like a box of chocolates. You never know what you re gonna get. Forrest Gump. Exercise 4.1 Let X be a random variable with p(x)

More information

Mathematics 426 Robert Gross Homework 9 Answers

Mathematics 426 Robert Gross Homework 9 Answers Mathematics 4 Robert Gross Homework 9 Answers. Suppose that X is a normal random variable with mean µ and standard deviation σ. Suppose that PX > 9 PX

More information

Chain Rule. MATH 311, Calculus III. J. Robert Buchanan. Spring Department of Mathematics

Chain Rule. MATH 311, Calculus III. J. Robert Buchanan. Spring Department of Mathematics 3.33pt Chain Rule MATH 311, Calculus III J. Robert Buchanan Department of Mathematics Spring 2019 Single Variable Chain Rule Suppose y = g(x) and z = f (y) then dz dx = d (f (g(x))) dx = f (g(x))g (x)

More information

Math 416 Lecture 2 DEFINITION. Here are the multivariate versions: X, Y, Z iff P(X = x, Y = y, Z =z) = p(x, y, z) of X, Y, Z iff for all sets A, B, C,

Math 416 Lecture 2 DEFINITION. Here are the multivariate versions: X, Y, Z iff P(X = x, Y = y, Z =z) = p(x, y, z) of X, Y, Z iff for all sets A, B, C, Math 416 Lecture 2 DEFINITION. Here are the multivariate versions: PMF case: p(x, y, z) is the joint Probability Mass Function of X, Y, Z iff P(X = x, Y = y, Z =z) = p(x, y, z) PDF case: f(x, y, z) is

More information

The Multivariate Normal Distribution 1

The Multivariate Normal Distribution 1 The Multivariate Normal Distribution 1 STA 302 Fall 2014 1 See last slide for copyright information. 1 / 37 Overview 1 Moment-generating Functions 2 Definition 3 Properties 4 χ 2 and t distributions 2

More information

OBJECTIVES Use the area under a graph to find total cost. Use rectangles to approximate the area under a graph.

OBJECTIVES Use the area under a graph to find total cost. Use rectangles to approximate the area under a graph. 4.1 The Area under a Graph OBJECTIVES Use the area under a graph to find total cost. Use rectangles to approximate the area under a graph. 4.1 The Area Under a Graph Riemann Sums (continued): In the following

More information

dy = f( x) dx = F ( x)+c = f ( x) dy = f( x) dx

dy = f( x) dx = F ( x)+c = f ( x) dy = f( x) dx Antiderivatives and The Integral Antiderivatives Objective: Use indefinite integral notation for antiderivatives. Use basic integration rules to find antiderivatives. Another important question in calculus

More information

Random Variables (Continuous Case)

Random Variables (Continuous Case) Chapter 6 Random Variables (Continuous Case) Thus far, we have purposely limited our consideration to random variables whose ranges are countable, or discrete. The reason for that is that distributions

More information

Practice Examination # 3

Practice Examination # 3 Practice Examination # 3 Sta 23: Probability December 13, 212 This is a closed-book exam so do not refer to your notes, the text, or any other books (please put them on the floor). You may use a single

More information