Introduction to Probability Theory for Graduate Economics Fall 2008

Similar documents
Chapter 2. Random Variable. Define single random variables in terms of their PDF and CDF, and calculate moments such as the mean and variance.

ECON 5350 Class Notes Review of Probability and Distribution Theory

n px p x (1 p) n x. p x n(n 1)... (n x + 1) x!

Chapter 2. Discrete Distributions

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

EE/CpE 345. Modeling and Simulation. Fall Class 5 September 30, 2002

Things to remember when learning probability distributions:

Probability Distributions Columns (a) through (d)

Plotting data is one method for selecting a probability distribution. The following

Chapter 2: The Random Variable

Lecture Notes 2 Random Variables. Discrete Random Variables: Probability mass function (pmf)

Lecture Notes 2 Random Variables. Random Variable

Computer Science, Informatik 4 Communication and Distributed Systems. Simulation. Discrete-Event System Simulation. Dr.

Chapter 3 Single Random Variables and Probability Distributions (Part 1)

Random Variable. Discrete Random Variable. Continuous Random Variable. Discrete Random Variable. Discrete Probability Distribution

STAT 3610: Review of Probability Distributions

Random Variables and Their Distributions

Random variable X is a mapping that maps each outcome s in the sample space to a unique real number x, x. X s. Real Line

Topic 3: The Expectation of a Random Variable

Class 26: review for final exam 18.05, Spring 2014

Stochastic processes Lecture 1: Multiple Random Variables Ch. 5

IEOR 3106: Introduction to Operations Research: Stochastic Models. Professor Whitt. SOLUTIONS to Homework Assignment 1

1 Review of Probability

Lecture 1: August 28

Chapter 5. Chapter 5 sections

15 Discrete Distributions

Chapter 5. Statistical Models in Simulations 5.1. Prof. Dr. Mesut Güneş Ch. 5 Statistical Models in Simulations

Actuarial Science Exam 1/P

Econ 371 Problem Set #1 Answer Sheet

1 Presessional Probability

STAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)

3. Probability and Statistics

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

Probability and Distributions

Brief Review of Probability

Stochastic Processes. Review of Elementary Probability Lecture I. Hamid R. Rabiee Ali Jalali

ACM 116: Lectures 3 4

Review 1: STAT Mark Carpenter, Ph.D. Professor of Statistics Department of Mathematics and Statistics. August 25, 2015

Ch. 5 Joint Probability Distributions and Random Samples

SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416)

Random variables, distributions and limit theorems

Review for the previous lecture

18.440: Lecture 28 Lectures Review

Random variable X is a mapping that maps each outcome s in the sample space to a unique real number x, < x <. ( ) X s. Real Line

Random variables. DS GA 1002 Probability and Statistics for Data Science.

Chapter 2: Discrete Distributions. 2.1 Random Variables of the Discrete Type

Contents 1. Contents

Week 1 Quantitative Analysis of Financial Markets Distributions A

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

1.1 Review of Probability Theory

Chapter 3 Common Families of Distributions

Week 2. Review of Probability, Random Variables and Univariate Distributions

BASICS OF PROBABILITY

Midterm Exam 1 (Solutions)

Review of Probability Theory

Notes 12 Autumn 2005

7 Random samples and sampling distributions

RS Chapter 2 Random Variables 9/28/2017. Chapter 2. Random Variables

Discrete Distributions

Review of Elementary Probability Lecture I Hamid R. Rabiee

Chapter 1. Sets and probability. 1.3 Probability space

Probability. Table of contents

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

Chapter 5 continued. Chapter 5 sections

Discrete Distributions

Transform Techniques - CF

ECON Fundamentals of Probability

Binomial and Poisson Probability Distributions

Mathematical statistics

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

CMPSCI 240: Reasoning Under Uncertainty

Random variables (discrete)

Midterm Exam 1 Solution

LIST OF FORMULAS FOR STK1100 AND STK1110

Lecture 4: Probability and Discrete Random Variables

Lecture 2: Repetition of probability theory and statistics

Chapter 4 : Expectation and Moments

ELEG 3143 Probability & Stochastic Process Ch. 2 Discrete Random Variables

Probability Theory and Statistics. Peter Jochumzen

18440: Probability and Random variables Quiz 1 Friday, October 17th, 2014

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

Stochastic Models of Manufacturing Systems

Course: ESO-209 Home Work: 1 Instructor: Debasis Kundu

Chapter 1 Statistical Reasoning Why statistics? Section 1.1 Basics of Probability Theory

Definition: A random variable X is a real valued function that maps a sample space S into the space of real numbers R. X : S R

TMA4265: Stochastic Processes

18.440: Lecture 28 Lectures Review

Statistical Methods in Particle Physics

Statistics STAT:5100 (22S:193), Fall Sample Final Exam B

Probability & Statistics - FALL 2008 FINAL EXAM

1 Random Variable: Topics

Ching-Han Hsu, BMES, National Tsing Hua University c 2015 by Ching-Han Hsu, Ph.D., BMIR Lab. = a + b 2. b a. x a b a = 12

Lecture 13 (Part 2): Deviation from mean: Markov s inequality, variance and its properties, Chebyshev s inequality

Review of Probability. CS1538: Introduction to Simulations

Institute of Actuaries of India

1: PROBABILITY REVIEW

Introduction. Probability and distributions

UC Berkeley Department of Electrical Engineering and Computer Science. EE 126: Probablity and Random Processes. Problem Set 9 Fall 2007

Moments. Raw moment: February 25, 2014 Normalized / Standardized moment:

Lecture 6: Special probability distributions. Summarizing probability distributions. Let X be a random variable with probability distribution

Transcription:

Introduction to Probability Theory for Graduate Economics Fall 008 Yiğit Sağlam October 10, 008 CHAPTER - RANDOM VARIABLES AND EXPECTATION 1 1 Random Variables A random variable (RV) is a real-valued function over a sample space. Some eamples of a RVare the outcome of a fair coin toss (head or tails), the outcome when we roll a dice (an integer between 1 and 6). In Economics, a RV may represent a wide range of factors: In Job-Search literature, the value of a wage offer is a RV. In Macroeconomics, production of a commodity may include a stochastic component, such as total factor productivity, and this stochastic part is a RV. In Auctions literature, bidders have a private or a common value which also is a RV defined over a specified domain. Random variables are commonly represented by capital letters, whereas lowercase letters denote a realization of a RV. For eample, X represents the outcome of a coin toss, where equals the particular outcome heads. Random variables are of two types: discrete, and continuous. It is noteworthy that a random may be a miture of both discrete and continuous RVs. The feature Discrete or Continuous describes the nature of the domain (or the sample space) of the RV. For instance, the outcome of a coin toss is a discrete RV, while the value of wage offer is a continuous RV. Since a RV has multiple outcomes, one needs a probability model for observing those outcomes. So, the probability distribution of a RV provides information about what values the RV can take and how to assign probabilities to those values. The probability mass function (pmf) of a discrete RV X lists the possible values i for the RV and their respective probabilities p i. The only restrictions on the probabilities are the following: Every probability must be positive and less than 1: The sum of the probabilities must equal 1: 0 p i 1; i = 1,,..., n. n p i = 1. i=1 PhD Candidate in Economics, Department of Economics, University of Iowa, W10 Pappajohn Business Building, Iowa City, IA 54, yigit-saglam@uiowa.edu, Phone: 1(319)335089, Fa: 1(319)3351956 1 I would like to thank Brent Hickman for his helpful comments. 1

Figure 1 - Density Curve of a Continuous Random Variable The probability distribution of a continuous RV X is described by a density curve. The probability of any event is the area under the density curve and above the values of X that make up the event. Suppose a RV X may take all values over an interval of real numbers. Then the probability that X is in the set of outcomes A, f X (A), is defined to be the area above A and under a curve. The curve, which represents a function f X (X), must satisfy the following: The curve has no negative values: p() > 0;. The total area under the curve is equal to 1: f X () d = 1. Discrete Random Variables.1 Bernoulli Distribution A A Bernoulli Eperiment is a random eperiment the outcome of which is one of the two mutually eclusive events such as failure and success, or head and tail, etc.. Denote the probability for these outcomes; ie, success and failure, are θ and (1 θ). The probabilities remain the same over trials. Let X be a Bernoulli RV. Furthermore, let X equal 1 if the trial is a success, and X equal 0 if the trial is a failure. The probability mass function (pmf), the mean, and the variance of X are: Applications: f(; θ) = θ (1 θ) 1 ; = 0, 1 1 µ = E() = θ (1 θ) 1 = θ, =0 σ = V ar() = 1 ( θ) θ (1 θ) 1 = θ(1 θ). =0 In the Asset Pricing literature, in the case of discontinuous interest rates, the jump in the interest rate is a Bernoulli RV which either occurs in a period with a certain probability or not; see Das (00).

. Binomial Distribution Binomial distribution is general case of Bernoulli distribution when there is a fied number of trials. Let n denote the number of trials. The trials are independent, and the probabilities remain the same over trials. Let X be a Binomial RV that shows the number of successes in n trials. Since the probability of success is θ and the failure is (1 θ) for each trial, the probability mass function of X is the sum of probabilities of ( n ) mutually eclusive events. The probability mass function (pmf), the mean, and the variance of X are: ( ) n f(; θ, n) = θ (1 θ) n ; = 0, 1,..., n n ( ) n µ = E() = θ (1 θ) n = nθ, =0 σ = V ar() = n =0 ( ) n ( nθ) θ (1 θ) n = nθ(1 θ). It is noteworthy that one can derive the Geometric and Hypergeometric distributions using the Bernoulli distribution. Applications: In the Labor Supply literature, the number of days a person is absent from work is a Binomial RV assuming that the days are not serially correlated; see Johansson and Palme (1996). In their paper, under some assumptions, Johansson and Palme maimized the Binomial log likelihood function to estimate the parameters. Eercise 1.1 Suppose that an airplane engine will fail, when in flight, with probability (1 θ) independently from engine to engine; suppose that the airplane will make a successful flight if at least 50 percent of its engines remain operative. For what values of θ is a four-engine plane preferable to a two-engine plane? Answer 1.1 As each engine is assumed to fail or function independently of what happens with the other engines, it follows that the number of engines remaining operative is a binomial RV. Hence, the probability that a four-engine plane makes a successful flight is: P r(x ) = ( ) 4 θ (1 θ) 4 + ( ) 4 3 = 6θ (1 θ) + 4θ 3 (1 θ) + θ 4 θ 3 (1 θ) 4 3 + ( ) 4 4 θ 4 (1 θ) 4 4 whereas the corresponding probability for a two-engine plane is: ( ) ( ) P r(x 1) = θ 1 (1 θ) 1 + θ (1 θ) = θ 1 (1 θ) 1 + θ 1 Hence, the four-engine plane is safer if: 6θ (1 θ) + 4θ 3 (1 θ) + θ 4 θ 1 (1 θ) 1 + θ 6θ 1 (1 θ) + 4θ (1 θ) + θ 3 θ θ 3 3

.3 Multinomial Distribution The multinomial distribution is a generalization of the binomial distribution. The binomial distribution is the probability distribution of the number of successes in n independent Bernoulli trials, with the same probability of success on each trial. Instead of each trial resulting in success or failure, we assume that each trial results in one of some fied finite number k of possible outcomes, with probabilities p 1,..., p k, and there are n independent trials. We can use a RV X i to indicate the number of times outcome number i was observed over the n trials. The probability mass function (pmf), the mean, and the variance of X 1,..., X k are: f( 1,..., k ; p 1,..., p k, n) = µ i = E( i ) = np i ; i = 1,..., k, σ i = V ar( i ) = np i (1 p i ); i = 1,..., k. {( ) n k 1,..., k i=1 p i i 0 otherwise. if k i=1 i = n, It is noteworthy that multinomial distribution is a joint distribution, which will be covered more thoroughly in Chapter 4. However, it is introduced here merely as another discrete probability distribution..4 Poisson Distribution Poisson distribution is particularly useful for modeling time instants at which events occur. Assume that we are going to observe the interested event for a period of time T. The time instant at which we start to observe the events will be labeled 0, the origin of time scale. The number of events in this time interval (0, T ) is a RV X. The probability mass function (pmf), the mean, and the variance of X are: f(; λ) = λ e λ ; = 0, 1,...! µ = E() = λ e λ = λ,! =0 σ = V ar() = ( λ) λ e λ = λ.! =0 Approimation to Binomial Distribution: An important property of the Poisson rv is that it may be used to approimate a binomial RV when the number of trials n is large, and the probability of success θ is small. To see this, let λ = nθ, and consider the following equations: n! P r(x = i) = (n i)! i! θi (1 θ) n i ( ) n! λ i = (1 λ (n i)! i! n n )n i For n large, θ small: = (1 λ/n) n ep( λ); n(n 1)... (n i + 1) λ i n i i! (1 λ/n) n ; for i = 0, 1,..., n. (1 λ/n) i n(n 1)... (n i + 1) n i 1; (1 λ/n) i 1 Finally, we can approimate the probability in the following way: P r(x = i) ep( λ) λi i!. 4

Applications: There is a certain relationship between Eponential and Poisson distributions. To wit, eponential distribution can be derived from Poisson distribution. We will eplore this relationship in the section for the Eponential distribution. Eponential and Poisson distribution are both useful in modeling stochastic processes with certain features. In particular, eponentially distributed RVs are memoryless processes. As an eample from the Industrial Organization literature, the distribution of increments for an innovation can be a Poisson distribution, if it a memoryless process ; see Hopenhayn and Squintani (004). We will see more about the stochastic processes in Chapter 5. In the Labor literature, the distribution of job offers in a specific time length can be a Poisson distribution; see Van Den Berg (1990). Eercise 1. Suppose that the number of typological errors on a single page of a book has a Poisson distribution with parameter λ = 1. Calculate the probability that there is at least one errors on a page. Answer 1. To calculate the probability that a page has at least one errors, we first calculate the probability that there is no error on the page: ] P r(x 1) = 1 P r(x = 0) = 1 [ep( λ) λ0 0! = 1 ep( 1) 0.6333 Eercise 1.3 If the number of wage offers that a person gets each day is a Poisson RV with parameter λ = 3, what is the probability that no wages are offered today? Answer 1.3 To calculate the probability that a person has no wage offers: ] P r(x = 0) = [ep( λ) λ0 0! = ep( 3) 0.05 5

3 Continuous Random Variables 3.1 Uniform Distribution The uniform distribution defines equal probability over a given range for a continuous distribution. For this reason, it is important as a reference distribution. The probability distribution function, mean, and variance of X are: f() = 1 ; [A, B], B A B ( ) 1 µ = E() = d = A + B, A B A B ( σ = V ar() = A + B ) ( ) 1 d = B A A (B A). 1 Applications: One of the most important applications of the uniform distribution is in the generation of random numbers. That is, almost all random number generators generate random numbers on the [0,1] interval. For other distributions, some transformation is applied to the uniform random numbers. This is called Inverse Transform Sampling Method. The procedure is as follows: First, one generates random draws y from the uniform distribution defined over [0,1]. Then, if the cumulative distribution function is known and continuous over the domain, then one can find the value for which the cumulative probability equals y. This method is very useful in theoretical work. However, this method may not work efficiently for some distributions such as the normal distribution. More details about this method will be covered in Chapter 3. 3. Eponential Distribution We developed the distribution of the number of occurrences in the interval (0, T ), which is a Poisson distribution with parameter λ. Now, let T be the time at which the first event occurs. Then, the RV T is continuous. Consider the event T > t, that the time of the first event is greater than t. The probability that there is zero event until time t becomes: P r (T > t) = P r ( = 0; λ) = λ0 e λ 0! = e λ = e ct where λ equals ct, proportional to t. Then, the cumulative distribution of T is: F T (T ) = P r (T < t) = 1 e ct The probability distribution function, mean, and variance of T are: { 1 f(t; θ) = θ e t θ ; > 0 0 otherwise µ = E() = σ = V ar() = 0 0 t 1 θ e t θ dt = θ, (t θ) 1 θ e t θ dt = θ. 6

Applications: A very important feature of the Eponential distribution is that it is a memoryless process. To wit, let X be an Eponentially distributed RV, and consider the probability that X t. P r(x t) = P r (X t o + t X t o ) ; t o 0, and for t > 0. In words, the probability that the first occurrence happens at a time X t is equivalent to the probability that the first occurrence happens at time X t o + t, given that it has not yet occurred until time t o. Whenever it is appropriate, the memoryless property of Eponential distribution is useful in economics, as one does not have to keep track of the whole history to compute the probability distribution of a variable in the current state. 3.3 Gamma Distribution Suppose that we wish to eamine the continuous RV X measuring the required length for r events to occur. Suppose that we are observing a Poisson process starting at time 0 and let T r be the time of an occurrence of the r th event. Also, let t any fied positive number and consider the event T r > t; ie, the time of the r th event is greater than t. This event {T r > t} is equivalent to {X r 1}, where X is the number of events that occur in [0, t]. Thus, X is a Poisson RV with parameter λ = ct: P r (T r > t) = P r (X r 1) = k = 0 r 1 ep( λ) λk k! F Tr (t) = 1 P r (T r t) = 1 k = 0 r 1 ep( ct) (ct)k k! f Tr (t) = cr t r 1 ep( ct), t > 0 (r 1)! where f Tr (t) is the pdf of X, and it is still discrete due to r. This is a special case of Gamma Probability Law and is called Erlang Law. However, since our aim is to make this epression valid for any positive real number r, we will employ the Gamma Function Γ(r). To derive the pdf of a Gamma RV, let y = /β and consider the following: Γ(r) = 0 f() = α 1 ep( /β) Γ(α)β α y r 1 ep( y)dy; for any r > 0. where {r, t, c} equal {α,, β 1}. Finally, α and β are usually known as shape and scale parameters, respectively. One can also see the Gamma distribution in the following way: The sum of independently and eponentially distributed RVs is a Gamma distributed RV. The probability distribution function, mean, and variance of X are: Applications: f(, α, β) = α 1 ep( /β) Γ(α)β α ; for r > 0, µ = E() = σ = V ar() = 0 0 f(, α, β) d = αβ, ( αβ) f(, α, β) d = αβ. Chi-square distribution is a special case of Gamma distribution: X Gamma (α = 1/, 1, 3/,, 5/,..., β = ) Eponential distribution, which we will cover net, is a special case of Gamma distribution: X Gamma (α = 1, β) 7

3.4 Normal Distribution The normal distribution, also called the Gaussian distribution, is an important family of continuous probability distributions. For both theoretical and practical reasons, the normal distribution is probably the most important distribution in statistics. For eample, many classical statistical tests are based on the assumption that the data follow a normal distribution. In modeling applications, such as linear and non-linear regression, the error term is often assumed to follow a normal distribution with fied location and scale. Also, the normal distribution is used for inference; ie, to find significance levels in many hypothesis tests and confidence intervals. The probability distribution function, mean, and variance of X are: Applications: f(; µ, σ) = 1 σ ( µ) π e σ ; [A, B], µ = E() = σ = V ar() = 1 σ π e ( µ) σ d = µ, ( µ) 1 σ ( µ) π e σ d = σ. The normal distribution is widely used. Part of the appeal is that it is well behaved and mathematically tractable. However, the central limit theorem provides a theoretical basis for why it has wide applicability. The central limit theorem basically states that as the sample size (N) becomes large, the following occur: * The sampling distribution of the mean becomes approimately normal regardless of the distribution of the original variable. * The sampling distribution of the mean is centered at the population mean, µ, of the original variable. In addition, the standard deviation of the sampling distribution of the mean approaches σ/ n. As mentioned above, normal distribution is widely used in econometrics in linear and nonlinear regression, and in inference. 3.5 Log-Normal Distribution The log-normal distribution is the single-tailed probability distribution of any random variablerv whose logarithm is normally distributed. If X is a RV with a normal distribution, then Y = ep(x) has a log-normal distribution; likewise, if Y is log-normally distributed, then log(y ) is normally distributed. The probability distribution function, mean, and variance of X are: Applications: 1 f(; µ, σ) = σ (log µ) π e σ ; [A, B], µ = E() = σ = V ar() = 1 σ (log µ) π e σ d = e µ+σ /, ( ) e µ+σ / 1 σ (log µ) π e σ d = ( ) e σ 1 e µ+σ. In Macroeconomics, the productivity shock to the production function of a commodity is taken as a log-normal RV. This assumption may be convenient as the log-normal distribution is defined over the positive real numbers, and the distribution function is tractable. 8

4 Epectation, Moments, and Moment Generating Functions 4.1 Epectation Definition Epectation Let X be a RV. If X is a continuous RV with a pdf f() and of X is E(X) = f() d <, then the epectation f() d. If X is a discrete RV with a pmf f() and f() <, then the epectation of X is E(X) = f() Epectation of X is also known as the epected value, or the mean of X. Theorem Let X be a RV, and g(.) be a function of X. Then: If X is a continuous RV with a pdf f() and g() f() d <, then E(g(X)) = g() f() d. If X is a discrete RV with a pmf f() and g() f() <, then E(g(X)) = g() f() Theorem Epectation is a linear operator. Let X be a RV, and g 1 (.) and g (.) be functions of X. Suppose that the epectations of g 1 (X) and g (X) eist. Then, for any constants k 1 and k, the epectation of k 1 g 1 (X) + k g (X) eists and is given by:: E [k 1 g 1 (X) + k g (X)] = k 1 E [g 1 (X)] + k E [g (X)]. Proposition Law of Total Epectation (or Law of Iterated Epectations) Let X and Y be two RVs. Define the conditional epectation Then the epectation of X satisfies E(X Y )(y) = E(X Y = y). E(X) = E [E(X Y )]. 9

Proof: Discrete Case: E [E(X Y )] = y = y = y = = = E(X Y = y) P r(y = y) ( ) P r(x = Y = y) P r(y = y) P r(x = Y = y) P r(y = y) P r(y = y X = ) P r(x = ) y P r(x = ) P r(x = ) ( ) P r(y = y X = ) y 4. Some Special Epectations = E(X) Definition Mean Let X be a RV whose epectation eists. The mean value of X is the defined to be µ = E(X). [ Definition Variance Let X be a RV with finite mean µ and such that E ( µ) ] eists. Then, [ the variance of X is defined to be σ = E ( µ) ] = V ar(x). Moreover, V ar(x) = σ = E(X ) [E(X)] = E(X ) µ. Definition Covariance Covariance is a measure of how much two variables change together. In particular, the variance is a special case of the covariance when the two variables are identical. Let X and Y be two RVs whose epectations eist. The covariance of X and Y is the defined to be Cov(X, Y ) = E(XY ) E(X) E(Y ) = E(XY ) µ X µ Y. It is noteworthy that if two RVs are independent, then the covariance equals 0. However, the converse is not true in general. Definition Moment Generating Functions Let X be a RV such that for some h > 0, the epectation of e tx eists for h < t < h. The moment generating function of X (henceforth: mgf) is defined to be the function M(t) = E(e tx ); for t [ h, h]. 10

Theorem: While a distribution has a unique mgf, mgf s uniquely identify distributions. Let X and Y be two RVs with mgf s M X and M Y, respectively, eisting in open interval about 0. Then F X (z) = F Y (z); z R M X (t) = M Y (t); t [ h, h] for some h > 0 A nice feature of the mgf s is that one can derive the mean and variance using the mgf s. To wit: M(t) = E(X), M (t) = d dt M(t) = E ( d dt etx ) = E ( X e tx), M (0) = E(X) = µ, M (0) = E(X ) = σ + µ,. { M (m) (0) = E(X m ) = m f() d if X is a continuous RV, m f() if X is a discrete RV where M m (0) is the m th derivative of the mgf of X evaluated at t = 0, which is also known as the m th raw moment around the origin. Definition Central Moments The m th central moment of the probability distribution of a random variablerv X is the moment around the mean and is denoted in the following way: 1. µ 1 = E [ (X µ) 1] = 0. µ m = E [(X µ) m ].. Variance = V ar(x) = σ = µ = E [ (X µ) ]. 3. Skewness = µ 3 = E [ (X µ) 3]. 4. Kurtosis = µ 4 = E [ (X µ) 4]. 11

4.3 Some Eamples for Moment Generating Functions Eercise 3.1 MGF of Binomial Distribution with Parameters n and θ M(t) = E(e tx ) = = n k=0 n ( ) n e tk θ k (1 θ) n k k k=0 ( ) n (θe t ) k (1 θ) n k k = ( θe t + 1 θ ) n. E(X) = µ = M (t = 0) = n ( θe 0 + 1 θ ) n 1 θ e 0 = n θ. E(X ) = M (t = 0) = n (n 1) θ + n θ. V ar(x) = σ = n θ (1 θ). Eercise 3. MGF of Poisson Distribution with Parameter λ M(t) = E(e tx e tk e λ λ k ) = k! k=0 ( λe = e λ t ) k k! k=0 = e λ e λet = e λ(et 1). E(X) = µ = M (t = 0) = λ e 0 e λ(e0 1) = λ. [ (λe E(X ) = M 0 (t = 0) = ) ] + λe 0 e λ(e0 1) = λ + λ. V ar(x) = σ = λ. Eercise 3.3 MGF of Eponential Distribution with Parameter θ M(t) = E(e tx ) = 0 t e /θ e θ = 1 e (1/θ t) d θ 0 1 = 1 θ t ; for t < θ 1. E(X) = µ = M (t = 0) = E(X ) = M (t = 0) = V ar(x) = σ = θ. d θ (1 θ 0) = θ. θ (1 θ 0) 3 = θ. 1

Eercise 3.4 MGF of Normal Distribution with Parameters µ and σ M(t) = E(e tx ) = 1 πσ e t e ( µ) σ d [ ( µµ σ t ) = 1 ] ep πσ σ d = 1 [ ( ( µ + σ t )) ] + σ 4 t + µσ t ep πσ σ d = 1 [ σ 4 t + µσ ] [ t ( ( µ + σ t )) ] ep πσ σ ep σ d [ σ t ] { [ = ep + µt 1 ( ( µ + σ t )) ] } ep πσ σ d [ σ t ] = ep + µt {1} [ σ t ] = ep + µt. E(X) = µ = M (t = 0) = ( µ + σ 0 ) [ σ 0 ] ep + µ0 = µ. [ (µ E(X ) = M (t = 0) = + σ 0 ) ] [ + σ σ 0 ] ep + µ0 = µ + σ. V ar(x) = σ = σ. Eercise 3.5 MGF of LogNormal Distribution does not eist. Even though the LogNormal distribution has finite moments of all orders, the moment generating function is infinite at any positive number. This property is one of the reasons for the fame of the LogNormal distribution. M(t) = E(e tx ) ; for any t > 0. 13

5 References Das, Sanjiv R. The Surprise Element: Jumps in Interest Rates. Journal of Econometrics, 106 (00), 7 65. Johansson, P. and M. Palme. Do Economic Incentives Affect Work Absence? Empirical Evidence Using Swedish Micro Data. Journal of Public Economics, 59 (1996), 195 18. Squintani, F. and H. Hopenhayn. Preemption Games with Private Information. 005 Meeting Papers 80, Society for Economic Dynamics. Van Der Berg, Gerard J. NonStationarity in Job Search Theory. Review of Economic Studies, 57 (1990), 55-77. Ross, Sheldon M. Introduction to Probability Models. Fifth Edition. San Diego: Academic Press, 1993. 14