Method of Moments. which we usually denote by X or sometimes by X n to emphasize that there are n observations.

Similar documents
Severity Models - Special Families of Distributions

Parameter Estimation

Practice Exam 1. (A) (B) (C) (D) (E) You are given the following data on loss sizes:

Exam C Solutions Spring 2005

SPRING 2007 EXAM C SOLUTIONS

STAT 6350 Analysis of Lifetime Data. Probability Plotting

Discrete Distributions Chapter 6

Definition 1.1 (Parametric family of distributions) A parametric distribution is a set of distribution functions, each of which is determined by speci

Errata and updates for ASM Exam C/Exam 4 Manual (Sixth Edition) sorted by date

Errata and updates for ASM Exam C/Exam 4 Manual (Sixth Edition) sorted by page

Mixture distributions in Exams MLC/3L and C/4

STAT 3610: Review of Probability Distributions

SOLUTION FOR HOMEWORK 8, STAT 4372

TAYLOR SERIES [SST 8.8]

Solutions to the Spring 2015 CAS Exam ST

Chapter 5. Continuous Probability Distributions

Frequency Analysis & Probability Plots

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

Continuous Distributions

IEOR 3106: Introduction to Operations Research: Stochastic Models. Professor Whitt. SOLUTIONS to Homework Assignment 2

On the Comparison of Fisher Information of the Weibull and GE Distributions

Applied Statistics and Probability for Engineers. Sixth Edition. Chapter 4 Continuous Random Variables and Probability Distributions.

Chapter 4 Continuous Random Variables and Probability Distributions

Probability Distributions Columns (a) through (d)

SOLUTIONS TO MATH68181 EXTREME VALUES AND FINANCIAL RISK EXAM

Modelling the risk process

Functions of Random Variables

3 Continuous Random Variables

CIVL 7012/8012. Continuous Distributions

ECONOMICS 207 SPRING 2006 LABORATORY EXERCISE 5 KEY. 8 = 10(5x 2) = 9(3x + 8), x 50x 20 = 27x x = 92 x = 4. 8x 2 22x + 15 = 0 (2x 3)(4x 5) = 0

Continuous Distributions

Some worked exercises from Ch 4: Ex 10 p 151

RMSC 2001 Introduction to Risk Management

Chapter 5: Generating Random Numbers from Distributions

Conditional densities, mass functions, and expectations

ACTEX CAS EXAM 3 STUDY GUIDE FOR MATHEMATICAL STATISTICS

Power Series. Part 2 Differentiation & Integration; Multiplication of Power Series. J. Gonzalez-Zugasti, University of Massachusetts - Lowell

Test Problems for Probability Theory ,

ON THE TRUNCATED COMPOSITE WEIBULL-PARETO MODEL

Continuous Probability Distributions. Uniform Distribution

CS Lecture 19. Exponential Families & Expectation Propagation

Exam C. Exam C. Exam C. Exam C. Exam C

Subject CT6. CMP Upgrade 2013/14. CMP Upgrade

Toward a unified approach to fitting loss models

A = (a + 1) 2 = a 2 + 2a + 1

RMSC 2001 Introduction to Risk Management

Basic concepts of probability theory

, find P(X = 2 or 3) et) 5. )px (1 p) n x x = 0, 1, 2,..., n. 0 elsewhere = 40

Basic concepts of probability theory

Actuarial Science Exam 1/P

Spatially Smoothed Kernel Density Estimation via Generalized Empirical Likelihood

Lecture 11: Probability, Order Statistics and Sampling

Exponential Family and Maximum Likelihood, Gaussian Mixture Models and the EM Algorithm. by Korbinian Schwinger

Course 4 Solutions November 2001 Exams

A world-wide investigation of the probability distribution of daily rainfall

SOLVED PROBLEMS ON TAYLOR AND MACLAURIN SERIES

Lecture 5: Moment generating functions

Reliable Inference in Conditions of Extreme Events. Adriana Cornea

The Type I Generalized Half Logistic Distribution Downloaded from jirss.irstat.ir at 11: on Monday August 20th 2018

Math 362, Problem set 1

Chapter (a) (b) f/(n x) = f/(69 10) = f/690

MFM Practitioner Module: Quantitative Risk Management. John Dodson. September 23, 2015

UC Berkeley Department of Electrical Engineering and Computer Sciences. EECS 126: Probability and Random Processes

Chapter 4: CONTINUOUS RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS

Continuous RVs. 1. Suppose a random variable X has the following probability density function: π, zero otherwise. f ( x ) = sin x, 0 < x < 2

Dr. Maddah ENMG 617 EM Statistics 10/15/12. Nonparametric Statistics (2) (Goodness of fit tests)

Estimation of Quantiles

Copula Regression RAHUL A. PARSA DRAKE UNIVERSITY & STUART A. KLUGMAN SOCIETY OF ACTUARIES CASUALTY ACTUARIAL SOCIETY MAY 18,2011

SOCIETY OF ACTUARIES EXAM STAM SHORT-TERM ACTUARIAL MODELS EXAM STAM SAMPLE SOLUTIONS

EDUCATION AND EXAMINATION COMMITTEE OF THE SOCIETY OF ACTUARIES ACTUARIAL MODELS LIFE CONTINGENCIES SEGMENT POISSON PROCESSES

Multivariate distributions

IE 303 Discrete-Event Simulation

Basic concepts of probability theory

Sampling Distributions

Chapter 2. Continuous random variables

Massachusetts Institute of Technology

Integration by Substitution

Introduction to Probability and Statistics (Continued)

y = x 3 and y = 2x 2 x. 2x 2 x = x 3 x 3 2x 2 + x = 0 x(x 2 2x + 1) = 0 x(x 1) 2 = 0 x = 0 and x = (x 3 (2x 2 x)) dx

t x 1 e t dt, and simplify the answer when possible (for example, when r is a positive even number). In particular, confirm that EX 4 = 3.

Continuous random variables

Kernel density estimation for heavy-tailed distributions...

Math RE - Calculus II Antiderivatives and the Indefinite Integral Page 1 of 5

MATH2715: Statistical Methods

Partial Solutions for h4/2014s: Sampling Distributions

2. Conditional Expectation (9/15/12; cf. Ross)

Posted June 26, 2013 In the solution of Problem 4 in Practice Examination 8, the fourth item in the list of six events was mistyped as { X 1

Basic methods for simulation of random variables: 1. Inversion

DIFFERENTIAL EQUATIONS

Probability and Distributions

Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama

DIFFERENTIAL EQUATIONS

1. Point Estimators, Review

Chapter 3.3 Continuous distributions

On the number of ways of writing t as a product of factorials

Page Max. Possible Points Total 100

ARCONES MANUAL FOR THE SOA EXAM P/CAS EXAM 1, PROBABILITY, SPRING 2010 EDITION.

On the Lindley family distribution: quantile function, limiting distributions of sample minima and maxima and entropies

Exam P Review Sheet. for a > 0. ln(a) i=0 ari = a. (1 r) 2. (Note that the A i s form a partition)

Continuous Random Variables and Continuous Distributions

Transcription:

Method of Moments Definition. If {X 1,..., X n } is a sample from a population, then the empirical k-th moment of this sample is defined to be X k 1 + + Xk n n Example. For a sample {X 1, X, X 3 } the empirical second moment is X 1 +X +X 3 3. Example. The empirical first moment of a sample {X 1,..., X n } is X 1+ +X n n which we usually denote by X or sometimes by X n to emphasize that there are n observations. Assume a parametric distribution defined in terms of r parameters (θ 1,..., θ r ; for example the distribution Exponential(θ is define in term of only one parameter, while the Normal(µ, σ has two parameters. Suppose that we have observed n data points from the population under study. We want to use this data set to estimate the parameters of the model. Based on this sample we can calculate the empirical moments. Then to estimate the parameters of the model, we match the first r empirical moments with their theoretical counterparts: theoretical k-moment = (sample or empirical k-th moment, k=1,..., r So, here we have r equations in r unknowns (θ 1,..., θ r. By solving this system of equations we get the estimates for the parameters. Note. Some manuscripts use the notation E(M,..., E(M r to denote the sample moments. Example. Four losses are observed from a Gamma distribution. The observed losses are 00, 300, 350, and 450. Find the method of moments estimate for α. First Step: The Gamma distribution has two parameters α and θ. The theoretical 1

first moment is E(X = αθ and the theoretical second moment is E(X = α(α + 1θ. Now we calculate the empirical counterparts: Second Step: Calculate the sample moments: sample first moment = 00 + 300 + 350 + 450 4 = 35 sample second moment = 00 + 300 + 350 + 450 4 = 113, 750 Third Step: Form the equations: E(X = E(M αθ = 35 E(X = E(M α(α + 1θ = 113, 750 αθ = 35 (1 α(α + 1θ = 113, 750 ( = α(α + 1θ 113, 750 (αθ = (35 α + 1 = 1.0769 α = 13 α Example. You are given the following sample of five claims: 4, 5, 1, 99, 41 You fit a Pareto distribution using the method of moments. Determine the 95 th percentile of the fitted distribution. For the Pareto distribution we have E(X = other hand: E(M = 4+5+1+99+41 5 = 110 E(M = 4 +5 +1 +99 +41 5 = 37504.8 θ α 1 and E(X θ = (α 1(α. On the

Next Step: Form the equations: E(X = E(M θ α 1 = 110 E(X = E(M α(α + 1θ = 37504.8 θ α 1 = 110 (1 θ (α 1(α = 37504.8 ( θ / ( θ = 37504.8 (α 1(α α 1 110 = 3.0996 (α 1 α = 3.0996 cross multiplying α = 3.819 putting in (1 θ = 310.08 ( θ α ( 310.08 3.819.95 = F (x = 1 = 1 x = 369 x + θ x + 310.08 Example. For a sample of dental claims x 1, x,..., x 10, you are given: (i. x i = 3860 and x i = 4, 574, 80. (ii. Claims are assumed to follow a lognormal distribution with parameters µ and σ. (iii. µ and σ are estimated using the method of moments. Calculate E(X 500 for the fitted distribution. For the lognormal distribution we have E(X = exp(µ + 1 σ and E(X = exp(µ + σ. We equate these with their sample counterparts: exp(µ + 1 σ = 3860 10 = 386 exp(µ + σ = 4,574,80 10 = 457, 480. exp(µ + σ / ( exp(µ + 1 457, 480. σ = (386 = 3.07 exp(σ = 3.07 3

σ = ln(3.07 = 1.1 exp(µ + (1.1 = 457, 480. µ = 5.39 Now (from the table: ( [ ( ] E(X r = e µ+ 1 ln r µ σ ln r µ σ Φ + r 1 Φ σ σ ( [ ( ] ln 500 5.39 1.1 ln 500 5.39 E(X 500 = 386 Φ + 500 1 Φ 1.1 σ [ ] = 386 Φ( 0.9 + 500 1 Φ(0.77 = 59 Here is an example for mixture distribution: Example. You are given the following: The random variable X has the density function f(x = wf 1 (x + (l wf (x, 0 < x <, 0 w 1. A single observation of the random variable X yields the value 1. 0 xf 1 (xdx = 1 0 xf 1 (xdx = f (1 = f 1 (1 0 Determine the method of moments estimate of w. E(X = w E(X 1 + (1 we(x = w(1 + (1 w( = w But the single observation x 1 = 1, so the sample mean is also 1. By equating the theoretical mean with the sample mean, we get: 4

w = 1 w = 1 Here is an example for non-classical distributions: Example. You are given the following: The random variable X has the density function f(x = θ (θ x, 0 < x < θ. A random sample of two observations of X yields values 0.50 and 0.90. Determine θ, the method of moments estimator of θ. E(X = θ θ 0 x(θ xdx = θ θ (θ θ x dx x dx = θ ( θ 3 0 0 θ3 = θ 3 3 On the other hand, the sample first moment is: 0.5 + 0.9 = 0.7 Matching the two values gives us: θ = 0.7 θ =.1 3 Here is an example for dealing with discrete distributions: Example. We want to estimate the parameters β and r in the negative binomial distribution. The first and second empirical moments are 6 and 60. Find the method of moment estimate of P (N. 5

E(N = rβ E(N = Var(N + E(N = rβ(1 + β + (rβ rβ = 6 rβ(1 + β + (rβ = 60 By putting rβ = 6 in the second equation, we get 6(1 + β + 36 = 60 β = 3 r = P (N = 1 P (N = 0 P (N = 1 = 1 1 (1 + β r rβ (1 + β r+1 = 1 1 4 6 4 3 = 0.8438 Below is an example for dealing with negative moments: Example. We have observed the following claim sizes 10, 13, 16, 0, 3 We want to fit an inverse exponential model to this data. Calculate the method of moments estimate for the probability of claim being higher than 1. For this distribution only the negative moments exist. If θ is the parameter of this distribution, then we have E(X 1 = θ 1 On the other hand, the sample negative moment is: 1 10 + 1 13 + 1 16 + 1 0 + 1 3 5 Matching the two values gives us: = Harmonic average = 0.0666 6

E(X 1 = E(M 1 θ 1 = 0.0666 θ = 15.0195 Using the table, we will have: F (x = exp( θx P (X > 1 = 1 F (1 = 1 exp( 15.0195 = 0.7140 1 Below is an example dealing with censoring. But before getting to the example, note that in presence of right censoring the expected value of claim is actually the limited expected value E(X u where u is the policy limit. The moments for this distribution are E[(X u k ]. These moments will be used for the purpose of method of moments estimation. So, the model distribution and the sample distribution are both censored. Example. We have observed the following 10 values of claim sizes: 100, 100, 150, 170, 170, 00, 0, 300 +, 300 +, 300 + where the last three are censored observations. We want to fit the distribution Pareto(α =, θ to this data. What is the method of moments estimation for θ?. E(M 300 = 100 + 100 + 150 + 170 + 170 + 00 + 0 + 300 + 300 + 300 10 = 01 Using the table, we will have: E(X 300 = [ θ ( ] θ α 1 1 = 300θ α 1 300 + θ 300 + θ E(X 300 = E(M 300 300θ 300 + θ = 01 300θ = 01(300 + θ t = (01(300 99 = 609.1 7

Note. If we had α = 3, then we would need to solve the equation θ which would require a numerical method to solve it. [ ( ] 1 θ 300+θ = 01 Below is an example dealing with left truncation. For this type of situations, the model distribution and the sample distribution must be truncated. Note that in this process, an estimation for the underlying distribution is found, but not for the truncated one. Example. The following values of a random variable have been observed 0.55, 0.60, 0.73, 0.8, 0.95 We want to fit the distribution with density f(x = 1 + c + cx 0 x 1 where 0 c 1. What is the MME for c if the distribution is assumed to be truncated at x = E(M M > = 0.73 + 0.8 + 0.95 3 = 0.8333 E(X X > = x f(x dx = f(x dx x (1 + c + cx dx = (1 + c + cx dx (x + cx + cx dx (1 + c + cx dx = [ ( 1+c x + 3 cx3 ] 1 [(1 + cx + cx ] 1 = 0.3016 + 3.1349 c 0.37 + 0.9731 c E(X X > = E(M M > 0.3016 + 3.1349 c 0.37 + 0.9731 c = 0.8333 c = 0.003 Below is an example dealing with (complete grouped data. So in general assume that we have k intervals formed by the boundary points: c 0 < c 1 < < c k 8

and that n is the total number of observations, and that n j is the number of observations falling in j-th interval (c j 1, c j ]. So we have n = k j=1. For a grouped data no adjustment is needed for the model being fitted, but on the other hand since the exact values of the sample points are not known, we need to approximate the sample moments. If in each interval (c j 1, c j ] no point has an advantage over another point, then we can assume that the the sample points have uniform distribution over this interval. Then the average value in each interval is just the midpoint of the interval. Then the weighted average k j=1 is taken for the sample mean x. c j 1 + c j n j n = 1 n k j=1 c j 1 + c j n j Example. Consider the following grouped data we had sometime ago.. Interval Number of observations (0, ] 5 (, 10] 10 (10, 100] 10 (100, 1000] 5 We want to fit an Exponential(θ to this data. Find the MME for θ. E(M = 1 n = 1 50 k j=1 [( 0 + c j 1 + c j (5 + n j ( + 10 E(X = E(M θ = 66.58 (10 + ( 10 + 100 (10 + ( 100 + 1000 ] (5 = 67.7 9

Finally we see here in an example how to calculate the limited expected value E(X u for a grouped data. Suppose you are given the following grouped data: Interval Number of observations (0, ] 5 (, 10] 10 (10, 100] 10 at 100 45 This data is right-censored at 100. Then the limited expected value E(X 100 is calculated through: E(X 100 = 1 90 [( 0 + (5 + ( + 10 (10 + ( 10 + 100 ] (10 + (100(45 = 57.06 10