Exam C Solutions Spring 2005

Similar documents
SPRING 2007 EXAM C SOLUTIONS

Course 4 Solutions November 2001 Exams

Practice Exam 1. (A) (B) (C) (D) (E) You are given the following data on loss sizes:

SOCIETY OF ACTUARIES EXAM STAM SHORT-TERM ACTUARIAL MODELS EXAM STAM SAMPLE SOLUTIONS

Method of Moments. which we usually denote by X or sometimes by X n to emphasize that there are n observations.

Parameter Estimation

Solutions to the Spring 2015 CAS Exam ST

Suggested solutions to written exam Jan 17, 2012

This does not cover everything on the final. Look at the posted practice problems for other topics.

Practice Problems Section Problems

Distribution Fitting (Censored Data)

ACTEX CAS EXAM 3 STUDY GUIDE FOR MATHEMATICAL STATISTICS

4 Testing Hypotheses. 4.1 Tests in the regression setting. 4.2 Non-parametric testing of survival between groups

Spring 2012 Math 541B Exam 1

f (1 0.5)/n Z =

Probability Distributions Columns (a) through (d)

MAY 2005 SOA EXAM C/CAS 4 SOLUTIONS

Solutions to the Spring 2018 CAS Exam MAS-1

Mixture distributions in Exams MLC/3L and C/4

Eco517 Fall 2004 C. Sims MIDTERM EXAM

STAT 6350 Analysis of Lifetime Data. Probability Plotting

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY

Chapter 1 Statistical Inference

Math 562 Homework 1 August 29, 2006 Dr. Ron Sahoo

GEOMETRIC -discrete A discrete random variable R counts number of times needed before an event occurs

TABLE OF CONTENTS CHAPTER 1 COMBINATORIAL PROBABILITY 1

ASM Study Manual for Exam P, Second Edition By Dr. Krzysztof M. Ostaszewski, FSA, CFA, MAAA Errata

Name: Partners: PreCalculus. Review 5 Version A

Solutions to the Fall 2018 CAS Exam MAS-1

Three hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER.

Exam C. Exam C. Exam C. Exam C. Exam C

Review. December 4 th, Review

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

November 2002 Course 4 solutions

ECON 4130 Supplementary Exercises 1-4

Announcements. Topics: Homework: - sections , 6.1 (extreme values) * Read these sections and study solved examples in your textbook!

11 Survival Analysis and Empirical Likelihood

Review for the previous lecture

Risk Models and Their Estimation

Introduction to Statistical Analysis

Modelling the risk process

Practice Exam #1 CAS Exam 3L

Exercises and Answers to Chapter 1

Algebra II. A2.1.1 Recognize and graph various types of functions, including polynomial, rational, and algebraic functions.

MS 2001: Test 1 B Solutions

Test Problems for Probability Theory ,

Errata and updates for ASM Exam C/Exam 4 Manual (Sixth Edition) sorted by date

Statistics 100A Homework 5 Solutions

Errata and updates for ASM Exam C/Exam 4 Manual (Sixth Edition) sorted by page

Errata for the ASM Study Manual for Exam P, Fourth Edition By Dr. Krzysztof M. Ostaszewski, FSA, CFA, MAAA

Preface Introduction to Statistics and Data Analysis Overview: Statistical Inference, Samples, Populations, and Experimental Design The Role of

Toward a unified approach to fitting loss models

Statistics 135 Fall 2007 Midterm Exam

ASM Study Manual for Exam P, First Edition By Dr. Krzysztof M. Ostaszewski, FSA, CFA, MAAA Errata

Severity Models - Special Families of Distributions

The Chi-Square Distributions

ECONOMICS 207 SPRING 2006 LABORATORY EXERCISE 5 KEY. 8 = 10(5x 2) = 9(3x + 8), x 50x 20 = 27x x = 92 x = 4. 8x 2 22x + 15 = 0 (2x 3)(4x 5) = 0

Consequences of Continuity and Differentiability

Lectures 5 & 6: Hypothesis Testing

Continuous RVs. 1. Suppose a random variable X has the following probability density function: π, zero otherwise. f ( x ) = sin x, 0 < x < 2

Estimation for Modified Data

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45

MFM Practitioner Module: Quantitative Risk Management. John Dodson. September 23, 2015

Final Examination a. STA 532: Statistical Inference. Wednesday, 2015 Apr 29, 7:00 10:00pm. Thisisaclosed bookexam books&phonesonthefloor.

INSTITUTE AND FACULTY OF ACTUARIES. Curriculum 2019 SPECIMEN SOLUTIONS

Objective Experiments Glossary of Statistical Terms

Stat 5102 Final Exam May 14, 2015

CSE 312 Final Review: Section AA

Basic Concepts of Inference

Department of Statistical Science FIRST YEAR EXAM - SPRING 2017

TMA 4275 Lifetime Analysis June 2004 Solution


Nonparametric Model Construction

Identifying the Graphs of Polynomial Functions

Lecture 32: Taylor Series and McLaurin series We saw last day that some functions are equal to a power series on part of their domain.

Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama

SOLUTION FOR HOMEWORK 8, STAT 4372

DRAFT - Math 101 Lecture Note - Dr. Said Algarni

Simulation. Alberto Ceselli MSc in Computer Science Univ. of Milan. Part 4 - Statistical Analysis of Simulated Data

Midterm Examination. STA 215: Statistical Inference. Due Wednesday, 2006 Mar 8, 1:15 pm

A Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.

MVE055/MSG Lecture 8

Randomized Algorithms

Math 106: Calculus I, Spring 2018: Midterm Exam II Monday, April Give your name, TA and section number:

Two-stage Adaptive Randomization for Delayed Response in Clinical Trials

A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints

Final Exam A Name. 20 i C) Solve the equation by factoring. 4) x2 = x + 30 A) {-5, 6} B) {5, 6} C) {1, 30} D) {-5, -6} -9 ± i 3 14

Introduction to Probability and Statistics (Continued)

A Very Brief Summary of Statistical Inference, and Examples

375 PU M Sc Statistics

M(t) = 1 t. (1 t), 6 M (0) = 20 P (95. X i 110) i=1

Review of Probabilities and Basic Statistics

A Very Brief Summary of Statistical Inference, and Examples

Machine Learning CSE546 Carlos Guestrin University of Washington. September 30, What about continuous variables?

BTRY 4090: Spring 2009 Theory of Statistics

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

1 Introduction. P (n = 1 red ball drawn) =

Question Points Score Total: 76

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

Model Fitting. Jean Yves Le Boudec

Transcription:

Exam C Solutions Spring 005 Question # The CDF is F( x) = 4 ( + x) Observation (x) F(x) compare to: Maximum difference 0. 0.58 0, 0. 0.58 0.7 0.880 0., 0.4 0.680 0.9 0.93 0.4, 0.6 0.53. 0.949 0.6, 0.8 0.349.3 0.964 0.8,.0 0.64 The K-S statistic is the maximum from the last column, 0.680. Critical values are: 0.546, 0.608, 0.66, and 0.79 for the given levels of significance. The test statistic is between 0.66 (.5%) and 0.79 (.0%) and therefore the test is rejected at 0.05 and not at 0.0. Question # Key: E For claim severity, μs = (0.4) + 0(0.4) + 00(0.) = 4.4, σ S = (0.4) + 0 (0.4) + 00 (0.) 4.4 =,445.04. For claim frequency, μf = rβ = 3, r σf = rβ( + β) =. r For aggregate losses, μ = μsμf = 4.4(3 r) = 73. r, σ = μsσf + σsμf = 4.4 ( r) +,445.04(3 r) =,479.44 r. For the given probability and tolerance, λ 0 = (.96 / 0.) = 384.6. The number of observations needed is λσ 0 / μ = 384.6(, 479.44 r) /(73. r) = 83.0 / r. The average observation produces 3r claims and so the required number of claims is (83.0 / r)(3 r ) =, 469.

Question #3 ˆ n 39 H ( t ) 39 = + = = => n 799n + 380 = 0 => n = 0, n = n n n( n ) 380 Discard the non-integer solution to have n = 0. The Kaplan-Meier Product-Limit Estimate is: 9 8 St ˆ( 9 ) =... = = 0.55. 0 9 0 0.487. Question #4 Key: E There are 7 possible bootstrap samples, which produce four different results. The results, their probabilities, and the values of g are: Bootstrap Sample Prob g,, 8/7 0,, 4 /7, 4, 4 6/7-4, 4, 4 /7 0 The third central moment of the original sample is. Then, 8 6 44 + + + = 7 7 7 7. 9 MSE = ( 0 ) ( ) ( ) ( 0 ) Question #5 Pick one of the points, say the fifth one. The vertical coordinate is F(30) from the model and should be slightly less than 0.6. Inserting 30 into the five answers produces 0.573, 0.096, 0.93, 0.950, and something less than 0.5. Only the model in answer A is close.

Question #6 or E This question was considered ambiguous, hence two solutions were accepted. Had the question begun with Claim amounts are then the correct interpretation is that there are five observations (whose values are not stated), each from the Poisson distribution. In that case, the solution proceeds as follows: The distribution of Θ is Pareto with parameters and.6. Then, v = EVPV = E( Θ ) = = 0.65, a = VHM = Var( Θ ) = 0.65 =.697,.6.6(0.6) 5 k = v/ a= 0.65/.697 = 0.369, Z = = 0.93. 5 + 0.369 Alternatively, had the question begun with The number of claims is then the correct interpretation is that there was a single observation whose value was 5. The only change is that now n = rather than 5 and so Z = = 0.7304. + 0.369 Question #7 Key: E At 300, there are 400 policies available, of which 350 survive to 500. At 500 the risk set increases to 875, of which 750 survive to,000. Of the 750 at,000, 450 survive to 5,000. The probability of surviving to 5,000 is 350 750 450 = 0.45. The distribution function is 0.45 = 400 875 750 0.55. Alternatively, the formulas in Loss Models could be applied as follows: j Interval ( cj, c j + ] d j u j x j P j r j q j Fc ˆ ( ) 0 (300, 500] 400 0 50 0 400 0.5 0 (500,,000] 55 0 5 350 875 0.4857 0.5 (,000, 5,000] 0 0 300 750 750 0.4 0.5 3 (5,000, 0,000] 0 30 300 330 330 0.90909 0.55 where d j = number of observations with a lower truncation point of c j, u j = number of observations censored from above at c j+, x j = number of uncensored observations in the interval ( cj, c j + ]. Then, P j P j d + j u j x j rj = Pj + d j (using = and β = 0 ), and qj = xj / rj. Finally, Fc ˆ ( ) = ( q) j j i. i= 0 j

Question #8 3 Let the function be f ( x) = a + bx + cx + dx. The four conditions imply the following: f(0) = 0 a= 0, f() = 0 b+ 4c+ 8d = 0, f '(0) = b=, f '() = b+ 4c+ d =. The solution is c = and d = 0 and thus the function is 0 0 [ f "( x)] dx = dx = 8. f ( x) x x =. The curvature measure is Question #9 For an exponential distribution, the maximum likelihood estimate of is the sample mean, 6. Let Y = X+ X where each X has an exponential distribution with mean 6. The sample mean is Y/ and Y has a gamma distribution with parameters and 6. Then x /6 xe Pr( Y / > 0) = Pr( Y > 0) = dx 0 36 x /6 0/6 xe x /6 0e 0/6 = e = + e = 0.546. 6 6 0 Question #0 0/ 0e 0/ 0/ From Question 9, F(0) = e = e ( + 0 ) = g( ). 0/ 0 0/ 0/ 0 400e g'( ) = e ( + 0 ) + e =. 3 At the maximum likelihood estimate of 6, g '(6) = 0.066063. The maximum likelihood estimator is the sample mean. Its variance is the variance of one observation divided by the sample size. For the exponential distribution the variance is the square of the mean, so the estimated variance of the sample mean is 36/ = 8. The answer is ( 0.066063) (8) = 0.079.

Question # Key B μ( λ, ) = ES ( λ, ) = λ, v( λ, ) = Var( S λ, ) = λ, v = EVPV = E( λ ) = ()(+ ) = 4, a= VHM = Var = E E E E = = k = v/ a= 4/3. ( λ ) ( λ ) ( ) [ ( λ) ( )] () 3, Question # Key: B The distribution is binomial with m = 00 and q = 0.03. The first three probabilities are: 00 99 p0 = 0.97 = 0.04755, p = 00(0.97) (0.03) = 0.4707, 00(99) 98 p = (0.97) (0.03) = 0.55. Values between 0 and 0.04755 simulate a 0, between 0.04755 and 0.946 simulate a, and between 0.946 and 0.4977 simulate a. The three simulated values are, 0, and. The mean is. Question #3 A mixture of two Poissons or negative binomials will always have a variance greater than its mean. A mixture of two binomials could have a variance equal to its mean, because a single binomial has a variance less than its mean.

Question #4 The posterior distribution can be found from λ 0 e λ 0.4 λ/6 0.6 λ/ 0 7 λ/6 3 λ/ πλ ( 0) e + e λ ( 0.8e + 0.6 e ). 0! 6 The required constant is found from 0 7 / 6 3 / ( 0.8e λ λ λ + 0.6e ) dλ = 0.8(0!)(6 / 7) + 0.6(0!)( /3) = 0.395536(0!). 0 The posterior mean is 7 / 6 3 / E( 0) ( 0.8e λ λ λ = 0.6 ) 0.395536(0!) λ + e dλ 0 0.8(!)(6 / 7) + 0.6(!)( /3) = = 9.88. 0.395536(0!) Question #5 H ˆ (4.5) = 0.77460. + 0 + 9 + 7 = The variance estimate is + + + = 0.089397. 0 9 7 The confidence interval is 0.77460 ±.96 0.089397 = 0.77460 ± 0.58603. The interval is from 0.886 to.3606. Question #6 ˆ k bias = E( ) = =, k+ k+ ( ˆ k k Var ) = Var =, k + 5( k + ) ˆ k MSE = Var( ) + bias = +, 5( k+ ) ( k+ ) MSE = bias =. ( k + ) Setting the last two equal and canceling the common terms gives k + = for k = 5. 5

Question #7 For the geometric distribution μβ ( ) = βand v( β) = β( + β). The prior density is Pareto with parameters and. Then, μ = E( β) =, v = EVPV = E[ β( + β)] = + =, ( )( ) ( )( ) a = VHM = Var( β ) = =, ( )( ) ( ) ( ) ( ) k = v/ a=, Z = =. + k The estimate is x + x + =. Question #8 z z The hazard rate function is hx ( ) h e β + = = β, a constant. For the four observations, those.(0) +.0(0).(0) +.0(30) e = 0.06786, e = 0.07449, 8 8 constants are.() +.0(30).() +.0(40) e = 0.0888, e = 0.0960. 8 8 hx Each observation is from an exponential distribution with density function he and survival hx function e. The contributions to the loglikelihood function are ln( h) hx and hx, respectively. The answer is, ln(0.06786) 0.06786(3) 0.07499(6) + ln(0.0888) 0.0888(7) 0.0960(8) = 7.47. Question #9 Key: E A is false. Using sample data gives a better than expected fit and therefore a test statistic that favors the null hypothesis, thus increasing the Type II error probability. The K-S test works only on individual data and so B is false. The A-D test emphasizes the tails, thus C is false. D is false because the critical value depends on the degrees of freedom which in turn depends on the number of cells, not the sample size.

Question #0 Key: B E( ) = 0.05(0.8) + 0.3(0.) = 0., E( ) = 0.05 (0.8) + 0.3 (0.) = 0.0, μ ( ) = 0( ) + ( ) + ( 3 ) = 5, v( ) = 0 ( ) + ( ) + ( 3 ) ( 5 ) = 9 5, μ = E( 5 ) = 5(0.) =.5, v = EVPV = E(9 5 ) = 9(0.) 5(0.0) = 0.4, a= VHM = Var( 5 ) = 5 Var( ) = 5(0.0 0. ) = 0.5, 5 k = v/ a= 0.4 / 0.5 =.6, Z = =, +.6 3 5 8 P = +.5 =.693. 3 3 Question # Key: B λ λ 3 5 5 λ e λ e λ λ e 4λ f ( λ 5,3) λ e. This is a gamma distribution with parameters 3 5! 3! 4λ and 0.5. The expected value is 3(0.5) = 3.5. Alternatively, if the Poisson-gamma relationships are known, begin with the prior parameters = 5 and β = where β = / if the parameterization from Loss Models is considered. Then the posterior parameters are ' = 5+ 5+ 3= 3 where the second 5 and the 3 are the observations and β ' = + = 4 where the second is the number of observations. The posterior mean is then 3/4 = 3.5. Question # Because the kernel extends one unit each direction there is no overlap. The result will be three replications of the kernel. The middle one will be twice has high due to having two observations of 3 while only one observation at and 5. Only graphs A and D fit this description. The kernel function is smooth, which rules out graph A.

Question #3 A is false, the collocation polynomial is too wiggly. Both collocation polynomials and cubic splines pass through all points, so B is false. C is false because the curvature adjusted spline sets the values, but not necessarily to zero. The inequality in E should be the other direction. Answer D is true. Question #4 The two moment equations are 508 =, 70, 40.6 =. Dividing the square of the first equation into the ( )( ) 70, 40.6 ( ) second equation gives =.779366 =. The solution is = 4.78576. From 508 the first equation, =,93.67. The requested LEV is 3.78576,93.67,93.67 EX ( 500) = 96.. 3.78576 =,93.67 + 500 Question #5 Key: B 50(00 7.7) + 60(50 7.7) + 00(60 78.95) + 90(00 78.95) vˆ = EVPV = + = 7,985.647, kˆ = 7,985.647 / 65.03 = 0.57, ˆ 0 Z = = 0.499. 0 + 0.57 Question #6 Key: B F00(, 000) = 0.6, F00(3, 000) = 0.38, F00(5, 000) = 0.63, F00(0, 000) = 0.8, F00(, 000) = 0.5(0.6) + 0.5(0.38) = 0.7, F00(6, 000) = 0.8(0.63) + 0.(0.8) = 0.666. Pr(, 000 < X < 6, 000) = 0.666 0.7 = 0.396.

Question #7 3 4 f(750) 3 4 6 f(400) L= f(00) f(300) [ F(0, 000)] F(00) F(300) 3 3 4 6 4 = 0,00 0,000 0,000 0,000 0,300 + + + + 0,750 0,00 0,300 0,000 0, 400 4 = 0,00 0,000 0,300 0,750 0,000 0, 400 3 3 4 3 3 6 4 4 4 3 3 6 4 0,000 0,750 0,000 0,400, ln L = 4ln + 3ln(0,000) 3ln(0,750) 6ln(0,000) 4ln(0,400) = 4ln 4.537. The derivative is 4 / 4.537 and setting it equal to zero gives ˆ = 3.089. Question #8 30 + 30 + + 4 vˆ = EVPV = x = = 0.76. 00 50(0 0.76) + 30( 0.76) + 5( 0.76) + 4(3 0.76) + (4 0.76) aˆ = VHM = 0.76 99 = 0.090909, ˆ 0.76 ˆ k = = 8.36, Z = = 0.0684, 0.090909 + 8.36 P = 0.0684() + 0.8936(0.76) = 0.78564. The above analysis was based on the distribution of total claims for two years. Thus 0.78564 is the expected number of claims for the next two years. For the next one year the expected number is 0.78564/ = 0.398.

Question #9 For members of Class A, x / x / A A β x γ / x γ / γ B B γ hx ( ) = x/, S( x) = e, f( x) = ( x/ ) e and for members of Class B, hx ( ) = xe / = x /, S( x) = e, f( x) = ( x / ) e where γ = e β. The likelihood function is L( γ, ) = fa() fa(3) fb() fb(4) / 9/ 4 γ / 6 γ / = e 6 e 4γ e 8γ e 4 (0+ 0 γ)/ γ e. The logarithm and its derivatives are l( γ, ) = 4ln + ln γ (0 + 0 γ) l / γ = γ + 0 l / = 4 + (0+ 0 γ). Setting each partial derivative equal to zero and solving the first equation for gives = 0γ. Substituting in the second equation yields 4 0+ 0γ + = 0, 400γ = 00γ + 00 γ, γ = 0.5. Then, β = ln( γ) = ln(0.5) = 0.69. 0γ 00γ Question #30 Let the two spline functions be 3 f0( x) = a+ b( x+ ) + c( x+ ) + d( x+ ), 3 f( x) = e+ fx+ gx + hx. The eight spline conditions produce the following eight equations. f0( ) = 3 a= 3, f(0) = 0 e= 0, '' f0( ) = 0 c= 0, f0(0) = 0 3+ b+ 4c+ 8d = 0, ' ' f() = 3 f + 4g+ 8h= 3, f0(0) = f(0) b+ 4c+ d = f, '' '' '' f0(0) = f (0) c+ d = g, f () = 0 g+ h= 0. Eliminating one variable at a time leads to b = 6, d = 0, f = 6, g = 0, and h = 0. Then, f ( x ) = 3 + 6( x + ) = 6 x and the second derivative is zero. 0 The above work could be avoided by noting that the three knots are (-, -3), (0,0), and (, 3). These points lie on a straight line and so the spline will be that straight line.

Question #3 Key: E 0.8 0.x 0. ( x / ) The density function is f( x) = e. The likelihood function is 0. L( ) = f(30) f(40) f(300) f(540)[ F(000)] 0.8 0.8 0.8 0.8 0.(30) 0.(40) 0.(300) 0.(540) = e e e e 0. 0. 0. 0. e e 0. 0. 0. 0. 0. 0. 0. 0.8 (30 + 40 + 300 + 540 + 000 + 000 ) e, 0. 0. 0. 0. 0. 0. 0. l( ) = 0.8ln( ) (30 + 40 + 300 + 540 + 000 + 000 ) 0. = 0.8ln( ) 0.505,. l '( ) = 0.8 + 0.(0.505) = 0, 0. = 0.9756, ˆ = 3,35.67. 0. 0. 0. 0. 0. 0. (30/ ) (40/ ) (300/ ) (540/ ) (000/ ) (000/ ) Question #3 Buhlmann estimates are on a straight line, which eliminates E. Bayes estimates are never outside the range of the prior distribution. Because graphs B and D include values outside the range to 4, they cannot be correct. The Buhlmann estimates are the linear least squares approximation to the Bayes estimates. In graph C the Bayes estimates are consistently higher and so the Buhlmann estimates are not the best approximation. This leaves A as the only feasible choice. Question #33 The expected counts are 300(0.035) = 0.5, 300(0.095) = 8.5, 300(0.5) = 50, 300(0.) = 60, and 300(0.7) = 5 for the five groups. The test statistic is (5 0.5) (4 8.5) (37 50) (66 60) (50 5) + + + + =.0. 0.5 8.5 50 60 5 There are 5 = 4 degrees of freedom. From the table, the critical value for a 5% test is 9.488 and for a.5% test is.43. The hypothesis is rejected at 5%, but not at.5%.

Question #34 To simulate a lognormal variate, first convert the uniform random number to a standard normal. This can be done be using the table of normal distribution values. For the six values given, the corresponding standard normal values are 0.3, -0.,.6, -.4, 0.8, and -0.. Next, multiply each number by the standard deviation of 0.75 and add the mean of 5.6. This produces random observations from the normal 5.6, 0.75 distribution. These values are 5.85, 5,55, 6.8, 4.55, 6., and 5.45. To create lognormal observations, exponentiate these values. The results are 339, 5, 898, 95, 493, and 33. After imposing the policy limit of 400, the mean is (339 + 5 + 400 + 95 + 400 + 33)/6 = 86. Question #35 β For the geometric distribution, Pr( X = β ) = and the expected value is β. 3 ( + β ) Pr( X = β = )Pr( β = ) Pr( β = X = ) = Pr( X = β = )Pr( β = ) + Pr( X = β = 5)Pr( β = 5) 4 = 7 3 = 0.3904. 4 5 + 7 3 6 3 The expected value is then 0.3904() + 0.60976(5) = 3.83.