Chain Ladder Method: Bayesian Bootstrap versus Classical Bootstrap

Size: px
Start display at page:

Download "Chain Ladder Method: Bayesian Bootstrap versus Classical Bootstrap"

Transcription

1 Chan Ladder Method: Bayesan Bootstrap versus Classcal Bootstrap Gareth W. Peters 1,2 Maro V. Wüthrch 3 Pavel V. Shevchenko 2 1 UNSW Mathematcs and Statstcs Department, Sydney, 2052, Australa; emal: peterga@maths.unsw.edu.au 2 CSIRO Sydney, Locked Bag 17, North Ryde, NSW, 1670, Australa 3 ETH Zurch, Department of Mathematcs, CH-8092 Zurch, Swtzerland Preprnt submtted to Elsever February 28, 2009

2 Chan Ladder Method: Bayesan Bootstrap versus Classcal Bootstrap Gareth W. Peters 1,2 Maro V. Wüthrch 3 Pavel V. Shevchenko 2 Abstract The ntenton of ths paper s to analyse the mean square error of predcton MSEP) under the dstrbuton-free chan ladder DFCL) clams reservng method. We compare the estmaton obtaned from the classcal bootstrap method wth the one obtaned from a Bayesan bootstrap. To acheve ths n the DFCL model we develop a novel approxmate Bayesan computaton ABC) samplng algorthm to obtan the emprcal posteror dstrbuton. We need an ABC samplng algorthm because we work n a dstrbuton-free settng. The use of ths ABC methodology combned wth bootstrap allows us to obtan samples from the ntractable posteror dstrbuton wthout the requrement of any dstrbutonal assumptons. Ths then enables us to calculate the MSEP and other rsk measures lke Value-at-Rsk. Key words: Clams reservng, dstrbuton-free chan ladder, mean square error of predcton, Bayesan chan ladder, approxmate Bayesan computaton, Markov chan Monte Carlo, adapton, annealng, bootstrap Preprnt submtted to Elsever February 28, 2009

3 1. Motvaton The dstrbuton-free chan ladder model DFCL) of Mack [13] s probably the most popular model for stochastc clams reservng. In ths paper we use a tme seres formulaton of the DFCL model whch allows for bootstrappng the clams reserves. An mportant aspect of ths model s that t can provde a ustfcaton for the classcal determnstc chan ladder CL) algorthm whch s not founded on an underlyng stochastc model. Moreover, t allows for the study of predcton uncertantes. Note that there are dfferent stochastc models that lead to the CL reserves see for example Wüthrch-Merz [22], Secton 3.2). In the present paper we use the DFCL formulaton to reproduce the CL reserves. Ths paper analyses the parameter estmates n the DFCL model, the assocated clams reserves and the mean square errors of predcton MSEP) from both the frequentst perspectve and a contrastng Bayesan vew. In dong so we analyse CL pont estmators for parameters of the DFCL model, the resultng estmated reserves and the assocated MSEP from the classcal perspectve. These nclude bootstrap estmated predcton errors obtaned va a process of condtonal back propagaton. These classcal frequentst estmators are compared to Bayesan pont estmators. The Bayesan estmates consdered are the maxmum a posteror MAP) and the mnmum mean square error MMSE) estmators and the assocated Bayesan estmated reserves condtonal on such pont estmators. In addton, snce n the Bayesan settng we obtan samples from the posteror on the parameters we use these along wth the MSEP obtaned by the pont estmators to obtan assocated posteror predctve ntervals to be compared wth the classcal bootstrap procedures. We then robustfy the predcton of reserves and analyse the resultant MSEP when we ntegrate out the nfluence of the predcton of unknown varance parameters n the DFCL model, agan achevable snce n the Bayesan settng we obtan samples from the ont posteror for the CL factors and the varances. Ths requres that we develop a Bayesan CL model for the DFCL model whch makes no parametrc assumptons on the form of the lkelhood functon, see also Gsler-Wüthrch [11]. Ths s unlke the works of Yao [23] and Peters-Shevchenko-Wüthrch [19] that assume explct dstrbutons n order to construct the posteror dstrbutons n the Bayesan context. Instead we demonstrate how to work drectly wth the ntractable lkelhood functons and the resultng ntractable posteror dstrbuton, usng approprate dstance functons. In ths regard we demonstrate that we do not need to make any parametrc assumptons to perform posteror nference, avodng the often poor model assumptons made, as for example n the paper of Yao 3

4 [23]. To acheve ths we develop a novel approxmate Bayesan computaton ABC) samplng algorthm to obtan samples from the posteror. We develop the frst use of ABC methodology combned wth bootstrap whch allows us to obtan samples from the ntractable posteror dstrbuton wthout the requrement of any dstrbutonal assumptons or even the requrement on generatng of observatons from the ntractable lkelhood model. Instead, we smply requre that certan model assumptons are not volated n our condtonal back propagaton bootstrap procedure. In ths context we analyse several aspects of the ABC methodology, ncludng senstvty studes and convergence dagnostcs, before we make use of the algorthm to perform a detaled comparson wth classcal models. Outlne of ths paper. The paper begns wth a presentaton of the clams reservng problem and then presents the model we shall consder. Ths s followed by the descrpton of the classcal chan ladder algorthm and the constructon of a Bayesan model that can be used to estmate the parameters of the model. The Bayesan model s constructed n a dstrbuton-free settng. Followng ths s a dscusson on classcal versus Bayesan parameter estmators along wth a bootstrap based procedure for estmaton of parameter uncertanty n the classcal settng. The next secton presents the methodology of ABC coupled wth a novel bootstrap based samplng procedure whch wll allow us to work drectly wth the dstrbuton-free Bayesan model. We then llustrate the developed algorthm on a synthetc data set and the real data set, comparng performance to the classcal results and those obtaned va credblty theory. 2. Clams development trangle and DFCL model We brefly outlne the clams development trangle structure we utlse n formulaton of our models. Assume there s a run-off trangle contanng clams development data wth the structure gven n Table 1. Assume that C, are cumulatve clams wth ndces {0,...,I} and {0,...,J}, where denotes the accdent year and denotes the development year cumulatve clams can refer to payments, clams ncurred, etc). We make the smplfyng assumpton that the number of accdent years s equal to the number of observed development perods.e. I = J. At tme I, we have observatons D I = {C, ; + I}, 2.1) 4

5 accdent development years year I 0 1 observed random varables C, D I.. to be predcted C, D c I I 1 I Table 1: Clams development trangles. and for clams reservng at tme I we need to predct the future clams DI c = {C, ; + > I, I, J}. 2.2) Moreover, we defne for {0,...,I} the set B = {C,k ; + k I,0 k }, e.g. B 0 s the frst column n Table Classcal chan ladder algorthm In the classcal determnstc) chan ladder algorthm there s no underlyng stochastc model. It s rather a recursve algorthm that was used to estmate the clams reserves and whch has proved to gve good practcal results. It smply nvolves the followng recursve steps to predct unobserved cumulatve clams n D c I. Set Ĉ,I = C,I and for > I Ĉ, = Ĉ, 1 f CL) 1 wth CL factor estmates fcl) 1 = I =0 C, I =0 C. 2.3), 1 Snce ths s a determnstc algorthm t does not allow for quantfcaton of the uncertanty assocated wth the predcted reserves. To analyse the assocated uncertanty there are several stochastc models that reproduce the CL reserves: for example Mack s dstrbuton-free chan ladder model [13], the over-dspersed Posson model see England-Verrall [5]) or the Bayesan chan ladder model see Gsler-Wüthrch [11]). We use a tme seres formulaton of the Bayesan chan ladder model n order to use bootstrap methods and Bayesan nference Bayesan DFCL model We use an addtve tme seres verson of the Bayes chan ladder model Model Assumptons 3.1, n Gsler-Wüthrch [11]). 5

6 Model Assumptons We defne the CL factors by F = F 0,...,F ) and the standard devaton parameters by Ξ = Ξ 0,...,Ξ ). We assume ndependence between all these parameters,.e., the pror densty of F,Ξ) s gven by πf, σ) = =0 πf ) πσ ), 2.4) where πf ) denotes the densty of F and πσ ) denotes the densty of Ξ. 2. Condtonally, gven F = f = f 0,...,f ) and Ξ = σ = σ 0,...,σ ), we have: Cumulatve clams C, n dfferent accdent years are ndependent. Cumulatve clams satsfy the followng tme seres representaton C,+1 = f C, + σ C, ε,+1, 2.5) where condtonally, gven B 0, we have that the resduals ε, are..d. satsfyng E [ε, B 0, F,Ξ] = 0 and Var [ε, B 0, F,Ξ] = 1, 2.6) and P [C, > 0 B 0, F,Ξ] = 1 for all,. Remark. Note that the assumptons on the resduals are slghtly nvolved n order to guarantee that cumulatve clams C, are postve P-a.s. Corollary 2.2. Under Model Assumptons 2.1 we have that condtonally, gven D I, the random varables F 0,...,F, Ξ 0,...,Ξ are ndependent. Proof of Corollary 2.2. The proof s completely analogous to the proof of Theorem 3.2 n Gsler-Wüthrch [11] and follows from pror ndependence of the parameters and the fact that C,+1 only depends on F, Ξ and C, Markov property). In partcular, Corollary 2.2 says that we obtan the followng posteror dstrbuton for F, Ξ), gven D I, π f, σ D I ) = =0 π f D I )π σ D I ). 2.7) Ths has mportant mplcatons for the ABC samplng algorthm developed below. In order to perform the Bayesan analyss we make explct assumptons on the pror dstrbutons of F,Ξ). 6

7 Model Assumptons 2.3. In addton to Model Assumptons 2.1 we assume that the pror model for all parameters {0,...,J 1} s gven by: F Γ α, β ), where Γ α, β ) s a gamma dstrbuton wth mean E [F ] = α β = see 2.3)) and large varance to have dffuse prors. f CL) The varances Ξ 2 IGa, b ), where IGa, b ) s an nverse gamma dstrbuton wth [ ] mean E Ξ 2 = b /a 1) = σ 2CL) see 2.8), below) and large varance. Remarks. 1) The lkelhood model s ntractable, meanng that no densty can be wrtten down analytcally n the DFCL model. We have only made dstrbutonal assumptons on the parameters F,Ξ) but not on the observable cumulatve clams C,. Therefore, a full Bayesan analyss usng analytc posteror dstrbutons cannot be performed. One way out of ths dlemma would be to formulate a full Bayesan model by makng dstrbutonal assumptons ths s, e.g., done n Yao [23]) but then the model s no longer dstrbuton-free. Another approach would be to use credblty methods see Gsler-Wüthrch [11]) but ths only gves statements for second moments. In the present set up we develop ABC methods that allow for a full dstrbutonal answer for the posteror wthout makng explct dstrbutonal assumptons for the cumulatve clams C,. 2) We select the prors to ensure that we mantan several relevant aspects of the DFCL model. Frstly, t s mportant to utlse prors that enforce the strct postvty of the parameters f, σ > 0. We note here that the parametrc Bayesan model developed n Yao [23] faled n ths aspect and therefore we develop an alternatve pror structure that satsfes these requred propertes of the DFCL model. Secondly, our prors are chosen such that they do not contan nformaton,.e. we assume dffuse prors wth large varances Classcal and Bayesan parameter estmators In the classcal CL method, the CL factors are estmated by parameters are estmated by, see e.g. 3.4) Wüthrch-Merz [22], f CL) gven n 2.3). The varance σ 2CL) = 1 I 1 I 1 =0 C, C,+1 C, ) 2 CL) f. 2.8) Note that ths estmator s only well-defned for < I 1. There s a vast lterature and dscusson on the estmaton of tal parameters. We do not enter ths dscusson here but we 7

8 smply choose the estmator gven n Mack [13] for the last varance parameter whch s defned by σ 2CL) = mn { σ 4CL) J 2 σ 2CL) J 3, σ 2CL) J 3, σ2cl) J 2 }. 2.9) In a Bayesan nference context one calculates the posteror dstrbuton of the parameters, gven D I. As n 2.7) we denote ths posteror by π f, σ D I ). Then, there are two commonly used pont estmators n Bayesan analyss that correspond to the posteror mode MAP) and the posteror mean MMSE), respectvely. Gven the posteror ndependence see Corollary 2.2) they are gven by: f MAP) = arg max f π f D I ), 2.10) σ MAP) = arg max σ π σ D I ), 2.11) and f MMSE) = σ MMSE) = f π f D I ) df = E [F D I ], 2.12) σ π σ D I )dσ = E [Ξ D I ]. 2.13) Note that for dffuse pror we fnd see Corollary 5.1 n Gsler-Wüthrch [11]) f MMSE) f CL). 2.14) Hence, usng Corollary 2.2, we obtan the approxmaton E [C,J D I ] = E [E [C,J D I, F,Ξ] D I ] = C,I E = C,I =I C,I =I E [F D I ] = C,I f CL) = Ĉ,J, =I =I F D I f MMSE) 2.15) where on the last lne we have an asymptotc equalty f the dffusvty of the prors πf ) tends to nfnty. Ths s exactly the argument why the Bayesan CL model can be used to ustfy the CL predctors, see Gsler-Wüthrch [11]. 3. Bootstrap and mean square error of predcton Assume that we have calculated the Bayesan predctor or the CL predctor gven n 2.15). Then we would lke to determne the predcton uncertanty,.e., we would lke to study the 8

9 devaton of C,J around ts predctor. If one s only nterested n second moments, the socalled condtonal mean square error of predcton MSEP), one can often estmate the error terms analytcally. However, other uncertanty measures lke Value-at-Rsk VaR) can only be determned numercally. A popular numercal method s the bootstrap method. The bootstrap technque was developed by Efron [3] and extended by Efron-Tbshran [4] and Davson-Hnkley [1]. Ths procedure allows one to obtan nformaton regardng an aggregated dstrbuton gven a sngle realsaton of the data. To apply the bootstrap procedure one ntroduces a mnmal amount of model structure such that resamplng observatons can be acheved usng observed samples of the data. In ths secton we present a bootstrap algorthm n the classcal frequentsts approach,.e., we assume that the CL factors F = f and the standard devaton parameters Ξ = σ gven n Model Assumptons 2.1 are unknown constants. The bootstrap then generates synthetc data denoted by DI that allow for the study of the fluctuatons of f CL) and σ 2CL) for detals see Wüthrch-Merz [22], Secton 7.4). In the presented text we restrct ourselves to the condtonal resamplng approach presented n Secton of Wüthrch-Merz [22]. Non-parametrc classcal bootstrap condtonal verson). 1. Calculate estmated resduals ε, for + I, > 0, condtonal on the estmators and σ 2CL) 0: and the observed data D I: CL) ε, = ε, f 1, σcl) 1 ) = C, CL) f 1 C, 1 σ CL) 1 C, These resduals ε, ) + I gve the emprcal bootstrap dstrbuton F DI. 3. Sample..d. resduals ε, F DI for + I, > Generate bootstrap observatons condtonal resamplng) f CL) 0: C, = f CL) 1 C, 1 + σ CL) 1 C, 1 ε,, whch defnes DI = DI f CL), σ CL) ). Note that for the uncondtonal verson of bootstrap we should generate C, CL) = f 1 C, 1 + σcl) 1 C, 1 ε,. For a dscusson on ths approach see Secton of [22]. 5. Calculate bootstrapped CL parameters f f = σ 2 = I 1 =0 C,+1 I 1, =0 C, 1 I 1 =0 and σ2 I 1 9 by C, C,+1 C, f ) 2.

10 6. Repeat steps 3-5 and obtan emprcal dstrbutons from the bootstrap samples Ĉ,J, f and σ 2. These are then used to quantfy the parameter estmaton uncertanty. Ths non-parametrc classcal bootstrap method can be seen as a frequentst approach. Ths means that we do not express our parameter uncertanty by the choce of an approprate pror dstrbuton. We rather use a pont estmator for the unknown parameters and then study the possble fluctuatons of ths pont estmator. The man dffculty now s that the non-parametrc bootstrap method, as descrbed above, underestmates the true uncertanty. Ths comes from the fact that the estmated resduals ε,, n general, have varance smaller than 1 see formula 7.23) n Wüthrch-Merz [22]). Ths means that our estmated resduals are not approprately scaled. Therefore, frequentsts use several dfferent scalngs to correct ths fact see formula 7.24) n Wüthrch-Merz [22] or England- Verrall [5]). Here, we use a dfferent approach by ntroducng the Bayesan bootstrap method, see Secton 4 below. Frequentst bootstrap estmates. Let us for the tme-beng concentrate on the condtonal MSEP gven by msep C,J D I Ĉ,J ) [ ) ] 2 = E C,J Ĉ,J D I Ĉ,J) 2 = Var C,J D I ) + E [C,J D I ]. 3.1) The frst term s known as the condtonal process varance and the second term as the parameter estmaton uncertanty. In the frequentsts approach,.e. for gven determnstc F = f and Ξ = σ, ths can be calculated, see Wüthrch-Merz [22], Secton 3.2. Namely, the terms are gven by and Var C,J D I ) = E [C,J C,I ] ) 2 ) 2 E [C,J D I ] Ĉ,J = C 2,I =I =I f σ 2/f2 def. = C,I Γ I, 3.2) E [C, C,I ] f CL) =I 2 def. = C 2,I I. 3.3) The process varance 3.2) s estmated by replacng the parameters by ts estmators, VarC,J D I ) = Ĉ,J ) 2 =I σ 2CL) CL) / f ) 2 Ĉ, def. = C,I Γfreq I. 3.4) 10

11 The parameter estmaton error s more nvolved and there we need the bootstrap algorthm. Assume that the bootstrap method gves T bootstrap samples 1) f,..., T) f. Then the parameter estmaton error 3.3) s estmated by the sample varance of the product of the bootstrap observaton chan ladder parameter estmates 1) f,..., T) f, whch gves the estmator C,I 2 freq I. Bayesan estmates. In the Bayesan setup,.e. choosng pror dstrbutons for the unknown parameters F and Ξ, we obtan a natural decomposton of the condtonal MSEP. msep C,J D I E [C,J D I ]) = Var C,J D I ) 3.5) = E [Var C,J D I, F,Ξ) D I ] + Var E [C,J D I, F,Ξ] D I ). The average process varance s gven by see Wüthrch-Merz [22], Lemma 3.6) 1 E [Var C,J D I, F,Ξ) D I ] = C,I E F m Ξ 2 Fn 2 D I 3.6) = C,I 1 =I m=i =I m=i E [F m D I ] E [ Ξ 2 ] D I n=+1 n=+1 E [ Fn 2 ] def. D I = C,I ΓBayes, where we have used Corollary 2.2. The parameter estmaton error s gven by Var E [C,J D I, F,Ξ] D I ) = C,I 2 Var F D def. I = C,I 2 Bayes I, 3.7) =I where we have used 2.15). Usng posteror ndependence, see Corollary 2.2, we obtan for the last term C,I 2 Bayes I = C,I 2 =I E [ F 2 ] D I =I E [F D I ] ) In order to calculate these two terms gven n 3.6) and 3.8) we need to calculate the posteror dstrbuton of F,Ξ), gven D I. Snce we do not have a full dstrbutonal model, we cannot wrte down the lkelhood functon, whch would allow for analytcal solutons or Markov chan Monte Carlo MCMC) smulatons. Therefore we ntroduce the ABC framework whch allows for dstrbuton-free smulatons usng approprate bootstrap samples and a dstance metrc. Ths we are gong to dscuss n the next secton. Credblty Estmates. As mentoned prevously, we can also consder the credblty estmates gven n Gsler-Wüthrch 11

12 [11]. As long as we are only nterested n the second moments,.e. condtonal MSEP, we can also use credblty estmators, whch are mnmum varance estmators that are lnear n the observatons. For dffuse prors we obtan the approxmaton gven n Corollary 7.2 of Gsler-Wüthrch [11] where Γ cred I = cred I = msep C,J D I E [C,J D I ]) = C,I Γcred + C,I 2 cred I, 3.9) =I =I 1 m=i f CL) m σ 2CL) n=+1 CL) f ) 2 σ 2CL) + I 1 =0 C, ) ) CL) f n ) 2 σ n 2CL) + I n 1 =0 C,n, 3.10) =I CL) f ) ) In the results secton we compare the frequentst bootstrap approach, the credblty approach and the ABC bootstrap approach that s descrbed below see Table 7 below). Note that the ABC bootstrap approach can also be appled to other rsk measures, not ust MSEP. Ths s unlke the credblty approach whch only apples to MSEP estmates. For example, f we fx a securty level 95% we can ask for the VaR on that level, whch s defned by ) { [ ] } VaR 0.95 C,J E [C,J D I ] D I = mn x; P C,J E [C,J D I ] > x D I ) 4. ABC for ntractable lkelhoods Obtanng samples { f t), σ 2t)} whch are realsatons of a random vector dstrbuted wth t=1:t a posteror dstrbuton π f, σ D I ) n the DFCL model s dffcult snce the lkelhood s ntractable. Hence, standard approaches such as Markov chan Monte Carlo MCMC) algorthms see, e.g., Glks et al. [10]) cannot be drectly used. Other authors have avoded ths dffculty by makng dstrbutonal assumptons for the form of the lkelhood. Ths then volates a dstrbuton-free CL model assumpton but allows for relatvely standard samplng procedures to be appled. Note that a specfc Gaussan assumpton does not allow for skewness. Here, we do not make any such assumptons and nstead we work n a truly dstrbuton-free model usng approxmate Bayesan computaton ABC) to facltate samplng from an ntractable posteror dstrbuton. To sample from the posteror n our DFCL model we develop a novel formulaton of the ABC methodology based on the bootstrap and condtonal back transformaton procedure, smlar to that dscussed n Secton 3. ABC methods am to sample from posteror dstrbutons n the presence of computatonally ntractable lkelhood functons. For an applcaton n rsk modellng of ABC technques see 12

13 Peters-Ssson [15]. In ths artcle we present a novel ABC-MCMC algorthm. Alternatvely, sequental Monte Carlo SMC) based algorthms whch can mprove smulaton effcency can be found n Del Moral et al. [2], Ssson et al. [21], Peters et al. [16],[17] and Maroram et al. [14]. These algorthms are beyond the scope of the present paper. We provde a basc argument for how the ABC methodology works n Appendx A. For a gven observaton y we want to sample from πθ y) wth an ntractable lkelhood functon. We assume that Sy) s a suffcent statstc for the model from whch we assume data y s a realsaton. We defne a hard decson functon gx, y) = I{ρSx), Sy)) < ǫ}x) for a gven tolerance level ǫ > 0 and a dstance metrc ρ, ), where I{ } s the ndcator functon whch equals 1 f the event s true and 0 otherwse. As demonstrated n Appendx A we use the approxmaton, see A.3)-A.4), πθ y) gx, y) πθ) πx θ) dx gx, y) πθ) πx θ) dxdθ = πθ)e [gx, y) θ], 4.1) E [gx, y)] where X πx θ) and θ πθ). In the next step the numerator of 4.1) s approxmated usng the emprcal dstrbuton,.e. πθ)e [gx, y) θ] πθ) 1 L L l=1 ) g x l) θ), y, 4.2) where x l) θ)..d. πx θ). Fnally, we need to consder the denomnator E [gx, y)]. In general ths has a non-trval form that cannot be calculated analytcally. However, snce we use an MCMC based method the denomnators cancel n the accept-reect stage of the algorthm. Therefore the ntractablty of the denomnator does not mpede samplng from the posteror. Therefore, we use πθ y) gx, y) πθ) πx θ) dx gx, y) πθ) πx θ) dxdθ πθ)e [gx, y) θ] πθ) 1 L L l=1 ) g x l) θ), y 4.3) n order to obtan samples from πθ y). Almost unversally, L = 1 s adopted to reduce computaton but on the other hand ths wll slow down the rate of convergence to the statonary dstrbuton. Note that sometmes one also uses softer decson functons for gx, y). The role of the dstance measure ρ s evaluated by Fan et al. [6], we further extend ths analyss to the class of models consdered n ths paper. We analyse several choces for the dstance measure ρ such as Mahlanobs dstance, scaled Eucldean dstance and the Manhattan Cty Block dstance. Fan et al. [6] demonstrate that t s not effcent to utlse the standard Eucldean dstance, especally when summary statstcs consdered are on dfferent scales. 13

14 Addtonally, usng an ABC-MCMC algorthm, t s mportant to assess convergence dagnostcs. Partcularly when usng ABC-MCMC where seral correlaton n the Markov chan samples can be sgnfcant f the sampler s not desgned carefully. We assess autocorrelaton of the smulated Markov chan, the Geweke [9] tme seres statstc and the Gelman-Rubn [7] R- statstc convergence dagnostc n an ABC settng. Concludng: We apply three dfferent technques n order to treat the ntractable lkelhood: 1) we use ABC to get a handle on the lkelhood, 2) therefore we need synthetc samples from the DFCL model gven realsatons of the parameters, these come from the bootstrap algorthm and 3) by usng a well understood MCMC based samplng algorthm we obtan cancellaton of the non-analytc normalzng constants. Next we dscuss the ABC-MCMC algorthm n detal ABC algorthmc choces for the tme seres DFCL model We start wth the choces of the ABC components. Data set: The observatons are gven by the cumulatve clams denoted by D I. Generaton of a synthetc data set: Note that n ths settng not only s the lkelhood ntractable but addtonally the generaton of a synthetc data set D I parameter values F,Ξ s not straghtforward. The synthetc data set D I gven the current s generated usng the bootstrap procedure descrbed n Secton 3. Note that both the bootstrap resdual ε, and the bootstrap samples DI are functons of the parameter choces, see page 9. Therefore we generate for gven F = f and Ξ = σ the bootstrap resduals ε, = ε, f 1, σ 1 ) and the bootstrap samples DI = D I f, σ) accordng to the non-parametrc bootstrap see page 9) where we replace the CL parameter estmates f CL), σ CL) ) by the parameters θ = F,Ξ). Summary statstcs: In order to defne the decson functon g we ntroduce summary statstcs, see Appendx A for detals of the role these summary statstcs play n the ABC approxmaton. For the observed data D I we defne the vector S D I ; 0, 1) = S 1,...,S n+2 ) = C 0,1,...,C 0,J, C 1,1,...,C 0,,...,C I 2,1, C I 2,2, C I 1,1 ; 0, 1), where n denotes the number of resduals ε,. For gven θ = F,Ξ) we generate the bootstrap sample D I = D I F,Ξ) as descrbed above. The correspondng resduals ε, = 14

15 ε, F 1, Ξ 1 ) should also be close to the standardzed observatons. Therefore we defne ts emprcal mean and standard devaton by µ = µ F,Ξ) = 1 n s = s F,Ξ) = ε, F 1, Ξ 1 ), 4.4), 1 n 1 ε, F 1, Ξ 1 ) µ F,Ξ)) 2 Hence, the summary statstcs for the synthetc data s gven by, 1/2. 4.5) S D I; µ, s ) = C 0,1,...,C 0,J, C 1,1,...,C 0,,...,C I 2,1, C I 2,2, C I 1,1; µ, s ). Dstance metrcs: Mahlanobs dstance and scaled Eucldean dstance Here we draw on the analyss of Ssson et al. [6] who propose the use of the Mahlanobs dstance metrc gven by, ρ S D I ; 0, 1),S D I; µ, s )) = [S D I ; 0, 1) S D I; µ, s )] Σ 1 D I [S D I ; 0, 1) S D I; µ, s )], where the covarance matrx Σ DI s an approprate scalng descrbed n Appendx B. The scaled Eucldean dstance s obtaned when we only consder the dagonal elements of the covarance matrx Σ DI. Note, the covarance matrx Σ DI provdes a weghtng on each element of the vector of summary statstcs to ensure they are scaled approprately accordng to ther nfluence on the ABC approxmaton. There are many other such weghtng schemes one could conceve. Manhattan Cty Block dstance We consder the L 1 -dstance gven by n+2 ρ S D I ; 0, 1),S DI; µ, s )) = S D I ; 0, 1) S DI; µ, s ). Decson functon: We work wth a hard decson functon gven by =1 g D I, D I) = I {ρ S D I ; 0, 1),S D I; µ, s )) < ǫ}. 15

16 Tolerance schedule: We use the sequence ǫ t = max { 20, t, ǫ mn}. Note, the use of an ABC-MCMC algorthm can result n stckng of the chan for extended perods. Therefore, one should carefully montor convergence dagnostcs of the resultng Markov chan for a gven tolerance schedule. There s a trade-off between the length of the Markov chan requred for samples approxmately from the statonary dstrbuton and the bas ntroduced by non zero tolerance. In ths paper we set ǫ mn va prelmnary analyss of the Markov chan sampler mxng rates for a transton kernel wth coeffcent of varaton set to one. We note that n general practtoners wll have a requred precson n posteror estmates, ths precson can be drectly used to determne, for a gven computatonal budget, a sutable tolerance ǫ mn. Convergence dagnostcs: We stress that when usng an ABC-MCMC algorthm, t s crucal to carefully montor the convergence dagnostcs of the Markov chan. Ths s more mportant n the ABC context than n the general MCMC context due to the possblty of extended reectons where the Markov chan can stck n a gven state for long perods. Ths can be combatted n several ways whch wll be dscussed once the algorthm s presented. The convergence dagnostcs we consder are evaluated only on samples post annealng of the tolerance threshold and after an ntal burnn perod once tolerance of ǫ mn s reached. If the total chan has length T, the ntal burnn stage wll correspond to the frst T b samples and we defne T = T T b. We denote by {θ t) } t=1: T the Markov chan of the -th parameter after burnn. The dagnostcs we consder are gven by, Autocorrelaton. Ths convergence dagnostc wll montor seral correlaton n the Markov chan. For gven Markov chan samples for the -th parameter {θ t) } t=1: T we defne the based autocorrelaton estmate at lag τ by ÂCFθ, τ) = 1 T τ)ˆσ θ ) T τ t=1 [θ t) µθ )][θ t+τ) µθ )], 4.6) where µθ ) and ˆσ θ ) are the estmated mean and standard devaton of θ. Geweke [9] tme seres dagnostc. For parameter θ we calculate: 16

17 1. Splt the Markov chan samples nto two sequences, {θ t) } t=1:t1 and {θ t) } t=t : T, such that T = T T 2 + 1, and wth ratos T 1 / T and T 2 / T fxed such that T 1 + T 2 )/ T < 1 for all T. ) 2. Evaluate µ and µ sequence. θ T 1 θ T 2 ) correspondng to the sample means on each sub 3. Evaluate consstent spectral densty estmates for each sub sequence, at frequency 0, denoted ŜD0; T 1, θ ) and ŜD0; T 2, θ ). The spectral densty estmator consdered n ths paper s the classcal non-parametrc perodogram or power spectral densty estmator. We use Welch s method wth a Hannng wndow, for detals see Appendx C. 4. Evaluate convergence dagnostc gven by Z T = µ ) µ θ T 2 θ T 1 ŜD0;T 1,θ )+T 1 ). For T one has accordng to the central T ŜD0;T 2,θ ) lmt theorem that Z T N0, 1) f the sequence {θ t) } t=1: T s statonary. Gelman-Rubn [7] R-statstc dagnostc. Ths approach to convergence analyss requres that one runs multple parallel ndependent Markov chans each startng at randomly selected ntal startng ponts, we run 5 chans. For comparson purposes we splt the total computatonal budget of T nto T1 = T 2 =... = T 5 = T 5. We compute the convergence dagnostc for parameter θ wth the followng steps: 1. Generate 5 ndependent Markov chan sequences, producng the chans for parameter θ, denoted {θ t),k } t=1:t k for k {1,...,5}. ) 2. Calculate the sample means µ θ T k for each sequence, and the overall mean ) µ θ T. 3. Calculate the varance of the sequence means, ) )) µ µ θ T 2 def. = B /T k k=1 θ T k ) 4. Calculate the wthn-sequence varances ŝ θ 2 T k for each sequence. 5. Calculate the average wthn-sequence varance, k=1 ŝ2 θ T k ) def. = W. 6. Estmate the target posteror varance for parameter θ by the weghted lnear ) combnaton, σ 2 θ T = T k 1 T k W + 1 T k B. Ths estmate s unbased for samples whch are from the statonary dstrbuton. In the case n whch not all sub chans have reached statonarty, ths overestmates the posteror varance for a fnte T but t asymptotcally, T, converges to the posteror varance. 17

18 7. Improve on the Gaussan estmate of the target posteror gven by ) N µ θ T, σ 2 θ T )) by accountng for samplng varablty n the estmates of the posteror mean and varance. Ths can be acheved by makng a Student-t ) approxmaton wth locaton µ θ T, scale V and degrees of freedom df, each gven respectvely by; ) V = σ 2 θ T + B T and df = 2 V 2 ) Var V = 1 5 Var V ) ) T1 1 2 Var T T 1 1) Ĉov 25T 1 24T 1 1) µ 25T 1, where the varance s estmated as, ŝ 2 θ T k ŝ 2 θ T k ) θ T Ĉov ), µ )) ) B 2 T 2 ŝ 2 θ T k θ T )) ) )), µ θ T. 4.7) Note, the covarance terms are estmated emprcally usng the wthn sequence estmates of the mean and varance obtaned for each sequence. 8. Calculate the convergence dagnostc, R = V df W df 2), where as T one can prove that R 1. Ths convergence dagnostc montors the scale factor by whch the current dstrbuton for θ may be reduced f smulatons are contnued for T. We begn by presentng the ABC-MCMC based algorthm we propose for samplng from the posteror dstrbuton presented n Secton ABC-MCMC to sample from π f, σ D I ) We develop an ABC-MCMC algorthm whch has an adaptve proposal mechansm and annealng of the tolerance durng burnn of the Markov chan. Havng reached the fnal tolerance post annealng, denoted ǫ mn, we utlze the remanng burnn samples to tune the proposal dstrbuton to ensure an acceptance probablty between the range of 0.3 and 0.5 s acheved. The optmal acceptance probablty when posteror parameters are..d. Gaussan was proven to be at 0.234, see Gelman et al. [8]. Though our problem does not match the requred condtons for ths proof, t provdes a practcal gude. To acheve ths, we tune the coeffcent of varaton of the proposal, n our case the shape parameter of the gamma proposal dstrbuton. We mpose an addtonal constrant that the mnmum shape parameter value s set at γ mn for {1,...,2J}. ABC-MCMC algorthm usng bootstrap samples. 18

19 ) 1. For t = 0 ntalze the parameter vector randomly, ths gves θ 0) 1:2J = f 0) 0:, σ0) 0:. Intalze the proposal shape parameters, γ γ mn 2. For t = 1,...,T ) a) Set θ t) 1:2J = b) For = 1,...,2J θ t 1) 1:2J ) for all {1,...,2J}.. Sample proposal θ from a Γγ, θ t) /γ )-dstrbuton. We denote the gamma proposal densty by K θ ; γ, θ t) /γ ). Ths gves proposed parameter vector ) θ = θ t) 1: 1, θ, θt). +1:2J. Condtonal on θ = θ t) 1: 1, θ, θt) +1:2J ), generate synthetc bootstrap data set D I = D I θ ) usng the bootstrap procedure detaled n Secton 3 where we replace the CL parameter estmates f CL), σ CL) ) by the parameters θ.. Evaluate summary statstcs S D I ; 0, 1) and S D I ; µ ; s ) and correspondng decson functon gd I, DI ) as descrbed n Secton 4.1. v. Accept proposal wth ABC acceptance probablty ) ) π θ A θ t) 1:2J, θ ) = mn 1, K θ t) ; γ, θ /γ ) ) gd π θ t) K θ ; γ, θ t) I, DI) /γ. That s, smulate U U0, 1) and set θ t) = θ f U < A θ t) 1:2J, θ ). v. If 100 t T b and ǫ t = ǫ mn then check to see f requre tunng of the proposal. Defne the average acceptance probablty over the last 100 teratons of updates for parameter by ā t 100:t) 0.9γ f ā t 100:t) γ = and consder the adapton, 1.1γ f ā t 100:t) > 0.5, γ otherwse. Then set the proposal shape parameter as, γ = max{γ, γmn }. < 0.3 and γ > γ mn, The ABC-MCMC algorthm presented can be enhanced by utlzng an dea of Gramacy et al. [12] n an ABC settng. Ths nvolves a combnaton of temperng the tolerance {ǫ t } t=1:t and mportance samplng correctons. 19

20 5. Example 1: Synthetc data The tunng of the proposal dstrbuton n ths study s done for the smplest base dstance metrc, the weghted Eucldean dstance. To study the effect of the dstance metrc n a comparatve fashon we shall keep the proposal dstrbuton unchanged. The frst example we present has a clams trangle of sze I = J = 9. In ths example we fx the true model parameters, denoted by f = f 0,...,f ) and σ 2 = σ0 2,.. ).,σ2 and gven n Table 2, used to generate the synthetc data set Generaton of synthetc data To generate the synthetc observatons for D I we generate randomly the frst column, denoted B 0. Then condtonal on ths realzaton of B 0 we make use of the model gven n 2.1) to generate the remanng columns of D I, ensurng the model assumptons are satsfed. Ths requres settng C,0 suffcently large, for approprate choces of f and σ 2, and then sample..d. realzatons of ε, U [ 3, 3 ] whch are used to obtan D I, see the observatons n Table Senstvty analyss and convergence assessment We perform a senstvty analyss, studyng the mpact of the dstance metrc on the mxng of the Markov chan n the case of ont estmaton of the chan ladder factors and the varance parameters. The pre-tuned coeffcent of varaton of the Gamma proposal dstrbuton for each parameter of the posteror was performed usng the followng settngs; T b = 50, 000, T = 200, 000, ǫ mn = 0.1 and ntal values γ = 1 for all {1,...,2J}. Addtonally, the pror parameters for the chan ladder factors F were set as α, β) = 2, 1.2/2) and the parameters for the varance parameters Ξ 2 were set as a, b) = 2, 1/2). After tunng the proposal dstrbutons durng burnn and roundng the shape parameters we found that γ = 10 for all {1,...,2J} produced average acceptance probabltes for each parameter between 0.3 and 0.5. Then keepng the proposal dstrbuton constant and usng a common data set D I we ran 3 versons of the ABC-MCMC algorthm for 200,000 samples correspondng to; 1. Scaled Eucldean dstance and ont estmaton of posteror for F,Ξ 2 2. Mahlanobs dstance modfed) and ont estmaton of posteror for F,Ξ 2 3. Manhattan Cty Block dstance and ont estmaton of posteror for F,Ξ 2. 20

21 We estmate the three convergence dagnostcs we presented n Secton 4.1. The results of ths analyss are presented as a functon of Markov chan teraton t post burn of 50,000 samples. Autocorrelaton Functon: In Fgure 1 we demonstrate the estmated autocorrelaton functons for the Markov chans of the random varables F 0 and Ξ 2 0. Snce the posteror parameters n ths model are ndependent t s sutable to analyze ust the margnal parameters to get a reasonable estmate of the mxng behavor of the MCMC-ABC algorthm on all the posteror parameters. The results demonstrate the degree of seral correlaton n the Markov chans generated for these parameters as a functon of lag tme τ. The hgher the decay rate n the tal of the estmated ACF as a functon of τ, the better the mxng of the MCMC algorthm. Due to the ndependence propertes of ths model there s lttle dfference between results obtaned for Scaled Eucldean and Mahlanobs dstances. As shown n Appendx A the estmate of the covarance matrx s dagonal on all but the rght lower 2 2 block. Hence, we recommend usng the smple Scaled Eucldean dstance metrc as t provded the best trade-off between smplcty and mxng performance. Geweke Tme Seres Dagnostc: Fgure 2 presents results for the Geweke tme seres dagnostc. Agan, we present the results for the random varables F 0 and Ξ 2 0. Note, we used the posteror mean as the sample functon and a set of ncreasng values for T from T b + 5, 000 ncreasng n steps of 5,000 samples to T. In each case we splt the chan n each wndow gven by {θ t) } t=1:t1 and {θ t) } t=t : T accordng to recommendatons from Geweke et al. [9]. We then calculate the convergence dagnostc Z T whch s the dfference between these 2 means dvded by the asymptotc standard error of ther dfference. As the chan length ncreases T, then the samplng dstrbuton of Z N0, 1) f the chan has converged. Hence values of Z T n the tals of a standard normal dstrbuton suggest that the chan was not fully converged early on.e. durng the 1st wndow). Hence, we plot Z T scores versus ncreasng T and montor f they le wthn a 95% confdence nterval Z T [ 1.96, 1.96]. The results n Fgure 2 demonstrate that clearly the convergence propertes of the dstance functons dffer. Agan ths s more materal n the Markov chan for the varance parameter when compared to the Markov chan results for the chan ladder factor. The man pont we note s that agan one would advse aganst use of the Cty block dstance metrc. Gelman and Rubn R statstc: In Fgure 3 we present the Gelman and Rubn convergence dagnostc. To calculate ths we ran 20 chans n parallel, each of length 10,000 samples and for each chan we dscarded 250 samples as burnn. We then estmated the R statstc as a functon of smulaton tme post burnn. In Fgure 3 we demonstrate the convergence rate of 21

22 the R statstc to 1 for each dstance metrc on ncreasng blocks of 200 samples. Usng ths summary statstc all three dstance metrcs are very smlar n terms of convergence rate of the R statstc to 1. Overall, these three convergence dagnostcs demonstrate the smple scaled Eucldean dstance metrc s the superor choce. Secondly, we see approprate convergence of the Markov chans under three convergence dagnostcs whch test dfferent aspects of the mxng of the Markov chans, gvng confdence n the performance of the MCMC-ABC algorthm n ths model Parameter estmaton Bayesan estmates In ths secton we present results for the Scaled Eucldean dstance metrc, wth a Markov chan of length 200,000 samples dscardng the frst 50,000 samples as burnn. In Tables 4 and?? we present the Chan Ladder parameter estmates for the DFCL model and the assocated parameter estmaton error. We defne the followng quanttes: MAP) f σ 0:, MMSE) f σ 0:, σ f σ 0: and [ˆq 0.05, ˆq 0.95 ] σ 0: denote respectvely the Maxmum a-posteror, Mnmum Mean Square Error, posteror standard devaton of the condtonal dstrbuton of chan ladder factor F and the posteror coverage probablty estmates at 5% of the condtonal dstrbuton of chan ladder factor F, each of these estmates s condtonal on knowledge of the true σ 0:. MAP) f, fmmse), σ f and [ˆq 0.05, ˆq 0.95 ] denote the same quanttes for the uncondtonal dstrbuton after ont estmaton of F 0:, Ξ 0:. Ave.[A θ 1:2J, f )] and Ave.[A θ 1:2J, σ )] denotes the average acceptance probablty of the Markov chan. σ 2MAP), σ 2MMSE), σ σ 2 and [ˆq 0.05, ˆq 0.95 ] denotes the same quanttes for the chan ladder varances as those defned above for chan ladder factors. For the frequentst approach we obtan the standard error n the estmates by usng 1,000 { } { } bootstrap realzatons of D s) I to obtan fccl) s=1:1,000 s), σ 2CCL) s). We use these s=1:1,000 bootstrap samples to calculate the standard devaton n the estmates of the parameters n the classcal frequentst CL approach, present n brackets.) next to ther correspondng estmators. The standard errors n the Bayesan parameter estmates are obtaned by blockng the Markov chan nto 100 blocks of length 1,500 samples and estmatng the posteror quanttes on each block. 22

23 6. Example 2: Real Clams Reservng data In ths example we consder estmaton usng real clams reservng data from Wüthrch-Merz [22], see Table 4. Ths yearly loss data s turned nto annual cumulatve clams and dvded by 10,000 for the analyss n ths example. We use the analyss from the prevous study to ustfy use of the ont MCMC-ABC smulaton algorthm wth a Scaled Eucldean dstance metrc. We pre-tuned the coeffcent of varaton of the Gamma proposal dstrbuton for each parameter of the posteror. Ths was performed usng the followng settngs; T b = 50, 000, T = 200, 000, ǫ mn = 10 5 and ntal values γ = 1 for all {1,...,2J}. Here we make a strct requrement of the tolerance level to ensure we have accurate results from our ABC approxmaton. Addtonally, ) CL) the pror parameters for the chan ladder factors F were set as α, β ) = 1, f and the ) parameters for the varance Ξ 2 prors were set as a, b ) = 1, σ CL). The code for ths problem was wrtten n Matlab and t took approxmately 10 mn to smulate 200,000 samples from the MCMC-ABC algorthm, on a Intel Xeon 3.4GHz processor wth 2Gb RAM. After tunng the proposal dstrbutons durng burnn we obtaned rounded shape parameters γ 1:9 = [50, 100, 500, 500, 5, 000, 20, 000, 100, 000, 2, 000, 000, 3, 000, 000] provded average acceptance probabltes between 0.3 and 0.5. In Fgures 4 we present box-whsker plots of estmates of the dstrbutons of the parameters F 0:, Ξ 0: obtaned from the MCMC-ABC algorthm, post burnn. Fgure 5 presents the Bayesan MCMC-ABC emprcal dstrbutons of the ultmate clams, C,J for = 1,...,I. In Table 5 we present the predcted cumulatve clams for each year along wth the estmates for the chan ladder factors and chan ladder varances under both the classcal approach and the Bayesan model. We see that wth ths farly vague pror specfed, we do ndeed obtan convergence of the MCMC-ABC based Bayesan estmates f MMSE), σ MMSE) to the classcal estmates f CL), σ CL). In Fgure 6 we present a study of the hstogram estmate of the margnal posteror dstrbuton for chan ladder factor π f 0 D I, ǫ mn). The plot was obtaned by samplng from the full posteror π f, σ D I, ǫ mn) for each specfed tolerance value, ǫ mn. Then the samples for the partcular chan ladder parameter n each plot are turned nto a smoothed hstogram estmate for each epslon and plotted. We observe a sgnfcant dfference n the parameter uncertanty as reflected by the change n the posteror precson as the tolerance decreases. Hence, the tolerance wll mpact the msep. Mnmzng ths mpact, nvolves managng the computatonal budget to acheve mxng at a gven tolerance level whlst mnmzng the tolerance level. 23

24 Ultmately, we would lke an algorthm whch could work well for any ǫ mn, the smaller the better. However, we note that wth a decreasng ǫ mn n the sampler we present n ths paper, one must take addtonal care to ensure the Markov chan s stll mxng and not stuck n a partcular state, as s observed to be the case n all MCMC-ABC algorthms. To avod ths acknowledged dffculty wth MCMC-ABC requres that one ether runs much longer MCMC chans or t requres the use of more sophstcated samplng algorthms such as SMC Samplers PRC-ABC based algorthms, see Ssson, Peters, Fan and Brers 2008). Ths s well beyond the scope of ths paper and wll be the subect of future papers and nvestgaton n ths context. In Table 6 we present the predctve VaR at 95% and 99% levels for the ultmate predcted clams, obtaned from the MCMC-ABC algorthm. These are easly obtaned under the Bayesan settng, we smply used the MCMC-ABC posteror samples to explctly obtan samples from the full predctve dstrbuton of the cumulatve clams after ntegratng out the parameter uncertanty numercally. In addton to ths we present the analyss of the MSEP under the bootstrap frequentst procedure and the Bayesan MCMC-ABC and credblty estmates for the total predcted cumulatve clams for each accdent year. We also present results for the sum of the total cumulatve clams for each accdent year. We can make the followng conclusons from these results: 1. The process varance for each C,J demonstrate that the unscaled condtonal frequentst bootstrap and the credblty estmates are very close for all accdent years. The Bayesan results compare favorably. 2. The results for the parameter estmaton error for the predcted cumulatve clams C,J demonstrate for small the Bayesan approach outperforms the frequentst approach. However, the Bayesan approach produces performance whch worsens as ncreases, relatve to the credblty approach. Ths could be a result of the tolerance level settng n the ABC algorthm. In future work t wll be nterestng to study the effect of the tolerance level on the parameter estmaton error. Fgure 6 demonstrates numercally that the tolerance ǫ should affect the parameter estmaton error n the Bayesan approach. Ths study requres sophstcated algorthms such as those found n Ssson, Peters, Fan and Brers 2008). 3. The total results for the process varance for C = C,J demonstrate that the frequentst and credblty results are very close. Addtonally, Bayesan total results are largest followed by credblty and then frequentst estmates whch s n agreement wth theoretcal bounds. 24

25 4. The total results for the parameter estmaton error for C = C,J demonstrate that frequentst uncondtonal bootstrap procedure results n the lowest total error. The Bayesan approach and credblty total parameter errors are close. Addtonally, we note that the results n Wüthrch-Merz, 2008) Table 7.1 for the total parameter estmaton error under an uncondtonal frequentst bootstrap wth unscaled resduals s also very close to the total obtaned for the frequentst approach. 7. Dscusson Ths paper has presented a dstrbuton-free clams reservng model under a Bayesan paradgm. We have then present a novel advanced MCMC-ABC algorthm to obtan estmates from the resultng ntractable posteror dstrbuton of the chan ladder factors and chan ladder varances. We assessed several aspects of ths algorthm, ncludng the propertes of the convergence of the MCMC algorthm as a functon of the dstance metrc approxmaton n the ABC component. We studed the performance on a synthetc data set generated from known parameters, n order to demonstrate the accuracy of ths methodology. Next we assessed a real clams reservng data set and compared the results we obtaned for predcted cumulatve ultmate clams to those obtaned va classcal chan ladder methods and va credblty theory. We note that ths clearly demonstrates our algorthm s workng accurately and provdes us not only wth the ablty to obtan pont estmates for the frst and second moments of the ultmate cumulatve clams, but an accurate emprcal approxmaton of the entre dstrbuton of the ultmate clams. Ths s valuable for many reasons, ncludng predcton of reserves whch are not based on centralty measures, such as the tal based VaR results we present. We have demonstrated a unque and accurate way n whch one can bypass the need to make dstrbutonal approxmatons. Acknowledgements The fst author thanks ETH FIM and ETH Rsk Lab for ther generous fnancal assstance whlst completng aspects of ths work at ETH. The frst author also thanks the Department of Mathematcs and Statstcs at the Unversty of NSW for support through an Australan Postgraduate Award and to CSIRO for support through a postgraduate research top up scholarshp. Fnally, ths materal was based upon work partally supported by the Natonal Scence Foundaton under Grant DMS to the Statstcal and Appled Mathematcal Scences Insttute, North Carolna, USA. Any opnons, fndngs, and conclusons or recommendatons expressed n ths materal are those of the authors) and do not necessarly reflect the vews of the Natonal 25

26 Scence Foundaton References [1] Davson, A.C., Hnkley, D.V. 1997). Bootstrap Methods and Ther Applcaton. Cambrdge Unversty Press, Cambrdge. [2] Del Moral, P., Doucet, A., Jasra, A. 2006). Sequental Monte Carlo samplers. Journal of the Royal Statstcal Socety Seres. B. 683), [3] Efron, B. 1979). Bootstrap methods: another look at the ackknfe. Annals of Statstcs. 71), [4] Efron, B., Tbshran, R.J. 1993). An Introducton to the Bootstrap. Chapman & Hall, NY. [5] England, P.D., Verrall, R.J. 2002). Stochastc clams reservng n general nsurance. Brtsh Actuaral Journal. 83), [6] Fan, Y., Ssson, S.A., Peters, G.W. 2008). Improved effcency n approxmate Bayesan computaton. Techncal report, Statstcs Department, Unversty of New South Wales. [7] Gelman, A., Rubn, D.B. 1992). Inference from teratve smulaton usng multple sequences. Statstcal Scence 4, [8] Gelman, A., Glks, W.R., Roberts, G.O. 1997). Weak convergence and optmal scalng of random walk metropols algorthm. Annals of Appled Probablty 7, [9] Geweke, J.F. 1991). Evaluatng the accuracy of samplng-based approaches to the calculaton of posteror moments. In J.M. Bernardo, J.O. Berger, A.P. Dawd and A.F.M. Smth eds.) Bayesan Statstcs, 4, Oxford Unversty Press, Oxford. [10] Glks, W.R., Rchardson, S., Spegelhalter, D.J. 1996). Markov Chan Monte Carlo n Practce. Chapman & Hall, London. [11] Gsler, A., Wüthrch, M.V. 2008). Credblty for the chan ladder reservng method. ASTIN Bulletn 382), [12] Gramacy, R.B., Samworth, R.J., Kng, R. 2008). Importance temperng. Preprnt, arxv: v5 [stat.co]. 26

27 [13] Mack, T. 1993). Dstrbuton-free calculaton of the standard error of chan ladder reserve estmates. ASTIN Bulletn 23, [14] Maroram, P., Moltor, J., Plagnol, V., Tavare, S. 2003). Markov chan Monte Carlo wthout lkelhoods. Proceedngs of the Natonal Academy of Scence USA 100, [15] Peters, G.W., Ssson, S.A. 2006). Bayesan nference, Monte Carlo samplng and operatonal rsk. Journal of Operatonal Rsk 13), [16] Peters, G.W., Fan, Y., Ssson, S.A. 2008). On Sequental Monte Carlo, partal reecton control and approxmate Bayesan computaton. Preprnt, Statstcs Department, Unversty of New South Wales. [17] Peters, G.W., Ssson, S.A., Fan, Y. 2008). Desgn effcency for lkelhood free Sequental Monte Carlo samplers. Preprnt, Statstcs Department, Unversty of New South Wales. [18] Ssson, S.A., Peters, G.W., Fan, Y., Brers, M., 2008). Lkelhood free samplers. Preprnt, Statstcs Department, Unversty of New South Wales. [19] Peters, G.W., Wüthrch, M.V., Shevchenko, P., 2008). Model rsk n clams reservng wthn Tweede s compound Posson models. to appear, ASTIN Bulletn. [20] Proaks, J.G., Manolaks, D.G., 1996). Dgtal Sgnal Processng. Upper Saddle Rver, N.J. Prentce Hall. [21] Ssson, S.A., Fan, Y., Tanaka, M. 2007). Sequental Monte Carlo wthout lkelhoods. Proceedngs of the Natonal Academy of Scence USA 104, [22] Wüthrch, M.V., Merz, M. 2008). Stochastc Clams Reservng Methods n Insurance. Wley Fnance. [23] Yao, J. 2008). Bayesan approach for predcton error n chan-ladder clams reservng. Conference paper presented at the ASTIN Colloqum, Manchester UK. 27

Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement

Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement Markov Chan Monte Carlo MCMC, Gbbs Samplng, Metropols Algorthms, and Smulated Annealng 2001 Bonformatcs Course Supplement SNU Bontellgence Lab http://bsnuackr/ Outlne! Markov Chan Monte Carlo MCMC! Metropols-Hastngs

More information

An adaptive SMC scheme for ABC. Bayesian Computation (ABC)

An adaptive SMC scheme for ABC. Bayesian Computation (ABC) An adaptve SMC scheme for Approxmate Bayesan Computaton (ABC) (ont work wth Prof. Mke West) Department of Statstcal Scence - Duke Unversty Aprl/2011 Approxmate Bayesan Computaton (ABC) Problems n whch

More information

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4) I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes

More information

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U) Econ 413 Exam 13 H ANSWERS Settet er nndelt 9 deloppgaver, A,B,C, som alle anbefales å telle lkt for å gøre det ltt lettere å stå. Svar er gtt . Unfortunately, there s a prntng error n the hnt of

More information

On an Extension of Stochastic Approximation EM Algorithm for Incomplete Data Problems. Vahid Tadayon 1

On an Extension of Stochastic Approximation EM Algorithm for Incomplete Data Problems. Vahid Tadayon 1 On an Extenson of Stochastc Approxmaton EM Algorthm for Incomplete Data Problems Vahd Tadayon Abstract: The Stochastc Approxmaton EM (SAEM algorthm, a varant stochastc approxmaton of EM, s a versatle tool

More information

Statistics for Economics & Business

Statistics for Economics & Business Statstcs for Economcs & Busness Smple Lnear Regresson Learnng Objectves In ths chapter, you learn: How to use regresson analyss to predct the value of a dependent varable based on an ndependent varable

More information

Markov Chain Monte Carlo Lecture 6

Markov Chain Monte Carlo Lecture 6 where (x 1,..., x N ) X N, N s called the populaton sze, f(x) f (x) for at least one {1, 2,..., N}, and those dfferent from f(x) are called the tral dstrbutons n terms of mportance samplng. Dfferent ways

More information

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems Numercal Analyss by Dr. Anta Pal Assstant Professor Department of Mathematcs Natonal Insttute of Technology Durgapur Durgapur-713209 emal: anta.bue@gmal.com 1 . Chapter 5 Soluton of System of Lnear Equatons

More information

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010 Parametrc fractonal mputaton for mssng data analyss Jae Kwang Km Survey Workng Group Semnar March 29, 2010 1 Outlne Introducton Proposed method Fractonal mputaton Approxmaton Varance estmaton Multple mputaton

More information

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number

More information

BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS. M. Krishna Reddy, B. Naveen Kumar and Y. Ramu

BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS. M. Krishna Reddy, B. Naveen Kumar and Y. Ramu BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS M. Krshna Reddy, B. Naveen Kumar and Y. Ramu Department of Statstcs, Osmana Unversty, Hyderabad -500 007, Inda. nanbyrozu@gmal.com, ramu0@gmal.com

More information

Composite Hypotheses testing

Composite Hypotheses testing Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter

More information

Global Sensitivity. Tuesday 20 th February, 2018

Global Sensitivity. Tuesday 20 th February, 2018 Global Senstvty Tuesday 2 th February, 28 ) Local Senstvty Most senstvty analyses [] are based on local estmates of senstvty, typcally by expandng the response n a Taylor seres about some specfc values

More information

Linear Approximation with Regularization and Moving Least Squares

Linear Approximation with Regularization and Moving Least Squares Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...

More information

A Robust Method for Calculating the Correlation Coefficient

A Robust Method for Calculating the Correlation Coefficient A Robust Method for Calculatng the Correlaton Coeffcent E.B. Nven and C. V. Deutsch Relatonshps between prmary and secondary data are frequently quantfed usng the correlaton coeffcent; however, the tradtonal

More information

Comparison of Regression Lines

Comparison of Regression Lines STATGRAPHICS Rev. 9/13/2013 Comparson of Regresson Lnes Summary... 1 Data Input... 3 Analyss Summary... 4 Plot of Ftted Model... 6 Condtonal Sums of Squares... 6 Analyss Optons... 7 Forecasts... 8 Confdence

More information

Uncertainty as the Overlap of Alternate Conditional Distributions

Uncertainty as the Overlap of Alternate Conditional Distributions Uncertanty as the Overlap of Alternate Condtonal Dstrbutons Olena Babak and Clayton V. Deutsch Centre for Computatonal Geostatstcs Department of Cvl & Envronmental Engneerng Unversty of Alberta An mportant

More information

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 30 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 2 Remedes for multcollnearty Varous technques have

More information

Development Pattern and Prediction Error for the Stochastic Bornhuetter-Ferguson Claims Reserving Method

Development Pattern and Prediction Error for the Stochastic Bornhuetter-Ferguson Claims Reserving Method Development Pattern and Predcton Error for the Stochastc Bornhuetter-Ferguson Clams Reservng Method Annna Saluz, Alos Gsler, Maro V. Wüthrch ETH Zurch ASTIN Colloquum Madrd, June 2011 Overvew 1 Notaton

More information

Chapter 11: Simple Linear Regression and Correlation

Chapter 11: Simple Linear Regression and Correlation Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests

More information

Negative Binomial Regression

Negative Binomial Regression STATGRAPHICS Rev. 9/16/2013 Negatve Bnomal Regresson Summary... 1 Data Input... 3 Statstcal Model... 3 Analyss Summary... 4 Analyss Optons... 7 Plot of Ftted Model... 8 Observed Versus Predcted... 10 Predctons...

More information

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction ECONOMICS 5* -- NOTE (Summary) ECON 5* -- NOTE The Multple Classcal Lnear Regresson Model (CLRM): Specfcaton and Assumptons. Introducton CLRM stands for the Classcal Lnear Regresson Model. The CLRM s also

More information

Chapter 5 Multilevel Models

Chapter 5 Multilevel Models Chapter 5 Multlevel Models 5.1 Cross-sectonal multlevel models 5.1.1 Two-level models 5.1.2 Multple level models 5.1.3 Multple level modelng n other felds 5.2 Longtudnal multlevel models 5.2.1 Two-level

More information

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA 4 Analyss of Varance (ANOVA) 5 ANOVA 51 Introducton ANOVA ANOVA s a way to estmate and test the means of multple populatons We wll start wth one-way ANOVA If the populatons ncluded n the study are selected

More information

Generalized Linear Methods

Generalized Linear Methods Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set

More information

Bayesian predictive Configural Frequency Analysis

Bayesian predictive Configural Frequency Analysis Psychologcal Test and Assessment Modelng, Volume 54, 2012 (3), 285-292 Bayesan predctve Confgural Frequency Analyss Eduardo Gutérrez-Peña 1 Abstract Confgural Frequency Analyss s a method for cell-wse

More information

Computing MLE Bias Empirically

Computing MLE Bias Empirically Computng MLE Bas Emprcally Kar Wa Lm Australan atonal Unversty January 3, 27 Abstract Ths note studes the bas arses from the MLE estmate of the rate parameter and the mean parameter of an exponental dstrbuton.

More information

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6 Department of Quanttatve Methods & Informaton Systems Tme Seres and Ther Components QMIS 30 Chapter 6 Fall 00 Dr. Mohammad Zanal These sldes were modfed from ther orgnal source for educatonal purpose only.

More information

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:

More information

Durban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications

Durban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications Durban Watson for Testng the Lack-of-Ft of Polynomal Regresson Models wthout Replcatons Ruba A. Alyaf, Maha A. Omar, Abdullah A. Al-Shha ralyaf@ksu.edu.sa, maomar@ksu.edu.sa, aalshha@ksu.edu.sa Department

More information

Simulated Power of the Discrete Cramér-von Mises Goodness-of-Fit Tests

Simulated Power of the Discrete Cramér-von Mises Goodness-of-Fit Tests Smulated of the Cramér-von Mses Goodness-of-Ft Tests Steele, M., Chaselng, J. and 3 Hurst, C. School of Mathematcal and Physcal Scences, James Cook Unversty, Australan School of Envronmental Studes, Grffth

More information

Errors for Linear Systems

Errors for Linear Systems Errors for Lnear Systems When we solve a lnear system Ax b we often do not know A and b exactly, but have only approxmatons  and ˆb avalable. Then the best thng we can do s to solve ˆx ˆb exactly whch

More information

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition)

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition) Count Data Models See Book Chapter 11 2 nd Edton (Chapter 10 1 st Edton) Count data consst of non-negatve nteger values Examples: number of drver route changes per week, the number of trp departure changes

More information

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton

More information

Lecture 12: Discrete Laplacian

Lecture 12: Discrete Laplacian Lecture 12: Dscrete Laplacan Scrbe: Tanye Lu Our goal s to come up wth a dscrete verson of Laplacan operator for trangulated surfaces, so that we can use t n practce to solve related problems We are mostly

More information

Limited Dependent Variables

Limited Dependent Variables Lmted Dependent Varables. What f the left-hand sde varable s not a contnuous thng spread from mnus nfnty to plus nfnty? That s, gven a model = f (, β, ε, where a. s bounded below at zero, such as wages

More information

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Analyss of Varance and Desgn of Experment-I MODULE VII LECTURE - 3 ANALYSIS OF COVARIANCE Dr Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Any scentfc experment s performed

More information

Econ Statistical Properties of the OLS estimator. Sanjaya DeSilva

Econ Statistical Properties of the OLS estimator. Sanjaya DeSilva Econ 39 - Statstcal Propertes of the OLS estmator Sanjaya DeSlva September, 008 1 Overvew Recall that the true regresson model s Y = β 0 + β 1 X + u (1) Applyng the OLS method to a sample of data, we estmate

More information

x i1 =1 for all i (the constant ).

x i1 =1 for all i (the constant ). Chapter 5 The Multple Regresson Model Consder an economc model where the dependent varable s a functon of K explanatory varables. The economc model has the form: y = f ( x,x,..., ) xk Approxmate ths by

More information

Chapter 13: Multiple Regression

Chapter 13: Multiple Regression Chapter 13: Multple Regresson 13.1 Developng the multple-regresson Model The general model can be descrbed as: It smplfes for two ndependent varables: The sample ft parameter b 0, b 1, and b are used to

More information

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands Content. Inference on Regresson Parameters a. Fndng Mean, s.d and covarance amongst estmates.. Confdence Intervals and Workng Hotellng Bands 3. Cochran s Theorem 4. General Lnear Testng 5. Measures of

More information

Lecture Notes on Linear Regression

Lecture Notes on Linear Regression Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume

More information

Testing for seasonal unit roots in heterogeneous panels

Testing for seasonal unit roots in heterogeneous panels Testng for seasonal unt roots n heterogeneous panels Jesus Otero * Facultad de Economía Unversdad del Rosaro, Colomba Jeremy Smth Department of Economcs Unversty of arwck Monca Gulett Aston Busness School

More information

GARETH W. PETERS, PAVEL V. SHEVCHENKO AND MARIO V. WÜTHRICH ABSTRACT

GARETH W. PETERS, PAVEL V. SHEVCHENKO AND MARIO V. WÜTHRICH ABSTRACT MODEL UNCERTAINTY IN CLAIMS RESERVING WITHIN TWEEDIE S COMPOUND POISSON MODELS BY GARETH W. PETERS, PAVEL V. SHEVCHENKO AND MARIO V. WÜTHRICH ABSTRACT In ths paper we examne the clams reservng problem

More information

NUMERICAL DIFFERENTIATION

NUMERICAL DIFFERENTIATION NUMERICAL DIFFERENTIATION 1 Introducton Dfferentaton s a method to compute the rate at whch a dependent output y changes wth respect to the change n the ndependent nput x. Ths rate of change s called the

More information

Basic Business Statistics, 10/e

Basic Business Statistics, 10/e Chapter 13 13-1 Basc Busness Statstcs 11 th Edton Chapter 13 Smple Lnear Regresson Basc Busness Statstcs, 11e 009 Prentce-Hall, Inc. Chap 13-1 Learnng Objectves In ths chapter, you learn: How to use regresson

More information

Parametric fractional imputation for missing data analysis

Parametric fractional imputation for missing data analysis Secton on Survey Research Methods JSM 2008 Parametrc fractonal mputaton for mssng data analyss Jae Kwang Km Wayne Fuller Abstract Under a parametrc model for mssng data, the EM algorthm s a popular tool

More information

Lecture 10 Support Vector Machines II

Lecture 10 Support Vector Machines II Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed

More information

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering / Theory and Applcatons of Pattern Recognton 003, Rob Polkar, Rowan Unversty, Glassboro, NJ Lecture 4 Bayes Classfcaton Rule Dept. of Electrcal and Computer Engneerng 0909.40.0 / 0909.504.04 Theory & Applcatons

More information

Convergence of random processes

Convergence of random processes DS-GA 12 Lecture notes 6 Fall 216 Convergence of random processes 1 Introducton In these notes we study convergence of dscrete random processes. Ths allows to characterze phenomena such as the law of large

More information

On Outlier Robust Small Area Mean Estimate Based on Prediction of Empirical Distribution Function

On Outlier Robust Small Area Mean Estimate Based on Prediction of Empirical Distribution Function On Outler Robust Small Area Mean Estmate Based on Predcton of Emprcal Dstrbuton Functon Payam Mokhtaran Natonal Insttute of Appled Statstcs Research Australa Unversty of Wollongong Small Area Estmaton

More information

Economics 130. Lecture 4 Simple Linear Regression Continued

Economics 130. Lecture 4 Simple Linear Regression Continued Economcs 130 Lecture 4 Contnued Readngs for Week 4 Text, Chapter and 3. We contnue wth addressng our second ssue + add n how we evaluate these relatonshps: Where do we get data to do ths analyss? How do

More information

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise. Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where y + = β + β e for =,..., y and are observable varables e s a random error How can an estmaton rule be constructed for the

More information

Estimation: Part 2. Chapter GREG estimation

Estimation: Part 2. Chapter GREG estimation Chapter 9 Estmaton: Part 2 9. GREG estmaton In Chapter 8, we have seen that the regresson estmator s an effcent estmator when there s a lnear relatonshp between y and x. In ths chapter, we generalzed the

More information

Basically, if you have a dummy dependent variable you will be estimating a probability.

Basically, if you have a dummy dependent variable you will be estimating a probability. ECON 497: Lecture Notes 13 Page 1 of 1 Metropoltan State Unversty ECON 497: Research and Forecastng Lecture Notes 13 Dummy Dependent Varable Technques Studenmund Chapter 13 Bascally, f you have a dummy

More information

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution Department of Statstcs Unversty of Toronto STA35HS / HS Desgn and Analyss of Experments Term Test - Wnter - Soluton February, Last Name: Frst Name: Student Number: Instructons: Tme: hours. Ads: a non-programmable

More information

Goodness of fit and Wilks theorem

Goodness of fit and Wilks theorem DRAFT 0.0 Glen Cowan 3 June, 2013 Goodness of ft and Wlks theorem Suppose we model data y wth a lkelhood L(µ) that depends on a set of N parameters µ = (µ 1,...,µ N ). Defne the statstc t µ ln L(µ) L(ˆµ),

More information

e i is a random error

e i is a random error Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where + β + β e for,..., and are observable varables e s a random error How can an estmaton rule be constructed for the unknown

More information

Lecture 12: Classification

Lecture 12: Classification Lecture : Classfcaton g Dscrmnant functons g The optmal Bayes classfer g Quadratc classfers g Eucldean and Mahalanobs metrcs g K Nearest Neghbor Classfers Intellgent Sensor Systems Rcardo Guterrez-Osuna

More information

Time-Varying Systems and Computations Lecture 6

Time-Varying Systems and Computations Lecture 6 Tme-Varyng Systems and Computatons Lecture 6 Klaus Depold 14. Januar 2014 The Kalman Flter The Kalman estmaton flter attempts to estmate the actual state of an unknown dscrete dynamcal system, gven nosy

More information

Kernel Methods and SVMs Extension

Kernel Methods and SVMs Extension Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general

More information

U-Pb Geochronology Practical: Background

U-Pb Geochronology Practical: Background U-Pb Geochronology Practcal: Background Basc Concepts: accuracy: measure of the dfference between an expermental measurement and the true value precson: measure of the reproducblty of the expermental result

More information

Lecture 4 Hypothesis Testing

Lecture 4 Hypothesis Testing Lecture 4 Hypothess Testng We may wsh to test pror hypotheses about the coeffcents we estmate. We can use the estmates to test whether the data rejects our hypothess. An example mght be that we wsh to

More information

/ n ) are compared. The logic is: if the two

/ n ) are compared. The logic is: if the two STAT C141, Sprng 2005 Lecture 13 Two sample tests One sample tests: examples of goodness of ft tests, where we are testng whether our data supports predctons. Two sample tests: called as tests of ndependence

More information

Computation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models

Computation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models Computaton of Hgher Order Moments from Two Multnomal Overdsperson Lkelhood Models BY J. T. NEWCOMER, N. K. NEERCHAL Department of Mathematcs and Statstcs, Unversty of Maryland, Baltmore County, Baltmore,

More information

Uncertainty in measurements of power and energy on power networks

Uncertainty in measurements of power and energy on power networks Uncertanty n measurements of power and energy on power networks E. Manov, N. Kolev Department of Measurement and Instrumentaton, Techncal Unversty Sofa, bul. Klment Ohrdsk No8, bl., 000 Sofa, Bulgara Tel./fax:

More information

STAT 511 FINAL EXAM NAME Spring 2001

STAT 511 FINAL EXAM NAME Spring 2001 STAT 5 FINAL EXAM NAME Sprng Instructons: Ths s a closed book exam. No notes or books are allowed. ou may use a calculator but you are not allowed to store notes or formulas n the calculator. Please wrte

More information

Comparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method

Comparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method Appled Mathematcal Scences, Vol. 7, 0, no. 47, 07-0 HIARI Ltd, www.m-hkar.com Comparson of the Populaton Varance Estmators of -Parameter Exponental Dstrbuton Based on Multple Crtera Decson Makng Method

More information

Hidden Markov Models

Hidden Markov Models Hdden Markov Models Namrata Vaswan, Iowa State Unversty Aprl 24, 204 Hdden Markov Model Defntons and Examples Defntons:. A hdden Markov model (HMM) refers to a set of hdden states X 0, X,..., X t,...,

More information

Appendix B: Resampling Algorithms

Appendix B: Resampling Algorithms 407 Appendx B: Resamplng Algorthms A common problem of all partcle flters s the degeneracy of weghts, whch conssts of the unbounded ncrease of the varance of the mportance weghts ω [ ] of the partcles

More information

Lecture 21: Numerical methods for pricing American type derivatives

Lecture 21: Numerical methods for pricing American type derivatives Lecture 21: Numercal methods for prcng Amercan type dervatves Xaoguang Wang STAT 598W Aprl 10th, 2014 (STAT 598W) Lecture 21 1 / 26 Outlne 1 Fnte Dfference Method Explct Method Penalty Method (STAT 598W)

More information

Winter 2008 CS567 Stochastic Linear/Integer Programming Guest Lecturer: Xu, Huan

Winter 2008 CS567 Stochastic Linear/Integer Programming Guest Lecturer: Xu, Huan Wnter 2008 CS567 Stochastc Lnear/Integer Programmng Guest Lecturer: Xu, Huan Class 2: More Modelng Examples 1 Capacty Expanson Capacty expanson models optmal choces of the tmng and levels of nvestments

More information

A Comparative Study for Estimation Parameters in Panel Data Model

A Comparative Study for Estimation Parameters in Panel Data Model A Comparatve Study for Estmaton Parameters n Panel Data Model Ahmed H. Youssef and Mohamed R. Abonazel hs paper examnes the panel data models when the regresson coeffcents are fxed random and mxed and

More information

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute

More information

( ) ( ) ( ) ( ) STOCHASTIC SIMULATION FOR BLOCKED DATA. Monte Carlo simulation Rejection sampling Importance sampling Markov chain Monte Carlo

( ) ( ) ( ) ( ) STOCHASTIC SIMULATION FOR BLOCKED DATA. Monte Carlo simulation Rejection sampling Importance sampling Markov chain Monte Carlo SOCHASIC SIMULAIO FOR BLOCKED DAA Stochastc System Analyss and Bayesan Model Updatng Monte Carlo smulaton Rejecton samplng Importance samplng Markov chan Monte Carlo Monte Carlo smulaton Introducton: If

More information

JAB Chain. Long-tail claims development. ASTIN - September 2005 B.Verdier A. Klinger

JAB Chain. Long-tail claims development. ASTIN - September 2005 B.Verdier A. Klinger JAB Chan Long-tal clams development ASTIN - September 2005 B.Verder A. Klnger Outlne Chan Ladder : comments A frst soluton: Munch Chan Ladder JAB Chan Chan Ladder: Comments Black lne: average pad to ncurred

More information

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore Sesson Outlne Introducton to classfcaton problems and dscrete choce models. Introducton to Logstcs Regresson. Logstc functon and Logt functon. Maxmum Lkelhood Estmator (MLE) for estmaton of LR parameters.

More information

4.3 Poisson Regression

4.3 Poisson Regression of teratvely reweghted least squares regressons (the IRLS algorthm). We do wthout gvng further detals, but nstead focus on the practcal applcaton. > glm(survval~log(weght)+age, famly="bnomal", data=baby)

More information

More metrics on cartesian products

More metrics on cartesian products More metrcs on cartesan products If (X, d ) are metrc spaces for 1 n, then n Secton II4 of the lecture notes we defned three metrcs on X whose underlyng topologes are the product topology The purpose of

More information

Chapter 14 Simple Linear Regression

Chapter 14 Simple Linear Regression Chapter 4 Smple Lnear Regresson Chapter 4 - Smple Lnear Regresson Manageral decsons often are based on the relatonshp between two or more varables. Regresson analss can be used to develop an equaton showng

More information

Relevance Vector Machines Explained

Relevance Vector Machines Explained October 19, 2010 Relevance Vector Machnes Explaned Trstan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introducton Ths document has been wrtten n an attempt to make Tppng s [1] Relevance Vector Machnes

More information

A New Method for Estimating Overdispersion. David Fletcher and Peter Green Department of Mathematics and Statistics

A New Method for Estimating Overdispersion. David Fletcher and Peter Green Department of Mathematics and Statistics A New Method for Estmatng Overdsperson Davd Fletcher and Peter Green Department of Mathematcs and Statstcs Byron Morgan Insttute of Mathematcs, Statstcs and Actuaral Scence Unversty of Kent, England Overvew

More information

Stat260: Bayesian Modeling and Inference Lecture Date: February 22, Reference Priors

Stat260: Bayesian Modeling and Inference Lecture Date: February 22, Reference Priors Stat60: Bayesan Modelng and Inference Lecture Date: February, 00 Reference Prors Lecturer: Mchael I. Jordan Scrbe: Steven Troxler and Wayne Lee In ths lecture, we assume that θ R; n hgher-dmensons, reference

More information

Small Area Interval Estimation

Small Area Interval Estimation .. Small Area Interval Estmaton Partha Lahr Jont Program n Survey Methodology Unversty of Maryland, College Park (Based on jont work wth Masayo Yoshmor, Former JPSM Vstng PhD Student and Research Fellow

More information

Chapter 6. Supplemental Text Material

Chapter 6. Supplemental Text Material Chapter 6. Supplemental Text Materal S6-. actor Effect Estmates are Least Squares Estmates We have gven heurstc or ntutve explanatons of how the estmates of the factor effects are obtaned n the textboo.

More information

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012 MLE and Bayesan Estmaton Je Tang Department of Computer Scence & Technology Tsnghua Unversty 01 1 Lnear Regresson? As the frst step, we need to decde how we re gong to represent the functon f. One example:

More information

LOW BIAS INTEGRATED PATH ESTIMATORS. James M. Calvin

LOW BIAS INTEGRATED PATH ESTIMATORS. James M. Calvin Proceedngs of the 007 Wnter Smulaton Conference S G Henderson, B Bller, M-H Hseh, J Shortle, J D Tew, and R R Barton, eds LOW BIAS INTEGRATED PATH ESTIMATORS James M Calvn Department of Computer Scence

More information

Tracking with Kalman Filter

Tracking with Kalman Filter Trackng wth Kalman Flter Scott T. Acton Vrgna Image and Vdeo Analyss (VIVA), Charles L. Brown Department of Electrcal and Computer Engneerng Department of Bomedcal Engneerng Unversty of Vrgna, Charlottesvlle,

More information

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements CS 750 Machne Learnng Lecture 5 Densty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square CS 750 Machne Learnng Announcements Homework Due on Wednesday before the class Reports: hand n before

More information

Homework Assignment 3 Due in class, Thursday October 15

Homework Assignment 3 Due in class, Thursday October 15 Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.

More information

Motion Perception Under Uncertainty. Hongjing Lu Department of Psychology University of Hong Kong

Motion Perception Under Uncertainty. Hongjing Lu Department of Psychology University of Hong Kong Moton Percepton Under Uncertanty Hongjng Lu Department of Psychology Unversty of Hong Kong Outlne Uncertanty n moton stmulus Correspondence problem Qualtatve fttng usng deal observer models Based on sgnal

More information

STAT 3008 Applied Regression Analysis

STAT 3008 Applied Regression Analysis STAT 3008 Appled Regresson Analyss Tutoral : Smple Lnear Regresson LAI Chun He Department of Statstcs, The Chnese Unversty of Hong Kong 1 Model Assumpton To quantfy the relatonshp between two factors,

More information

Difference Equations

Difference Equations Dfference Equatons c Jan Vrbk 1 Bascs Suppose a sequence of numbers, say a 0,a 1,a,a 3,... s defned by a certan general relatonshp between, say, three consecutve values of the sequence, e.g. a + +3a +1

More information

On the correction of the h-index for career length

On the correction of the h-index for career length 1 On the correcton of the h-ndex for career length by L. Egghe Unverstet Hasselt (UHasselt), Campus Depenbeek, Agoralaan, B-3590 Depenbeek, Belgum 1 and Unverstet Antwerpen (UA), IBW, Stadscampus, Venusstraat

More information

ANOMALIES OF THE MAGNITUDE OF THE BIAS OF THE MAXIMUM LIKELIHOOD ESTIMATOR OF THE REGRESSION SLOPE

ANOMALIES OF THE MAGNITUDE OF THE BIAS OF THE MAXIMUM LIKELIHOOD ESTIMATOR OF THE REGRESSION SLOPE P a g e ANOMALIES OF THE MAGNITUDE OF THE BIAS OF THE MAXIMUM LIKELIHOOD ESTIMATOR OF THE REGRESSION SLOPE Darmud O Drscoll ¹, Donald E. Ramrez ² ¹ Head of Department of Mathematcs and Computer Studes

More information

Joint Statistical Meetings - Biopharmaceutical Section

Joint Statistical Meetings - Biopharmaceutical Section Iteratve Ch-Square Test for Equvalence of Multple Treatment Groups Te-Hua Ng*, U.S. Food and Drug Admnstraton 1401 Rockvlle Pke, #200S, HFM-217, Rockvlle, MD 20852-1448 Key Words: Equvalence Testng; Actve

More information

CSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography

CSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography CSc 6974 and ECSE 6966 Math. Tech. for Vson, Graphcs and Robotcs Lecture 21, Aprl 17, 2006 Estmatng A Plane Homography Overvew We contnue wth a dscusson of the major ssues, usng estmaton of plane projectve

More information

COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS

COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS Avalable onlne at http://sck.org J. Math. Comput. Sc. 3 (3), No., 6-3 ISSN: 97-537 COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS

More information

4DVAR, according to the name, is a four-dimensional variational method.

4DVAR, according to the name, is a four-dimensional variational method. 4D-Varatonal Data Assmlaton (4D-Var) 4DVAR, accordng to the name, s a four-dmensonal varatonal method. 4D-Var s actually a drect generalzaton of 3D-Var to handle observatons that are dstrbuted n tme. The

More information

Psychology 282 Lecture #24 Outline Regression Diagnostics: Outliers

Psychology 282 Lecture #24 Outline Regression Diagnostics: Outliers Psychology 282 Lecture #24 Outlne Regresson Dagnostcs: Outlers In an earler lecture we studed the statstcal assumptons underlyng the regresson model, ncludng the followng ponts: Formal statement of assumptons.

More information