Gschlößl, Czado: Does a Gibbs sampler approach to spatial Poisson regression models outperform a single site MH sampler?

Size: px
Start display at page:

Download "Gschlößl, Czado: Does a Gibbs sampler approach to spatial Poisson regression models outperform a single site MH sampler?"

Transcription

1 Gschlößl, Czado: Does a Gbbs sampler approach to spatal Posson regresson models outperform a sngle ste MH sampler? Sonderforschungsberech 386, Paper 46 (25) Onlne unter: Projektpartner

2 Does a Gbbs sampler approach to spatal Posson regresson models outperform a sngle ste MH sampler? Susanne Gschlößl Clauda Czado December 6, 25 Abstract In ths paper we present and evaluate a Gbbs sampler for a Posson regresson model ncludng spatal effects. The approach s based on Frühwrth-Schnatter and Wagner (24b) who show that by data augmentaton usng the ntroducton of two sequences of latent varables a Posson regresson model can be transformed nto an approxmate normal lnear model. We show how ths methodology can be extended to spatal Posson regresson models and gve detals of the resultng Gbbs sampler. In partcular, the nfluence of model parametersaton and dfferent update strateges on the mxng of the MCMC chans s dscussed. The developed Gbbs samplers are analysed n two smulaton studes and appled to model the expected number of clams for polcyholders of a German car nsurance company. The mxng of the Gbbs samplers depends crucally on the model parametersaton and the update schemes. The best mxng s acheved when collapsed algorthms are used, reasonable low autocorrelatons for the spatal effects are obtaned n ths case. For the regresson effects however, autocorrelatons are rather hgh, especally for data wth very low heterogenety. For comparson a sngle component Metropols Hastngs algorthms s appled whch dsplays very good mxng for all components. Although the Metropols Hastngs sampler requres a hgher computatonal effort, t outperforms the Gbbs samplers whch would have to be run consderably longer n order to obtan the same precson of the parameters. Key words: block updates, collapsng, data augmentaton, Gbbs sampler, model parametersaton, spatal Posson count data Both at Center of Mathematcal Scences, Munch Unversty of Technology, Boltzmannstr.3, D Garchng, Germany,emal: susanne@ma.tum.de, cczado@ma.tum.de,

3 Introducton In ths paper we present a straghtforward Gbbs sampler for spatal Posson regresson models usng data augmentaton technques. In partcular, we am to nvestgate whether ths Gbbs sampler s found to be superor to a conventonal sngle ste Metropols Hastngs (MH) sampler. The ssue of model parametersaton and several update schemes for the parameters n the Gbbs sampler s thoroughly addressed. The performance of the developed Gbbs sampler schemes and the MH sampler s nvestgated n two smulaton studes as well as on real data from a German car nsurance company. Performance of the samplers s measured n the computatonal costs requred to obtan the same precson of the posteror means of the parameters. Snce the full condtonal dstrbutons of a spatal Posson regresson model do not follow any standard dstrbuton, often sngle ste MH steps are performed n a MCMC settng, see for example Dggle et al. (998), Dmakos and Frgess (22) or Gschlößl and Czado (25). However, ths requres the choce of approprate proposal dstrbutons n order to acheve reasonable acceptance rates and a good mxng of the MCMC chans. Advanced ndependence proposals, lke for example a normal proposal wth the same mode and nverse curvature at the mode as the target dstrbuton, can lead to hgh acceptance rates and low autocorrelatons but nvolve consderable computatonal efforts. Frühwrth-Schnatter and Wagner (24a) developed a Gbbs sampler for Posson regresson models for small counts. They show that by data augmentaton va the ntroducton of two sequences of latent varables a lnear normal model s obtaned. In Frühwrth-Schnatter and Wagner (24b) an applcaton of ths Gbbs sampler to state space models s gven, n Frühwrth-Schnatter and Wagner (24a) the same methodology s appled for standard Posson regresson models and Posson regresson models wth overdsperson. Usng smlar technques, a Gbbs sampler for logstc models s developed n Frühwrth-Schnatter and Waldl (24). The am of ths paper s to show that ths methodology can be extended to spatal Posson regresson models n a straghtforward manner allowng for a Gbbs update of both regresson parameters and spatal effects. Although we only consder spatal Posson data dstrbuted on regons n ths paper, the presented methodology could also be appled on geostatstcal Posson models, see Dggle et al. (998). It s well known, that mxng and convergence of the Gbbs sampler depends crucally on several mplementaton ssues, see for example Roberts and Sahu (997) for a detaled dscusson. Hgh autocorrelatons can be reduced by updatng several parameters n one block or usng collapsed algorthms, another mportant ssue s model parametersaton. Gelfand et al. (995) dscuss the effcency of centered and non-centered parametersatons for herarchcal normal lnear models, Papasplopoulos et al. (23) address parametersaton ssues for several classes of herarchcal models and ntroduce partally non-centered parametersatons. Chrstensen et al. (25) propose the standardzaton and orthogonalzaton of all model components leadng to effcent and robust MCMC algorthms. In ths paper both centered and non-centered model parametersatons are consdered, varous 2

4 algorthmc schemes, n partcular a jont block update of the ntercept and the spatal effects as well as collapsed algorthms, see Lu et al. (994), are dscussed. The performance of the samplers s examned and compared to a sngle ste MH sampler wth ndependence proposals n two smulaton studes. In the frst study, the samplers are appled on data wth both large and small spatal effects, whle the second study consders the nfluence of the data heterogenety on the performance of the samplers. The performance of the samplers s measured n the computatonal costs requred n order to obtan a certan precson of the posteror means of the regresson parameters and spatal effects. Ths s done by takng both the Monte Carlo error of the posteror means of the parameters and the computatonal tme requred for one teraton nto account. A very smlar approach for comparng the performance of MCMC samplers s conducted by Chrstensen and Waagepetersen (22). Among the Gbbs samplers collapsed algorthms perform best. In partcular for data wth small spatal effects, the Monte Carlo errors of the spatal effects are consderably reduced when collapsed samplers and model parametersatons wth non-centered scale or varance are used. The Monte Carlo errors of the regresson parameters however are rather hgh, especally for data wth low heterogenety. The MH ndependence sampler n contrast, exhbts very low Monte Carlo errors and good mxng for both regresson and spatal effects n all settngs. Although the MH sampler requres a hgher computatonal effort, ths drawback s compensated by the hgh precson of the posteror means of the parameters. In order to obtan the same precson the Gbbs samplers would have to be run consderably longer, dmnshng the computatonal advantage n comparson to the MH sampler. Therefore we have to conclude that the proposed Gbbs sampler for spatal Posson regresson models can not outperform a sngle ste MH sampler usng ndependence proposals. Ths paper s organzed as follows. In Secton 2 the spatal Posson regresson model s specfed and the two steps of the data augmentaton scheme are descrbed for ths specfc model. Detals on several algorthmc schemes for updatng the regresson and spatal effects are gven n Secton 3. In Secton 4 the developed Gbbs sampler schemes are examned and compared to a sngle component MH sampler wth ndependence proposals n two smulaton studys. We also apply the Gbbs samplers to model the expected number of clams n a real data set from a German car nsurance company. Secton 5 gves a summary and draws conclusons. 2 Data augmentaton and Gbbs sampler for spatal Posson regresson models We assume that observatons Y, =,..,n observed at J regons follow a Posson model y Posson(µ ). (2.) The mean µ s specfed by µ = t exp(z α) := t exp(x β + v γ) = t exp(x β + γ R() ) (2.2) 3

5 where z = (x,v ) denotes the covarate vector x = (,x,..,x p ) and the ncdence vector v = (v,..,v J ), f R() = j for the regons,.e. v j =, wth R(),..,J} denotng otherwse the regon of the -th observaton. Note, that we do not only observe one sngle but several observatons n each regon. Further α = (β,γ) denotes the vector of regresson parameters β = (β,β,..,β p ) and spatal random effects γ = (γ,..γ J ). By the ncluson of spatal effects we allow for geographcal dfferences n the J regons. The quantty t gves the exposure tme for the -th observaton. We assume a normal pror dstrbuton centered around zero wth a large standard devaton for the regresson parameters β, n partcular β N p+ (,V ) where V = τ 2 I p+ wth τ 2 =. Here N p (µ,σ) denotes the p-varate Normal dstrbuton wth mean µ and covarance matrx Σ. For the spatal effects a condtonal autoregressve (CAR) pror based on Petttt et al. (22) s used. In partcular, we assume γ ψ,σ 2 N J (,σ 2 Q ) where the elements of the precson matrx Q = (Q j ),,j =,..,J are gven by + ψ N = j Q j = ψ j, j otherwse. (2.3) We wrte j for regons and j whch are contguous and assume regons to be neghbours f they share a common border. N denotes the number of neghbours of regon. The spatal hyperparameter ψ determnes the degree of spatal dependence, for ψ = ndependence of the spatal effects s obtaned whereas for ψ the degree of spatal dependency ncreases. Note, that ths pror s a proper dstrbuton n contrast to the well known ntrnsc CAR model ntroduced by Besag and Kooperberg (995). Other proper spatal pror dstrbutons have been consdered, see for example Czado and Prokopenko (24) who use a modfcaton of Model (2.3) and Sun et al. (2). Therefore, we have a multvarate normal pror dstrbuton for the regresson and spatal parameters α whch s gven by wth Σ =. For the spatal hyperparameters θ = (ψ,σ 2 ) the proper pror dstrbutons ( V σ 2 Q ) ψ α θ N p++j (,Σ) (2.4) ( + ψ) 2 and σ 2 IGamma(,.5) 4

6 are assumed. The parametersaton of ths model descrbed by Observaton Equaton (2.2) and Pror Specfcaton (2.4) s called non-centered n the mean, snce the ntercept β appears n the observaton equaton, but not n the spatal pror formulaton. Other possble model parametersatons nclude parametersatons addtonally non-centered n the scale or varance of the spatal pror as well as a centered parametersaton, where the ntercept β only appears as the mean of the spatal pror. These parametersatons are summarzed n Table. For a summary on exstng parametersaton technques see for example Frühwrth-Schnatter (24). Intally, our nvestgatons are based on the non-centered mean parametersaton gven by (2.2) parametersaton spatal pror observaton equaton centered γ c N(β,σ 2 Q ) µ = t exp(x β + v γc ) non-centered mean γ N(,σ 2 Q ) µ = t exp(β + x β + v γ) non-centered mean and scale γ N(,Q ) µ = t exp(β + x β + σv γ ) non-centered mean and varance γ N(,I) µ = t exp(β + x β + σv Lγ ) where LL = Q Table : Spatal pror and observaton equaton for dfferent model parametersatons, where x := (x,..,x p ) and β := (β,..,β p ) and (2.4). Necessary changes when other parametersatons are used wll be ndcated specfcally. 2. Step : Introducton of hdden nter-arrval tmes The basc dea of the data augmentaton scheme developed by Frühwrth-Schnatter and Wagner (24b) s to regard the Posson observatons y, =,..,n, as the number of jumps of an unobserved Posson process wth ntensty µ wthn the unt nterval. The frst step of the data augmentaton conssts n the ntroducton of y + hdden nter-arrval tmes τ j,j =,..,y + for each observaton y. Usng that the nter-arrval tmes are ndependent and follow an exponental dstrbuton wth parameter µ, see for example Mkosch (24),.e. τ j α Exponental(µ ) = Exponental() µ, we obtan log τ j α = log t z α + ǫ j, ǫ j log(exponental()). (2.5) Denote by τ = τ j, =,..,n,j =,..,y + } the collecton of all nter-arrval tmes. Snce the posteror dstrbuton of α condtonal on τ s ndependent of y, condtonal on τ we are now dealng wth model (2.5) whch s lnear n the parameters α, but stll has a non-normal error term. 2.2 Step 2: Mxture approxmaton for error term The second step of the data augmentaton scheme elmnates the non-normalty of model (2.5). As shown by Frühwrth-Schnatter and Wagner (24b), the error term n (2.5) can be approx- 5

7 mated suffcently close to a normal dstrbuton by a mxture of fve normal dstrbutons,.e. p(ǫ j ) = exp(ǫ j exp(ǫ j )) 5 w r f N (ǫ j ;m r,s 2 r), where f N ( ;m r,s 2 r ) denotes the densty of the normal dstrbuton wth mean m r and varance s 2 r. Frühwrth-Schnatter and Wagner (24b) also gve the correspondng values for m r,s 2 r and the weghts w r. In the second step of the data augmentaton the component ndcators r j,..,5} are ntroduced as latent varables. Denotng the set of all component ndcators by R = r j, =,..,n,j =,..,y + }, we have condtonal on R r= log τ j α,r j = log t z α + m rj + ǫ j, ǫ j N(,s 2 r j ),.e. we are dealng wth a normal model whch s lnear n α now. Snce the pror dstrbuton π(α θ) s normal as well, the resultng posteror dstrbuton s multvarate normal and a Gbbs sampler can be appled. Note, that by performng ths data augmentaton we are no longer dealng wth n but wth n = (y + ) observatons. Therefore ths Gbbs Sampler s manly useful for count data wth small counts only, otherwse the data set mght get very large. 2.3 Algorthmc scheme The algorthmc scheme for the above Gbbs Sampler s the followng: Choose approprate startng values for the component ndcators R and the nter-arrval tmes τ. () sample regresson and spatal parameters α = (β,γ) gven τ,r,θ (2) sample spatal hyperparameters θ = (ψ,σ 2 ) gven α (3) sample the nter-arrval tmes τ j gven α,y (4) sample the component ndcators r j gven τ,α Step () conssts of samplng from a multvarate normal dstrbuton. Ths can be done n one block, however t mght be computatonally more effcent to perform an update n several smaller blocks. We wll consder several update strateges for step () n Secton 3 n more detal. The spatal hyperparameter ψ s updated usng a MH step, whereas σ 2 can be updated usng a Gbbs step. Steps (3) and (4), elaborated n Frühwrth-Schnatter and Wagner (24b), are descrbed n the Appendx A and B, for the choce of startng values see Appendx C. 6

8 3 Updatng schemes for the regresson and spatal parameters n the Gbbs Sampler For α several update schemes are possble and wll be dscussed n ths secton. For notatonal convenence we defne wth N := n = (y + ) τ = ( τ,.., τ N ) := (τ,..,τ,y +,τ 2,..,τ 2,y2 +,..,τ n,..,τ n,yn+), ǫ = ( ǫ,.., ǫ N ) := (ǫ,..,ǫ,y +,ǫ 2,..,ǫ 2,y2 +,..,ǫ n,..,ǫ n,yn+), m = ( m,.., m N ) := (m r,..,m,m r,y + r 2,..,m,..,m r2,y2 + r n,..,m rn,yn+ ) and s 2 = ( s 2,.., s 2 N) := (s 2 r,..,s 2 r,y +,s2 r 2,..,s 2 r 2,y2 +,..,s2 r n,..,s 2 r n,yn+ ). Let t = ( t,.., t N ) denote the vector where t s repeated y + tmes. Further defne ỹ = (ỹ,..,ỹ N ) := (log τ m + log t,..,log τ N m N + log t N ). Usng ths notaton we have accordng to (2.5) where z = z. z N ỹ α,r N( z α, s 2 ) s a N (p + + J)-matrx where z s repeated y + tmes. 3. Block update of α = (β, γ) For a jont update of the regresson parameters β and the spatal effects γ n one block we have to consder the full condtonal of α = (β,γ) whch s gven by p(α θ, τ, R) π(α θ) exp 2 exp 2 N = ( exp 2 s 2 (ỹ + z α)2) [ α Σ α + N = (ỹ + z α) 2]} s 2 [α Σ α α 2α µ α ]}, where Σ α := Σ + N = s z 2 z and µ α := N = s z 2 ỹ. Hence, α θ,τ,r N p++j (Σ α µ α,σ α ). 7

9 3.2 Separate update of β and γ The calculaton of the posteror covarance matrx Σ α n Secton 3. can be computatonally expensve f the number of regresson parameters and spatal effects s large as s the case n most spatal applcatons. Therefore t mght be more effcent to update β and γ n two separate blocks. The full condtonal dstrbutons of β and γ are gven by β γ,θ,τ,r N p+ (Σ β µ β,σ β ) and γ β,θ,τ,r N J(Σ γ µ γ,σ γ ). The explct formulas for Σ β, µ β, Σ γ and µ γ are gven n Table Block update of the ntercept β and γ (block) Due to dentfablty problems between the ntercept β and the spatal effects γ mxng and convergence s not very good when β and γ are updated n two separate blocks. Better results are acheved f a jont block update of β and γ s performed, whereas the remanng parameters β = (β,..,β p ) are stll updated n one separate block. Wth ths settng the posteror dstrbutons are gven by and β β,γ,θ,τ,r N p (Σ β µ β,σ β ) γ,β β,θ,τ,r N J+ (Σ γβ µ γβ,σ γβ ) wth Σ β, µ β, Σ γβ and µ γβ as gven n Table Collapsed algorthm for a model parametersaton wth a non-centered mean (coll) Another possblty s to use a collapsed algorthm. Ths means, that partcular components of the posteror are ntegrated out and an update based on the margnal dstrbuton s performed. In our context the jont posteror dstrbuton of β and γ can be wrtten as p(β,γ θ,τ,r) p(β τ,r)p(γ β,θ,τ,r) where p(β τ,r) = p(β,γ θ,τ,r)dγ s the margnalsed posteror densty of β wth γ ntegrated out. It s shown n the Appendx D that wth Σ col and µ col as gven n Table 2. β τ,r N p+ (Σ col µ col,σ col ) Step () n the algorthmc scheme presented n Secton 2.3 s then the followng for the collapsed algorthm: sample β from N p+ (Σ col µ col,σ col ) sample γ β, θ, τ, R as n Secton 3.2 8

10 Secton 3.2 Σ β := V + N = s x x 2 µ β := N = s x 2 (ỹ + γ R() ) Σ γ := σ 2 Q + N = s 2 ṽ ṽ µ γ := N = s 2 ṽ (ỹ + x β) 3.3 Σ β := V β + N = s x 2 β x β (block) µ β := N x β (ỹ + γ R() + β ) Σ γβ := ( = s 2 τ 2 Q σ 2 µ γβ := N = s 2 ) + N = s (,ṽ 2 )(,ṽ ) (,ṽ )(ỹ + x β β ) V β = τ 2 I p x β = ( x,.., x p ) 3.4 Σ col := τ 2 I + N = s x x 2 ( N = s ṽ x 2 ) A ( N = s ṽ x 2 ) (coll) µ col := ( N = s ṽ x 2 ) A ( N = s ṽ 2 ỹ ) N = s x 2 ỹ A := N = s ṽ 2 ṽ + σ 2 Q 3.5 Σ col := τ 2 I + N = s x x 2 (σ N = s ṽ x 2 ) (A ) (σ N = s ṽ x 2 ) (coll2) µ col := (σ N = s ṽ x 2 )(A ) (σ N = s ṽ 2 ỹ ) N = s x 2 ỹ A := σ 2 N = s ṽ 2 ṽ + Q Σ γ := σ2 N = s ṽ 2 ṽ + Q µ γ := σ N = s ṽ 2 (ỹ + x β) Σ σ := N = s (γ 2 R() )2 + τσ 2 µ σ := N (ỹ + x β) = γ R() s Σ col := τ 2 I + N = s x x 2 (σ N = s 2 (coll3) µ col := (σ N = s L ṽ 2 x )(A ) (σ N = s 2 L v x ) (A ) (σ N A := σ 2 N = s 2 L ṽ ṽ L + I Σ γ := σ 2 N = s 2 L ṽ ṽ L + I µ γ := σ N = s 2 L ṽ (ỹ + x β) = s 2 L v ỹ ) N = s x 2 ỹ Σ σ := N = s (γ ( N 2 = s L ṽ 2 ṽ L)γ + τσ 2 µ σ := N = v Lγ s (ỹ 2 + x β) 3.7 µ cent γ := β σ 2 Q N = s 2 ṽ (ỹ + x β β ) (centered) Σ β := σ 2 J,j= Q j + τ 2 µ β := σ 2 Qγ c L v x ) Table 2: Covarance and mean specfcatons for the update strateges n Sectons

11 3.5 Collapsed algorthm for a model parametersaton wth a non-centered mean and scale (coll2) Up to now, we only consdered models wth the non-centered mean parametersaton specfed by (2.2) and the spatal pror γ ψ,σ N J (,σ 2 Q ). In ths secton we consder a model where the pror of the spatal effects s not only non-centered n the mean, but n the scale as well,.e. the thrd model parametersaton gven n Table. By assumng γ ψ N J (,Q ), σ appears as an unknown parameter n the observaton equaton, n partcular we have µ = t exp(x β + σγ R() ). For ths parametersaton and π( ) denotng the pror dstrbutons, the jont posteror of β,γ,ψ and σ s gven by p(β,γ,ψ,σ ỹ,τ,r) exp 2 n = (ỹ + x β + σṽ γ ) 2} π(β)π(γ ψ)π(ψ)π(σ). s 2 Followng the lnes of Secton 3.4 we obtan for β the margnalzed posteror dstrbuton β σ,τ,r N p+ ((Σ col ) µ col,(σ col ) ). The full condtonal dstrbuton for γ s gven by γ β,τ,r,σ,ψ N J ((Σ γ) µ γ,(σ γ) ). The defntons of Σ col, µ col, Σ γ and µ γ can be found n Table 2. The spatal hyperparameter ψ s agan updated usng a MH step snce the full condtonal dstrbuton can not be sampled from drectly. For ths model parametersaton we choose a normal pror for σ, n partcular σ N(,τ 2 σ). Note, that σ s not restrcted to take postve values, leadng to nondentfablty, snce the same lkelhood results for (σ,γ ) and ( σ, γ ). However, as ponted out by Frühwrth- Schnatter (24), ths leads to an mproved mxng for models wth small scales σ 2 snce boundary problems for σ are avoded. The full condtonal dstrbuton of σ s then agan normal, n partcular see Table 2 for detals on Σ σ and µ σ. σ β,γ,τ,r N((Σ σ) µ σ,(σ σ) ), 3.6 Collapsed algorthm for a model parametersaton wth a non-centered mean and varance (coll3) In ths secton we consder the model parametersaton non-centered n both mean and varance, also gven n Table. In contrast to the non-centered parametersaton n scale only consdered n the prevous secton, we now assume the pror γ N J (,I).

12 The spatal structure ncorporated n the precson matrx Q s now moved to the observaton equaton gven by µ = t exp(x β + σv Lγ ), where L s a lower trangular matrx resultng from the Cholesky decomposton Q = LL. The resultng jont posteror dstrbuton of β,γ,ψ and σ s gven by p(β,γ,ψ,σ ỹ,τ,r) exp 2 n = The margnalzed posteror dstrbuton of β changes to (ỹ + x β + σṽ Lγ ) 2} π(β)π(γ )π(ψ)π(σ). s 2 β σ,τ,r N p+ ((Σ col ) µ col,(σ col ) ), the full condtonal dstrbuton of γ s gven by γ β,τ,r,σ,ψ N J ((Σ γ ) µ γ,(σ γ ) ), wth Σ col, µ col, Σ γ and µ γ as gven n Table 2. Whle ψ s agan updated usng a MH step, the full condtonal dstrbuton of σ s gven by σ β,γ,ψ,τ,r N((Σ σ ) µ σ,(σ σ ) ), see Table 2 for detals on Σ σ and µ σ. Here agan the normal pror σ N(,τ2 σ ) s assumed. 3.7 Centered CAR-Model (centered) Alternatvely, the centered spatal pror γ c β N(β,σ 2 Q ) wth β N(,τ 2 ) and β N(,τ 2 I p ) can be used. For ths model the posteror dstrbuton for β s the same as n Secton 3.3 but wth µ β replaced by N = s x 2 β (ỹ + γr() c ). The posteror dstrbuton for γ c s gven by γ c β,β,θ,τ,r,y N J (Σ γ µcent γ,σ γ ) where Σ γ s the same as n Secton 3.2 and µ cent γ s gven n Table 2. β s updated n an extra Gbbs step, n partcular wth Σ β and µ β defned as n Table 2. β β,γ,θ,τ,r,y N(Σ β µ β,σ β ) 4 Smulaton studes and applcaton We am to apply the developed Gbbs samplers to analyse the expected number of clams n a data set from a German car nsurance company. The data nclude 637 polcyholders n Bavara wth

13 full comprehensve car nsurance wthn one year and contan nformaton on several covarates lke age and gender of the polcyholders, klometers drven per year and the geographcal regon each polcyholder s lvng n. Bavara s dvded nto 96 regons. The varablty of these data s very small, 95% of the observatons are zero observatons, the hghest number of clams observed s only four. The data have been already analysed by Gschlößl and Czado (25) who consdered both a spatal Posson regresson model as well as spatal models takng overdsperson nto account. They show that the spatal effects are very small for these data and have no sgnfcant contrbuton to explanng the expected clam number. In ths secton, the performance of the Gbbs sampler schemes developed n Sectons 2 and 3 wll be examned on smulated data frst. For comparson, we addtonally use a sngle ste Metropols Hastngs algorthm for spatal Posson regresson models wth an ndependence proposal where both β and γ are updated component by component. In partcular, we use a t-dstrbuton wth 2 degrees of freedom as proposal whch has the same mode and nverse curvature at the mode as the target dstrbuton. The performance of the samplers s measured n terms of the computaton tme requred n order to obtan a certan precson of the estmated posteror means of the parameters. The posteror mean of a varable θ s gven by θ := R ˆθ j= j wth ˆθ j,j =,..,R denotng the MCMC terates of θ after burnn. The precson of θ s gven by the Monte Carlo standard error of θ whch s defned as σ MC ( θ) := σasy( θ) R where ( σasy( θ) 2 := V ar(θ) + 2 k= ) ρ k (θ) denotes the asymptotc varance of θ, V ar(θ) the sample varance and ρ k (θ) the autocorrelaton of the MCMC terates ˆθ,.., ˆθ R at lag k. The asymptotc varance wll be estmated usng the ntal monotone sequence estmator (see Geyer (992)), defned by 2m+ ˆσ asy( θ) 2 := V ˆ ar(θ)( + 2 j= ˆρ k (θ)), where m s chosen to be the largest nteger such that the sequence Γ m = ˆρ 2m (θ) + ˆρ 2m+ (θ) s postve and monotone. Here Vˆar(θ) := ˆγ, ˆρ k (θ) := ˆγ k, ˆγ k := R k ˆγ R j= (ˆθ j θ)(ˆθ j+k θ). We addtonally requre the estmated emprcal autocorrelatons ˆρ 2m+ (θ) to fall below.. In order to obtan a certan precson k, R = ˆσ2 asy samples are needed. Hence, the computaton k 2 tme requred to obtan a precson k for an algorthm wth computatonal costs m per teraton, s gven by R m. For a drect comparson of the Gbbs sampler schemes to the MH ndependence sampler we consder the computatonal costs relatve to the costs of the MH sampler requred to obtan the same precson of the posteror means of the parameters. Ths s gven by R rel m rel := ˆσ2 asy m ˆσ asy,nd 2 m nd, where ˆσ asy,nd 2 and m nd denote the estmated asymptotc varance and the computatonal costs for one teraton of the MH ndependence sampler. We consder two studes. In the frst study the nfluence of the sze of the spatal effects on mxng behavour s examned, whle n the second study the mpact of data heterogenety s 2

14 nvestgated. In both studes the Gbbs samplers descrbed n Sectons ,.e. the followng model parametersatons and update schemes are assumed: non-centered mean: block update of β β,γ and (β,γ) β gven n Secton 3.3 (block) collapsed algorthm gven n Secton 3.4 (coll) non-centered mean and scale: collapsed algorthm gven n Secton 3.5 (coll2) non-centered mean and varance: collapsed algorthm gven n Secton 3.6 (coll3) centered parametersaton: algorthm gven n Secton 3.7 (centered) In the followng we wll refer to these samplers as block, coll, coll2, coll3 and centered. 4. Computatonal costs Recall, that by usng the data augmentaton scheme descrbed above, we are no longer dealng wth n observatons, but wth N = n = (y + ) latent nter-arrval tmes τ j and mxture component ndcators r j. Both τ and R have to be updated, therefore the number of varables to sample from n each teraton s 2N +J +p+(+2 hyperparameters) n comparson to J +p+(+2 hyperparameters) varables n the MH ndependence sampler. The MH ndependence sampler n contrast requres the calculaton of the posteror mode and the nverse curvature at the posteror mode for each of the J + p + components n every teraton. The posteror mode may be obtaned usng the bsecton method for example. In our smulaton studes, except the sampler coll3, the Gbbs samplers are always faster than the MH ndependence sampler. However, the computatonal advantage of the Gbbs samplers depends on the complexty of the model. The computatonal costs m rel relatve to the costs of the MH sampler for one teraton are reported n Table 3. For the settng n Study wth 5 observatons, an ntercept and two covarates for example, the centered Gbbs sampler only takes.86 tmes as long as the MH ndependence sampler. For the settng n Study 2 wth a larger data set the centered Gbbs sampler even takes only.26 tmes as long. Among the Gbbs samplers the centered Gbbs sampler s the fastest, followed closely by the Gbbs sampler usng a block update. The collapsed Gbbs samplers noncentered n the mean (coll) and non-centered n mean and scale (coll2) requre slghtly more tme than the centered Gbbs sampler. The computatonal effort for the Gbbs sampler n the model parametersaton non-centered n the mean and the varance (coll3) however s more than twce as large. In ths algorthm a Cholesky decomposton of the precson matrx Q has to be performed n every teraton. 4.2 Study : Influence of the sze of the spatal effects We consder two smulated data sets of sze 5 wth y Posson(µ ), =,..,5. For both data sets the mean µ s specfed by µ = exp(β + x β + x 2 β 2 + γ R() ) 3

15 sampler Study Study 2 ndependence block centered coll.96.3 coll coll Table 3: Computaton tmes m rel for the dfferent samplers relatve to the MH ndependence sampler for the settngs n Study and Study 2. where x s an ndcator varable and x 2 a contnuous standardzed varable. The exposure s assumed to be t = for all observatons. We assume a smple spatal structure, namely regons on a grd. The spatal effects γ are generated accordng to the CAR pror γ N(,σ 2 Q ) wth spatal dependence parameter ψ = 3. For the frst smulated data set y we assume σ 2 = resultng n a range of [mn(γ)max(γ)] = [.86,.85] for the spatal effects, whereas for the second data set y 2 we take σ 2 =. resultng n a range of [mn(γ)max(γ)] = [.8,.8]. The Gbbs samplers block, coll, coll2, coll3 and centered as well as the ndependence MH sampler are run for 5 teratons, a burnn of teratons s taken. As descrbed above, the performance of the samplers s measured n terms of the Monte Carlo standard error of the posteror means of the parameters and the requred computaton tmes. Snce estmaton of the Monte Carlo error s based on the estmated emprcal autocorrelatons, ths quantty also depends on the mxng of the samplers. For a far comparson of the Monte Carlo error of the spatal effects the model parametersaton of each sampler has to be taken nto account. Therefore we compute the Monte Carlo error for β + γ for the MH ndependence sampler and the samplers block and collapsed, whle for the centered sampler the standard error of γ s consdered snce here the ntercept s the spatal pror mean and therefore already ncluded n γ. For the coll2 and coll3 samplers the Monte Carlo errors for β + σγ and β + σlγ, respectvely, are computed. In the left panel of Table 4, for each sampler the Monte Carlo standard errors and the performance relatve to the MH ndependence sampler R rel m rel are reported for the regresson parameters β,β 2 and the spatal effects n data set y. For the spatal effects the average error, taken over all J components, s gven. Addtonally plots of the emprcal estmated autocorrelatons are presented n Fgure. In the left panel the autocorrelatons for 25 of the spatal effects, n the rght panel autocorrelatons for the regresson effects are plotted. Mxng for all Gbbs samplers s reasonable well, n average autocorrelatons of the spatal effects are below. at a lag of about 6 to 8. The average Monte Carlo error for the spatal effects s around. for all Gbbs samplers. The Monte Carlo error of the regresson parameters however s lower for the collapsed Gbbs samplers, for the block and the centered Gbbs sampler especally the autocorrelatons of β decrease rather slowly. The ndependence MH sampler n contrast, dsplays the smallest Monte Carlo error for both 4

16 spatal effects and regresson parameters. In average the autocorrelatons of β + γ j are below. at a lag 3 already, the autocorrelatons for the regresson parameters decrease rapdly as well. Consderng the computatonal effort relatve to the MH ndependence sampler, gven by R rel m rel, the MH ndependence sampler outperforms the Gbbs samplers consderably. The computatonal effort requred to obtan the same precson of the posteror means of the spatal effects s more than 5 tmes as large for the Gbbs samplers compared to the ndependence sampler. Data set y Data set y 2 sampler spatal β β 2 spatal β β 2 effects effects ndependence block centered coll coll coll Table 4: Estmated ˆσ MC (upper row) for the regresson parameters β,β 2 and average estmated ˆσ MC for the spatal effects γ + β n the ndependence, block, coll sampler, γ n the centered, β + σγ n the coll2 and β + σlγ n the coll3 sampler, as well as R rel m rel (lower row) for all parameters for data set y and y 2 usng dfferent update strateges n Study. The correspondng results for data set y 2 wth small spatal effects are reported n the rght panel n Table 4, plots of the estmated emprcal autocorrelatons are gven n Fgure. Here, clearly the lowest precson and worst mxng s obtaned f the Gbbs sampler based on the centered model parametersaton s used. Ths confrms the results gven n Gelfand et al. (995). They show that for a herarchcal normal lnear model wth random effects the centered parametersaton s effcent f the varance of the random effects domnates the varance n the data. However, f the varance of the random effects s very small n contrast to the varablty of the data (as t s the case n data set y 2 ), hgh posteror correlatons result. For the block and partcularly the collapsed Gbbs samplers a consderably lower Monte Carlo error s obtaned. The average Monte Carlo error of the spatal effects n the collapsed sampler coll s almost as small as n the MH ndependence sampler. For the regresson effects however, the MH ndependence sampler 5

17 exhbts lower Monte Carlo standard errors. The computatonal costs R rel m rel relatve to the MH sampler, whch are requred to obtan the same precson of the posteror means of the parameters are greater for all Gbbs samplers for both spatal effects and regresson parameters. Hence, the ndependence sampler gves the best performance for data set y 2 as well. The varance of the two smulated data sets y and y 2 takes the values var(y ) =.5 and var(y 2 ) =.49. However, the varablty of our real data from a car nsurance company s very small, the varance of these data s only.5. Therefore we conduct a second smulaton study where we examne whether the heterogenety of the data nfluences the performance of the samplers. 4.3 Study 2: Influence of data heterogenety We smulate two data sets based on the desgn of the real data where, accordng to Gschlößl and Czado (25), eght covarates sgnfcant for explanng the expected clam number y were observed,.e. y Posson(µ ), =,..,637 wth µ = t exp(x β + γ R()). Here x = (,x,..,x 8 ) and x k,k =,..,8 are standardzed categorcal and metrcal covarates, the observaton specfc exposure t takes values up to one year. In ths settng we have 96 rregular regons n Bavara. The spatal effects γ agan are generated accordng to the CAR pror γ N(,σ 2 Q ) wth ψ = 8 and σ 2 =.. Ths results n small spatal effects wth a range of [.6.8],.e. spatal effects smlar to the ones observed n our real data set. For the frst data set y 3 the ntercept β s taken to be, whereas for the second data set y 4 we take β = 2.5. For the remanng regresson parameters the same values are assumed for both data sets. The resultng varances of y 3 and y 4 are V ar(y 3 ) =.46 and V ar(y 4 ) =.5,.e. data set y 4 has very low heterogenety and s close to our real data. The varance of data set y 3 s not partcularly hgh ether, but n comparson to data set y 4 we wll refer to ths data set as data wth hgh heterogenety. The block, centered, coll, coll2 and coll3 Gbbs samplers are run for 5 teratons, the frst teratons are dscarded for burnn. For comparson agan the MH ndependence sampler s appled. The Monte Carlo errors for the posteror means of the regresson parameters β,..,β 8, the spatal effects γ n the centered, β +γ n the non-centered mean, β +σγ n the non-centered mean and scale and β + σlγ n the non-centered mean and varance model parametersaton and the quanttes R rel m rel are reported n Table 5. For the hgh heterogenety data set y 3 the collapsed Gbbs samplers coll2 and coll3 exhbt the lowest Monte Carlo errors for the spatal effects among the Gbbs samplers. The sampler coll2 even only requres 38 % of the computatonal effort of the MH sampler n order to obtan the same precson for the spatal effects. The precson and autocorrelatons of the regresson effects however are consderably smaller n the ndependence sampler compared to all Gbbs samplers. The precson and autocorrelatons (see Fgure 2) of the regresson effects however are consderably smaller n the ndependence sampler compared to all Gbbs samplers. In order to acheve a hgh precson lke n the MH sampler 6

18 for all parameters, for each Gbbs sampler the maxmum relatve effort R rel m rel, occurrng for spatal and regresson parameters, s requred. Snce the maxmum values R rel m rel are consderably greater than for each Gbbs sampler, the MH sampler s clearly superor to the Gbbs samplers. The average Monte Carlo error for the spatal effects n data set y 4 wth low heterogenety s rather hgh for the three Gbbs sampler schemes block, centered and coll for both spatal effects and regresson parameters, the estmated emprcal autocorrelatons plotted n Fgure 2 decrease very slowly. Whle for the hgh heterogenety data y 3 the computatonal costs n order to obtan the same precson for the spatal effects of the block Gbbs sampler are only.65 tmes as large as of the MH sampler, for the data y 4 the performance of the Gbbs sampler s clearly worse wth R rel m rel = Results are mproved for the collapsed algorthms based on the model parametersatons non-centered n the scale (coll2) and n the varance (coll3). The sampler coll2 performs even better than the MH sampler (R rel m rel =.45). As ndcated n Secton 3.5, the model parametersaton wth non-centered scale s supposed to mprove mxng partcularly for models wth small scale σ 2 whch s the case for data sets y 3 and y 4. However, the Monte Carlo errors for the regresson parameters are rather hgh for all Gbbs samplers and n partcular consderably hgher than for the hgh heterogenety data y 3. The MH ndependence sampler n contrast exhbts a hgh precson for all parameters agan. Compared to data set y 3, the standard errors for all parameters resultng from the MH sampler are about twce as large for data set y 4, ths loss of precson however s much smaller than for the Gbbs samplers. Accordng to the performance measure R rel m rel for the regresson parameters, the MH sampler outperforms the Gbbs samplers consderably. For example, although the Gbbs sampler coll2 sampler only requres 3 % of the computaton tme of the MH sampler for one teraton (see Table 3), 3.33 (R rel m rel for β 2 ) tmes the effort of the MH sampler for data set y 4 would be needed n order to obtan for all parameters a precson comparable to the MH sampler. Note that, compared to the collapsed algorthm coll2, the collapsed algorthm coll3 does not dsplay sgnfcantly lower standard errors, nether n Study nor n Study 2. The addtonal computatonal effort requred for coll3 whch s more than twce as large as for coll2, see Table 3, does not pay off. 4.4 Applcaton to car nsurance data Fnally we apply the dscussed Gbbs samplers as well as the ndependence MH sampler on the car nsurance data set descrbed at the begnnng of ths secton. The Monte Carlo errors for the posteror means of the regresson and the spatal effects as well as the correspondng values of R rel m rel are reported n Table 6. Smlar results as for data set y 4 whch s very close to our real data, are observed. In partcular for the regresson parameters, the performance of all Gbbs samplers s consderably worse than the performance of the MH ndependence sampler. When usng the non-centered scale and varance parametersatons at least for the spatal effects reasonable low errors are obtaned, however, accordng to the relatve effort R rel m rel the MH sampler s stll superor. 7

19 data sampler spatal β β 2 β 3 β 4 β 5 β 6 β 7 β 8 effects nd y 3 block centered coll coll coll nd y 4 block centered coll coll coll Table 5: Estmated ˆσ MC 2 (upper row) for the regresson parameters β,..,β 8 and estmated average ˆσ MC 2 for the spatal effects γ + β n the ndependence, block, coll sampler, γ n the centered, β + σγ n the coll2 and β + σlγ n the coll3 sampler, as well as R rel m rel (lower row) for data set y 3 and y 4 usng dfferent update strateges n Study 2. 5 Summary and conclusons We have presented a new MCMC methodology for spatal Posson regresson models, extendng the approach by Frühwrth-Schnatter and Wagner (24b). Usng data augmentaton we have shown that a straghtforward Gbbs sampler for spatal Posson models s avalable. Several update schemes lke a jont block update of the ntercept and the spatal effects as well as collapsed algorthms have been dscussed. Further we have addressed the ssue of model parametersaton, centered as well as non-centered model parametersatons n the mean, the scale and the varance have been consdered. The performance of the Gbbs sampler based on dfferent model parametersatons and update schemes has been compared to a sngle ste MH ndependence sampler on smulated and real data. Performance s measured n terms of the computatonal 8

20 sampler spatal β β 2 β 3 β 4 β 5 β 6 β 7 β 8 effects ndependence block centered coll coll coll Table 6: Estmated ˆσ MC (upper row) for the regresson parameters β,..,β 8 and average estmated ˆσ MC for the spatal effects γ + β n the ndependence, block, coll sampler, γ n the centered, β + σγ n the coll2 and β + σlγ n the coll3 sampler, as well as R rel m rel (lower row) for the car nsurance data usng dfferent update strateges. costs requred n order to obtan the same precson of the posteror means of the parameters. For data whch are not too homogeneous, the Gbbs samplers dsplay good mxng and reasonable small Monte Carlo errors. In partcular for data wth small spatal random effects, the performance s mproved when collapsed Gbbs samplers are used, whle the centered parametersaton s not very effcent any more n ths case. The MH ndependence sampler however exhbts the smallest Monte Carlo errors for all parameters for data wth both small and large spatal effects. Takng addtonally the requred computaton tmes of the samplers nto account, the MH sampler gves the best performance. For data wth low heterogenety the Monte Carlo errors ncrease sgnfcantly for all Gbbs samplers, mxng of the samplers s much worse. The MH sampler n contrast also mxes well for low heterogenety data, the precson of the posteror means of the parameters s consderably hgher than for the Gbbs samplers. Consderng the computaton tmes of the samplers and the requred MCMC teratons n order to obtan the same precson for all parameters, the MH sampler clearly outperforms the Gbbs samplers for low heterogenety data. Smlar results are observed for the real data whch also dsplay low heterogenety. In the lterature varous approaches for MCMC estmaton n spatal Posson models are provded. Knorr-Held and Rue (22) dscuss effcent block samplng MH algorthms for Markov random feld models n dsease mappng, based on the methodology developed n Rue (2). Haran et al. (23) study MH algorthms wth proposal dstrbutons based on Structured MCMC, ntroduced by Sargent et al. (2), for spatal Posson models, whle Chrstensen et al. (25) dscuss Langevn-Hastngs updates n spatal GLMM s. Rue et al. (24) present non-gaussan 9

21 approxmatons to hdden Markov random felds and gve applcatons n dsease mappng and geostatstcal models. These methods have been found to be superor to a conventonal MH sampler only performng ndvdual updates of the parameters. Therefore, snce a sngle ste MH sampler clearly outperformed the Gbbs samplers developed n ths paper, a comparson of the Gbbs samplers to these methods seems to be unnecessary. However, the performance of the Gbbs samplers mght be mproved by applyng the reparametersaton technques presented n Chrstensen et al. (25), whch s subject of current research. A Samplng the nter-arrval tmes Gven y and α, the nter-arrval tmes for dfferent observatons =,.., n are ndependent. For fxed however, τ,..,τ,y + are stochastcally dependent, but ndependent of the component ndcators R. The nter-arrval tmes τ,..,τ y are ndependent of α and only depend on the number of jumps, whereas τ,y + depends on the model parameters. Usng ths we have p(τ y,α,r) = n p(τ,y + y,α,τ,..,τ y )p(τ,..,τ y y ) = Gven y = n, the n arrval tmes of a Posson process are dstrbuted as the order statstcs of n U([, ]) dstrbuted random varables, see for example Mkosch (24). The last nterarrval tme τ,y +, gven y,τ,..,τ y, s exponentally dstrbuted wth mean µ = t exp(z α) condtonally on beng greater than y j= τ j. Usng the lack of memory property of the exponental dstrbuton ths corresponds to samplng τ,y + from an exponental dstrbuton wth mean µ plus an offset y j= τ j. Therefore the nter-arrval tmes can be sampled as follows: If y > sample y random numbers u,..,u y U([,]) sort these random numbers: u,(),..,u,(y ) defne τ j as the ncrements τ j = u,(j) u,(j ), j =,..,y where u j,() := sample τ,y + = y j= τ j + ζ, where ζ Exponental(µ ) If y = sample τ = + ζ, where ζ Exponental(µ ) B Samplng the component ndcators The component ndcators R are mutually ndependent gven τ,α, therefore p(r τ,α) = n = y + j= p(r j τ j,α). Further p(r j = k τ j,α) = p(r j = k,τ j,α) p(τ j,α) = p(τ j r j = k,α)p(r j = k) p(τ j α) p(τ j r j = k,α)w k (2.) 2

22 snce w k = p(r j = k). Snce log τ j α,r j N( log µ +m rj,s 2 r j ), τ j s log normal dstrbuted,.e. p(τ j r j = k,α) [ exp ( log(τj ) + log µ m ) k 2 ]. s k τ j 2 s k r j can therefore be sampled from the dscrete dstrbuton (2.) wth fve categores. C Startng values Startng values for the component ndcators r j are obtaned by drawng random numbers from to 5. For τ j startng values are generated accordng to the samplng procedure descrbed n Appendx A. For observatons equal to zero we sample ζ Exponental(.), for observatons greater than zero ζ Exponental(y ), as suggested n Frühwrth-Schnatter and Wagner (24b). D Detals on algorthm n Secton 3.4 For the collapsed algorthm n Secton 3.4 we consder p(β τ,r) = p(β,γ θ,τ,r)dγ. We have p(β,γ θ,τ,r) exp [ N 2 = = exp 2 s 2 [ β τ 2 Iβ + exp [γ ( N 2 = := c(β) exp 2 ]} (ỹ + x β + ṽ γ)2 + γ σ 2 Qγ + β τ 2 Iβ s 2 N = (ỹ + x β) 2]} s 2 ṽ ṽ + σ 2 Q )γ N + 2γ ]} ṽ (ỹ + x β) [ γ Aγ + 2γ a ]} s 2 = (4.2) where A := N = s 2 ṽ ṽ + σ 2 Q. Further exp [ ]} γ Aγ + 2γ a 2 exp [ ]} γ Aγ + 2γ A(A a) + (A a) A(A a) (A a) A(A a) 2 exp [ ]} (γ + A a) A(γ + A a) (A a) A(A a) 2 and therefore exp [ ]} γ Aγ + 2γ a dγ (2π) J 2 A } 2 exp 2 2 (A a) A(A a) } exp 2 (A a) A(A a) (4.3) 2

23 From (4.2) and (4.3) t then follows that Fnally, wth p(β,γ θ,τ,r)dγ } c(β)exp 2 (A a) A(A a) exp [ β (τ 2 I + 2 N = x x )β + 2β s 2 N = s 2 ]} x ỹ a A a a A a = ( N ṽ ỹ + = s 2 = s 2 N ) A ( ṽ x β N ṽ ỹ + = s 2 = s 2 = s 2 = N ) ṽ x β β ( N ) A ( N ṽ x ṽ x )β + 2β ( N ) A ( N ṽ x ) ṽ ỹ = s 2 s 2 = s 2 t follows that p(β τ, R) exp [ β ( τ 2 I + 2 = s 2 N N x x ( N ṽ x ) A ( ṽ x )β ) s 2 = 2β ( N N ( ṽ x ) A ( ṽ ỹ ) = s 2 = N = s 2 x ỹ )]}, s 2 = s 2.e. β τ,r N(Σ col µ col,σ col ) wth N Σ col := τ 2 N I + x x N ( ṽ x ) A ( ṽ x ) = s 2 = s 2 = s 2 and Acknowledgement N N µ col := ( ṽ x ) A ( ṽ ỹ ) s 2 = = s 2 N x ỹ. We would lke to thank Sylva Frühwrth-Schnatter for frutful dscussons and helpful comments and suggestons. The frst author s supported by a doctoral fellowshp wthn the Graduertenkolleg Angewandte Algorthmsche Mathematk, whle the second author s supported by Sonderforschungsberech 386 Statstsche Analyse Dskreter Strukturen, both sponsored by the Deutsche Forschungsgemenschaft. = s 2 22

24 References Besag, J. and C. Kooperberg (995). On condtonal and ntrnsc autoregressons. Bometrka 82, Chrstensen, O., G. Roberts, and M. Sköld (25). Robust MCMC methods for spatal GLMM s. to appear n Journal of Computatonal and Graphcal Statstcs. Chrstensen, O. and R. Waagepetersen (22). Bayesan predcton of spatal count data usng generalzed lnear mxed models. Bometrcs 58, Czado, C. and S. Prokopenko (24). Modelng transport mode decsons usng herarchcal bnary spatal regresson models wth cluster effects. Dscusson paper 46, SFB 386 Statstsche Analyse dskreter Strukturen. Dggle, P. J., J. A. Tawn, and R. A. Moyeed (998). Model-based geostatstcs. J. Roy. Statst. Soc. Ser. C 47(3), Wth dscusson and a reply by the authors. Dmakos, X. and A. Frgess (22). Bayesan premum ratng wth latent structure. Scandnavan Actuaral Journal (3), Frühwrth-Schnatter, S. (24). Effcent Bayesan parameter estmaton. In S. K. A.C. Harvey and N. Shephard (Eds.), State space and unobserved component models, pp Cambrdge: Cambrdge Unv. Press. Frühwrth-Schnatter, S. and H. Wagner (24a). Data augmentaton and Gbbs samplng for regresson models of small counts. IFAS Research Paper Seres Frühwrth-Schnatter, S. and H. Wagner (24b). Gbbs samplng for parameter-drven models of tme seres of small counts wth applcatons to state space modellng. IFAS Research Paper Seres 24-. Frühwrth-Schnatter, S. and H. Waldl (24). Data augmentaton and Gbbs samplng for logstc models. IFAS Research Paper Seres Gelfand, A., S. Sahu, and B.P.Carln (995). Effcent parametrsatons for normal lnear mxed models. Bometrka 82 (3), Geyer, C. (992). Practcal Markov Chan Monte Carlo. Statstcal Scence 7(4), Gschlößl, S. and C. Czado (25). Modellng count data wth overdsperson and spatal effects. Dscusson paper 42, SFB 386 Statstsche Analyse dskreter Strukturen. Haran, M., J. Hodges, and B. Carln (23). Acceleratng computaton n Markov random feld models for spatal data va structured MCMC. Journal of Computatonal and Graphcal Statstcs 2(2), Knorr-Held, L. and H. Rue (22). On block updatng n Markov random feld models for dsease mappng. Scandnavan Journal of Statstcs. Theory and Applcatons 29,

25 Lu, J., W. Wong, and A. Kong (994). Covarance structure of the Gbbs sampler wth applcatons to the comparsons of estmators and augmentaton schemes. Bometrka 8 (), Mkosch, T. (24). Non-Lfe Insurance Mathematcs. An Introducton wth Stochastc Processes. New York: Sprnger. Papasplopoulos, O., G. O. Roberts, and M. Sköld (23). Non-centered parameterzatons for herarchcal models and data augmentaton. In Bayesan statstcs, 7 (Tenerfe, 22), pp New York: Oxford Unv. Press. Wth a dscusson by Alan E. Gelfand, Ole F. Chrstensen and Darren J. Wlknson, and a reply by the authors. Petttt, A., I. Wer, and A. Hart (22). A condtonal autoregressve Gaussan process for rregularly spaced multvarate data wth applcaton to modellng large sets of bnary data. Statstcs and Computng 2 (4), Roberts, G. and S. Sahu (997). Updatng schemes, correlaton structure, blockng and parameterzaton for the Gbbs sampler. Journal of the Royal Statstcal Socety, B 59 (2), Rue, H. (2). Fast samplng of Gaussan Markov random felds. Journal of the Royal Statstcal Socety. Seres B. Statstcal Methodology 63(4), Rue, H., I. Stensland, and S. Erland (24). Approxmatng hdden Gaussan Markov random felds. Journal of the Royal Statstcal Socety. Seres B. Statstcal Methodology 66(4), Sargent, D., J. Hodges, and B. Carln (2). Structured Markov Chan Monte Carlo. Journal of Computatonal and Graphcal Statstcs 9, Sun, D., R. K. Tsutakawa, H. Km, and Z. He (2). Bayesan analyss of mortalty rates wth dsease maps. Statstcs n Medcne 9,

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4) I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes

More information

Negative Binomial Regression

Negative Binomial Regression STATGRAPHICS Rev. 9/16/2013 Negatve Bnomal Regresson Summary... 1 Data Input... 3 Statstcal Model... 3 Analyss Summary... 4 Analyss Optons... 7 Plot of Ftted Model... 8 Observed Versus Predcted... 10 Predctons...

More information

Comparison of Regression Lines

Comparison of Regression Lines STATGRAPHICS Rev. 9/13/2013 Comparson of Regresson Lnes Summary... 1 Data Input... 3 Analyss Summary... 4 Plot of Ftted Model... 6 Condtonal Sums of Squares... 6 Analyss Optons... 7 Forecasts... 8 Confdence

More information

A Robust Method for Calculating the Correlation Coefficient

A Robust Method for Calculating the Correlation Coefficient A Robust Method for Calculatng the Correlaton Coeffcent E.B. Nven and C. V. Deutsch Relatonshps between prmary and secondary data are frequently quantfed usng the correlaton coeffcent; however, the tradtonal

More information

Computation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models

Computation of Higher Order Moments from Two Multinomial Overdispersion Likelihood Models Computaton of Hgher Order Moments from Two Multnomal Overdsperson Lkelhood Models BY J. T. NEWCOMER, N. K. NEERCHAL Department of Mathematcs and Statstcs, Unversty of Maryland, Baltmore County, Baltmore,

More information

Outline for today. Markov chain Monte Carlo. Example: spatial statistics (Christensen and Waagepetersen 2001)

Outline for today. Markov chain Monte Carlo. Example: spatial statistics (Christensen and Waagepetersen 2001) Markov chan Monte Carlo Rasmus Waagepetersen Department of Mathematcs Aalborg Unversty Denmark November, / Outlne for today MCMC / Condtonal smulaton for hgh-dmensonal U: Markov chan Monte Carlo Consder

More information

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010 Parametrc fractonal mputaton for mssng data analyss Jae Kwang Km Survey Workng Group Semnar March 29, 2010 1 Outlne Introducton Proposed method Fractonal mputaton Approxmaton Varance estmaton Multple mputaton

More information

On an Extension of Stochastic Approximation EM Algorithm for Incomplete Data Problems. Vahid Tadayon 1

On an Extension of Stochastic Approximation EM Algorithm for Incomplete Data Problems. Vahid Tadayon 1 On an Extenson of Stochastc Approxmaton EM Algorthm for Incomplete Data Problems Vahd Tadayon Abstract: The Stochastc Approxmaton EM (SAEM algorthm, a varant stochastc approxmaton of EM, s a versatle tool

More information

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems Numercal Analyss by Dr. Anta Pal Assstant Professor Department of Mathematcs Natonal Insttute of Technology Durgapur Durgapur-713209 emal: anta.bue@gmal.com 1 . Chapter 5 Soluton of System of Lnear Equatons

More information

How its computed. y outcome data λ parameters hyperparameters. where P denotes the Laplace approximation. k i k k. Andrew B Lawson 2013

How its computed. y outcome data λ parameters hyperparameters. where P denotes the Laplace approximation. k i k k. Andrew B Lawson 2013 Andrew Lawson MUSC INLA INLA s a relatvely new tool that can be used to approxmate posteror dstrbutons n Bayesan models INLA stands for ntegrated Nested Laplace Approxmaton The approxmaton has been known

More information

Simulated Power of the Discrete Cramér-von Mises Goodness-of-Fit Tests

Simulated Power of the Discrete Cramér-von Mises Goodness-of-Fit Tests Smulated of the Cramér-von Mses Goodness-of-Ft Tests Steele, M., Chaselng, J. and 3 Hurst, C. School of Mathematcal and Physcal Scences, James Cook Unversty, Australan School of Envronmental Studes, Grffth

More information

Durban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications

Durban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications Durban Watson for Testng the Lack-of-Ft of Polynomal Regresson Models wthout Replcatons Ruba A. Alyaf, Maha A. Omar, Abdullah A. Al-Shha ralyaf@ksu.edu.sa, maomar@ksu.edu.sa, aalshha@ksu.edu.sa Department

More information

Chapter 13: Multiple Regression

Chapter 13: Multiple Regression Chapter 13: Multple Regresson 13.1 Developng the multple-regresson Model The general model can be descrbed as: It smplfes for two ndependent varables: The sample ft parameter b 0, b 1, and b are used to

More information

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction ECONOMICS 5* -- NOTE (Summary) ECON 5* -- NOTE The Multple Classcal Lnear Regresson Model (CLRM): Specfcaton and Assumptons. Introducton CLRM stands for the Classcal Lnear Regresson Model. The CLRM s also

More information

Markov Chain Monte Carlo Lecture 6

Markov Chain Monte Carlo Lecture 6 where (x 1,..., x N ) X N, N s called the populaton sze, f(x) f (x) for at least one {1, 2,..., N}, and those dfferent from f(x) are called the tral dstrbutons n terms of mportance samplng. Dfferent ways

More information

Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement

Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement Markov Chan Monte Carlo MCMC, Gbbs Samplng, Metropols Algorthms, and Smulated Annealng 2001 Bonformatcs Course Supplement SNU Bontellgence Lab http://bsnuackr/ Outlne! Markov Chan Monte Carlo MCMC! Metropols-Hastngs

More information

Hidden Markov Models & The Multivariate Gaussian (10/26/04)

Hidden Markov Models & The Multivariate Gaussian (10/26/04) CS281A/Stat241A: Statstcal Learnng Theory Hdden Markov Models & The Multvarate Gaussan (10/26/04) Lecturer: Mchael I. Jordan Scrbes: Jonathan W. Hu 1 Hdden Markov Models As a bref revew, hdden Markov models

More information

Bayesian predictive Configural Frequency Analysis

Bayesian predictive Configural Frequency Analysis Psychologcal Test and Assessment Modelng, Volume 54, 2012 (3), 285-292 Bayesan predctve Confgural Frequency Analyss Eduardo Gutérrez-Peña 1 Abstract Confgural Frequency Analyss s a method for cell-wse

More information

Comparison on Modelling the Relative. Risk Estimation: Bayesian Study

Comparison on Modelling the Relative. Risk Estimation: Bayesian Study Appled Mathematcal Scences, Vol. 4, 00, no. 54, 663-68 Comparson on Modellng the Relatve Rsk Estmaton: Bayesan Study Rafda M. Elobad and Noor Akma I. Laboratory of Appled & Computatonal Statstcs Insttute

More information

Chapter 11: Simple Linear Regression and Correlation

Chapter 11: Simple Linear Regression and Correlation Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests

More information

Uncertainty as the Overlap of Alternate Conditional Distributions

Uncertainty as the Overlap of Alternate Conditional Distributions Uncertanty as the Overlap of Alternate Condtonal Dstrbutons Olena Babak and Clayton V. Deutsch Centre for Computatonal Geostatstcs Department of Cvl & Envronmental Engneerng Unversty of Alberta An mportant

More information

x = , so that calculated

x = , so that calculated Stat 4, secton Sngle Factor ANOVA notes by Tm Plachowsk n chapter 8 we conducted hypothess tests n whch we compared a sngle sample s mean or proporton to some hypotheszed value Chapter 9 expanded ths to

More information

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models ECO 452 -- OE 4: Probt and Logt Models ECO 452 -- OE 4 Maxmum Lkelhood Estmaton of Bnary Dependent Varables Models: Probt and Logt hs note demonstrates how to formulate bnary dependent varables models

More information

Psychology 282 Lecture #24 Outline Regression Diagnostics: Outliers

Psychology 282 Lecture #24 Outline Regression Diagnostics: Outliers Psychology 282 Lecture #24 Outlne Regresson Dagnostcs: Outlers In an earler lecture we studed the statstcal assumptons underlyng the regresson model, ncludng the followng ponts: Formal statement of assumptons.

More information

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition)

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition) Count Data Models See Book Chapter 11 2 nd Edton (Chapter 10 1 st Edton) Count data consst of non-negatve nteger values Examples: number of drver route changes per week, the number of trp departure changes

More information

Composite Hypotheses testing

Composite Hypotheses testing Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter

More information

NUMERICAL DIFFERENTIATION

NUMERICAL DIFFERENTIATION NUMERICAL DIFFERENTIATION 1 Introducton Dfferentaton s a method to compute the rate at whch a dependent output y changes wth respect to the change n the ndependent nput x. Ths rate of change s called the

More information

Chapter 6. Supplemental Text Material

Chapter 6. Supplemental Text Material Chapter 6. Supplemental Text Materal S6-. actor Effect Estmates are Least Squares Estmates We have gven heurstc or ntutve explanatons of how the estmates of the factor effects are obtaned n the textboo.

More information

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore Sesson Outlne Introducton to classfcaton problems and dscrete choce models. Introducton to Logstcs Regresson. Logstc functon and Logt functon. Maxmum Lkelhood Estmator (MLE) for estmaton of LR parameters.

More information

since [1-( 0+ 1x1i+ 2x2 i)] [ 0+ 1x1i+ assumed to be a reasonable approximation

since [1-( 0+ 1x1i+ 2x2 i)] [ 0+ 1x1i+ assumed to be a reasonable approximation Econ 388 R. Butler 204 revsons Lecture 4 Dummy Dependent Varables I. Lnear Probablty Model: the Regresson model wth a dummy varables as the dependent varable assumpton, mplcaton regular multple regresson

More information

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 30 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 2 Remedes for multcollnearty Varous technques have

More information

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Analyss of Varance and Desgn of Experment-I MODULE VII LECTURE - 3 ANALYSIS OF COVARIANCE Dr Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Any scentfc experment s performed

More information

Number of cases Number of factors Number of covariates Number of levels of factor i. Value of the dependent variable for case k

Number of cases Number of factors Number of covariates Number of levels of factor i. Value of the dependent variable for case k ANOVA Model and Matrx Computatons Notaton The followng notaton s used throughout ths chapter unless otherwse stated: N F CN Y Z j w W Number of cases Number of factors Number of covarates Number of levels

More information

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands Content. Inference on Regresson Parameters a. Fndng Mean, s.d and covarance amongst estmates.. Confdence Intervals and Workng Hotellng Bands 3. Cochran s Theorem 4. General Lnear Testng 5. Measures of

More information

Hidden Markov Models

Hidden Markov Models Hdden Markov Models Namrata Vaswan, Iowa State Unversty Aprl 24, 204 Hdden Markov Model Defntons and Examples Defntons:. A hdden Markov model (HMM) refers to a set of hdden states X 0, X,..., X t,...,

More information

Non-Mixture Cure Model for Interval Censored Data: Simulation Study ABSTRACT

Non-Mixture Cure Model for Interval Censored Data: Simulation Study ABSTRACT Malaysan Journal of Mathematcal Scences 8(S): 37-44 (2014) Specal Issue: Internatonal Conference on Mathematcal Scences and Statstcs 2013 (ICMSS2013) MALAYSIAN JOURNAL OF MATHEMATICAL SCIENCES Journal

More information

Statistics for Economics & Business

Statistics for Economics & Business Statstcs for Economcs & Busness Smple Lnear Regresson Learnng Objectves In ths chapter, you learn: How to use regresson analyss to predct the value of a dependent varable based on an ndependent varable

More information

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:

More information

2016 Wiley. Study Session 2: Ethical and Professional Standards Application

2016 Wiley. Study Session 2: Ethical and Professional Standards Application 6 Wley Study Sesson : Ethcal and Professonal Standards Applcaton LESSON : CORRECTION ANALYSIS Readng 9: Correlaton and Regresson LOS 9a: Calculate and nterpret a sample covarance and a sample correlaton

More information

Comparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method

Comparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method Appled Mathematcal Scences, Vol. 7, 0, no. 47, 07-0 HIARI Ltd, www.m-hkar.com Comparson of the Populaton Varance Estmators of -Parameter Exponental Dstrbuton Based on Multple Crtera Decson Makng Method

More information

Linear Approximation with Regularization and Moving Least Squares

Linear Approximation with Regularization and Moving Least Squares Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...

More information

BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS. M. Krishna Reddy, B. Naveen Kumar and Y. Ramu

BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS. M. Krishna Reddy, B. Naveen Kumar and Y. Ramu BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS M. Krshna Reddy, B. Naveen Kumar and Y. Ramu Department of Statstcs, Osmana Unversty, Hyderabad -500 007, Inda. nanbyrozu@gmal.com, ramu0@gmal.com

More information

Conjugacy and the Exponential Family

Conjugacy and the Exponential Family CS281B/Stat241B: Advanced Topcs n Learnng & Decson Makng Conjugacy and the Exponental Famly Lecturer: Mchael I. Jordan Scrbes: Bran Mlch 1 Conjugacy In the prevous lecture, we saw conjugate prors for the

More information

Lecture Notes on Linear Regression

Lecture Notes on Linear Regression Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume

More information

STATS 306B: Unsupervised Learning Spring Lecture 10 April 30

STATS 306B: Unsupervised Learning Spring Lecture 10 April 30 STATS 306B: Unsupervsed Learnng Sprng 2014 Lecture 10 Aprl 30 Lecturer: Lester Mackey Scrbe: Joey Arthur, Rakesh Achanta 10.1 Factor Analyss 10.1.1 Recap Recall the factor analyss (FA) model for lnear

More information

Chapter 5 Multilevel Models

Chapter 5 Multilevel Models Chapter 5 Multlevel Models 5.1 Cross-sectonal multlevel models 5.1.1 Two-level models 5.1.2 Multple level models 5.1.3 Multple level modelng n other felds 5.2 Longtudnal multlevel models 5.2.1 Two-level

More information

Discussion of Extensions of the Gauss-Markov Theorem to the Case of Stochastic Regression Coefficients Ed Stanek

Discussion of Extensions of the Gauss-Markov Theorem to the Case of Stochastic Regression Coefficients Ed Stanek Dscusson of Extensons of the Gauss-arkov Theorem to the Case of Stochastc Regresson Coeffcents Ed Stanek Introducton Pfeffermann (984 dscusses extensons to the Gauss-arkov Theorem n settngs where regresson

More information

4DVAR, according to the name, is a four-dimensional variational method.

4DVAR, according to the name, is a four-dimensional variational method. 4D-Varatonal Data Assmlaton (4D-Var) 4DVAR, accordng to the name, s a four-dmensonal varatonal method. 4D-Var s actually a drect generalzaton of 3D-Var to handle observatons that are dstrbuted n tme. The

More information

Chapter 20 Duration Analysis

Chapter 20 Duration Analysis Chapter 20 Duraton Analyss Duraton: tme elapsed untl a certan event occurs (weeks unemployed, months spent on welfare). Survval analyss: duraton of nterest s survval tme of a subject, begn n an ntal state

More information

A New Method for Estimating Overdispersion. David Fletcher and Peter Green Department of Mathematics and Statistics

A New Method for Estimating Overdispersion. David Fletcher and Peter Green Department of Mathematics and Statistics A New Method for Estmatng Overdsperson Davd Fletcher and Peter Green Department of Mathematcs and Statstcs Byron Morgan Insttute of Mathematcs, Statstcs and Actuaral Scence Unversty of Kent, England Overvew

More information

Chapter 8 Indicator Variables

Chapter 8 Indicator Variables Chapter 8 Indcator Varables In general, e explanatory varables n any regresson analyss are assumed to be quanttatve n nature. For example, e varables lke temperature, dstance, age etc. are quanttatve n

More information

STAT 511 FINAL EXAM NAME Spring 2001

STAT 511 FINAL EXAM NAME Spring 2001 STAT 5 FINAL EXAM NAME Sprng Instructons: Ths s a closed book exam. No notes or books are allowed. ou may use a calculator but you are not allowed to store notes or formulas n the calculator. Please wrte

More information

Chapter 15 - Multiple Regression

Chapter 15 - Multiple Regression Chapter - Multple Regresson Chapter - Multple Regresson Multple Regresson Model The equaton that descrbes how the dependent varable y s related to the ndependent varables x, x,... x p and an error term

More information

Generalized Linear Methods

Generalized Linear Methods Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set

More information

Homework Assignment 3 Due in class, Thursday October 15

Homework Assignment 3 Due in class, Thursday October 15 Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.

More information

Appendix B. The Finite Difference Scheme

Appendix B. The Finite Difference Scheme 140 APPENDIXES Appendx B. The Fnte Dfference Scheme In ths appendx we present numercal technques whch are used to approxmate solutons of system 3.1 3.3. A comprehensve treatment of theoretcal and mplementaton

More information

The Geometry of Logit and Probit

The Geometry of Logit and Probit The Geometry of Logt and Probt Ths short note s meant as a supplement to Chapters and 3 of Spatal Models of Parlamentary Votng and the notaton and reference to fgures n the text below s to those two chapters.

More information

RELIABILITY ASSESSMENT

RELIABILITY ASSESSMENT CHAPTER Rsk Analyss n Engneerng and Economcs RELIABILITY ASSESSMENT A. J. Clark School of Engneerng Department of Cvl and Envronmental Engneerng 4a CHAPMAN HALL/CRC Rsk Analyss for Engneerng Department

More information

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U) Econ 413 Exam 13 H ANSWERS Settet er nndelt 9 deloppgaver, A,B,C, som alle anbefales å telle lkt for å gøre det ltt lettere å stå. Svar er gtt . Unfortunately, there s a prntng error n the hnt of

More information

Lecture 12: Discrete Laplacian

Lecture 12: Discrete Laplacian Lecture 12: Dscrete Laplacan Scrbe: Tanye Lu Our goal s to come up wth a dscrete verson of Laplacan operator for trangulated surfaces, so that we can use t n practce to solve related problems We are mostly

More information

Notes on Frequency Estimation in Data Streams

Notes on Frequency Estimation in Data Streams Notes on Frequency Estmaton n Data Streams In (one of) the data streamng model(s), the data s a sequence of arrvals a 1, a 2,..., a m of the form a j = (, v) where s the dentty of the tem and belongs to

More information

Time-Varying Systems and Computations Lecture 6

Time-Varying Systems and Computations Lecture 6 Tme-Varyng Systems and Computatons Lecture 6 Klaus Depold 14. Januar 2014 The Kalman Flter The Kalman estmaton flter attempts to estmate the actual state of an unknown dscrete dynamcal system, gven nosy

More information

Lecture 3. Ax x i a i. i i

Lecture 3. Ax x i a i. i i 18.409 The Behavor of Algorthms n Practce 2/14/2 Lecturer: Dan Spelman Lecture 3 Scrbe: Arvnd Sankar 1 Largest sngular value In order to bound the condton number, we need an upper bound on the largest

More information

Introduction to Generalized Linear Models

Introduction to Generalized Linear Models INTRODUCTION TO STATISTICAL MODELLING TRINITY 00 Introducton to Generalzed Lnear Models I. Motvaton In ths lecture we extend the deas of lnear regresson to the more general dea of a generalzed lnear model

More information

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA

4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA 4 Analyss of Varance (ANOVA) 5 ANOVA 51 Introducton ANOVA ANOVA s a way to estmate and test the means of multple populatons We wll start wth one-way ANOVA If the populatons ncluded n the study are selected

More information

Limited Dependent Variables

Limited Dependent Variables Lmted Dependent Varables. What f the left-hand sde varable s not a contnuous thng spread from mnus nfnty to plus nfnty? That s, gven a model = f (, β, ε, where a. s bounded below at zero, such as wages

More information

Econometrics of Panel Data

Econometrics of Panel Data Econometrcs of Panel Data Jakub Mućk Meetng # 8 Jakub Mućk Econometrcs of Panel Data Meetng # 8 1 / 17 Outlne 1 Heterogenety n the slope coeffcents 2 Seemngly Unrelated Regresson (SUR) 3 Swamy s random

More information

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Regression Analysis

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Regression Analysis Resource Allocaton and Decson Analss (ECON 800) Sprng 04 Foundatons of Regresson Analss Readng: Regresson Analss (ECON 800 Coursepak, Page 3) Defntons and Concepts: Regresson Analss statstcal technques

More information

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family IOSR Journal of Mathematcs IOSR-JM) ISSN: 2278-5728. Volume 3, Issue 3 Sep-Oct. 202), PP 44-48 www.osrjournals.org Usng T.O.M to Estmate Parameter of dstrbutons that have not Sngle Exponental Famly Jubran

More information

ANOVA. The Observations y ij

ANOVA. The Observations y ij ANOVA Stands for ANalyss Of VArance But t s a test of dfferences n means The dea: The Observatons y j Treatment group = 1 = 2 = k y 11 y 21 y k,1 y 12 y 22 y k,2 y 1, n1 y 2, n2 y k, nk means: m 1 m 2

More information

Appendix B: Resampling Algorithms

Appendix B: Resampling Algorithms 407 Appendx B: Resamplng Algorthms A common problem of all partcle flters s the degeneracy of weghts, whch conssts of the unbounded ncrease of the varance of the mportance weghts ω [ ] of the partcles

More information

Testing for seasonal unit roots in heterogeneous panels

Testing for seasonal unit roots in heterogeneous panels Testng for seasonal unt roots n heterogeneous panels Jesus Otero * Facultad de Economía Unversdad del Rosaro, Colomba Jeremy Smth Department of Economcs Unversty of arwck Monca Gulett Aston Busness School

More information

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6 Department of Quanttatve Methods & Informaton Systems Tme Seres and Ther Components QMIS 30 Chapter 6 Fall 00 Dr. Mohammad Zanal These sldes were modfed from ther orgnal source for educatonal purpose only.

More information

Notes prepared by Prof Mrs) M.J. Gholba Class M.Sc Part(I) Information Technology

Notes prepared by Prof Mrs) M.J. Gholba Class M.Sc Part(I) Information Technology Inverse transformatons Generaton of random observatons from gven dstrbutons Assume that random numbers,,, are readly avalable, where each tself s a random varable whch s unformly dstrbuted over the range(,).

More information

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results. Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson

More information

18.1 Introduction and Recap

18.1 Introduction and Recap CS787: Advanced Algorthms Scrbe: Pryananda Shenoy and Shjn Kong Lecturer: Shuch Chawla Topc: Streamng Algorthmscontnued) Date: 0/26/2007 We contnue talng about streamng algorthms n ths lecture, ncludng

More information

EEE 241: Linear Systems

EEE 241: Linear Systems EEE : Lnear Systems Summary #: Backpropagaton BACKPROPAGATION The perceptron rule as well as the Wdrow Hoff learnng were desgned to tran sngle layer networks. They suffer from the same dsadvantage: they

More information

Lecture 10 Support Vector Machines II

Lecture 10 Support Vector Machines II Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed

More information

Semiparametric geographically weighted generalised linear modelling in GWR 4.0

Semiparametric geographically weighted generalised linear modelling in GWR 4.0 Semparametrc geographcally weghted generalsed lnear modellng n GWR 4.0 T. Nakaya 1, A. S. Fotherngham 2, M. Charlton 2, C. Brunsdon 3 1 Department of Geography, Rtsumekan Unversty, 56-1 Tojn-kta-mach,

More information

Supplementary Notes for Chapter 9 Mixture Thermodynamics

Supplementary Notes for Chapter 9 Mixture Thermodynamics Supplementary Notes for Chapter 9 Mxture Thermodynamcs Key ponts Nne major topcs of Chapter 9 are revewed below: 1. Notaton and operatonal equatons for mxtures 2. PVTN EOSs for mxtures 3. General effects

More information

Web Appendix B Estimation. We base our sampling procedure on the method of data augmentation (e.g., Tanner and Wong,

Web Appendix B Estimation. We base our sampling procedure on the method of data augmentation (e.g., Tanner and Wong, Web Appendx B Estmaton Lkelhood and Data Augmentaton We base our samplng procedure on the method of data augmentaton (eg anner and Wong 987) here e treat the unobserved ndvdual choces as parameters Specfcally

More information

This column is a continuation of our previous column

This column is a continuation of our previous column Comparson of Goodness of Ft Statstcs for Lnear Regresson, Part II The authors contnue ther dscusson of the correlaton coeffcent n developng a calbraton for quanttatve analyss. Jerome Workman Jr. and Howard

More information

Errors for Linear Systems

Errors for Linear Systems Errors for Lnear Systems When we solve a lnear system Ax b we often do not know A and b exactly, but have only approxmatons  and ˆb avalable. Then the best thng we can do s to solve ˆx ˆb exactly whch

More information

Relevance Vector Machines Explained

Relevance Vector Machines Explained October 19, 2010 Relevance Vector Machnes Explaned Trstan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introducton Ths document has been wrtten n an attempt to make Tppng s [1] Relevance Vector Machnes

More information

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number

More information

BIO Lab 2: TWO-LEVEL NORMAL MODELS with school children popularity data

BIO Lab 2: TWO-LEVEL NORMAL MODELS with school children popularity data Lab : TWO-LEVEL NORMAL MODELS wth school chldren popularty data Purpose: Introduce basc two-level models for normally dstrbuted responses usng STATA. In partcular, we dscuss Random ntercept models wthout

More information

Lecture 4 Hypothesis Testing

Lecture 4 Hypothesis Testing Lecture 4 Hypothess Testng We may wsh to test pror hypotheses about the coeffcents we estmate. We can use the estmates to test whether the data rejects our hypothess. An example mght be that we wsh to

More information

Chapter 3 Describing Data Using Numerical Measures

Chapter 3 Describing Data Using Numerical Measures Chapter 3 Student Lecture Notes 3-1 Chapter 3 Descrbng Data Usng Numercal Measures Fall 2006 Fundamentals of Busness Statstcs 1 Chapter Goals To establsh the usefulness of summary measures of data. The

More information

A Particle Filter Algorithm based on Mixing of Prior probability density and UKF as Generate Importance Function

A Particle Filter Algorithm based on Mixing of Prior probability density and UKF as Generate Importance Function Advanced Scence and Technology Letters, pp.83-87 http://dx.do.org/10.14257/astl.2014.53.20 A Partcle Flter Algorthm based on Mxng of Pror probablty densty and UKF as Generate Importance Functon Lu Lu 1,1,

More information

UNR Joint Economics Working Paper Series Working Paper No Further Analysis of the Zipf Law: Does the Rank-Size Rule Really Exist?

UNR Joint Economics Working Paper Series Working Paper No Further Analysis of the Zipf Law: Does the Rank-Size Rule Really Exist? UNR Jont Economcs Workng Paper Seres Workng Paper No. 08-005 Further Analyss of the Zpf Law: Does the Rank-Sze Rule Really Exst? Fungsa Nota and Shunfeng Song Department of Economcs /030 Unversty of Nevada,

More information

Target tracking example Filtering: Xt. (main interest) Smoothing: X1: t. (also given with SIS)

Target tracking example Filtering: Xt. (main interest) Smoothing: X1: t. (also given with SIS) Target trackng example Flterng: Xt Y1: t (man nterest) Smoothng: X1: t Y1: t (also gven wth SIS) However as we have seen, the estmate of ths dstrbuton breaks down when t gets large due to the weghts becomng

More information

BAYESIAN CURVE FITTING USING PIECEWISE POLYNOMIALS. Dariusz Biskup

BAYESIAN CURVE FITTING USING PIECEWISE POLYNOMIALS. Dariusz Biskup BAYESIAN CURVE FITTING USING PIECEWISE POLYNOMIALS Darusz Bskup 1. Introducton The paper presents a nonparaetrc procedure for estaton of an unknown functon f n the regresson odel y = f x + ε = N. (1) (

More information

COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS

COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS Avalable onlne at http://sck.org J. Math. Comput. Sc. 3 (3), No., 6-3 ISSN: 97-537 COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS

More information

Lecture 17 : Stochastic Processes II

Lecture 17 : Stochastic Processes II : Stochastc Processes II 1 Contnuous-tme stochastc process So far we have studed dscrete-tme stochastc processes. We studed the concept of Makov chans and martngales, tme seres analyss, and regresson analyss

More information

CS-433: Simulation and Modeling Modeling and Probability Review

CS-433: Simulation and Modeling Modeling and Probability Review CS-433: Smulaton and Modelng Modelng and Probablty Revew Exercse 1. (Probablty of Smple Events) Exercse 1.1 The owner of a camera shop receves a shpment of fve cameras from a camera manufacturer. Unknown

More information

Global Sensitivity. Tuesday 20 th February, 2018

Global Sensitivity. Tuesday 20 th February, 2018 Global Senstvty Tuesday 2 th February, 28 ) Local Senstvty Most senstvty analyses [] are based on local estmates of senstvty, typcally by expandng the response n a Taylor seres about some specfc values

More information

Multinomial logit regression

Multinomial logit regression 07/0/6 Multnomal logt regresson Introducton We now turn our attenton to regresson models for the analyss of categorcal dependent varables wth more than two response categores: Y car owned (many possble

More information

Statistics II Final Exam 26/6/18

Statistics II Final Exam 26/6/18 Statstcs II Fnal Exam 26/6/18 Academc Year 2017/18 Solutons Exam duraton: 2 h 30 mn 1. (3 ponts) A town hall s conductng a study to determne the amount of leftover food produced by the restaurants n the

More information

PHYS 450 Spring semester Lecture 02: Dealing with Experimental Uncertainties. Ron Reifenberger Birck Nanotechnology Center Purdue University

PHYS 450 Spring semester Lecture 02: Dealing with Experimental Uncertainties. Ron Reifenberger Birck Nanotechnology Center Purdue University PHYS 45 Sprng semester 7 Lecture : Dealng wth Expermental Uncertantes Ron Refenberger Brck anotechnology Center Purdue Unversty Lecture Introductory Comments Expermental errors (really expermental uncertantes)

More information

Basically, if you have a dummy dependent variable you will be estimating a probability.

Basically, if you have a dummy dependent variable you will be estimating a probability. ECON 497: Lecture Notes 13 Page 1 of 1 Metropoltan State Unversty ECON 497: Research and Forecastng Lecture Notes 13 Dummy Dependent Varable Technques Studenmund Chapter 13 Bascally, f you have a dummy

More information