Robust Bayesian Regression

Size: px
Start display at page:

Download "Robust Bayesian Regression"

Transcription

1 Readings: Hoff Chapter 9, West JRSSB 1984, Fúquene, Pérez & Pericchi 2015 Duke University November 17, 2016

2 Body Fat Data: Intervals w/ All Data Response % Body Fat and Predictor Waist Circumference 95% confidence and prediction intervals for bodyfat.lm 95% confidence and prediction intervals for bodyfat.lm2 xbar xbar Bodyfat 30 observed fit conf int pred int Bodyfat observed fit conf int pred int xbar Abdomen xbar Abdomen Which analysis do we use? with Case 39 or not or something different?

3 Cook s Distance Leverage Standardized residuals Cook's distance Residuals vs Leverage

4 Options for Handling Influential Cases Are there scientific grounds for eliminating the case?

5 Options for Handling Influential Cases Are there scientific grounds for eliminating the case? Test if the case has a different mean than population

6 Options for Handling Influential Cases Are there scientific grounds for eliminating the case? Test if the case has a different mean than population Report results with and without the case

7 Options for Handling Influential Cases Are there scientific grounds for eliminating the case? Test if the case has a different mean than population Report results with and without the case Model Averaging to Account for Model Uncertainty?

8 Options for Handling Influential Cases Are there scientific grounds for eliminating the case? Test if the case has a different mean than population Report results with and without the case Model Averaging to Account for Model Uncertainty? Full model Y = Xβ + I n δ + ɛ

9 Options for Handling Influential Cases Are there scientific grounds for eliminating the case? Test if the case has a different mean than population Report results with and without the case Model Averaging to Account for Model Uncertainty? Full model Y = Xβ + I n δ + ɛ 2 n submodels γ i = 0 δ i = 0 If γ i = 1 then case i has a different mean mean shift outliers.

10 Mean Shift = Variance Inflation Model Y = Xβ + I n δ + ɛ Prior δ i γ i N(0, V σ 2 γ i ) γ i Ber(π) Then ɛ i given σ 2 is independent of δ i and ɛ i ɛ i + δ i σ 2 { N(0, σ 2 ) wp (1 π) N(0, σ 2 (1 + V )) wp π Model Y = Xβ + ɛ variance inflation V + 1 = K = 7 in the paper by Hoeting et al. package BMA

11 Simultaneous Outlier and Variable Selection MC3.REG(all.y = bodyfat$bodyfat, all.x = as.matrix(bodyfat$abdom num.its = 10000, outliers = TRUE) Model parameters: PI=0.02 K=7 nu=2.58 lambda=0.28 phi= models were selected Best 5 models (cumulative posterior probability = ): prob model 1 model 2 model 3 model 4 model 5 variables all.x 1 x x x x x outliers x x. x x x.. x post prob

12 Change Error Assumptions Y i ind t(ν, α + βx i, 1/φ)

13 Change Error Assumptions Y i L(α, β, φ) ind t(ν, α + βx i, 1/φ) n ( φ 1/2 1 + φ(y i α βx i ) 2 ν i=1 ) (ν+1) 2

14 Change Error Assumptions Y i L(α, β, φ) ind t(ν, α + βx i, 1/φ) n ( φ 1/2 1 + φ(y i α βx i ) 2 ν i=1 ) (ν+1) 2 Use Prior p(α, β, φ) 1/φ

15 Change Error Assumptions Y i L(α, β, φ) ind t(ν, α + βx i, 1/φ) n ( φ 1/2 1 + φ(y i α βx i ) 2 ν i=1 ) (ν+1) 2 Use Prior p(α, β, φ) 1/φ Posterior distribution p(α, β, φ Y ) φ n/2 1 n (1 + φ(y i α βx i ) 2 i=1 ν ) (ν+1) 2

16 Change Error Assumptions Y i L(α, β, φ) ind t(ν, α + βx i, 1/φ) n ( φ 1/2 1 + φ(y i α βx i ) 2 ν i=1 ) (ν+1) 2 Use Prior p(α, β, φ) 1/φ Posterior distribution p(α, β, φ Y ) φ n/2 1 n (1 + φ(y i α βx i ) 2 i=1 ν ) (ν+1) 2

17 Bounded Influence - West 1984 (and references within) Treat σ 2 as given, then influence of individual observations on the posterior distribution of β in the model where E[Y i ] = x T i β is investigated through the score function:

18 Bounded Influence - West 1984 (and references within) Treat σ 2 as given, then influence of individual observations on the posterior distribution of β in the model where E[Y i ] = x T i β is investigated through the score function: d d n log p(β Y) = dβ dβ log p(β) + xg(y i x T i β) i=1

19 Bounded Influence - West 1984 (and references within) Treat σ 2 as given, then influence of individual observations on the posterior distribution of β in the model where E[Y i ] = x T i β is investigated through the score function: where d d n log p(β Y) = dβ dβ log p(β) + xg(y i x T i β) g(ɛ) = d log p(ɛ) dɛ is the influence function of the error distribution (unimodal, continuous, differentiable, symmetric) i=1

20 Bounded Influence - West 1984 (and references within) Treat σ 2 as given, then influence of individual observations on the posterior distribution of β in the model where E[Y i ] = x T i β is investigated through the score function: where d d n log p(β Y) = dβ dβ log p(β) + xg(y i x T i β) g(ɛ) = d log p(ɛ) dɛ is the influence function of the error distribution (unimodal, continuous, differentiable, symmetric) i=1 An outlying observation y j is accommodated if the posterior distribution for p(β Y (i) ) converges to p(β Y) for all β as Y i. Requires error models with influence functions that go to zero such as the Student t (O Hagan, 1979)

21 Choice of df Score function for t with α degrees of freedom has turning points at ± α

22 Choice of df Score function for t with α degrees of freedom has turning points at ± α g(eps, 9) eps g (ɛ) is negative when ɛ 2 > α (standardized errors)

23 Choice of df Score function for t with α degrees of freedom has turning points at ± α g(eps, 9) eps g (ɛ) is negative when ɛ 2 > α (standardized errors) Contribution of observation to information matrix is negative and the observation is doubtful

24 Choice of df Score function for t with α degrees of freedom has turning points at ± α g(eps, 9) eps g (ɛ) is negative when ɛ 2 > α (standardized errors) Contribution of observation to information matrix is negative and the observation is doubtful Suggest taking α = 8 or α = 9 to reject errors larger than 8 or 3 sd.

25 Choice of df Score function for t with α degrees of freedom has turning points at ± α g(eps, 9) eps g (ɛ) is negative when ɛ 2 > α (standardized errors) Contribution of observation to information matrix is negative and the observation is doubtful Suggest taking α = 8 or α = 9 to reject errors larger than 8 or 3 sd.

26 Scale-Mixtures of Normal Representation Z i iid t(ν, 0, σ 2 )

27 Scale-Mixtures of Normal Representation Z i iid t(ν, 0, σ 2 ) Z i λ i ind N(0, σ 2 /λ i )

28 Scale-Mixtures of Normal Representation Z i iid t(ν, 0, σ 2 ) Z i λ i ind N(0, σ 2 /λ i ) λ i iid G(ν/2, ν/2)

29 Scale-Mixtures of Normal Representation Z i iid t(ν, 0, σ 2 ) Z i λ i ind N(0, σ 2 /λ i ) λ i iid G(ν/2, ν/2) Integrate out latent λ s to obtain marginal distribution.

30 Latent Variable Model Y i α, β, φ, λ ind 1 N(α + βx i, ) φλ i

31 Latent Variable Model Y i α, β, φ, λ ind 1 N(α + βx i, ) φλ i λ i iid G(ν/2, ν/2)

32 Latent Variable Model Y i α, β, φ, λ ind 1 N(α + βx i, ) φλ i λ i iid G(ν/2, ν/2) p(α, β, φ) 1/φ

33 Latent Variable Model Y i α, β, φ, λ Joint Posterior Distribution: ind 1 N(α + βx i, ) φλ i λ i iid G(ν/2, ν/2) p(α, β, φ) 1/φ

34 Latent Variable Model Y i α, β, φ, λ ind 1 N(α + βx i, ) φλ i λ i iid G(ν/2, ν/2) p(α, β, φ) 1/φ Joint Posterior Distribution: p((α, β, φ, λ 1,..., λ n Y ) { φ n/2 exp φ } λi (y i α βx i ) 2 2

35 Latent Variable Model Y i α, β, φ, λ ind 1 N(α + βx i, ) φλ i λ i iid G(ν/2, ν/2) p(α, β, φ) 1/φ Joint Posterior Distribution: p((α, β, φ, λ 1,..., λ n Y ) { φ n/2 exp φ } λi (y i α βx i ) 2 2 φ 1

36 Latent Variable Model Y i α, β, φ, λ ind 1 N(α + βx i, ) φλ i λ i iid G(ν/2, ν/2) p(α, β, φ) 1/φ Joint Posterior Distribution: p((α, β, φ, λ 1,..., λ n Y ) { φ n/2 exp φ } λi (y i α βx i ) 2 2 φ 1 n i=1 λ ν/2 1 i exp( λ i ν/2)

37 Programs BUGS: Bayesian inference Using Gibbs Sampling

38 Programs BUGS: Bayesian inference Using Gibbs Sampling WinBUGS is the Windows implementation

39 Programs BUGS: Bayesian inference Using Gibbs Sampling WinBUGS is the Windows implementation can be called from R with R2WinBUGS package

40 Programs BUGS: Bayesian inference Using Gibbs Sampling WinBUGS is the Windows implementation can be called from R with R2WinBUGS package can be run on any intel-based computer using VMware, wine

41 Programs BUGS: Bayesian inference Using Gibbs Sampling WinBUGS is the Windows implementation can be called from R with R2WinBUGS package can be run on any intel-based computer using VMware, wine OpenBUGS open source version of WinBUGS

42 Programs BUGS: Bayesian inference Using Gibbs Sampling WinBUGS is the Windows implementation can be called from R with R2WinBUGS package can be run on any intel-based computer using VMware, wine OpenBUGS open source version of WinBUGS LinBUGS is the Linux implementation of OpenBUGS.

43 Programs BUGS: Bayesian inference Using Gibbs Sampling WinBUGS is the Windows implementation can be called from R with R2WinBUGS package can be run on any intel-based computer using VMware, wine OpenBUGS open source version of WinBUGS LinBUGS is the Linux implementation of OpenBUGS. JAGS: Just Another Gibbs Sampler is an alternative program that uses the (almost) same model description as BUGS (Linux, MAC OS X, Windows) Can call from R using library(r2jags) or library(rjags)

44 Programs BUGS: Bayesian inference Using Gibbs Sampling WinBUGS is the Windows implementation can be called from R with R2WinBUGS package can be run on any intel-based computer using VMware, wine OpenBUGS open source version of WinBUGS LinBUGS is the Linux implementation of OpenBUGS. JAGS: Just Another Gibbs Sampler is an alternative program that uses the (almost) same model description as BUGS (Linux, MAC OS X, Windows) Can call from R using library(r2jags) or library(rjags) Include more than just Gibbs Sampling

45 JAGS Model

46 JAGS Model Data

47 JAGS Model Data Initial values (optional)

48 JAGS Model Data Initial values (optional) May do this through ordinary text files or use the functions in R2jags to specify model, data, and initial values then call jags.

49 Model Specification via R2jags rr.model = function() { for (i in 1:n) { mu[i] <- alpha0 + alpha1*(x[i] - Xbar) lambda[i] ~ dgamma(9/2, 9/2) prec[i] <- phi*lambda[i] Y[i] ~ dnorm(mu[i], prec[i]) } phi ~ dgamma(1.0e-6, 1.0E-6) alpha0 ~ dnorm(0, 1.0E-6) alpha1 ~ dnorm(0,1.0e-6) }

50 Notes on Models Distributions of stochastic nodes are specified using

51 Notes on Models Distributions of stochastic nodes are specified using Assignment of deterministic nodes uses <- (NOT =)

52 Notes on Models Distributions of stochastic nodes are specified using Assignment of deterministic nodes uses <- (NOT =) JAGS allows expressions as arguments in distributions (WinBUGS does not)

53 Notes on Models Distributions of stochastic nodes are specified using Assignment of deterministic nodes uses <- (NOT =) JAGS allows expressions as arguments in distributions (WinBUGS does not) Normal distributions are parameterized using precisions, so dnorm(0, 1.0E-6) is a N(0, )

54 Notes on Models Distributions of stochastic nodes are specified using Assignment of deterministic nodes uses <- (NOT =) JAGS allows expressions as arguments in distributions (WinBUGS does not) Normal distributions are parameterized using precisions, so dnorm(0, 1.0E-6) is a N(0, ) uses for loop structure as in R for model description but coded in C++ so is fast!

55 Data A list or rectangular data structure for all data and summaries of data used in the model bf.data = list(y = bodyfat$bodyfat, X=bodyfat$Abdomen) bf.data$n = length(bf.data$y) bf.data$xbar = mean(bf.data$x)

56 Specifying which Parameters to Save The parameters to be monitored and returned to R are specified with the variable parameters parameters = c("beta0", "beta1", "sigma", "mu34", "y34", "lambda[39]") All of the above (except lambda) are calculated from the other parameters. (See R-code for definitions of these parameters.)

57 Specifying which Parameters to Save The parameters to be monitored and returned to R are specified with the variable parameters parameters = c("beta0", "beta1", "sigma", "mu34", "y34", "lambda[39]") All of the above (except lambda) are calculated from the other parameters. (See R-code for definitions of these parameters.) lambda[39] saves only the 39th case of λ

58 Specifying which Parameters to Save The parameters to be monitored and returned to R are specified with the variable parameters parameters = c("beta0", "beta1", "sigma", "mu34", "y34", "lambda[39]") All of the above (except lambda) are calculated from the other parameters. (See R-code for definitions of these parameters.) lambda[39] saves only the 39th case of λ To save a whole vector (for example all lambdas, just give the vector name)

59 Specifying which Parameters to Save The parameters to be monitored and returned to R are specified with the variable parameters parameters = c("beta0", "beta1", "sigma", "mu34", "y34", "lambda[39]") All of the above (except lambda) are calculated from the other parameters. (See R-code for definitions of these parameters.) lambda[39] saves only the 39th case of λ To save a whole vector (for example all lambdas, just give the vector name)

60 Running jags from R bf.sim = jags(bf.data, inits=null, par=parameters, model=rr.model, n.chains=2, n.iter=5000, )

61 Output mean sd 2.5% 50% 97.5% beta beta sigma mu y lambda[39] % HPD interval for expected bodyfat (14.5, 15.8) 95% HPD interval for bodyfat (5.1, 25.3)

62 Comparison 95% Probability Interval for β is (0.60, 0.71) with t 9 errors

63 Comparison 95% Probability Interval for β is (0.60, 0.71) with t 9 errors 95% Confidence Interval for β is (0.58, 0.69) (all data normal model)

64 Comparison 95% Probability Interval for β is (0.60, 0.71) with t 9 errors 95% Confidence Interval for β is (0.58, 0.69) (all data normal model) 95% Confidence Interval for β is (0.61, 0.73) ( normal model without case 39)

65 Comparison 95% Probability Interval for β is (0.60, 0.71) with t 9 errors 95% Confidence Interval for β is (0.58, 0.69) (all data normal model) 95% Confidence Interval for β is (0.61, 0.73) ( normal model without case 39) Results intermediate without having to remove any observations

66 Comparison 95% Probability Interval for β is (0.60, 0.71) with t 9 errors 95% Confidence Interval for β is (0.58, 0.69) (all data normal model) 95% Confidence Interval for β is (0.61, 0.73) ( normal model without case 39) Results intermediate without having to remove any observations Case 39 down weighted by λ 39

67 Full Conditional for λ j p(λ j rest, Y ) p(α, β, φ, λ 1,..., λ n Y )

68 Full Conditional for λ j p(λ j rest, Y ) p(α, β, φ, λ 1,..., λ n Y ) n φ n/2 1 exp { φ2 } λ i(y i α βx i ) 2 n i=1 λ ν i i=1 exp( λ i ν 2 )

69 Full Conditional for λ j p(λ j rest, Y ) p(α, β, φ, λ 1,..., λ n Y ) n φ n/2 1 exp { φ2 } λ i(y i α βx i ) 2 n i=1 λ ν i i=1 exp( λ i ν 2 ) Ignore all terms except those that involve λ j

70 Full Conditional for λ j p(λ j rest, Y ) p(α, β, φ, λ 1,..., λ n Y ) n φ n/2 1 exp { φ2 } λ i(y i α βx i ) 2 n i=1 λ ν i i=1 exp( λ i ν 2 ) Ignore all terms except those that involve λ j λ j rest, Y G ( ν + 1 2, φ(y j α βx j ) 2 ) + ν 2

71 Weights Under prior E[λ i ] = 1

72 Weights Under prior E[λ i ] = 1 Under posterior, large residuals are down-weighted (approximately those bigger than ν) Posterior Distribution Density λ 39

73 Prior Distributions on Parameter As a general recommendation, the prior distribution should have heavier tails than the likelihood

74 Prior Distributions on Parameter As a general recommendation, the prior distribution should have heavier tails than the likelihood with t 9 errors use a t α with α < 9

75 Prior Distributions on Parameter As a general recommendation, the prior distribution should have heavier tails than the likelihood with t 9 errors use a t α with α < 9 also represent via scale mixture of normals

76 Prior Distributions on Parameter As a general recommendation, the prior distribution should have heavier tails than the likelihood with t 9 errors use a t α with α < 9 also represent via scale mixture of normals Horseshoe, Double Pareto, Cauchy all have heavier tails

77 Prior Distributions on Parameter As a general recommendation, the prior distribution should have heavier tails than the likelihood with t 9 errors use a t α with α < 9 also represent via scale mixture of normals Horseshoe, Double Pareto, Cauchy all have heavier tails See Stack-loss code

78 Prior Distributions on Parameter As a general recommendation, the prior distribution should have heavier tails than the likelihood with t 9 errors use a t α with α < 9 also represent via scale mixture of normals Horseshoe, Double Pareto, Cauchy all have heavier tails See Stack-loss code

Robust Bayesian Simple Linear Regression

Robust Bayesian Simple Linear Regression Robust Bayesian Simple Linear Regression October 1, 2008 Readings: GIll 4 Robust Bayesian Simple Linear Regression p.1/11 Body Fat Data: Intervals w/ All Data 95% confidence and prediction intervals for

More information

Simple Linear Regression

Simple Linear Regression Simple Linear Regression Reading: Hoff Chapter 9 November 4, 2009 Problem Data: Observe pairs (Y i,x i ),i = 1,... n Response or dependent variable Y Predictor or independent variable X GOALS: Exploring

More information

BUGS Bayesian inference Using Gibbs Sampling

BUGS Bayesian inference Using Gibbs Sampling BUGS Bayesian inference Using Gibbs Sampling Glen DePalma Department of Statistics May 30, 2013 www.stat.purdue.edu/~gdepalma 1 / 20 Bayesian Philosophy I [Pearl] turned Bayesian in 1971, as soon as I

More information

Linear Regression. Data Model. β, σ 2. Process Model. ,V β. ,s 2. s 1. Parameter Model

Linear Regression. Data Model. β, σ 2. Process Model. ,V β. ,s 2. s 1. Parameter Model Regression: Part II Linear Regression y~n X, 2 X Y Data Model β, σ 2 Process Model Β 0,V β s 1,s 2 Parameter Model Assumptions of Linear Model Homoskedasticity No error in X variables Error in Y variables

More information

WinBUGS : part 2. Bruno Boulanger Jonathan Jaeger Astrid Jullion Philippe Lambert. Gabriele, living with rheumatoid arthritis

WinBUGS : part 2. Bruno Boulanger Jonathan Jaeger Astrid Jullion Philippe Lambert. Gabriele, living with rheumatoid arthritis WinBUGS : part 2 Bruno Boulanger Jonathan Jaeger Astrid Jullion Philippe Lambert Gabriele, living with rheumatoid arthritis Agenda 2! Hierarchical model: linear regression example! R2WinBUGS Linear Regression

More information

Motivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University

Motivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University Econ 690 Purdue University In virtually all of the previous lectures, our models have made use of normality assumptions. From a computational point of view, the reason for this assumption is clear: combined

More information

Bayesian Inference for Regression Parameters

Bayesian Inference for Regression Parameters Bayesian Inference for Regression Parameters 1 Bayesian inference for simple linear regression parameters follows the usual pattern for all Bayesian analyses: 1. Form a prior distribution over all unknown

More information

36-463/663Multilevel and Hierarchical Models

36-463/663Multilevel and Hierarchical Models 36-463/663Multilevel and Hierarchical Models From Bayes to MCMC to MLMs Brian Junker 132E Baker Hall brian@stat.cmu.edu 1 Outline Bayesian Statistics and MCMC Distribution of Skill Mastery in a Population

More information

Multivariate Normal & Wishart

Multivariate Normal & Wishart Multivariate Normal & Wishart Hoff Chapter 7 October 21, 2010 Reading Comprehesion Example Twenty-two children are given a reading comprehsion test before and after receiving a particular instruction method.

More information

DAG models and Markov Chain Monte Carlo methods a short overview

DAG models and Markov Chain Monte Carlo methods a short overview DAG models and Markov Chain Monte Carlo methods a short overview Søren Højsgaard Institute of Genetics and Biotechnology University of Aarhus August 18, 2008 Printed: August 18, 2008 File: DAGMC-Lecture.tex

More information

Bayesian linear regression

Bayesian linear regression Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding

More information

Statistical Models. ref: chapter 1 of Bates, D and D. Watts (1988) Nonlinear Regression Analysis and its Applications, Wiley. Dave Campbell 2009

Statistical Models. ref: chapter 1 of Bates, D and D. Watts (1988) Nonlinear Regression Analysis and its Applications, Wiley. Dave Campbell 2009 Statistical Models ref: chapter 1 of Bates, D and D. Watts (1988) Nonlinear Regression Analysis and its Applications, Wiley Dave Campbell 2009 Today linear regression in terms of the response surface geometry

More information

Simple Linear Regression

Simple Linear Regression Simple Linear Regression September 24, 2008 Reading HH 8, GIll 4 Simple Linear Regression p.1/20 Problem Data: Observe pairs (Y i,x i ),i = 1,...n Response or dependent variable Y Predictor or independent

More information

Statistics in Environmental Research (BUC Workshop Series) II Problem sheet - WinBUGS - SOLUTIONS

Statistics in Environmental Research (BUC Workshop Series) II Problem sheet - WinBUGS - SOLUTIONS Statistics in Environmental Research (BUC Workshop Series) II Problem sheet - WinBUGS - SOLUTIONS 1. (a) The posterior mean estimate of α is 14.27, and the posterior mean for the standard deviation of

More information

Bayesian Graphical Models

Bayesian Graphical Models Graphical Models and Inference, Lecture 16, Michaelmas Term 2009 December 4, 2009 Parameter θ, data X = x, likelihood L(θ x) p(x θ). Express knowledge about θ through prior distribution π on θ. Inference

More information

Hierarchical Linear Models

Hierarchical Linear Models Hierarchical Linear Models Statistics 220 Spring 2005 Copyright c 2005 by Mark E. Irwin The linear regression model Hierarchical Linear Models y N(Xβ, Σ y ) β σ 2 p(β σ 2 ) σ 2 p(σ 2 ) can be extended

More information

36-463/663: Multilevel & Hierarchical Models HW09 Solution

36-463/663: Multilevel & Hierarchical Models HW09 Solution 36-463/663: Multilevel & Hierarchical Models HW09 Solution November 15, 2016 Quesion 1 Following the derivation given in class, when { n( x µ) 2 L(µ) exp, f(p) exp 2σ 2 0 ( the posterior is also normally

More information

Advanced Statistical Modelling

Advanced Statistical Modelling Markov chain Monte Carlo (MCMC) Methods and Their Applications in Bayesian Statistics School of Technology and Business Studies/Statistics Dalarna University Borlänge, Sweden. Feb. 05, 2014. Outlines 1

More information

Why Bayesian approaches? The average height of a rare plant

Why Bayesian approaches? The average height of a rare plant Why Bayesian approaches? The average height of a rare plant Estimation and comparison of averages is an important step in many ecological analyses and demographic models. In this demonstration you will

More information

Gibbs Sampling in Latent Variable Models #1

Gibbs Sampling in Latent Variable Models #1 Gibbs Sampling in Latent Variable Models #1 Econ 690 Purdue University Outline 1 Data augmentation 2 Probit Model Probit Application A Panel Probit Panel Probit 3 The Tobit Model Example: Female Labor

More information

Hypothesis Testing and Estimation under a Bayesian Approach

Hypothesis Testing and Estimation under a Bayesian Approach Hypothesis Testing and Estimation under a Bayesian Approach L.R. Pericchi* and M.E. Pérez 1 1 Department of Mathematics Universidad de Puerto Rico, Río Piedras Campus *Co-Leader of Biostatistics, Epidemiology

More information

Stat 544 Final Exam. May 2, I have neither given nor received unauthorized assistance on this examination.

Stat 544 Final Exam. May 2, I have neither given nor received unauthorized assistance on this examination. Stat 544 Final Exam May, 006 I have neither given nor received unauthorized assistance on this examination. signature date 1 1. Below is a directed acyclic graph that represents a joint distribution for

More information

Bayesian Model Averaging

Bayesian Model Averaging Bayesian Model Averaging Hoff Chapter 9, Hoeting et al 1999, Clyde & George 2004, Liang et al 2008 October 24, 2017 Bayesian Model Choice Models for the variable selection problem are based on a subset

More information

INTRODUCING LINEAR REGRESSION MODELS Response or Dependent variable y

INTRODUCING LINEAR REGRESSION MODELS Response or Dependent variable y INTRODUCING LINEAR REGRESSION MODELS Response or Dependent variable y Predictor or Independent variable x Model with error: for i = 1,..., n, y i = α + βx i + ε i ε i : independent errors (sampling, measurement,

More information

Metropolis-Hastings Algorithm

Metropolis-Hastings Algorithm Strength of the Gibbs sampler Metropolis-Hastings Algorithm Easy algorithm to think about. Exploits the factorization properties of the joint probability distribution. No difficult choices to be made to

More information

DIC: Deviance Information Criterion

DIC: Deviance Information Criterion (((( Welcome Page Latest News DIC: Deviance Information Criterion Contact us/bugs list WinBUGS New WinBUGS examples FAQs DIC GeoBUGS DIC (Deviance Information Criterion) is a Bayesian method for model

More information

Mixtures of Prior Distributions

Mixtures of Prior Distributions Mixtures of Prior Distributions Hoff Chapter 9, Liang et al 2007, Hoeting et al (1999), Clyde & George (2004) November 10, 2016 Bartlett s Paradox The Bayes factor for comparing M γ to the null model:

More information

Weakness of Beta priors (or conjugate priors in general) They can only represent a limited range of prior beliefs. For example... There are no bimodal beta distributions (except when the modes are at 0

More information

Mixtures of Prior Distributions

Mixtures of Prior Distributions Mixtures of Prior Distributions Hoff Chapter 9, Liang et al 2007, Hoeting et al (1999), Clyde & George (2004) November 9, 2017 Bartlett s Paradox The Bayes factor for comparing M γ to the null model: BF

More information

Bayesian Linear Regression

Bayesian Linear Regression Bayesian Linear Regression Sudipto Banerjee 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. September 15, 2010 1 Linear regression models: a Bayesian perspective

More information

Bayesian course - problem set 5 (lecture 6)

Bayesian course - problem set 5 (lecture 6) Bayesian course - problem set 5 (lecture 6) Ben Lambert November 30, 2016 1 Stan entry level: discoveries data The file prob5 discoveries.csv contains data on the numbers of great inventions and scientific

More information

Lecture 6. Prior distributions

Lecture 6. Prior distributions Summary Lecture 6. Prior distributions 1. Introduction 2. Bivariate conjugate: normal 3. Non-informative / reference priors Jeffreys priors Location parameters Proportions Counts and rates Scale parameters

More information

Partial factor modeling: predictor-dependent shrinkage for linear regression

Partial factor modeling: predictor-dependent shrinkage for linear regression modeling: predictor-dependent shrinkage for linear Richard Hahn, Carlos Carvalho and Sayan Mukherjee JASA 2013 Review by Esther Salazar Duke University December, 2013 Factor framework The factor framework

More information

Robust estimation of skew-normal distribution with location and scale parameters via log-regularly varying functions

Robust estimation of skew-normal distribution with location and scale parameters via log-regularly varying functions Robust estimation of skew-normal distribution with location and scale parameters via log-regularly varying functions Shintaro Hashimoto Department of Mathematics, Hiroshima University October 5, 207 Abstract

More information

November 2002 STA Random Effects Selection in Linear Mixed Models

November 2002 STA Random Effects Selection in Linear Mixed Models November 2002 STA216 1 Random Effects Selection in Linear Mixed Models November 2002 STA216 2 Introduction It is common practice in many applications to collect multiple measurements on a subject. Linear

More information

Weakness of Beta priors (or conjugate priors in general) They can only represent a limited range of prior beliefs. For example... There are no bimodal beta distributions (except when the modes are at 0

More information

Package horseshoe. November 8, 2016

Package horseshoe. November 8, 2016 Title Implementation of the Horseshoe Prior Version 0.1.0 Package horseshoe November 8, 2016 Description Contains functions for applying the horseshoe prior to highdimensional linear regression, yielding

More information

Part 8: GLMs and Hierarchical LMs and GLMs

Part 8: GLMs and Hierarchical LMs and GLMs Part 8: GLMs and Hierarchical LMs and GLMs 1 Example: Song sparrow reproductive success Arcese et al., (1992) provide data on a sample from a population of 52 female song sparrows studied over the course

More information

Bayesian Model Diagnostics and Checking

Bayesian Model Diagnostics and Checking Earvin Balderama Quantitative Ecology Lab Department of Forestry and Environmental Resources North Carolina State University April 12, 2013 1 / 34 Introduction MCMCMC 2 / 34 Introduction MCMCMC Steps in

More information

Penalized Loss functions for Bayesian Model Choice

Penalized Loss functions for Bayesian Model Choice Penalized Loss functions for Bayesian Model Choice Martyn International Agency for Research on Cancer Lyon, France 13 November 2009 The pure approach For a Bayesian purist, all uncertainty is represented

More information

PIER HLM Course July 30, 2011 Howard Seltman. Discussion Guide for Bayes and BUGS

PIER HLM Course July 30, 2011 Howard Seltman. Discussion Guide for Bayes and BUGS PIER HLM Course July 30, 2011 Howard Seltman Discussion Guide for Bayes and BUGS 1. Classical Statistics is based on parameters as fixed unknown values. a. The standard approach is to try to discover,

More information

Bayesian Networks in Educational Assessment

Bayesian Networks in Educational Assessment Bayesian Networks in Educational Assessment Estimating Parameters with MCMC Bayesian Inference: Expanding Our Context Roy Levy Arizona State University Roy.Levy@asu.edu 2017 Roy Levy MCMC 1 MCMC 2 Posterior

More information

GOV 2001/ 1002/ E-2001 Section 3 Theories of Inference

GOV 2001/ 1002/ E-2001 Section 3 Theories of Inference GOV 2001/ 1002/ E-2001 Section 3 Theories of Inference Solé Prillaman Harvard University February 11, 2015 1 / 48 LOGISTICS Reading Assignment- Unifying Political Methodology chs 2 and 4. Problem Set 3-

More information

Statistics/Mathematics 310 Larget April 14, Solution to Assignment #9

Statistics/Mathematics 310 Larget April 14, Solution to Assignment #9 Solution to Assignment #9 1. A sample s = (753.0, 1396.9, 1528.2, 1646.9, 717.6, 446.5, 848.0, 1222.0, 748.5, 610.1) is modeled as an i.i.d sample from a Gamma(α, λ) distribution. You may think of this

More information

Model Checking and Improvement

Model Checking and Improvement Model Checking and Improvement Statistics 220 Spring 2005 Copyright c 2005 by Mark E. Irwin Model Checking All models are wrong but some models are useful George E. P. Box So far we have looked at a number

More information

Bayesian Linear Models

Bayesian Linear Models Eric F. Lock UMN Division of Biostatistics, SPH elock@umn.edu 03/07/2018 Linear model For observations y 1,..., y n, the basic linear model is y i = x 1i β 1 +... + x pi β p + ɛ i, x 1i,..., x pi are predictors

More information

Bayesian inference. Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark. April 10, 2017

Bayesian inference. Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark. April 10, 2017 Bayesian inference Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark April 10, 2017 1 / 22 Outline for today A genetic example Bayes theorem Examples Priors Posterior summaries

More information

A Lightweight Guide on Gibbs Sampling and JAGS

A Lightweight Guide on Gibbs Sampling and JAGS A Lightweight Guide on Gibbs Sampling and JAGS Gianpaolo Coro Istituto di Scienza e Tecnologie dell Informazione A. Faedo, Pisa, Italy 1 Summary In this document we give some insight about how Gibbs Sampling

More information

Bayesian Inference in a Normal Population

Bayesian Inference in a Normal Population Bayesian Inference in a Normal Population September 20, 2007 Casella & Berger Chapter 7, Gelman, Carlin, Stern, Rubin Sec 2.6, 2.8, Chapter 3. Bayesian Inference in a Normal Population p. 1/16 Normal Model

More information

Bayesian Inference in a Normal Population

Bayesian Inference in a Normal Population Bayesian Inference in a Normal Population September 17, 2008 Gill Chapter 3. Sections 1-4, 7-8 Bayesian Inference in a Normal Population p.1/18 Normal Model IID observations Y = (Y 1,Y 2,...Y n ) Y i N(µ,σ

More information

Lecture 13 Fundamentals of Bayesian Inference

Lecture 13 Fundamentals of Bayesian Inference Lecture 13 Fundamentals of Bayesian Inference Dennis Sun Stats 253 August 11, 2014 Outline of Lecture 1 Bayesian Models 2 Modeling Correlations Using Bayes 3 The Universal Algorithm 4 BUGS 5 Wrapping Up

More information

Model Choice. Hoff Chapter 9, Clyde & George Model Uncertainty StatSci, Hoeting et al BMA StatSci. October 27, 2015

Model Choice. Hoff Chapter 9, Clyde & George Model Uncertainty StatSci, Hoeting et al BMA StatSci. October 27, 2015 Model Choice Hoff Chapter 9, Clyde & George Model Uncertainty StatSci, Hoeting et al BMA StatSci October 27, 2015 Topics Variable Selection / Model Choice Stepwise Methods Model Selection Criteria Model

More information

Problem Selected Scores

Problem Selected Scores Statistics Ph.D. Qualifying Exam: Part II November 20, 2010 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. Problem 1 2 3 4 5 6 7 8 9 10 11 12 Selected

More information

Linear models and their mathematical foundations: Simple linear regression

Linear models and their mathematical foundations: Simple linear regression Linear models and their mathematical foundations: Simple linear regression Steffen Unkel Department of Medical Statistics University Medical Center Göttingen, Germany Winter term 2018/19 1/21 Introduction

More information

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns

More information

STA 2201/442 Assignment 2

STA 2201/442 Assignment 2 STA 2201/442 Assignment 2 1. This is about how to simulate from a continuous univariate distribution. Let the random variable X have a continuous distribution with density f X (x) and cumulative distribution

More information

Module 17: Bayesian Statistics for Genetics Lecture 4: Linear regression

Module 17: Bayesian Statistics for Genetics Lecture 4: Linear regression 1/37 The linear regression model Module 17: Bayesian Statistics for Genetics Lecture 4: Linear regression Ken Rice Department of Biostatistics University of Washington 2/37 The linear regression model

More information

The lmm Package. May 9, Description Some improved procedures for linear mixed models

The lmm Package. May 9, Description Some improved procedures for linear mixed models The lmm Package May 9, 2005 Version 0.3-4 Date 2005-5-9 Title Linear mixed models Author Original by Joseph L. Schafer . Maintainer Jing hua Zhao Description Some improved

More information

A Parameter Expansion Approach to Bayesian SEM Estimation

A Parameter Expansion Approach to Bayesian SEM Estimation A Parameter Expansion Approach to Bayesian SEM Estimation Ed Merkle and Yves Rosseel Utrecht University 24 June 2016 Yves Rosseel A Parameter Expansion Approach to Bayesian SEM Estimation 1 / 51 overview

More information

" M A #M B. Standard deviation of the population (Greek lowercase letter sigma) σ 2

 M A #M B. Standard deviation of the population (Greek lowercase letter sigma) σ 2 Notation and Equations for Final Exam Symbol Definition X The variable we measure in a scientific study n The size of the sample N The size of the population M The mean of the sample µ The mean of the

More information

Package lmm. R topics documented: March 19, Version 0.4. Date Title Linear mixed models. Author Joseph L. Schafer

Package lmm. R topics documented: March 19, Version 0.4. Date Title Linear mixed models. Author Joseph L. Schafer Package lmm March 19, 2012 Version 0.4 Date 2012-3-19 Title Linear mixed models Author Joseph L. Schafer Maintainer Jing hua Zhao Depends R (>= 2.0.0) Description Some

More information

An Introduction to Bayesian Linear Regression

An Introduction to Bayesian Linear Regression An Introduction to Bayesian Linear Regression APPM 5720: Bayesian Computation Fall 2018 A SIMPLE LINEAR MODEL Suppose that we observe explanatory variables x 1, x 2,..., x n and dependent variables y 1,

More information

Information. Hierarchical Models - Statistical Methods. References. Outline

Information. Hierarchical Models - Statistical Methods. References. Outline Information Hierarchical Models - Statistical Methods Sarah Filippi 1 University of Oxford Hilary Term 2015 Webpage: http://www.stats.ox.ac.uk/~filippi/msc_ hierarchicalmodels_2015.html Lectures: Week

More information

Measurement error as missing data: the case of epidemiologic assays. Roderick J. Little

Measurement error as missing data: the case of epidemiologic assays. Roderick J. Little Measurement error as missing data: the case of epidemiologic assays Roderick J. Little Outline Discuss two related calibration topics where classical methods are deficient (A) Limit of quantification methods

More information

Module 11: Linear Regression. Rebecca C. Steorts

Module 11: Linear Regression. Rebecca C. Steorts Module 11: Linear Regression Rebecca C. Steorts Announcements Today is the last class Homework 7 has been extended to Thursday, April 20, 11 PM. There will be no lab tomorrow. There will be office hours

More information

A discussion on multiple regression models

A discussion on multiple regression models A discussion on multiple regression models In our previous discussion of simple linear regression, we focused on a model in which one independent or explanatory variable X was used to predict the value

More information

Inference for a Population Proportion

Inference for a Population Proportion Al Nosedal. University of Toronto. November 11, 2015 Statistical inference is drawing conclusions about an entire population based on data in a sample drawn from that population. From both frequentist

More information

10. Exchangeability and hierarchical models Objective. Recommended reading

10. Exchangeability and hierarchical models Objective. Recommended reading 10. Exchangeability and hierarchical models Objective Introduce exchangeability and its relation to Bayesian hierarchical models. Show how to fit such models using fully and empirical Bayesian methods.

More information

Latent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent

Latent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent Latent Variable Models for Binary Data Suppose that for a given vector of explanatory variables x, the latent variable, U, has a continuous cumulative distribution function F (u; x) and that the binary

More information

Math 423/533: The Main Theoretical Topics

Math 423/533: The Main Theoretical Topics Math 423/533: The Main Theoretical Topics Notation sample size n, data index i number of predictors, p (p = 2 for simple linear regression) y i : response for individual i x i = (x i1,..., x ip ) (1 p)

More information

Collaborative Statistics: Symbols and their Meanings

Collaborative Statistics: Symbols and their Meanings OpenStax-CNX module: m16302 1 Collaborative Statistics: Symbols and their Meanings Susan Dean Barbara Illowsky, Ph.D. This work is produced by OpenStax-CNX and licensed under the Creative Commons Attribution

More information

Part 6: Multivariate Normal and Linear Models

Part 6: Multivariate Normal and Linear Models Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of

More information

9 Correlation and Regression

9 Correlation and Regression 9 Correlation and Regression SW, Chapter 12. Suppose we select n = 10 persons from the population of college seniors who plan to take the MCAT exam. Each takes the test, is coached, and then retakes the

More information

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

Bayesian Inference. Chapter 4: Regression and Hierarchical Models Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Advanced Statistics and Data Mining Summer School

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters

More information

ACCOUNTING FOR INPUT-MODEL AND INPUT-PARAMETER UNCERTAINTIES IN SIMULATION. <www.ie.ncsu.edu/jwilson> May 22, 2006

ACCOUNTING FOR INPUT-MODEL AND INPUT-PARAMETER UNCERTAINTIES IN SIMULATION. <www.ie.ncsu.edu/jwilson> May 22, 2006 ACCOUNTING FOR INPUT-MODEL AND INPUT-PARAMETER UNCERTAINTIES IN SIMULATION Slide 1 Faker Zouaoui Sabre Holdings James R. Wilson NC State University May, 006 Slide From American

More information

Contents. 1 Introduction: what is overdispersion? 2 Recognising (and testing for) overdispersion. 1 Introduction: what is overdispersion?

Contents. 1 Introduction: what is overdispersion? 2 Recognising (and testing for) overdispersion. 1 Introduction: what is overdispersion? Overdispersion, and how to deal with it in R and JAGS (requires R-packages AER, coda, lme4, R2jags, DHARMa/devtools) Carsten F. Dormann 07 December, 2016 Contents 1 Introduction: what is overdispersion?

More information

36-463/663: Multilevel & Hierarchical Models

36-463/663: Multilevel & Hierarchical Models 36-463/663: Multilevel & Hierarchical Models (P)review: in-class midterm Brian Junker 132E Baker Hall brian@stat.cmu.edu 1 In-class midterm Closed book, closed notes, closed electronics (otherwise I have

More information

Gibbs Sampling in Endogenous Variables Models

Gibbs Sampling in Endogenous Variables Models Gibbs Sampling in Endogenous Variables Models Econ 690 Purdue University Outline 1 Motivation 2 Identification Issues 3 Posterior Simulation #1 4 Posterior Simulation #2 Motivation In this lecture we take

More information

Module 4: Bayesian Methods Lecture 5: Linear regression

Module 4: Bayesian Methods Lecture 5: Linear regression 1/28 The linear regression model Module 4: Bayesian Methods Lecture 5: Linear regression Peter Hoff Departments of Statistics and Biostatistics University of Washington 2/28 The linear regression model

More information

Lecture 18: Simple Linear Regression

Lecture 18: Simple Linear Regression Lecture 18: Simple Linear Regression BIOS 553 Department of Biostatistics University of Michigan Fall 2004 The Correlation Coefficient: r The correlation coefficient (r) is a number that measures the strength

More information

Example using R: Heart Valves Study

Example using R: Heart Valves Study Example using R: Heart Valves Study Goal: Show that the thrombogenicity rate (TR) is less than two times the objective performance criterion R and WinBUGS Examples p. 1/27 Example using R: Heart Valves

More information

Model Choice. Hoff Chapter 9. Dec 8, 2010

Model Choice. Hoff Chapter 9. Dec 8, 2010 Model Choice Hoff Chapter 9 Dec 8, 2010 Topics Variable Selection / Model Choice Stepwise Methods Model Selection Criteria Model Averaging Variable Selection Reasons for reducing the number of variables

More information

9. Linear Regression and Correlation

9. Linear Regression and Correlation 9. Linear Regression and Correlation Data: y a quantitative response variable x a quantitative explanatory variable (Chap. 8: Recall that both variables were categorical) For example, y = annual income,

More information

9.1 Analysis of Variance models

9.1 Analysis of Variance models 9.1 Analysis of Variance models Assume a continuous response variable y with categorical explanatory variable x (i.e. factor). Assume x has an inuence on expected values of y, hence there would be dierent

More information

Contents. 1 Review of Residuals. 2 Detecting Outliers. 3 Influential Observations. 4 Multicollinearity and its Effects

Contents. 1 Review of Residuals. 2 Detecting Outliers. 3 Influential Observations. 4 Multicollinearity and its Effects Contents 1 Review of Residuals 2 Detecting Outliers 3 Influential Observations 4 Multicollinearity and its Effects W. Zhou (Colorado State University) STAT 540 July 6th, 2015 1 / 32 Model Diagnostics:

More information

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

Bayesian Inference. Chapter 4: Regression and Hierarchical Models Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Master in Business Administration and Quantitative

More information

A Bayesian Treatment of Linear Gaussian Regression

A Bayesian Treatment of Linear Gaussian Regression A Bayesian Treatment of Linear Gaussian Regression Frank Wood December 3, 2009 Bayesian Approach to Classical Linear Regression In classical linear regression we have the following model y β, σ 2, X N(Xβ,

More information

Bayesian Econometrics - Computer section

Bayesian Econometrics - Computer section Bayesian Econometrics - Computer section Leandro Magnusson Department of Economics Brown University Leandro Magnusson@brown.edu http://www.econ.brown.edu/students/leandro Magnusson/ April 26, 2006 Preliminary

More information

Horseshoe, Lasso and Related Shrinkage Methods

Horseshoe, Lasso and Related Shrinkage Methods Readings Chapter 15 Christensen Merlise Clyde October 15, 2015 Bayesian Lasso Park & Casella (JASA 2008) and Hans (Biometrika 2010) propose Bayesian versions of the Lasso Bayesian Lasso Park & Casella

More information

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 EPSY 905: Intro to Bayesian and MCMC Today s Class An

More information

Approximating high-dimensional posteriors with nuisance parameters via integrated rotated Gaussian approximation (IRGA)

Approximating high-dimensional posteriors with nuisance parameters via integrated rotated Gaussian approximation (IRGA) Approximating high-dimensional posteriors with nuisance parameters via integrated rotated Gaussian approximation (IRGA) Willem van den Boom Department of Statistics and Applied Probability National University

More information

Inferences for Regression

Inferences for Regression Inferences for Regression An Example: Body Fat and Waist Size Looking at the relationship between % body fat and waist size (in inches). Here is a scatterplot of our data set: Remembering Regression In

More information

Markov Chain Monte Carlo

Markov Chain Monte Carlo Markov Chain Monte Carlo Jamie Monogan University of Georgia Spring 2013 For more information, including R programs, properties of Markov chains, and Metropolis-Hastings, please see: http://monogan.myweb.uga.edu/teaching/statcomp/mcmc.pdf

More information

The linear model is the most fundamental of all serious statistical models encompassing:

The linear model is the most fundamental of all serious statistical models encompassing: Linear Regression Models: A Bayesian perspective Ingredients of a linear model include an n 1 response vector y = (y 1,..., y n ) T and an n p design matrix (e.g. including regressors) X = [x 1,..., x

More information

MCMC Methods: Gibbs and Metropolis

MCMC Methods: Gibbs and Metropolis MCMC Methods: Gibbs and Metropolis Patrick Breheny February 28 Patrick Breheny BST 701: Bayesian Modeling in Biostatistics 1/30 Introduction As we have seen, the ability to sample from the posterior distribution

More information

Multiple Linear Regression

Multiple Linear Regression Multiple Linear Regression University of California, San Diego Instructor: Ery Arias-Castro http://math.ucsd.edu/~eariasca/teaching.html 1 / 42 Passenger car mileage Consider the carmpg dataset taken from

More information

ST 740: Linear Models and Multivariate Normal Inference

ST 740: Linear Models and Multivariate Normal Inference ST 740: Linear Models and Multivariate Normal Inference Alyson Wilson Department of Statistics North Carolina State University November 4, 2013 A. Wilson (NCSU STAT) Linear Models November 4, 2013 1 /

More information

MCMC 2: Lecture 2 Coding and output. Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham

MCMC 2: Lecture 2 Coding and output. Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham MCMC 2: Lecture 2 Coding and output Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham Contents 1. General (Markov) epidemic model 2. Non-Markov epidemic model 3. Debugging

More information

Count Predicted Variable & Contingency Tables

Count Predicted Variable & Contingency Tables Count Predicted Variable & Contingency Tables Tim Frasier Copyright Tim Frasier This work is licensed under the Creative Commons Attribution 4.0 International license. Click here for more information.

More information