Robust Bayesian Regression
|
|
- Hugo Rich
- 5 years ago
- Views:
Transcription
1 Readings: Hoff Chapter 9, West JRSSB 1984, Fúquene, Pérez & Pericchi 2015 Duke University November 17, 2016
2 Body Fat Data: Intervals w/ All Data Response % Body Fat and Predictor Waist Circumference 95% confidence and prediction intervals for bodyfat.lm 95% confidence and prediction intervals for bodyfat.lm2 xbar xbar Bodyfat 30 observed fit conf int pred int Bodyfat observed fit conf int pred int xbar Abdomen xbar Abdomen Which analysis do we use? with Case 39 or not or something different?
3 Cook s Distance Leverage Standardized residuals Cook's distance Residuals vs Leverage
4 Options for Handling Influential Cases Are there scientific grounds for eliminating the case?
5 Options for Handling Influential Cases Are there scientific grounds for eliminating the case? Test if the case has a different mean than population
6 Options for Handling Influential Cases Are there scientific grounds for eliminating the case? Test if the case has a different mean than population Report results with and without the case
7 Options for Handling Influential Cases Are there scientific grounds for eliminating the case? Test if the case has a different mean than population Report results with and without the case Model Averaging to Account for Model Uncertainty?
8 Options for Handling Influential Cases Are there scientific grounds for eliminating the case? Test if the case has a different mean than population Report results with and without the case Model Averaging to Account for Model Uncertainty? Full model Y = Xβ + I n δ + ɛ
9 Options for Handling Influential Cases Are there scientific grounds for eliminating the case? Test if the case has a different mean than population Report results with and without the case Model Averaging to Account for Model Uncertainty? Full model Y = Xβ + I n δ + ɛ 2 n submodels γ i = 0 δ i = 0 If γ i = 1 then case i has a different mean mean shift outliers.
10 Mean Shift = Variance Inflation Model Y = Xβ + I n δ + ɛ Prior δ i γ i N(0, V σ 2 γ i ) γ i Ber(π) Then ɛ i given σ 2 is independent of δ i and ɛ i ɛ i + δ i σ 2 { N(0, σ 2 ) wp (1 π) N(0, σ 2 (1 + V )) wp π Model Y = Xβ + ɛ variance inflation V + 1 = K = 7 in the paper by Hoeting et al. package BMA
11 Simultaneous Outlier and Variable Selection MC3.REG(all.y = bodyfat$bodyfat, all.x = as.matrix(bodyfat$abdom num.its = 10000, outliers = TRUE) Model parameters: PI=0.02 K=7 nu=2.58 lambda=0.28 phi= models were selected Best 5 models (cumulative posterior probability = ): prob model 1 model 2 model 3 model 4 model 5 variables all.x 1 x x x x x outliers x x. x x x.. x post prob
12 Change Error Assumptions Y i ind t(ν, α + βx i, 1/φ)
13 Change Error Assumptions Y i L(α, β, φ) ind t(ν, α + βx i, 1/φ) n ( φ 1/2 1 + φ(y i α βx i ) 2 ν i=1 ) (ν+1) 2
14 Change Error Assumptions Y i L(α, β, φ) ind t(ν, α + βx i, 1/φ) n ( φ 1/2 1 + φ(y i α βx i ) 2 ν i=1 ) (ν+1) 2 Use Prior p(α, β, φ) 1/φ
15 Change Error Assumptions Y i L(α, β, φ) ind t(ν, α + βx i, 1/φ) n ( φ 1/2 1 + φ(y i α βx i ) 2 ν i=1 ) (ν+1) 2 Use Prior p(α, β, φ) 1/φ Posterior distribution p(α, β, φ Y ) φ n/2 1 n (1 + φ(y i α βx i ) 2 i=1 ν ) (ν+1) 2
16 Change Error Assumptions Y i L(α, β, φ) ind t(ν, α + βx i, 1/φ) n ( φ 1/2 1 + φ(y i α βx i ) 2 ν i=1 ) (ν+1) 2 Use Prior p(α, β, φ) 1/φ Posterior distribution p(α, β, φ Y ) φ n/2 1 n (1 + φ(y i α βx i ) 2 i=1 ν ) (ν+1) 2
17 Bounded Influence - West 1984 (and references within) Treat σ 2 as given, then influence of individual observations on the posterior distribution of β in the model where E[Y i ] = x T i β is investigated through the score function:
18 Bounded Influence - West 1984 (and references within) Treat σ 2 as given, then influence of individual observations on the posterior distribution of β in the model where E[Y i ] = x T i β is investigated through the score function: d d n log p(β Y) = dβ dβ log p(β) + xg(y i x T i β) i=1
19 Bounded Influence - West 1984 (and references within) Treat σ 2 as given, then influence of individual observations on the posterior distribution of β in the model where E[Y i ] = x T i β is investigated through the score function: where d d n log p(β Y) = dβ dβ log p(β) + xg(y i x T i β) g(ɛ) = d log p(ɛ) dɛ is the influence function of the error distribution (unimodal, continuous, differentiable, symmetric) i=1
20 Bounded Influence - West 1984 (and references within) Treat σ 2 as given, then influence of individual observations on the posterior distribution of β in the model where E[Y i ] = x T i β is investigated through the score function: where d d n log p(β Y) = dβ dβ log p(β) + xg(y i x T i β) g(ɛ) = d log p(ɛ) dɛ is the influence function of the error distribution (unimodal, continuous, differentiable, symmetric) i=1 An outlying observation y j is accommodated if the posterior distribution for p(β Y (i) ) converges to p(β Y) for all β as Y i. Requires error models with influence functions that go to zero such as the Student t (O Hagan, 1979)
21 Choice of df Score function for t with α degrees of freedom has turning points at ± α
22 Choice of df Score function for t with α degrees of freedom has turning points at ± α g(eps, 9) eps g (ɛ) is negative when ɛ 2 > α (standardized errors)
23 Choice of df Score function for t with α degrees of freedom has turning points at ± α g(eps, 9) eps g (ɛ) is negative when ɛ 2 > α (standardized errors) Contribution of observation to information matrix is negative and the observation is doubtful
24 Choice of df Score function for t with α degrees of freedom has turning points at ± α g(eps, 9) eps g (ɛ) is negative when ɛ 2 > α (standardized errors) Contribution of observation to information matrix is negative and the observation is doubtful Suggest taking α = 8 or α = 9 to reject errors larger than 8 or 3 sd.
25 Choice of df Score function for t with α degrees of freedom has turning points at ± α g(eps, 9) eps g (ɛ) is negative when ɛ 2 > α (standardized errors) Contribution of observation to information matrix is negative and the observation is doubtful Suggest taking α = 8 or α = 9 to reject errors larger than 8 or 3 sd.
26 Scale-Mixtures of Normal Representation Z i iid t(ν, 0, σ 2 )
27 Scale-Mixtures of Normal Representation Z i iid t(ν, 0, σ 2 ) Z i λ i ind N(0, σ 2 /λ i )
28 Scale-Mixtures of Normal Representation Z i iid t(ν, 0, σ 2 ) Z i λ i ind N(0, σ 2 /λ i ) λ i iid G(ν/2, ν/2)
29 Scale-Mixtures of Normal Representation Z i iid t(ν, 0, σ 2 ) Z i λ i ind N(0, σ 2 /λ i ) λ i iid G(ν/2, ν/2) Integrate out latent λ s to obtain marginal distribution.
30 Latent Variable Model Y i α, β, φ, λ ind 1 N(α + βx i, ) φλ i
31 Latent Variable Model Y i α, β, φ, λ ind 1 N(α + βx i, ) φλ i λ i iid G(ν/2, ν/2)
32 Latent Variable Model Y i α, β, φ, λ ind 1 N(α + βx i, ) φλ i λ i iid G(ν/2, ν/2) p(α, β, φ) 1/φ
33 Latent Variable Model Y i α, β, φ, λ Joint Posterior Distribution: ind 1 N(α + βx i, ) φλ i λ i iid G(ν/2, ν/2) p(α, β, φ) 1/φ
34 Latent Variable Model Y i α, β, φ, λ ind 1 N(α + βx i, ) φλ i λ i iid G(ν/2, ν/2) p(α, β, φ) 1/φ Joint Posterior Distribution: p((α, β, φ, λ 1,..., λ n Y ) { φ n/2 exp φ } λi (y i α βx i ) 2 2
35 Latent Variable Model Y i α, β, φ, λ ind 1 N(α + βx i, ) φλ i λ i iid G(ν/2, ν/2) p(α, β, φ) 1/φ Joint Posterior Distribution: p((α, β, φ, λ 1,..., λ n Y ) { φ n/2 exp φ } λi (y i α βx i ) 2 2 φ 1
36 Latent Variable Model Y i α, β, φ, λ ind 1 N(α + βx i, ) φλ i λ i iid G(ν/2, ν/2) p(α, β, φ) 1/φ Joint Posterior Distribution: p((α, β, φ, λ 1,..., λ n Y ) { φ n/2 exp φ } λi (y i α βx i ) 2 2 φ 1 n i=1 λ ν/2 1 i exp( λ i ν/2)
37 Programs BUGS: Bayesian inference Using Gibbs Sampling
38 Programs BUGS: Bayesian inference Using Gibbs Sampling WinBUGS is the Windows implementation
39 Programs BUGS: Bayesian inference Using Gibbs Sampling WinBUGS is the Windows implementation can be called from R with R2WinBUGS package
40 Programs BUGS: Bayesian inference Using Gibbs Sampling WinBUGS is the Windows implementation can be called from R with R2WinBUGS package can be run on any intel-based computer using VMware, wine
41 Programs BUGS: Bayesian inference Using Gibbs Sampling WinBUGS is the Windows implementation can be called from R with R2WinBUGS package can be run on any intel-based computer using VMware, wine OpenBUGS open source version of WinBUGS
42 Programs BUGS: Bayesian inference Using Gibbs Sampling WinBUGS is the Windows implementation can be called from R with R2WinBUGS package can be run on any intel-based computer using VMware, wine OpenBUGS open source version of WinBUGS LinBUGS is the Linux implementation of OpenBUGS.
43 Programs BUGS: Bayesian inference Using Gibbs Sampling WinBUGS is the Windows implementation can be called from R with R2WinBUGS package can be run on any intel-based computer using VMware, wine OpenBUGS open source version of WinBUGS LinBUGS is the Linux implementation of OpenBUGS. JAGS: Just Another Gibbs Sampler is an alternative program that uses the (almost) same model description as BUGS (Linux, MAC OS X, Windows) Can call from R using library(r2jags) or library(rjags)
44 Programs BUGS: Bayesian inference Using Gibbs Sampling WinBUGS is the Windows implementation can be called from R with R2WinBUGS package can be run on any intel-based computer using VMware, wine OpenBUGS open source version of WinBUGS LinBUGS is the Linux implementation of OpenBUGS. JAGS: Just Another Gibbs Sampler is an alternative program that uses the (almost) same model description as BUGS (Linux, MAC OS X, Windows) Can call from R using library(r2jags) or library(rjags) Include more than just Gibbs Sampling
45 JAGS Model
46 JAGS Model Data
47 JAGS Model Data Initial values (optional)
48 JAGS Model Data Initial values (optional) May do this through ordinary text files or use the functions in R2jags to specify model, data, and initial values then call jags.
49 Model Specification via R2jags rr.model = function() { for (i in 1:n) { mu[i] <- alpha0 + alpha1*(x[i] - Xbar) lambda[i] ~ dgamma(9/2, 9/2) prec[i] <- phi*lambda[i] Y[i] ~ dnorm(mu[i], prec[i]) } phi ~ dgamma(1.0e-6, 1.0E-6) alpha0 ~ dnorm(0, 1.0E-6) alpha1 ~ dnorm(0,1.0e-6) }
50 Notes on Models Distributions of stochastic nodes are specified using
51 Notes on Models Distributions of stochastic nodes are specified using Assignment of deterministic nodes uses <- (NOT =)
52 Notes on Models Distributions of stochastic nodes are specified using Assignment of deterministic nodes uses <- (NOT =) JAGS allows expressions as arguments in distributions (WinBUGS does not)
53 Notes on Models Distributions of stochastic nodes are specified using Assignment of deterministic nodes uses <- (NOT =) JAGS allows expressions as arguments in distributions (WinBUGS does not) Normal distributions are parameterized using precisions, so dnorm(0, 1.0E-6) is a N(0, )
54 Notes on Models Distributions of stochastic nodes are specified using Assignment of deterministic nodes uses <- (NOT =) JAGS allows expressions as arguments in distributions (WinBUGS does not) Normal distributions are parameterized using precisions, so dnorm(0, 1.0E-6) is a N(0, ) uses for loop structure as in R for model description but coded in C++ so is fast!
55 Data A list or rectangular data structure for all data and summaries of data used in the model bf.data = list(y = bodyfat$bodyfat, X=bodyfat$Abdomen) bf.data$n = length(bf.data$y) bf.data$xbar = mean(bf.data$x)
56 Specifying which Parameters to Save The parameters to be monitored and returned to R are specified with the variable parameters parameters = c("beta0", "beta1", "sigma", "mu34", "y34", "lambda[39]") All of the above (except lambda) are calculated from the other parameters. (See R-code for definitions of these parameters.)
57 Specifying which Parameters to Save The parameters to be monitored and returned to R are specified with the variable parameters parameters = c("beta0", "beta1", "sigma", "mu34", "y34", "lambda[39]") All of the above (except lambda) are calculated from the other parameters. (See R-code for definitions of these parameters.) lambda[39] saves only the 39th case of λ
58 Specifying which Parameters to Save The parameters to be monitored and returned to R are specified with the variable parameters parameters = c("beta0", "beta1", "sigma", "mu34", "y34", "lambda[39]") All of the above (except lambda) are calculated from the other parameters. (See R-code for definitions of these parameters.) lambda[39] saves only the 39th case of λ To save a whole vector (for example all lambdas, just give the vector name)
59 Specifying which Parameters to Save The parameters to be monitored and returned to R are specified with the variable parameters parameters = c("beta0", "beta1", "sigma", "mu34", "y34", "lambda[39]") All of the above (except lambda) are calculated from the other parameters. (See R-code for definitions of these parameters.) lambda[39] saves only the 39th case of λ To save a whole vector (for example all lambdas, just give the vector name)
60 Running jags from R bf.sim = jags(bf.data, inits=null, par=parameters, model=rr.model, n.chains=2, n.iter=5000, )
61 Output mean sd 2.5% 50% 97.5% beta beta sigma mu y lambda[39] % HPD interval for expected bodyfat (14.5, 15.8) 95% HPD interval for bodyfat (5.1, 25.3)
62 Comparison 95% Probability Interval for β is (0.60, 0.71) with t 9 errors
63 Comparison 95% Probability Interval for β is (0.60, 0.71) with t 9 errors 95% Confidence Interval for β is (0.58, 0.69) (all data normal model)
64 Comparison 95% Probability Interval for β is (0.60, 0.71) with t 9 errors 95% Confidence Interval for β is (0.58, 0.69) (all data normal model) 95% Confidence Interval for β is (0.61, 0.73) ( normal model without case 39)
65 Comparison 95% Probability Interval for β is (0.60, 0.71) with t 9 errors 95% Confidence Interval for β is (0.58, 0.69) (all data normal model) 95% Confidence Interval for β is (0.61, 0.73) ( normal model without case 39) Results intermediate without having to remove any observations
66 Comparison 95% Probability Interval for β is (0.60, 0.71) with t 9 errors 95% Confidence Interval for β is (0.58, 0.69) (all data normal model) 95% Confidence Interval for β is (0.61, 0.73) ( normal model without case 39) Results intermediate without having to remove any observations Case 39 down weighted by λ 39
67 Full Conditional for λ j p(λ j rest, Y ) p(α, β, φ, λ 1,..., λ n Y )
68 Full Conditional for λ j p(λ j rest, Y ) p(α, β, φ, λ 1,..., λ n Y ) n φ n/2 1 exp { φ2 } λ i(y i α βx i ) 2 n i=1 λ ν i i=1 exp( λ i ν 2 )
69 Full Conditional for λ j p(λ j rest, Y ) p(α, β, φ, λ 1,..., λ n Y ) n φ n/2 1 exp { φ2 } λ i(y i α βx i ) 2 n i=1 λ ν i i=1 exp( λ i ν 2 ) Ignore all terms except those that involve λ j
70 Full Conditional for λ j p(λ j rest, Y ) p(α, β, φ, λ 1,..., λ n Y ) n φ n/2 1 exp { φ2 } λ i(y i α βx i ) 2 n i=1 λ ν i i=1 exp( λ i ν 2 ) Ignore all terms except those that involve λ j λ j rest, Y G ( ν + 1 2, φ(y j α βx j ) 2 ) + ν 2
71 Weights Under prior E[λ i ] = 1
72 Weights Under prior E[λ i ] = 1 Under posterior, large residuals are down-weighted (approximately those bigger than ν) Posterior Distribution Density λ 39
73 Prior Distributions on Parameter As a general recommendation, the prior distribution should have heavier tails than the likelihood
74 Prior Distributions on Parameter As a general recommendation, the prior distribution should have heavier tails than the likelihood with t 9 errors use a t α with α < 9
75 Prior Distributions on Parameter As a general recommendation, the prior distribution should have heavier tails than the likelihood with t 9 errors use a t α with α < 9 also represent via scale mixture of normals
76 Prior Distributions on Parameter As a general recommendation, the prior distribution should have heavier tails than the likelihood with t 9 errors use a t α with α < 9 also represent via scale mixture of normals Horseshoe, Double Pareto, Cauchy all have heavier tails
77 Prior Distributions on Parameter As a general recommendation, the prior distribution should have heavier tails than the likelihood with t 9 errors use a t α with α < 9 also represent via scale mixture of normals Horseshoe, Double Pareto, Cauchy all have heavier tails See Stack-loss code
78 Prior Distributions on Parameter As a general recommendation, the prior distribution should have heavier tails than the likelihood with t 9 errors use a t α with α < 9 also represent via scale mixture of normals Horseshoe, Double Pareto, Cauchy all have heavier tails See Stack-loss code
Robust Bayesian Simple Linear Regression
Robust Bayesian Simple Linear Regression October 1, 2008 Readings: GIll 4 Robust Bayesian Simple Linear Regression p.1/11 Body Fat Data: Intervals w/ All Data 95% confidence and prediction intervals for
More informationSimple Linear Regression
Simple Linear Regression Reading: Hoff Chapter 9 November 4, 2009 Problem Data: Observe pairs (Y i,x i ),i = 1,... n Response or dependent variable Y Predictor or independent variable X GOALS: Exploring
More informationBUGS Bayesian inference Using Gibbs Sampling
BUGS Bayesian inference Using Gibbs Sampling Glen DePalma Department of Statistics May 30, 2013 www.stat.purdue.edu/~gdepalma 1 / 20 Bayesian Philosophy I [Pearl] turned Bayesian in 1971, as soon as I
More informationLinear Regression. Data Model. β, σ 2. Process Model. ,V β. ,s 2. s 1. Parameter Model
Regression: Part II Linear Regression y~n X, 2 X Y Data Model β, σ 2 Process Model Β 0,V β s 1,s 2 Parameter Model Assumptions of Linear Model Homoskedasticity No error in X variables Error in Y variables
More informationWinBUGS : part 2. Bruno Boulanger Jonathan Jaeger Astrid Jullion Philippe Lambert. Gabriele, living with rheumatoid arthritis
WinBUGS : part 2 Bruno Boulanger Jonathan Jaeger Astrid Jullion Philippe Lambert Gabriele, living with rheumatoid arthritis Agenda 2! Hierarchical model: linear regression example! R2WinBUGS Linear Regression
More informationMotivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University
Econ 690 Purdue University In virtually all of the previous lectures, our models have made use of normality assumptions. From a computational point of view, the reason for this assumption is clear: combined
More informationBayesian Inference for Regression Parameters
Bayesian Inference for Regression Parameters 1 Bayesian inference for simple linear regression parameters follows the usual pattern for all Bayesian analyses: 1. Form a prior distribution over all unknown
More information36-463/663Multilevel and Hierarchical Models
36-463/663Multilevel and Hierarchical Models From Bayes to MCMC to MLMs Brian Junker 132E Baker Hall brian@stat.cmu.edu 1 Outline Bayesian Statistics and MCMC Distribution of Skill Mastery in a Population
More informationMultivariate Normal & Wishart
Multivariate Normal & Wishart Hoff Chapter 7 October 21, 2010 Reading Comprehesion Example Twenty-two children are given a reading comprehsion test before and after receiving a particular instruction method.
More informationDAG models and Markov Chain Monte Carlo methods a short overview
DAG models and Markov Chain Monte Carlo methods a short overview Søren Højsgaard Institute of Genetics and Biotechnology University of Aarhus August 18, 2008 Printed: August 18, 2008 File: DAGMC-Lecture.tex
More informationBayesian linear regression
Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding
More informationStatistical Models. ref: chapter 1 of Bates, D and D. Watts (1988) Nonlinear Regression Analysis and its Applications, Wiley. Dave Campbell 2009
Statistical Models ref: chapter 1 of Bates, D and D. Watts (1988) Nonlinear Regression Analysis and its Applications, Wiley Dave Campbell 2009 Today linear regression in terms of the response surface geometry
More informationSimple Linear Regression
Simple Linear Regression September 24, 2008 Reading HH 8, GIll 4 Simple Linear Regression p.1/20 Problem Data: Observe pairs (Y i,x i ),i = 1,...n Response or dependent variable Y Predictor or independent
More informationStatistics in Environmental Research (BUC Workshop Series) II Problem sheet - WinBUGS - SOLUTIONS
Statistics in Environmental Research (BUC Workshop Series) II Problem sheet - WinBUGS - SOLUTIONS 1. (a) The posterior mean estimate of α is 14.27, and the posterior mean for the standard deviation of
More informationBayesian Graphical Models
Graphical Models and Inference, Lecture 16, Michaelmas Term 2009 December 4, 2009 Parameter θ, data X = x, likelihood L(θ x) p(x θ). Express knowledge about θ through prior distribution π on θ. Inference
More informationHierarchical Linear Models
Hierarchical Linear Models Statistics 220 Spring 2005 Copyright c 2005 by Mark E. Irwin The linear regression model Hierarchical Linear Models y N(Xβ, Σ y ) β σ 2 p(β σ 2 ) σ 2 p(σ 2 ) can be extended
More information36-463/663: Multilevel & Hierarchical Models HW09 Solution
36-463/663: Multilevel & Hierarchical Models HW09 Solution November 15, 2016 Quesion 1 Following the derivation given in class, when { n( x µ) 2 L(µ) exp, f(p) exp 2σ 2 0 ( the posterior is also normally
More informationAdvanced Statistical Modelling
Markov chain Monte Carlo (MCMC) Methods and Their Applications in Bayesian Statistics School of Technology and Business Studies/Statistics Dalarna University Borlänge, Sweden. Feb. 05, 2014. Outlines 1
More informationWhy Bayesian approaches? The average height of a rare plant
Why Bayesian approaches? The average height of a rare plant Estimation and comparison of averages is an important step in many ecological analyses and demographic models. In this demonstration you will
More informationGibbs Sampling in Latent Variable Models #1
Gibbs Sampling in Latent Variable Models #1 Econ 690 Purdue University Outline 1 Data augmentation 2 Probit Model Probit Application A Panel Probit Panel Probit 3 The Tobit Model Example: Female Labor
More informationHypothesis Testing and Estimation under a Bayesian Approach
Hypothesis Testing and Estimation under a Bayesian Approach L.R. Pericchi* and M.E. Pérez 1 1 Department of Mathematics Universidad de Puerto Rico, Río Piedras Campus *Co-Leader of Biostatistics, Epidemiology
More informationStat 544 Final Exam. May 2, I have neither given nor received unauthorized assistance on this examination.
Stat 544 Final Exam May, 006 I have neither given nor received unauthorized assistance on this examination. signature date 1 1. Below is a directed acyclic graph that represents a joint distribution for
More informationBayesian Model Averaging
Bayesian Model Averaging Hoff Chapter 9, Hoeting et al 1999, Clyde & George 2004, Liang et al 2008 October 24, 2017 Bayesian Model Choice Models for the variable selection problem are based on a subset
More informationINTRODUCING LINEAR REGRESSION MODELS Response or Dependent variable y
INTRODUCING LINEAR REGRESSION MODELS Response or Dependent variable y Predictor or Independent variable x Model with error: for i = 1,..., n, y i = α + βx i + ε i ε i : independent errors (sampling, measurement,
More informationMetropolis-Hastings Algorithm
Strength of the Gibbs sampler Metropolis-Hastings Algorithm Easy algorithm to think about. Exploits the factorization properties of the joint probability distribution. No difficult choices to be made to
More informationDIC: Deviance Information Criterion
(((( Welcome Page Latest News DIC: Deviance Information Criterion Contact us/bugs list WinBUGS New WinBUGS examples FAQs DIC GeoBUGS DIC (Deviance Information Criterion) is a Bayesian method for model
More informationMixtures of Prior Distributions
Mixtures of Prior Distributions Hoff Chapter 9, Liang et al 2007, Hoeting et al (1999), Clyde & George (2004) November 10, 2016 Bartlett s Paradox The Bayes factor for comparing M γ to the null model:
More informationWeakness of Beta priors (or conjugate priors in general) They can only represent a limited range of prior beliefs. For example... There are no bimodal beta distributions (except when the modes are at 0
More informationMixtures of Prior Distributions
Mixtures of Prior Distributions Hoff Chapter 9, Liang et al 2007, Hoeting et al (1999), Clyde & George (2004) November 9, 2017 Bartlett s Paradox The Bayes factor for comparing M γ to the null model: BF
More informationBayesian Linear Regression
Bayesian Linear Regression Sudipto Banerjee 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. September 15, 2010 1 Linear regression models: a Bayesian perspective
More informationBayesian course - problem set 5 (lecture 6)
Bayesian course - problem set 5 (lecture 6) Ben Lambert November 30, 2016 1 Stan entry level: discoveries data The file prob5 discoveries.csv contains data on the numbers of great inventions and scientific
More informationLecture 6. Prior distributions
Summary Lecture 6. Prior distributions 1. Introduction 2. Bivariate conjugate: normal 3. Non-informative / reference priors Jeffreys priors Location parameters Proportions Counts and rates Scale parameters
More informationPartial factor modeling: predictor-dependent shrinkage for linear regression
modeling: predictor-dependent shrinkage for linear Richard Hahn, Carlos Carvalho and Sayan Mukherjee JASA 2013 Review by Esther Salazar Duke University December, 2013 Factor framework The factor framework
More informationRobust estimation of skew-normal distribution with location and scale parameters via log-regularly varying functions
Robust estimation of skew-normal distribution with location and scale parameters via log-regularly varying functions Shintaro Hashimoto Department of Mathematics, Hiroshima University October 5, 207 Abstract
More informationNovember 2002 STA Random Effects Selection in Linear Mixed Models
November 2002 STA216 1 Random Effects Selection in Linear Mixed Models November 2002 STA216 2 Introduction It is common practice in many applications to collect multiple measurements on a subject. Linear
More informationWeakness of Beta priors (or conjugate priors in general) They can only represent a limited range of prior beliefs. For example... There are no bimodal beta distributions (except when the modes are at 0
More informationPackage horseshoe. November 8, 2016
Title Implementation of the Horseshoe Prior Version 0.1.0 Package horseshoe November 8, 2016 Description Contains functions for applying the horseshoe prior to highdimensional linear regression, yielding
More informationPart 8: GLMs and Hierarchical LMs and GLMs
Part 8: GLMs and Hierarchical LMs and GLMs 1 Example: Song sparrow reproductive success Arcese et al., (1992) provide data on a sample from a population of 52 female song sparrows studied over the course
More informationBayesian Model Diagnostics and Checking
Earvin Balderama Quantitative Ecology Lab Department of Forestry and Environmental Resources North Carolina State University April 12, 2013 1 / 34 Introduction MCMCMC 2 / 34 Introduction MCMCMC Steps in
More informationPenalized Loss functions for Bayesian Model Choice
Penalized Loss functions for Bayesian Model Choice Martyn International Agency for Research on Cancer Lyon, France 13 November 2009 The pure approach For a Bayesian purist, all uncertainty is represented
More informationPIER HLM Course July 30, 2011 Howard Seltman. Discussion Guide for Bayes and BUGS
PIER HLM Course July 30, 2011 Howard Seltman Discussion Guide for Bayes and BUGS 1. Classical Statistics is based on parameters as fixed unknown values. a. The standard approach is to try to discover,
More informationBayesian Networks in Educational Assessment
Bayesian Networks in Educational Assessment Estimating Parameters with MCMC Bayesian Inference: Expanding Our Context Roy Levy Arizona State University Roy.Levy@asu.edu 2017 Roy Levy MCMC 1 MCMC 2 Posterior
More informationGOV 2001/ 1002/ E-2001 Section 3 Theories of Inference
GOV 2001/ 1002/ E-2001 Section 3 Theories of Inference Solé Prillaman Harvard University February 11, 2015 1 / 48 LOGISTICS Reading Assignment- Unifying Political Methodology chs 2 and 4. Problem Set 3-
More informationStatistics/Mathematics 310 Larget April 14, Solution to Assignment #9
Solution to Assignment #9 1. A sample s = (753.0, 1396.9, 1528.2, 1646.9, 717.6, 446.5, 848.0, 1222.0, 748.5, 610.1) is modeled as an i.i.d sample from a Gamma(α, λ) distribution. You may think of this
More informationModel Checking and Improvement
Model Checking and Improvement Statistics 220 Spring 2005 Copyright c 2005 by Mark E. Irwin Model Checking All models are wrong but some models are useful George E. P. Box So far we have looked at a number
More informationBayesian Linear Models
Eric F. Lock UMN Division of Biostatistics, SPH elock@umn.edu 03/07/2018 Linear model For observations y 1,..., y n, the basic linear model is y i = x 1i β 1 +... + x pi β p + ɛ i, x 1i,..., x pi are predictors
More informationBayesian inference. Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark. April 10, 2017
Bayesian inference Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark April 10, 2017 1 / 22 Outline for today A genetic example Bayes theorem Examples Priors Posterior summaries
More informationA Lightweight Guide on Gibbs Sampling and JAGS
A Lightweight Guide on Gibbs Sampling and JAGS Gianpaolo Coro Istituto di Scienza e Tecnologie dell Informazione A. Faedo, Pisa, Italy 1 Summary In this document we give some insight about how Gibbs Sampling
More informationBayesian Inference in a Normal Population
Bayesian Inference in a Normal Population September 20, 2007 Casella & Berger Chapter 7, Gelman, Carlin, Stern, Rubin Sec 2.6, 2.8, Chapter 3. Bayesian Inference in a Normal Population p. 1/16 Normal Model
More informationBayesian Inference in a Normal Population
Bayesian Inference in a Normal Population September 17, 2008 Gill Chapter 3. Sections 1-4, 7-8 Bayesian Inference in a Normal Population p.1/18 Normal Model IID observations Y = (Y 1,Y 2,...Y n ) Y i N(µ,σ
More informationLecture 13 Fundamentals of Bayesian Inference
Lecture 13 Fundamentals of Bayesian Inference Dennis Sun Stats 253 August 11, 2014 Outline of Lecture 1 Bayesian Models 2 Modeling Correlations Using Bayes 3 The Universal Algorithm 4 BUGS 5 Wrapping Up
More informationModel Choice. Hoff Chapter 9, Clyde & George Model Uncertainty StatSci, Hoeting et al BMA StatSci. October 27, 2015
Model Choice Hoff Chapter 9, Clyde & George Model Uncertainty StatSci, Hoeting et al BMA StatSci October 27, 2015 Topics Variable Selection / Model Choice Stepwise Methods Model Selection Criteria Model
More informationProblem Selected Scores
Statistics Ph.D. Qualifying Exam: Part II November 20, 2010 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. Problem 1 2 3 4 5 6 7 8 9 10 11 12 Selected
More informationLinear models and their mathematical foundations: Simple linear regression
Linear models and their mathematical foundations: Simple linear regression Steffen Unkel Department of Medical Statistics University Medical Center Göttingen, Germany Winter term 2018/19 1/21 Introduction
More informationBayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence
Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns
More informationSTA 2201/442 Assignment 2
STA 2201/442 Assignment 2 1. This is about how to simulate from a continuous univariate distribution. Let the random variable X have a continuous distribution with density f X (x) and cumulative distribution
More informationModule 17: Bayesian Statistics for Genetics Lecture 4: Linear regression
1/37 The linear regression model Module 17: Bayesian Statistics for Genetics Lecture 4: Linear regression Ken Rice Department of Biostatistics University of Washington 2/37 The linear regression model
More informationThe lmm Package. May 9, Description Some improved procedures for linear mixed models
The lmm Package May 9, 2005 Version 0.3-4 Date 2005-5-9 Title Linear mixed models Author Original by Joseph L. Schafer . Maintainer Jing hua Zhao Description Some improved
More informationA Parameter Expansion Approach to Bayesian SEM Estimation
A Parameter Expansion Approach to Bayesian SEM Estimation Ed Merkle and Yves Rosseel Utrecht University 24 June 2016 Yves Rosseel A Parameter Expansion Approach to Bayesian SEM Estimation 1 / 51 overview
More information" M A #M B. Standard deviation of the population (Greek lowercase letter sigma) σ 2
Notation and Equations for Final Exam Symbol Definition X The variable we measure in a scientific study n The size of the sample N The size of the population M The mean of the sample µ The mean of the
More informationPackage lmm. R topics documented: March 19, Version 0.4. Date Title Linear mixed models. Author Joseph L. Schafer
Package lmm March 19, 2012 Version 0.4 Date 2012-3-19 Title Linear mixed models Author Joseph L. Schafer Maintainer Jing hua Zhao Depends R (>= 2.0.0) Description Some
More informationAn Introduction to Bayesian Linear Regression
An Introduction to Bayesian Linear Regression APPM 5720: Bayesian Computation Fall 2018 A SIMPLE LINEAR MODEL Suppose that we observe explanatory variables x 1, x 2,..., x n and dependent variables y 1,
More informationInformation. Hierarchical Models - Statistical Methods. References. Outline
Information Hierarchical Models - Statistical Methods Sarah Filippi 1 University of Oxford Hilary Term 2015 Webpage: http://www.stats.ox.ac.uk/~filippi/msc_ hierarchicalmodels_2015.html Lectures: Week
More informationMeasurement error as missing data: the case of epidemiologic assays. Roderick J. Little
Measurement error as missing data: the case of epidemiologic assays Roderick J. Little Outline Discuss two related calibration topics where classical methods are deficient (A) Limit of quantification methods
More informationModule 11: Linear Regression. Rebecca C. Steorts
Module 11: Linear Regression Rebecca C. Steorts Announcements Today is the last class Homework 7 has been extended to Thursday, April 20, 11 PM. There will be no lab tomorrow. There will be office hours
More informationA discussion on multiple regression models
A discussion on multiple regression models In our previous discussion of simple linear regression, we focused on a model in which one independent or explanatory variable X was used to predict the value
More informationInference for a Population Proportion
Al Nosedal. University of Toronto. November 11, 2015 Statistical inference is drawing conclusions about an entire population based on data in a sample drawn from that population. From both frequentist
More information10. Exchangeability and hierarchical models Objective. Recommended reading
10. Exchangeability and hierarchical models Objective Introduce exchangeability and its relation to Bayesian hierarchical models. Show how to fit such models using fully and empirical Bayesian methods.
More informationLatent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent
Latent Variable Models for Binary Data Suppose that for a given vector of explanatory variables x, the latent variable, U, has a continuous cumulative distribution function F (u; x) and that the binary
More informationMath 423/533: The Main Theoretical Topics
Math 423/533: The Main Theoretical Topics Notation sample size n, data index i number of predictors, p (p = 2 for simple linear regression) y i : response for individual i x i = (x i1,..., x ip ) (1 p)
More informationCollaborative Statistics: Symbols and their Meanings
OpenStax-CNX module: m16302 1 Collaborative Statistics: Symbols and their Meanings Susan Dean Barbara Illowsky, Ph.D. This work is produced by OpenStax-CNX and licensed under the Creative Commons Attribution
More informationPart 6: Multivariate Normal and Linear Models
Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of
More information9 Correlation and Regression
9 Correlation and Regression SW, Chapter 12. Suppose we select n = 10 persons from the population of college seniors who plan to take the MCAT exam. Each takes the test, is coached, and then retakes the
More informationBayesian Inference. Chapter 4: Regression and Hierarchical Models
Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Advanced Statistics and Data Mining Summer School
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters
More informationACCOUNTING FOR INPUT-MODEL AND INPUT-PARAMETER UNCERTAINTIES IN SIMULATION. <www.ie.ncsu.edu/jwilson> May 22, 2006
ACCOUNTING FOR INPUT-MODEL AND INPUT-PARAMETER UNCERTAINTIES IN SIMULATION Slide 1 Faker Zouaoui Sabre Holdings James R. Wilson NC State University May, 006 Slide From American
More informationContents. 1 Introduction: what is overdispersion? 2 Recognising (and testing for) overdispersion. 1 Introduction: what is overdispersion?
Overdispersion, and how to deal with it in R and JAGS (requires R-packages AER, coda, lme4, R2jags, DHARMa/devtools) Carsten F. Dormann 07 December, 2016 Contents 1 Introduction: what is overdispersion?
More information36-463/663: Multilevel & Hierarchical Models
36-463/663: Multilevel & Hierarchical Models (P)review: in-class midterm Brian Junker 132E Baker Hall brian@stat.cmu.edu 1 In-class midterm Closed book, closed notes, closed electronics (otherwise I have
More informationGibbs Sampling in Endogenous Variables Models
Gibbs Sampling in Endogenous Variables Models Econ 690 Purdue University Outline 1 Motivation 2 Identification Issues 3 Posterior Simulation #1 4 Posterior Simulation #2 Motivation In this lecture we take
More informationModule 4: Bayesian Methods Lecture 5: Linear regression
1/28 The linear regression model Module 4: Bayesian Methods Lecture 5: Linear regression Peter Hoff Departments of Statistics and Biostatistics University of Washington 2/28 The linear regression model
More informationLecture 18: Simple Linear Regression
Lecture 18: Simple Linear Regression BIOS 553 Department of Biostatistics University of Michigan Fall 2004 The Correlation Coefficient: r The correlation coefficient (r) is a number that measures the strength
More informationExample using R: Heart Valves Study
Example using R: Heart Valves Study Goal: Show that the thrombogenicity rate (TR) is less than two times the objective performance criterion R and WinBUGS Examples p. 1/27 Example using R: Heart Valves
More informationModel Choice. Hoff Chapter 9. Dec 8, 2010
Model Choice Hoff Chapter 9 Dec 8, 2010 Topics Variable Selection / Model Choice Stepwise Methods Model Selection Criteria Model Averaging Variable Selection Reasons for reducing the number of variables
More information9. Linear Regression and Correlation
9. Linear Regression and Correlation Data: y a quantitative response variable x a quantitative explanatory variable (Chap. 8: Recall that both variables were categorical) For example, y = annual income,
More information9.1 Analysis of Variance models
9.1 Analysis of Variance models Assume a continuous response variable y with categorical explanatory variable x (i.e. factor). Assume x has an inuence on expected values of y, hence there would be dierent
More informationContents. 1 Review of Residuals. 2 Detecting Outliers. 3 Influential Observations. 4 Multicollinearity and its Effects
Contents 1 Review of Residuals 2 Detecting Outliers 3 Influential Observations 4 Multicollinearity and its Effects W. Zhou (Colorado State University) STAT 540 July 6th, 2015 1 / 32 Model Diagnostics:
More informationBayesian Inference. Chapter 4: Regression and Hierarchical Models
Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Master in Business Administration and Quantitative
More informationA Bayesian Treatment of Linear Gaussian Regression
A Bayesian Treatment of Linear Gaussian Regression Frank Wood December 3, 2009 Bayesian Approach to Classical Linear Regression In classical linear regression we have the following model y β, σ 2, X N(Xβ,
More informationBayesian Econometrics - Computer section
Bayesian Econometrics - Computer section Leandro Magnusson Department of Economics Brown University Leandro Magnusson@brown.edu http://www.econ.brown.edu/students/leandro Magnusson/ April 26, 2006 Preliminary
More informationHorseshoe, Lasso and Related Shrinkage Methods
Readings Chapter 15 Christensen Merlise Clyde October 15, 2015 Bayesian Lasso Park & Casella (JASA 2008) and Hans (Biometrika 2010) propose Bayesian versions of the Lasso Bayesian Lasso Park & Casella
More informationIntroduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016
Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 EPSY 905: Intro to Bayesian and MCMC Today s Class An
More informationApproximating high-dimensional posteriors with nuisance parameters via integrated rotated Gaussian approximation (IRGA)
Approximating high-dimensional posteriors with nuisance parameters via integrated rotated Gaussian approximation (IRGA) Willem van den Boom Department of Statistics and Applied Probability National University
More informationInferences for Regression
Inferences for Regression An Example: Body Fat and Waist Size Looking at the relationship between % body fat and waist size (in inches). Here is a scatterplot of our data set: Remembering Regression In
More informationMarkov Chain Monte Carlo
Markov Chain Monte Carlo Jamie Monogan University of Georgia Spring 2013 For more information, including R programs, properties of Markov chains, and Metropolis-Hastings, please see: http://monogan.myweb.uga.edu/teaching/statcomp/mcmc.pdf
More informationThe linear model is the most fundamental of all serious statistical models encompassing:
Linear Regression Models: A Bayesian perspective Ingredients of a linear model include an n 1 response vector y = (y 1,..., y n ) T and an n p design matrix (e.g. including regressors) X = [x 1,..., x
More informationMCMC Methods: Gibbs and Metropolis
MCMC Methods: Gibbs and Metropolis Patrick Breheny February 28 Patrick Breheny BST 701: Bayesian Modeling in Biostatistics 1/30 Introduction As we have seen, the ability to sample from the posterior distribution
More informationMultiple Linear Regression
Multiple Linear Regression University of California, San Diego Instructor: Ery Arias-Castro http://math.ucsd.edu/~eariasca/teaching.html 1 / 42 Passenger car mileage Consider the carmpg dataset taken from
More informationST 740: Linear Models and Multivariate Normal Inference
ST 740: Linear Models and Multivariate Normal Inference Alyson Wilson Department of Statistics North Carolina State University November 4, 2013 A. Wilson (NCSU STAT) Linear Models November 4, 2013 1 /
More informationMCMC 2: Lecture 2 Coding and output. Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham
MCMC 2: Lecture 2 Coding and output Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham Contents 1. General (Markov) epidemic model 2. Non-Markov epidemic model 3. Debugging
More informationCount Predicted Variable & Contingency Tables
Count Predicted Variable & Contingency Tables Tim Frasier Copyright Tim Frasier This work is licensed under the Creative Commons Attribution 4.0 International license. Click here for more information.
More information