The Calibrated Bayes Factor for Model Comparison

Similar documents
1 Hypothesis Testing and Model Selection

Foundations of Nonparametric Bayesian Methods

Transformations and Bayesian Density Estimation

ST440/540: Applied Bayesian Statistics. (9) Model selection and goodness-of-fit checks

Invariant HPD credible sets and MAP estimators

Non-Parametric Bayes

(1) Introduction to Bayesian statistics

Inconsistency of Bayesian inference when the model is wrong, and how to repair it

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007

Should all Machine Learning be Bayesian? Should all Bayesian models be non-parametric?

Hypothesis Testing. Econ 690. Purdue University. Justin L. Tobias (Purdue) Testing 1 / 33

STAT Advanced Bayesian Inference

Construction of Dependent Dirichlet Processes based on Poisson Processes

Bayesian Econometrics

Bayesian Semiparametric GARCH Models

Seminar über Statistik FS2008: Model Selection

Statistics: Learning models from data

Bayesian Semiparametric GARCH Models

Recent Advances in Bayesian Inference Techniques

Curve Fitting Re-visited, Bishop1.2.5

Machine Learning. Gaussian Mixture Models. Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall

Bayesian Models in Machine Learning

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS

STAT 425: Introduction to Bayesian Analysis

Lecture : Probabilistic Machine Learning

Lecture 6: Model Checking and Selection

Uniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence

Divergence Based priors for the problem of hypothesis testing

Noninformative Priors for the Ratio of the Scale Parameters in the Inverted Exponential Distributions

Bayesian estimation of the discrepancy with misspecified parametric models

Construction of an Informative Hierarchical Prior Distribution: Application to Electricity Load Forecasting

Minimum Message Length Analysis of the Behrens Fisher Problem

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.

7. Estimation and hypothesis testing. Objective. Recommended reading

Semi-Nonparametric Inferences for Massive Data

MODEL COMPARISON CHRISTOPHER A. SIMS PRINCETON UNIVERSITY

Principles of Bayesian Inference

Expectation Propagation for Approximate Bayesian Inference

Bayesian Methods for Machine Learning

Multiple Sample Categorical Data

Quantitative Biology II Lecture 4: Variational Methods

Quantifying the Price of Uncertainty in Bayesian Models

Bayesian Analysis for Natural Language Processing Lecture 2

Bayesian Model Diagnostics and Checking

CS 540: Machine Learning Lecture 2: Review of Probability & Statistics

Power-Expected-Posterior Priors for Variable Selection in Gaussian Linear Models

Hierarchical Models & Bayesian Model Selection

Harrison B. Prosper. CMS Statistics Committee

Why Try Bayesian Methods? (Lecture 5)

Accounting for Complex Sample Designs via Mixture Models

CPSC 540: Machine Learning

An Extended BIC for Model Selection

Lecture 35: December The fundamental statistical distances

7. Estimation and hypothesis testing. Objective. Recommended reading

Model comparison. Christopher A. Sims Princeton University October 18, 2016

Machine Learning Linear Classification. Prof. Matteo Matteucci

Stable Limit Laws for Marginal Probabilities from MCMC Streams: Acceleration of Convergence

Simple Linear Regression

Clustering. Léon Bottou COS 424 3/4/2010. NEC Labs America

Bayesian Regression Linear and Logistic Regression

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =

Hybrid Censoring; An Introduction 2

Bayesian Learning (II)

Latent Dirichlet Alloca/on

CS 6140: Machine Learning Spring What We Learned Last Week. Survey 2/26/16. VS. Model

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

Lecture 16-17: Bayesian Nonparametrics I. STAT 6474 Instructor: Hongxiao Zhu

Some Curiosities Arising in Objective Bayesian Analysis

Nonparametric Bayesian Methods (Gaussian Processes)

CS 6140: Machine Learning Spring 2016

Introduction: MLE, MAP, Bayesian reasoning (28/8/13)

Inference for a Population Proportion

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution

Optimization. The value x is called a maximizer of f and is written argmax X f. g(λx + (1 λ)y) < λg(x) + (1 λ)g(y) 0 < λ < 1; x, y X.

Infer relationships among three species: Outgroup:

Model comparison: Deviance-based approaches

Density Estimation. Seungjin Choi

Computational statistics

the long tau-path for detecting monotone association in an unspecified subpopulation

Machine Learning CSE546 Carlos Guestrin University of Washington. September 30, 2013

PATTERN RECOGNITION AND MACHINE LEARNING

Principles of Bayesian Inference

Introduction. Start with a probability distribution f(y θ) for the data. where η is a vector of hyperparameters

Principles of Bayesian Inference

Bayesian Methods: Naïve Bayes

Objective Prior for the Number of Degrees of Freedom of a t Distribution

Robustness to Parametric Assumptions in Missing Data Models

Hybrid Censoring Scheme: An Introduction

Machine Learning CSE546 Carlos Guestrin University of Washington. September 30, What about continuous variables?

Confidence Intervals of the Simple Difference between the Proportions of a Primary Infection and a Secondary Infection, Given the Primary Infection

Introduction to Bayesian Methods

Machine Learning. Classification. Bayes Classifier. Representing data: Choosing hypothesis class. Learning: h:x a Y. Eric Xing

Introduction to Bayesian Statistics

Penalized Loss functions for Bayesian Model Choice

Statistical Models with Uncertain Error Parameters (G. Cowan, arxiv: )

Machine Learning for OR & FE

13: Variational inference II

Fast Approximate MAP Inference for Bayesian Nonparametrics

Topic Models. Brandon Malone. February 20, Latent Dirichlet Allocation Success Stories Wrap-up

Using Bayes Factors for Model Selection in High-Energy Astrophysics

Transcription:

The Calibrated Bayes Factor for Model Comparison Steve MacEachern The Ohio State University Joint work with Xinyi Xu, Pingbo Lu and Ruoxi Xu Supported by the NSF and NSA Bayesian Nonparametrics Workshop ICERM 2012

Outline The Bayes factor when it works, and when it doesn t The calibrated Bayes factor Ohio Family Health Survey (OFHS) analysis Wrap-up

Outline Bayes factors and prior distributions The calibrated Bayes factor OFHS analysis Wrap-up

Bayes factors The Bayes factor is one of the most important and most widely used tools for Bayesian hypothesis testing and model comparison. Given two models M 1 and M 2, we have BF = m(y;m 1) m(y;m 2 ), Some rules of thumb for using Bayes factors (Jeffreys 1961) 1 < Bayes factor 3: weak evidence for M 1 3 < Bayes factor 10: substantial evidence for M 1 10 < Bayes factor 100: strong evidence for M 1 100 < Bayes factor: decisive evidence for M 1

Monotonicity and the Bayes factor The Bayes factor is best examined by consideration of log(bf) = log(m(y;m 1 )) log(m(y;m 2 )) = n 1 log m(y i+1 Y 0:i,M 1 ) m(y i+1 Y 0:i,M 2 ). i=0 The expectation under M 1 is non-negative, and is postive if M 1 M 2. Consider examining the data set one observation at a time. If M 1 is right, each obs n makes a positive contribution in expectation. Trace of GMSS is similar to Brownian motion with non-linear drift.

Bayes factors (Cont.) log(bf) versus sample size log(bayes factor) 1.0 0.5 0.0 0.5 1.0 1.5 2.0 0 10 20 30 40 50 sample size

Example 1. Suppose that we have n = 176 i.i.d. observations from a skew-normal(location=0, scale=1.5, shape=2.5). Compare a Gaussian parametric model vs. a Mixture of Dirichlet processes (MDP) nonparametric model. Gaussian parametric model: y i θ,σ 2 DP nonparametric model: y i θ i,σ 2 θ i G iid N(θ,σ 2 ), i = 1,...,n θ N(µ,τ 2 ) iid N(θi,σ 2 ), i = 1,...,n iid G G DP(M = 2,N(µ,τ 2 )) Common priors on hyper-parameters: µ N(0,500), σ 2 IG(7,0.3), τ 2 IG(11,9.5)

Model comparison results Using the Bayes factor: B P,NP = e 4.92 137 Decisive evidence for the parametric model! Using posterior predictive performance: E[logm(Y n Y 1:(n 1) ;P)] = 1.4267 E[logm(Y n Y 1:(n 1) ;NP)] = 1.3977 The nonparametric model is better! Given the same sample size, why do the Bayes factor and the posterior marginal likelihoods provide such very different results?

A motivating example (Cont.) Here is the whole story... We randomly select subsamples of smaller size, compute the log Bayes factor and log posterior predictive distribution based on each subsample, and then take averages to smooth the plot E[log(Bayes factor)] vs. sample size log(bayes Factor) 0 1 2 3 4 5 1 21 41 56 71 86 101 126 151 176 Sample size

A motivating example(cont.) E[log(posterior predictive density)] vs. sample size E(log(Marginal Predictive Likelihood)) 2.0 1.9 1.8 1.7 1.6 1.5 1.4 1.3 1.4265 Para model Non P model 1.3977 1 20 40 55 70 85 100 125 150 175 Sample size

Bayes factors and predictive distributions The small model accumulates an enormous lead when the sample size is small. When the large model starts to have better predictive performance at larger sample sizes, the Bayes factor is slow to reflect the change in predictive performances! The Bayes factor follows from Bayes Theorem. It is, of course, exactly right, provided the inputs are right right likelihood right prior right loss

Bayes factors do they work? The Bayes factor works well for the subjective Bayesian The within-model prior distributions are meaningful The calculation follows from Bayes theorem Estimation, testing, and all other inference fit as part of a comprehensive analysis The Bayes factor breaks down for rule-based priors Objective priors (noninformative priors) High-dimensional settings (too much to specify) Infinite dimensional models (nonparametric Bayes) Many, many variants on the Bayes factor Most change the prior specifically for model comparison/choice One class of modified Bayes factors stands out within-model priors specified by some rule partial update is performed then the Bayes factor calculation commences

Bayes factors and partial updates Several different partial update methods (e.g., Lempers (1970), Geisser and Eddy (1979, PRESS), Berger and Pericchi (1995, 1996, IBF), O Hagan (1995, FBF)) Training data / test data split Rotation through different splits to stabiliize calculation Prior before updating is generally noninformative Minimal training sets have been advocated Questions! Why a minimal update? What if there is no noninformative prior? Do the methods work in more complex settings? In search of a principled solution: The question is not Bayesian model selection, but Bayesian model selection which is consistent with the rest of the analysis.

Outline Bayes factors and prior distributions The calibrated Bayes factor OFHS analysis Wrap-up

Toward a solution - the calibrated Bayes factor Subjective Bayesian analyses work well in high-information settings, much less well in low-information settings (try elicitation for yourself) We begin with the situation where all (Bayesians) agree on the solution and use this to drive the technique We propose that one start the Bayes factor calculation after the partial posterior resembles a high-info subjective prior Elements of the problem Measure the information content of the partial posterior Benchmark prior to describe adequate prior information Criterion for whether partial posterior matches benchmark We recommend calibration of the Bayes factor in any low-information or rule-based prior setting In these settings, elicited priors are unstable (Psychology)

Measurement of prior information Measure the proximity of f θ1 and f θ2 through the the Symmetric - Kullback-Leibler (SKL) divergence SKL(f θ1,f θ2 ) = 1 2 [ E θ 1 log f θ 1 f θ2 + E θ 2 log f θ 2 f θ1 SKL driven by likelihood, appropriate for Bayesians The distribution on (θ 1,θ 2 ) induces a distribution on SKL(f θ1,f θ2 ) This works for infinite-dimensional models too, unlike alternatives such as Fisher information Criterion: Evaluate the information contained in π using the percentiles of the distribution of SKL divergence ].

A benchmark prior To calibrate the Bayes factor and select a training sample size, we choose a benchmark prior and then require the updated priors to contain at least as much information as this benchmark prior. In order to perform a reasonable analysis where subjective input has little impact on the final conclusion, we set the benchmark to be a minimally informative" prior the unit information prior (Kass and Wasserman 1995), which contains the amount of information in a single observation Under the Gaussian model Y N(θ,σ 2 ), a unit information prior on θ is N(µ,σ 2 ), inducing a χ 2 1 distribution on SKL(f θ 1,f θ2 ).

The overall strategy: Calibrating the priors Step 1: For a single model, randomly draw a training sample of a pre-specified sample size from the data Step 2: Update the prior based on this training sample. Take M pairs of (θ j 1,θj 2 ), where j = 1,,M, and compute SKL(f j θ,f j 1 θ ) based on each pair 2 Step 3: Repeat Steps 1 and 2 N times. Pool all MN values of the SKLs to evaluate the information in the posterior Step 4: Compare the amount of information in the posterior to that in the benchmark distribution. If the amount of information is comparable, terminate the search and report the current sample size as the calibration sample size. Otherwise reset the sample size and repeat Steps 1 to 4.

Calibrated Bayes factors Let s 1 and s 2 represent the calibration sample sizes for models M 1 and M 2. Take s = max(s 1,s 2 ). Based on a training sample y (s), the updated Bayes factor satisfies logb 12 (y y (s)) = logb 12 (y) logb 12 (y (s) ), Let {y 1 (s),y2 (s),,yh (s)} denote all possible subsets of y of size s. Then the calibrated Bayes factor is defined by logcb 12 (y) = logb 12 (y) 1 H H logb 12 (y h (s) ) h=1

Asymptotic properties The calibrated Bayes factor shares the qualitative asymptotic properties of the Bayes factor. The main condition is that the calibration sample size (under both models) be finite with probability one. This removes an expected log(bayes factor) based on the calibration sample size. If log(bf) tends to some infinite limit, then so does log(cbf). If log(bf) tends to some finite number (odd, but examples exist), then log(cbf) will tend to a number differing by the offset.

Perils avoided We have avoided two huge pitfalls: Change the form of the prior distribution? Destroys the cohesiveness of the analysis, including both selection and estimation Hampers use of low-information priors Adding an extra model may change the priors! Adjust the hyper-parameters in the prior? Where should this more concentrated prior be centered? We use the data to drive the centering via the training sample Instead, we use training samples to update the priors What within-model prior do you want to use for estimation? Training sample size is chosen to mimic Bayes factor when they work well Driven by actual data without double use of the data

The motivating example revisited In the skew-normal example, our search leads to a calibration sample size of 50, based on the MDP model. The calibrated Bayes factor e 4.92 4.54 1.46, which is not worth more than a bare mention under Jeffrey s criterion. This result is consistent with the posterior predictive performances. Under the calibrated Bayes factor, the small parametric model never accumulates a significant lead!

The simulation setup To investigate the patterns of log Bayes factors and to illustrate the effect of calibration, we compare the Gaussian parametric model to the MDP model under the following distributions with various shapes: Skew-normal with varying shape parameter α (skewness) Student-t with varying degrees of freedom ν (thick-tails) Symmetric mixture of normals with varying component means ±δ (Bimodality) In all cases, the distributions have been centered and scaled to have mean 0 and standard deviation 1. By specifying α, ν and δ, we tune the KL distances from the true distributions to the best fitting Gaussian distributions.

Simulation results Small divergences from the Gaussian True standardized density 0.0 0.2 0.4 0.6 t skew normal mix of normal 4 2 0 2 4 Cumulative log BF 0 1 2 3 4 5 t skew normal mix of normal 1 21 41 56 71 86 101 126 151 176 201 251 301 Sample Size

Simulation results (Cont.) Moderate divergences from the Gaussian True standardized density 0.0 0.2 0.4 0.6 t skew normal mix of normal 4 2 0 2 4 Cumulative log BF 3 1 1 2 3 4 t skew normal mix of normal 1 21 41 56 71 86 101 126 151 176 201 251 301 Sample Size

Simulation results (Cont.) Large divergences from the Gaussian True standardized density 0.0 0.2 0.4 0.6 skew normal mix of normal 4 2 0 2 4 Cumulative log BF 25 15 5 0 skew normal mix of normal 1 21 41 56 71 86 101 126 151 176 201 251 301 Sample Size

Simulation results (Cont.) Very large divergences from the Gaussian True standardized density 0.0 0.2 0.4 0.6 mix of normal 4 2 0 2 4 Cumulative log BF 100 60 20 0 mix of normal 1 21 41 56 71 86 101 126 151 176 201 251 301 Sample Size

Simulation results summary In all cases, the calibration is driven by the MDP model rather than the Gaussian model (which is typically calibrated after two or three observations) In all cases, the peaks of the log calibrated Bayes factors remain below two, leading to better agreement between the Bayes factor and the models predictive performances. In the same scenario (the same KL divergence from the true distribution to the best fitting Gaussian distribution), the calibration sample size varies little Across different scenarios, the further the underlying true distribution is from normality, the larger the calibration sample size will be.

Outline Bayes factors and prior distributions The calibrated Bayes factor OFHS analysis Wrap-up

Model comparisons in OFHS analysis The OFHS (Ohio Family Health Survey) was conducted between August 2008 and January 2009 to study health insurance coverage of Ohioans. Data is largely self-reported. An important health measurement in this survey is BMI (Body Mass Index). We focus on the subpopulation consisting of male adults aged from 18 to 24. There are 895 non-missing BMI values in this group.

Model comparison in the OFHS analysis (Cont.) The log transformed data is close to a skew-normal distribution with the skewness parameter ˆα MLE = 2.41. Density Estimate for log(bmi) of Male Adult (Aged 18-24) Density 0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 4.0 log(bmi) of Male Adult (Aged 18-24)

Model Comparison in the OFHS analysis (Cont.) Based on the full data set, the log Bayes factor is 12.19, translating to a Bayes factor of 196, 811 favoring MDP. We further investigate the expected log Bayes factor for a range of smaller sample sizes. For each sample size, we generate 300 subsamples. If we only had a subset of the observations with size n = 106, the Bayes factor is B P;NP e 4.64 104, which provides strong evidence for the Gaussian parametric model

Model Comparison in the OFHS analysis (Cont.) After matching the prior concentration to the unit information prior, we calibrate the priors with training samples of size 50. At sample size n = 106, the calibrated Bayes factor is CB P;NP e 0.64 1.9, which provides a very weak model preference; at the full sample size, the eventual calibrated Bayes factor is CB P;NP e 16.18, or about 10.6 million to one in favor of the MDP model. We find the swing from inconclusive evidence for modest sample sizes to conclusive evidence in favor of the MDP model for the full sample far more palatable than the swing from very strong evidence in one direction to conclusive evidence in the opposite direction.

Outline Bayes factors and prior distributions The calibrated Bayes factor OFHS analysis Wrap-up

Concluding remarks Huge swings in the Bayes factor are not unique for parametric vs. nonparametric model comparisons. They are prevalent in small vs. large model comparisons. The original Bayes factor can be very misleading in this situation. There seems to be no sound remedy through alteration of the prior distributions. Such methods destroy the cohesiveness of the analysis. To make a fair comparison between small and large models, careful calibration of the Bayes factor is needed, and this can be done through the use of training samples. Such adjustment is also needed whenever priors are poorly specified or are rule-based. Regression models, discrete data models, dependence models (spatial, temporal, other), complex hierarchical models,...

A disturbing message The story has been that the Bayes factor is inadequate for model comparison in hard problems. But the Bayes factor is merely an expression of Bayes Theorem. Model comparison involves a contrast across submodels of a hyper-model So, what is to say that we should consider the MDP component of our hypermodel a submodel? If we split our model up differently, should we calibrate our Bayes Theorem calculations differently as well?