Bayesian parameter estimation with weak data and when combining evidence: the case of climate sensitivity

Similar documents
My talk concerns estimating a fixed but unknown, continuously valued parameter, linked to data by a statistical model. I focus on contrasting

Objectively combining AR5 instrumental period and paleoclimate climate sensitivity evidence

STAT 425: Introduction to Bayesian Analysis

Confidence Distribution

Parametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012

Statistical Theory MT 2007 Problems 4: Solution sketches

Supporting Information

Computational Perception. Bayesian Inference

Objective Inference for Climate Parameters: Bayesian, Transformation-of-Variables, and Profile Likelihood Approaches

Introduction. Start with a probability distribution f(y θ) for the data. where η is a vector of hyperparameters

Bayesian Phylogenetics:

Probability and Estimation. Alan Moses

Bayesian Inference. STA 121: Regression Analysis Artin Armagan

Lecture 2: From Linear Regression to Kalman Filter and Beyond

COMP90051 Statistical Machine Learning

Statistical Theory MT 2006 Problems 4: Solution sketches

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn

Bayesian analysis in nuclear physics

Estimation of reliability parameters from Experimental data (Parte 2) Prof. Enrico Zio

STAT 499/962 Topics in Statistics Bayesian Inference and Decision Theory Jan 2018, Handout 01

Overview. Probabilistic Interpretation of Linear Regression Maximum Likelihood Estimation Bayesian Estimation MAP Estimation

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007

Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak

Introduction into Bayesian statistics

Bayesian Methods. David S. Rosenberg. New York University. March 20, 2018

Part 2: One-parameter models

Bayesian Statistics. Debdeep Pati Florida State University. February 11, 2016

Unobservable Parameter. Observed Random Sample. Calculate Posterior. Choosing Prior. Conjugate prior. population proportion, p prior:

Inference Control and Driving of Natural Systems

Lecture 2: From Linear Regression to Kalman Filter and Beyond

The Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition.

The binomial model. Assume a uniform prior distribution on p(θ). Write the pdf for this distribution.

A Very Brief Summary of Bayesian Inference, and Examples

A Discussion of the Bayesian Approach

Density Estimation. Seungjin Choi

Bayesian Regression Linear and Logistic Regression

Noninformative Priors for the Ratio of the Scale Parameters in the Inverted Exponential Distributions

Introduction to Bayesian Methods

Approximate Bayesian Computation

Data Mining Chapter 4: Data Analysis and Uncertainty Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

Introduction to Bayesian Inference

Lecture 6: Model Checking and Selection

On the observational assessment of climate model performance

Minimum Message Length Analysis of the Behrens Fisher Problem

A Very Brief Summary of Statistical Inference, and Examples

Bayesian Inference. Chapter 1. Introduction and basic concepts

Probabilistic Reasoning in Deep Learning

Chapter 3: Maximum-Likelihood & Bayesian Parameter Estimation (part 1)

CS540 Machine learning L9 Bayesian statistics

Bayesian Methods in Positioning Applications

Foundations of Statistical Inference

Principles of Bayesian Inference

David Giles Bayesian Econometrics

Machine Learning 4771

Applied Bayesian Statistics STAT 388/488

Principles of Bayesian Inference

Expectation propagation as a way of life

Principles of Bayesian Inference

Lecture 6. Prior distributions

Inference for a Population Proportion

Data Analysis and Uncertainty Part 2: Estimation

Remarks on Improper Ignorance Priors

Bayesian Statistics. State University of New York at Buffalo. From the SelectedWorks of Joseph Lucke. Joseph F. Lucke

Bayesian Inference: Posterior Intervals

Bayesian Analysis (Optional)

Method of Feldman and Cousins for the construction of classical confidence belts

Lecture 1: Bayesian Framework Basics

Bayesian Inference: Concept and Practice

Bayesian Deep Learning

7. Estimation and hypothesis testing. Objective. Recommended reading

Persuading Skeptics and Reaffirming Believers

Statistical Tools and Techniques for Solar Astronomers

A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring

Lecture 5: Bayes pt. 1

Accouncements. You should turn in a PDF and a python file(s) Figure for problem 9 should be in the PDF

Statistics: Learning models from data

Introduction to Probabilistic Machine Learning

Bayesian Interpretations of Regularization

Module 22: Bayesian Methods Lecture 9 A: Default prior selection

SYDE 372 Introduction to Pattern Recognition. Probability Measures for Classification: Part I

Statistics of Small Signals

Classical and Bayesian inference

Frequentist Statistics and Hypothesis Testing Spring

Efficient Likelihood-Free Inference

A Bayesian Treatment of Linear Gaussian Regression

Statistical learning. Chapter 20, Sections 1 3 1

Statistical Data Analysis Stat 3: p-values, parameter estimation

Estimation Theory Fredrik Rusek. Chapter 11

Principles of Bayesian Inference

Machine Learning. Lecture 4: Regularization and Bayesian Statistics. Feng Li.

Linear Models A linear model is defined by the expression

Predictive Distributions

CS-E3210 Machine Learning: Basic Principles

Tutorial on Gaussian Processes and the Gaussian Process Latent Variable Model

A BAYESIAN MATHEMATICAL STATISTICS PRIMER. José M. Bernardo Universitat de València, Spain

Belief and Desire: On Information and its Value

Bayesian vs frequentist techniques for the analysis of binary outcome data

Bayes Factors, posterior predictives, short intro to RJMCMC. Thermodynamic Integration

Statistical Methods for Particle Physics Lecture 4: discovery, exclusion limits

Confidence Intervals Lecture 2

Transcription:

Bayesian parameter estimation with weak data and when combining evidence: the case of climate sensitivity Nicholas Lewis Independent climate scientist CliMathNet 5 July 2016

Main areas to be discussed Why Objective not Subjective Bayes is needed for parameter estimation How correctly to combine evidence in the Objective Bayesian case 2

Probability is not a settled field Probabilistic inference has a troubled history Bitter disputes and personality clashes in past Deep disagreements over fundamental issues A few main belief sets and multiple variants In parts more like philosophy than mathematics Bayesian inference disdained for much of 20 th C Subjective Bayes now strong: suits computers 3

Key figures in probability & statistics 4

Bayesian estimation: continuous case Usual, Subjective method: apply Bayes theorem Likelihood: probability density for data at observed value y, expressed as a function of the parameter θ Prior: estimated PDF for θ given current evidence Posterior PDF= Likelihood x Prior PDF, normalised p(θ y) = c p(y θ) p(θ); c set so that p( y) d 1 x% y% range: posterior CDF credible interval; CrI 5

Does Bayes theorem apply here? Bayes theorem restates conditional probability p(θ y) p(y) = p(y,θ) = p(y θ) p(θ); divide by p(y) Mathematically valid in the continuous case iff y and θ are Kolmogorov random variables (KRV) Bayes theorem gives mathematical probability iff parameter value is random with known prior PDF A fixed but unknown parameter is not a KRV Subjective Bayes provides coherent personal beliefs about parameters, not scientific validity 6

Objective Bayesian estimation Objective Bayes methods retain Bayesian form Objective Bayes uses noninformative prior (NIP) NIP lets info. in data dominate; typically CrI CI NIP is a weighting factor: no probability meaning 7

Noninformative priors Jeffreys prior: sqrt( Fisher information matrix ) NIP => inference invariant on reparameterisation NIP inter alia converts probability between data and parameter spaces: includes a Jacobian factor NIP is usually simple in data space often uniform Can use Bayes in data space and change variables to parameter space even if lower dimensional 8

Subjective vs Objective ECS estimation Many ECS estimates use Subjective Bayes Subjective ~ Objective Bayes if data strong ECS data too weak to dominate most priors 9

Subjective vs Objective ECS estimation Many ECS estimates use Subjective Bayes Subjective ~ Objective Bayes if data strong ECS data too weak to dominate most priors Subjective Bayesian posterior PDFs for ECS generally don t reflect data error distributions Uniform ECS & K v priors greatly bias estimates Expert ECS prior usually dominates over data 10

Estimating λ( 1/ECS): Gaussian errors 11

PDF converted (Jacobian) from λ to ECS 12

Jacobian for converting λ PDF to ECS 13

ECS PDF: effect of uniform prior vs NIP 14

3D: Jeffreys prior for ECS, & F aer Lewis (2013) An objective Bayesian improved approach... J Clim

Noninformative v Uniform priors: 3D 16

Dominance of expert priors 17

Combining ECS evidence Instrumental & paleo evidence ~ independent Standard Bayes method: combine by updating Final posterior= Π(likelihoods) x prior for 1 st est. 18

Bayesian updating is unsatisfactory Objective Bayes + updating: order-dependence! NIPs likely to differ for each source of evidence Objective Bayes & Bayesian updating incompatible 19

Avoiding Bayesian updating Solution: compute NIP for combined evidence Bayes on combined prior+combined likelihood To combine Jeffreys priors, add in quadrature Represent any prior info by data likelihood+nip 20

Objectively combining ECS evidence Bayes on combined prior+combined likelihood To combine Jeffreys priors, add in quadrature Apply to ECS estimation using parameterised distributions for which Jeffreys-like NIP known ECS ~ ratio of two normals: ΔT / (ΔF ΔQ) Good approximation to this distribution: RS93 Method gives CrI=CI for combined ratio-normals 21

Ratio-normal approximation fits 22

Testing RS93-based probability matching 23

Testing probability matching: combination 24

Combining recent & paleo evidence Represent each ECS estimate by RS93 formula RS93 3-parameter formula fits ECS PDFs well Median and range suffice to uniquely specify 25

Combining recent & paleo evidence Represent each ECS estimate by RS93 formula RS93 3-parameter formula fits ECS PDFs well Median and range suffice to uniquely specify Instrumental: Lewis & Curry 2015; Otto ea 2013 Paleo range: 1 6 K per AR5, as exactly 10 90% Paleo median: 10 AR5 estimates; median 2.75 K 26

Ratio-normal fits to ECS estimates 27

Original and combined likelihoods 28

Original+combined Jeffreys priors: log 29

Original and combined PDFs for ECS 30

Bayesian estimation: Conclusions Use Objective not Subjective Bayes if data weak Need to derive & compute noninformative prior Can infer true data values & change variables Add Jeffreys priors 2 to combine info: don t update Represent prior knowledge by data likelihood+nip 31

Thank you for listening Nicholas Lewis References Lewis N, 2016. Combining independent Bayesian posteriors into a confidence distribution, with application to estimating climate sensitivity. JSPI (accepted) Bernardo J, 2009. Modern Bayesian Inference: Foundations and Objective Methods. Lewis N, 2013. An objective Bayesian improved approach for applying optimal fingerprint techniques to estimate climate sensitivity. J Clim, 26, 7414-29 Lewis N, 2013. Modification of Bayesian Updating where Continuous Parameters have Differing Relationships with New and Existing Data. arxiv:1308.2791 Lewis N, Curry JA, 2015. The implications for climate sensitivity of AR5 forcing and heat uptake estimates. Clim Dyn, 45, 1009 1023 DOI 10.1007/s00382-014-2342-y Otto A et al, 2013. Energy budget constraints on climate response. Nat Geosci, 6, 415-6 32