Bayesian Evidence and Model Selection: A Tutorial in Two Acts

Similar documents
EXONEST The Exoplanetary Explorer. Kevin H. Knuth and Ben Placek Department of Physics University at Albany (SUNY) Albany NY

arxiv:astro-ph/ v1 14 Sep 2005

Strong Lens Modeling (II): Statistical Methods

Nested Sampling. Brendon J. Brewer. brewer/ Department of Statistics The University of Auckland

Multimodal Nested Sampling

2D Image Processing (Extended) Kalman and particle filter

Miscellany : Long Run Behavior of Bayesian Methods; Bayesian Experimental Design (Lecture 4)

Additional Keplerian Signals in the HARPS data for Gliese 667C from a Bayesian re-analysis

Expectation propagation for signal detection in flat-fading channels

Bayesian inference J. Daunizeau

Variational Methods in Bayesian Deconvolution

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision

Mobile Robot Localization

PATTERN RECOGNITION AND MACHINE LEARNING

Bayesian Dropout. Tue Herlau, Morten Morup and Mikkel N. Schmidt. Feb 20, Discussed by: Yizhe Zhang

arxiv: v1 [physics.ins-det] 18 Mar 2013

Nonparametric Bayesian Methods (Gaussian Processes)

A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring

Bayesian Model Selection & Extrasolar Planet Detection

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods

Unfolding techniques for neutron spectrometry

Bayesian Regression Linear and Logistic Regression

Hierarchical Bayesian Modeling

The Expectation Maximization or EM algorithm

An introduction to Bayesian inference and model comparison J. Daunizeau

Gaussian Process Approximations of Stochastic Differential Equations

Linear Dynamical Systems

SAMSI Astrostatistics Tutorial. More Markov chain Monte Carlo & Demo of Mathematica software

SRNDNA Model Fitting in RL Workshop

CIS 390 Fall 2016 Robotics: Planning and Perception Final Review Questions

Expectation Propagation for Approximate Bayesian Inference

Expectation Propagation Algorithm

Lecture 13 : Variational Inference: Mean Field Approximation

Robotics. Lecture 4: Probabilistic Robotics. See course website for up to date information.

Bayesian inference J. Daunizeau

Bayesian Inference. Chris Mathys Wellcome Trust Centre for Neuroimaging UCL. London SPM Course

Why Try Bayesian Methods? (Lecture 5)

Statistics. Lent Term 2015 Prof. Mark Thomson. 2: The Gaussian Limit

Efficient Likelihood-Free Inference

Expectation Maximization

13: Variational inference II

CSC 2541: Bayesian Methods for Machine Learning

Computer Vision Group Prof. Daniel Cremers. 14. Sampling Methods

Bayesian search for other Earths

ECE295, Data Assimila0on and Inverse Problems, Spring 2015

Integrated Non-Factorized Variational Inference

A Bayesian Analysis of Extrasolar Planet Data for HD 73526

Approximate Inference Part 1 of 2

Unsupervised Learning

STA 4273H: Statistical Machine Learning

Approximate Inference Part 1 of 2

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA

Part 1: Expectation Propagation

Recent advances in cosmological Bayesian model comparison

Approximate Inference

Foundations of Statistical Inference

Modeling a Sensor to Improve its Efficacy

Kalman filtering and friends: Inference in time series models. Herke van Hoof slides mostly by Michael Rubinstein

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions

Lecture 6: Model Checking and Selection

Cheng Soon Ong & Christian Walder. Canberra February June 2017

Statistical Methods for Astronomy

Approximate inference in Energy-Based Models

Lecture: Gaussian Process Regression. STAT 6474 Instructor: Hongxiao Zhu

Variational Principal Components

Advanced Statistical Methods. Lecture 6

Bayesian room-acoustic modal analysis

Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a

9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering

Lecture 5. G. Cowan Lectures on Statistical Data Analysis Lecture 5 page 1

Introduction to Machine Learning

IEOR E4570: Machine Learning for OR&FE Spring 2015 c 2015 by Martin Haugh. The EM Algorithm

Methods of Data Analysis Learning probability distributions

Physics 403. Segev BenZvi. Choosing Priors and the Principle of Maximum Entropy. Department of Physics and Astronomy University of Rochester

Week 3: The EM algorithm

Astronomy. Astrophysics. DIAMONDS: A new Bayesian nested sampling tool. Application to peak bagging of solar-like oscillations

Recent Advances in Bayesian Inference Techniques

Introduction to Bayesian Data Analysis

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Mobile Robot Localization

Machine Learning! in just a few minutes. Jan Peters Gerhard Neumann

Independent Component Analysis. Contents

Robotics. Mobile Robotics. Marc Toussaint U Stuttgart

Bayesian Inference in Astronomy & Astrophysics A Short Course

CSci 8980: Advanced Topics in Graphical Models Gaussian Processes

Quantitative Biology II Lecture 4: Variational Methods

Probabilistic Graphical Models for Image Analysis - Lecture 1

Lecture : Probabilistic Machine Learning

Clustering with k-means and Gaussian mixture distributions

Lecture 2: From Linear Regression to Kalman Filter and Beyond

REVEAL. Receiver Exploiting Variability in Estimated Acoustic Levels Project Review 16 Sept 2008

Computer Vision Group Prof. Daniel Cremers. 6. Mixture Models and Expectation-Maximization

Extra-solar Planets via Bayesian Fusion MCMC

Testing Restrictions and Comparing Models

Fast Likelihood-Free Inference via Bayesian Optimization

COMP90051 Statistical Machine Learning

Clustering with k-means and Gaussian mixture distributions

Mixtures of Gaussians. Sargur Srihari

Series 7, May 22, 2018 (EM Convergence)

ECE 5984: Introduction to Machine Learning

Transcription:

Bayesian Evidence and Model Selection: A Tutorial in Two Acts Kevin H. Knuth Depts. of Physics and Informatics, University at Albany, Albany NY USA Based on the paper: Knuth K.H., Habeck M., Malakar N.K., Mubeen A.M., Placek B. 2015. Bayesian evidence and model selection. In press at Digital Signal Processing. doi:10.1016/j.dsp.2015.06.012 DOWNLOAD TALK NOW: Google knuthlab Click Talks 7/19/2015 MaxEnt 2015 Tutorial 1

This tutorial follows the paper: Knuth K.H., Habeck M., Malakar N.K., Mubeen A.M., Placek B. 2015. Bayesian evidence and model selection. In press at Digital Signal Processing. doi:10.1016/j.dsp.2015.06.012 References are not provided in the talk slides, please consult the paper Equations in the talk are numbered in accordance with the paper When referencing anything from Act 1 of this talk, please reference this paper When referencing anything from Act 2 of this talk, please reference the slides 7/19/2015 MaxEnt 2015 Tutorial 2

Bayesian Evidence Odds Ratios Evidence, Model Order and Priors Numerical Techniques Laplace Approximation Importance Sampling Annealed Importance Sampling Variational Bayes Nested Sampling Applications Signal Detection : Brain Computer Interface / Neuroscience Sensor Characterization : Robotics / Signal Processing Exoplanet Characterization : Astronomy / Astrophysics Examples Nested Sampling Demo Nested Sampling and Phase Transitions 7/19/2015 MaxEnt 2015 Tutorial 3

7/19/2015 MaxEnt 2015 Tutorial 4

Bayesian Evidence 7/19/2015 MaxEnt 2015 Tutorial 5

Bayesian Evidence : Odds Ratios : oooo Bayes Theorem Posterior Probability Prior Probability Likelihood Evidence or Marginal Likelihood M represents a class of models represented by a set of model parameters m represents a particular model defined by a set of particular model parameter values d represents the acquired data 7/19/2015 MaxEnt 2015 Tutorial 6

Bayesian Evidence : Odds Ratios : oooo Bayesian Evidence The Bayesian evidence can be found by marginalizing the joint distribution P m, d M, I over all model parameter values. M represents a class of models represented by a set of model parameters m represents a particular model defined by a set of particular model parameter values d represents the acquired data I represents the dependence on any relevant prior information 7/19/2015 MaxEnt 2015 Tutorial 7

Bayesian Evidence : Odds Ratios : oooo Model Comparison We derive the ratio of probabilities of two models given data If the prior probabilities of the models are equal then this is the ratio of evidences 7/19/2015 MaxEnt 2015 Tutorial 8

Bayesian Evidence : Odds Ratios : oooo Odds Ratio or Bayes Factor The ratio of probabilities of the models given the data is proportional to the Odds Ratio 7/19/2015 MaxEnt 2015 Tutorial 9

Bayesian Evidence : Evidence, Model Order and Priors : ooooo Evidence: Model Order and Priors It is instructive to see how the evidence depends on both the model order and prior probabilities Consider a model with a single parameter: x ε [x min, x max ] with a width of x = x max x min Define the effective width where L max is the maximum likelihood value. 7/19/2015 MaxEnt 2015 Tutorial 10

Bayesian Evidence : Evidence, Model Order and Priors : ooooo Model with a Single Parameter Consider a model with a single parameter: x ε [x min, x max ] with a width of x = x max x min Given the effective width The evidence is 7/19/2015 MaxEnt 2015 Tutorial 11

Bayesian Evidence : Evidence, Model Order and Priors : ooooo Occam Factor The evidence is proportional to the ratio of the effective width of the likelihood and the width of the prior This ratio δx x is called the Occam factor after Occam s Razor: "Non sunt multiplicanda entia sine necessitate", " Entities must not be multiplied beyond necessity - William of Ockham 7/19/2015 MaxEnt 2015 Tutorial 12

Bayesian Evidence : Evidence, Model Order and Priors : ooooo Model Order For models with multiple parameters this generalizes to the ratio of the volume of the models that are compatible with both the data and the prior and the prior volume. If we assume that each of the K parameters has prior width x δx K then the Occam factor scales as. As model parameters added, eventually one fits the data asymptotically well so that δx attains a maximum value and further model parameters can only decrease the Occam factor. If we increase the flexibility of our model by the introduction of more model parameters, we reduce the Occam factor. x 7/19/2015 MaxEnt 2015 Tutorial 13

Bayesian Evidence : Evidence, Model Order and Priors : ooooo Odds Ratios and Occam Factors We compute the odds ratio for a model M 0 without model parameters to a model M 1 with a single model parameter. L max ratio Occam Factor The likelihood ratio is a classical statistic in frequentist model selection. If we only consider the likelihood ratio in model comparison problems, we fail to acknowledge the importance of Occam factors. 7/19/2015 MaxEnt 2015 Tutorial 14

Numerical Techniques 7/19/2015 MaxEnt 2015 Tutorial 15

Numerical Techniques : o Numerical Techniques There are a wide variety of techniques that can be used to estimate the Bayesian evidence: Laplace Approximation Importance Sampling Path Sampling Thermodynamic Integration Simulated Annealing Annealed Importance Sampling Variational Bayes (Ensemble Learning) Nested Sampling 7/19/2015 MaxEnt 2015 Tutorial 16

Numerical Techniques : Laplace Approximation : oooo Laplace Approximation is a simple and useful method for approximating a unimodal probability density function with a Gaussian Consider a function p x with a peak at x = x 0 We write a Taylor series expansion of ln p x about x = x 0 which can be simplified to 7/19/2015 MaxEnt 2015 Tutorial 17

Numerical Techniques : Laplace Approximation : oooo Laplace Approximation Previously, we had By defining We can write 7/19/2015 MaxEnt 2015 Tutorial 18

Numerical Techniques : Laplace Approximation : oooo Laplace Approximation Previously, we had By taking the exponential we can approximate the density by with an integral (evidence) of: 7/19/2015 MaxEnt 2015 Tutorial 19

Numerical Techniques : Laplace Approximation : oooo Laplace Approximation In the case of a multidimensional posterior we have where The evidence is then 7/19/2015 MaxEnt 2015 Tutorial 20

Numerical Techniques : Importance Sampling : oooo Importance Sampling allows one to find expectation values with respect to one distribution p(x) by computing expectation values with respect to a second distribution q(x) that is easier to sample from. The expectation value of f x with respect to p x is given by One can write p x as p x q x whenever p x is nonzero. q x as long as q x is non-zero 7/19/2015 MaxEnt 2015 Tutorial 21

Numerical Techniques : Importance Sampling : oooo Importance Sampling Writing p x as p x q x q x, we have: As long as the ratio p x q x estimate this with samples from q x by does not attain extreme values we can 7/19/2015 MaxEnt 2015 Tutorial 22

Numerical Techniques : Importance Sampling : oooo Importance Sampling Importance sampling can be used to compute ratios of evidence values in a similar fashion by writing 7/19/2015 MaxEnt 2015 Tutorial 23

Numerical Techniques : Importance Sampling : oooo Importance Sampling The evidence ratio can be found by sampling from q x as long as p x is sufficiently close to q x to avoid extreme ratios of p x q x 7/19/2015 MaxEnt 2015 Tutorial 24

Numerical Techniques : Variational Bayes : ooo Variational Bayes which is also known as ensemble learning, relies on approximating The posterior P m M, I with another distribution Q m. By defining the negative Free Energy And the Kullback-Leibler (KL) Divergence We can write 7/19/2015 MaxEnt 2015 Tutorial 25

Numerical Techniques : Variational Bayes : ooo Variational Bayes With this expression in hand We can show that the negative Free Energy is a lower bound to the evidence By minimizing the negative Free Energy, we can approximate the evidence 7/19/2015 MaxEnt 2015 Tutorial 26

Numerical Techniques : Variational Bayes : ooo Variational Bayes By choosing a distribution Q m that factorizes into where the set of parameters m 0 is disjoint from m 1 we can minimize the negative Free Energy and estimate the evidence by choosing where 7/19/2015 MaxEnt 2015 Tutorial 27

Numerical Techniques : Nested Sampling : ooooooo Nested Sampling was developed by John Skilling to stochastically integrate the posterior probability to obtain the evidence. Posterior estimates are used to obtain model parameter estimates. Nested sampling aims to estimate the cumulative distribution function of the density or states (DOS), which is the prior probability mass enclosed within a likelihood boundary. 7/19/2015 MaxEnt 2015 Tutorial 28

Numerical Techniques : Nested Sampling : ooooooo Nested Sampling Given a likelihood L, one can find the prior mass such that the likelihood of those states is greater than L Parameter Space 7/19/2015 MaxEnt 2015 Tutorial 29

Numerical Techniques : Nested Sampling : ooooooo Nested Sampling One can then estimate the evidence via stochastic integration using samples distributed according to the prior Likelihood integrated over Prior 7/19/2015 MaxEnt 2015 Tutorial 30

Numerical Techniques : Nested Sampling : ooooooo Nested Sampling One begins with a set of N samples. Use the sample with the lowest likelihood to define an implicit likelihood boundary. This results in an average decrease of the prior volume by 1/N Sample from the prior (uniformly is easiest) from within the implicit likelihood boundary to maintain N samples to estimate the evidence Z Keep track of L i (X i+1 X i ) i 7/19/2015 MaxEnt 2015 Tutorial 31

Numerical Techniques : Nested Sampling : ooooooo Nested Sampling Note how the prior volume contracts by 1/N each time. Early steps contribute little to the integral (Z) since the likelihood is very low. Later steps contribute little to (Z) since the prior volume change is very small. The steps that contribute most are in the middle of the sequence. 7/19/2015 MaxEnt 2015 Tutorial 32

Numerical Techniques : Nested Sampling : ooooooo Nested Sampling Since nested sampling contracts along the prior volume, it is relatively unaffected by local maxima in evidence (phase transitions). (See Figure A) Methods based on tempering, such as simulated annealing follow the slope of the log L curve and as such, get stuck at phase transitions. (See Figure B) 7/19/2015 MaxEnt 2015 Tutorial 33

Numerical Techniques : Nested Sampling : ooooooo Nested Sampling The great challenge is sampling uniformly (from the prior) within the implicit Likelihood boundaries. Several versions of Nested Sampling now exist: MultiNest (developed by Feroz and Hobson): clusters samples (K-means) and fits clusters with ellipsoids. Samples uniformly from within those ellipsoids. Very fast. Excellent performance for multi-modal distributions. Clustering limits this to 10s of parameters and the ellipsoids may not cover the high likelihood regions. Galilean Monte Carlo (Developed by Feroz and Skilling): moves a new sample with momentum reflecting off of logl boundaries. Excellent at handling ridges both angled and curved. Constrained Hamilton Monte Carlo (developed by M. Betancourt): similar to Galilean Monte Carlo. Diffusive Nested Sampling (developed by Brewer): allows samples to diffuse to lower likelihood nested levels and takes a weighted average. Nested Sampling with Demons (developed by M. Habeck): utilizes demon variables that smooth the constraint boundary and push the samples away from it. 7/19/2015 MaxEnt 2015 Tutorial 34

7/19/2015 MaxEnt 2015 Tutorial 35

Applications: Signal Detection: oooooooo Signal Detection Brain Computer Interface / Neuroscience 7/19/2015 MaxEnt 2015 Tutorial 36

Applications: Signal Detection: oooooooo Signal Detection (Mubeen and Knuth) We consider a practical signal detection problem where the log odds-ratio can be analytically derived. The specific application was originally for the detection of evoked brain responses The signal absent case models the recording x in channel m as noise The signal present case models the recording x in channel m as signal plus where the signal has a amplitude parameter α and can be coupled differently to different detectors (via C) 7/19/2015 MaxEnt 2015 Tutorial 37

Applications: Signal Detection: oooooooo Considering the Evidence The odds ratio can be written as For the noise only case, the evidence is the likelihood (Gaussian) 7/19/2015 MaxEnt 2015 Tutorial 38

Applications: Signal Detection: oooooooo Considering the Evidence In the signal plus noise case, the evidence is Assigning a Gaussian Likelihood and Prior for α 7/19/2015 MaxEnt 2015 Tutorial 39

Applications: Signal Detection: oooooooo Considering the Evidence We can then write the evidence as where 7/19/2015 MaxEnt 2015 Tutorial 40

Applications: Signal Detection: oooooooo Considering the Evidence If the signal amplitude must be positive: α ε 0, + then: If amplitude can be positive or negative: α ε, + then: 7/19/2015 MaxEnt 2015 Tutorial 41

Applications: Signal Detection: oooooooo Considering the Evidence Look at: The expression E (86) contains the cross-correlation term, which is what is typically used for the detection of a target signal in ongoing recordings. The log OR detection filters incorporate more information that leads to extra terms, which serve to aid in target signal detection. 7/19/2015 MaxEnt 2015 Tutorial 42

Applications: Signal Detection: oooooooo Detecting Signals A. The P300 template target signal. B. An example of three channels (Cz, Pz, Fz) of synthetic ongoing EEG with two P300 target signal events (indicated by the arrows) at an SNR of 5 db. 7/19/2015 MaxEnt 2015 Tutorial 43

Applications: Signal Detection: oooooooo Signal Detection Performance Detection performance measured by that area under the ROC curve as a function of signal SNR. Both OR techniques outperform cross-correlation! 7/19/2015 MaxEnt 2015 Tutorial 44

Applications: Signal Detection: oooo Sensor Characterization Robotics / Signal Processing 7/19/2015 MaxEnt 2015 Tutorial 45

Applications: Signal Detection: oooo Modeling a Robotic Sensor (Malakar, Gladkov, Knuth) In this project, we aim to model the spatial sensitivity function of a LEGO light sensor for use on a robotic system. Here the application is to develop a robotic arm that can characterize the white circle by measuring light intensities are various locations. By modeling the light sensor, we aim to increase the robot s performance. 7/19/2015 MaxEnt 2015 Tutorial 46

Applications: Signal Detection: oooo Modeling a Robotic Sensor The LEGO light sensor was slowly moved over a black-and-white albedo pattern on the surface of a table to obtain calibration data. Sensor orientation was varied as well. 7/19/2015 MaxEnt 2015 Tutorial 47

Applications: Signal Detection: oooo Modeling a Robotic Sensor Mixture of Gaussians models were used. Four model orders were tested using Nested Sampling. The 1-MoG model was slightly favored. Note the increasing uncertainty as the model becomes more complex. This suggests that the permutation space was not fully explored. 7/19/2015 MaxEnt 2015 Tutorial 48

Applications: Signal Detection: oooo Examining the Sensor Model Performance Here we show a comparison between the 1-MoG model and the data 7/19/2015 MaxEnt 2015 Tutorial 49

Applications: Star System Characterization: ooooooo Star System Characterization Astronomy / Astrophysics 7/19/2015 MaxEnt 2015 Tutorial 50

Applications: Star System Characterization: ooooooo Star System Characterization (Placek and Knuth) In our DSP paper, we give an example of Bayesian model testing applied to exoplanet characterization. Ben Placek also has a paper and poster here at MaxEnt 2015 on the topic. Here I will apply these model testing concepts to determining the orbital configuration of a triple star system. Digital Sky Survey (DSS) 7/19/2015 MaxEnt 2015 Tutorial 51

Applications: Star System Characterization: ooooooo KIC 5436161: Two Periods This star exhibits oscillations of two commensurate periods in its Light curve: 6.45 days and 0.645 days (a rare 10:1 resonance!) Digital Sky Survey (DSS) Photometric data obtained from the Kepler mission (A) Quarter 13 light curve folded on the P1 = 6.45 day period, (B) Quarter 13 light curve folded on the P2 = 0.645 day period (C) is the entire Q13 light curve. 7/19/2015 MaxEnt 2015 Tutorial 52

Applications: Star System Characterization: ooooooo KIC 5436161: Radial Velocity Measurements Eleven radial velocity measurements taken over the span of a week. The 6.45 day period is visible, but not the 0.645 day period. Courtesy of Geoff Marcy and Howard Issacson 7/19/2015 MaxEnt 2015 Tutorial 53

Applications: Star System Characterization: ooooooo KIC 5436161: Models Two possible models of the system. The main star is a G-star (like our sun), at least one of the other companions (C1) is M-dwarf. (A) A hierarchical arrangement (C1 and C2 orbit G with 6.45 day period, and orbit one another with 0.645 day period) (B) A planetary arrangement (C1 orbits with 6.45 day period, and C2 orbits with 0.645 day period) 7/19/2015 MaxEnt 2015 Tutorial 54

Applications: Star System Characterization: ooooooo KIC 5436161: Results Testing the Hierarchical Model against the Planetary Model using the Radial Velocity Data. The Circular Hierarchical Model has the greatest evidence (by a factor of exp 3.73 42) 7/19/2015 MaxEnt 2015 Tutorial 55

Applications: Star System Characterization: ooooooo KIC 5436161 This system is a hierarchical triple system consisting of a G-star with two co-orbiting M-dwarfs in a 1:10 resonance (P1 = 6.45 day, P2 = 0.645 day) 7/19/2015 MaxEnt 2015 Tutorial 56

Applications: Star System Characterization: ooooooo KIC 5436161 7/19/2015 MaxEnt 2015 Tutorial 57

Demonstrations: Nested Sampling: Lighthouse Problem: ooooo Nested Sampling Demo (sans model testing) 7/19/2015 MaxEnt 2015 Tutorial 58

Demonstrations: Nested Sampling: Lighthouse Problem: ooooo Nested Sampling Demo The Lighthouse Problem (Gull) Consider a Lighthouse located just off of a straight shore that extends a great distance. Imagine that the lighthouse has a laser beam that it fires at random times as it rotates with a uniform speed. Along the shore are light detectors that detect laser beam hits. Based on this data, where is the lighthouse? 7/19/2015 MaxEnt 2015 Tutorial 59

Demonstrations: Nested Sampling: Lighthouse Problem: ooooo The Likelihood Function It is a useful exercise to derive the likelihood via a change of variables. p x I = β π β 2 + α x 2 We assign a uniform prior for the location parameters α and β x = β tan θ + α 7/19/2015 MaxEnt 2015 Tutorial 60

y position Demonstrations: Nested Sampling: Lighthouse Problem: ooooo Nested Sampling Run using D = 64 data points (recorded flashes) and N = 100 samples Iteration is halted when Δ log Z = 10 7 o Live Samples + Used Samples Location of Lighthouse # Iterations = 1193 Log Z = -0.401 +- 0.076 mean(x) = 0.48 ± 0.26 mean(y) = 0.51 ± 0.28 x position 7/19/2015 MaxEnt 2015 Tutorial 61

Demonstrations: Nested Sampling: Lighthouse Problem: ooooo Nested Sampling Run This shows the relationship between log L and log Prior Volume 7/19/2015 MaxEnt 2015 Tutorial 62

Demonstrations: Nested Sampling: Lighthouse Problem: ooooo Nested Sampling Run with a Gaussian Likelihood This shows the relationship between log L and log Prior Volume 7/19/2015 MaxEnt 2015 Tutorial 63

Applications: Nested Sampling and Phase Transitions: oooo Nested Sampling and Phase Transitions 7/19/2015 MaxEnt 2015 Tutorial 64

Applications: Nested Sampling and Phase Transitions: oooo Peaks on Peaks Here is a Gaussian Likelihood with a taller peak on the side 7/19/2015 MaxEnt 2015 Tutorial 65

Applications: Nested Sampling and Phase Transitions: oooo Nested Sampling with Phase Transitions Phase Transitions represent local peaks in the evidence Phase Transition 7/19/2015 MaxEnt 2015 Tutorial 66

Applications: Nested Sampling and Phase Transitions: oooo Acoustic Source Localization: One Detector Consider an acoustic source localization problem using a single detector. There is a low (red) and a high (blue) frequency source. Note how the high frequency source is found first inducing a phase transition: 7/19/2015 MaxEnt 2015 Tutorial 67

Applications: Nested Sampling and Phase Transitions: oooo Acoustic Source Localization: Two Detectors In this example, we have two detectors, which allow us to localize the sources to rings. Again, the low frequency source is found first. 7/19/2015 MaxEnt 2015 Tutorial 68

Acknowledgements Michael Habeck Nabin Malakar Asim Mubeen Ben Placek 7/19/2015 MaxEnt 2015 Tutorial 69