Luminosity Functions

Similar documents
X-ray Dark Sources Detection

pyblocxs Bayesian Low-Counts X-ray Spectral Analysis in Sherpa

Time Delay Estimation: Microlensing

Two Statistical Problems in X-ray Astronomy

Overlapping Astronomical Sources: Utilizing Spectral Information

Bayesian Estimation of log N log S

Accounting for Calibration Uncertainty

Bayesian Estimation of log N log S

Statistical Tools and Techniques for Solar Astronomers

Markov Chain Monte Carlo

Fitting Narrow Emission Lines in X-ray Spectra

Bayesian Modeling of log N log S

Accounting for Calibration Uncertainty in Spectral Analysis. David A. van Dyk

Part 7: Hierarchical Modeling

Lecturer: David Blei Lecture #3 Scribes: Jordan Boyd-Graber and Francisco Pereira October 1, 2007

Accounting for Calibration Uncertainty in Spectral Analysis. David A. van Dyk

Making rating curves - the Bayesian approach

Bayesian data analysis in practice: Three simple examples

Embedding Supernova Cosmology into a Bayesian Hierarchical Model

Bayesian Computation in Color-Magnitude Diagrams

Closed-form Gibbs Sampling for Graphical Models with Algebraic constraints. Hadi Mohasel Afshar Scott Sanner Christfried Webers

Ages of stellar populations from color-magnitude diagrams. Paul Baines. September 30, 2008

Bayesian Model Comparison:

Classical and Bayesian inference

F denotes cumulative density. denotes probability density function; (.)

Robust Bayesian Simple Linear Regression

Monte Carlo integration

Large Scale Bayesian Inference

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

Learning the hyper-parameters. Luca Martino

Part 6: Multivariate Normal and Linear Models

Dynamic Factor Models and Factor Augmented Vector Autoregressions. Lawrence J. Christiano

Chapter 8: Sampling distributions of estimators Sections

eqr094: Hierarchical MCMC for Bayesian System Reliability

CTDL-Positive Stable Frailty Model

Advances and Applications in Perfect Sampling

STA 4273H: Statistical Machine Learning

Principles of Bayesian Inference

CS281A/Stat241A Lecture 22

Computational statistics

Fitting Narrow Emission Lines in X-ray Spectra

Metropolis Hastings. Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601. Module 9

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution

Embedding Astronomical Computer Models into Complex Statistical Models. David A. van Dyk

High Energy Astrophysics:

Markov Chain Monte Carlo Methods

MCMC 2: Lecture 2 Coding and output. Phil O Neill Theo Kypraios School of Mathematical Sciences University of Nottingham

A short diversion into the theory of Markov chains, with a view to Markov chain Monte Carlo methods

Embedding Astronomical Computer Models into Principled Statistical Analyses. David A. van Dyk

Computer intensive statistical methods

Multivariate Normal & Wishart

Point spread function reconstruction from the image of a sharp edge

Introduction to Machine Learning CMU-10701

Introduction to Stochastic Gradient Markov Chain Monte Carlo Methods

Bayesian Analysis of RR Lyrae Distances and Kinematics

Statistics/Mathematics 310 Larget April 14, Solution to Assignment #9

Labor-Supply Shifts and Economic Fluctuations. Technical Appendix

HPD Intervals / Regions

Down by the Bayes, where the Watermelons Grow

LECTURE 15 Markov chain Monte Carlo

Hierarchical Bayesian Modeling with Approximate Bayesian Computation

Bayesian Nonparametric Regression for Diabetes Deaths

Calibration of Stochastic Volatility Models using Particle Markov Chain Monte Carlo Methods

Doing Bayesian Integrals

Part 2: One-parameter models

Outline. Clustering. Capturing Unobserved Heterogeneity in the Austrian Labor Market Using Finite Mixtures of Markov Chain Models

Embedding Astronomical Computer Models into Complex Statistical Models. David A. van Dyk

Bayesian Graphical Models

Introduction to Bayesian Methods

MCMC: Markov Chain Monte Carlo

MARKOV CHAIN MONTE CARLO

Measurement Error and Linear Regression of Astronomical Data. Brandon Kelly Penn State Summer School in Astrostatistics, June 2007

Bayes via forward simulation. Approximate Bayesian Computation

Bayesian Estimation of Input Output Tables for Russia

Hierarchical models. Dr. Jarad Niemi. August 31, Iowa State University. Jarad Niemi (Iowa State) Hierarchical models August 31, / 31

Using Bayes Factors for Model Selection in High-Energy Astrophysics

Lecture 13 Fundamentals of Bayesian Inference

The Jackknife-Like Method for Assessing Uncertainty of Point Estimates for Bayesian Estimation in a Finite Gaussian Mixture Model

MCMC algorithms for fitting Bayesian models

Markov Chain Monte Carlo (MCMC)

Markov Chain Monte Carlo, Numerical Integration

Bayesian Linear Regression

David Giles Bayesian Econometrics

Introduction to Bayesian methods in inverse problems

Forward Problems and their Inverse Solutions

Spatial Normalized Gamma Process

Parameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1

Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC

Bayesian Inference and Decision Theory

Hierarchical Bayesian Modeling

Advanced Machine Learning

36-463/663Multilevel and Hierarchical Models

Stat 516, Homework 1

CMPS 242: Project Report

Bayesian Prediction of Code Output. ASA Albuquerque Chapter Short Course October 2014

Bayesian Methods and Uncertainty Quantification for Nonlinear Inverse Problems

Conjugate Analysis for the Linear Model

Dynamic models. Dependent data The AR(p) model The MA(q) model Hidden Markov models. 6 Dynamic models

18 : Advanced topics in MCMC. 1 Gibbs Sampling (Continued from the last lecture)

Example: Ground Motion Attenuation

Transcription:

Department of Statistics, Harvard University May 15th, 2012

Introduction: What is a Luminosity Function? Figure: A galaxy cluster.

Introduction: Project Goal The Luminosity Function specifies the relative number of sources at each luminosity.

Introduction: Project Goal The Luminosity Function specifies the relative number of sources at each luminosity. Goal of the Project: To develop a fully Bayesian model to infer the distribution of the luminosities of all the sources in a population.

Introduction: Data Y i, observed photon counts, contaminated with background in a source exposure. X, observed photon counts in the exposure of pure background.

Level I model: X ξ Pois(ξ), Y i = Y ib + Y is, where Y ib ξ Pois(a i ξ), { δ 0, if λ i = 0; Y is λ i Pois(b i λ i ) Pois(b i λ i ), if λ i 0. ξ is the background intensity, λ i is the intensity of source i, a i is ratio of source area to background area (known constant), b i is the telescope effective area (known constant).

Level II model: ξ Gamma(α 0, β 0 ), { = 0, with probability 1 π; λ i α, β, π Gamma(α, β), with probability π. Level III model: P(α, β, π) 1 β 3 πc 1 1 (1 π) c 2 1.

: Summary Model X ξ Pois(ξ), Y i = Y ib + Y is, Y is ξ Pois(a i ξ), Y ib λ i Pois(b i λ i ), ξ Gamma(α 0, β 0 ), { = 0, with probability 1 π, λ i α, β, π Gamma(α, β), with probability π, P(α, β, π) 1 β 3. Research interest: The posterior distribution of intensities λ, The posterior distribution of 1 π, the proportion of dark sources.

Simulation Study 1: Setup Number of sources, N=500. Distribution of λ s: λ i iid { δ 0, with prob 1 π = 0.15, Gamma[10, 35] with prob 0.85. Distribution of Y is : Y is λ i Pois(b i λ i ). Distribution of background noise Y ib : Y i = Y is + Y ib. 0 10 20 30 40 Observed Photon Counts 0 10 20 30 40 Y ib ξ Pois(4), approximately.

Posterior Distributions of the Hyper-parameters posterior draws of alpha/beta posterior draws of alpha/beta^2 posteriro draws of 1-pi Density Density 0.00 0.02 0.04 0.06 0.08 0.10 Density 0 5 10 15 9.0 9.5 10.0 10.5 11.0 11.5 25 30 35 40 45 50 55 0.10 0.15 0.20 0.25

Z Posterior Distributions of the Hyper-parameters Scatter Plots of Posterior Draws Joint Posterior Density of pi and alpha/beta The True Value Posterior Draws pi 0.75 0.80 0.85 0.90 pi alpha/beta 9.0 9.5 10.0 10.5 11.0 11.5 alpha/beta

Distribution of source intensities Distribution of source intensities λ i iid { δ 0, with prob 1 π, Gamma(α, β), with prob π. 0.00 0.05 0.10 0.15 0.20 0.25 0 5 10 15 20 25

Simulation Study 2: Setup Number of sources, N=500. Distribution of λ s: λ i iid { δ 0, with prob 1 π = 0.15, Gamma[2, 60] with prob 0.85. Distribution of Y is : Y is λ i Pois(b i λ i ). Y ib ξ Pois(2), approximately. 0 50 100 150 200 250 Observed Photon Counts 0 20 40 60 80 100

Simulation Study 2 posterior draws of alpha/beta posterior draws of alpha/beta^2 posteriro draws of 1-pi Density 0.00 0.05 0.10 0.15 Density 0.0000 0.0005 0.0010 0.0015 0.0020 0.0025 0.0030 0.0035 Density 0.0 0.5 1.0 1.5 2.0 5 10 15 20 0 500 1000 1500 0.0 0.2 0.4 0.6 0.8

Z Z Z Z Simulation Study 2 Scatter Plots of Posterior Draws pi 0.2 0.4 0.6 0.8 1.0 The True Value Posterior Draws alpha/beta pi alpha/beta pi 5 10 15 20 pi alpha/beta pi alpha/beta alpha/beta

Simulation Study 2 Distribution of source intensities 0.0 0.5 1.0 1.5 2.0 2.5

Choices of Priors for the Hyper-parameters Recall λ i α, β, π iid { δ0, with prob 1 π, Gamma[ α β, α ] = Gamma[µ, µ β 2 β ] with prob π.

Choices of Priors for the Hyper-parameters Recall λ i α, β, π iid { δ0, with prob 1 π, Gamma[ α β, α ] = Gamma[µ, µ β 2 β ] with prob π. P(µ, β, π)dµdβdπ P(β)P(π)dµdβdπ 1 β P(β)P(π)dαdβdπ, P(α, β, π)dαdβdπ 1 β c 3+1 πc 1 1 (1 π) c 2 1 dαdβdπ

Choices of Priors: Simulation Study π Beta(c 1, c 2 ), we can use informative prior if we have some prior information about the distribution of π. Otherwise, we can let c 1 = c 2 = 1, so π Unif (0, 1). Priors for (α, β): Prior 0: P(α, β) 1, Prior 1: P(α, β) 1 β, Prior 2: P(α, β) 1 β 2, Prior 3: P(α, β) 1 β 3.

Choices of Priors: Coverage for π prior0 prior1 prior2 prior3 Gamma[2,60] Gamma[4,100] Gamma[8,200] Gamma[2,20] Gamma[4,30] Gamma[8,50] Gamma[2,2] Gamma[4,5] Gamma[8,8]

Choices of Priors: Coverage for α β prior0 prior1 prior2 prior3 Gamma[2,60] Gamma[4,100] Gamma[8,200] Gamma[2,20] Gamma[4,30] Gamma[8,50] Gamma[2,2] Gamma[4,5] Gamma[8,8]

Choices of Priors Conclusion: the prior P(α, β) 1 β 3 gives the highest frequency coverage in most simulation studies. This prior is called Stein s Harmonic Prior. The SHP prior is shown to provide estimators that have adequate frequency coverage.

Speeding up MCMC It takes about 3 hours to get 100,000 draws. Metropolis-Hastings algorithm within Gibbs Sampler: P(α β, π, ξ, λ, Y B, X, Ỹ ) where K = 1 λi 0, and λ = ( i,λ i 0 λ i) 1/K. ( (βλ ) α ) K, Γ(α) Is there a good way to sample from the conditional posterior distribution?

Acknowledgement Vinay Kashyap, David van Dyk, Andreas Zezas.