Doing Cosmology with Balls and Envelopes

Similar documents
New Approaches to False Discovery Control

A Large-Sample Approach to Controlling the False Discovery Rate

Confidence Thresholds and False Discovery Control

New Procedures for False Discovery Control

Controlling the False Discovery Rate: Understanding and Extending the Benjamini-Hochberg Method

Exceedance Control of the False Discovery Proportion Christopher Genovese 1 and Larry Wasserman 2 Carnegie Mellon University July 10, 2004

Uniform Confidence Sets for Nonparametric Regression with Application to Cosmology

A Stochastic Process Approach to False Discovery Rates Christopher Genovese 1 and Larry Wasserman 2 Carnegie Mellon University January 7, 2003

Uniform Confidence Sets for Nonparametric Regression with Application to Cosmology

False Discovery Control in Spatial Multiple Testing

Nonparametric Inference in Cosmology and Astrophysics: Biases and Variants

FDR-CONTROLLING STEPWISE PROCEDURES AND THEIR FALSE NEGATIVES RATES

Weighted Adaptive Multiple Decision Functions for False Discovery Rate Control

Estimation of a Two-component Mixture Model

A GENERAL DECISION THEORETIC FORMULATION OF PROCEDURES CONTROLLING FDR AND FNR FROM A BAYESIAN PERSPECTIVE

False discovery rate and related concepts in multiple comparisons problems, with applications to microarray data

False Discovery Rates for Random Fields

Post-Selection Inference

More powerful control of the false discovery rate under dependence

FALSE DISCOVERY AND FALSE NONDISCOVERY RATES IN SINGLE-STEP MULTIPLE TESTING PROCEDURES 1. BY SANAT K. SARKAR Temple University

High-throughput Testing

Table of Outcomes. Table of Outcomes. Table of Outcomes. Table of Outcomes. Table of Outcomes. Table of Outcomes. T=number of type 2 errors

Multiple Testing Procedures under Dependence, with Applications

The miss rate for the analysis of gene expression data

Modified Simes Critical Values Under Positive Dependence

Simultaneous Testing of Grouped Hypotheses: Finding Needles in Multiple Haystacks

Heterogeneity and False Discovery Rate Control

arxiv: v1 [math.st] 31 Mar 2009

Statistical Applications in Genetics and Molecular Biology

Non-specific filtering and control of false positives

STAT 461/561- Assignments, Year 2015

Estimation and Confidence Sets For Sparse Normal Mixtures

Alpha-Investing. Sequential Control of Expected False Discoveries

Statistical testing. Samantha Kleinberg. October 20, 2009

Large-Scale Hypothesis Testing

Nonparametric Inference and the Dark Energy Equation of State

Stat 710: Mathematical Statistics Lecture 31

Journal of Statistical Software

Peak Detection for Images

Applying the Benjamini Hochberg procedure to a set of generalized p-values

3 Comparison with Other Dummy Variable Methods

Controlling Bayes Directional False Discovery Rate in Random Effects Model 1

A Bayesian Determination of Threshold for Identifying Differentially Expressed Genes in Microarray Experiments

Two-stage stepup procedures controlling FDR

A Sequential Bayesian Approach with Applications to Circadian Rhythm Microarray Gene Expression Data

Rejoinder on: Control of the false discovery rate under dependence using the bootstrap and subsampling

Improved Astronomical Inferences via Nonparametric Density Estimation

Multiple Testing. Hoang Tran. Department of Statistics, Florida State University

Optimal Rates of Convergence for Estimating the Null Density and Proportion of Non-Null Effects in Large-Scale Multiple Testing

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model

On Methods Controlling the False Discovery Rate 1

Summary and discussion of: Controlling the False Discovery Rate: A Practical and Powerful Approach to Multiple Testing

The optimal discovery procedure: a new approach to simultaneous significance testing

Hunting for significance with multiple testing

Asymptotic Results on Adaptive False Discovery Rate Controlling Procedures Based on Kernel Estimators

Control of Generalized Error Rates in Multiple Testing

Optional Stopping Theorem Let X be a martingale and T be a stopping time such

Bayesian Determination of Threshold for Identifying Differentially Expressed Genes in Microarray Experiments

Sanat Sarkar Department of Statistics, Temple University Philadelphia, PA 19122, U.S.A. September 11, Abstract

PROCEDURES CONTROLLING THE k-fdr USING. BIVARIATE DISTRIBUTIONS OF THE NULL p-values. Sanat K. Sarkar and Wenge Guo

The Pennsylvania State University The Graduate School Eberly College of Science GENERALIZED STEPWISE PROCEDURES FOR

Controlling the Proportion of Falsely Rejected Hypotheses when Conducting Multiple Tests with Climatological Data

Multiple testing with the structure-adaptive Benjamini Hochberg algorithm

EMPIRICAL BAYES METHODS FOR ESTIMATION AND CONFIDENCE INTERVALS IN HIGH-DIMENSIONAL PROBLEMS

Aliaksandr Hubin University of Oslo Aliaksandr Hubin (UIO) Bayesian FDR / 25

Controlling the proportion of falsely-rejected hypotheses. when conducting multiple tests with climatological data

Selective Inference for Effect Modification

Large-Scale Multiple Testing of Correlations

False Discovery Rate

The Pennsylvania State University The Graduate School A BAYESIAN APPROACH TO FALSE DISCOVERY RATE FOR LARGE SCALE SIMULTANEOUS INFERENCE

High-Throughput Sequencing Course. Introduction. Introduction. Multiple Testing. Biostatistics and Bioinformatics. Summer 2018

Model Identification for Wireless Propagation with Control of the False Discovery Rate

Probabilistic Inference for Multiple Testing

H 2 : otherwise. that is simply the proportion of the sample points below level x. For any fixed point x the law of large numbers gives that

Stat 206: Estimation and testing for a mean vector,

Statistical Applications in Genetics and Molecular Biology

Incorporation of Sparsity Information in Large-scale Multiple Two-sample t Tests

Estimating False Discovery Proportion Under Arbitrary Covariance Dependence

FDR and ROC: Similarities, Assumptions, and Decisions

Rank conditional coverage and confidence intervals in high dimensional problems

Political Science 236 Hypothesis Testing: Review and Bootstrapping

Proportion of non-zero normal means: universal oracle equivalences and uniformly consistent estimators

On adaptive procedures controlling the familywise error rate

A Simple, Graphical Procedure for Comparing Multiple Treatment Effects

Resampling-Based Control of the FDR

Control of Directional Errors in Fixed Sequence Multiple Testing

Large-Scale Multiple Testing of Correlations

Spring 2012 Math 541B Exam 1

Positive false discovery proportions: intrinsic bounds and adaptive control

Testing Jumps via False Discovery Rate Control

Better Bootstrap Confidence Intervals

Factor-Adjusted Robust Multiple Test. Jianqing Fan (Princeton University)

STEPDOWN PROCEDURES CONTROLLING A GENERALIZED FALSE DISCOVERY RATE. National Institute of Environmental Health Sciences and Temple University

Adaptive False Discovery Rate Control for Heterogeneous Data

arxiv: v2 [stat.me] 14 Mar 2011

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided

Step-down FDR Procedures for Large Numbers of Hypotheses

STAT 263/363: Experimental Design Winter 2016/17. Lecture 1 January 9. Why perform Design of Experiments (DOE)? There are at least two reasons:

Marginal Screening and Post-Selection Inference

Nonconcave Penalized Likelihood with A Diverging Number of Parameters

Transcription:

Doing Cosmology with Balls and Envelopes Christopher R. Genovese Department of Statistics Carnegie Mellon University http://www.stat.cmu.edu/ ~ genovese/ Larry Wasserman Department of Statistics Carnegie Mellon University This work partially supported by NSF Grant SES 9866147 and ITR 0104016.

Overview: Two Talks in One Envelopes Situation: Performing many simultaneous hypothesis tests Problem: Attain needed power while still controlling false discoveries in some principled way. Approach: Bound the proportion of false discoveries among rejected nulls with high probability. Balls Situation: Have noisy samples of an unknown function. Problem: Make inferences about various features of the function. Approach: Construct uniformly valid confidence sets for the unknown function.

Notation Also, focus on the blue stuff. EX X θ estimate of θ sup max inf min

Road Map: Envelopes 1. The Multiple Testing Problem Idea and Examples Error Criteria 2. Controlling FDR The Benjamini-Hochberg Procedure Increasing Power 3. Confidence Envelopes and Thresholds Exact Confidence Envelopes for the False Discovery Proportion Choice of Tests 4. False Discovery Control for Random Fields Confidence Supersets and Thresholds Controlling the Proportion of False Clusters

Road Map: Envelopes 1. The Multiple Testing Problem Idea and Examples Error Criteria 2. Controlling FDR The Benjamini-Hochberg Procedure Increasing Power 3. Confidence Envelopes and Thresholds Exact Confidence Envelopes for the False Discovery Proportion Choice of Tests 4. False Discovery Control for Random Fields Confidence Supersets and Thresholds Controlling the Proportion of False Clusters

The Multiple Testing Problem Perform m simultaneous hypothesis tests with a common procedure. For any given threshold, classify the results as follows: H 0 Retained H 0 Rejected Total H 0 True TN FD T 0 H 0 False FN TD T 1 Total N D m Mnemonics: T/F = True/False, D/N = Discovery/Nondiscovery All quantities except m, D, and N are unobserved. The problem is to choose a threshold that balances the competing demands of sensitivity and specificity.

Motivating Examples fmri Data Astronomical Source Detection DNA Microarrays Scan Statistics These all involve many thousands of tests and interesting spatial structure.

How to Choose a Threshold? Control Per-Comparison Type I Error a.k.a. uncorrected testing, many type I errors { Gives P 0 FDi > 0 } α marginally for all 1 i m Strong Control of Familywise Type I Error e.g.: Bonferroni: use per-comparison significance level α/m Guarantees P 0 { FD > 0 } α False Discovery Control e.g.: Benjamini & Hochberg (BH, 1995, 2000): False Discovery Rate (FDR) ( ) FD Guarantees FDR E α D

Road Map: Envelopes 1. The Multiple Testing Problem Idea and Examples Error Criteria 2. Controlling FDR The Benjamini-Hochberg Procedure Increasing Power 3. Confidence Envelopes and Thresholds Exact Confidence Envelopes for the False Discovery Proportion Choice of Tests 4. False Discovery Control for Random Fields Confidence Supersets and Thresholds Controlling the Proportion of False Clusters

The Benjamini-Hochberg Procedure Convenient to work with p-values p/2 p/2 p Test Statistic Test Statistic Test Statistic Given m p-values ordered 0 P (0) <P (1) < <P (m), the BH procedure rejects any null hypothesis with P j T BH where { T BH = max P (i) : P (i) α i }. m

The Benjamini-Hochberg Procedure (cont d) p value 0.0 0.2 0.4 0.6 0.8 m = 50, α =0.1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 1 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0.0 0.2 0.4 0.6 0.8 1.0 Index/m

The Benjamini-Hochberg Procedure (cont d) ( ) FD BH guarantees that FDR E T 0 D m α. Gives more power than Bonferroni, fewer Type I errors than uncorrected testing. If Ĝ is the empirical cdf of the m p-values, Ĝ(P (i)) =i/m, so T BH = max { t: Ĝ(t) = t } α = max { t: t Ĝ(t) α Note that FDR(t) (1 a)t G(t), so BH bounds FDR taking a =0. BH performs best in very sparse cases (T 0 m); power can be improved in non-sparse cases by more complicated procedures. }.

Simulated Example: Bonferroni

Simulated Example: BH

Astronomical Examples (PiCA Group) Baryon wiggles (Miller, Nichol, Batuski 2001) Radio Source Detection (Hopkins et al. 2002) Dark Energy (Scranton et al. 2003) 2 1.5 z ~ 0.57 z ~ 0.49 1 w gt (θ) [µk] 0.5 0-0.5 1.5 z ~ 0.43 z ~ 0.35 1 0.5 0-0.5-1 1 10 θ (degrees) 1 10

Mixture Model for Multiple Testing Let P m =(P 1,...,P m ) be the p-values for the m tests, drawn independently from G =(1 a)u + af, where: 1. 0 a 1 is the frequency of alternatives, 2. U is the Uniform 0, 1 cdf, and 3. F = ξdl F (ξ) is a distribution on [0,1]. Let H m =(H 1,...,H m ) where H i = 0 (or 1) if the i th null hypothesis is true (or false). Assume the H i s are independent Bernoulli a, but everything works with the H i s fixed as well.

Mixture Model for Multiple Testing (cont d) We assume the following model (Efron et al., 2001; Efron, 2003): H 1,...,H m iid Bernoulli a Ξ 1,...,Ξ m iid L F P i H i =0, Ξ i = ξ i Uniform 0, 1 P i H i =1, Ξ i = ξ i ξ i. where L F denotes a probability distribution on a class F of distributions on [0, 1]. Typical examples: Parametric family: F Θ = {F θ : θ Θ} Concave, continuous distributions F C = {F : F concave, continuous cdf with F U}.

Multiple Testing Procedures A multiple testing procedure T is a map [0, 1] m [0, 1], where the null hypotheses are rejected in all those tests for which P i T (P m ). We call T a threshold. Examples: Uncorrected testing T U (P m )=α Bonferroni T B (P m )=α/m Fixed threshold at t T t (P m )=t First r T (r) (P m )=P (r) Benjamini-Hochberg T BH (P m ) = sup{t: Ĝ(t) =t/α} Oracle T O (P m ) = sup{t: G(t) =(1 a)t/α} Plug-In T PI (P m ) = sup{t: Ĝ(t) =(1 â)t/α} Regression Classifier T Reg (P m ) = sup{t: P{H 1 =1 P 1 =t}>1/2}

The False Discovery Process Define two stochastic processes as a function of threshold t: the False Discovery Proportion FDP(t) and False Nondiscovery Proportion FNP(t). FDP(t; P m,h m )= FNP(t; P m,h m )= i i i 1 { P i t } (1 H i ) 1 { P i t } +1 { all P i >t } = #False Discoveries #Discoveries i 1 { P i >t } H i 1 { P i >t } +1 { all P i t } = #False Nondiscoveries #Nondiscoveries

The False Discovery Rate For a given procedure T, let FDP and FNP denote the value of these processes at T (P m ). Then, the False Discovery Rate (FDR) and the False Nondiscovery Rate (FNR) are given by FDR = E(FDP) FNR = E(FNP). The BH guarantee becomes FDR (1 a)α α. This bound holds at least under positive dependence. Replacing α by α/ m i=1 1/i extends FDR bound to any distribution, but this is typically very conservative.

Selected Recent Work on FDR Abromovich, Benjamini, Donoho, & Johnstone (2000) Benjamini & Hochberg (1995, 2000) Benjamini & Yekutieli (2001) Efron, Tibshirani, & Storey (2001) Efron, Tibshirani, Storey, & Tusher (2002) Finner & Roters (2001, 2002) Hochberg & Benjamini (1999) Genovese & Wasserman (2001,2002,2003) Pacifico, Genovese, Verdinelli, & Wasserman (2003) Sarkar (2002) Seigmund, Taylor, & Storey (2003) Storey (2002,2003) Storey & Tibshirani (2001) Tusher, Tibshirani, Chu (2001) Yekutieli & Benjamini (2001)

Road Map: Envelopes 1. The Multiple Testing Problem Idea and Examples Error Criteria 2. Controlling FDR The Benjamini-Hochberg Procedure Increasing Power 3. Confidence Envelopes and Thresholds Exact Confidence Envelopes for the False Discovery Proportion Choice of Tests 4. False Discovery Control for Random Fields Confidence Supersets and Thresholds Controlling the Proportion of False Clusters

Confidence Envelopes and Thresholds D α need not bound the # of false discoveries. In practice, it would be useful to control quantiles of FDP. We want a procedure T that for specified A and γ guarantees P { FDP(T ) >A } γ We call this an (A, 1 γ) confidence-threshold procedure. Three methods: (i) asymptotic closed-form threshold, (ii) asymptotic confidence envelope, and (iii) exact small-sample confidence envelope. (See Genovese & Wasserman 2003, to appear Annals of Statistics.) I ll focus here on (iii).

t Confidence Envelopes and Thresholds (cont d) A1 γ confidence envelope for FDP is a random function FDP(t) on [0, 1] such that P { FDP(t) FDP(t) for all t } 1 γ. Given such an envelope, we can construct confidence thresholds. Two special cases have proven useful. Fixed-ceiling: T = sup{t: FDP(t) α}. Minimum-envelope: T = sup{t: FDP(t) = min t FDP(t)}. FDP

Exact Confidence Envelopes Short version: take max FDP over all subsets that look Uniform. Given V 1,...,V j, let ϕ j (v 1,...,v j ) be a level γ test of the null hypothesis that V 1,...,V j are iid Uniform(0, 1). Define p m 0 (hm )=(p i : h i =0, 1 i m) m 0 (h m )= m i=1 (1 h i ) and U γ (p m )= { h m {0, 1} m : ϕ m0 (h m ) (p m 0 (hm ))=0 }. Note that as defined, U γ always contains the vector (1, 1,...,1). Let G γ (p m )= { FDP( ; h m,p m ): h m U γ (p m ) } M γ (p m )= { m 0 (h m ): h m U γ (p m ) }.

Exact Confidence Envelopes (cont d) Short version: it works. Theorem. For all 0 <a<1, F, and positive integers m, P { H m U γ (P m ) } 1 γ P { M 0 M γ (P m ) } 1 γ P { FDP( ; H m,p m ) G γ } 1 γ. Define FDP to be the pointwise sup over G γ. This is a 1 γ confidence envelope for FDP. Confidence thresholds follow directly. For example, T α = sup { t : FDP(t) α } is an (α, 1 γ) confidence threshold.

Choice of Tests The confidence envelopes depend strongly on choice of tests. Want an automatic way to choose a good test Two desiderata for selecting uniformity tests: Power, such that FDP is close to FDP, and Computability, given that there are 2 m subsets to test. Traditional uniformity tests, such as the (one-sided) Kolmogorov- Smirnov (KS) test, do not meet both conditions. For example, the KS test is sensitive to deviations from uniformity equally though all the p-values.

The P (k) Tests In contrast, using the kth order statistic as a one-sided test statistic meets both desiderata. For small k, these are sensitive to departures that have a large impact on FDP. Good power. Computing the confidence envelopes is linear in m. We call these the P (k) tests. They form a sub-family of weighted, one-sided KS tests.

Results: P (k) 90% Confidence Envelopes For k =1, 10, 25, 50, 100, with 0.05 FDP level marked. FDP 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.0000 0.0005 0.0010 0.0015 0.0020 0.0025 p value threshold

Results: (0.05,0.9) Confidence Threshold Frontal Eye Field Supplementary Eye Field Inferior Prefrontal Inferior Prefrontal Superior Parietal 14.28 Temporal-parietal junction Extrastriate Visual Cortex Temporal-parietal junction t 0

Results: (0.05,0.9) Threshold versus BH Frontal Eye Field Supplementary Eye Field Inferior Prefrontal Inferior Prefrontal Superior Parietal Temporal-parietal junction Extrastriate Visual Cortex Temporal-parietal junction

Results: (0.05,0.9) Threshold versus Bonferroni Frontal Eye Field Supplementary Eye Field Inferior Prefrontal Inferior Prefrontal Superior Parietal Temporal-parietal junction Extrastriate Visual Cortex Temporal-parietal junction

Choosing k Direct (Simulation) Approach Simulate from pre-specified parametric family or mixtures of these. Compute the optimal k, k (θ, m). Data-dependent approaches Estimate a and F, and simulate from corresponding mixture. Parametric estimate k ( θ, m). Solve for optimal k distribution using smoothed estimate of G. The data-dependence only has a small effect on coverage.

Results: Direct versus Fitting Approach FDP 0.00 0.05 0.10 0.15 0.20 0.25 0.30 0.0000 0.0005 0.0010 0.0015 0.0020 0.0025 p value threshold

Road Map: Envelopes 1. The Multiple Testing Problem Idea and Examples Error Criteria 2. Controlling FDR The Benjamini-Hochberg Procedure Increasing Power 3. Confidence Envelopes and Thresholds Exact Confidence Envelopes for the False Discovery Proportion Choice of Tests 4. False Discovery Control for Random Fields Confidence Supersets and Thresholds Controlling the Proportion of False Clusters

Results: False Region Control Threshold P { prop n false regions 0.1 } 0.95 where false means null overlap 10% Frontal Eye Field Supplementary Eye Field Inferior Prefrontal Inferior Prefrontal Superior Parietal Temporal-parietal junction Extrastriate Visual Cortex Temporal-parietal junction

Take-Home Points: Envelopes Confidence thresholds have advantages for False Discovery Control. In particular, we gain a stronger inferential guarantee with little effective loss of power. Dependence complicates the analysis greatly, but confidence envelopes appear to be valid under positive dependence. For spatial applications, we care about clusters/regions/sources not pixels. Current methods ignore spatial information. Controlling proportion of false regions is a first step. Region-based false discovery control is the next step. (work in progress)

Road Map: Balls 1. Inferences about Functions CMB Spectrum Example The Statistical Problem in General Form Criteria for Effective Inferences 2. Nonparametric Confidence Balls Features and Extensions 3. Keeping Your Eyes on the Ball Parametric Probes Model Checking Confidence Catalogs

Road Map: Balls 1. Inferences about Functions CMB Spectrum Example The Statistical Problem in General Form Criteria for Effective Inferences 2. Nonparametric Confidence Balls Features and Extensions 3. Keeping Your Eyes on the Ball Parametric Probes Model Checking Confidence Catalogs

C l l(l + 1) 2π (µ K 2 ) CMB Power Spectrum: WMAP Data 0 200 400 600 800 0 200 400 600 800 Multipole Moment l Multipole Moment l 10000 5000 0 5000 10000 15000 20000 C l l(l + 1) 2π (µ K 2 ) 0 2000 4000 6000

CMB Power Spectrum: WMAP Variances log 10 (σ l ) (log 10 µ K 2 ) 5.0 5.5 6.0 6.5 7.0 7.5 8.0 0 200 400 600 800 Multipole Moment l Noise is correlated and heteroskedastic

CMB Power Spectrum: Models C l l(l + 1) 2π 0 1000 2000 3000 4000 5000 6000 0 200 400 600 800 Multipole Moment l 11(7)-dimensional model maps cosmological parameters to spectra. Ultimate goal: inferences about these cosmological parameters. Subsidiary goal: identify location, height, widths of peaks

The Statistical Problem Observe noisy samples of an unknown function. Data of the form Y i = f(x i )+ɛ i, i =1,...,n, where f is a function on [0, 1] and ɛ is a possibly correlated vector of (Gaussian) noise. We assume f lies in some pre-specified space of functions F, such as a Besov ball. Assume for the moment that the noise covariance is known. Goal: Make inferences about (often complicated) functionals of f.

Approaches to Function Inference Common Estimate plus Goodness of Fit Pointwise confidence bands Confidence intervals on pre-specified features Another Idea A. Generate a confidence set for the whole object. B. Restrict by imposing constraints, if desired. C. Probe confidence set to address specific questions of interest.

What Do We Want from an Inference? Frequentist Confidence Set C Bayesian Posterior Region B Can have (2) hold and yet have in nonparametric problems. min f P { C f } 1 α. (1) P { f B Data } 1 α. (2) min f P { B f } 0 (3)

Road Map: Balls 1. Inferences about Functions CMB Spectrum Example The Statistical Problem in General Form Criteria for Effective Inferences 2. Nonparametric Confidence Balls Features and Extensions 3. Keeping Your Eyes on the Ball Parametric Probes Model Checking Confidence Catalogs

Our Approach Construct (asymptotic) confidence set for f that is uniform sup P { C n f } (1 α) 0, f F that provides post-hoc protection: we can constrain or probe the ball to address any set of questions. Construction based on Stein-Beran-Dümbgen pivot method. Extended to wavelet bases (GW, 2003b), weighted loss functions (GW, 2003c), and density estimation (GJW, 2003). Confidence set takes form of ball (or ellipsoid)

CMB: Center of Ball vs Concordance Model 0 200 400 600 800 Multipole Moment l C l l(l + 1) 2π 0 1000 2000 3000 4000 5000 6000

Road Map: Balls 1. Inferences about Functions CMB Spectrum Example The Statistical Problem in General Form Criteria for Effective Inferences 2. Nonparametric Confidence Balls Features and Extensions 3. Keeping Your Eyes on the Ball Parametric Probes Model Checking Confidence Catalogs

Eyes on the Ball I: Parametric Probes Peak Heights Peak Locations Ratios of Peak Heights

Eyes on the Ball I: Parametric Probes (cont d) Varied baryon fraction in CMBfast keeping Ω total 1 C l l(l + 1) 2π 1000 2000 3000 4000 5000 0 200 400 600 800 Multipole Moment l Range [0.034,0.0586] in ball

Eyes on the Ball I: Parametric Probes (cont d) Probe from center with boxcars of given width centered at each l. Maximum boxcar height in 95% ball, relative to Concordance Model Boxcar Height/Concordance Model 0 2 4 6 8 Width 50 Width 100 0 200 400 600 800 Multipole Moment l

Eyes on the Ball II: Model Checking Inclusion in the confidence ball provides simultaneous goodness-of-fit tests for parametric (or other) models. Concordance α=0.16 Just WMAP α=0.73 C l l(l + 1) 2π 0 1000 2000 3000 4000 5000 6000 C l l(l + 1) 2π 0 1000 2000 3000 4000 5000 6000 0 200 400 600 800 Multipole Moment l 0 200 400 600 800 Multipole Moment l Extremal 1 α=0.95 Extremal 2 α=0.95 C l l(l + 1) 2π 0 1000 2000 3000 4000 5000 6000 C l l(l + 1) 2π 0 1000 2000 3000 4000 5000 6000 0 200 400 600 800 Multipole Moment l 0 200 400 600 800 Multipole Moment l

Eyes on the Ball III: Confidence Catalogs Our confidence set construction does not impose constraints based on prior knowledge. Instead: form ball first and impose constraints at will. Raises the possibility of viewing inferences as a function of prior assumptions. The confidence ball creates a mapping from prior assumptions to inferences; we call this a confidence catalog. Ex: Constraints on peak curvature over range defined by reasonable parametric models.

Take-Home Points: Balls Uniformity makes the asymptotic approximations more useful. Post-hoc protection allows snooping. Can make inferences about any set of functionals with simultaneous validity. Nonparametric approach provides check on physical models. Embedding parametric model in constrained nonparametric model gives flexibility when model is uncertain. Beginning with a confidence set on the whole object makes it easy to compare different sets of assumptions.