Controlling the False Discovery Rate: Understanding and Extending the Benjamini-Hochberg Method

Similar documents
A Large-Sample Approach to Controlling the False Discovery Rate

Doing Cosmology with Balls and Envelopes

New Approaches to False Discovery Control

Confidence Thresholds and False Discovery Control

New Procedures for False Discovery Control

Exceedance Control of the False Discovery Proportion Christopher Genovese 1 and Larry Wasserman 2 Carnegie Mellon University July 10, 2004

Estimation of a Two-component Mixture Model

A Stochastic Process Approach to False Discovery Rates Christopher Genovese 1 and Larry Wasserman 2 Carnegie Mellon University January 7, 2003

False discovery rate and related concepts in multiple comparisons problems, with applications to microarray data

High-throughput Testing

The miss rate for the analysis of gene expression data

FDR-CONTROLLING STEPWISE PROCEDURES AND THEIR FALSE NEGATIVES RATES

FALSE DISCOVERY AND FALSE NONDISCOVERY RATES IN SINGLE-STEP MULTIPLE TESTING PROCEDURES 1. BY SANAT K. SARKAR Temple University

More powerful control of the false discovery rate under dependence

A GENERAL DECISION THEORETIC FORMULATION OF PROCEDURES CONTROLLING FDR AND FNR FROM A BAYESIAN PERSPECTIVE

A Sequential Bayesian Approach with Applications to Circadian Rhythm Microarray Gene Expression Data

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

High-Throughput Sequencing Course. Introduction. Introduction. Multiple Testing. Biostatistics and Bioinformatics. Summer 2018

Table of Outcomes. Table of Outcomes. Table of Outcomes. Table of Outcomes. Table of Outcomes. Table of Outcomes. T=number of type 2 errors

Post-Selection Inference

Weighted Adaptive Multiple Decision Functions for False Discovery Rate Control

Simultaneous Testing of Grouped Hypotheses: Finding Needles in Multiple Haystacks

False Discovery Control in Spatial Multiple Testing

Statistical testing. Samantha Kleinberg. October 20, 2009

Controlling Bayes Directional False Discovery Rate in Random Effects Model 1

Summary and discussion of: Controlling the False Discovery Rate: A Practical and Powerful Approach to Multiple Testing

Statistical Applications in Genetics and Molecular Biology

Non-specific filtering and control of false positives

A Bayesian Determination of Threshold for Identifying Differentially Expressed Genes in Microarray Experiments

Modified Simes Critical Values Under Positive Dependence

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided

Performance Evaluation and Comparison

Alpha-Investing. Sequential Control of Expected False Discoveries

Some General Types of Tests

Multiple testing: Intro & FWER 1

Multiple Testing Procedures under Dependence, with Applications

Bayesian Determination of Threshold for Identifying Differentially Expressed Genes in Microarray Experiments

Rank conditional coverage and confidence intervals in high dimensional problems

The optimal discovery procedure: a new approach to simultaneous significance testing

EMPIRICAL BAYES METHODS FOR ESTIMATION AND CONFIDENCE INTERVALS IN HIGH-DIMENSIONAL PROBLEMS

Stat 206: Estimation and testing for a mean vector,

Multiple Testing. Hoang Tran. Department of Statistics, Florida State University

Probabilistic Inference for Multiple Testing

PROCEDURES CONTROLLING THE k-fdr USING. BIVARIATE DISTRIBUTIONS OF THE NULL p-values. Sanat K. Sarkar and Wenge Guo

Looking at the Other Side of Bonferroni

Large-Scale Hypothesis Testing

STAT 461/561- Assignments, Year 2015

arxiv: v2 [stat.me] 14 Mar 2011

Journal of Statistical Software

Advanced Statistical Methods: Beyond Linear Regression

Two-stage stepup procedures controlling FDR

Association studies and regression

arxiv: v1 [math.st] 31 Mar 2009

Math 494: Mathematical Statistics

Applying the Benjamini Hochberg procedure to a set of generalized p-values

Aliaksandr Hubin University of Oslo Aliaksandr Hubin (UIO) Bayesian FDR / 25

On Procedures Controlling the FDR for Testing Hierarchically Ordered Hypotheses

Announcements. Proposals graded

Biostatistics Advanced Methods in Biostatistics IV

The Pennsylvania State University The Graduate School Eberly College of Science GENERALIZED STEPWISE PROCEDURES FOR

The Pennsylvania State University The Graduate School A BAYESIAN APPROACH TO FALSE DISCOVERY RATE FOR LARGE SCALE SIMULTANEOUS INFERENCE

Marginal Screening and Post-Selection Inference

FDR and ROC: Similarities, Assumptions, and Decisions

Lecture 2: Statistical Decision Theory (Part I)

Statistical Applications in Genetics and Molecular Biology

Improving the Performance of the FDR Procedure Using an Estimator for the Number of True Null Hypotheses

Stat 5101 Lecture Notes

Introduction to Machine Learning. Lecture 2

Uniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence

Lecture 8 Inequality Testing and Moment Inequality Models

Introduction to Bayesian Statistics

Notes on Decision Theory and Prediction

Journal de la Société Française de Statistique Vol. 152 No. 2 (2011) Type I error rate control for testing many hypotheses: a survey with proofs

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Sanat Sarkar Department of Statistics, Temple University Philadelphia, PA 19122, U.S.A. September 11, Abstract

Estimation and Confidence Sets For Sparse Normal Mixtures

Model Identification for Wireless Propagation with Control of the False Discovery Rate

A union of Bayesian, frequentist and fiducial inferences by confidence distribution and artificial data sampling

Nonparametric Inference and the Dark Energy Equation of State

MINIMUM EXPECTED RISK PROBABILITY ESTIMATES FOR NONPARAMETRIC NEIGHBORHOOD CLASSIFIERS. Maya Gupta, Luca Cazzanti, and Santosh Srivastava

Bayesian Econometrics

Optional Stopping Theorem Let X be a martingale and T be a stopping time such

Asymptotic Results on Adaptive False Discovery Rate Controlling Procedures Based on Kernel Estimators

Sequential Analysis & Testing Multiple Hypotheses,

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

A multiple testing procedure for input variable selection in neural networks

Uniformly and Restricted Most Powerful Bayesian Tests

On Methods Controlling the False Discovery Rate 1

Effects of dependence in high-dimensional multiple testing problems. Kyung In Kim and Mark van de Wiel

Can we do statistical inference in a non-asymptotic way? 1

Data Mining. CS57300 Purdue University. March 22, 2018

Hunting for significance with multiple testing

Bayesian estimation of the discrepancy with misspecified parametric models

Rejoinder on: Control of the false discovery rate under dependence using the bootstrap and subsampling

arxiv: v1 [stat.me] 25 Aug 2016

Political Science 236 Hypothesis Testing: Review and Bootstrapping

A NEW APPROACH FOR LARGE SCALE MULTIPLE TESTING WITH APPLICATION TO FDR CONTROL FOR GRAPHICALLY STRUCTURED HYPOTHESES

Estimation of the False Discovery Rate

Spring 2012 Math 541B Exam 1

P Values and Nuisance Parameters

Transcription:

Controlling the False Discovery Rate: Understanding and Extending the Benjamini-Hochberg Method Christopher R. Genovese Department of Statistics Carnegie Mellon University joint work with Larry Wasserman This work partially supported by NSF Grant SES 9866147.

Motivating Example #1: fmri fmri Data: Time series of 3-d images acquired while subject performs specified tasks. 1850 1900 1950 2000 Task Control Task Control Task Control Task Control 0 24 48 72 96 120 144 168 Goal: Characterize task-related signal changes caused (indirectly) by neural activity. [See, for example, Genovese (2000), JASA 95, 691.]

fmri (cont d) Perform hypothesis tests at many thousands of volume elements to identify loci of activation.

Motivating Example #2: Source Detection Interferometric radio telescope observations processed into digital image of the sky in radio frequencies. Signal at each pixel is a mixture of source and background signals.

Motivating Example #3: DNA Microarrays New technologies allow measurement of gene expression for thousands of genes simultaneously. Subject Subject 1 2 3... 1 2 3... 1 X 111 X 121 X 131... X 112 X 122 X 132... 2 X 211 X 221 X 231... X 212 X 222 X 232... 3............ Gene 4 5 6. Condition 1 Condition 2 Goal: Identify genes associated with differences among conditions. Typical analysis: hypothesis test at each gene.

The Multiple Testing Problem Perform m simultaneous hypothesis tests. Classify results as follows: H 0 Retained H 0 Rejected Total H 0 True N 0 0 N 1 0 M 0 H 0 False N 0 1 N 1 1 M 1 Total m R R m Only R is observed here. Assess outcome through combined error measure. This binds the separate decision rules together.

Multiple Testing (cont d) Traditional methods seek strong control of familywise Type I error (FWER). Weak Control: If all nulls true, P { N 1 0 > 0 } α. Strong Control: Corresponding statement holds for any subset of tests for which all nulls are true. For example, Bonferroni correction provides strong control but is quite conservative. Can power be improved while maintaining control over a meaningful measure of error? Enter Benjamini & Hochberg...

FDR and the BH Procedure Define the realized False Discovery Rate (FDR) by FDR = N 1 0 R if R>0, 0, if R =0. Benjamini & Hochberg (1995) define a sequential p-value procedure that controls expected FDR. Specifically, the BH procedure guarantees E(FDR) M 0 m α α for a pre-specified 0 <α<1. (The first inequality is an equality in the continuous case.)

p value 0.0 0.2 0.4 0.6 0.8 m = 50, α =0.1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 1 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0.0 0.2 0.4 0.6 0.8 1.0 Index/m

The BH procedure for p-values P 1,...,P m : 0. Select 0 <α<1. 1. Define P (0) 0 and { R BH = max 0 i m: P (i) α i m }. 2. Reject H 0 for every test where P j P (RBH ). Several variant procedures also control E(FDR). Bound on E(FDR) holds if p-values are independent or positively dependent (Benjamini & Yekutieli, 2001). Storey (2001) shows it holds under a possibly weaker condition. By replacing α with α/ m i=1 1/i, control E(FDR) at level α for any joint distribution on the p-values. (Very conservative!)

Road Map 1. Preliminaries Considering both types of errors: The False Nondiscovery Rate (FNR) Models for realized FDR and FNR FDR and FNR as stochastic processes 2. Understanding BH Re-express BH procedure as plug-in estimator Asymptotic behavior of BH Improving the power more general plug-ins Asymptotic risk comparisons 3. Extensions to BH Conditional risk FDR control as an estimation problem Confidence intervals for realized FDR Confidence thresholds

Recent Work on FDR Benjamini & Hochberg (1995) Benjamini & Liu (1999) Benjamini & Hochberg (2000) Benjamini & Yekutieli (2001) Storey (2001a,b) Efron, et al. (2001) Storey & Tibshirani (2001) Tusher, Tibshirani, Chu (2001) Abromovich, et al. (2000) Genovese & Wasserman (2001a,b)

The False Nondiscovery Rate Controlling FDR alone only deals with Type I errors. Define the realized False Nondiscovery Rate as follows: FNR = N 0 1 m R if R<m, 0 if R = m. This is the proportion of false non-rejections among those tests whose null hypothesis is not rejected. Idea: Combine FDR and FNR in assessment of procedures.

Basic Models Let H i = 0 (or 1) if the i th null hypothesis is true (or false). These are unobserved. Let P i be the i th p-value. We assume that (P 1,H 1 ),...,(P m,h m ) are independent. Under the conditional model, H 1,...,H m are fixed, unknown. Under the mixture model, we assume each H i has Bernoulli a, P i { H i =0 } Uniform 0, 1, and P i { H i =1 } F F Here, F is a class of alternative p-value distributions. Define M 0 = i(1 H i ) and M 1 = i H i = m M 0. Under the conditional model, these are fixed. Under the mixture model, M 1 Binomial m, a.

Recurring Notation m, M 0,N 1 0 # of tests, true nulls, false discoveries a Mixture weight on alternative H m =(H 1,...,H m ) Unobserved true classifications P m =(P 1,...,P m ) Observed p-values P m () =(P (1),...,P (m) ) Sorted p-values (define P (0) 0) U CDF of Uniform 0, 1 F, f Alternative CDF and density G =(1 a)u + af Marginal CDF of P i (mixture model) Ĝ Empirical CDF of P m ɛ m = 1 ( ) 2 2m log DKW bound 1 β quantile of Ĝ β

Multiple Testing Procedures A multiple testing procedure T is a map [0, 1] m [0, 1], where the null hypotheses are rejected in all those tests for which P i T (P m ). Examples: Uncorrected testing T U (P m ) = α Bonferroni T B (P m ) = α/m Benjamini-Hochberg T BH (P m )=P (RBH ) Fixed Threshold T t (P m ) = t First-r T (r) (P m )=P (r)

FDR and FNR as Stochastic Processes Define the realized FDR and FNR processes, respectively, by FDR(t) FDR(t; P m,h m )= FNR(t) FNR(t; P m,h m )= i i i 1 { P i t } (1 H i ) 1 { P i t } + i i 1 { P i >t } H i 1 { P i >t } + i 1 { P i >t } 1 { P i t }. For procedure T, the realized FDR and FNR are obtained by evaluating these processes at T (P m ). Both these processes converge to Gaussian processes outside a neighborhood of 0 and 1 respectively.

BH as a Plug-in Procedure Let Ĝ be the empirical cdf of P m under the mixture model. Ignoring ties, Ĝ(P (i)) =i/m, so BH equivalent to T BH (P m { ) = arg max t: Ĝ(t) = t α We can think of this as a plug-in procedure for estimating u { (a, F ) = arg max t: G(t) = t } α = arg max {t: F (t) =βt}, where β =(1 α + αa)/αa. }.

Asymptotic Behavior of BH Procedure This yields the following picture: α m, α, u F (u) u 0 α u m Bonferroni FDR Uncorrected α

Optimal Thresholds Under the mixture model and in the continuous case, E(FDR(T BH (P m )))=(1 a)α. The BH procedure overcontrols E(FDR) and thus will not in general minimize E(FNR). This suggests finding a plug-in estimator for { t (1 a)t (a, F ) = arg max t: G(t) = α = arg max {t: F (t) =(β 1/α)t}, where β 1/α =(1 a)(1 α)/aα. Note that t u. }

Optimal Thresholds (cont d) For each 0 t 1, E(FDR(t)) = (1 a) t G(t) + O E(FNR(t)) = a 1 F (t) 1 G(t) + O ( ) 1 m ( ) 1. m Ignoring O(m 1/2 ) terms and choosing t to minimize E(FNR(t)) subject to E(FDR(t)) α, yields t (a, F ) as the optimal threshold. Can the potential improvement in power be achieved when estimating t? Yes, if F sufficiently far from U.

Operating Characteristics of the BH Method Define the misclassification risk of a procedure T by R M (T )= 1 m m i=1 E 1 { Pi T (P m ) } H i. This is the average fraction of errors of both types. Then R M (T BH ) R(a, F )asm, where R(a, F )=(1 a)u + a(1 F (u ))=(1 a)u + a(1 βu ). Compare this to Uncorrected and Bonferroni and the oracle rule T O (P m )=b where b solves f(b) =(1 a)/a. R M (T U )=(1 a) α + a (1 F (α)) R M (T B )=(1 a) α m + a (1 F R M (T O )=(1 a) b + a (1 F (b)). ( α m))

theta = 2 0.0 0.2 0.4 0.6 0.8 1.0 A0 theta = 3.125 0.0 0.2 0.4 0.6 0.8 1.0 A0 theta = 4.25 0.0 0.2 0.4 0.6 0.8 1.0 A0 Risk 0.01 0.03 0.05 Risk 0.02 0.04 0.06 0.08 Risk 0.0 0.1 0.2 0.3 0.4 Risk Risk Risk 0.0 0.01 0.03 0.05 0.01 0.02 0.03 0.04 0.05 0.0 0.10 0.20 theta = 2.375 0.0 0.2 0.4 0.6 0.8 1.0 A0 theta = 3.5 0.0 0.2 0.4 0.6 0.8 1.0 A0 theta = 4.625 0.0 0.2 0.4 0.6 0.8 1.0 A0 Risk Risk Risk 0.0 0.01 0.03 0.05 0.01 0.03 0.05 0.02 0.06 0.10 0.14 theta = 2.75 0.0 0.2 0.4 0.6 0.8 1.0 A0 theta = 3.875 0.0 0.2 0.4 0.6 0.8 1.0 A0 theta = 5 0.0 0.2 0.4 0.6 0.8 1.0 A0

theta = 2 0.0 0.2 0.4 0.6 0.8 1.0 A0 theta = 3.125 0.0 0.2 0.4 0.6 0.8 1.0 A0 theta = 4.25 0.0 0.2 0.4 0.6 0.8 1.0 A0 Risk 0.005 0.010 0.015 Risk 0.02 0.04 0.06 0.08 Risk 0.0 0.1 0.2 0.3 0.4 Risk Risk Risk 0.002 0.006 0.010 0.014 0.01 0.02 0.03 0.04 0.0 0.10 0.20 theta = 2.375 0.0 0.2 0.4 0.6 0.8 1.0 A0 theta = 3.5 0.0 0.2 0.4 0.6 0.8 1.0 A0 theta = 4.625 0.0 0.2 0.4 0.6 0.8 1.0 A0 Risk Risk Risk 0.002 0.006 0.010 0.014 0.005 0.015 0.025 0.02 0.06 0.10 0.14 theta = 2.75 0.0 0.2 0.4 0.6 0.8 1.0 A0 theta = 3.875 0.0 0.2 0.4 0.6 0.8 1.0 A0 theta = 5 0.0 0.2 0.4 0.6 0.8 1.0 A0

Extension 1: Conditional Risk It is intuitively appealing (cf. Kiefer, 1977) to assess the performance of a procedure conditionally given the ordered p-values. When conditioning, we need only consider the m + 1 procedures T (r) (P m )=P (r) for r =0,...,m. Under the conditional model, once P m () is observed, only the randomness in the labelling of the true classifications remains. Consider a parametric family F = {F θ : θ Θ} of alternative p-value distributions. Then, (M 0,θ) becomes the unknown parameter. Begin by treating this as known.

Conditional Risk (cont d) Define a conditional risk for λ 0by R λ (r; M 0,θ P m () )=E M0,θ [ FNR(P (r) )+λ FDR(P (r) ) P m ] (), where M 0 and r are in {0,...,m} and θ Θ. Here λ determines the balance between the two error types. It also serves as a Lagrange multiplier for the optimization problem: r = arg min E M 0 r m 0,θ(FNR(P (r) ) P m () ) subject to E M0,θ(FDR(P (r) ) P m () ) α.

Conditional Risk (cont d) This problem can be solved exactly: Closed form for conditional distribution of FDR and FNR based on expressions for { P M0,θ N1 0 = k P m } () and EM0,θ(N 1 0 P m () ) derived via generating function methods. Find R λ -minimizer explicitly. Select λ to satisfy the constraint. Remark: The R λ minimizing conditional procedure also minimizes the unconditional R λ risk, but the constrained optimization problem is harder to solve unconditionally.

Conditional Risk (cont d) For M 0 unknown case, R λ dominated by extremes, M 0 =0orM 0 = m. One approach: minimize conditional Bayes risk based on R λ R λ (r; θ P m () )= m m 0 =0 R λ (r; m 0,θ P m () ) p θ (m 0 P m () ), where p θ (m 0 P m () ) is derived from a specified (e.g., Uniform) prior on {0,...,m}. This minimizes the unconditional Bayes risk.

Compare R λ for BH, optimal, and naive: Theta = 2 0.0 0.4 0.8 M0/m Theta = 4 0.0 0.4 0.8 M0/m R_lambda 0.0 0.4 0.8 R_lambda 0.0 0.4 0.8 R_lambda 0.0 0.4 0.8 R_lambda 0.0 0.4 0.8 Theta = 3 0.0 0.4 0.8 M0/m Theta = 5 0.0 0.4 0.8 M0/m

Bayesian FDR These conditional results yield the posterior distribution of FDR and FNR (and related quantities). No simulation necessary: can compute full posterior directly. Suggests the procedure T Bayes (P m )=P (r ), where r = arg min E(FNR(P (r)) P m () ) 0 r m subject to E(FDR(P (r) ) P m () ) α. This procedure has good asympotic frequentist performance.

Extension 2: Estimating a and F To compute plug-in estimates that approximate the optimal threshold, we need a good estimate of a. For instance, t = arg max { t: } â)t Ĝ(t) =(1. α For confidence thresholds, need estimate of a and F. Identifiability f If min f = b>0, can write F =(1 b)u+bf 0, so many (a, F ) pairs yield the same G. If f = F is decreasing with f(1) = 0, then (a, F ) is identifiable. b

Estimating a and F (cont d) Even when non-identifiable, a can be bounded from below by a. a a is typically small. For example, a a = ae nθ2 /2 in the two-sided test of θ = 0 versus θ 0 in the Normal θ, 1 model. Parametric Case: (a, θ) typically identifiable; use MLE. Non-parametric case: Derived a 1 β confidence interval for a and thus a. When F concave, get â LCM = a + O P (m 1/3 ). Can do better with further smoothness assumptions. In general, requires density estimate of g. Can estimate F by: F m = argmin H Ĝ (1 â)u âh. Consistent for reduced F if â consistent for a. Note: Assumption of concavity has a big effect.

Extension 3: Confidence Intervals Beyond controlling FDR and FNR on average, we would like to be able to make inferences about the realized quantities. Want to find c(p m,t), for any procedure T, such that at least asymptotically. P a,f { FDR(T (P m )) c(p m,t) } 1 α, Let r(p m,t)= i 1 { P i T (P m ) } be the number of rejections. Template: c(p m,t)isa1 β quantile of the sum of r(p m,t) independent Bernoulli q i variables. Here, the q i bound q(p (i) ) with high probability, where q(t) = (1 a)/g(t) gives the conditional distribution of H 1 given P 1. The q i depend on the assumed class F of alternative p-value distributions.

Confidence Intervals (cont d) Case 1: F = {F θ : θ Θ} Asymptotic: β = α and q i = 1 â 1 â + âf θ (P (i)). Exact: Let β =1 1 α and let Ψ m bea1 β confidence set for (a, θ). 1 a q i = sup Ψ m 1 a + af θ (P (i) ). Example: Invert DKW Envelope Ψ m = {(a, θ): G a,θ Ĝ ɛ m}.

Confidence Intervals (cont d) Case 2: F = {F : F concave, continuous cdf and F U}. Can find a minimal concave cdf G in DKW envelope. Define q i = 1 â g(p i ), and use β =1 (1 α) 1/3. May be possible to obtain nonparametric results in non-concave case, but the intervals appear to be hopelessly wide in practice. Bayesian posterior intervals also have asymptotically valid frequentist coverage. All these results extend to give joint confidence intervals for FDR and FNR.

Extension 4: Confidence Thresholds In practice, it would be useful to have a procedure T C that guarantees P G { FDR(TC ) >c } α for some specified c and α. We call this a (1 α, c) confidence threshold procedure. Two approaches: an asymptotic threshold using the Bootstrap, and an exact (small-sample) threshold requiring numerical search. Here, I ll discuss the case where a is known. In general, can use an estimate of a, but this introduces additional complexity.

Bootstrap Confidence Thresholds First guess: Choose T such that Unfortunately, this fails. PĜ{ FDR (T ) c } 1 α. The problem is an additional bias term: { 1 α =PĜ FDR (T ) c } { P G FDR(T ) c +(Q(T ) Q(T )) } P G { FDR(T ) c }, where Q =(1 a)u/g and Q =(1 a)u/ĝ.

Bootstrap Confidence Thresholds (cont d) Let β = α/2 and ɛ m = Procedure 1 2m log ( ) 2. β 1. Draw H1...,H m iid Bernoulli a 2. Draw Pi H i from (1 H i )U + H i F. 3. Define Ω c(t) = i I{Pi t}(1 H i c). 4. Use threshold defined by T C = max { } } t:pĝ{ Ω c (t) cɛ m 1 β. Then, P G { FDR(TC ) c } 1 α + O ( 1 m ).

Exact Confidence Thresholds Let M β be a 1 β confidence set for M 0, derived from the Binomial m, 1 a. Define S(t; h m,p m i )= 1 { p i t } (1 h i ) i (1 h, i) { U = (h m,p m ): } (1 h i ) M α and S( ; h m,p m ) U ɛ m, i where ɛ m = log(2/β)/2m as above. Then, if β =1 1 α, P G { (H m,p m ) U } 1 α and T C = sup {t : FDR(t; h m,p m ) c and h m :(h m,p m ) U} is a(1 α, c) confidence threshold procedure. That is, P G { FDR(TC ) c } 1 α.

Exact Confidence Thresholds (cont d) The confidence set U directly yields a confidence set for the FDR(t) sample paths.

Take-Home Points Realized versus Expected FDR Considering both FDR and FNR yields greater power Multiple testing problem is transformed to an estimation problem. Must control FDR and FNR as stochastic processes. In general, the threshold and the FDR are coupled, and these correlations can have a large effect.