New Procedures for False Discovery Control

Similar documents
New Approaches to False Discovery Control

Confidence Thresholds and False Discovery Control

Doing Cosmology with Balls and Envelopes

A Large-Sample Approach to Controlling the False Discovery Rate

Controlling the False Discovery Rate: Understanding and Extending the Benjamini-Hochberg Method

Exceedance Control of the False Discovery Proportion Christopher Genovese 1 and Larry Wasserman 2 Carnegie Mellon University July 10, 2004

False Discovery Control in Spatial Multiple Testing

Peak Detection for Images

False Discovery Rates for Random Fields

PROCEDURES CONTROLLING THE k-fdr USING. BIVARIATE DISTRIBUTIONS OF THE NULL p-values. Sanat K. Sarkar and Wenge Guo

More powerful control of the false discovery rate under dependence

Department of Statistics University of Central Florida. Technical Report TR APR2007 Revised 25NOV2007

The General Linear Model. Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London

Sanat Sarkar Department of Statistics, Temple University Philadelphia, PA 19122, U.S.A. September 11, Abstract

Multiple Testing Procedures under Dependence, with Applications

Table of Outcomes. Table of Outcomes. Table of Outcomes. Table of Outcomes. Table of Outcomes. Table of Outcomes. T=number of type 2 errors

STEPDOWN PROCEDURES CONTROLLING A GENERALIZED FALSE DISCOVERY RATE. National Institute of Environmental Health Sciences and Temple University

Control of Generalized Error Rates in Multiple Testing

Step-down FDR Procedures for Large Numbers of Hypotheses

ON STEPWISE CONTROL OF THE GENERALIZED FAMILYWISE ERROR RATE. By Wenge Guo and M. Bhaskara Rao

Resampling-Based Control of the FDR

STEPUP PROCEDURES FOR CONTROL OF GENERALIZATIONS OF THE FAMILYWISE ERROR RATE

Heterogeneity and False Discovery Rate Control

Rejoinder on: Control of the false discovery rate under dependence using the bootstrap and subsampling

Applying the Benjamini Hochberg procedure to a set of generalized p-values

On Methods Controlling the False Discovery Rate 1

On Procedures Controlling the FDR for Testing Hierarchically Ordered Hypotheses

High-Throughput Sequencing Course. Introduction. Introduction. Multiple Testing. Biostatistics and Bioinformatics. Summer 2018

Procedures controlling generalized false discovery rate

Lecture 7 April 16, 2018

Hunting for significance with multiple testing

Two-stage stepup procedures controlling FDR

Statistical Applications in Genetics and Molecular Biology

The Pennsylvania State University The Graduate School Eberly College of Science GENERALIZED STEPWISE PROCEDURES FOR

Chapter 1. Stepdown Procedures Controlling A Generalized False Discovery Rate

False discovery rate and related concepts in multiple comparisons problems, with applications to microarray data

Multiple Testing. Hoang Tran. Department of Statistics, Florida State University

Modelling temporal structure (in noise and signal)

arxiv: v1 [math.st] 31 Mar 2009

Supplementary Information. Brain networks involved in tactile speed classification of moving dot patterns: the. effects of speed and dot periodicity

Specific Differences. Lukas Meier, Seminar für Statistik

Research Article Sample Size Calculation for Controlling False Discovery Proportion

Neuroimaging for Machine Learners Validation and inference

Weighted Adaptive Multiple Decision Functions for False Discovery Rate Control

Controlling the proportion of falsely-rejected hypotheses. when conducting multiple tests with climatological data

Jan 16: The Visual System

False Discovery Rate

STAT 263/363: Experimental Design Winter 2016/17. Lecture 1 January 9. Why perform Design of Experiments (DOE)? There are at least two reasons:

Looking at the Other Side of Bonferroni

Lecture 28. Ingo Ruczinski. December 3, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University

Estimating False Discovery Proportion Under Arbitrary Covariance Dependence

Control of Directional Errors in Fixed Sequence Multiple Testing

Positive false discovery proportions: intrinsic bounds and adaptive control

A Stochastic Process Approach to False Discovery Rates Christopher Genovese 1 and Larry Wasserman 2 Carnegie Mellon University January 7, 2003

High-throughput Testing

Controlling Bayes Directional False Discovery Rate in Random Effects Model 1

Estimation of a Two-component Mixture Model

Optional Stopping Theorem Let X be a martingale and T be a stopping time such

Non-specific filtering and control of false positives

Post-Selection Inference

On adaptive procedures controlling the familywise error rate

Statistical inference for MEG

arxiv: v3 [stat.me] 29 Oct 2016

Controlling the Proportion of Falsely Rejected Hypotheses when Conducting Multiple Tests with Climatological Data

Alpha-Investing. Sequential Control of Expected False Discoveries

Statistical testing. Samantha Kleinberg. October 20, 2009

False discovery control for multiple tests of association under general dependence

Modified Simes Critical Values Under Positive Dependence

Simultaneous Testing of Grouped Hypotheses: Finding Needles in Multiple Haystacks

Summary and discussion of: Controlling the False Discovery Rate: A Practical and Powerful Approach to Multiple Testing

The optimal discovery procedure: a new approach to simultaneous significance testing

Multiple testing: Intro & FWER 1

Lecture 6 April

FDR-CONTROLLING STEPWISE PROCEDURES AND THEIR FALSE NEGATIVES RATES

EMPIRICAL BAYES METHODS FOR ESTIMATION AND CONFIDENCE INTERVALS IN HIGH-DIMENSIONAL PROBLEMS

Dynamic Causal Modelling for fmri

Marginal Screening and Post-Selection Inference

Advanced Statistical Methods: Beyond Linear Regression

Week 5 Video 1 Relationship Mining Correlation Mining

Statistical Inference

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided

Lecture 27. December 13, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

Comments on: Control of the false discovery rate under dependence using the bootstrap and subsampling

Improving the Performance of the FDR Procedure Using an Estimator for the Number of True Null Hypotheses

POSITIVE FALSE DISCOVERY PROPORTIONS: INTRINSIC BOUNDS AND ADAPTIVE CONTROL

arxiv:math/ v1 [math.st] 29 Dec 2006 Jianqing Fan Peter Hall Qiwei Yao

Performance Evaluation and Comparison

The University of Warwick. Department of Statistics. Dissertation. Controls for Multiplicity in Resting State Functional Connectivity Analysis

Model Identification for Wireless Propagation with Control of the False Discovery Rate

3 Comparison with Other Dummy Variable Methods

Statistical Inference

Experimental design of fmri studies

Introduction to Statistical Data Analysis III

Statistical Applications in Genetics and Molecular Biology

Estimating False Discovery Proportion Under Arbitrary Covariance Dependence

FALSE DISCOVERY AND FALSE NONDISCOVERY RATES IN SINGLE-STEP MULTIPLE TESTING PROCEDURES 1. BY SANAT K. SARKAR Temple University

The General Linear Model (GLM)

Familywise Error Rate Controlling Procedures for Discrete Data

A GENERAL DECISION THEORETIC FORMULATION OF PROCEDURES CONTROLLING FDR AND FNR FROM A BAYESIAN PERSPECTIVE

Granger Mediation Analysis of Functional Magnetic Resonance Imaging Time Series

Transcription:

New Procedures for False Discovery Control Christopher R. Genovese Department of Statistics Carnegie Mellon University http://www.stat.cmu.edu/ ~ genovese/ Elisha Merriam Department of Neuroscience University of Pittsburgh This work partially supported by NIH Grant 1 R01 NS047493-01 and NSF Grant SES 9866147..

Multiple Testing In fmri In fmri, we attach statistics to brain locations voxels, triangles, regions but our inferences combine many locations. For m simultaneous hypothesis tests, get p-values P 1,...,P m. Consider rejecting null hypotheses i with P i t. For any rejection threshold t, classify results as follows: H 0 Retained H 0 Rejected Total H 0 True TN FD T 0 H 0 False FN TD T 1 Total N D m Mnemonics: T/F = True/False, D/N = Discovery/Nondiscovery Want to find t that yields both sensitivity (small FN) and specificity (small FD), but these goals are in conflict. Must find a desirable balance.

How to Choose a Threshold? Uncorrected Testing (Control Per-Comparison Type I Error) high sensitivity, low specificity (equivalently: high power, many type I errors) bounds probability of false discovery for each test separately may be useful for small ROIs Strong Control of Familywise Type I Error Rate (FWER) low sensitivity, high specificity (equivalently: low power, few type I errors) guarantees P { FD > 0 } α for any subset of true nulls methods: traditional (e.g., Bonferroni), resampling, random field Move the Slider Until the Pattern Looks Right Uhhh, no.

More Power To You Strong control of FWER is effective but often want more power. The problem is how to get more power while remaining principled. Alternative criteria: Cluster-size Threshold (Forman et al. 1996): Use spatial information to separate signal from noise. Control k-fwer: Fix k and guarantee P { FD > k } α. False Discovery Control: Bound ratio FD/D in some way. Leading example (Benjamini and Hochberg 1995): guarantee that FDR E FD D +1 { D =0 } α. FDR is called the False Discovery Rate.

Controlling k-fwer k-fwer generalizes FWER by not penalizing some fixed number of errors. Several approaches developed (van der Laan et al. 2004, Romano and Lehmann 2004, Finner and Roters 2002, Sarkar 2002, Genovese and Wasserman 2002), but easiest is augmentation approach of van der Laan et al. (2004). Begin with FWER-controlling procedure and reject k additional null hypotheses. Typically use the next k smallest p-values, but can split k in any way by slice, anatomical region,... whatever. But wait, shouldn t we keep k small relative to the total number of rejections (e.g., active voxels)? Ah, that s where False Discovery Control comes in.

The Benjamini-Hochberg (BH) Procedure Sort the m p-values 0 P (0) <P (1) < <P (m). The BH procedure has rejection threshold T BH where T BH = max {P (i) : P (i) α im }, 0 i m. Almost trivial to compute. BH method guarantees that FDR T 0 m α. Proven for independent and positive regression dependent p-values. Hard to break in general. Distribution-free variant is simple but very conservative. Yields more power than Bonferroni (FWER control) and fewer Type I errors than uncorrected testing. BH method overcontrols FDR by factor T 0 /m.

The Benjamini-Hochberg Procedure (cont d) p value 0.0 0.2 0.4 0.6 0.8 m = 50, α =0.1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 0 1 0 0 1 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0.0 0.2 0.4 0.6 0.8 1.0 Index/m

Issues for fmri Performance and Interpretation How do we choose α? k? How do we interpret the bounds? Do these bounds meet our needs? How can we get more power? Dependence Are fmri p-values positive regression dependent? How do we test this? How robust is the BH method to dependence assumptions? Smoothing How should the data be smoothed? Spatial Structure These methods ignore location information. How can we make regions rather than voxels the unit of inference?

Adaptive FDR Control If we knew the proportion of true nulls T 0 /m, we could get FDR = α by using BH with β = α/(t 0 /m). Instead: plug in an estimate. Several approaches developed (Benjamini and Hochberg 2000, Storey 2003, Genovese and Wasserman 2003, Benjamini et al. 2004). Easiest is Benjamini et al. (2004): A. Apply BH procedure at level α. If nothing rejected, STOP. B. Define µ j =(m +1 j)/(1 P (j) )for1 j m. C. Let k = min{j 2:µ j >µ j 1 }. D. Let T 0 /m = min ( 1, µ k /m ). E. Apply BH procedure at level β = α/( T 0 /m).

Controlling False Discovery Exceedance When controlling FDR, it is tempting to use the results to bound the # of false discoveries, but D α need not bound FD. Goal: Find a procedure that keeps FD/D low with high probability. Useful to define the False Discovery Proportion (FDP) at threshold t by FDP(t) = FD(t) D(t)+1 { D(t) =0 }. Note: FDR(t) = E(FDP(t))

Confidence Thresholds In practice, it is useful to be able to make a confidence statement about FDP. A(α, 1 γ) confidence threshold T is a statistic that satisfies P { FDP(T ) α } 1 γ. This implies that FDR α(1 γ)+γ, though this is conservative. Can obtain such a procedure via a confidence envelope (Genovese and Wasserman 2003, 2004), a step-down testing proceudre (Romano and Lehmann 2004), or by augmenting a FWER-controlling procedure (van der Laan et al. 2004).

Confidence Envelopes A1 γ confidence envelope for FDP is a random function FDP(t) on [0, 1] such that P { FDP(t) FDP(t) for all t } 1 γ. Given such an envelope, we can construct confidence thresholds. Two special cases have proved useful. Fixed-ceiling: T = sup{t: FDP(t) α}. Minimum-envelope: T = sup{t: FDP(t) = min t FDP(t)}. FDP

Exact Confidence Envelopes Let FDP maximize FDP over all subsets that look Uniform according to a hypothesis test. Can compute in linear time for suitable uniformity tests. Traditional uniformity tests have little power in the direction needed here. In contrast, using the kth order statistic as a one-sided test statistic the P (k) test has what we need: For small k, sensitive to departures that have a large impact on FDP. Good power. Computing the confidence envelopes is linear in m. Can choose k automatically.

Data Example Monkeys exhibit visual remapping in parietal cortex When the eyes move so that the receptive field of a neuron lands on a previously stimulated location, the neuron fires even though no stimulus is present. Implies transformation in neural representation with eye movements. (Duhamel et al. 1992) Seek evidence for remapping in human cortex. See Merriam, Genovese, and Colby (2003). Neuron, 39, 361 373 for more details. EPI-RT acquisition, TR 2s, TE 30ms, 20 oblique slices, 3.125mm 3.125mm 3mm voxels.

Results: P (k) 90% Confidence Envelopes For k =1, 10, 25, 50, 100, with 0.05 FDP level marked. FDP 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.0000 0.0005 0.0010 0.0015 0.0020 0.0025 p value threshold

Results: Confidence Threshold FDP bound: α =0.05 and Confidence Level: 1 γ =0.9 Frontal Eye Field Supplementary Eye Field Inferior Prefrontal Inferior Prefrontal Superior Parietal 14.28 Temporal-parietal junction Extrastriate Visual Cortex Temporal-parietal junction t 0

Results: Confidence Threshold versus BH FDP bound: α =0.05 and Confidence Level: 1 γ =0.9 Frontal Eye Field Supplementary Eye Field Inferior Prefrontal Inferior Prefrontal Superior Parietal Temporal-parietal junction Extrastriate Visual Cortex Temporal-parietal junction Red = BH + Confidence Blue = BH only

Results: Confidence Threshold versus Bonferroni FDP bound: α =0.05 and Confidence Level: 1 γ =0.9 Frontal Eye Field Supplementary Eye Field Inferior Prefrontal Inferior Prefrontal Superior Parietal Temporal-parietal junction Extrastriate Visual Cortex Temporal-parietal junction Red = Confidence + Bonferroni Blue = Confidence only

False Discovery Control for Random Fields Multiple testing methods based on the excursions of random fields are widely used in functional neuroimaging (e.g., Worsley 2003, Worsley et al. 2002, Cao and Worsley 2001). False Discovery Control extends to this setting under the same basic assumptions (Pacifico Perone, Genovese, et al. 2004). For a set S and a random field X = {X(s): s S} with mean function µ(s), use the realized value of X to test the collection of one-sided hypotheses Let S 0 = {s S : µ(s) =0}. H 0,s : µ(s) = 0 versus H 1,s : µ(s) > 0.

False Discovery Control for Random Fields Define a spatial version of FDP by FDP(t) = Volume(S 0 {s S : X(s) t}). Volume({s S : X(s) t}) Approach analogous to confidence thresholds: i. Find a confidence superset U such that P { U S 0 } 1 γ. ii. Compute a confidence envelope for FDP: FDP(t) = Volume(U {s S : X(s) >t}), Volume({s S : X(s) >t}) iii. Select threshold from confidence envelope. With appropriate choice of α (FDP bound) and 1 γ (confidence level), can control FDR or FDP exceedance.

Controlling the Proportion of False Regions We can use this to make inferences at the region level. (See also Taylor 2004 for another good idea on this.) Say a region R is false at tolerance ɛ if more than an ɛ proportion of its area is in null set S 0 : error(r) = λ(r S 0) ɛ. λ(r) Decompose the t-level set of X into its connected components C t1,...,c tkt. For each level t, let ξ(t) denote the proportion of false regions (at tolerance ɛ) out of k t regions. Then, # { 1 i k t : ξ(t) = k t gives a 1 γ confidence envelope for ξ. λ(c ti U) λ(c ti ) } ɛ

Results: False Region Control Threshold P { prop n false regions 0.1 } 0.95 where false means null overlap 10% Frontal Eye Field Supplementary Eye Field Inferior Prefrontal Inferior Prefrontal Superior Parietal Temporal-parietal junction Extrastriate Visual Cortex Temporal-parietal junction

Take Home Points Though multiple testing is only one among many problems in fmri, recent research has produced a variety of useful error criteria and efficient procedures. Confidence thresholds have advantages for False Discovery Control. In particular, we gain a stronger inferential guarantee with little effective loss of power. Dependence tends to make all these procedures more conservative, but the bounds are still valid for practically relevant dependence. False Discovery Control procedures are available for smoothed or unsmoothed data. For spatial applications, we care about clusters/regions/areas not voxels. Current methods are extending to such region-based inferences.