High-Throughput Sequencing Course. Introduction. Introduction. Multiple Testing. Biostatistics and Bioinformatics. Summer 2018

Similar documents
Looking at the Other Side of Bonferroni

High-throughput Testing

Multiple testing: Intro & FWER 1

False discovery rate and related concepts in multiple comparisons problems, with applications to microarray data

Table of Outcomes. Table of Outcomes. Table of Outcomes. Table of Outcomes. Table of Outcomes. Table of Outcomes. T=number of type 2 errors

Sta$s$cs for Genomics ( )

Alpha-Investing. Sequential Control of Expected False Discoveries

Multiple Testing. Hoang Tran. Department of Statistics, Florida State University

Lecture 7 April 16, 2018

Statistical testing. Samantha Kleinberg. October 20, 2009

Familywise Error Rate Controlling Procedures for Discrete Data

Hunting for significance with multiple testing

University of California, Berkeley

Advanced Statistical Methods: Beyond Linear Regression

Statistical Applications in Genetics and Molecular Biology

Summary and discussion of: Controlling the False Discovery Rate: A Practical and Powerful Approach to Multiple Testing

Week 5 Video 1 Relationship Mining Correlation Mining

The Pennsylvania State University The Graduate School Eberly College of Science GENERALIZED STEPWISE PROCEDURES FOR

Non-specific filtering and control of false positives

False Discovery Rate

Step-down FDR Procedures for Large Numbers of Hypotheses

Sample Size Estimation for Studies of High-Dimensional Data

Biostatistics Advanced Methods in Biostatistics IV

Lecture 28. Ingo Ruczinski. December 3, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University

Linear Combinations. Comparison of treatment means. Bruce A Craig. Department of Statistics Purdue University. STAT 514 Topic 6 1

A Sequential Bayesian Approach with Applications to Circadian Rhythm Microarray Gene Expression Data

Family-wise Error Rate Control in QTL Mapping and Gene Ontology Graphs

Lecture 21: October 19

Controlling Bayes Directional False Discovery Rate in Random Effects Model 1

REPRODUCIBLE ANALYSIS OF HIGH-THROUGHPUT EXPERIMENTS

Specific Differences. Lukas Meier, Seminar für Statistik

Lecture 6 April

Improving the Performance of the FDR Procedure Using an Estimator for the Number of True Null Hypotheses

HYPOTHESIS TESTING. Hypothesis Testing

Lecture 3: A basic statistical concept

On Procedures Controlling the FDR for Testing Hierarchically Ordered Hypotheses

Stat 206: Estimation and testing for a mean vector,

Let us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided

Some General Types of Tests

Class 4: Classification. Quaid Morris February 11 th, 2011 ML4Bio

Statistical Applications in Genetics and Molecular Biology

ON STEPWISE CONTROL OF THE GENERALIZED FAMILYWISE ERROR RATE. By Wenge Guo and M. Bhaskara Rao

Exam: high-dimensional data analysis January 20, 2014

Single gene analysis of differential expression

PROCEDURES CONTROLLING THE k-fdr USING. BIVARIATE DISTRIBUTIONS OF THE NULL p-values. Sanat K. Sarkar and Wenge Guo

IEOR165 Discussion Week 12

Applying the Benjamini Hochberg procedure to a set of generalized p-values

Multiple Hypothesis Testing in Microarray Data Analysis

Sequential Analysis & Testing Multiple Hypotheses,

Modified Simes Critical Values Under Positive Dependence

Lecture 12: Multiple Hypothesis Testing

The optimal discovery procedure: a new approach to simultaneous significance testing

False discovery control for multiple tests of association under general dependence

FDR and ROC: Similarities, Assumptions, and Decisions

MULTISTAGE AND MIXTURE PARALLEL GATEKEEPING PROCEDURES IN CLINICAL TRIALS

Peak Detection for Images

Controlling the False Discovery Rate: Understanding and Extending the Benjamini-Hochberg Method

Lecture: Mixture Models for Microbiome data

Lecture 27. December 13, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

STAT 263/363: Experimental Design Winter 2016/17. Lecture 1 January 9. Why perform Design of Experiments (DOE)? There are at least two reasons:

Lecture 3: Mixture Models for Microbiome data. Lecture 3: Mixture Models for Microbiome data

Statistical Applications in Genetics and Molecular Biology

Outline. Inference in regression. Critical values GEOGRAPHICALLY WEIGHTED REGRESSION

EMPIRICAL BAYES METHODS FOR ESTIMATION AND CONFIDENCE INTERVALS IN HIGH-DIMENSIONAL PROBLEMS

Multiple Dependent Hypothesis Tests in Geographically Weighted Regression

Research Article Sample Size Calculation for Controlling False Discovery Proportion

Inferential Statistical Analysis of Microarray Experiments 2007 Arizona Microarray Workshop

Aliaksandr Hubin University of Oslo Aliaksandr Hubin (UIO) Bayesian FDR / 25

Statistical Applications in Genetics and Molecular Biology

False Discovery Control in Spatial Multiple Testing

Multiple hypothesis testing using the excess discovery count and alpha-investing rules

22s:152 Applied Linear Regression. Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA)

The miss rate for the analysis of gene expression data

Multiple Testing Issues & K-Means Clustering. Definitions related to the significance level (or type I error) of multiple tests

Chapter Seven: Multi-Sample Methods 1/52

Lecture 2: Descriptive statistics, normalizations & testing

Class 19. Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science. Marquette University MATH 1700

Biochip informatics-(i)

Lesson 11. Functional Genomics I: Microarray Analysis

Two-stage stepup procedures controlling FDR

FALSE DISCOVERY AND FALSE NONDISCOVERY RATES IN SINGLE-STEP MULTIPLE TESTING PROCEDURES 1. BY SANAT K. SARKAR Temple University

22s:152 Applied Linear Regression. 1-way ANOVA visual:

A Large-Sample Approach to Controlling the False Discovery Rate

Adaptive Filtering Multiple Testing Procedures for Partial Conjunction Hypotheses

Estimation of the False Discovery Rate

Probability and Statistics. Terms and concepts

Announcements. Proposals graded

Resampling-Based Control of the FDR

New Procedures for False Discovery Control

Association studies and regression

arxiv: v2 [stat.me] 14 Mar 2011

High-Throughput Sequencing Course

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

A NEW APPROACH FOR LARGE SCALE MULTIPLE TESTING WITH APPLICATION TO FDR CONTROL FOR GRAPHICALLY STRUCTURED HYPOTHESES

Performance Evaluation and Comparison

STAT 5200 Handout #7a Contrasts & Post hoc Means Comparisons (Ch. 4-5)

Normalization and differential analysis of RNA-seq data

Statistical Emerging Pattern Mining with Multiple Testing Correction

Superchain Procedures in Clinical Trials. George Kordzakhia FDA, CDER, Office of Biostatistics Alex Dmitrienko Quintiles Innovation

Quick Calculation for Sample Size while Controlling False Discovery Rate with Application to Microarray Analysis

Transcription:

High-Throughput Sequencing Course Multiple Testing Biostatistics and Bioinformatics Summer 2018 Introduction You have previously considered the significance of a single gene Introduction You have previously considered the significance of a single gene The analysis of high-dimensional data is concerned with assessing the significance of multiple loci/genes Microarray : 20,000-50,000 probe sets GWAS: 500,000-5,000,000 typed SNPs RNA-Seq: 22,000 genes (humans)

Introduction You have previously considered the significance of a single gene The analysis of high-dimensional data is concerned with assessing the significance of multiple loci/genes Microarray : 20,000-50,000 probe sets GWAS: 500,000-5,000,000 typed SNPs RNA-Seq: 22,000 genes (humans) This leads to the Multiple Testing problem Introduction You have previously considered the significance of a single gene The analysis of high-dimensional data is concerned with assessing the significance of multiple loci/genes Microarray : 20,000-50,000 probe sets GWAS: 500,000-5,000,000 typed SNPs RNA-Seq: 22,000 genes (humans) This leads to the Multiple Testing problem Before we address this problem, let s quickly review the single hypothesis case Introduction: Type I and II Errors Recall that in hypothesis testing with a single hypothesis (gene), errors can be classified as: Type I error - rejecting H 0 when it is true Type II error - not rejecting H 0 when it is false

Introduction: Type I and II Errors Recall that in hypothesis testing with a single hypothesis (gene), errors can be classified as: Type I error - rejecting H 0 when it is true Type II error - not rejecting H 0 when it is false We can define probabilities associated with each of these errors: α = P r(reject H 0 H 0 true) β = P r(do not reject H 0 H 0 false) Introduction: Type I and II Errors Decision H 0 True H 0 False Do not reject H 0 1 α β Reject H 0 α 1 β Introduction: Type I and II Errors Always a compromise between type I and type II error

Introduction: Type I and II Errors Always a compromise between type I and type II error An α-level test has probability α of rejecting H 0 when H 0 is true Introduction: Type I and II Errors Always a compromise between type I and type II error An α-level test has probability α of rejecting H 0 when H 0 is true Usually try to use the most powerful (smallest β) test for a given α-level Introduction: Type I and II Errors Always a compromise between type I and type II error An α-level test has probability α of rejecting H 0 when H 0 is true Usually try to use the most powerful (smallest β) test for a given α-level control type I error

Introduction: Multiple Tests Single-hypothesis type-i-error-control proves inadequate when experiment is characterized by a collection of hypotheses Introduction: Multiple Tests Single-hypothesis type-i-error-control proves inadequate when experiment is characterized by a collection of hypotheses Genomics Introduction: Multiple Tests Single-hypothesis type-i-error-control proves inadequate when experiment is characterized by a collection of hypotheses Genomics Why?

Introduction: Multiple Tests Assume we have m hypotheses we wish to test as part of a given experiment Introduction: Multiple Tests Assume we have m hypotheses we wish to test as part of a given experiment These hypotheses could correspond to m genes that we are investigating for differential expression between two groups Introduction: Multiple Tests Assume we have m hypotheses we wish to test as part of a given experiment These hypotheses could correspond to m genes that we are investigating for differential expression between two groups Assume that each hypothesis is tested using a α-level test

Introduction: Multiple Tests Assume we have m hypotheses we wish to test as part of a given experiment These hypotheses could correspond to m genes that we are investigating for differential expression between two groups Assume that each hypothesis is tested using a α-level test Assume that the tests are INDEPENDENT and that the null is true for each of the hypothesis Introduction: Multiple Tests What is the probability of rejecting a null hypothesis even though all hypotheses are actually null? Introduction: Multiple Tests What is the probability of rejecting a null hypothesis even though all hypotheses are actually null? For a given test, we have an α chance of rejecting its null

Introduction: Multiple Tests What is the probability of rejecting a null hypothesis even though all hypotheses are actually null? For a given test, we have an α chance of rejecting its null Therefore we have a 1 α chance of not rejecting Introduction: Multiple Tests What is the probability of rejecting a null hypothesis even though all hypotheses are actually null? For a given test, we have an α chance of rejecting its null Therefore we have a 1 α chance of not rejecting We have m independent α-level tests, so the chance of not rejecting any of them is: (1 α) m Introduction: Multiple Tests What is the probability of rejecting a null hypothesis even though all hypotheses are actually null? For a given test, we have an α chance of rejecting its null Therefore we have a 1 α chance of not rejecting We have m independent α-level tests, so the chance of not rejecting any of them is: (1 α) m Therefore, the chance of rejecting at least one hypothesis is: 1 (1 α) m

Introduction: Multiple Tests 1 (1.05)^m 0.2 0.4 0.6 0.8 1.0 0 20 40 60 80 100 Number of tests (m) Introduction: Notation Gene j (among the m genes) is either associated with the outcome or not The truth is unknown to us The null hypothesis for gene j is denoted by H j H j : gene j is not associated with the outcome of interest The alternative hypothesis is denoted by H j Hj : gene j is associated with the outcome of interest H j and H j are called marginal or local hypotheses Introduction: Marginal and Global Hypotheses H j and H j are called marginal or local hypotheses A global null hypothesis: None of the m genes is associated with the outcome A global alternative: At least one of the m genes is is associated with the outcome Using notation Global Null: H 0 : H 1 and H 2 and... H m Global Alternative: H 1 : H 1 or H 2 or... H m

Introduction: Unadjusted vs Adjusted P-values Suppose that we only test a single gene, say gene j, among the m genes Let p j (lower case p) denote P-value corresponding to H j p j is called the marginal or unadjusted P-value If m hypotheses are tested, inference on H j on the basis of p j is inappropriate The P-value for H j has to account for testing the other m 1 hypotheses We will denote the adjusted P-value by P j (upper case P) When testing multiple genes, using the marginal P-value is inappropriate Why? Introduction: More Notation Suppose that gene j is not associated with the outcome of interest (H j is true) Then Decision rule rejects False-Positive (FP) Decision rule fails to reject True-Negative (TN) Suppose that gene j is associated with the outcome of interest (H j is false) Decision rule rejects True-Positive (TP) Decision rule fails to reject False-Negative (FN) Introduction: Summarizing a Multiple Testing Procedure The results from any multiple testing procedure can be summarized as the following table Accept Reject Total Truth Null A 0 R 0 m 0 Alt. A 1 R 1 m 1 A R m Notation: m: Number of tests, m 0, m 1 number of null/true genes R: Number of genes rejected according to the decision rule A: Number of genes accepted according to the decision rule R 0 /R 1 number of TN/FP A 0 /A 1 number of FN/TP

Introduction: Example Results from an analysis based on m = 10 genes: ## gene truth pvalue ## 1 gene1 0 0.29070 ## 2 gene2 1 0.61630 ## 3 gene3 1 0.00320 ## 4 gene4 0 0.01641 ## 5 gene5 0 0.25150 ## 6 gene6 0 0.58450 ## 7 gene7 0 0.22890 ## 8 gene8 1 0.12630 ## 9 gene9 0 0.26080 ## 10 gene10 0 0.04980 Investigator decides to use following decision rule: Any gene with a corresponding unadjusted P-value of less than 0.05 will be rejected. Reject H j if p j < 0.05 or accept H j otherwise Exercise: Fill in the 2x2 table Accept Reject Total Truth Null A 0 =? R 0 =? m 0 =? Alt. A 1 =? R 1 =? m 1 =? A =? R =? m =? Example: Fill in the 2x2 table Accept Reject Total Truth Null A 0 = 5 R 0 = 2 m 0 = 7 Alt. A 1 = 2 R 1 = 1 m 1 = 3 A = 7 R = 3 m = 10 m 0 = 7 and m 1 = 3 R = 3 will be rejected based on the decision rule Consequently A = m R = 7 will be accepted R 0 = 2, R 1 = 1, A 0 = 5 and A 1 = 2

The Truth What know or observe is this ## gene pvalue ## 1 gene1 0.29070 ## 2 gene2 0.61630 ## 3 gene3 0.00320 ## 4 gene4 0.01641 ## 5 gene5 0.25150 ## 6 gene6 0.58450 ## 7 gene7 0.22890 ## 8 gene8 0.12630 ## 9 gene9 0.26080 ## 10 gene10 0.04980 and not (truth colum is not known to us): dat ## gene truth pvalue ## 1 gene1 0 0.29070 ## 2 gene2 1 0.61630 ## 3 gene3 1 0.00320 ## 4 gene4 0 0.01641 ## 5 gene5 0 0.25150 ## 6 gene6 0 0.58450 ## 7 gene7 0 0.22890 ## 8 gene8 1 0.12630 ## 9 gene9 0 0.26080 ## 10 gene10 0 0.04980 Example: Fill in the 2x2 table (based on what we observe) We can only fill in the bottom row of the table Accept Reject Total Truth Null A 0 R 0 m 0 Alt. A 1 R 1 m 1 A = 7 R = 3 m = 10 The remaining quantities are fixed unknown quantities or unobservable random variables. Comments Accept Reject Total Truth Null A 0 R 0 m 0 Alt. A 1 R 1 m 1 A R m

Comments Accept Reject Total Truth Null A 0 R 0 m 0 Alt. A 1 R 1 m 1 A R m m is a known constant Comments Accept Reject Total Truth Null A 0 R 0 m 0 Alt. A 1 R 1 m 1 A R m m is a known constant m 0 and m 1 are unknown constants Comments Accept Reject Total Truth Null A 0 R 0 m 0 Alt. A 1 R 1 m 1 A R m m is a known constant m 0 and m 1 are unknown constants R and A are determined on the basis of applying the decision rule to the data

Comments Accept Reject Total Truth Null A 0 R 0 m 0 Alt. A 1 R 1 m 1 A R m m is a known constant m 0 and m 1 are unknown constants R and A are determined on the basis of applying the decision rule to the data They are observable random quantities Comments Accept Reject Total Truth Null A 0 R 0 m 0 Alt. A 1 R 1 m 1 A R m m is a known constant m 0 and m 1 are unknown constants R and A are determined on the basis of applying the decision rule to the data They are observable random quantities The true states of the genes of the genes are unknown Comments Accept Reject Total Truth Null A 0 R 0 m 0 Alt. A 1 R 1 m 1 A R m m is a known constant m 0 and m 1 are unknown constants R and A are determined on the basis of applying the decision rule to the data They are observable random quantities The true states of the genes of the genes are unknown A 0, A 1, R 0 and R 1 are unobservable random quantities

Introduction: Multiple Testing Problem Control error rate(s) in multiple testing context Introduction: Multiple Testing Problem Control error rate(s) in multiple testing context Multiple testing methods are designed to control a particular error rate Introduction: Multiple Testing Problem Control error rate(s) in multiple testing context Multiple testing methods are designed to control a particular error rate Multiple error rates exist

Introduction: Multiple Testing Problem Control error rate(s) in multiple testing context Multiple testing methods are designed to control a particular error rate Multiple error rates exist need to chose error rate to control and then method to control it Introduction: Error Rates Family-wise error rate (FWER): the probability of at least one type I error Introduction: Error Rates Family-wise error rate (FWER): the probability of at least one type I error False discovery rate (FDR): the expected proportion of type I errors among the rejected hypotheses.

Family-wise Error Rate (FWER) Probability of committing at least one false-rejection (among m) given that all genes are null Family-wise Error Rate (FWER) Probability of committing at least one false-rejection (among m) given that all genes are null FWER = P (R 1 m = m0) Family-wise Error Rate (FWER) Probability of committing at least one false-rejection (among m) given that all genes are null FWER = P (R 1 m = m0) Note that when m = 1 (single gene), this definition is identical to the type I error we have previously considered

Controlling FWER: Sidak s method Recall that we showed that with m independent α-level tests: FWER = 1 (1 α) m Controlling FWER: Sidak s method Recall that we showed that with m independent α-level tests: FWER = 1 (1 α) m Therefore, 1 FWER = (1 α) m (1 FWER) 1/m = 1 α 1 (1 FWER) 1/m = α Controlling FWER: Sidak s method Recall that we showed that with m independent α-level tests: Therefore, FWER = 1 (1 α) m 1 FWER = (1 α) m (1 FWER) 1/m = 1 α 1 (1 FWER) 1/m = α This suggests that we can control FWER by choosing α for each individual test to be 1 (1 FWER) 1/m

Controlling FWER: Bonferroni Note that (1 x) m 1 mx (for x close to zero) Controlling FWER: Bonferroni Note that (1 x) m 1 mx (for x close to zero) function 0.4 0.5 0.6 0.7 0.8 0.9 1.0 0.00 0.02 0.04 0.06 0.08 0.10 alpha Controlling FWER: Bonferroni Note that (1 x) m 1 mx (for x close to zero) Therefore 1 FWER = (1 α) m 1 FWER 1 mα FWER mα

Controlling FWER: Bonferroni Note that (1 x) m 1 mx (for x close to zero) Therefore 1 FWER = (1 α) m 1 FWER 1 mα FWER mα This suggests that we can control FWER by choosing α for each individual test to be FWER m Controlling FWER: Bonferroni The Bonferroni adjusted P-value is defined as P j = m p j Technical note: P j is defined above could be larger than 1 so a more technically rigorous definition is P j = min{m p j, 1} In other words, if m p j is larger than 1, then truncate P j at 1. Controlling FWER: Holm s Sidak and Bonferroni are one-step approaches

Controlling FWER: Holm s Sidak and Bonferroni are one-step approaches Some power can be gained by sequential methods Controlling FWER: Holm s Sidak and Bonferroni are one-step approaches Some power can be gained by sequential methods The simplest sequential method is Holm s method: Controlling FWER: Holm s Sidak and Bonferroni are one-step approaches Some power can be gained by sequential methods The simplest sequential method is Holm s method: Order the unadjusted P -values such that p 1 p 2... p m

Controlling FWER: Holm s Sidak and Bonferroni are one-step approaches Some power can be gained by sequential methods The simplest sequential method is Holm s method: Order the unadjusted P -values such that p 1 p 2... p m To control FWER, the step-down Holm adjusted P -values are P j = min{(m j + 1) p j, 1} Controlling FWER: Holm s Sidak and Bonferroni are one-step approaches Some power can be gained by sequential methods The simplest sequential method is Holm s method: Order the unadjusted P -values such that p 1 p 2... p m To control FWER, the step-down Holm adjusted P -values are P j = min{(m j + 1) p j, 1} Note that every unadjusted P -value is not multiplied by same factor Controlling FWER: Permutation These methods work well when tests are independent

Controlling FWER: Permutation These methods work well when tests are independent (not generally the case in genomics experiments) Controlling FWER: Permutation These methods work well when tests are independent (not generally the case in genomics experiments) When tests are correlated, these methods are conservative Controlling FWER: Permutation These methods work well when tests are independent (not generally the case in genomics experiments) When tests are correlated, these methods are conservative Permutation approaches are useful in this context

Controlling FWER: Permutation Assume we are interested in assessing differential expression between 2 groups : 1. Compute minimum unadjusted P -value for all genes from the observed data (call it p 1 ) 2. Randomly permute the group labels Breaks relationship between group and expression Reflects sample from global null hypothesis 3. Compute minimum P -value from data set generated in 2 (call it p 1 1 ) 4. Repeat 2 and 3 B times to get p 1 1, p2 1,..., pb 1 5. Compute the proportion of p 1 1, p2 1,..., pb 1 that are p 1 6. This proportion is the permutation adjusted P 1 Controlling FWER: Permutation Can get P 2,..., P m in a similar way (a Holm s-like permutation proceedure) Controlling FWER: Permutation Can get P 2,..., P m in a similar way (a Holm s-like permutation proceedure) Correlation among the genes is accounted for

False Discovery Rate (FDR) Consider the quantity R0 R This is the proportion of of false discoveries among the genes rejected This is an unobservable random quantity (R 0 is not observable) In the FDR framework is based on controlling the expected value of this ratio FDR E[ R0 R ] Expectation is set to zero if R = 0, therefore FDR = E[ R0 R R > 0]P r(r > 0) Note that when m 0 = m (none of the genes are true), FWER=FDR Controlling FDR: Benjamini and Hochberg 1. Order the unadjusted P -values p 1 p 2... p m Controlling FDR: Benjamini and Hochberg 1. Order the unadjusted P -values p 1 p 2... p m 2. Find the largest j such that p j jα/m

Controlling FDR: Benjamini and Hochberg 1. Order the unadjusted P -values p 1 p 2... p m 2. Find the largest j such that p j jα/m 3. Reject null hypotheses affiliated with p 1, p 2,..., p j Controlling FDR: Benjamini and Hochberg 1. Order the unadjusted P -values p 1 p 2... p m 2. Find the largest j such that p j jα/m 3. Reject null hypotheses affiliated with p 1, p 2,..., p j This proceedure will control FDR at αm 0 /m when the tests are independent and continuous Controlling FDR: Benjamini and Hochberg 1. Order the unadjusted P -values p 1 p 2... p m 2. Find the largest j such that p j jα/m 3. Reject null hypotheses affiliated with p 1, p 2,..., p j This proceedure will control FDR at αm 0 /m when the tests are independent and continuous Benjamini and Yekutieli (2001) proposed a modification of the BH procedure that always controls FDR (no larger than αm 0 /m)

Controlling FDR: Benjamini and Hochberg 1. Order the unadjusted P -values p 1 p 2... p m 2. Find the largest j such that p j jα/m 3. Reject null hypotheses affiliated with p 1, p 2,..., p j This proceedure will control FDR at αm 0 /m when the tests are independent and continuous Benjamini and Yekutieli (2001) proposed a modification of the BH procedure that always controls FDR (no larger than αm 0 /m) can be quite conservative Controlling FDR: Benjamini and Hochberg 1. Order the unadjusted P -values p 1 p 2... p m 2. Find the largest j such that p j jα/m 3. Reject null hypotheses affiliated with p 1, p 2,..., p j This proceedure will control FDR at αm 0 /m when the tests are independent and continuous Benjamini and Yekutieli (2001) proposed a modification of the BH procedure that always controls FDR (no larger than αm 0 /m) can be quite conservative Note that when m 0 = m (i.e., all hypotheses are null), these procedures maintain FWER at α Controlling pfdr: q-values FDR = E[ R0 R R > 0]P r(r > 0)

Controlling pfdr: q-values FDR = E[ R0 R R > 0]P r(r > 0) pfdr = E[ R0 R R > 0] positive FDR Controlling pfdr: q-values FDR = E[ R0 R R > 0]P r(r > 0) pfdr = E[ R0 R R > 0] positive FDR Since P r(r > 0) is often 1 in most genomics experiments, FDR and pfdr are ver similar Controlling pfdr: q-values q-value is the minimum pfdr for which the affiliated hypothesis is rejected

Controlling pfdr: q-values q-value is the minimum pfdr for which the affiliated hypothesis is rejected q-value can be interpreted as the expected proportion of false positives incurred when calling that test significant Genome-wide Significance In GWAS papers, α = 5 10 8 is typically considered the threshold for genome-wide significance It is based on a Bonferroni correction: If you consider testing m = 1, 000, 000 SNPs at the FWER level of 0.05, then each SNP should be tested at the α = 0.05 1, 000, 000 = 5 10 8, level Suppose that the unadjusted P=value for a SNP is 5 10 7 Is this reaching genome-wide significance? The term suggestive is also used Reaching Genome-wide Significance Suppose that your m = 1, 000, 000 SNPs are independent The adjusted P-value is P = 5 10 7 m = 5 10 7 10 6 = 0.5, This is off by an order of magnitude (0.5 = 0.05 10) It is not reaching Note: Due to linkage disequiblirium among SNPs the adjusted P-value is likely to be smaller than 0.5 The point is that while 5 10 7 is small number, it may not be small enough when tesing a large number of hypotheses

Conclusions Multiple testing must be accounted for when testing for associations in the context of high-dimensional data FWER and FDR are the two common frameworks for quantifying error Error rate estimates can be used to compute adjusted p-values Resampling-based methods can increase power in controlling error when sample sizes are sufficient for their use. When large-scale patterns of differential expression are observed, it is important to consider if such effects are biologically reasonable, and if technical factors can be attributed to the variation.