Intuitive Biostatistics: Choosing a statistical test

Similar documents
Introduction and Descriptive Statistics p. 1 Introduction to Statistics p. 3 Statistics, Science, and Observations p. 5 Populations and Samples p.

Small n, σ known or unknown, underlying nongaussian

Analyzing Data with GraphPad Prism

Rank-Based Methods. Lukas Meier

Turning a research question into a statistical question.

Background to Statistics

NON-PARAMETRIC STATISTICS * (

Introduction to Statistical Analysis

CHAPTER 17 CHI-SQUARE AND OTHER NONPARAMETRIC TESTS FROM: PAGANO, R. R. (2007)

Everything is not normal

Data are sometimes not compatible with the assumptions of parametric statistical tests (i.e. t-test, regression, ANOVA)

CHI SQUARE ANALYSIS 8/18/2011 HYPOTHESIS TESTS SO FAR PARAMETRIC VS. NON-PARAMETRIC

Glossary. The ISI glossary of statistical terms provides definitions in a number of different languages:

Non-parametric tests, part A:

APPENDIX B Sample-Size Calculation Methods: Classical Design

Textbook Examples of. SPSS Procedure

Statistics: revision

Physics 509: Non-Parametric Statistics and Correlation Testing

ANOVA - analysis of variance - used to compare the means of several populations.

Lecture 7: Hypothesis Testing and ANOVA

DETAILED CONTENTS PART I INTRODUCTION AND DESCRIPTIVE STATISTICS. 1. Introduction to Statistics

This is particularly true if you see long tails in your data. What are you testing? That the two distributions are the same!

CDA Chapter 3 part II

psychological statistics

Types of Statistical Tests DR. MIKE MARRAPODI

Nonparametric Statistics. Leah Wright, Tyler Ross, Taylor Brown

Chapter 15: Nonparametric Statistics Section 15.1: An Overview of Nonparametric Statistics

Nonparametric statistic methods. Waraphon Phimpraphai DVM, PhD Department of Veterinary Public Health

Intro to Parametric & Nonparametric Statistics

Transition Passage to Descriptive Statistics 28

What is a Hypothesis?

NONPARAMETRICS. Statistical Methods Based on Ranks E. L. LEHMANN HOLDEN-DAY, INC. McGRAW-HILL INTERNATIONAL BOOK COMPANY

Statistics and Measurement Concepts with OpenStat

Introduction to Statistics with GraphPad Prism 7

Analysis of variance (ANOVA) Comparing the means of more than two groups

Inferences About the Difference Between Two Means

Power and nonparametric methods Basic statistics for experimental researchersrs 2017

Non-parametric (Distribution-free) approaches p188 CN

Module 9: Nonparametric Statistics Statistics (OA3102)

Basic Business Statistics, 10/e

Nemours Biomedical Research Statistics Course. Li Xie Nemours Biostatistics Core October 14, 2014

QUANTITATIVE TECHNIQUES

Contents. Acknowledgments. xix

Exam details. Final Review Session. Things to Review

Hypothesis testing, part 2. With some material from Howard Seltman, Blase Ur, Bilge Mutlu, Vibha Sazawal

SEVERAL μs AND MEDIANS: MORE ISSUES. Business Statistics

My data doesn t look like that..

3. Nonparametric methods

Parametric versus Nonparametric Statistics-when to use them and which is more powerful? Dr Mahmoud Alhussami

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n =

HYPOTHESIS TESTING II TESTS ON MEANS. Sorana D. Bolboacă

Statistics Handbook. All statistical tables were computed by the author.

Nonparametric Statistics

t-test for b Copyright 2000 Tom Malloy. All rights reserved. Regression

Violating the normal distribution assumption. So what do you do if the data are not normal and you still need to perform a test?

Basic Statistical Analysis

Glossary for the Triola Statistics Series

SAS/STAT 14.1 User s Guide. Introduction to Nonparametric Analysis

Non-parametric methods

SPSS Guide For MMI 409

Non-Parametric Statistics: When Normal Isn t Good Enough"

N Utilization of Nursing Research in Advanced Practice, Summer 2008

Physics 509: Bootstrap and Robust Parameter Estimation

PSY 305. Module 3. Page Title. Introduction to Hypothesis Testing Z-tests. Five steps in hypothesis testing

The Inductive Proof Template

Inferential Statistics

LOOKING FOR RELATIONSHIPS

What Are Nonparametric Statistics and When Do You Use Them? Jennifer Catrambone

Statistics Boot Camp. Dr. Stephanie Lane Institute for Defense Analyses DATAWorks 2018

Course Review. Kin 304W Week 14: April 9, 2013

Formulas and Tables by Mario F. Triola

Rama Nada. -Ensherah Mokheemer. 1 P a g e

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing

Nonparametric tests, Bootstrapping

z and t tests for the mean of a normal distribution Confidence intervals for the mean Binomial tests

Selection should be based on the desired biological interpretation!

Data Analysis and Statistical Methods Statistics 651

Introduction to Inferential Statistics. Jaranit Kaewkungwal, Ph.D. Faculty of Tropical Medicine Mahidol University

Introduction to Nonparametric Analysis (Chapter)

Mathematical Notation Math Introduction to Applied Statistics

Degrees of freedom df=1. Limitations OR in SPSS LIM: Knowing σ and µ is unlikely in large

Regression, Part I. - In correlation, it would be irrelevant if we changed the axes on our graph.

Module 03 Lecture 14 Inferential Statistics ANOVA and TOI

Correlation and Regression

Statistics Toolbox 6. Apply statistical algorithms and probability models

GROUPED DATA E.G. FOR SAMPLE OF RAW DATA (E.G. 4, 12, 7, 5, MEAN G x / n STANDARD DEVIATION MEDIAN AND QUARTILES STANDARD DEVIATION

Chapter 24. Comparing Means. Copyright 2010 Pearson Education, Inc.

6.041SC Probabilistic Systems Analysis and Applied Probability, Fall 2013 Transcript Tutorial:A Random Number of Coin Flips

Finding Relationships Among Variables

Dealing with the assumption of independence between samples - introducing the paired design.

CHAPTER 10 Comparing Two Populations or Groups

Unit 14: Nonparametric Statistical Methods

Analysis of categorical data S4. Michael Hauptmann Netherlands Cancer Institute Amsterdam, The Netherlands

One-sample categorical data: approximate inference

Chapter 1 Statistical Inference

Do students sleep the recommended 8 hours a night on average?

Formulas and Tables. for Elementary Statistics, Tenth Edition, by Mario F. Triola Copyright 2006 Pearson Education, Inc. ˆp E p ˆp E Proportion

We're in interested in Pr{three sixes when throwing a single dice 8 times}. => Y has a binomial distribution, or in official notation, Y ~ BIN(n,p).

COMP6053 lecture: Sampling and the central limit theorem. Markus Brede,

Transcription:

pagina 1 van 5 < BACK Intuitive Biostatistics: Choosing a statistical This is chapter 37 of Intuitive Biostatistics (ISBN 0-19-508607-4) by Harvey Motulsky. Copyright 1995 by Oxfd University Press Inc. All rights reserved. You may der the book from GraphPad Software with a software purchase, from any academic bookste, from amazon.com. Learn how to interpret the results of statistical s and about our programs GraphPad InStat and GraphPad Prism. REVIEW OF AVAILABLE STATISTICAL TESTS This book has discussed many different statistical s. To select the right, ask yourself two questions: What kind of data have you collected? What is your goal? Then refer to Table 37.1. All s are described in this book and are perfmed by InStat, except f s marked with asterisks. Tests labeled with a single asterisk are briefly mentioned in this book, and s labeled with two asterisks are not mentioned at all. Table 37.1. Selecting a statistical Goal Describe one group Compare one group to a hypothetical value Compare two unpaired groups Compare two paired groups Compare three me unmatched groups Compare three me matched groups Type of Data Measurement (from Gaussian Population) Mean, SD One-sample t Rank, Sce, Measurement (from Non- Gaussian Population) Median, interquartile range Wilcoxon Binomial (Two Possible Outcomes) Proption Chi-square Binomial ** Unpaired t Mann-Whitney Fisher's (chi-square f large samples) Paired t Wilcoxon McNemar's One-way ANOVA Repeatedmeasures ANOVA Kruskal-Wallis Chi-square Survival Time Kaplan Meier survival curve Log-rank Mantel- Haenszel* Conditional s * Friedman Cochrane Q** Conditional s *

pagina 2 van 5 Quantify association between two variables Pearson crelation Spearman crelation Contingency coefficients** Predict value from another measured variable Simple linear regression Nonlinear regression Nonparametric * Simple logistic Predict value from several measured binomial variables Multiple linear Multiple nonlinear * Multiple logistic REVIEW OF NONPARAMETRIC TESTS Choosing the right to compare measurements is a bit tricky, as you must choose between two families of s: parametric and nonparametric. Many -statistical are based upon the assumption that the data are sampled from a Gaussian distribution. These s are referred to as parametric s. Commonly used parametric s are listed in the first column of the table and include the t and analysis of variance. Tests that do not make assumptions about the population distribution are referred to as nonparametric- s. You've already learned a bit about nonparametric s in previous chapters. All commonly used nonparametric s rank the outcome variable from low to high and then analyze the ranks. These s are listed in the second column of the table and include the Wilcoxon, Mann-Whitney, and Kruskal-Wallis s. These s are also called distribution-free s. CHOOSING BETWEEN PARAMETRIC AND NONPARAMETRIC TESTS: THE EASY CASES Choosing between parametric and nonparametric s is sometimes easy. You should definitely choose a parametric if you are sure that your data are sampled from a population that follows a Gaussian distribution (at least approximately). You should definitely select a nonparametric in three situations: The outcome is a rank a sce and the population is clearly not Gaussian. Examples include class ranking of students, the Apgar sce f the health of newbn babies (measured on a scale of 0 to IO and where all sces are integers), the visual analogue sce f pain (measured on a continuous scale where 0 is no pain and 10 is unbearable pain), and the star scale commonly used by movie and restaurant critics (* is OK, ***** is fantastic). Some values are "off the scale," that is, too high too low to measure. Even if the population is Gaussian, it is impossible to analyze such data with a parametric since you don't know all of the values. Using a nonparametric with these data is simple. Assign values too low to measure an arbitrary very low value and assign values too high to measure an arbitrary very high value. Then perfm a nonparametric. Since the nonparametric only knows about the relative ranks of the values, it won't matter that you didn't know all the values exactly. The data ire measurements, and you are sure that the population is not distributed in a Gaussian manner. If the data are not sampled from a Gaussian distribution, consider whether you can transfmed the values to make the distribution become Gaussian. F example, you might take the logarithm reciprocal of all values. There are often biological chemical reasons (as well as statistical ones) f perfming a

pagina 3 van 5 particular transfm. CHOOSING BETWEEN PARAMETRIC AND NONPARAMETRIC TESTS: THE HARD CASES It is not always easy to decide whether a sample comes from a Gaussian population. Consider these points: If you collect many data points (over a hundred so), you can look at the distribution of data and it will be fairly obvious whether the distribution is approximately bell shaped. A fmal statistical (Kolmogov-Smirnoff, not explained in this book) can be used to whether the distribution of the data differs significantly from a Gaussian distribution. With few data points, it is difficult to tell whether the data are Gaussian by inspection, and the fmal has little power to discriminate between Gaussian and non-gaussian distributions. You should look at previous data as well. Remember, what matters is the distribution of the overall population, not the distribution of your sample. In deciding whether a population is Gaussian, look at all available data, not just data in the current experiment. Consider the source of scatter. When the scatter comes from the sum of numerous sources (with no one source contributing most of the scatter), you expect to find a roughly Gaussian distribution. When in doubt, some people choose a parametric (because they aren't sure the Gaussian assumption is violated), and others choose a nonparametric (because they aren't sure the Gaussian assumption is met). CHOOSING BETWEEN PARAMETRIC AND NONPARAMETRIC TESTS: DOES IT MATTER? Does it matter whether you choose a parametric nonparametric? The answer depends on sample size. There are four cases to think about: Large sample. What happens when you use a parametric with data from a nongaussian population? The central limit theem (discussed in Chapter 5) ensures that parametric s wk well with large samples even if the population is non- Gaussian. In other wds, parametric s are robust to deviations from Gaussian distributions, so long as the samples are large. The snag is that it is impossible to say how large is large enough, as it depends on the nature of the particular non-gaussian distribution. Unless the population distribution is really weird, you are probably safe choosing a parametric when there are at least two dozen data points in each group. Large sample. What happens when you use a nonparametric with data from a Gaussian population? Nonparametric s wk well with large samples from Gaussian populations. The P values tend to be a bit too large, but the discrepancy is small. In other wds, nonparametric s are only slightly less powerful than parametric s with large samples. Small samples. What happens when you use a parametric with data from nongaussian populations? You can't rely on the central limit theem, so the P value may be inaccurate. Small samples. When you use a nonparametric with data from a Gaussian population, the P values tend to be too high. The nonparametric s lack statistical power with small samples. Thus, large data sets present no problems. It is usually easy to tell if the data come from a Gaussian population, but it doesn't really matter because the nonparametric s are so powerful and the parametric s are so robust. Small data sets present a dilemma. It is difficult to tell if the data come from a Gaussian population, but it matters a lot. The nonparametric s are not powerful and the parametric s are not robust.

pagina 4 van 5 ONE- OR TWO-SIDED P VALUE? With many s, you must choose whether you wish to calculate a one- two-sided P value (same as one- two-tailed P value). The difference between one- and two-sided P values was discussed in Chapter 10. Let's review the difference in the context of a t. The P value is calculated f the null hypothesis that the two population means are equal, and any discrepancy between the two sample means is due to chance. If this null hypothesis is true, the one-sided P value is the probability that two sample means would differ as much as was observed ( further) in the direction specified by the hypothesis just by chance, even though the means of the overall populations are actually equal. The twosided P value also includes the probability that the sample means would differ that much in the opposite direction (i.e., the other group has the larger mean). The two-sided P value is twice the one-sided P value. A one-sided P value is appropriate when you can state with certainty (and befe collecting any data) that there either will be no difference between the means that the difference will go in a direction you can specify in advance (i.e., you have specified which group will have the larger mean). If you cannot specify the direction of any difference befe collecting data, then a two-sided P value is me appropriate. If in doubt, select a two-sided P value. If you select a one-sided, you should do so befe collecting any data and you need to state the direction of your experimental hypothesis. If the data go the other way, you must be willing to attribute that difference ( association crelation) to chance, no matter how striking the data. If you would be intrigued, even a little, by data that goes in the "wrong" direction, then you should use a two-sided P value. F reasons discussed in Chapter 10, I recommend that you always calculate a two-sided P value. PAIRED OR UNPAIRED TEST? When comparing two groups, you need to decide whether to use a paired. When comparing three me groups, the term paired is not apt and the term repeated measures is used instead. Use an unpaired to compare groups when the individual values are not paired matched with one another. Select a paired repeated-measures when values represent repeated measurements on one subject (befe and after an intervention) measurements on matched subjects. The paired repeated-measures s are also appropriate f repeated labaty experiments run at different times, each with its own control. You should select a paired when values in one group are me closely crelated with a specific value in the other group than with random values in the other group. It is only appropriate to select a paired when the subjects were matched paired befe the data were collected. You cannot base the pairing on the data you are analyzing. FISHER'S TEST OR THE CHI-SQUARE TEST? When analyzing contingency tables with two rows and two columns, you can use either Fisher's exact the chi-square. The Fisher's is the best choice as it always gives the exact P value. The chi-square is simpler to calculate but yields only an approximate P value. If a computer is doing the calculations, you should choose Fisher's unless you prefer the familiarity of the chi-square. You should definitely avoid the chi-square when the numbers in the contingency table are very small (any number less than about six). When the numbers are larger, the P values repted by the chi-square and Fisher's will he very similar. The chi-square calculates approximate P values, and the Yates' continuity crection is designed to make the approximation better. Without the Yates' crection, the P values are

pagina 5 van 5 too low. However, the crection goes too far, and the resulting P value is too high. Statisticians give different recommendations regarding Yates' crection. With large sample sizes, the Yates' crection makes little difference. If you select Fisher's, the P value is exact and Yates' crection is not needed and is not available. REGRESSION OR CORRELATION? Linear regression and crelation are similar and easily confused. In some situations it makes sense to perfm both calculations. Calculate linear crelation if you measured both X and Y in each subject and wish to quantity how well they are associated. Select the Pearson (parametric) crelation coefficient if you can assume that both X and Y are sampled from Gaussian populations. Otherwise choose the Spearman nonparametric crelation coefficient. Don't calculate the crelation coefficient ( its confidence interval) if you manipulated the X variable. Calculate linear regressions only if one of the variables (X) is likely to precede cause the other variable (Y). Definitely choose linear regression if you manipulated the X variable. It makes a big difference which variable is called X and which is called Y, as linear regression calculations are not symmetrical with respect to X and Y. If you swap the two variables, you will obtain a different regression line. In contrast, linear crelation calculations are symmetrical with respect to X and Y. If you swap the labels X and Y, you will still get the same crelation coefficient. Learn how to interpret the results of statistical s and about our programs GraphPad InStat and GraphPad Prism. Visit the GraphPad home page