Statistical Inference Part 2. t test for single mean. t tests useful for PSYCH 710. Review of Statistical Inference (Part 2) Week 2
|
|
- Wilfrid Dorsey
- 5 years ago
- Views:
Transcription
1 Statistical Inference Part PSYCH 70 eview Statistical Inference (Part ) Week Pr. Patrick ennett t-tests Effect Size Equivalence Tests Consequences Low Power s P t tests useful for t test for single mean t statistic (Ȳ µ) t = (s/ p n) - comparing mean sample to some expected value - comparing samples statistic t - sampling distribution is t distribution - unimodal, symmetric about mean - shape governed by parameter degrees--freedom (df) - when NULL hyposis is true, mean t will be zero code set.seed(55540) mu <- 0 # population mean n <- 0 # sample size stdev <- # population sd < t.val <- rep(0,) p.val <- rep(0,) for(kk in :){.sample <- rnorm(n,mu,stdev) t.results <- t.test(.sample) t.val[kk] <- t.results$statistic p.val[kk] <- t.results$p.value } null hyposis is true null hyposis is true t.test evaluates null hyposis that mu=0 p value t value
2 is true. If null hyposis is true, n z will be between ±.96 95% time and between ±.56 99% time. Therefore, using criteria ±.96 to reject null hyposis will yield a Type I error rate 0.05, whereas criteria ±.56 corresponds to a Type I error rate 0.0. Our observed z score falls outside both sets criteria, and so null hyposis is rejected. It used to be standard practice to indicate which level was used by writing that null hyposis was rejected (p <.05) or (p <.0). Nowadays, scientists are encouraged to publish exact p value for ir observed statistic. In our case, probability drawing a z that was outside range ±3.3 is , and so we would report statistical result by writing sample mean di ered significantly from 00 (z = 3.3, p = ) and so null hyposis µ = 00 was rejected. 0.3 ennett, PJ t tests ennett, PJ Hyposis Testing PSYCH 70 In previous example, we knew that must be estimated from data: di erence between two sample PSYCH 70 Hyposis Testing two independent samples = 0. ut in most cases we we do not know, and refore it sp Y ) i (Yi ˆ=s= (n ) The estimate, s can be used to calculate t, which is similar to a z score: t testsor single mean t statistic Comparing far we have considered how to decide wher a single sample mean was drawn from a population with (Y µ) di erence between two sample p = p =.36 two independent samples a mean µ. More commonly, we want to compare p t =to each (6) or to decide wher or (/ n) (/ two 0) sample (s/ n) populations scores Given So far we have considered how to decide wher single sample mean was drawn from a population not ya di er. If population with two samples are µ and µ, n two-sided null and a mean µ. More commonly, we want to compare two sample to each or to decide wher or - : µa & µb alternative hyposes are: not y di er. If population code two samples are µ and µ, n two-sided null and 4 null hyposis is false mu=0.5 null hyposis is false mu=0.5 - variances: σa & σb alternative hyposes are: set.seed(55540) H0 : µ = µ mu <- 0.5 # population Distributions sample : H0 : µmean = µ H : µ 6= µ n <- 0 # sample size Hennett, : µ 6= PJ µ - : µ a & µb PSYCH 70 Hyposis Testing stdev <- # population sd < a - variances: These hyposes can be re-written as:these hyposes can be re-written as: = Y a t.val <- rep(0,) n b n Y b = 0. 6 di erence between two sample will be normal: p.val <- rep(0,) H0 : µ µ = 0 for(kk in :){ H0 : µ µ = 0 - shape: normal (via Central Limit Theorem) : µ µ 6= 0.sample <-H rnorm(n,mu,stdev) 0.7. two independent samples H : µ µ 6= 0 t.results <- t.test(.sample) Distribution difference between sample Each sample mean is a random variable drawn from sampling distribution mean. The t.val[kk] <- t.results$statistic /n and /n. (We will sampling distributions will be µ<µso and we have variances will be mean: u d = µ a & µb and, far to decide wher a single sample mean was drawn from a population with p.val[kk] t.results$p.value Each sample mean isconsidered adistributions randomhow variable from sampling distribution mean. The- take advantage Central Limit Theorem and assume are normal.)drawn } a mean that µ. More commonly, we wantnow, to let s compare two sample to each or to decide wher or 0 /n and /n. (We - variance: create a new variable that is di erence between sample, d = Y will Y. How sampling distributions be isµd distributed? and µ, and variances will be will di er. If variables population two samples µ 0. a Normal variable, are µ 4 and 6 8,0n two-sided null and It can be shown that sum evaluates (or di erence) not twoy or more Normal is also t.test null hyposis that mu=0 take advantage Central 0.0 Limit Theorem and.0assume that distributions are normal.) Now, let s so d is distributed Normally. Furrmore, alternative mean a sum (or di erence) two variables hyposes are: p valueis sum createa mean new variable thatdistribution is di erence between sample, dt value = Y Y. How is d distributed? (or di erence) two. Therefore, di erence equals di erence between two samplingitdistributions : µ. If two can be shown that Y sum (or di erence) are two variables is also a Normal variable, d = µ Y Y H0 : or µmore = Normal µ equal, n Y d = 0. Finally, variance (or di erence) two independent random variables is so d aissum distributed Normally. Furrmore, mean sumµ (or di erence) two variables is sum H : µ a 6= sum two variances: d = Y + Y. (N.. The variance a di erence between independent (or di erence) two. Therefore, mean di erence distribution equals di erence variables also is sum, not di erence, two variances.) Therefore, if we know form between sampling distributions : Y d = µy µy. If two are populations and, n statistical ory allows us to predict form distribution between These hyposes cantwo be re-written as: samples drawn from those populations (see Figure equal, n Y d =5).0. Finally, variance a sum (or di erence) two independent random variables is or, and refore In general, we will not know eir will have to from : µ be estimated H0 µ = 0 Y sum two Y variances: d = Y + Y. (N.. The variance a di erence between independent data with formula (n )s + (n is )s H : µ µ 6= 0 Therefore, if we know form sum, variables also not di erence, two variances.) ˆY Y = + (7) n + n n n d = Y a Y b + populations and, n statistical ory allows us to predict form distribution between variable drawn from The samples drawn from those populations (see Figure 5).sampling distribution mean. /n. (We will Y sampling distributions will be µ and µ, and variances will be /n and or (Y general, (µwe µ ) ) not know eir In will, and refore will have to be estimated from -sample t-test (8) and assume that Simulation t= Y Y distributions are normal.) takey advantage Central Limit Theorem Now, let s Y data withˆ formula + (n sample, d = Y create a new variable that is di erence between Y. How is d distributed? (n )s )s underlying similarity: with df = n + n. close comparison Equations 6 and 8 will revealˆan = di erence)in two or more + variables is also a Normal variable, (7) It can be shown that parameter, sum Normal -sample t-test H0 is true Y (or both cases, t equals di erence between an observed statistic and a populationy divided n by + n n n so d is distributed Normally. Furrmore, mean a sum (or di erence) two variables is sum standard error statistic: sample pop where (or n and n are sizes Therefore, two groups. The t statistic is sample two. mean two-group di erence distribution equals di erence statisticdi erence) parameter t= µy. If two are estimated between standard error statistic two sampling distributions : Y d = µy code (Y ay ) (or (µdi erence) µ ) equal, n Y d = 0. Finally, variance sum two independent random variables t= (8) is This general formula applies to all t tests. set.seed(65540) + ˆ.Y (N.. Y The variance a di erence between sum variance, two variances: we estimate population so parameter In, a two-sample t testnormally, is done with t.test() command. Notice that mu <- independent 00 # population mean d = var.equal Y Y ourtells test will a tis distribution mu <- 00 # population mean is set to TUE. Using this setting statistic to assume thatfollow samples and have same variables also sum, not variance. di erence, two variances.) Therefore, if we know form standard an errorunderlying n <5 # samplein size = degrees n + n. close comparison Equations 6 and 8estimated will reveal similarity: withwith n + df n - freedom difference form distribution > sample. <- c(95, 9, 93, 96, 98, 99) populations and, n statistical ory allows us to predict stdev between <- 0 # population sd both t equals di erence between an observed statistic and a population parameter,stdev divided by# > sample. <- c(0, 99, 06, 00, 98, cases, ) <0 population sd samples drawn from those populations (see Figure 5). standard error statistic: > t.test(x=sample.,y=sample.,alternative="two.sided", var.equal=tue) < In general, we will not know eir or, and refore Y Y will have to be estimated from t.val <- rep(0,) bove statement t assumes data re: with formula statistic parameter p.val <- rep(0,) t = that are distributed (n )s + (n )s for(kk in :){ 0 standard error statistic + ˆY estimated (7) normally with equal variance Y =.sample. <- rnorm(n,mu,stdev) n + n n n p value.sample. <- rnorm(n,mu,stdev) This general formula applies to all t tests. t.results <- t.test(.sample.,.sample.,var.equal=t) where n and n are sample sizes two groups. The two-group t statistic is t.val[kk] <- t.results$statistic In, a two-sample t test is done with t.test() command. Notice that parameter var.equal p.val[kk] <- t.results$p.value is set to TUE. Using this setting tells to assume and (Y thaty samples ) (µ µ ) have same variance. } t= (8) ˆY Y > sample. <- c(95, 9, 93, 96, 98, 99) > sample. 00, 98, ) with df<= c(0, n + n 99,. 06, close comparison Equations 6 and 8 will reveal an underlying similarity: in > t.test(x=sample.,y=sample.,alternative="two.sided", both cases, t equals di erence between an observed statistic var.equal=tue) and a population parameter, divided by standard error statistic: where n and n are sample sizes two groups. The two-group t statistic is Each sample mean is a random t test for comparing independent t= statistic parameter 0 estimated standard error statistic This general formula applies to all t tests sample t-test H0 is true -4-0 t value 4
3 Simulation -sample t-test (equal variances) -sample t-test H0 is false -sample t-test H0 is false Simulation -sample t-test (unequal variances) -samp t unequal Var H0 is true code set.seed(65540) mu <- 00 # population mean mu <- 05 # population mean n <- 5 # sample size stdev <- 0 # population sd stdev <- 0 # population sd < t.val <- rep(0,) p.val <- rep(0,) for(kk in :){ sample. <- rnorm(n,mu,stdev).sample. <- rnorm(n,mu,stdev) p value t value t.results <- t.test(.sample.,.sample.,var.equal=t) t.val[kk] <- t.results$statistic p.val[kk] <- t.results$p.value } stdev_diff <- sqrt(( ( (5-)*00 + (5-)*00 ) / (5+5-) ) * ( (/5)+(/5) ) ) =.88 Expected Value t = (00-05)/.88 = -.76 (actual mean = -.8) code set.seed(65540) mu <- 00 # population mean mu <- 00 # population mean n <- 5 # sample size stdev <- 0 # population sd stdev <- 3 # population sd < t.val <- rep(0,) p.val <- rep(0,) for(kk in :){.sample. <- rnorm(n,mu,stdev).sample. <- rnorm(n,mu,stdev) t.results <- t.test(.sample.,.sample.,var.equal=t) t.val[kk] <- t.results$statistic p.val[kk] <- t.results$p.value } t value Effect Size Cohen s ds what makes a p-value significant or non-significant? - alpha, sample size (power), effect size classes effect size - d - standardized differences (distances) btwn - r - measures association (variance accounted for) Used for between-subjects designs: d s = X X q (n )SD +(n )SD /(n + n ) d s = t r n + n t/ p N
4 > m<-0 > m<- > n = 0 > stdev <- 4 > s <- rnorm(n,m,stdev) > s <- rnorm(n,m,stdev) > (t.test. <- t.test(s,s,var.equal=t)) Two Sample t-test data: s and s t = , df = 38, p-value = H: true difference in is not equal to 0 95 percent confidence interval: sample estimates: mean x mean y > t <- t.test.$statistic > ( ds <- t*sqrt( (/n)+(/n) ) ) > m<-0 > m<- > n = 00 > stdev <- 4 > s <- rnorm(n,m,stdev) > s <- rnorm(n,m,stdev) > (t.test. <- t.test(s,s,var.equal=t)) Two Sample t-test data: s and s t = , df = 98, p-value =.764e-05 H: true difference in is not equal to 0 95 percent confidence interval: sample estimates: mean x mean y > t <- t.test.$statistic > ( ds <- t*sqrt( (/n)+(/n) ) ) Hedges gs unbiased estimate population ds g s = d s 3 4(n + n ) 9 Convert ds to r Cohen s dz r = d s p d s +(N N)/(n n ) used for within-subjects design (paired scores) - Md = mean difference scores - Di = difference score i - N = number difference scores d z = M d p ( P (Di M d )) /(N ) d z = t p N
5 eta-squared measure association between independent & dependent variables omega-squared unbiased estimate association in population = SS e ect SS total proportion variation in Y accounted for by group membership! = df e ect (MS e ect MS error ) SS total + MS error p = SS e ect SS e ect + SS error for factorial experiments (more than independent variable)! p = df e ect (MS e ect MS error ) df error MS e ect +(N df error ) MS error p = F df e ect F df e ect + df error Why report effect size? Why not just report p-values as index effect? - p-values depend on number subjects/events/measures - increased power leads to lower p values Would like to have measure doesn t depend on experiment N - or on particular aspects experimental design Effect size measures try to do this - and also give information about magnitude effect Sample size affects precision ds estimation simulation parameters: < # 5K samples m <- 0 # mean pop m <- # mean pop stdev <- 4 # sd population simulation : n <- 0 # sample size simulation : n <- 70 # sample size n=n= t (-sample test) Cohens ds n=n= t (-sample test) Cohens ds effect size measures vary across samples variance sampling distribution is larger for smaller samples
6 Effect sizes in published reports (winner s curse) effect size inflation (winner s curse) n=n=0 n=n=70 n = n = 0 apply statistical threshold (e.g., p<.05) to effect size - what is median significant ds? median = depends on sample size t (-sample test) t (-sample test) Significant Cohen ds (p<.05) - for small samples, only big ds are significant - large samples more likely to get significant large & small ds published studies usually report significant results n = n = 70 median = small n studies lead to overestimation true effect size Cohens ds Cohens ds Significant Cohen ds (p<.05) effect size measures vary across samples variance sampling distribution is larger for smaller samples significant ds are more extreme with small samples What does a significant p-value mean? probability replicating p=.05 effect a ejecting null hyposis Null Observed effect p = sample t-test H0 is false significant p-value indicates that result is unusual - assuming null hyposis is true and assumptions are correct That is LL it - it is not equal to probability that H0 is true - it is not equal to probability replicating result.96 x sem observed effect significant p= x sem b The sampling distribution critical value t(8)= x sem expect 50% replications to yield smaller effect t value n=9, mu=.86*sem
7 What does a p-value mean? non-significant p-value indicates that result is NOT unusual assuming null hyposis is true. It does NOT mean that null hyposis is true - e.g., a non-significant -sample t-test does NOT mean that populations are equal - You do not accept null hyposis, you simply fail to reject it. To accept null hyposis, perform an equivalence test Equivalence Tests reverse H0 & H: - H0: re is an effect - H: re is no effect try to reject H0 in favour H define Smallest Effect Sizes Interest (SESOI) - test upper & lower bounds (SESOI) with -tailed t-tests - H0-u: observed effect unusually smaller than upper SESOI? - H0-l: observed effect unusually larger than lower SESOI? if both -tailed tests are significant, n we say observed effect is smaller than SESOI - e.g., two groups are equivalent Equivalence Tests reverse H0 & H: - H0: re is an effect; H: re is no effect - try to reject H0 in favour H define Smallest Effect Sizes Interest (SESOI; L & U) - evaluate hyposis with two -tailed t-tests: H0-u: is observed effect unusually smaller than U? H0-l: observed effect unusually larger than L? if both -tailed tests are significant, n observed effect is smaller than SESOI - e.g., two groups are equivalent Null Hyposis vs Equivalence Tests Four possible outcomes when evaluating difference between groups Null Hyposis Significance Test Equivalence Test Equivalent to Zero Not Equivalent to Zero Not Different from Zero + - Different from Zero - +
8 NHST vs Equivalence Tests (4 outcomes) Spurious findings? equivalence region difference between L 0 U Statistically Equivalent & Different Statistically Equivalent & Not Different Not Equivalent & Different Not Equivalent & Not Different Over time: - re has been a dramatic increase in ease with which we can analyze a single data set in multiple ways - also, effects/phenomena interest typically become increasingly smaller & more subtle as a research area matures - but, sample sizes tend not to change very much - nor have unwritten rules about deciding if a paper is publishable se trends make it likely that number papers reporting spurious statistically significant findings has increased Lakens, D. Equivalence tests: practical primer for t tests, correlations, and metaanalyses. Social Psychological & Personality Science, 8(4), Power (Type I & Type II Errors) Positive Predictive Value () Table : Possible outcomes hyposis testing. decision H0 is True H0 is False reject H0: Type I (p = ) Correct (p = =power) do not reject H0: Correct (p = ) Type II error (p = ) Given that we ve rejected HO in favour H - what is probability that our conclusion is correct? decision H0 is True H0 is False reject H0: p Type I (p = ) Correct (p = =power) do not reject H0: Correct (p = ) Type II error (p = ) p Type I Error: reject H0 when it is true (alpha) Type II Error: fail to reject H0 when it is false (beta) = p / (p + p) Power = Probability rejecting false H0 (-beta)
9 Ioannidis 005 depends on a priori probability that H0 is true Ioannidis 005 Table. esearch Findings and True elationships esearch Finding True elationship Yes No Total Yes c( β)/( + ) cα/( + ) c( + α β)/( + ) No cβ/( + ) c( α)/( + ) c( α + β)/( + ) Total c/( + ) c/( + ) c DOI: 0.37/journal.pmed.0004.t00 alpha is Type error rate (-eta) is power ( minus Type II error rate) c is number effects begin examined is ratio number true effects divided number no effects Positive Predictive Value () is probability that a statistically significant finding is true = (-eta) / ( - *eta + alpha) reduced by low power - small studies & small effects reduced by low - testing more false hyposes - increasing bias greater flexibility in design & analysis selective reporting positive results N.. higher beta = lower power beta = 0. alpha = 0.05 Statistically significant finding >50% chance being true/correct if: (-eta) > alpha Ioannidis 005 effect alpha on Ioannidis 005 effect bias on bias: proportion null results findings that are reported as true effects beta = 0. alpha = 0.05 beta = 0. alpha = 0.0 beta = 0. bias = 0 alpha = 0.0 beta = 0. bias = 0. alpha = 0.0
10 Ioannidis 005 effect bias on bias: proportion null results findings that are reported as true effects Factors that lower beta = 0. bias = 0 alpha = 0.05 beta = 0. bias = 0. alpha = 0.05 small studies & small effects (low power) testing more hyposes (lowering ) greater flexibility in design and analysis (increasing bias) more independent research teams studying phenomenon - due to selective reporting positive results
Population Variance. Concepts from previous lectures. HUMBEHV 3HB3 one-sample t-tests. Week 8
Concepts from previous lectures HUMBEHV 3HB3 one-sample t-tests Week 8 Prof. Patrick Bennett sampling distributions - sampling error - standard error of the mean - degrees-of-freedom Null and alternative/research
More informationPSY 305. Module 3. Page Title. Introduction to Hypothesis Testing Z-tests. Five steps in hypothesis testing
Page Title PSY 305 Module 3 Introduction to Hypothesis Testing Z-tests Five steps in hypothesis testing State the research and null hypothesis Determine characteristics of comparison distribution Five
More informationAn inferential procedure to use sample data to understand a population Procedures
Hypothesis Test An inferential procedure to use sample data to understand a population Procedures Hypotheses, the alpha value, the critical region (z-scores), statistics, conclusion Two types of errors
More informationHYPOTHESIS TESTING. Hypothesis Testing
MBA 605 Business Analytics Don Conant, PhD. HYPOTHESIS TESTING Hypothesis testing involves making inferences about the nature of the population on the basis of observations of a sample drawn from the population.
More informationBusiness Statistics. Lecture 10: Course Review
Business Statistics Lecture 10: Course Review 1 Descriptive Statistics for Continuous Data Numerical Summaries Location: mean, median Spread or variability: variance, standard deviation, range, percentiles,
More informationStatistics Primer. ORC Staff: Jayme Palka Peter Boedeker Marcus Fagan Trey Dejong
Statistics Primer ORC Staff: Jayme Palka Peter Boedeker Marcus Fagan Trey Dejong 1 Quick Overview of Statistics 2 Descriptive vs. Inferential Statistics Descriptive Statistics: summarize and describe data
More informationAdvanced Experimental Design
Advanced Experimental Design Topic Four Hypothesis testing (z and t tests) & Power Agenda Hypothesis testing Sampling distributions/central limit theorem z test (σ known) One sample z & Confidence intervals
More informationCIVL /8904 T R A F F I C F L O W T H E O R Y L E C T U R E - 8
CIVL - 7904/8904 T R A F F I C F L O W T H E O R Y L E C T U R E - 8 Chi-square Test How to determine the interval from a continuous distribution I = Range 1 + 3.322(logN) I-> Range of the class interval
More informationThe t-test: A z-score for a sample mean tells us where in the distribution the particular mean lies
The t-test: So Far: Sampling distribution benefit is that even if the original population is not normal, a sampling distribution based on this population will be normal (for sample size > 30). Benefit
More informationSampling distribution of t. 2. Sampling distribution of t. 3. Example: Gas mileage investigation. II. Inferential Statistics (8) t =
2. The distribution of t values that would be obtained if a value of t were calculated for each sample mean for all possible random of a given size from a population _ t ratio: (X - µ hyp ) t s x The result
More informationStatistical inference (estimation, hypothesis tests, confidence intervals) Oct 2018
Statistical inference (estimation, hypothesis tests, confidence intervals) Oct 2018 Sampling A trait is measured on each member of a population. f(y) = propn of individuals in the popn with measurement
More informationIntroduction to Statistical Data Analysis III
Introduction to Statistical Data Analysis III JULY 2011 Afsaneh Yazdani Preface Major branches of Statistics: - Descriptive Statistics - Inferential Statistics Preface What is Inferential Statistics? The
More informationT.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS
ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS In our work on hypothesis testing, we used the value of a sample statistic to challenge an accepted value of a population parameter. We focused only
More informationT-tests for 2 Independent Means
T-tests for 2 Independent Means January 15, 2019 Contents t-test for 2 independent means Tutorial Example 1: one-tailed test for independent means, equal sample sizes Error Bars Example 2: two-tailed test
More informationChapter 23: Inferences About Means
Chapter 3: Inferences About Means Sample of Means: number of observations in one sample the population mean (theoretical mean) sample mean (observed mean) is the theoretical standard deviation of the population
More informationOutline. PubH 5450 Biostatistics I Prof. Carlin. Confidence Interval for the Mean. Part I. Reviews
Outline Outline PubH 5450 Biostatistics I Prof. Carlin Lecture 11 Confidence Interval for the Mean Known σ (population standard deviation): Part I Reviews σ x ± z 1 α/2 n Small n, normal population. Large
More informationT-tests for 2 Independent Means
T-tests for 2 Independent Means February 22, 208 Contents t-test for 2 independent means Tutorial Example : one-tailed test for independent means, equal sample sizes Error Bars Example 2: two-tailed test
More information" M A #M B. Standard deviation of the population (Greek lowercase letter sigma) σ 2
Notation and Equations for Final Exam Symbol Definition X The variable we measure in a scientific study n The size of the sample N The size of the population M The mean of the sample µ The mean of the
More informationPSY 307 Statistics for the Behavioral Sciences. Chapter 20 Tests for Ranked Data, Choosing Statistical Tests
PSY 307 Statistics for the Behavioral Sciences Chapter 20 Tests for Ranked Data, Choosing Statistical Tests What To Do with Non-normal Distributions Tranformations (pg 382): The shape of the distribution
More informationSingle Sample Means. SOCY601 Alan Neustadtl
Single Sample Means SOCY601 Alan Neustadtl The Central Limit Theorem If we have a population measured by a variable with a mean µ and a standard deviation σ, and if all possible random samples of size
More informationDETAILED CONTENTS PART I INTRODUCTION AND DESCRIPTIVE STATISTICS. 1. Introduction to Statistics
DETAILED CONTENTS About the Author Preface to the Instructor To the Student How to Use SPSS With This Book PART I INTRODUCTION AND DESCRIPTIVE STATISTICS 1. Introduction to Statistics 1.1 Descriptive and
More informationINTERVAL ESTIMATION AND HYPOTHESES TESTING
INTERVAL ESTIMATION AND HYPOTHESES TESTING 1. IDEA An interval rather than a point estimate is often of interest. Confidence intervals are thus important in empirical work. To construct interval estimates,
More informationCOMPARING SEVERAL MEANS: ANOVA
LAST UPDATED: November 15, 2012 COMPARING SEVERAL MEANS: ANOVA Objectives 2 Basic principles of ANOVA Equations underlying one-way ANOVA Doing a one-way ANOVA in R Following up an ANOVA: Planned contrasts/comparisons
More informationThe t-statistic. Student s t Test
The t-statistic 1 Student s t Test When the population standard deviation is not known, you cannot use a z score hypothesis test Use Student s t test instead Student s t, or t test is, conceptually, very
More informationChapter 23. Inferences About Means. Monday, May 6, 13. Copyright 2009 Pearson Education, Inc.
Chapter 23 Inferences About Means Sampling Distributions of Means Now that we know how to create confidence intervals and test hypotheses about proportions, we do the same for means. Just as we did before,
More informationTwo-Sample Inferential Statistics
The t Test for Two Independent Samples 1 Two-Sample Inferential Statistics In an experiment there are two or more conditions One condition is often called the control condition in which the treatment is
More informationMultiple Regression Analysis
Multiple Regression Analysis y = β 0 + β 1 x 1 + β 2 x 2 +... β k x k + u 2. Inference 0 Assumptions of the Classical Linear Model (CLM)! So far, we know: 1. The mean and variance of the OLS estimators
More information16.400/453J Human Factors Engineering. Design of Experiments II
J Human Factors Engineering Design of Experiments II Review Experiment Design and Descriptive Statistics Research question, independent and dependent variables, histograms, box plots, etc. Inferential
More informationLast two weeks: Sample, population and sampling distributions finished with estimation & confidence intervals
Past weeks: Measures of central tendency (mean, mode, median) Measures of dispersion (standard deviation, variance, range, etc). Working with the normal curve Last two weeks: Sample, population and sampling
More information1/24/2008. Review of Statistical Inference. C.1 A Sample of Data. C.2 An Econometric Model. C.4 Estimating the Population Variance and Other Moments
/4/008 Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University C. A Sample of Data C. An Econometric Model C.3 Estimating the Mean of a Population C.4 Estimating the Population
More informationProbability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institution of Technology, Kharagpur
Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institution of Technology, Kharagpur Lecture No. # 36 Sampling Distribution and Parameter Estimation
More informationQuantitative Methods for Economics, Finance and Management (A86050 F86050)
Quantitative Methods for Economics, Finance and Management (A86050 F86050) Matteo Manera matteo.manera@unimib.it Marzio Galeotti marzio.galeotti@unimi.it 1 This material is taken and adapted from Guy Judge
More informationHypothesis Testing and Confidence Intervals (Part 2): Cohen s d, Logic of Testing, and Confidence Intervals
Hypothesis Testing and Confidence Intervals (Part 2): Cohen s d, Logic of Testing, and Confidence Intervals Lecture 9 Justin Kern April 9, 2018 Measuring Effect Size: Cohen s d Simply finding whether a
More informationLab #12: Exam 3 Review Key
Psychological Statistics Practice Lab#1 Dr. M. Plonsky Page 1 of 7 Lab #1: Exam 3 Review Key 1) a. Probability - Refers to the likelihood that an event will occur. Ranges from 0 to 1. b. Sampling Distribution
More informationCHAPTER 7. Hypothesis Testing
CHAPTER 7 Hypothesis Testing A hypothesis is a statement about one or more populations, and usually deal with population parameters, such as means or standard deviations. A research hypothesis is a conjecture
More information10/31/2012. One-Way ANOVA F-test
PSY 511: Advanced Statistics for Psychological and Behavioral Research 1 1. Situation/hypotheses 2. Test statistic 3.Distribution 4. Assumptions One-Way ANOVA F-test One factor J>2 independent samples
More informationSimple Linear Regression
Simple Linear Regression ST 430/514 Recall: A regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates)
More informationVariance. Standard deviation VAR = = value. Unbiased SD = SD = 10/23/2011. Functional Connectivity Correlation and Regression.
10/3/011 Functional Connectivity Correlation and Regression Variance VAR = Standard deviation Standard deviation SD = Unbiased SD = 1 10/3/011 Standard error Confidence interval SE = CI = = t value for
More informationLast week: Sample, population and sampling distributions finished with estimation & confidence intervals
Past weeks: Measures of central tendency (mean, mode, median) Measures of dispersion (standard deviation, variance, range, etc). Working with the normal curve Last week: Sample, population and sampling
More informationAnalysis of Variance
Analysis of Variance Blood coagulation time T avg A 62 60 63 59 61 B 63 67 71 64 65 66 66 C 68 66 71 67 68 68 68 D 56 62 60 61 63 64 63 59 61 64 Blood coagulation time A B C D Combined 56 57 58 59 60 61
More information+ Specify 1 tail / 2 tail
Week 2: Null hypothesis Aeroplane seat designer wonders how wide to make the plane seats. He assumes population average hip size μ = 43.2cm Sample size n = 50 Question : Is the assumption μ = 43.2cm reasonable?
More information10/4/2013. Hypothesis Testing & z-test. Hypothesis Testing. Hypothesis Testing
& z-test Lecture Set 11 We have a coin and are trying to determine if it is biased or unbiased What should we assume? Why? Flip coin n = 100 times E(Heads) = 50 Why? Assume we count 53 Heads... What could
More informationStatistical Analysis for QBIC Genetics Adapted by Ellen G. Dow 2017
Statistical Analysis for QBIC Genetics Adapted by Ellen G. Dow 2017 I. χ 2 or chi-square test Objectives: Compare how close an experimentally derived value agrees with an expected value. One method to
More informationThe Welch-Satterthwaite approximation
The Welch-Satterthwaite approximation Aidan Rocke Charles University of Prague aidanrocke@gmail.com October 19, 2018 Aidan Rocke (CU) Welch-Satterthwaite October 19, 2018 1 / 25 Overview 1 Motivation 2
More informationBasics of Experimental Design. Review of Statistics. Basic Study. Experimental Design. When an Experiment is Not Possible. Studying Relations
Basics of Experimental Design Review of Statistics And Experimental Design Scientists study relation between variables In the context of experiments these variables are called independent and dependent
More informationAP Statistics Cumulative AP Exam Study Guide
AP Statistics Cumulative AP Eam Study Guide Chapters & 3 - Graphs Statistics the science of collecting, analyzing, and drawing conclusions from data. Descriptive methods of organizing and summarizing statistics
More informationOne-Sample and Two-Sample Means Tests
One-Sample and Two-Sample Means Tests 1 Sample t Test The 1 sample t test allows us to determine whether the mean of a sample data set is different than a known value. Used when the population variance
More informationSTAT 135 Lab 5 Bootstrapping and Hypothesis Testing
STAT 135 Lab 5 Bootstrapping and Hypothesis Testing Rebecca Barter March 2, 2015 The Bootstrap Bootstrap Suppose that we are interested in estimating a parameter θ from some population with members x 1,...,
More informationThe t-distribution. Patrick Breheny. October 13. z tests The χ 2 -distribution The t-distribution Summary
Patrick Breheny October 13 Patrick Breheny Biostatistical Methods I (BIOS 5710) 1/25 Introduction Introduction What s wrong with z-tests? So far we ve (thoroughly!) discussed how to carry out hypothesis
More informationTwo sample Hypothesis tests in R.
Example. (Dependent samples) Two sample Hypothesis tests in R. A Calculus professor gives their students a 10 question algebra pretest on the first day of class, and a similar test towards the end of the
More informationAMS7: WEEK 7. CLASS 1. More on Hypothesis Testing Monday May 11th, 2015
AMS7: WEEK 7. CLASS 1 More on Hypothesis Testing Monday May 11th, 2015 Testing a Claim about a Standard Deviation or a Variance We want to test claims about or 2 Example: Newborn babies from mothers taking
More informationTwo Sample Problems. Two sample problems
Two Sample Problems Two sample problems The goal of inference is to compare the responses in two groups. Each group is a sample from a different population. The responses in each group are independent
More informationSampling Distributions: Central Limit Theorem
Review for Exam 2 Sampling Distributions: Central Limit Theorem Conceptually, we can break up the theorem into three parts: 1. The mean (µ M ) of a population of sample means (M) is equal to the mean (µ)
More informationData analysis and Geostatistics - lecture VI
Data analysis and Geostatistics - lecture VI Statistical testing with population distributions Statistical testing - the steps 1. Define a hypothesis to test in statistics only a hypothesis rejection is
More informationDifference in two or more average scores in different groups
ANOVAs Analysis of Variance (ANOVA) Difference in two or more average scores in different groups Each participant tested once Same outcome tested in each group Simplest is one-way ANOVA (one variable as
More informationOHSU OGI Class ECE-580-DOE :Statistical Process Control and Design of Experiments Steve Brainerd Basic Statistics Sample size?
ECE-580-DOE :Statistical Process Control and Design of Experiments Steve Basic Statistics Sample size? Sample size determination: text section 2-4-2 Page 41 section 3-7 Page 107 Website::http://www.stat.uiowa.edu/~rlenth/Power/
More information22s:152 Applied Linear Regression. Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA)
22s:152 Applied Linear Regression Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA) We now consider an analysis with only categorical predictors (i.e. all predictors are
More informationSampling Distributions
Sampling Distributions Sampling Distribution of the Mean & Hypothesis Testing Remember sampling? Sampling Part 1 of definition Selecting a subset of the population to create a sample Generally random sampling
More informationGov Univariate Inference II: Interval Estimation and Testing
Gov 2000-5. Univariate Inference II: Interval Estimation and Testing Matthew Blackwell October 13, 2015 1 / 68 Large Sample Confidence Intervals Confidence Intervals Example Hypothesis Tests Hypothesis
More informationLECTURE 5. Introduction to Econometrics. Hypothesis testing
LECTURE 5 Introduction to Econometrics Hypothesis testing October 18, 2016 1 / 26 ON TODAY S LECTURE We are going to discuss how hypotheses about coefficients can be tested in regression models We will
More informationExam 2 (KEY) July 20, 2009
STAT 2300 Business Statistics/Summer 2009, Section 002 Exam 2 (KEY) July 20, 2009 Name: USU A#: Score: /225 Directions: This exam consists of six (6) questions, assessing material learned within Modules
More informationpsyc3010 lecture 2 factorial between-ps ANOVA I: omnibus tests
psyc3010 lecture 2 factorial between-ps ANOVA I: omnibus tests last lecture: introduction to factorial designs next lecture: factorial between-ps ANOVA II: (effect sizes and follow-up tests) 1 general
More informationCENTRAL LIMIT THEOREM (CLT)
CENTRAL LIMIT THEOREM (CLT) A sampling distribution is the probability distribution of the sample statistic that is formed when samples of size n are repeatedly taken from a population. If the sample statistic
More informationCOMPLETELY RANDOM DESIGN (CRD) -Design can be used when experimental units are essentially homogeneous.
COMPLETELY RANDOM DESIGN (CRD) Description of the Design -Simplest design to use. -Design can be used when experimental units are essentially homogeneous. -Because of the homogeneity requirement, it may
More informationCh 2: Simple Linear Regression
Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component
More informationHuman-Centered Fidelity Metrics for Virtual Environment Simulations
Human-Centered Fidelity Metrics for Virtual Environment Simulations Three Numbers from Standard Experimental Design and Analysis: α, power, effect magnitude VR 2005 Tutorial J. Edward Swan II, Mississippi
More informationChapter 7 Comparison of two independent samples
Chapter 7 Comparison of two independent samples 7.1 Introduction Population 1 µ σ 1 1 N 1 Sample 1 y s 1 1 n 1 Population µ σ N Sample y s n 1, : population means 1, : population standard deviations N
More informationConfidence Interval Estimation
Department of Psychology and Human Development Vanderbilt University 1 Introduction 2 3 4 5 Relationship to the 2-Tailed Hypothesis Test Relationship to the 1-Tailed Hypothesis Test 6 7 Introduction In
More informationt-test for 2 matched/related/ dependent samples
HUMBEHV 3HB3 two-sample t-tests & statistical power Week 9 Prof. Patrick Bennett Concepts from previous lectures t distribution standard error of the mean degrees-of-freedom Null and alternative/research
More informationData analysis and Geostatistics - lecture VII
Data analysis and Geostatistics - lecture VII t-tests, ANOVA and goodness-of-fit Statistical testing - significance of r Testing the significance of the correlation coefficient: t = r n - 2 1 - r 2 with
More informationStat 135, Fall 2006 A. Adhikari HOMEWORK 6 SOLUTIONS
Stat 135, Fall 2006 A. Adhikari HOMEWORK 6 SOLUTIONS 1a. Under the null hypothesis X has the binomial (100,.5) distribution with E(X) = 50 and SE(X) = 5. So P ( X 50 > 10) is (approximately) two tails
More information4/6/16. Non-parametric Test. Overview. Stephen Opiyo. Distinguish Parametric and Nonparametric Test Procedures
Non-parametric Test Stephen Opiyo Overview Distinguish Parametric and Nonparametric Test Procedures Explain commonly used Nonparametric Test Procedures Perform Hypothesis Tests Using Nonparametric Procedures
More informationCan you tell the relationship between students SAT scores and their college grades?
Correlation One Challenge Can you tell the relationship between students SAT scores and their college grades? A: The higher SAT scores are, the better GPA may be. B: The higher SAT scores are, the lower
More informationStatistics for Managers Using Microsoft Excel Chapter 9 Two Sample Tests With Numerical Data
Statistics for Managers Using Microsoft Excel Chapter 9 Two Sample Tests With Numerical Data 999 Prentice-Hall, Inc. Chap. 9 - Chapter Topics Comparing Two Independent Samples: Z Test for the Difference
More information22s:152 Applied Linear Regression. 1-way ANOVA visual:
22s:152 Applied Linear Regression 1-way ANOVA visual: Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA) 0.00 0.05 0.10 0.15 0.20 0.25 0.30 0.35 Y We now consider an analysis
More informationReview of Statistics 101
Review of Statistics 101 We review some important themes from the course 1. Introduction Statistics- Set of methods for collecting/analyzing data (the art and science of learning from data). Provides methods
More informationSMAM 314 Exam 3 Name. F A. A null hypothesis that is rejected at α =.05 will always be rejected at α =.01.
SMAM 314 Exam 3 Name 1. Indicate whether the following statements are true (T) or false (F) (6 points) F A. A null hypothesis that is rejected at α =.05 will always be rejected at α =.01. T B. A course
More informationREED TUTORIALS (Pty) LTD ECS3706 EXAM PACK
REED TUTORIALS (Pty) LTD ECS3706 EXAM PACK 1 ECONOMETRICS STUDY PACK MAY/JUNE 2016 Question 1 (a) (i) Describing economic reality (ii) Testing hypothesis about economic theory (iii) Forecasting future
More informationMaking Inferences About Parameters
Making Inferences About Parameters Parametric statistical inference may take the form of: 1. Estimation: on the basis of sample data we estimate the value of some parameter of the population from which
More informationStatistics: revision
NST 1B Experimental Psychology Statistics practical 5 Statistics: revision Rudolf Cardinal & Mike Aitken 29 / 30 April 2004 Department of Experimental Psychology University of Cambridge Handouts: Answers
More informationPreview from Notesale.co.uk Page 3 of 63
Stem-and-leaf diagram - vertical numbers on far left represent the 10s, numbers right of the line represent the 1s The mean should not be used if there are extreme scores, or for ranks and categories Unbiased
More informationRelating Graph to Matlab
There are two related course documents on the web Probability and Statistics Review -should be read by people without statistics background and it is helpful as a review for those with prior statistics
More informationParameter Estimation, Sampling Distributions & Hypothesis Testing
Parameter Estimation, Sampling Distributions & Hypothesis Testing Parameter Estimation & Hypothesis Testing In doing research, we are usually interested in some feature of a population distribution (which
More information2 and F Distributions. Barrow, Statistics for Economics, Accounting and Business Studies, 4 th edition Pearson Education Limited 2006
and F Distributions Lecture 9 Distribution The distribution is used to: construct confidence intervals for a variance compare a set of actual frequencies with expected frequencies test for association
More informationStudio 8: NHST: t-tests and Rejection Regions Spring 2014
Studio 8: NHST: t-tests and Rejection Regions 18.05 Spring 2014 You should have downloaded studio8.zip and unzipped it into your 18.05 working directory. January 2, 2017 2 / 10 Left-side vs Right-side
More informationHypothesis Testing. Mean (SDM)
Confidence Intervals and Hypothesis Testing Readings: Howell, Ch. 4, 7 The Sampling Distribution of the Mean (SDM) Derivation - See Thorne & Giesen (T&G), pp. 169-171 or online Chapter Overview for Ch.
More informationHYPOTHESIS TESTING II TESTS ON MEANS. Sorana D. Bolboacă
HYPOTHESIS TESTING II TESTS ON MEANS Sorana D. Bolboacă OBJECTIVES Significance value vs p value Parametric vs non parametric tests Tests on means: 1 Dec 14 2 SIGNIFICANCE LEVEL VS. p VALUE Materials and
More informationChap The McGraw-Hill Companies, Inc. All rights reserved.
11 pter11 Chap Analysis of Variance Overview of ANOVA Multiple Comparisons Tests for Homogeneity of Variances Two-Factor ANOVA Without Replication General Linear Model Experimental Design: An Overview
More informationThis gives us an upper and lower bound that capture our population mean.
Confidence Intervals Critical Values Practice Problems 1 Estimation 1.1 Confidence Intervals Definition 1.1 Margin of error. The margin of error of a distribution is the amount of error we predict when
More informationN J SS W /df W N - 1
One-Way ANOVA Source Table ANOVA MODEL: ij = µ* + α j + ε ij H 0 : µ = µ =... = µ j or H 0 : Σα j = 0 Source Sum of Squares df Mean Squares F J Between Groups nj( j * ) J - SS B /(J ) MS B /MS W = ( N
More informationPower Analysis. Ben Kite KU CRMDA 2015 Summer Methodology Institute
Power Analysis Ben Kite KU CRMDA 2015 Summer Methodology Institute Created by Terrence D. Jorgensen, 2014 Recall Hypothesis Testing? Null Hypothesis Significance Testing (NHST) is the most common application
More information9/2/2010. Wildlife Management is a very quantitative field of study. throughout this course and throughout your career.
Introduction to Data and Analysis Wildlife Management is a very quantitative field of study Results from studies will be used throughout this course and throughout your career. Sampling design influences
More informationQuestions 3.83, 6.11, 6.12, 6.17, 6.25, 6.29, 6.33, 6.35, 6.50, 6.51, 6.53, 6.55, 6.59, 6.60, 6.65, 6.69, 6.70, 6.77, 6.79, 6.89, 6.
Chapter 7 Reading 7.1, 7.2 Questions 3.83, 6.11, 6.12, 6.17, 6.25, 6.29, 6.33, 6.35, 6.50, 6.51, 6.53, 6.55, 6.59, 6.60, 6.65, 6.69, 6.70, 6.77, 6.79, 6.89, 6.112 Introduction In Chapter 5 and 6, we emphasized
More informationInferential statistics
Inferential statistics Inference involves making a Generalization about a larger group of individuals on the basis of a subset or sample. Ahmed-Refat-ZU Null and alternative hypotheses In hypotheses testing,
More informationProbability and Statistics
Probability and Statistics Kristel Van Steen, PhD 2 Montefiore Institute - Systems and Modeling GIGA - Bioinformatics ULg kristel.vansteen@ulg.ac.be CHAPTER 4: IT IS ALL ABOUT DATA 4a - 1 CHAPTER 4: IT
More informationPower. January 12, 2019
Power January 12, 2019 Contents Definition of power Z-test example If H 0 is true If H 0 is false The 2x2 matrix of All Things That Can Happen Pr(Type I error) = α Pr(Type II error) = β power = 1 β Things
More informationPSY 216. Assignment 9 Answers. Under what circumstances is a t statistic used instead of a z-score for a hypothesis test
PSY 216 Assignment 9 Answers 1. Problem 1 from the text Under what circumstances is a t statistic used instead of a z-score for a hypothesis test The t statistic should be used when the population standard
More informationDistribution-Free Procedures (Devore Chapter Fifteen)
Distribution-Free Procedures (Devore Chapter Fifteen) MATH-5-01: Probability and Statistics II Spring 018 Contents 1 Nonparametric Hypothesis Tests 1 1.1 The Wilcoxon Rank Sum Test........... 1 1. Normal
More informationCHAPTER 9, 10. Similar to a courtroom trial. In trying a person for a crime, the jury needs to decide between one of two possibilities:
CHAPTER 9, 10 Hypothesis Testing Similar to a courtroom trial. In trying a person for a crime, the jury needs to decide between one of two possibilities: The person is guilty. The person is innocent. To
More informationSampling Distributions
Sampling Error As you may remember from the first lecture, samples provide incomplete information about the population In particular, a statistic (e.g., M, s) computed on any particular sample drawn from
More information