Statistical Inference Part 2. t test for single mean. t tests useful for PSYCH 710. Review of Statistical Inference (Part 2) Week 2

Size: px
Start display at page:

Download "Statistical Inference Part 2. t test for single mean. t tests useful for PSYCH 710. Review of Statistical Inference (Part 2) Week 2"

Transcription

1 Statistical Inference Part PSYCH 70 eview Statistical Inference (Part ) Week Pr. Patrick ennett t-tests Effect Size Equivalence Tests Consequences Low Power s P t tests useful for t test for single mean t statistic (Ȳ µ) t = (s/ p n) - comparing mean sample to some expected value - comparing samples statistic t - sampling distribution is t distribution - unimodal, symmetric about mean - shape governed by parameter degrees--freedom (df) - when NULL hyposis is true, mean t will be zero code set.seed(55540) mu <- 0 # population mean n <- 0 # sample size stdev <- # population sd < t.val <- rep(0,) p.val <- rep(0,) for(kk in :){.sample <- rnorm(n,mu,stdev) t.results <- t.test(.sample) t.val[kk] <- t.results$statistic p.val[kk] <- t.results$p.value } null hyposis is true null hyposis is true t.test evaluates null hyposis that mu=0 p value t value

2 is true. If null hyposis is true, n z will be between ±.96 95% time and between ±.56 99% time. Therefore, using criteria ±.96 to reject null hyposis will yield a Type I error rate 0.05, whereas criteria ±.56 corresponds to a Type I error rate 0.0. Our observed z score falls outside both sets criteria, and so null hyposis is rejected. It used to be standard practice to indicate which level was used by writing that null hyposis was rejected (p <.05) or (p <.0). Nowadays, scientists are encouraged to publish exact p value for ir observed statistic. In our case, probability drawing a z that was outside range ±3.3 is , and so we would report statistical result by writing sample mean di ered significantly from 00 (z = 3.3, p = ) and so null hyposis µ = 00 was rejected. 0.3 ennett, PJ t tests ennett, PJ Hyposis Testing PSYCH 70 In previous example, we knew that must be estimated from data: di erence between two sample PSYCH 70 Hyposis Testing two independent samples = 0. ut in most cases we we do not know, and refore it sp Y ) i (Yi ˆ=s= (n ) The estimate, s can be used to calculate t, which is similar to a z score: t testsor single mean t statistic Comparing far we have considered how to decide wher a single sample mean was drawn from a population with (Y µ) di erence between two sample p = p =.36 two independent samples a mean µ. More commonly, we want to compare p t =to each (6) or to decide wher or (/ n) (/ two 0) sample (s/ n) populations scores Given So far we have considered how to decide wher single sample mean was drawn from a population not ya di er. If population with two samples are µ and µ, n two-sided null and a mean µ. More commonly, we want to compare two sample to each or to decide wher or - : µa & µb alternative hyposes are: not y di er. If population code two samples are µ and µ, n two-sided null and 4 null hyposis is false mu=0.5 null hyposis is false mu=0.5 - variances: σa & σb alternative hyposes are: set.seed(55540) H0 : µ = µ mu <- 0.5 # population Distributions sample : H0 : µmean = µ H : µ 6= µ n <- 0 # sample size Hennett, : µ 6= PJ µ - : µ a & µb PSYCH 70 Hyposis Testing stdev <- # population sd < a - variances: These hyposes can be re-written as:these hyposes can be re-written as: = Y a t.val <- rep(0,) n b n Y b = 0. 6 di erence between two sample will be normal: p.val <- rep(0,) H0 : µ µ = 0 for(kk in :){ H0 : µ µ = 0 - shape: normal (via Central Limit Theorem) : µ µ 6= 0.sample <-H rnorm(n,mu,stdev) 0.7. two independent samples H : µ µ 6= 0 t.results <- t.test(.sample) Distribution difference between sample Each sample mean is a random variable drawn from sampling distribution mean. The t.val[kk] <- t.results$statistic /n and /n. (We will sampling distributions will be µ<µso and we have variances will be mean: u d = µ a & µb and, far to decide wher a single sample mean was drawn from a population with p.val[kk] t.results$p.value Each sample mean isconsidered adistributions randomhow variable from sampling distribution mean. The- take advantage Central Limit Theorem and assume are normal.)drawn } a mean that µ. More commonly, we wantnow, to let s compare two sample to each or to decide wher or 0 /n and /n. (We - variance: create a new variable that is di erence between sample, d = Y will Y. How sampling distributions be isµd distributed? and µ, and variances will be will di er. If variables population two samples µ 0. a Normal variable, are µ 4 and 6 8,0n two-sided null and It can be shown that sum evaluates (or di erence) not twoy or more Normal is also t.test null hyposis that mu=0 take advantage Central 0.0 Limit Theorem and.0assume that distributions are normal.) Now, let s so d is distributed Normally. Furrmore, alternative mean a sum (or di erence) two variables hyposes are: p valueis sum createa mean new variable thatdistribution is di erence between sample, dt value = Y Y. How is d distributed? (or di erence) two. Therefore, di erence equals di erence between two samplingitdistributions : µ. If two can be shown that Y sum (or di erence) are two variables is also a Normal variable, d = µ Y Y H0 : or µmore = Normal µ equal, n Y d = 0. Finally, variance (or di erence) two independent random variables is so d aissum distributed Normally. Furrmore, mean sumµ (or di erence) two variables is sum H : µ a 6= sum two variances: d = Y + Y. (N.. The variance a di erence between independent (or di erence) two. Therefore, mean di erence distribution equals di erence variables also is sum, not di erence, two variances.) Therefore, if we know form between sampling distributions : Y d = µy µy. If two are populations and, n statistical ory allows us to predict form distribution between These hyposes cantwo be re-written as: samples drawn from those populations (see Figure equal, n Y d =5).0. Finally, variance a sum (or di erence) two independent random variables is or, and refore In general, we will not know eir will have to from : µ be estimated H0 µ = 0 Y sum two Y variances: d = Y + Y. (N.. The variance a di erence between independent data with formula (n )s + (n is )s H : µ µ 6= 0 Therefore, if we know form sum, variables also not di erence, two variances.) ˆY Y = + (7) n + n n n d = Y a Y b + populations and, n statistical ory allows us to predict form distribution between variable drawn from The samples drawn from those populations (see Figure 5).sampling distribution mean. /n. (We will Y sampling distributions will be µ and µ, and variances will be /n and or (Y general, (µwe µ ) ) not know eir In will, and refore will have to be estimated from -sample t-test (8) and assume that Simulation t= Y Y distributions are normal.) takey advantage Central Limit Theorem Now, let s Y data withˆ formula + (n sample, d = Y create a new variable that is di erence between Y. How is d distributed? (n )s )s underlying similarity: with df = n + n. close comparison Equations 6 and 8 will revealˆan = di erence)in two or more + variables is also a Normal variable, (7) It can be shown that parameter, sum Normal -sample t-test H0 is true Y (or both cases, t equals di erence between an observed statistic and a populationy divided n by + n n n so d is distributed Normally. Furrmore, mean a sum (or di erence) two variables is sum standard error statistic: sample pop where (or n and n are sizes Therefore, two groups. The t statistic is sample two. mean two-group di erence distribution equals di erence statisticdi erence) parameter t= µy. If two are estimated between standard error statistic two sampling distributions : Y d = µy code (Y ay ) (or (µdi erence) µ ) equal, n Y d = 0. Finally, variance sum two independent random variables t= (8) is This general formula applies to all t tests. set.seed(65540) + ˆ.Y (N.. Y The variance a di erence between sum variance, two variances: we estimate population so parameter In, a two-sample t testnormally, is done with t.test() command. Notice that mu <- independent 00 # population mean d = var.equal Y Y ourtells test will a tis distribution mu <- 00 # population mean is set to TUE. Using this setting statistic to assume thatfollow samples and have same variables also sum, not variance. di erence, two variances.) Therefore, if we know form standard an errorunderlying n <5 # samplein size = degrees n + n. close comparison Equations 6 and 8estimated will reveal similarity: withwith n + df n - freedom difference form distribution > sample. <- c(95, 9, 93, 96, 98, 99) populations and, n statistical ory allows us to predict stdev between <- 0 # population sd both t equals di erence between an observed statistic and a population parameter,stdev divided by# > sample. <- c(0, 99, 06, 00, 98, cases, ) <0 population sd samples drawn from those populations (see Figure 5). standard error statistic: > t.test(x=sample.,y=sample.,alternative="two.sided", var.equal=tue) < In general, we will not know eir or, and refore Y Y will have to be estimated from t.val <- rep(0,) bove statement t assumes data re: with formula statistic parameter p.val <- rep(0,) t = that are distributed (n )s + (n )s for(kk in :){ 0 standard error statistic + ˆY estimated (7) normally with equal variance Y =.sample. <- rnorm(n,mu,stdev) n + n n n p value.sample. <- rnorm(n,mu,stdev) This general formula applies to all t tests. t.results <- t.test(.sample.,.sample.,var.equal=t) where n and n are sample sizes two groups. The two-group t statistic is t.val[kk] <- t.results$statistic In, a two-sample t test is done with t.test() command. Notice that parameter var.equal p.val[kk] <- t.results$p.value is set to TUE. Using this setting tells to assume and (Y thaty samples ) (µ µ ) have same variance. } t= (8) ˆY Y > sample. <- c(95, 9, 93, 96, 98, 99) > sample. 00, 98, ) with df<= c(0, n + n 99,. 06, close comparison Equations 6 and 8 will reveal an underlying similarity: in > t.test(x=sample.,y=sample.,alternative="two.sided", both cases, t equals di erence between an observed statistic var.equal=tue) and a population parameter, divided by standard error statistic: where n and n are sample sizes two groups. The two-group t statistic is Each sample mean is a random t test for comparing independent t= statistic parameter 0 estimated standard error statistic This general formula applies to all t tests sample t-test H0 is true -4-0 t value 4

3 Simulation -sample t-test (equal variances) -sample t-test H0 is false -sample t-test H0 is false Simulation -sample t-test (unequal variances) -samp t unequal Var H0 is true code set.seed(65540) mu <- 00 # population mean mu <- 05 # population mean n <- 5 # sample size stdev <- 0 # population sd stdev <- 0 # population sd < t.val <- rep(0,) p.val <- rep(0,) for(kk in :){ sample. <- rnorm(n,mu,stdev).sample. <- rnorm(n,mu,stdev) p value t value t.results <- t.test(.sample.,.sample.,var.equal=t) t.val[kk] <- t.results$statistic p.val[kk] <- t.results$p.value } stdev_diff <- sqrt(( ( (5-)*00 + (5-)*00 ) / (5+5-) ) * ( (/5)+(/5) ) ) =.88 Expected Value t = (00-05)/.88 = -.76 (actual mean = -.8) code set.seed(65540) mu <- 00 # population mean mu <- 00 # population mean n <- 5 # sample size stdev <- 0 # population sd stdev <- 3 # population sd < t.val <- rep(0,) p.val <- rep(0,) for(kk in :){.sample. <- rnorm(n,mu,stdev).sample. <- rnorm(n,mu,stdev) t.results <- t.test(.sample.,.sample.,var.equal=t) t.val[kk] <- t.results$statistic p.val[kk] <- t.results$p.value } t value Effect Size Cohen s ds what makes a p-value significant or non-significant? - alpha, sample size (power), effect size classes effect size - d - standardized differences (distances) btwn - r - measures association (variance accounted for) Used for between-subjects designs: d s = X X q (n )SD +(n )SD /(n + n ) d s = t r n + n t/ p N

4 > m<-0 > m<- > n = 0 > stdev <- 4 > s <- rnorm(n,m,stdev) > s <- rnorm(n,m,stdev) > (t.test. <- t.test(s,s,var.equal=t)) Two Sample t-test data: s and s t = , df = 38, p-value = H: true difference in is not equal to 0 95 percent confidence interval: sample estimates: mean x mean y > t <- t.test.$statistic > ( ds <- t*sqrt( (/n)+(/n) ) ) > m<-0 > m<- > n = 00 > stdev <- 4 > s <- rnorm(n,m,stdev) > s <- rnorm(n,m,stdev) > (t.test. <- t.test(s,s,var.equal=t)) Two Sample t-test data: s and s t = , df = 98, p-value =.764e-05 H: true difference in is not equal to 0 95 percent confidence interval: sample estimates: mean x mean y > t <- t.test.$statistic > ( ds <- t*sqrt( (/n)+(/n) ) ) Hedges gs unbiased estimate population ds g s = d s 3 4(n + n ) 9 Convert ds to r Cohen s dz r = d s p d s +(N N)/(n n ) used for within-subjects design (paired scores) - Md = mean difference scores - Di = difference score i - N = number difference scores d z = M d p ( P (Di M d )) /(N ) d z = t p N

5 eta-squared measure association between independent & dependent variables omega-squared unbiased estimate association in population = SS e ect SS total proportion variation in Y accounted for by group membership! = df e ect (MS e ect MS error ) SS total + MS error p = SS e ect SS e ect + SS error for factorial experiments (more than independent variable)! p = df e ect (MS e ect MS error ) df error MS e ect +(N df error ) MS error p = F df e ect F df e ect + df error Why report effect size? Why not just report p-values as index effect? - p-values depend on number subjects/events/measures - increased power leads to lower p values Would like to have measure doesn t depend on experiment N - or on particular aspects experimental design Effect size measures try to do this - and also give information about magnitude effect Sample size affects precision ds estimation simulation parameters: < # 5K samples m <- 0 # mean pop m <- # mean pop stdev <- 4 # sd population simulation : n <- 0 # sample size simulation : n <- 70 # sample size n=n= t (-sample test) Cohens ds n=n= t (-sample test) Cohens ds effect size measures vary across samples variance sampling distribution is larger for smaller samples

6 Effect sizes in published reports (winner s curse) effect size inflation (winner s curse) n=n=0 n=n=70 n = n = 0 apply statistical threshold (e.g., p<.05) to effect size - what is median significant ds? median = depends on sample size t (-sample test) t (-sample test) Significant Cohen ds (p<.05) - for small samples, only big ds are significant - large samples more likely to get significant large & small ds published studies usually report significant results n = n = 70 median = small n studies lead to overestimation true effect size Cohens ds Cohens ds Significant Cohen ds (p<.05) effect size measures vary across samples variance sampling distribution is larger for smaller samples significant ds are more extreme with small samples What does a significant p-value mean? probability replicating p=.05 effect a ejecting null hyposis Null Observed effect p = sample t-test H0 is false significant p-value indicates that result is unusual - assuming null hyposis is true and assumptions are correct That is LL it - it is not equal to probability that H0 is true - it is not equal to probability replicating result.96 x sem observed effect significant p= x sem b The sampling distribution critical value t(8)= x sem expect 50% replications to yield smaller effect t value n=9, mu=.86*sem

7 What does a p-value mean? non-significant p-value indicates that result is NOT unusual assuming null hyposis is true. It does NOT mean that null hyposis is true - e.g., a non-significant -sample t-test does NOT mean that populations are equal - You do not accept null hyposis, you simply fail to reject it. To accept null hyposis, perform an equivalence test Equivalence Tests reverse H0 & H: - H0: re is an effect - H: re is no effect try to reject H0 in favour H define Smallest Effect Sizes Interest (SESOI) - test upper & lower bounds (SESOI) with -tailed t-tests - H0-u: observed effect unusually smaller than upper SESOI? - H0-l: observed effect unusually larger than lower SESOI? if both -tailed tests are significant, n we say observed effect is smaller than SESOI - e.g., two groups are equivalent Equivalence Tests reverse H0 & H: - H0: re is an effect; H: re is no effect - try to reject H0 in favour H define Smallest Effect Sizes Interest (SESOI; L & U) - evaluate hyposis with two -tailed t-tests: H0-u: is observed effect unusually smaller than U? H0-l: observed effect unusually larger than L? if both -tailed tests are significant, n observed effect is smaller than SESOI - e.g., two groups are equivalent Null Hyposis vs Equivalence Tests Four possible outcomes when evaluating difference between groups Null Hyposis Significance Test Equivalence Test Equivalent to Zero Not Equivalent to Zero Not Different from Zero + - Different from Zero - +

8 NHST vs Equivalence Tests (4 outcomes) Spurious findings? equivalence region difference between L 0 U Statistically Equivalent & Different Statistically Equivalent & Not Different Not Equivalent & Different Not Equivalent & Not Different Over time: - re has been a dramatic increase in ease with which we can analyze a single data set in multiple ways - also, effects/phenomena interest typically become increasingly smaller & more subtle as a research area matures - but, sample sizes tend not to change very much - nor have unwritten rules about deciding if a paper is publishable se trends make it likely that number papers reporting spurious statistically significant findings has increased Lakens, D. Equivalence tests: practical primer for t tests, correlations, and metaanalyses. Social Psychological & Personality Science, 8(4), Power (Type I & Type II Errors) Positive Predictive Value () Table : Possible outcomes hyposis testing. decision H0 is True H0 is False reject H0: Type I (p = ) Correct (p = =power) do not reject H0: Correct (p = ) Type II error (p = ) Given that we ve rejected HO in favour H - what is probability that our conclusion is correct? decision H0 is True H0 is False reject H0: p Type I (p = ) Correct (p = =power) do not reject H0: Correct (p = ) Type II error (p = ) p Type I Error: reject H0 when it is true (alpha) Type II Error: fail to reject H0 when it is false (beta) = p / (p + p) Power = Probability rejecting false H0 (-beta)

9 Ioannidis 005 depends on a priori probability that H0 is true Ioannidis 005 Table. esearch Findings and True elationships esearch Finding True elationship Yes No Total Yes c( β)/( + ) cα/( + ) c( + α β)/( + ) No cβ/( + ) c( α)/( + ) c( α + β)/( + ) Total c/( + ) c/( + ) c DOI: 0.37/journal.pmed.0004.t00 alpha is Type error rate (-eta) is power ( minus Type II error rate) c is number effects begin examined is ratio number true effects divided number no effects Positive Predictive Value () is probability that a statistically significant finding is true = (-eta) / ( - *eta + alpha) reduced by low power - small studies & small effects reduced by low - testing more false hyposes - increasing bias greater flexibility in design & analysis selective reporting positive results N.. higher beta = lower power beta = 0. alpha = 0.05 Statistically significant finding >50% chance being true/correct if: (-eta) > alpha Ioannidis 005 effect alpha on Ioannidis 005 effect bias on bias: proportion null results findings that are reported as true effects beta = 0. alpha = 0.05 beta = 0. alpha = 0.0 beta = 0. bias = 0 alpha = 0.0 beta = 0. bias = 0. alpha = 0.0

10 Ioannidis 005 effect bias on bias: proportion null results findings that are reported as true effects Factors that lower beta = 0. bias = 0 alpha = 0.05 beta = 0. bias = 0. alpha = 0.05 small studies & small effects (low power) testing more hyposes (lowering ) greater flexibility in design and analysis (increasing bias) more independent research teams studying phenomenon - due to selective reporting positive results

Population Variance. Concepts from previous lectures. HUMBEHV 3HB3 one-sample t-tests. Week 8

Population Variance. Concepts from previous lectures. HUMBEHV 3HB3 one-sample t-tests. Week 8 Concepts from previous lectures HUMBEHV 3HB3 one-sample t-tests Week 8 Prof. Patrick Bennett sampling distributions - sampling error - standard error of the mean - degrees-of-freedom Null and alternative/research

More information

PSY 305. Module 3. Page Title. Introduction to Hypothesis Testing Z-tests. Five steps in hypothesis testing

PSY 305. Module 3. Page Title. Introduction to Hypothesis Testing Z-tests. Five steps in hypothesis testing Page Title PSY 305 Module 3 Introduction to Hypothesis Testing Z-tests Five steps in hypothesis testing State the research and null hypothesis Determine characteristics of comparison distribution Five

More information

An inferential procedure to use sample data to understand a population Procedures

An inferential procedure to use sample data to understand a population Procedures Hypothesis Test An inferential procedure to use sample data to understand a population Procedures Hypotheses, the alpha value, the critical region (z-scores), statistics, conclusion Two types of errors

More information

HYPOTHESIS TESTING. Hypothesis Testing

HYPOTHESIS TESTING. Hypothesis Testing MBA 605 Business Analytics Don Conant, PhD. HYPOTHESIS TESTING Hypothesis testing involves making inferences about the nature of the population on the basis of observations of a sample drawn from the population.

More information

Business Statistics. Lecture 10: Course Review

Business Statistics. Lecture 10: Course Review Business Statistics Lecture 10: Course Review 1 Descriptive Statistics for Continuous Data Numerical Summaries Location: mean, median Spread or variability: variance, standard deviation, range, percentiles,

More information

Statistics Primer. ORC Staff: Jayme Palka Peter Boedeker Marcus Fagan Trey Dejong

Statistics Primer. ORC Staff: Jayme Palka Peter Boedeker Marcus Fagan Trey Dejong Statistics Primer ORC Staff: Jayme Palka Peter Boedeker Marcus Fagan Trey Dejong 1 Quick Overview of Statistics 2 Descriptive vs. Inferential Statistics Descriptive Statistics: summarize and describe data

More information

Advanced Experimental Design

Advanced Experimental Design Advanced Experimental Design Topic Four Hypothesis testing (z and t tests) & Power Agenda Hypothesis testing Sampling distributions/central limit theorem z test (σ known) One sample z & Confidence intervals

More information

CIVL /8904 T R A F F I C F L O W T H E O R Y L E C T U R E - 8

CIVL /8904 T R A F F I C F L O W T H E O R Y L E C T U R E - 8 CIVL - 7904/8904 T R A F F I C F L O W T H E O R Y L E C T U R E - 8 Chi-square Test How to determine the interval from a continuous distribution I = Range 1 + 3.322(logN) I-> Range of the class interval

More information

The t-test: A z-score for a sample mean tells us where in the distribution the particular mean lies

The t-test: A z-score for a sample mean tells us where in the distribution the particular mean lies The t-test: So Far: Sampling distribution benefit is that even if the original population is not normal, a sampling distribution based on this population will be normal (for sample size > 30). Benefit

More information

Sampling distribution of t. 2. Sampling distribution of t. 3. Example: Gas mileage investigation. II. Inferential Statistics (8) t =

Sampling distribution of t. 2. Sampling distribution of t. 3. Example: Gas mileage investigation. II. Inferential Statistics (8) t = 2. The distribution of t values that would be obtained if a value of t were calculated for each sample mean for all possible random of a given size from a population _ t ratio: (X - µ hyp ) t s x The result

More information

Statistical inference (estimation, hypothesis tests, confidence intervals) Oct 2018

Statistical inference (estimation, hypothesis tests, confidence intervals) Oct 2018 Statistical inference (estimation, hypothesis tests, confidence intervals) Oct 2018 Sampling A trait is measured on each member of a population. f(y) = propn of individuals in the popn with measurement

More information

Introduction to Statistical Data Analysis III

Introduction to Statistical Data Analysis III Introduction to Statistical Data Analysis III JULY 2011 Afsaneh Yazdani Preface Major branches of Statistics: - Descriptive Statistics - Inferential Statistics Preface What is Inferential Statistics? The

More information

T.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS

T.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS In our work on hypothesis testing, we used the value of a sample statistic to challenge an accepted value of a population parameter. We focused only

More information

T-tests for 2 Independent Means

T-tests for 2 Independent Means T-tests for 2 Independent Means January 15, 2019 Contents t-test for 2 independent means Tutorial Example 1: one-tailed test for independent means, equal sample sizes Error Bars Example 2: two-tailed test

More information

Chapter 23: Inferences About Means

Chapter 23: Inferences About Means Chapter 3: Inferences About Means Sample of Means: number of observations in one sample the population mean (theoretical mean) sample mean (observed mean) is the theoretical standard deviation of the population

More information

Outline. PubH 5450 Biostatistics I Prof. Carlin. Confidence Interval for the Mean. Part I. Reviews

Outline. PubH 5450 Biostatistics I Prof. Carlin. Confidence Interval for the Mean. Part I. Reviews Outline Outline PubH 5450 Biostatistics I Prof. Carlin Lecture 11 Confidence Interval for the Mean Known σ (population standard deviation): Part I Reviews σ x ± z 1 α/2 n Small n, normal population. Large

More information

T-tests for 2 Independent Means

T-tests for 2 Independent Means T-tests for 2 Independent Means February 22, 208 Contents t-test for 2 independent means Tutorial Example : one-tailed test for independent means, equal sample sizes Error Bars Example 2: two-tailed test

More information

" M A #M B. Standard deviation of the population (Greek lowercase letter sigma) σ 2

 M A #M B. Standard deviation of the population (Greek lowercase letter sigma) σ 2 Notation and Equations for Final Exam Symbol Definition X The variable we measure in a scientific study n The size of the sample N The size of the population M The mean of the sample µ The mean of the

More information

PSY 307 Statistics for the Behavioral Sciences. Chapter 20 Tests for Ranked Data, Choosing Statistical Tests

PSY 307 Statistics for the Behavioral Sciences. Chapter 20 Tests for Ranked Data, Choosing Statistical Tests PSY 307 Statistics for the Behavioral Sciences Chapter 20 Tests for Ranked Data, Choosing Statistical Tests What To Do with Non-normal Distributions Tranformations (pg 382): The shape of the distribution

More information

Single Sample Means. SOCY601 Alan Neustadtl

Single Sample Means. SOCY601 Alan Neustadtl Single Sample Means SOCY601 Alan Neustadtl The Central Limit Theorem If we have a population measured by a variable with a mean µ and a standard deviation σ, and if all possible random samples of size

More information

DETAILED CONTENTS PART I INTRODUCTION AND DESCRIPTIVE STATISTICS. 1. Introduction to Statistics

DETAILED CONTENTS PART I INTRODUCTION AND DESCRIPTIVE STATISTICS. 1. Introduction to Statistics DETAILED CONTENTS About the Author Preface to the Instructor To the Student How to Use SPSS With This Book PART I INTRODUCTION AND DESCRIPTIVE STATISTICS 1. Introduction to Statistics 1.1 Descriptive and

More information

INTERVAL ESTIMATION AND HYPOTHESES TESTING

INTERVAL ESTIMATION AND HYPOTHESES TESTING INTERVAL ESTIMATION AND HYPOTHESES TESTING 1. IDEA An interval rather than a point estimate is often of interest. Confidence intervals are thus important in empirical work. To construct interval estimates,

More information

COMPARING SEVERAL MEANS: ANOVA

COMPARING SEVERAL MEANS: ANOVA LAST UPDATED: November 15, 2012 COMPARING SEVERAL MEANS: ANOVA Objectives 2 Basic principles of ANOVA Equations underlying one-way ANOVA Doing a one-way ANOVA in R Following up an ANOVA: Planned contrasts/comparisons

More information

The t-statistic. Student s t Test

The t-statistic. Student s t Test The t-statistic 1 Student s t Test When the population standard deviation is not known, you cannot use a z score hypothesis test Use Student s t test instead Student s t, or t test is, conceptually, very

More information

Chapter 23. Inferences About Means. Monday, May 6, 13. Copyright 2009 Pearson Education, Inc.

Chapter 23. Inferences About Means. Monday, May 6, 13. Copyright 2009 Pearson Education, Inc. Chapter 23 Inferences About Means Sampling Distributions of Means Now that we know how to create confidence intervals and test hypotheses about proportions, we do the same for means. Just as we did before,

More information

Two-Sample Inferential Statistics

Two-Sample Inferential Statistics The t Test for Two Independent Samples 1 Two-Sample Inferential Statistics In an experiment there are two or more conditions One condition is often called the control condition in which the treatment is

More information

Multiple Regression Analysis

Multiple Regression Analysis Multiple Regression Analysis y = β 0 + β 1 x 1 + β 2 x 2 +... β k x k + u 2. Inference 0 Assumptions of the Classical Linear Model (CLM)! So far, we know: 1. The mean and variance of the OLS estimators

More information

16.400/453J Human Factors Engineering. Design of Experiments II

16.400/453J Human Factors Engineering. Design of Experiments II J Human Factors Engineering Design of Experiments II Review Experiment Design and Descriptive Statistics Research question, independent and dependent variables, histograms, box plots, etc. Inferential

More information

Last two weeks: Sample, population and sampling distributions finished with estimation & confidence intervals

Last two weeks: Sample, population and sampling distributions finished with estimation & confidence intervals Past weeks: Measures of central tendency (mean, mode, median) Measures of dispersion (standard deviation, variance, range, etc). Working with the normal curve Last two weeks: Sample, population and sampling

More information

1/24/2008. Review of Statistical Inference. C.1 A Sample of Data. C.2 An Econometric Model. C.4 Estimating the Population Variance and Other Moments

1/24/2008. Review of Statistical Inference. C.1 A Sample of Data. C.2 An Econometric Model. C.4 Estimating the Population Variance and Other Moments /4/008 Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University C. A Sample of Data C. An Econometric Model C.3 Estimating the Mean of a Population C.4 Estimating the Population

More information

Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institution of Technology, Kharagpur

Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institution of Technology, Kharagpur Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institution of Technology, Kharagpur Lecture No. # 36 Sampling Distribution and Parameter Estimation

More information

Quantitative Methods for Economics, Finance and Management (A86050 F86050)

Quantitative Methods for Economics, Finance and Management (A86050 F86050) Quantitative Methods for Economics, Finance and Management (A86050 F86050) Matteo Manera matteo.manera@unimib.it Marzio Galeotti marzio.galeotti@unimi.it 1 This material is taken and adapted from Guy Judge

More information

Hypothesis Testing and Confidence Intervals (Part 2): Cohen s d, Logic of Testing, and Confidence Intervals

Hypothesis Testing and Confidence Intervals (Part 2): Cohen s d, Logic of Testing, and Confidence Intervals Hypothesis Testing and Confidence Intervals (Part 2): Cohen s d, Logic of Testing, and Confidence Intervals Lecture 9 Justin Kern April 9, 2018 Measuring Effect Size: Cohen s d Simply finding whether a

More information

Lab #12: Exam 3 Review Key

Lab #12: Exam 3 Review Key Psychological Statistics Practice Lab#1 Dr. M. Plonsky Page 1 of 7 Lab #1: Exam 3 Review Key 1) a. Probability - Refers to the likelihood that an event will occur. Ranges from 0 to 1. b. Sampling Distribution

More information

CHAPTER 7. Hypothesis Testing

CHAPTER 7. Hypothesis Testing CHAPTER 7 Hypothesis Testing A hypothesis is a statement about one or more populations, and usually deal with population parameters, such as means or standard deviations. A research hypothesis is a conjecture

More information

10/31/2012. One-Way ANOVA F-test

10/31/2012. One-Way ANOVA F-test PSY 511: Advanced Statistics for Psychological and Behavioral Research 1 1. Situation/hypotheses 2. Test statistic 3.Distribution 4. Assumptions One-Way ANOVA F-test One factor J>2 independent samples

More information

Simple Linear Regression

Simple Linear Regression Simple Linear Regression ST 430/514 Recall: A regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates)

More information

Variance. Standard deviation VAR = = value. Unbiased SD = SD = 10/23/2011. Functional Connectivity Correlation and Regression.

Variance. Standard deviation VAR = = value. Unbiased SD = SD = 10/23/2011. Functional Connectivity Correlation and Regression. 10/3/011 Functional Connectivity Correlation and Regression Variance VAR = Standard deviation Standard deviation SD = Unbiased SD = 1 10/3/011 Standard error Confidence interval SE = CI = = t value for

More information

Last week: Sample, population and sampling distributions finished with estimation & confidence intervals

Last week: Sample, population and sampling distributions finished with estimation & confidence intervals Past weeks: Measures of central tendency (mean, mode, median) Measures of dispersion (standard deviation, variance, range, etc). Working with the normal curve Last week: Sample, population and sampling

More information

Analysis of Variance

Analysis of Variance Analysis of Variance Blood coagulation time T avg A 62 60 63 59 61 B 63 67 71 64 65 66 66 C 68 66 71 67 68 68 68 D 56 62 60 61 63 64 63 59 61 64 Blood coagulation time A B C D Combined 56 57 58 59 60 61

More information

+ Specify 1 tail / 2 tail

+ Specify 1 tail / 2 tail Week 2: Null hypothesis Aeroplane seat designer wonders how wide to make the plane seats. He assumes population average hip size μ = 43.2cm Sample size n = 50 Question : Is the assumption μ = 43.2cm reasonable?

More information

10/4/2013. Hypothesis Testing & z-test. Hypothesis Testing. Hypothesis Testing

10/4/2013. Hypothesis Testing & z-test. Hypothesis Testing. Hypothesis Testing & z-test Lecture Set 11 We have a coin and are trying to determine if it is biased or unbiased What should we assume? Why? Flip coin n = 100 times E(Heads) = 50 Why? Assume we count 53 Heads... What could

More information

Statistical Analysis for QBIC Genetics Adapted by Ellen G. Dow 2017

Statistical Analysis for QBIC Genetics Adapted by Ellen G. Dow 2017 Statistical Analysis for QBIC Genetics Adapted by Ellen G. Dow 2017 I. χ 2 or chi-square test Objectives: Compare how close an experimentally derived value agrees with an expected value. One method to

More information

The Welch-Satterthwaite approximation

The Welch-Satterthwaite approximation The Welch-Satterthwaite approximation Aidan Rocke Charles University of Prague aidanrocke@gmail.com October 19, 2018 Aidan Rocke (CU) Welch-Satterthwaite October 19, 2018 1 / 25 Overview 1 Motivation 2

More information

Basics of Experimental Design. Review of Statistics. Basic Study. Experimental Design. When an Experiment is Not Possible. Studying Relations

Basics of Experimental Design. Review of Statistics. Basic Study. Experimental Design. When an Experiment is Not Possible. Studying Relations Basics of Experimental Design Review of Statistics And Experimental Design Scientists study relation between variables In the context of experiments these variables are called independent and dependent

More information

AP Statistics Cumulative AP Exam Study Guide

AP Statistics Cumulative AP Exam Study Guide AP Statistics Cumulative AP Eam Study Guide Chapters & 3 - Graphs Statistics the science of collecting, analyzing, and drawing conclusions from data. Descriptive methods of organizing and summarizing statistics

More information

One-Sample and Two-Sample Means Tests

One-Sample and Two-Sample Means Tests One-Sample and Two-Sample Means Tests 1 Sample t Test The 1 sample t test allows us to determine whether the mean of a sample data set is different than a known value. Used when the population variance

More information

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing STAT 135 Lab 5 Bootstrapping and Hypothesis Testing Rebecca Barter March 2, 2015 The Bootstrap Bootstrap Suppose that we are interested in estimating a parameter θ from some population with members x 1,...,

More information

The t-distribution. Patrick Breheny. October 13. z tests The χ 2 -distribution The t-distribution Summary

The t-distribution. Patrick Breheny. October 13. z tests The χ 2 -distribution The t-distribution Summary Patrick Breheny October 13 Patrick Breheny Biostatistical Methods I (BIOS 5710) 1/25 Introduction Introduction What s wrong with z-tests? So far we ve (thoroughly!) discussed how to carry out hypothesis

More information

Two sample Hypothesis tests in R.

Two sample Hypothesis tests in R. Example. (Dependent samples) Two sample Hypothesis tests in R. A Calculus professor gives their students a 10 question algebra pretest on the first day of class, and a similar test towards the end of the

More information

AMS7: WEEK 7. CLASS 1. More on Hypothesis Testing Monday May 11th, 2015

AMS7: WEEK 7. CLASS 1. More on Hypothesis Testing Monday May 11th, 2015 AMS7: WEEK 7. CLASS 1 More on Hypothesis Testing Monday May 11th, 2015 Testing a Claim about a Standard Deviation or a Variance We want to test claims about or 2 Example: Newborn babies from mothers taking

More information

Two Sample Problems. Two sample problems

Two Sample Problems. Two sample problems Two Sample Problems Two sample problems The goal of inference is to compare the responses in two groups. Each group is a sample from a different population. The responses in each group are independent

More information

Sampling Distributions: Central Limit Theorem

Sampling Distributions: Central Limit Theorem Review for Exam 2 Sampling Distributions: Central Limit Theorem Conceptually, we can break up the theorem into three parts: 1. The mean (µ M ) of a population of sample means (M) is equal to the mean (µ)

More information

Data analysis and Geostatistics - lecture VI

Data analysis and Geostatistics - lecture VI Data analysis and Geostatistics - lecture VI Statistical testing with population distributions Statistical testing - the steps 1. Define a hypothesis to test in statistics only a hypothesis rejection is

More information

Difference in two or more average scores in different groups

Difference in two or more average scores in different groups ANOVAs Analysis of Variance (ANOVA) Difference in two or more average scores in different groups Each participant tested once Same outcome tested in each group Simplest is one-way ANOVA (one variable as

More information

OHSU OGI Class ECE-580-DOE :Statistical Process Control and Design of Experiments Steve Brainerd Basic Statistics Sample size?

OHSU OGI Class ECE-580-DOE :Statistical Process Control and Design of Experiments Steve Brainerd Basic Statistics Sample size? ECE-580-DOE :Statistical Process Control and Design of Experiments Steve Basic Statistics Sample size? Sample size determination: text section 2-4-2 Page 41 section 3-7 Page 107 Website::http://www.stat.uiowa.edu/~rlenth/Power/

More information

22s:152 Applied Linear Regression. Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA)

22s:152 Applied Linear Regression. Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA) 22s:152 Applied Linear Regression Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA) We now consider an analysis with only categorical predictors (i.e. all predictors are

More information

Sampling Distributions

Sampling Distributions Sampling Distributions Sampling Distribution of the Mean & Hypothesis Testing Remember sampling? Sampling Part 1 of definition Selecting a subset of the population to create a sample Generally random sampling

More information

Gov Univariate Inference II: Interval Estimation and Testing

Gov Univariate Inference II: Interval Estimation and Testing Gov 2000-5. Univariate Inference II: Interval Estimation and Testing Matthew Blackwell October 13, 2015 1 / 68 Large Sample Confidence Intervals Confidence Intervals Example Hypothesis Tests Hypothesis

More information

LECTURE 5. Introduction to Econometrics. Hypothesis testing

LECTURE 5. Introduction to Econometrics. Hypothesis testing LECTURE 5 Introduction to Econometrics Hypothesis testing October 18, 2016 1 / 26 ON TODAY S LECTURE We are going to discuss how hypotheses about coefficients can be tested in regression models We will

More information

Exam 2 (KEY) July 20, 2009

Exam 2 (KEY) July 20, 2009 STAT 2300 Business Statistics/Summer 2009, Section 002 Exam 2 (KEY) July 20, 2009 Name: USU A#: Score: /225 Directions: This exam consists of six (6) questions, assessing material learned within Modules

More information

psyc3010 lecture 2 factorial between-ps ANOVA I: omnibus tests

psyc3010 lecture 2 factorial between-ps ANOVA I: omnibus tests psyc3010 lecture 2 factorial between-ps ANOVA I: omnibus tests last lecture: introduction to factorial designs next lecture: factorial between-ps ANOVA II: (effect sizes and follow-up tests) 1 general

More information

CENTRAL LIMIT THEOREM (CLT)

CENTRAL LIMIT THEOREM (CLT) CENTRAL LIMIT THEOREM (CLT) A sampling distribution is the probability distribution of the sample statistic that is formed when samples of size n are repeatedly taken from a population. If the sample statistic

More information

COMPLETELY RANDOM DESIGN (CRD) -Design can be used when experimental units are essentially homogeneous.

COMPLETELY RANDOM DESIGN (CRD) -Design can be used when experimental units are essentially homogeneous. COMPLETELY RANDOM DESIGN (CRD) Description of the Design -Simplest design to use. -Design can be used when experimental units are essentially homogeneous. -Because of the homogeneity requirement, it may

More information

Ch 2: Simple Linear Regression

Ch 2: Simple Linear Regression Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component

More information

Human-Centered Fidelity Metrics for Virtual Environment Simulations

Human-Centered Fidelity Metrics for Virtual Environment Simulations Human-Centered Fidelity Metrics for Virtual Environment Simulations Three Numbers from Standard Experimental Design and Analysis: α, power, effect magnitude VR 2005 Tutorial J. Edward Swan II, Mississippi

More information

Chapter 7 Comparison of two independent samples

Chapter 7 Comparison of two independent samples Chapter 7 Comparison of two independent samples 7.1 Introduction Population 1 µ σ 1 1 N 1 Sample 1 y s 1 1 n 1 Population µ σ N Sample y s n 1, : population means 1, : population standard deviations N

More information

Confidence Interval Estimation

Confidence Interval Estimation Department of Psychology and Human Development Vanderbilt University 1 Introduction 2 3 4 5 Relationship to the 2-Tailed Hypothesis Test Relationship to the 1-Tailed Hypothesis Test 6 7 Introduction In

More information

t-test for 2 matched/related/ dependent samples

t-test for 2 matched/related/ dependent samples HUMBEHV 3HB3 two-sample t-tests & statistical power Week 9 Prof. Patrick Bennett Concepts from previous lectures t distribution standard error of the mean degrees-of-freedom Null and alternative/research

More information

Data analysis and Geostatistics - lecture VII

Data analysis and Geostatistics - lecture VII Data analysis and Geostatistics - lecture VII t-tests, ANOVA and goodness-of-fit Statistical testing - significance of r Testing the significance of the correlation coefficient: t = r n - 2 1 - r 2 with

More information

Stat 135, Fall 2006 A. Adhikari HOMEWORK 6 SOLUTIONS

Stat 135, Fall 2006 A. Adhikari HOMEWORK 6 SOLUTIONS Stat 135, Fall 2006 A. Adhikari HOMEWORK 6 SOLUTIONS 1a. Under the null hypothesis X has the binomial (100,.5) distribution with E(X) = 50 and SE(X) = 5. So P ( X 50 > 10) is (approximately) two tails

More information

4/6/16. Non-parametric Test. Overview. Stephen Opiyo. Distinguish Parametric and Nonparametric Test Procedures

4/6/16. Non-parametric Test. Overview. Stephen Opiyo. Distinguish Parametric and Nonparametric Test Procedures Non-parametric Test Stephen Opiyo Overview Distinguish Parametric and Nonparametric Test Procedures Explain commonly used Nonparametric Test Procedures Perform Hypothesis Tests Using Nonparametric Procedures

More information

Can you tell the relationship between students SAT scores and their college grades?

Can you tell the relationship between students SAT scores and their college grades? Correlation One Challenge Can you tell the relationship between students SAT scores and their college grades? A: The higher SAT scores are, the better GPA may be. B: The higher SAT scores are, the lower

More information

Statistics for Managers Using Microsoft Excel Chapter 9 Two Sample Tests With Numerical Data

Statistics for Managers Using Microsoft Excel Chapter 9 Two Sample Tests With Numerical Data Statistics for Managers Using Microsoft Excel Chapter 9 Two Sample Tests With Numerical Data 999 Prentice-Hall, Inc. Chap. 9 - Chapter Topics Comparing Two Independent Samples: Z Test for the Difference

More information

22s:152 Applied Linear Regression. 1-way ANOVA visual:

22s:152 Applied Linear Regression. 1-way ANOVA visual: 22s:152 Applied Linear Regression 1-way ANOVA visual: Chapter 8: 1-Way Analysis of Variance (ANOVA) 2-Way Analysis of Variance (ANOVA) 0.00 0.05 0.10 0.15 0.20 0.25 0.30 0.35 Y We now consider an analysis

More information

Review of Statistics 101

Review of Statistics 101 Review of Statistics 101 We review some important themes from the course 1. Introduction Statistics- Set of methods for collecting/analyzing data (the art and science of learning from data). Provides methods

More information

SMAM 314 Exam 3 Name. F A. A null hypothesis that is rejected at α =.05 will always be rejected at α =.01.

SMAM 314 Exam 3 Name. F A. A null hypothesis that is rejected at α =.05 will always be rejected at α =.01. SMAM 314 Exam 3 Name 1. Indicate whether the following statements are true (T) or false (F) (6 points) F A. A null hypothesis that is rejected at α =.05 will always be rejected at α =.01. T B. A course

More information

REED TUTORIALS (Pty) LTD ECS3706 EXAM PACK

REED TUTORIALS (Pty) LTD ECS3706 EXAM PACK REED TUTORIALS (Pty) LTD ECS3706 EXAM PACK 1 ECONOMETRICS STUDY PACK MAY/JUNE 2016 Question 1 (a) (i) Describing economic reality (ii) Testing hypothesis about economic theory (iii) Forecasting future

More information

Making Inferences About Parameters

Making Inferences About Parameters Making Inferences About Parameters Parametric statistical inference may take the form of: 1. Estimation: on the basis of sample data we estimate the value of some parameter of the population from which

More information

Statistics: revision

Statistics: revision NST 1B Experimental Psychology Statistics practical 5 Statistics: revision Rudolf Cardinal & Mike Aitken 29 / 30 April 2004 Department of Experimental Psychology University of Cambridge Handouts: Answers

More information

Preview from Notesale.co.uk Page 3 of 63

Preview from Notesale.co.uk Page 3 of 63 Stem-and-leaf diagram - vertical numbers on far left represent the 10s, numbers right of the line represent the 1s The mean should not be used if there are extreme scores, or for ranks and categories Unbiased

More information

Relating Graph to Matlab

Relating Graph to Matlab There are two related course documents on the web Probability and Statistics Review -should be read by people without statistics background and it is helpful as a review for those with prior statistics

More information

Parameter Estimation, Sampling Distributions & Hypothesis Testing

Parameter Estimation, Sampling Distributions & Hypothesis Testing Parameter Estimation, Sampling Distributions & Hypothesis Testing Parameter Estimation & Hypothesis Testing In doing research, we are usually interested in some feature of a population distribution (which

More information

2 and F Distributions. Barrow, Statistics for Economics, Accounting and Business Studies, 4 th edition Pearson Education Limited 2006

2 and F Distributions. Barrow, Statistics for Economics, Accounting and Business Studies, 4 th edition Pearson Education Limited 2006 and F Distributions Lecture 9 Distribution The distribution is used to: construct confidence intervals for a variance compare a set of actual frequencies with expected frequencies test for association

More information

Studio 8: NHST: t-tests and Rejection Regions Spring 2014

Studio 8: NHST: t-tests and Rejection Regions Spring 2014 Studio 8: NHST: t-tests and Rejection Regions 18.05 Spring 2014 You should have downloaded studio8.zip and unzipped it into your 18.05 working directory. January 2, 2017 2 / 10 Left-side vs Right-side

More information

Hypothesis Testing. Mean (SDM)

Hypothesis Testing. Mean (SDM) Confidence Intervals and Hypothesis Testing Readings: Howell, Ch. 4, 7 The Sampling Distribution of the Mean (SDM) Derivation - See Thorne & Giesen (T&G), pp. 169-171 or online Chapter Overview for Ch.

More information

HYPOTHESIS TESTING II TESTS ON MEANS. Sorana D. Bolboacă

HYPOTHESIS TESTING II TESTS ON MEANS. Sorana D. Bolboacă HYPOTHESIS TESTING II TESTS ON MEANS Sorana D. Bolboacă OBJECTIVES Significance value vs p value Parametric vs non parametric tests Tests on means: 1 Dec 14 2 SIGNIFICANCE LEVEL VS. p VALUE Materials and

More information

Chap The McGraw-Hill Companies, Inc. All rights reserved.

Chap The McGraw-Hill Companies, Inc. All rights reserved. 11 pter11 Chap Analysis of Variance Overview of ANOVA Multiple Comparisons Tests for Homogeneity of Variances Two-Factor ANOVA Without Replication General Linear Model Experimental Design: An Overview

More information

This gives us an upper and lower bound that capture our population mean.

This gives us an upper and lower bound that capture our population mean. Confidence Intervals Critical Values Practice Problems 1 Estimation 1.1 Confidence Intervals Definition 1.1 Margin of error. The margin of error of a distribution is the amount of error we predict when

More information

N J SS W /df W N - 1

N J SS W /df W N - 1 One-Way ANOVA Source Table ANOVA MODEL: ij = µ* + α j + ε ij H 0 : µ = µ =... = µ j or H 0 : Σα j = 0 Source Sum of Squares df Mean Squares F J Between Groups nj( j * ) J - SS B /(J ) MS B /MS W = ( N

More information

Power Analysis. Ben Kite KU CRMDA 2015 Summer Methodology Institute

Power Analysis. Ben Kite KU CRMDA 2015 Summer Methodology Institute Power Analysis Ben Kite KU CRMDA 2015 Summer Methodology Institute Created by Terrence D. Jorgensen, 2014 Recall Hypothesis Testing? Null Hypothesis Significance Testing (NHST) is the most common application

More information

9/2/2010. Wildlife Management is a very quantitative field of study. throughout this course and throughout your career.

9/2/2010. Wildlife Management is a very quantitative field of study. throughout this course and throughout your career. Introduction to Data and Analysis Wildlife Management is a very quantitative field of study Results from studies will be used throughout this course and throughout your career. Sampling design influences

More information

Questions 3.83, 6.11, 6.12, 6.17, 6.25, 6.29, 6.33, 6.35, 6.50, 6.51, 6.53, 6.55, 6.59, 6.60, 6.65, 6.69, 6.70, 6.77, 6.79, 6.89, 6.

Questions 3.83, 6.11, 6.12, 6.17, 6.25, 6.29, 6.33, 6.35, 6.50, 6.51, 6.53, 6.55, 6.59, 6.60, 6.65, 6.69, 6.70, 6.77, 6.79, 6.89, 6. Chapter 7 Reading 7.1, 7.2 Questions 3.83, 6.11, 6.12, 6.17, 6.25, 6.29, 6.33, 6.35, 6.50, 6.51, 6.53, 6.55, 6.59, 6.60, 6.65, 6.69, 6.70, 6.77, 6.79, 6.89, 6.112 Introduction In Chapter 5 and 6, we emphasized

More information

Inferential statistics

Inferential statistics Inferential statistics Inference involves making a Generalization about a larger group of individuals on the basis of a subset or sample. Ahmed-Refat-ZU Null and alternative hypotheses In hypotheses testing,

More information

Probability and Statistics

Probability and Statistics Probability and Statistics Kristel Van Steen, PhD 2 Montefiore Institute - Systems and Modeling GIGA - Bioinformatics ULg kristel.vansteen@ulg.ac.be CHAPTER 4: IT IS ALL ABOUT DATA 4a - 1 CHAPTER 4: IT

More information

Power. January 12, 2019

Power. January 12, 2019 Power January 12, 2019 Contents Definition of power Z-test example If H 0 is true If H 0 is false The 2x2 matrix of All Things That Can Happen Pr(Type I error) = α Pr(Type II error) = β power = 1 β Things

More information

PSY 216. Assignment 9 Answers. Under what circumstances is a t statistic used instead of a z-score for a hypothesis test

PSY 216. Assignment 9 Answers. Under what circumstances is a t statistic used instead of a z-score for a hypothesis test PSY 216 Assignment 9 Answers 1. Problem 1 from the text Under what circumstances is a t statistic used instead of a z-score for a hypothesis test The t statistic should be used when the population standard

More information

Distribution-Free Procedures (Devore Chapter Fifteen)

Distribution-Free Procedures (Devore Chapter Fifteen) Distribution-Free Procedures (Devore Chapter Fifteen) MATH-5-01: Probability and Statistics II Spring 018 Contents 1 Nonparametric Hypothesis Tests 1 1.1 The Wilcoxon Rank Sum Test........... 1 1. Normal

More information

CHAPTER 9, 10. Similar to a courtroom trial. In trying a person for a crime, the jury needs to decide between one of two possibilities:

CHAPTER 9, 10. Similar to a courtroom trial. In trying a person for a crime, the jury needs to decide between one of two possibilities: CHAPTER 9, 10 Hypothesis Testing Similar to a courtroom trial. In trying a person for a crime, the jury needs to decide between one of two possibilities: The person is guilty. The person is innocent. To

More information

Sampling Distributions

Sampling Distributions Sampling Error As you may remember from the first lecture, samples provide incomplete information about the population In particular, a statistic (e.g., M, s) computed on any particular sample drawn from

More information