Sampling Distributions: Central Limit Theorem

Similar documents
Hypothesis testing: Steps

Hypothesis testing: Steps

Introduction to the Analysis of Variance (ANOVA) Computing One-Way Independent Measures (Between Subjects) ANOVAs

The One-Way Independent-Samples ANOVA. (For Between-Subjects Designs)

The t-test: A z-score for a sample mean tells us where in the distribution the particular mean lies

Two-Sample Inferential Statistics

AMS7: WEEK 7. CLASS 1. More on Hypothesis Testing Monday May 11th, 2015

Sampling distribution of t. 2. Sampling distribution of t. 3. Example: Gas mileage investigation. II. Inferential Statistics (8) t =

Psych 230. Psychological Measurement and Statistics

The One-Way Repeated-Measures ANOVA. (For Within-Subjects Designs)

Chapter 7 Comparison of two independent samples

10/31/2012. One-Way ANOVA F-test

Student s t-distribution. The t-distribution, t-tests, & Measures of Effect Size

Factorial Analysis of Variance

HYPOTHESIS TESTING. Hypothesis Testing

COGS 14B: INTRODUCTION TO STATISTICAL ANALYSIS

Introduction to the Analysis of Variance (ANOVA)

Answer Key. 9.1 Scatter Plots and Linear Correlation. Chapter 9 Regression and Correlation. CK-12 Advanced Probability and Statistics Concepts 1

AP Statistics Ch 12 Inference for Proportions

The t-statistic. Student s t Test

OHSU OGI Class ECE-580-DOE :Design of Experiments Steve Brainerd

Sampling Distributions

" M A #M B. Standard deviation of the population (Greek lowercase letter sigma) σ 2

Independent Samples ANOVA

CBA4 is live in practice mode this week exam mode from Saturday!

PSY 216. Assignment 9 Answers. Under what circumstances is a t statistic used instead of a z-score for a hypothesis test

PSYC 331 STATISTICS FOR PSYCHOLOGISTS

Keppel, G. & Wickens, T.D. Design and Analysis Chapter 2: Sources of Variability and Sums of Squares

Statistics Introductory Correlation

One-factor analysis of variance (ANOVA)

Last week: Sample, population and sampling distributions finished with estimation & confidence intervals

8/23/2018. One-Way ANOVA F-test. 1. Situation/hypotheses. 2. Test statistic. 3.Distribution. 4. Assumptions

1 Descriptive statistics. 2 Scores and probability distributions. 3 Hypothesis testing and one-sample t-test. 4 More on t-tests

LECTURE 5. Introduction to Econometrics. Hypothesis testing

Inferences About Two Proportions

Section 9.4. Notation. Requirements. Definition. Inferences About Two Means (Matched Pairs) Examples

Chapter 9 Inferences from Two Samples

CIVL /8904 T R A F F I C F L O W T H E O R Y L E C T U R E - 8

First we look at some terms to be used in this section.

Statistical Inference. Why Use Statistical Inference. Point Estimates. Point Estimates. Greg C Elvers

Introduction to Business Statistics QM 220 Chapter 12

Sampling Distributions

COMPARING SEVERAL MEANS: ANOVA

STAT 135 Lab 9 Multiple Testing, One-Way ANOVA and Kruskal-Wallis

POLI 443 Applied Political Research

Chapter 5: HYPOTHESIS TESTING

Last two weeks: Sample, population and sampling distributions finished with estimation & confidence intervals

Statistics Primer. ORC Staff: Jayme Palka Peter Boedeker Marcus Fagan Trey Dejong

Multiple t Tests. Introduction to Analysis of Variance. Experiments with More than 2 Conditions

Hypothesis T e T sting w ith with O ne O One-Way - ANOV ANO A V Statistics Arlo Clark Foos -

An Analysis of College Algebra Exam Scores December 14, James D Jones Math Section 01

Probability and Samples. Sampling. Point Estimates

HYPOTHESIS TESTING II TESTS ON MEANS. Sorana D. Bolboacă

One-Way ANOVA. Some examples of when ANOVA would be appropriate include:

This document contains 3 sets of practice problems.

Lecture 10: Comparing two populations: proportions

Review. One-way ANOVA, I. What s coming up. Multiple comparisons

COSC 341 Human Computer Interaction. Dr. Bowen Hui University of British Columbia Okanagan

Data Analysis and Statistical Methods Statistics 651

Variance Estimates and the F Ratio. ERSH 8310 Lecture 3 September 2, 2009

SEVERAL μs AND MEDIANS: MORE ISSUES. Business Statistics

Correlation and Linear Regression

Written Exam (2 hours)

CHAPTER 10 ONE-WAY ANALYSIS OF VARIANCE. It would be very unusual for all the research one might conduct to be restricted to

COSC 341 Human Computer Interaction. Dr. Bowen Hui University of British Columbia Okanagan

Difference in two or more average scores in different groups

Unit 27 One-Way Analysis of Variance

16.400/453J Human Factors Engineering. Design of Experiments II

Lab #12: Exam 3 Review Key

7.2 One-Sample Correlation ( = a) Introduction. Correlation analysis measures the strength and direction of association between

The Difference in Proportions Test

1. The (dependent variable) is the variable of interest to be measured in the experiment.

y = a + bx 12.1: Inference for Linear Regression Review: General Form of Linear Regression Equation Review: Interpreting Computer Regression Output

Name: Exam: In-term Two Page: 1 of 8 Date: 12/07/2018. University of Texas at Austin, Department of Mathematics M358K - Applied Statistics TRUE/FALSE

Econometrics. 4) Statistical inference

Practice Questions: Statistics W1111, Fall Solutions

Week 14 Comparing k(> 2) Populations

Lecture 14. Analysis of Variance * Correlation and Regression. The McGraw-Hill Companies, Inc., 2000

Lecture 14. Outline. Outline. Analysis of Variance * Correlation and Regression Analysis of Variance (ANOVA)

Parameter Estimation, Sampling Distributions & Hypothesis Testing

Hypothesis Testing. We normally talk about two types of hypothesis: the null hypothesis and the research or alternative hypothesis.

Inferences for Correlation

PSY 305. Module 3. Page Title. Introduction to Hypothesis Testing Z-tests. Five steps in hypothesis testing

5 Basic Steps in Any Hypothesis Test

Two sample Hypothesis tests in R.

Analysis of Variance (ANOVA)

STA 101 Final Review

REVIEW 8/2/2017 陈芳华东师大英语系

Factorial Analysis of Variance

1. What does the alternate hypothesis ask for a one-way between-subjects analysis of variance?

Simple Linear Regression: One Quantitative IV

Stat 529 (Winter 2011) Experimental Design for the Two-Sample Problem. Motivation: Designing a new silver coins experiment

Hypothesis Testing and Confidence Intervals (Part 2): Cohen s d, Logic of Testing, and Confidence Intervals

10/4/2013. Hypothesis Testing & z-test. Hypothesis Testing. Hypothesis Testing

Single Sample Means. SOCY601 Alan Neustadtl

Soc 3811 Basic Social Statistics Second Midterm Exam Spring Your Name [50 points]: ID #: ANSWERS

Hypothesis Testing hypothesis testing approach formulation of the test statistic

Math 101: Elementary Statistics Tests of Hypothesis

Review of Statistics 101

Statistics: revision

Transcription:

Review for Exam 2

Sampling Distributions: Central Limit Theorem Conceptually, we can break up the theorem into three parts: 1. The mean (µ M ) of a population of sample means (M) is equal to the mean (µ) of the underlying population of scores 2. The standard deviation (σ M ) of a population of sample means (called the standard error) becomes smaller as the size of the sample (n, the number of scores used to compute M) increases 3. Whatever the distribution of the underlying scores, the sampling distribution of the mean will approach the normal distribution as the sample size (n) increases

p(m) Exam 2 Review Sample Size & Standard Error n 1 M 7 n 4 M 3. 5 4 n 16 M 1.75 16 (M)

Sampling Error & Sampling Distributions Exam 2 Review Sampling error prevents us from directly comparing sample statistics to population parameters to evaluate hypotheses The sampling distribution tells us how much variability (error) to expect in our sample statistics (e.g., M) The standard error of a statistic (the standard deviation of its sampling distribution) depends both on the variability of the underlying population scores (σ) and on the size of the sample (n).

Hypothesis Testing The purpose of a hypothesis test is to decide between two explanations: 1. H 0 : The difference between the sample and the population can be explained by sampling error (there does not appear to be an effect of treatment or condition) 2. H 1 : The difference between the sample and the population is too large to be explained by sampling error (there does appear to be an effect of treatment or condition).

Hypothesis Testing The null hypothesis is the default hypothesis that there is NO treatment effect We compute the sampling distribution of the statistic of interest (e.g., the mean exam score of the sample) under the assumption that the null hypothesis (H 0 ) is true We determine the critical values & rejection regions What values of the statistic, were they observed, would be extreme enough to make us reject the null hypothesis? Generally, we will compute the sampling distribution of a standardized test statistic (e.g., z, t, or F statistics).

0 1 Population distributions Raw scores (x) n 4 M n 2 Sampling distributions M crit Sample means (M)

Errors in Hypothesis Testing P = α P = 1-β P = 1-α P = β

Step 3: Make a Decision critical region (extreme 5%) reject H 0 middle 95% retain H 0 α= 0.05; 2-tailed test M µ from H 0 z -1.96 0 1.96 Z= 1.60 Retain H 0 : no difference

critical region (extreme 5%) reject H 0 middle 95% retain H 0 α= 0.05; 2-tailed test n = 16 M µ from H 0 z -1.96 0 1.96 z = 2.86 reject H 0 : significant difference

Hypothesis testing: Steps 1. Determine appropriate test and hypotheses 2. Use distribution table to find critical statistic value(s) representing rejection region Remember: the rejection region represents values that you would be unlikely to obtain if the null hypothesis were true 3. Compute appropriate test statistic from data 4. Make a decision: does the statistic for your sample fall into the critical (rejection) region or not?

Computing Mean-Difference Statistics 2 M 0 ˆM ˆ Test Sample Data Hypothesized Population Parameter Estimated Standard Error Estimated Variance Exam 2 Review Degrees of Freedom z-test M µ σ 2 n σ 2 n Singlesample t-test M µ s 2 n s 2 = SS df df = n 1 Relatedsamples t-test M D, where D = x 2 x 1 μ D =0 s D 2 n D s D 2 = SS D df D df = n D 1 Independentsamples t-test M 1 M 2 μ 1 μ 2 =0 s p 2 2 + s p s 2 p = SS 1 + SS 2 df = n n 1 n 2 df 1 + df 1 +n 2 2 2 Post-hoc tests M A M B μ A μ B =0 MS error n A + MS error n B MS error = SS error df error df error

yes z-test one Is σ provided? no One-sample t-test Number of Samples two Are scores matched across samples? yes no Related samples t-test Independent samples t-test >2 ANOVA

Example Scenario A professor thinks that this year s freshman class seems to be smarter than previous classes. To test this, she administers an IQ test to a sample of 36 freshman and computes the mean (M=114.5) and standard deviation (s = 18) of their scores. College records indicate that the mean IQ across previous years was 110.3. What is the appropriate statistical test for this problem? a) A z-test b) A one-sample t-test c) An independent-samples t-test d) A related-samples t-test

Example Scenario A psychologist examined the effect of exercise on a standardized memory test. Scores on this test for the general population form a normal distribution with a mean of 50 and a standard deviation of 8. A sample of 62 people who exercise at least 3 hours per week has a mean score of 57. What is the appropriate statistical test for this problem? a) A z-test b) A one-sample t-test c) An independent-samples t-test d) A related-samples t-test

Example Scenario A researcher studies the effect of a drug on nightmares in veterans with PTSD. A sample of clients with PTSD kept count of their nightmares for 1 month before treatment. They were then given the medication and asked to record counts of their nightmares again for a month following treatment. What is the appropriate statistical test for this problem? a) A z-test b) A one-sample t-test c) An independent-samples t-test d) A related-samples t-test

Example Scenario A neurologist had two groups of patients with different types of aphasia (a brain disorder) name objects presented to them as line drawings. He wanted to determine whether the number of objects correctly named differed across the two groups. What is the appropriate statistical test for this problem? a) A z-test b) A one-sample t-test c) An independent-samples t-test d) A related-samples t-test

Computational Example 1 A psychologist is investigating the effect of being an only child on personality. A sample of 30 only children is obtained and each child is given a standardized personality test. Population scores on the test form a normal distribution with µ = 50 and σ =15. a) If the mean for the sample is 58, what can the researcher conclude about his hypothesis? Use a two-tailed test with α = 0.05. b) Compute a 90% confidence interval for the mean of the only child population

Clicker Question What is the appropriate statistical test for part (a) of this problem? a) A z-test b) A one-sample t-test c) An independent-samples t-test d) A related-samples t-test

Example 1: Research Hypothesis H 1 : Only children have different personality scores than the general population I.e., µ only children µ general population Null Hypothesis H 0 : Only children have personality scores that are not significantly different than those of the general population I.e., µ only children = µ general population We have,, n, and M

Find Critical z value z 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.0 0.5000 0.4960 0.4920 0.4880 0.4840 0.4801 0.4761 0.4721 0.4681 0.4641 0.1 0.4602 0.4562 0.4522 0.4483 0.4443 0.4404 0.4364 0.4325 0.4286 0.4247 0.2 0.4207 0.4168 0.4129 0.4090 0.4052 0.4013 0.3974 0.3936 0.3897 0.3859 0.3 0.3821 0.3783 0.3745 0.3707 0.3669 0.3632 0.3594 0.3557 0.3520 0.3483 0.4 0.3446 0.3409 0.3372 0.3336 0.3300 0.3264 0.3228 0.3192 0.3156 0.3121 0.5 0.3085 0.3050 0.3015 0.2981 0.2946 0.2912 0.2877 0.2843 0.2810 0.2776 0.6 0.2743 0.2709 0.2676 0.2643 0.2611 0.2578 0.2546 0.2514 0.2483 0.2451 0.7 0.2420 0.2389 0.2358 0.2327 0.2296 0.2266 0.2236 0.2206 0.2177 0.2148 0.8 0.2119 0.2090 0.2061 0.2033 0.2005 0.1977 0.1949 0.1922 0.1894 0.1867 0.9 0.1841 0.1814 0.1788 0.1762 0.1736 0.1711 0.1685 0.1660 0.1635 0.1611 1.0 0.1587 0.1562 0.1539 0.1515 0.1492 0.1469 0.1446 0.1423 0.1401 0.1379 1.1 0.1357 0.1335 0.1314 0.1292 0.1271 0.1251 0.1230 0.1210 0.1190 0.1170 1.2 0.1151 0.1131 0.1112 0.1093 0.1075 0.1056 0.1038 0.1020 0.1003 0.0985 1.3 0.0968 0.0951 0.0934 0.0918 0.0901 0.0885 0.0869 0.0853 0.0838 0.0823 1.4 0.0808 0.0793 0.0778 0.0764 0.0749 0.0735 0.0721 0.0708 0.0694 0.0681 1.5 0.0668 0.0655 0.0643 0.0630 0.0618 0.0606 0.0594 0.0582 0.0571 0.0559 1.6 0.0548 0.0537 0.0526 0.0516 0.0505 0.0495 0.0485 0.0475 0.0465 0.0455 1.7 0.0446 0.0436 0.0427 0.0418 0.0409 0.0401 0.0392 0.0384 0.0375 0.0367 1.8 0.0359 0.0351 0.0344 0.0336 0.0329 0.0322 0.0314 0.0307 0.0301 0.0294 1.9 0.0287 0.0281 0.0274 0.0268 0.0262 0.0256 0.0250 0.0244 0.0239 0.0233 2.0 0.0228 0.0222 0.0217 0.0212 0.0207 0.0202 0.0197 0.0192 0.0188 0.0183 2.1 0.0179 0.0174 0.0170 0.0166 0.0162 0.0158 0.0154 0.0150 0.0146 0.0143 2.2 0.0139 0.0136 0.0132 0.0129 0.0125 0.0122 0.0119 0.0116 0.0113 0.0110 2.3 0.0107 0.0104 0.0102 0.0099 0.0096 0.0094 0.0091 0.0089 0.0087 0.0084 2.4 0.0082 0.0080 0.0078 0.0075 0.0073 0.0071 0.0069 0.0068 0.0066 0.0064 2.5 0.0062 0.0060 0.0059 0.0057 0.0055 0.0054 0.0052 0.0051 0.0049 0.0048 2.6 0.0047 0.0045 0.0044 0.0043 0.0041 0.0040 0.0039 0.0038 0.0037 0.0036 2.7 0.0035 0.0034 0.0033 0.0032 0.0031 0.0030 0.0029 0.0028 0.0027 0.0026 2.8 0.0026 0.0025 0.0024 0.0023 0.0023 0.0022 0.0021 0.0021 0.0020 0.0019 2.9 0.0019 0.0018 0.0018 0.0017 0.0016 0.0016 0.0015 0.0015 0.0014 0.0014 Upper-Tail Probabilities

Example1: Compute z-statistic 50.0 15.0 n 30 M 58.0 z crit 1.96 z M n 58 50 15 30 8 2.92 2.74 2.92 1.96; reject H 0 The personalities of only children differ significantly from those of the general population z = 2.92, p = 0.0036

Example 1: Compute Confidence Interval b) Compute a 90% confidence interval for the mean of the only child population Recall that to compute a confidence interval, you: Exam 2 Review 1. Select a level of confidence and look up the corresponding t α (or z α ) values in the t (or z) distribution table. 2. Use two-tailed probabilities (e.g., for z α, look up p(z>z) = α/2) 3. The confidence interval is computed by inverting the t (or z) transformation CI 0.1 M z M 0.90 0.1 M z n

z 0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09 0.0 0.5000 0.4960 0.4920 0.4880 0.4840 0.4801 0.4761 0.4721 0.4681 0.4641 0.1 0.4602 0.4562 0.4522 0.4483 0.4443 0.4404 0.4364 0.4325 0.4286 0.4247 0.2 0.4207 0.4168 0.4129 0.4090 0.4052 0.4013 0.3974 0.3936 0.3897 0.3859 0.3 0.3821 0.3783 0.3745 0.3707 0.3669 0.3632 0.3594 0.3557 0.3520 0.3483 0.4 0.3446 0.3409 0.3372 0.3336 0.3300 0.3264 0.3228 0.3192 0.3156 0.3121 0.5 0.3085 0.3050 0.3015 0.2981 0.2946 0.2912 0.2877 0.2843 0.2810 0.2776 0.6 0.2743 0.2709 0.2676 0.2643 0.2611 0.2578 0.2546 0.2514 0.2483 0.2451 0.7 0.2420 0.2389 0.2358 0.2327 0.2296 0.2266 0.2236 0.2206 0.2177 0.2148 0.8 0.2119 0.2090 0.2061 0.2033 0.2005 0.1977 0.1949 0.1922 0.1894 0.1867 0.9 0.1841 0.1814 0.1788 0.1762 0.1736 0.1711 0.1685 0.1660 0.1635 0.1611 1.0 0.1587 0.1562 0.1539 0.1515 0.1492 0.1469 0.1446 0.1423 0.1401 0.1379 1.1 0.1357 0.1335 0.1314 0.1292 0.1271 0.1251 0.1230 0.1210 0.1190 0.1170 1.2 0.1151 0.1131 0.1112 0.1093 0.1075 0.1056 0.1038 0.1020 0.1003 0.0985 1.3 0.0968 0.0951 0.0934 0.0918 0.0901 0.0885 0.0869 0.0853 0.0838 0.0823 1.4 0.0808 0.0793 0.0778 0.0764 0.0749 0.0735 0.0721 0.0708 0.0694 0.0681 1.5 0.0668 0.0655 0.0643 0.0630 0.0618 0.0606 0.0594 0.0582 0.0571 0.0559 1.6 0.0548 0.0537 0.0526 0.0516 0.0505 0.0495 0.0485 0.0475 0.0465 0.0455 1.7 0.0446 0.0436 0.0427 0.0418 0.0409 0.0401 0.0392 0.0384 0.0375 0.0367 1.8 0.0359 0.0351 0.0344 0.0336 0.0329 0.0322 0.0314 0.0307 0.0301 0.0294 1.9 0.0287 0.0281 0.0274 0.0268 0.0262 0.0256 0.0250 0.0244 0.0239 0.0233 2.0 0.0228 0.0222 0.0217 0.0212 0.0207 0.0202 0.0197 0.0192 0.0188 0.0183 2.1 0.0179 0.0174 0.0170 0.0166 0.0162 0.0158 0.0154 0.0150 0.0146 0.0143 2.2 0.0139 0.0136 0.0132 0.0129 0.0125 0.0122 0.0119 0.0116 0.0113 0.0110 2.3 0.0107 0.0104 0.0102 0.0099 0.0096 0.0094 0.0091 0.0089 0.0087 0.0084 2.4 0.0082 0.0080 0.0078 0.0075 0.0073 0.0071 0.0069 0.0068 0.0066 0.0064 2.5 0.0062 0.0060 0.0059 0.0057 0.0055 0.0054 0.0052 0.0051 0.0049 0.0048 2.6 0.0047 0.0045 0.0044 0.0043 0.0041 0.0040 0.0039 0.0038 0.0037 0.0036 2.7 0.0035 0.0034 0.0033 0.0032 0.0031 0.0030 0.0029 0.0028 0.0027 0.0026 2.8 0.0026 0.0025 0.0024 0.0023 0.0023 0.0022 0.0021 0.0021 0.0020 0.0019 2.9 0.0019 0.0018 0.0018 0.0017 0.0016 0.0016 0.0015 0.0015 0.0014 0.0014 Upper-Tail Probabilities

Example 1: Compute Confidence Interval Exam 2 Review 50.0 15.0 n 30 M 58.0 z 1.645 CI 0.90 M z 0. 1 n 1.645 15 58 30 24.68 58 5.48 58 4.51 [53.49, 62.51] 53.49 62.51

Top Incorrect Problems (Exam 2) If the population from which we sample is normal, the sampling distribution of the mean a) will approach normal for large sample sizes. b) will be normal. c) will be normal only for small samples. d) will be slightly positively skewed.

Top Incorrect Problems (Exam 2) If the population from which we sample is normal, the sampling distribution of the mean a) will approach normal for large sample sizes. b) will be normal. c) will be normal only for small samples. d) will be slightly positively skewed. Why? If the population is normal, then the distribution of the mean for any sample size (including a sample size of 1) will be normal.

Top Incorrect Problems (Exam 2) Sampling distributions help us test hypotheses about means by a) telling us what kinds of means to expect if the null hypothesis is false. b) telling us what kinds of means to expect if the null hypothesis is true. c) telling us how variable the population is. d) telling us exactly what the population mean is.

Top Incorrect Problems (Exam 2) Sampling distributions help us test hypotheses about means by a) telling us what kinds of means to expect if the null hypothesis is false. b) telling us what kinds of means to expect if the null hypothesis is true. c) telling us how variable the population is. d) telling us exactly what the population mean is. Why? Sampling distributions are the distributions of a sample statistic. In many hypothesis tests (e.g., t-tests and ANOVAs), we use these distributions to predict the distribution of means or mean differences under the null hypothesis.

Example 2 An educational psychologist studies the effect of frequent testing on retention of class material. In one section of an introductory course, students are given quizzes each week. A second section of the same course receives only two tests during the semester. At the end of the semester, a sample from each of the sections receives the same final exam, and the number of errors made are recorded. X 1 (quiz) X 2 (no quiz) 11 11 6 14 8 11 11 17 9 12 13 7 14 M 9.67 12.29 SS 31.33 59.43 Does frequent testing significantly affect retention of class material? Use a two-tailed test, with α = 0.05.

Clicker Question What is the appropriate statistical test for this problem? a) A z-test b) A one-sample t-test c) An independent-samples t-test d) A related-samples t-test

Example 2: Research Hypothesis H 1 : Testing frequency affects retention of class material I.e., µ quiz µ no quiz Null Hypothesis H 0 : Testing frequency does not significantly affect retention of class material I.e., µ quiz = µ no quiz We have no population data, and sample data from two independent samples, so we must use the independentsamples t-test df = n 1 + n 2 2 = 6+7-2 = 11

t-distribution Table t One-tailed test α/2 -t t Two-tailed test α α/2 Level of significance for one-tailed test 0.25 0.2 0.15 0.1 0.05 0.025 0.01 0.005 0.0005 Level of significance for two-tailed test df 0.5 0.4 0.3 0.2 0.1 0.05 0.02 0.01 0.001 1 1.000 1.376 1.963 3.078 6.314 12.706 31.821 63.657 636.619 2 0.816 1.061 1.386 1.886 2.920 4.303 6.965 9.925 31.599 3 0.765 0.978 1.250 1.638 2.353 3.182 4.541 5.841 12.924 4 0.741 0.941 1.190 1.533 2.132 2.776 3.747 4.604 8.610 5 0.727 0.920 1.156 1.476 2.015 2.571 3.365 4.032 6.869 6 0.718 0.906 1.134 1.440 1.943 2.447 3.143 3.707 5.959 7 0.711 0.896 1.119 1.415 1.895 2.365 2.998 3.499 5.408 8 0.706 0.889 1.108 1.397 1.860 2.306 2.896 3.355 5.041 9 0.703 0.883 1.100 1.383 1.833 2.262 2.821 3.250 4.781 10 0.700 0.879 1.093 1.372 1.812 2.228 2.764 3.169 4.587 11 0.697 0.876 1.088 1.363 1.796 2.201 2.718 3.106 4.437 12 0.695 0.873 1.083 1.356 1.782 2.179 2.681 3.055 4.318 13 0.694 0.870 1.079 1.350 1.771 2.160 2.650 3.012 4.221 14 0.692 0.868 1.076 1.345 1.761 2.145 2.624 2.977 4.140 15 0.691 0.866 1.074 1.341 1.753 2.131 2.602 2.947 4.073 16 0.690 0.865 1.071 1.337 1.746 2.120 2.583 2.921 4.015 17 0.689 0.863 1.069 1.333 1.740 2.110 2.567 2.898 3.965 18 0.688 0.862 1.067 1.330 1.734 2.101 2.552 2.878 3.922 19 0.688 0.861 1.066 1.328 1.729 2.093 2.539 2.861 3.883 20 0.687 0.860 1.064 1.325 1.725 2.086 2.528 2.845 3.850 21 0.686 0.859 1.063 1.323 1.721 2.080 2.518 2.831 3.819 22 0.686 0.858 1.061 1.321 1.717 2.074 2.508 2.819 3.792 23 0.685 0.858 1.060 1.319 1.714 2.069 2.500 2.807 3.768 24 0.685 0.857 1.059 1.318 1.711 2.064 2.492 2.797 3.745 25 0.684 0.856 1.058 1.316 1.708 2.060 2.485 2.787 3.725 26 0.684 0.856 1.058 1.315 1.706 2.056 2.479 2.779 3.707 27 0.684 0.855 1.057 1.314 1.703 2.052 2.473 2.771 3.690 28 0.683 0.855 1.056 1.313 1.701 2.048 2.467 2.763 3.674 29 0.683 0.854 1.055 1.311 1.699 2.045 2.462 2.756 3.659 30 0.683 0.854 1.055 1.310 1.697 2.042 2.457 2.750 3.646 40 0.681 0.851 1.050 1.303 1.684 2.021 2.423 2.704 3.551 50 0.679 0.849 1.047 1.299 1.676 2.009 2.403 2.678 3.496 100 0.677 0.845 1.042 1.290 1.660 1.984 2.364 2.626 3.390

Compute t Statistic M SS n 1 M SS n 2 1 1 2 2 9.67 31.33 6 12.29 59.43 7 df 11 t crit 2.201 s M Compute Pooled Variance: s SS SS 2 1 2 p df1 df2 31.33 59.43 8.25 11 Estimate Standard Error: M 1 2 s n 2 p s n 2 p 1 2 8.25 8.25 2.55 1. 60 6 7 Compute t-statistic: t df t 11 M s 1 2 M M M 1 2 9.67 12.29 1.6 1. 64 1.64 2.201; retain H 0 Frequent testing does not significantly affect the amount of information retained by students t(11) = 1.64, p>0.05.

Top Incorrect Problems (Exam 2) The reason why we need to solve for t instead of z in some situations is due to a) the sampling distribution of the mean. b) the size of our sample mean. c) the sampling distribution of the sample size. d) the sampling distribution of the variance.

Top Incorrect Problems (Exam 2) The reason why we need to solve for t instead of z in some situations is due to a) the sampling distribution of the mean. b) the size of our sample mean. c) the sampling distribution of the sample size. d) the sampling distribution of the variance. Why? We solve for the z-statistic using σ, which is a constant, but we solve for the t-statistic using the sample standard deviation s which varies from sample to sample.