Notes 3: Statistical Inference: Sampling, Sampling Distributions Confidence Intervals, and Hypothesis Testing

Similar documents
UNIVERSITY OF TORONTO MISSISSAUGA. SOC222 Measuring Society In-Class Test. November 11, 2011 Duration 11:15a.m. 13 :00p.m.

Lecture 5: Sampling Methods

Review of the Normal Distribution

9/2/2010. Wildlife Management is a very quantitative field of study. throughout this course and throughout your career.

Probability and Samples. Sampling. Point Estimates

Examine characteristics of a sample and make inferences about the population

Inferential Statistics. Chapter 5

FCE 3900 EDUCATIONAL RESEARCH LECTURE 8 P O P U L A T I O N A N D S A M P L I N G T E C H N I Q U E

Statistical Inference

Inferential Statistics

POPULATION AND SAMPLE

Now we will define some common sampling plans and discuss their strengths and limitations.

MN 400: Research Methods. CHAPTER 7 Sample Design

ECON1310 Quantitative Economic and Business Analysis A

CHOOSING THE RIGHT SAMPLING TECHNIQUE FOR YOUR RESEARCH. Awanis Ku Ishak, PhD SBM

Ch. 17. DETERMINATION OF SAMPLE SIZE

Last week: Sample, population and sampling distributions finished with estimation & confidence intervals

Module 9: Sampling IPDET. Sampling. Intro Concepts Types Confidence/ Precision? How Large? Intervention or Policy. Evaluation Questions

We would like to describe this population. Central tendency (mean) Variability (standard deviation) ( X X ) 2 N

Last two weeks: Sample, population and sampling distributions finished with estimation & confidence intervals

Prepared by: DR. ROZIAH MOHD RASDI Faculty of Educational Studies Universiti Putra Malaysia

Training and Technical Assistance Webinar Series Statistical Analysis for Criminal Justice Research

Chapter 2 Descriptive Statistics

Statistical Inference. Why Use Statistical Inference. Point Estimates. Point Estimates. Greg C Elvers

STAT Chapter 9: Two-Sample Problems. Paired Differences (Section 9.3)

Data Collection: What Is Sampling?

Questions 3.83, 6.11, 6.12, 6.17, 6.25, 6.29, 6.33, 6.35, 6.50, 6.51, 6.53, 6.55, 6.59, 6.60, 6.65, 6.69, 6.70, 6.77, 6.79, 6.89, 6.

Lecture 11. Data Description Estimation

Sampling. What is the purpose of sampling: Sampling Terms. Sampling and Sampling Distributions

Harvard University. Rigorous Research in Engineering Education

MALLOY PSYCH 3000 MEAN & VARIANCE PAGE 1 STATISTICS MEASURES OF CENTRAL TENDENCY. In an experiment, these are applied to the dependent variable (DV)

Sampling Distributions

SAMPLING- Method of Psychology. By- Mrs Neelam Rathee, Dept of Psychology. PGGCG-11, Chandigarh.

Fundamentals of Applied Sampling

Interpret Standard Deviation. Outlier Rule. Describe the Distribution OR Compare the Distributions. Linear Transformations SOCS. Interpret a z score

Vocabulary: Samples and Populations

Sampling (Statistics)

Probability and Inference. POLI 205 Doing Research in Politics. Populations and Samples. Probability. Fall 2015

One-Way ANOVA. Some examples of when ANOVA would be appropriate include:

Survey on Population Mean

Probability and Statistics

Sampling distributions:


Sample size and Sampling strategy

Chapter 18: Sampling Distributions

AP Statistics Cumulative AP Exam Study Guide

Chapter 22. Comparing Two Proportions 1 /29

Sampling. General introduction to sampling methods in epidemiology and some applications to food microbiology study October Hanoi

Lecturer: Dr. Adote Anum, Dept. of Psychology Contact Information:

Chapter 0: Some basic preliminaries

Z score indicates how far a raw score deviates from the sample mean in SD units. score Mean % Lower Bound

Interval estimation. October 3, Basic ideas CLT and CI CI for a population mean CI for a population proportion CI for a Normal mean

PSY 250 8/29/2011. Sampling. Choosing your Sample. Sampling: Selecting Research Participants. Selecting a sample of participants from the population

Lectures of STA 231: Biostatistics

What is Statistics? Statistics is the science of understanding data and of making decisions in the face of variability and uncertainty.

Probably About Probability p <.05. Probability. What Is Probability?

Chapter 1 Statistical Inference

Answer keys for Assignment 10: Measurement of study variables (The correct answer is underlined in bold text)

Mitosis Data Analysis: Testing Statistical Hypotheses By Dana Krempels, Ph.D. and Steven Green, Ph.D.

Understanding Inference: Confidence Intervals I. Questions about the Assignment. The Big Picture. Statistic vs. Parameter. Statistic vs.

Given a sample of n observations measured on k IVs and one DV, we obtain the equation

Statistic: a that can be from a sample without making use of any unknown. In practice we will use to establish unknown parameters.

Sampling Distribution Models. Chapter 17

Statistics and Quantitative Analysis U4320. Segment 5: Sampling and inference Prof. Sharyn O Halloran

10/4/2013. Hypothesis Testing & z-test. Hypothesis Testing. Hypothesis Testing

Two-Sample Inferential Statistics

PSY 250. Sampling. Representative Sample. Representativeness 7/23/2015. Sampling: Selecting Research Participants

Ch. 7. One sample hypothesis tests for µ and σ

where Female = 0 for males, = 1 for females Age is measured in years (22, 23, ) GPA is measured in units on a four-point scale (0, 1.22, 3.45, etc.

Figure Figure

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing

Simple Linear Regression: One Quantitative IV

T.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS

Distribution of sample means

Appendix A. Review of Basic Mathematical Operations. 22Introduction

Psych Jan. 5, 2005

Background to Statistics

Deciphering Math Notation. Billy Skorupski Associate Professor, School of Education

Statistics Introductory Correlation

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM

B. Weaver (24-Mar-2005) Multiple Regression Chapter 5: Multiple Regression Y ) (5.1) Deviation score = (Y i

Keppel, G. & Wickens, T.D. Design and Analysis Chapter 2: Sources of Variability and Sums of Squares

Unit 19 Formulating Hypotheses and Making Decisions

1 Descriptive statistics. 2 Scores and probability distributions. 3 Hypothesis testing and one-sample t-test. 4 More on t-tests

Part 7: Glossary Overview

Day 8: Sampling. Daniel J. Mallinson. School of Public Affairs Penn State Harrisburg PADM-HADM 503

Measurements and Data Analysis

Chapter 22. Comparing Two Proportions 1 /30

Descriptive Statistics-I. Dr Mahmoud Alhussami

MATH 1150 Chapter 2 Notation and Terminology

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

Business Statistics:

Part I. Sampling design. Overview. INFOWO Lecture M6: Sampling design and Experiments. Outline. Sampling design Experiments.

Prentice Hall Stats: Modeling the World 2004 (Bock) Correlated to: National Advanced Placement (AP) Statistics Course Outline (Grades 9-12)

Data Mining Chapter 4: Data Analysis and Uncertainty Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

Why do we Care About Forest Sampling?

Correlation and regression

Central Limit Theorem Confidence Intervals Worked example #6. July 24, 2017

Sampling. Module II Chapter 3

Week 11 Sample Means, CLT, Correlation

Transcription:

Notes 3: Statistical Inference: Sampling, Sampling Distributions Confidence Intervals, and Hypothesis Testing 1. Purpose of statistical inference Statistical inference provides a means of generalizing from sample data to population data; from sample findings to population relationships.. Population and Samples; Parameters and Statistics Population: Target group for our inferences, the group we to which we wish to generalize; each population must have defining characteristics. Target population is that group to which we generalize, accessible population is that group from which sample is actually drawn. Sample (n) : A subgroup, subset, or part of the population; the actual group used in the research. Parameter: Descriptive measure calculated for the population; usually denoted by Greek letters (e.g.,,, ) Statistic: Descriptive measure calculated for the sample; usually denoted by Roman letters (e.g., s, M, ) Estimate: Any statistic calculated for a sample is an estimate of its corresponding parameter; an estimator is some function of the scores that leads to an estimate 3. Randomness, Random Sampling, and Sampling Sampling: Process of selecting a subgroup from the defined population. Representativeness: The method of sampling is very important in judging the validity of inferences made from the sample to the population. At issue with the validity of inferences is the representativeness of the sample. Note that the size of a sample never compensates for its lack of representativeness of the population. it is very important that sample be representative of the population for statistics to work properly in estimating parameters no sampling procedure guarantees a single representative sample; however, some sampling procedures do guarantee representative samples over the long run--over many samples Type of samples: accidental or convenience samples: these samples do not involve any type of controlled or systematically applied random selection; these samples are frequently biased, and are therefore not representative of the population. Probability Sampling: selection in which some form of randomized selection occurs and the chances of groups of individuals can be calculated (thus elements of the population individuals within population are known and can be identified [a list of individuals can be formed]). Probability sampling includes simple random samples, stratified samples, and cluster samples. Sometimes systematic samples are included in this list. Simple Random Sample, (SRS): a SRS involves the selection of a subgroup from the population in a manner that includes identifying all elements (e.g., people) in the population (e.g., US citizens), assigning unique identification to all elements, and selecting elements based upon some randomized procedure (e.g., using table of random numbers). SRS, in the long run, will produce representative samples of the population if SRS are used, any discrepancy (sampling error) between sample and population is due to chance, not to bias

sampling error: difference between a statistic and its corresponding parameter for a given sample (i.e., M - ) with any difference due to chance and not due to systematic influences bias: difference between a statistic and its corresponding parameter for a given sample (i.e., M - ) with difference due systematic influences (e.g., selecting too many females from population due to selection strategy); a statistic is biased if it s expected value does not equal the corresponding population parameter. if SRS is used, the degree of sampling error (error due to chance) can be estimated, and this allows us to determine accuracy for statistical inference in short, the use of SRS eliminates systematic error (or bias) from existing in the population if SRS is used, then independence of observations is certain (in the long run) independence: observations in the sample are not correlated; knowledge of one observation will not enable us to predict another example of non-independence (i.e., dependence) We wish to determine the average age in a community. We randomly select five people and ask of them their ages and their spouses ages. True random sampling was not used since spouses were also used in the sample. Spouses were not randomly selected. The problem with this sample is that spouses' ages are not independent that is, often people marry others with similar ages, so spouses' ages will have some degree of positive correlation. In short, if we know your age, we can probably predict your spouse's age, so spouses' ages are not independent. example of independence We wish to determine the average age in a community. We randomly select ten people and ask of them their ages. With this selection process, we will not be able to predict the next person's age simply by looking at the previously selected person's age ages are not likely to be dependent. Stratified Sampling: similar to SRS except that population is first divided based upon some stratification variable, then SRS is performed within each category of the stratification variable (e.g., divide population of teachers according to the district in which they work, then randomly select 5 teachers from each and every district; in this example district is the stratification variable). Cluster Sampling: random selection of groups rather than individuals (e.g., randomly selecting 10 schools from all schools in state and then surveying teachers in those 10 schools selected) Systematic Sampling: with this sampling procedure, every element in the population is placed in some order (e.g., on a list), and every n th element is selected after randomly selecting a starting place near the beginning of the list. systematic sampling, while it can lead to biased results, usually is representative, thus sampling results will be generalizable when systematic sampling does result in biased samples, it is usually because of patterns in the lists of elements (e.g., listed created or sorted by some variable such as race, sex, IQ)

3 4. Point and Interval Estimates point estimate: a descriptive measure calculated from a sample, such as M, s, etc. interval estimate: a range or band within which the parameter is presumed to be included; e.g., the point estimate is M = 5, the interval estimate is 5, so the interval would be 3 to 7; so one would expect the parameter to be within this interval with a given level of confidence interval estimates convey the degree of accuracy of the estimate, therefore it is more informative than the point estimate note that the parameter one is trying to estimate does not vary it is fixed; what varies across samples are the point estimates and the intervals estimates 5. Types of Distributions population distribution: distribution of raw scores (s) in the population, such as all IQ scores in population of U.S. Note that when we make inferences from sample to population, what we are really doing in inferential statistics is making inferences about population descriptive characteristics, parameters (e.g., or ), based upon sample summary characteristics, statistics (e.g., M or s). sample distribution: distribution of raw scores in the sample, such as IQ scores in our sample of students, s. sampling distribution: distribution of all possible values of a statistic, such as the mean, M, median, mode, SD, variance, etc.. How is a sampling distribution created? If one were to randomly select a sample, of size n, from the population over and over again, we would get a slightly different M for each sample; these different Ms form a frequency distribution which is called the sampling distribution of M. Note the emphasis on sample size of n; this means that the samples must all be of the same size when constructing the sampling distribution; if the sample size changes, the sampling distribution will also change. All statistics (e.g.,, s, s ) have sampling distributions. 6. The Sampling Distribution of M ( ): As noted above, all statistics (e.g., mean, variance, standard deviation, correlation) have sampling distributions. The most commonly used, however, is the sampling distribution of the mean, M (or ). If the population distribution of s is normally distributed with a mean of and variance of, then the sampling distribution of M will be normally distributed with a mean of and a variance of /n. That the sampling distribution of M is normal in shape will play an important role in inferential tests to be discussed later. The shape of the sampling distribution is important because it will allow us to find probabilities based upon the distribution, just like finding probabilities for z scores. Some common properties of the sampling distribution of are: (a) The mean of the sampling distribution of, which may be denoted as saying the expected value of is :, is. This is the same as

4 E( ) = = In short, the mean of the sample means equals the population mean. (b) As just noted above, the shape of the sampling distribution of is normal if the population distribution is normal in shape. If the population distribution is not normal, then will not be normal, although it will approach normality rapidly as n increases. (c) The variance of the sampling distribution of is = /n. and this variance is referred to as the variance error of the mean. (d) The standard deviation of the sampling distribution of is calculated as = n = n This is called the standard error of the mean and is symbolized. Note that is just the standard deviation of the means, s. As the sample size increases (n ), the standard error of the mean will become smaller, which means that we are getting more precise information about the population from the sample. For example, if a SRS was taken of IQ scores with = 100 and = 15, and a sample size of (a) n = 9, then = n = = 15 = 5.00 n 9 (b) n = 16, then = n = 15 = = 3.75 n 16 (c) n = 5, then = n = 15 = = 3.00 n 5 So as sample size increases, the standard error of the mean will decrease. An Additional Example Calculation of : Suppose we have the following SAT scores: 450, 375, 550, 675, 300, 330. We know that SAT is suppose to have a population standard deviation,, equal to 100. The mean for the sample,, is 446.67. What is the standard error of the mean for this data?

5 The formula is, so the standard error of the mean is n = 100 100 = = 40.8. 6. 45 Note that averages (e.g., means) have less variability than individual observations/scores. Thus, sampling distributions will be tighter or demonstrate less spread than will their corresponding population distributions. 7. Central Limit Theorem As explained earlier, the sampling distribution of will be normally distributed only if the population distribution of s is normally distributed. It is very important that the sampling distribution of the s be normally distributed so we can calculate probabilities of the for inferential purposes. Fortunately, the central limit theorem is invoked for population distributions that are not normally distributed. The central limit theorem states: If one is sampling independently from a population that has mean of and variance of, then as the sample size n approaches infinity, the sampling distribution of the sample mean approaches normality without regard to the shape of the sampled population. In practice, this means that the larger the sample size used, the better the sampling distribution of M will approximate a normal distribution. Note, however, that the Central Limit Theorem applies only the. (Question: Which is it, the number of samples or the size of the sample which enables us to say that the sampling distribution of approaches normality?) An illustration of the central limit theorem appears in the document entitled Central Limit Theorem Illustrated : http://www.bwgriffin.com/gsu/courses/edur8131/content/edur_8131_clt_illustrated_one_page.pdf Why is it important that the sampling distribution of the mean be normally distributed? Later we will see that z scores can be calculated for sample means, and these z scores will aid in making inferential statements. 8. Confidence Intervals(CI) Recall the discussion of interval estimates. CIs represent the method used to develop interval estimates. Since M is only a point estimate of, we know that M varies from sample to sample. The CI will give us a better look at the possible values that could represent.

6 Some logic: (a) From studying z scores we know that about 95% of all scores in a normal distribution lie between ± 1.96 standard deviations from the mean. If, due to the central limit theorem, we know that the sampling distribution of is approximately normal in shape, then we know that the probability is about.95 that will be within ± 1.96 of the population mean. (b) To say that is within ± 1.96 of is that same as to say is within ± 1.96 of. (c) So 95% of all samples will capture the true in the interval of 1.96 to + 1.96 For example, suppose we randomly select n = 5 grade 8 students and collect their IQ scores with = 103. Since, in theory, = 15 for the population distribution of IQ scores, then the standard error for this mean is = 15 15 = = 3. 5 5 We can anticipate 95% of all such selected samples will produce a.95 confidence interval that will capture the population value of. For this particular example, we are 95% confident that lies within the interval ± 1.96, 1.96 to + 1.96 103 1.96 (3) to 103 + 1.96 (3) 103 5.88 to 103 + 5.88 97.1 to 108.88. We are 95% confident that the unknown population mean of IQ for all grade 8 students lies between 97.1 and 108.88. Only 5% of randomly selected samples will not provide a.95 confidence interval that captures. In constructing the CI, we need also to consider the degree of accuracy of our interval estimate. The most common degrees of accuracy used in interval estimation are.95 and.99.

7 As illustrated above, a.95 CI is calculated as:.95ci = 1.96 As a second example, the SAT scores presented earlier demonstrated = 446.67 with = 100. The standard error of the mean was = 40.8, so the 95% CI is.95ci = 1.96.95CI = 446.67 (1.96)(40.8).95CI = 446.67 80.01 The limits are: lower limit (LL) is 446.67-80.01 = 366.66 upper limit (UL) is 446.67 + 80.01 = 56.68 This CI can be interpreted as follows: We are 95% confident that the true population mean lies somewhere within the interval of 366.66 to 56.68. A 99% CI can be calculated using the following formula:.99ci =.576 (or, to keep it simpler, round.576 to.58) Confirm that a value of.58 produces approximately 99% of observations about the mean in a normal distribution using the Z table. What would be the corresponding value if we sought a 90% CI? The CI constructed above only hold when is known. When the population SD is not known, other formulas, presented later, can be used. it is important to note that CIs vary from sample to sample; what does not vary is the parameter if.95cis were constructed for 100 SRS, then on average the population parameter would lie within 95 of the CIs, and would not be within 5 of the CIs

8 9. Margin of Error All CI are constructed from three components: (a) An estimate of a population parameter (such as ) (b) Standard error of the estimate (e.g., ) (c) And a test statistics (e.g., z = 1.96), together, then we have this example: ± 1.96 where (1.96 ) is known as the margin of error. If we believe the margin of error is too large, thus providing a large CI, we can reduce it by using a lower level of confidence (say from.99 to.95 reduces the test statistics from.576 to 1.96) reduce (which is unlikely to be something we can adjust easily) increase sample size (larger n will reduce the standard error ) 10. Properties of Estimators (a) Unbiasedness An estimator of a parameter is said to be unbiased if the expected value of the estimator equals the parameter, i.e., E( ) = = In other words, if the mean of the sampling distribution of the statistic equals the parameter, then the statistic is unbiased. The term unbiased applies to all statistics, but not all statistics are unbiased. For example, E(s) =/ or s =/ This means that the expected value of the sample standard deviation (i.e., the mean of the sampling distribution of s) does not equal the population standard deviation,. Positive bias occurs when the expected value of a statistic is greater than the parameter, i.e., E(Mdn) > Negative bias results when the expected value of a statistic is less than the parameter, i.e., E(r) <

9 (b) Consistency An estimator is said to be consistent if, as the sample size increases, the expected value of the statistic approaches (get closer) to the parameter. As n, E(s) (c) Relative Efficiency Efficiency refers to the precision of an estimator; the variability of an estimator from sample to sample; the degree of sampling error associated with an estimator. The variance error and standard error of statistics are measures of efficiency. Efficiency is usually thought to be more important than either the biasness or consistency of an estimator. For example, suppose we wish to determine which of two unbiased, equally consistent estimators is better to use, such as M and Mdn for a normally distributed population of IQ scores. the standard error of the mean is = the standard error of the median is Mdn = 3 Since is smaller than Mdn ( < Mdn), the mean is declared to be more efficient than the median for estimating. Note that the assessment of efficiency is relative since it requires that the variance errors of the sampling distributions of two or more statistics be compared.