Unit5: Inferenceforcategoricaldata. 4. MT2 Review. Sta Fall Duke University, Department of Statistical Science

Similar documents
Annoucements. MT2 - Review. one variable. two variables

Announcements. Final Review: Units 1-7

Announcements. Final exam, Saturday 9AM to Noon, usual classroom cheat sheet (1 page, front&back) + calculator

STA 101 Final Review

2. Outliers and inference for regression

FinalExamReview. Sta Fall Provided: Z, t and χ 2 tables

Unit4: Inferencefornumericaldata. 1. Inference using the t-distribution. Sta Fall Duke University, Department of Statistical Science

Lecture 11 - Tests of Proportions

STATISTICS 141 Final Review

Lecture 9 Two-Sample Test. Fall 2013 Prof. Yao Xie, H. Milton Stewart School of Industrial Systems & Engineering Georgia Tech

Announcements. Unit 3: Foundations for inference Lecture 3: Decision errors, significance levels, sample size, and power.

Announcements. Statistics 104 (Mine C etinkaya-rundel) Diamonds. From last time.. = 22

Exam details. Final Review Session. Things to Review

Nicole Dalzell. July 2, 2014

Lecture 30. DATA 8 Summer Regression Inference

Dover- Sherborn High School Mathematics Curriculum Probability and Statistics

The t-distribution. Patrick Breheny. October 13. z tests The χ 2 -distribution The t-distribution Summary

Announcements. Unit 4: Inference for numerical variables Lecture 4: ANOVA. Data. Statistics 104

Review of Statistics 101

Course Review. Kin 304W Week 14: April 9, 2013

Nature vs. nurture? Lecture 18 - Regression: Inference, Outliers, and Intervals. Regression Output. Conditions for inference.

Occupy movement - Duke edition. Lecture 14: Large sample inference for proportions. Exploratory analysis. Another poll on the movement

Lecture 19: Inference for SLR & Transformations

Subject CS1 Actuarial Statistics 1 Core Principles

MidtermII Review. Sta Fall Office Hours Wednesday 12:30-2:30pm Watch linear regression videos before lab on Thursday

Tables Table A Table B Table C Table D Table E 675

Sociology 6Z03 Review II

Wolf River. Lecture 19 - ANOVA. Exploratory analysis. Wolf River - Data. Sta 111. June 11, 2014

Stat 412/512 REVIEW OF SIMPLE LINEAR REGRESSION. Jan Charlotte Wickham. stat512.cwick.co.nz

An introduction to biostatistics: part 1

AMS7: WEEK 7. CLASS 1. More on Hypothesis Testing Monday May 11th, 2015

Ch. 1: Data and Distributions

Business Statistics. Lecture 10: Course Review

Introduction to Statistics

:the actual population proportion are equal to the hypothesized sample proportions 2. H a

Difference in two or more average scores in different groups

DETAILED CONTENTS PART I INTRODUCTION AND DESCRIPTIVE STATISTICS. 1. Introduction to Statistics

Confidence Intervals, Testing and ANOVA Summary

Analysis of Variance. Contents. 1 Analysis of Variance. 1.1 Review. Anthony Tanbakuchi Department of Mathematics Pima Community College

Can you tell the relationship between students SAT scores and their college grades?

Class 24. Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science. Marquette University MATH 1700

SEVERAL μs AND MEDIANS: MORE ISSUES. Business Statistics

Review for Final. Chapter 1 Type of studies: anecdotal, observational, experimental Random sampling

Tests of Linear Restrictions

STA441: Spring Multiple Regression. This slide show is a free open source document. See the last slide for copyright information.

" M A #M B. Standard deviation of the population (Greek lowercase letter sigma) σ 2

PSYC 331 STATISTICS FOR PSYCHOLOGISTS

Example - Alfalfa (11.6.1) Lecture 16 - ANOVA cont. Alfalfa Hypotheses. Treatment Effect

Lecture 20: Multiple linear regression

HYPOTHESIS TESTING II TESTS ON MEANS. Sorana D. Bolboacă

Inference for Regression

Announcements. Lecture 5: Probability. Dangling threads from last week: Mean vs. median. Dangling threads from last week: Sampling bias

Chapter 16: Understanding Relationships Numerical Data

Introduction to Business Statistics QM 220 Chapter 12

UNIVERSITY OF TORONTO. Faculty of Arts and Science APRIL - MAY 2005 EXAMINATIONS STA 248 H1S. Duration - 3 hours. Aids Allowed: Calculator

Inference for the Regression Coefficient

Sampling Distributions: Central Limit Theorem

One-sample categorical data: approximate inference

Two-Sample Inference for Proportions and Inference for Linear Regression

Announcements. Lecture 18: Simple Linear Regression. Poverty vs. HS graduate rate

y ˆ i = ˆ " T u i ( i th fitted value or i th fit)

Contents. Acknowledgments. xix

Statistical methods for comparing multiple groups. Lecture 7: ANOVA. ANOVA: Definition. ANOVA: Concepts

7 Estimation. 7.1 Population and Sample (P.91-92)

Unit 7: Multiple linear regression 1. Introduction to multiple linear regression

BIOS 6222: Biostatistics II. Outline. Course Presentation. Course Presentation. Review of Basic Concepts. Why Nonparametrics.

Wolf River. Lecture 15 - ANOVA. Exploratory analysis. Wolf River - Data. Sta102 / BME102. October 22, 2014

MGEC11H3Y L01 Introduction to Regression Analysis Term Test Friday July 5, PM Instructor: Victor Yu

Lecture 5: ANOVA and Correlation

Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model

Stat 311: HW 9, due Th 5/27/10 in your Quiz Section

10.2: The Chi Square Test for Goodness of Fit

Harvard University. Rigorous Research in Engineering Education

Business Statistics: Lecture 8: Introduction to Estimation & Hypothesis Testing

Chapter 12 - Lecture 2 Inferences about regression coefficient

The goodness-of-fit test Having discussed how to make comparisons between two proportions, we now consider comparisons of multiple proportions.

Six Sigma Black Belt Study Guides

Wolf River. Lecture 15 - ANOVA. Exploratory analysis. Wolf River - Data. Sta102 / BME102. October 26, 2015

This document contains 3 sets of practice problems.

Hypothesis testing: Steps

ANALYTICAL COMPARISONS AMONG TREATMENT MEANS (CHAPTER 4)

ANOVA: Analysis of Variation

Bootstrapping, Permutations, and Monte Carlo Testing

Analysis of Covariance. The following example illustrates a case where the covariate is affected by the treatments.

Statistics Introductory Correlation

Lab #12: Exam 3 Review Key

Multiple t Tests. Introduction to Analysis of Variance. Experiments with More than 2 Conditions

Statistical Analysis for QBIC Genetics Adapted by Ellen G. Dow 2017

Simple Linear Regression: One Qualitative IV

Simple Linear Regression: One Quantitative IV

Project proposal feedback. Unit 4: Inference for numerical variables Lecture 3: t-distribution. Statistics 104

Math Review Sheet, Fall 2008

Extra Exam Empirical Methods VU University Amsterdam, Faculty of Exact Sciences , July 2, 2015

Example - Alfalfa (11.6.1) Lecture 14 - ANOVA cont. Alfalfa Hypotheses. Treatment Effect

Degrees of freedom df=1. Limitations OR in SPSS LIM: Knowing σ and µ is unlikely in large

Announcements. Unit 7: Multiple linear regression Lecture 3: Confidence and prediction intervals + Transformations. Uncertainty of predictions

STAT 4385 Topic 01: Introduction & Review

CBA4 is live in practice mode this week exam mode from Saturday!

Hypothesis testing: Steps

Business Analytics and Data Mining Modeling Using R Prof. Gaurav Dixit Department of Management Studies Indian Institute of Technology, Roorkee

Transcription:

Unit5: Inferenceforcategoricaldata 4. MT2 Review Sta 101 - Fall 2015 Duke University, Department of Statistical Science Dr. Çetinkaya-Rundel Slides posted at http://bit.ly/sta101_f15

Outline 1. Housekeeping 2. Map of concepts 3. Review exercises 4. Summary of methods

Announcements MT 2 next week Bring a calculator + cheat sheet + writing utensil Tables will be provided MT 2 review session: Sat, Nov 7, 4-5pm, Old Chem 116 + office hours as usual: https://stat.duke.edu/courses/fall15/sta101.002/info/#oh + extra office hours from Dr. Monod: Friday, 1:30-3pm (Old Chem 122A) MT 2 review materials posted on the course website Project 1 due Friday evening (+ work on it in lab on Thursday) PS 5 due Friday evening, PA 5 due Saturday evening (note day change to allow for review before midterm) 1

Outline 1. Housekeeping 2. Map of concepts 3. Review exercises 4. Summary of methods

inference HT CI theoretical Z, T, F, chi-sq Z, T simulation bootstrap centered at null randomization bootstrap 2

one variable numerical categorical 2 levels 3+ levels mean median proportion proportion H0: mu = mu_0 T bootstrap centered at mu_0 H0: median = med_0 randomization H0: p = p_0 H0: p follows hypothesized distribution S-F check -> Z S-F fail -> simulation centered at p_0 E >= 5 -> chi-sq GOF E < 5 -> randomization HT two variables y = numerical x = categorical y = categorical x = categorical x: 2 levels x: 3+ levels x&y: 2 levels mean median mean H0: mu_1 = mu_2 H0: p_1 = p_2 T randomization H0: med_1 = med_2 randomization H0: All mu_i are equal ANOVA, F randomization median H0: All med_i are equal randomization proportion S-F check -> Z S-F fail -> randomization x/y: 3+ levels proportion H0: x and y are independent E >= 5 -> chi-sq independence E < 5 -> randomization 3

one variable numerical categorical 2 levels mean median proportion mu median 3+ levels proportion N/A p T bootstrap bootstrap S-F check -> Z S-F fail -> bootstrap CI x: 2 levels mean mu_1 - mu_2 T bootstrap two variables y = numerical x = categorical y = categorical x = categorical x: 3+ levels x&y: 2 levels median med_1 - med_2 bootstrap mean N/A median N/A proportion p_1 - p_2 S-F check -> Z S-F fail -> bootstrap x/y: 3+ levels proportion N/A 4

Outline 1. Housekeeping 2. Map of concepts 3. Review exercises 4. Summary of methods

Bayesian inference Clicker question Which of the following is true? Design of studies Frequentist inference (CLT & simulation) Exploratory numerical data Inference analysis one mean & median two means & medians many means Probability categorical one proportion two proportions many proportions Modeling (numerical response) 1 explanatory many explanatory (a) If the sample size is large enough, conclusions can be generalized to the population. (b) If subjects are randomly assigned to treatments, conclusions can be generalized to the population. (c) Blocking in experiments serves a similar purpose as stratifying in observational studies. (d) Representative samples allow us to make causal conclusions. (e) Statistical inference requires normal distribution of the response variable. 5

Bayesian inference Clicker question Which of the following is true? Design of studies Frequentist inference (CLT & simulation) Exploratory numerical data Inference analysis one mean & median two means & medians many means Probability categorical one proportion two proportions many proportions Modeling (numerical response) 1 explanatory many explanatory (a) If the sample size is large enough, conclusions can be generalized to the population. (b) If subjects are randomly assigned to treatments, conclusions can be generalized to the population. (c) Blockinginexperimentsservesasimilarpurposeasstratifyingin observationalstudies. (d) Representative samples allow us to make causal conclusions. (e) Statistical inference requires normal distribution of the response variable. 5

Bayesian inference Clicker question Which of the following is the best visualization for evaluating the relationship between two categorical variables? Design of studies Frequentist inference (CLT & simulation) Exploratory numerical data Inference analysis one mean & median two means & medians many means Probability categorical one proportion two proportions many proportions Modeling (numerical response) 1 explanatory many explanatory (a) side-by-side box plots (b) mosaic plot (c) pie chart (d) segmented frequency bar plot (e) relative frequency histogram 6

Bayesian inference Clicker question Which of the following is the best visualization for evaluating the relationship between two categorical variables? Design of studies Frequentist inference (CLT & simulation) Exploratory numerical data Inference analysis one mean & median two means & medians many means Probability categorical one proportion two proportions many proportions Modeling (numerical response) 1 explanatory many explanatory (a) side-by-side box plots (b) mosaicplot (c) pie chart (d) segmented frequency bar plot (e) relative frequency histogram 6

Clicker question Two students in an introductory statistics class choose to conduct similar studies estimating the proportion of smokers at their school. Student A collects data from 100 students, and student B collects data from 50 students. How will the standard errors used by the two students compare? Assume both are simple random samples. Design of studies Exploratory data analysis Probability 1 explanatory (a) SE used by Student A < SE used as Student B. (b) SE used by Student A > SE used as Student B. (c) SE used by Student A = SE used as Student B. (d) SE used by Student A SE used as Student B. Bayesian inference Inference Modeling (numerical response) many explanatory Frequentist inference (CLT & simulation) (e) Cannot tell without knowing the true proportion of smokers at this school. numerical one mean & median two means & medians many means categorical one proportion two proportions many proportions 7

Clicker question Two students in an introductory statistics class choose to conduct similar studies estimating the proportion of smokers at their school. Student A collects data from 100 students, and student B collects data from 50 students. How will the standard errors used by the two students compare? Assume both are simple random samples. Design of studies Exploratory data analysis Probability 1 explanatory (a) SE usedbystudenta < SE usedasstudentb. (b) SE used by Student A > SE used as Student B. (c) SE used by Student A = SE used as Student B. (d) SE used by Student A SE used as Student B. Bayesian inference Inference Modeling (numerical response) many explanatory Frequentist inference (CLT & simulation) (e) Cannot tell without knowing the true proportion of smokers at this school. numerical one mean & median two means & medians many means categorical one proportion two proportions many proportions 7

Bayesian inference Clicker question Which of the following is the best method for evaluating the relationship between two categorical variables? Design of studies Frequentist inference (CLT & simulation) Exploratory numerical data Inference analysis one mean & median two means & medians many means Probability categorical one proportion two proportions many proportions Modeling (numerical response) (a) chi-square test of independence (b) chi-square test of goodness of fit (c) anova (d) t-test 1 explanatory many explanatory 8

Bayesian inference Clicker question Which of the following is the best method for evaluating the relationship between two categorical variables? Design of studies Frequentist inference (CLT & simulation) Exploratory numerical data Inference analysis one mean & median two means & medians many means Probability categorical one proportion two proportions many proportions Modeling (numerical response) (a) chi-squaretestofindependence (b) chi-square test of goodness of fit (c) anova (d) t-test 1 explanatory many explanatory 8

Bayesian inference Clicker question Which of the following is the best method for evaluating the relationship between a numerical and a categorical variable with many levels? Design of studies Frequentist inference (CLT & simulation) Exploratory numerical data Inference analysis one mean & median two means & medians many means Probability categorical one proportion two proportions many proportions Modeling (numerical response) (a) z-test (b) chi-square test of goodness of fit (c) anova (d) t-test 1 explanatory many explanatory 9

Bayesian inference Clicker question Which of the following is the best method for evaluating the relationship between a numerical and a categorical variable with many levels? Design of studies Frequentist inference (CLT & simulation) Exploratory numerical data Inference analysis one mean & median two means & medians many means Probability categorical one proportion two proportions many proportions Modeling (numerical response) (a) z-test (b) chi-square test of goodness of fit (c) anova (d) t-test 1 explanatory many explanatory 9

Bayesian inference Data are collected at a bank on 6 tellers randomly sampled transactions. Do average transaction times vary by teller? Design of studies Frequentist inference (CLT & simulation) Exploratory numerical data Inference analysis one mean & median two means & medians many means Probability categorical one proportion two proportions many proportions Modeling (numerical response) 1 explanatory many explanatory Response variable: numerical, Explanatory variable: categorical ANOVA Summary statistics: n_1 = 14, mean_1 = 65.7857, sd_1 = 15.2249 n_2 = 23, mean_2 = 79.9174, sd_2 = 23.284 n_3 = 15, mean_3 = 82.66, sd_3 = 18.1842 n_4 = 15, mean_4 = 77.9933, sd_4 = 23.2754 n_5 = 44, mean_5 = 81.7295, sd_5 = 21.5768 n_6 = 29, mean_6 = 75.3069, sd_6 = 20.4814 H_0: All means are equal. H_A: At least one mean is different. Analysis of Variance Table Response: data Df Sum Sq Mean Sq F value Pr(>F) group 5 3315 663.06 1.508 0.1914 Residuals 134 58919 439.69 10

20 40 60 80 100 120 1 2 3 4 5 6 1.508 11

Activity: Data are collected on download times at three different times during the day. We want to evaluate whether average download times vary by time of day. Fill in the??s in the ANOVA output below. Bayesian inference Design of studies Frequentist inference (CLT & simulation) Exploratory numerical data Inference analysis one mean & median two means & medians many means Probability categorical one proportion two proportions many proportions Modeling (numerical response) 1 explanatory many explanatory Response variable: numerical, Explanatory variable: categorical Summary statistics: n_early (7AM) = 16, mean_early (7AM) = 113.375, sd_early (7AM) = 47.6541 n_eve (5 PM) = 16, mean_eve (5 PM) = 273.3125, sd_eve (5 PM) = 52.1929 n_late (12 AM) = 16, mean_late (12 AM) = 193.0625, sd_late (12 AM) = 40.9023 Analysis of Variance Table Response: data Df Sum Sq Mean Sq F value Pr(>F) group???????? 1.306e-11 Residuals?? 100020?? Total?? 304661 12

What is the result of the ANOVA? 100 150 200 250 300 350 Early (7AM) Evening (5 PM) Late Night (12 AM) 46.0349 13

What is the result of the ANOVA? 100 150 200 250 300 350 Early (7AM) Evening (5 PM) Late Night (12 AM) 46.0349 Since 1.306e-11 < 0.05, we reject the null hypothesis. The data provide convincing evidence that the average download time is different for at least one pair of times of day. 13

Activity: The next step is to evaluate the pairwise tests. There are 3 pairs of times of day 1. Early vs. Evening: left side of class (facing the board) 2. Evening vs. Late Night: center of class 3. Early vs. Late Night: right side of class Determine the appropriate significance level for these tests, and then complete the test assigned to your team. 14

α = 0.05/3 = 0.0167 15

α = 0.05/3 = 0.0167 (1) Early vs. Evening 15

α = 0.05/3 = 0.0167 (1) Early vs. Evening T 45 = 113.375 273.3125 2223 16 + 2223 16 15

α = 0.05/3 = 0.0167 (1) Early vs. Evening 113.375 273.3125 T 45 = 2223 16 + 2223 16 = 159.9375 = 9.59 16.67 15

α = 0.05/3 = 0.0167 (1) Early vs. Evening 113.375 273.3125 T 45 = 2223 16 + 2223 16 = 159.9375 = 9.59 16.67 p val < 0.01 15

α = 0.05/3 = 0.0167 (1) Early vs. Evening (2) Evening vs. Late Night 113.375 273.3125 T 45 = 2223 16 + 2223 16 = 159.9375 = 9.59 16.67 p val < 0.01 15

α = 0.05/3 = 0.0167 (1) Early vs. Evening (2) Evening vs. Late Night 113.375 273.3125 T 45 = 2223 16 + 2223 16 = 159.9375 = 9.59 16.67 p val < 0.01 T 45 = 113.375 193.0625 2223 16 + 2223 16 15

α = 0.05/3 = 0.0167 (1) Early vs. Evening (2) Evening vs. Late Night 113.375 273.3125 T 45 = 2223 16 + 2223 16 = 159.9375 = 9.59 16.67 p val < 0.01 T 45 = 113.375 193.0625 2223 16 + 2223 16 = 79.6875 = 4.78 16.67 15

α = 0.05/3 = 0.0167 (1) Early vs. Evening (2) Evening vs. Late Night 113.375 273.3125 T 45 = 2223 16 + 2223 16 = 159.9375 = 9.59 16.67 p val < 0.01 T 45 = 113.375 193.0625 2223 16 + 2223 16 = 79.6875 = 4.78 16.67 p val < 0.01 15

α = 0.05/3 = 0.0167 (1) Early vs. Evening (2) Evening vs. Late Night (3) Early vs. Late Night 113.375 273.3125 T 45 = 2223 16 + 2223 16 = 159.9375 = 9.59 16.67 p val < 0.01 T 45 = 113.375 193.0625 2223 16 + 2223 16 = 79.6875 = 4.78 16.67 p val < 0.01 15

α = 0.05/3 = 0.0167 (1) Early vs. Evening (2) Evening vs. Late Night (3) Early vs. Late Night 113.375 273.3125 T 45 = 2223 16 + 2223 16 = 159.9375 = 9.59 16.67 p val < 0.01 T 45 = 113.375 193.0625 2223 16 + 2223 16 = 79.6875 = 4.78 16.67 p val < 0.01 T 45 = 273.3125 193.0625 2223 16 + 2223 16 15

α = 0.05/3 = 0.0167 (1) Early vs. Evening (2) Evening vs. Late Night (3) Early vs. Late Night 113.375 273.3125 T 45 = 2223 16 + 2223 16 = 159.9375 = 9.59 16.67 p val < 0.01 T 45 = 113.375 193.0625 2223 16 + 2223 16 = 79.6875 = 4.78 16.67 p val < 0.01 273.3125 193.0625 T 45 = 2223 16 + 2223 16 = 80.25 16.67 = 4.81 15

α = 0.05/3 = 0.0167 (1) Early vs. Evening (2) Evening vs. Late Night (3) Early vs. Late Night 113.375 273.3125 T 45 = 2223 16 + 2223 16 = 159.9375 = 9.59 16.67 p val < 0.01 T 45 = 113.375 193.0625 2223 16 + 2223 16 = 79.6875 = 4.78 16.67 p val < 0.01 273.3125 193.0625 T 45 = 2223 16 + 2223 16 = 80.25 16.67 = 4.81 p val < 0.01 15

Bayesian inference Clicker question What percent of variability in download times is explained by time of day? Design of studies Frequentist inference (CLT & simulation) Exploratory numerical data Inference analysis one mean & median two means & medians many means Probability categorical one proportion two proportions many proportions Modeling (numerical response) 1 explanatory many explanatory Response: data Df Sum Sq Mean Sq F value Pr(>F) group 2 204641 102320 46.035 1.306e-11 Residuals 45 100020 2223 (a) 204641 204641+100020 (b) 204641 100020 (c) 100020 204641 (d) 102320 102320+2223 16

Bayesian inference Clicker question What percent of variability in download times is explained by time of day? Design of studies Frequentist inference (CLT & simulation) Exploratory numerical data Inference analysis one mean & median two means & medians many means Probability categorical one proportion two proportions many proportions Modeling (numerical response) 1 explanatory many explanatory Response: data Df Sum Sq Mean Sq F value Pr(>F) group 2 204641 102320 46.035 1.306e-11 Residuals 45 100020 2223 (a) 204641 204641+100020 = 0.67 (b) 204641 100020 (c) 100020 204641 (d) 102320 102320+2223 16

Clicker question n = 50 and ˆp = 0.80. Hypotheses: H 0 : p = 0.82; H A : p 0.82. We use a randomization test because the sample size isn t large enough for ˆp to be distributed nearly normally (50 0.82 = 41 < 10; 50 0.18 = 9 < 10). Which of the following is the correct set up for this hypothesis test? Red: success, blue: failure, ˆp sim = proportion of reds in simulated samples. Design of studies Exploratory data analysis Probability 1 explanatory Bayesian inference Inference Modeling (numerical response) many explanatory Frequentist inference (CLT & simulation) numerical one mean & median two means & medians many means (a) Place 80 red and 20 blue chips in a bag. Sample, with replacement, 50 chips and calculate the proportion of reds. Repeat this many times and calculate the proportion of simulations where ˆp sim 0.82. (b) Place 82 red and 18 blue chips in a bag. Sample, without replacement, 50 chips and calculate the proportion of reds. Repeat this many times and calculate the proportion of simulations where ˆp sim 0.80. (c) Place 82 red and 18 blue chips in a bag. Sample, with replacement, 50 chips and calculate the proportion of reds. Repeat this many times and calculate the proportion of simulations where ˆp sim 0.80 or ˆp sim 0.84. categorical one proportion two proportions many proportions (d) Place 82 red and 18 blue chips in a bag. Sample, with replacement, 100 chips and calculate the proportion of reds. Repeat this many times and calculate the proportion of simulations where ˆp sim 0.80 or 17

Clicker question n = 50 and ˆp = 0.80. Hypotheses: H 0 : p = 0.82; H A : p 0.82. We use a randomization test because the sample size isn t large enough for ˆp to be distributed nearly normally (50 0.82 = 41 < 10; 50 0.18 = 9 < 10). Which of the following is the correct set up for this hypothesis test? Red: success, blue: failure, ˆp sim = proportion of reds in simulated samples. Design of studies Exploratory data analysis Probability 1 explanatory Bayesian inference Inference Modeling (numerical response) many explanatory Frequentist inference (CLT & simulation) numerical one mean & median two means & medians many means (a) Place 80 red and 20 blue chips in a bag. Sample, with replacement, 50 chips and calculate the proportion of reds. Repeat this many times and calculate the proportion of simulations where ˆp sim 0.82. (b) Place 82 red and 18 blue chips in a bag. Sample, without replacement, 50 chips and calculate the proportion of reds. Repeat this many times and calculate the proportion of simulations where ˆp sim 0.80. (c) Place82redand18bluechipsinabag. Sample, with replacement, 50chipsandcalculatetheproportionofreds. Repeatthismany timesandcalculatetheproportionofsimulationswhere ˆp sim 0.80 or ˆp sim 0.84. categorical one proportion two proportions many proportions (d) Place 82 red and 18 blue chips in a bag. Sample, with replacement, 100 chips and calculate the proportion of reds. Repeat this many times and calculate the proportion of simulations where ˆp sim 0.80 or 17

What is / should be the center of the randomization distribution? What is the result of the hypothesis test? observed 0.8 0.70 0.75 0.80 0.85 0.90 randomization statistic 18

Outline 1. Housekeeping 2. Map of concepts 3. Review exercises 4. Summary of methods

Inference for numerical data: 19

Inference for numerical data: One numerical: Parameter of interest: µ T HT and CI 19

Inference for numerical data: One numerical: Parameter of interest: µ T HT and CI One numerical vs. one categorical (with 2 levels): Parameter of interest: µ 1 µ 2 T HT and CI 19

Inference for numerical data: One numerical: Parameter of interest: µ T HT and CI One numerical vs. one categorical (with 2 levels): Parameter of interest: µ 1 µ 2 T HT and CI If samples are dependent (paired), first find differences between paired observations 19

Inference for numerical data: One numerical: Parameter of interest: µ T HT and CI One numerical vs. one categorical (with 2 levels): Parameter of interest: µ 1 µ 2 T HT and CI If samples are dependent (paired), first find differences between paired observations One numerical vs. one categorical (with 3+ levels) - mean: Parameter of interest: N/A ANOVA HT only 19

Inference for numerical data: One numerical: Parameter of interest: µ T HT and CI One numerical vs. one categorical (with 2 levels): Parameter of interest: µ 1 µ 2 T HT and CI If samples are dependent (paired), first find differences between paired observations One numerical vs. one categorical (with 3+ levels) - mean: Parameter of interest: N/A ANOVA HT only For all other parameters of interest: simulation 19

Inference for categorical data - binary outcome: Binaryoutcome: 20

Binaryoutcome: One categorical: Inference for categorical data - binary outcome: Parameter of interest: p S/F condition met Z, if not simulation HT and CI 20

Binaryoutcome: One categorical: Inference for categorical data - binary outcome: Parameter of interest: p S/F condition met Z, if not simulation HT and CI One categorical vs. one categorical, each with only 2 outcomes: Parameter of interest: p 1 p 2 S/F condition met Z, if not simulation HT and CI 20

Binaryoutcome: One categorical: Inference for categorical data - binary outcome: Parameter of interest: p S/F condition met Z, if not simulation HT and CI One categorical vs. one categorical, each with only 2 outcomes: Parameter of interest: p 1 p 2 S/F condition met Z, if not simulation HT and CI S/F: use observed S and F for CIs and expexted for HT 20

Inference for categorical data - 3+ outcomes: 3+ outcomes: 21

3+ outcomes: Inference for categorical data - 3+ outcomes: One categorical, compared to hypothetical distribution: Parameter of interest: N/A At least 5 expected successes in each cell χ 2 GOF, if not simulation HT only 21

3+ outcomes: Inference for categorical data - 3+ outcomes: One categorical, compared to hypothetical distribution: Parameter of interest: N/A At least 5 expected successes in each cell χ 2 GOF, if not simulation HT only One categorical vs. one categorical, either with 3+ outcomes: Parameter of interest: N/A At least 5 expected successes in each cell χ 2 Independence, if not simulation HT only 21