Stat 135, Fall 2006 A. Adhikari HOMEWORK 6 SOLUTIONS

Similar documents
STAT 135 Lab 5 Bootstrapping and Hypothesis Testing

TUTORIAL 8 SOLUTIONS #

STAT 830 Hypothesis Testing

Institute of Actuaries of India

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

1 Hypothesis testing for a single mean

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80

STAT 830 Hypothesis Testing

6.4 Type I and Type II Errors

Master s Written Examination - Solution

A Very Brief Summary of Statistical Inference, and Examples

STAT 801: Mathematical Statistics. Hypothesis Testing

Performance Evaluation and Comparison

Lecture 10: Generalized likelihood ratio test

Mathematical Statistics

Math 181B Homework 1 Solution

Review: General Approach to Hypothesis Testing. 1. Define the research question and formulate the appropriate null and alternative hypotheses.

1 Statistical inference for a population mean

Visual interpretation with normal approximation

Homework 7: Solutions. P3.1 from Lehmann, Romano, Testing Statistical Hypotheses.

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

Topic 19 Extensions on the Likelihood Ratio

Hypothesis testing (cont d)

Null Hypothesis Significance Testing p-values, significance level, power, t-tests Spring 2017

Asymptotic Statistics-VI. Changliang Zou

Topic 15: Simple Hypotheses

Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

STAT 461/561- Assignments, Year 2015

Lecture Testing Hypotheses: The Neyman-Pearson Paradigm

UNIVERSITY OF TORONTO Faculty of Arts and Science

Hypothesis Testing Chap 10p460

Loglikelihood and Confidence Intervals

Stat 5102 Final Exam May 14, 2015

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

Inference for Single Proportions and Means T.Scofield

Partitioning the Parameter Space. Topic 18 Composite Hypotheses

Master s Written Examination


Chapter 6. Hypothesis Tests Lecture 20: UMP tests and Neyman-Pearson lemma

Introduction 1. STA442/2101 Fall See last slide for copyright information. 1 / 33

Statistics 135 Fall 2008 Final Exam

z and t tests for the mean of a normal distribution Confidence intervals for the mean Binomial tests

Normal (Gaussian) distribution The normal distribution is often relevant because of the Central Limit Theorem (CLT):

Composite Hypotheses. Topic Partitioning the Parameter Space The Power Function

F79SM STATISTICAL METHODS

Chapters 10. Hypothesis Testing

Central Limit Theorem ( 5.3)

Summary of Chapters 7-9

Stat 5421 Lecture Notes Fuzzy P-Values and Confidence Intervals Charles J. Geyer March 12, Discreteness versus Hypothesis Tests

STA 2101/442 Assignment 2 1

MAT 122 Homework 7 Solutions

Probability theory and inference statistics! Dr. Paola Grosso! SNE research group!! (preferred!)!!

i=1 X i/n i=1 (X i X) 2 /(n 1). Find the constant c so that the statistic c(x X n+1 )/S has a t-distribution. If n = 8, determine k such that

Difference between means - t-test /25

Chapter 23: Inferences About Means

f (1 0.5)/n Z =

557: MATHEMATICAL STATISTICS II HYPOTHESIS TESTING: EXAMPLES

Statistical Inference

M(t) = 1 t. (1 t), 6 M (0) = 20 P (95. X i 110) i=1

Math 494: Mathematical Statistics

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45

CHAPTER 8. Test Procedures is a rule, based on sample data, for deciding whether to reject H 0 and contains:

One-sample categorical data: approximate inference

Chapter 9: Hypothesis Testing Sections

HYPOTHESIS TESTING: FREQUENTIST APPROACH.

14.30 Introduction to Statistical Methods in Economics Spring 2009

BTRY 4090: Spring 2009 Theory of Statistics

CSE 312 Final Review: Section AA

Prof. Thistleton MAT 505 Introduction to Probability Lecture 13

ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.

Lecture 18: Analysis of variance: ANOVA

STAT 285: Fall Semester Final Examination Solutions

Math 475. Jimin Ding. August 29, Department of Mathematics Washington University in St. Louis jmding/math475/index.

Ch. 5 Hypothesis Testing

simple if it completely specifies the density of x

Lecture 9 Two-Sample Test. Fall 2013 Prof. Yao Xie, H. Milton Stewart School of Industrial Systems & Engineering Georgia Tech

Testing Simple Hypotheses R.L. Wolpert Institute of Statistics and Decision Sciences Duke University, Box Durham, NC 27708, USA

Chapters 10. Hypothesis Testing

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

Hypothesis testing: theory and methods

CHAPTER 9, 10. Similar to a courtroom trial. In trying a person for a crime, the jury needs to decide between one of two possibilities:

40.530: Statistics. Professor Chen Zehua. Singapore University of Design and Technology

An inferential procedure to use sample data to understand a population Procedures

Statistical Inference

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Distribution-Free Procedures (Devore Chapter Fifteen)

GEOMETRIC -discrete A discrete random variable R counts number of times needed before an event occurs

STAT FINAL EXAM

1. (Regular) Exponential Family

FYST17 Lecture 8 Statistics and hypothesis testing. Thanks to T. Petersen, S. Maschiocci, G. Cowan, L. Lyons

Introduction to Statistics

ST495: Survival Analysis: Hypothesis testing and confidence intervals

exp{ (x i) 2 i=1 n i=1 (x i a) 2 (x i ) 2 = exp{ i=1 n i=1 n 2ax i a 2 i=1

Chapter 24. Comparing Means

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing

This paper is not to be removed from the Examination Halls

Statistics 100A Homework 5 Solutions

Transcription:

Stat 135, Fall 2006 A. Adhikari HOMEWORK 6 SOLUTIONS 1a. Under the null hypothesis X has the binomial (100,.5) distribution with E(X) = 50 and SE(X) = 5. So P ( X 50 > 10) is (approximately) two tails of the standard normal beyond the points z = (39.5 50)/5 = 2.1 and z = (60.5 50)/5) = 2.1. That s 3.58%. I know the answer in the book is 4.56%, but that s a normal approximation without the continuity correction. You know better than that. R computes the exact binomial probability of rejection to be 3.52%, so 3.58% is a better approximation than 4.56%. b. The rejection region is: X greater than 60 or less than 40. The distribution of X is approximately normal with mean 100p and SE 100p(1 p). So for each fixed p the is going to be the sum of two normal tails, beyond the points 40.5 and 60.5. I used the following code. > p seq(.001,.999, by =.01) > pnorm(39.5, 100*p, sqrt(100*p*(1-p))) + 1 - pnorm(60.5, 100*p, sqrt(100*p*(1-p)) > plot(p,, type= l ) > lines(c(.001,.999), c(.0358,.0358)) The last command plots the horizontal line at the level α = 0.0358. It s clear that the converges to the level as p converges to 0.5, and that it converges to 1 as p converges to 0 or 1. The horizontal line is at the level alpha = 3.58%. 0.0 0.4 0.8 0.0 0.2 0.4 0.6 0.8 1.0 p 2. We derived the relevant two-tailed test in lecture. It is a two-tailed test which rejects when X is far from 0 on either side. When n = 25, the distribution of X is normal with mean µ and SE 10/ 25 = 2. Therefore under H 0, the distribution of X is normal with mean 0 and SE 2. If the significance level is α = 0.1 the critical z is 1.65, so the rejection region is X < 3.3 or X > 3.3. For a fixed value µ in the alternative, the is the total area to the left of -3.3 and to the right of 3.3, under the normal curve with mean µ and SE 2. I chose to plot for µ in the range ( 10, 10) because that s a few SEs on either side of 0. > seq(-10, 10, by=0.1) > pnorm(-3.3,, 2) + 1 - pnorm(3.3,, 2) If α = 0.05 the critical z is 1.96 so the rejection region is X < 3.92 or X > 3.92. When n = 100 the disribution of X is normal with mean µ and SE 1. For α = 0.1 the rejection region is X < 1.65 or X > 1.65. For α = 0.05 the rejection region is X < 1.96 or X > 1.96. The against any µ in the alternative is the chance of the rejection region computed using the normal curve with mean µ and SE 1.

n=25, p=0.1 n=25, p=0.05 n=100, p=0.1 n=100, p=0.05 You can see that for fixed n any fixed value of µ in the alternative, the is less when α is small. That makes sense because if you allow yourself less Type I error you should expect to make more Type II error and hence have smaller. It is also clear from the plots that for fixed α the test based on the larger sample is more ful. 3. Since the MLE of θ is 1/ X, the generalized likelihood ratio is ni=1 θ 0 e θ 0X i ni=1 1 X e X i/ X = [θ 0e Xe θ 0 X] n This is small when Xe θ 0 X is small. 4a. Use the result of the previous problem. The function g(x) = xe θ 0x is small when x is very large (because e θ 0x gets small at a ch faster rate than x gets large) and when x is very small (because e θ 0x gets close to 1). That explains the form of the rejection region. b. Under θ 0 = 1, the given chance is the significance level.

c. X i has the gamma (10, 1) distribution because it is the sum of 10 i.i.d. exponentials with rate 1. So X is 1/10 times a gamma random variable. Multiplying a gamma random variable by a constant yields another gamma, with a change only in the scale parameter. So X has the gamma (10, λ) density where λ can be found by equating means: 1 = E( X) = 10 λ so λ = 1. Use this knowledge and R to find the cutoffs x 0 and x 1. Those should be respectively the 2.5th and 97.5th percentiles of the gamma (10, 10) distribution, which are about 0.48 and 1.71 according to R. d. If you forget the theory, silate. Draw 1000 i.i.d. samples, each consisting of 10 i.i.d. exponential (1) observations. Compute the mean of each sample, and use as your cutoffs the 2.5th and 97.5th quantiles of the observed distribution of means. 5. Both the hypotheses are simple, so by the Neyman-Pearson Lemma the likelihood ratio test is the most ful. So the problem amounts to finding the of the likelihood ratio test. The likelihood ratio test rejects when 1/2X is small, that is, when X is large. The null distribution of X is uniform[0,1], so to achieve the level α = 0.1, the test st reject when X > 0.9. The is the chance of the rejection region under the alternative, so it is equal to 1 2xdx = 1 2 0.9 2 = 0.19 0.9 6a. Because the maxim likelihood estimate of p is X/n, the generalized likelihood ratio is ( n ) X 0.5 n ( n ) X (X/n) X (1 X/n) n X = 0.5 n (X/n) X (1 X/n) n X b. The numerator in the ratio above does not depend on X, so the ratio will be small when the denominator is small. The denominator is (X/n) X (1 X/n) n X = 1 n n XX (n X) n X This is large when g(x) = X X (n X) n X is large. The function g is a function on [0, n]. It is symmetric about n/2 because g(u) = g(n u). And g is smallest at n/2, rising symmetrically on either side. If you re not convinced, compute a few values, plot g, or find its minim. So X X (n X) n X is large when X is far from n/2, i.e., if X n/2 is large. c. The null distribution of X is binomial (n, 0.5) which is a symmetric distribution. The rejection region is going to be two equal tails of the binomial, with the area of each tail equal to half the given significance level. For large n the binomial can be approximated by the normal with mean n/2 and SE 0.5 n. d. The rejection region is: X < 3 or X > 7. The significance level is α = 0.109375, obtained by the R command > 2*pbinom(2, 10, 0.5) e. This is exactly the same problem as 1b above, where you got the normal approximation of α. Use pbinom to check that the exact level is 0.0352, as discussed in the answer to 1b. 7a. R says the sample means are 0.554625 and 1.62398, so those are my estimates. The estimate of the difference is 1.07, roughly. Yes, you can say it is 1.069355 or 1.069355 if you like.

b. The pooled estimate for the variance, 0.7666932. My lists were called x and y so I did > (3*var(x) + 4*var(y))/7 c. s p 1/4 + 1/5 = 0.587, roughly. d. Use the upper 5% point of the t 7 distribution, which is 1.895. The interval is 1.07 ± 1.895 0.587, which is (-0.0424, 2.18). e. I ll accept either answer. An argument for the one-sided test is that you re already told that the means may be different so it s the direction of the difference that s more interesting. An argument for the two-sided test is that you re told that the means may be the same, so you should try to find out whether they were the same or not. f. The t-statistic is (1.07 0)/.587 = 1.82, roughly. The two-sided p-value is 0.11, roughly. I did > 2*(1 - pt(1.82, 7)) g. No, but only just no. Because we re doing a 10%-level test, this is consistent with the fact that the 90%-confidence interval in d contains 0, but only just. h. The variance of the difference would be exactly 0.670824 (use the known σ instead of s p ) and you would use the normal curve instead of the t. There s no need to do it all again. 8. Under H 0, X Ȳ has the normal distribution with mean 0 and variance 10 2/n. The critical value is z = 1.3, so the rejection region is X > 13 2/n. Under the particular alternative in the question, the distribution of X Ȳ is normal with mean 2. The is the area to the right of 13 2/n under this curve, and the area has to be 0.5. So 13 2/n has to be the mean of the curve. Therefore 13 2/n = 2, so n = 85. 9. The issue of statistical significance doesn t arise. There s nothing random, no probability model of any kind. The inner planets are simply more dense, on average, than the outer planets. There s no chance involved. 10a. Yes, it makes sense. The 55% is a percent from all Berkeley freshmen so there s no randomness there, but the randomness is in the fact that the 61% comes from a random sample. The test, therefore, is about the unknown proportion p of 2004 freshmen in the nation who would disagree with Statement B. H 0 : p = 0.55, same as Berkeley. The sample percent is higher due to chance. H A : p > 0.55. The high sample percent reflects a high population percent. [It s also fine to say H A : p 0.55.] Under the null, ˆp has approximately normal distribution with mean 0.55 and standard error.55.45/1000 = 0.0157. On this scale,.61 corresponds to z = 3.82 so the p-value is tiny and the null is rejected. The difference is statistically significant (indeed, highly statistically significant). b. The question doesn t make sense because neither of the percents is from a random sample. Those two percents are just different. No p-values are needed, nor are they appropriate. c. The question makes sense because both percents come from random samples. But because they come from the same random sample, they are dependent. In order to answer the question it is necessary to have some way of measuring the dependence. That can t be done based on the information given. You should note, however, that the size of the difference is very great: 28%. It s hard to imagine a reasonable measure of dependence that would not make such a big difference statistically significant. Still, you can t do a test with the information given.

d. Both percents are from random samples, so the question makes sense. The test compares the national percent of freshmen agreeing with Statement A, in 1990 and 2004. The data come from two independent random samples. H 0 : p 90 = p 04 versus H A : p 90 < p 04 (yes, you can say instead of <) Under H 0, ˆp 90 ˆp 04 has approximately normal distribution with mean 0. To compute the SE, first compute the pooled estimate of the common population proportion,.5.22 +.5.33 =.275. The SE of ˆp 90 ˆp 04 is approximately 0.275 (1 0.275) (1/1000 + 1/1000) = 0.02. So the difference in proportions is expected to be 0 with an SE of about 0.02. The observed difference was 0.11. That s more than 5 SEs away from the expected 0, so the difference is statistically significant. The data indicate that the two population proportions are not the same.