STA 2101/442 Assignment 3 1

Similar documents
STA 2101/442 Assignment Four 1

STA442/2101: Assignment 5

STA 302f16 Assignment Five 1

Random Vectors 1. STA442/2101 Fall See last slide for copyright information. 1 / 30

STA 2201/442 Assignment 2

The Multivariate Normal Distribution 1

Introduction 1. STA442/2101 Fall See last slide for copyright information. 1 / 33

STA 2101/442 Assignment 2 1

The Multivariate Normal Distribution 1

The Bootstrap 1. STA442/2101 Fall Sampling distributions Bootstrap Distribution-free regression example

The Multinomial Model

UNIVERSITY OF TORONTO Faculty of Arts and Science

Joint Distributions: Part Two 1

Generalized Linear Models 1

Categorical Data Analysis 1

STA 431s17 Assignment Eight 1

Introduction to Bayesian Statistics 1

STA441: Spring Multiple Regression. This slide show is a free open source document. See the last slide for copyright information.

Lecture 11. Multivariate Normal theory

Contingency Tables Part One 1

Lecture 1: August 28

The Delta Method and Applications

UNIVERSITY OF MASSACHUSETTS. Department of Mathematics and Statistics. Basic Exam - Applied Statistics. Tuesday, January 17, 2017

Part IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015

MAS223 Statistical Inference and Modelling Exercises

Delta Method. Example : Method of Moments for Exponential Distribution. f(x; λ) = λe λx I(x > 0)

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

STA 4322 Exam I Name: Introduction to Statistics Theory

BIOS 2083 Linear Models Abdus S. Wahed. Chapter 2 84

Probability Theory and Statistics. Peter Jochumzen

Lecture 15. Hypothesis testing in the linear model

This does not cover everything on the final. Look at the posted practice problems for other topics.

Continuous Random Variables 1

Asymptotic Statistics-VI. Changliang Zou

STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method.

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

Interactions and Factorial ANOVA

Statistics STAT:5100 (22S:193), Fall Sample Final Exam B

Interactions and Factorial ANOVA

Matrix Approach to Simple Linear Regression: An Overview

Exercises and Answers to Chapter 1

3. Probability and Statistics

Notes on Random Vectors and Multivariate Normal

Linear Models and Estimation by Least Squares

STAT 111 Recitation 7

Multivariate Analysis and Likelihood Inference

(a) (3 points) Construct a 95% confidence interval for β 2 in Equation 1.

Regression and Statistical Inference

Random Vectors and Multivariate Normal Distributions

STT 843 Key to Homework 1 Spring 2018

Simple and Multiple Linear Regression

Asymptotic Theory. L. Magee revised January 21, 2013

First Year Examination Department of Statistics, University of Florida

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

Regression. ECO 312 Fall 2013 Chris Sims. January 12, 2014

Lecture 9 SLR in Matrix Form

Lecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012

Chapter 5 continued. Chapter 5 sections

Advanced topics from statistics

1 Least Squares Estimation - multiple regression.

Stat 5101 Notes: Brand Name Distributions

Power and Sample Size for Log-linear Models 1

The purpose of this section is to derive the asymptotic distribution of the Pearson chi-square statistic. k (n j np j ) 2. np j.

01 Probability Theory and Statistics Review

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama

Advanced Econometrics I

Chapter 5 Matrix Approach to Simple Linear Regression

Regression Review. Statistics 149. Spring Copyright c 2006 by Mark E. Irwin

Stat 5101 Lecture Notes

Qualifying Exam in Probability and Statistics.

Lecture 15 Multiple regression I Chapter 6 Set 2 Least Square Estimation The quadratic form to be minimized is

So far our focus has been on estimation of the parameter vector β in the. y = Xβ + u

Inference for Regression

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Spring 2012 Math 541A Exam 1. X i, S 2 = 1 n. n 1. X i I(X i < c), T n =

Continuous Distributions

Stat 5101 Notes: Algorithms

Hypothesis Tests and Estimation for Population Variances. Copyright 2014 Pearson Education, Inc.

Probability and Distributions

Calculus I Sample Exam #01

STA441: Spring Multiple Regression. More than one explanatory variable at the same time

STA2603/205/1/2014 /2014. ry II. Tutorial letter 205/1/

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003

Topic 7 - Matrix Approach to Simple Linear Regression. Outline. Matrix. Matrix. Review of Matrices. Regression model in matrix form

Correlation 1. December 4, HMS, 2017, v1.1

Master s Written Examination

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing

Probability Background

4. Distributions of Functions of Random Variables

EEL 5544 Noise in Linear Systems Lecture 30. X (s) = E [ e sx] f X (x)e sx dx. Moments can be found from the Laplace transform as

Math 423/533: The Main Theoretical Topics

Outline. Remedial Measures) Extra Sums of Squares Standardized Version of the Multiple Regression Model

Stat 366 A1 (Fall 2006) Midterm Solutions (October 23) page 1

PhD Qualifying Examination Department of Statistics, University of Florida

HYPOTHESIS TESTING: FREQUENTIST APPROACH.

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011

Exercise I.1 I.2 I.3 I.4 II.1 II.2 III.1 III.2 III.3 IV.1 Question (1) (2) (3) (4) (5) (6) (7) (8) (9) (10) Answer

18.S096 Problem Set 3 Fall 2013 Regression Analysis Due Date: 10/8/2013

Transcription:

STA 2101/442 Assignment 3 1 These questions are practice for the midterm and final exam, and are not to be handed in. 1. Suppose X 1,..., X n are a random sample from a distribution with mean µ and variance σ 2. The central limit theorem says n ( X n µ ) d T N(0, σ 2 ). One version of the delta method says that if g(x) is a function whose derivative is continuous in a neighbourhood of x = µ, then n ( g(x n ) g(µ) ) d g (µ)t. In many applications, both µ and σ 2 are functions of some parameter θ. (a) Let X 1,..., X n be a random sample from a Bernoulli distribution with parameter θ. Find the limiting distribution of Z n = 2 n (sin 1 X n sin ) 1 θ. Hint: d dx sin 1 (x) = 1 1 x 2. The measurements are in radians, not degrees. (b) In a coffee taste test, 100 coffee drinkers tasted coffee made with two different blends of coffee beans, the old standard blend and a new blend. We will adopt a Bernoulli model for these data, with θ denoting the probability that a customer will prefer the new blend. Suppose 60 out of 100 consumers preferred the new blend of coffee beans. Using your answer to the first part of this question, test H 0 : θ = 1 using a variance-stabilized test statistic. Give the value of the test 2 statistic (a number), and state whether you reject H 0 at the usual α = 0.05 significance level. In plain, non-statistical language, what do you conclude? This is a statement about preference for types of coffee, and of course you will draw a directional conclusion if possible. (c) If the probability of an event is p, the odds of the event is (are?) defined as p/(1 p). Suppose again that X 1,..., X n are a random sample from a Bernoulli distribution with parameter θ. In this case the log odds of X i = 1 would be estimated by X n Y n = log. 1 X n Naturally, that s the natural log. Find the approximate large-sample distribution (that is, the asymptotic distribution) of Y n. It s normal, of course. Your job is to give the approximate (that is, asymptotic) mean and variance of Y n. (d) Again using the Taste Test data, give a 95% confidence interval for the log odds of preferring the new brand. Your answer is a pair of numbers. 1 This assignment was prepared by Jerry Brunner, Department of Statistics, University of Toronto. It is licensed under a Creative Commons Attribution - ShareAlike 3.0 Unported License. Use any part of it as you like and share the result freely. The L A TEX source code is available from the course website: http://www.utstat.toronto.edu/ brunner/oldclass/appliedf18 1

(e) Let X 1,..., X n be a random sample from an exponential distribution with parameter θ, so that E(X i ) = θ and V ar(x i ) = θ 2. i. Find a variance-stabilizing transformation. That is, find a function g(x) such that the limiting distribution of Y n = n ( g(x n ) g(θ) ) does not depend on θ. ii. According to a Poisson process model for calls answered by a service technician, service times (that is, time intervals between taking 2 successive calls; there is always somebody on hold) are independent exponential random variables with mean θ. In 50 successive calls, one technician s mean service time was 3.4 minutes. Test whether this technician s mean service time differs from the mandated average time of 3 minutes. Use your answer to the first part of this question. 2. Let X 1,..., X n be a random sample from a uniform distribution on (0, θ). (a) What is the limiting distribution of n ( X n µ )? Just give the answer; there is no need to show any work. (b) What is the limiting distribution of 2 n ( X n µ )? Just give the answer; there is no need to show any work. But what Slutsky Lemma are you using? Check the lecture slides if necessary. (c) Find a variance-stabilizing transformation that produces a standard normal distribution. That is, letting T n = 2X n, find a function g(x) such that the limiting distribution of Y n = n (g(t n ) g(θ)) is standard normal. 3. The label on the peanut butter jar says peanuts, partially hydrogenated peanut oil, salt and sugar. But we all know there is other stuff in there too. There is very good reason to assume that the number of rat hairs in a 500g jar of peanut butter has a Poisson distribution with mean λ, because it s easy to justify a Poisson process model for how the hairs get into the jars. A sample of 30 jars of Brand A yields X = 6.8, while an independent sample of 40 jars of Brand B yields Y = 7.275. (a) State the model for this problem. (b) What is the parameter space Θ? (c) State the null hypothesis in symbols. (d) Find a variance-stabilizing transformation for the Poisson distribution. (e) Using your variance-stabilizing transformation, derive a test statistic that has an approximate standard normal distribution under H 0. 2

(f) Calculate your test statistic for these data. Do you reject the null hypothesis at α = 0.05? Answer Yes or No. (g) In plain, non-statistical language, what do you conclude? Your answer is something about peanut butter and rat hairs. 4. If the p 1 random vector x has variance-covariance matrix Σ and A is an m p matrix of constants, prove that the variance-covariance matrix of Ax is AΣA. Start with the definition of a variance-covariance matrix: cov(z) = E(Z µ z )(Z µ z ). 5. If the p 1 random vector x has mean µ and variance-covariance matrix Σ, show Σ = E(xx ) µµ. 6. Let the p 1 random vector x have mean µ and variance-covariance matrix Σ, and let c be a p 1 vector of constants. Find cov(x + c). Show your work. 7. Let the p 1 random vector x have mean µ and variance-covariance matrix Σ; let A be a q p matrix of constants and let B be an r p matrix of constants. Derive a nice simple formula for cov(ax, Bx). 8. Let x be a p 1 random vector with mean µ x and variance-covariance matrix Σ x, and let y be a q 1 random vector with mean µ y and variance-covariance matrix Σ y. Let Σ xy denote the p q matrix cov(x, y) = E ( (x µ x )(y µ y ) ). (a) What is the (i, j) element of Σ xy? You don t need to show any work; just write down the answer. (b) Find an expression for cov(x + y) in terms of Σ x, Σ y and Σ xy. Show your work. (c) Simplify further for the special case where Cov(X i, Y j ) = 0 for all i and j. (d) Let c be a p 1 vector of constants and d be a q 1 vector of constants. Find cov(x + c, y + d). Show your work. 9. Let x = (X 1, X 2, X 3 ) be multivariate normal with 1 1 0 0 µ = 0 and Σ = 0 2 0 6 0 0 1 Let Y 1 = X 1 + X 2 and Y 2 = X 2 + X 3. Find the joint distribution of Y 1 and Y 2. 10. Let X 1 be Normal(µ 1, σ 2 1), and X 2 be Normal(µ 2, σ 2 2), independent of X 1. What is the joint distribution of Y 1 = X 1 + X 2 and Y 2 = X 1 X 2? What is required for Y 1 and Y 2 to be independent? Hint: Use matrices.. 3

11. Show that if w N p (µ, Σ) with Σ positive definite, Y = (w µ) Σ 1 (w µ) has a chi-squared distribution with p degrees of freedom. 12. You know that if w N p (µ, Σ), then Aw + c N r (Aµ + c, AΣA ). Use this result to obtain the distribution of the sample mean under normal random sampling. That is, let X 1,..., X n be a random sample from a N(µ, σ 2 ) distribution. Find the distribution of X. You might want to use 1 to represent an n 1 column vector of ones. 13. Let X 1,..., X n be independent and identically distributed random variables with E(X i ) = µ and V ar(x i ) = σ 2. (a) Show Cov(X, (X j X)) = 0 for j = 1,..., n. (b) Why does this imply that if X 1,..., X n are normal, X and S 2 are independent? 14. Recall that the chi-squared distribution with ν degrees of freedom is just Gamma with α = ν 2 and β = 2. So if X χ2 (ν), it has moment-generating function M X (t) = (1 2t) ν/2. (a) Let W 1 χ 2 (ν 1 ) and W 2 χ 2 (ν 2 ) be independent, and W = W 1 + W 2. Find the distribution of W. Show your work (there s not much). (b) Let W = W 1 + W 2, where W 1 and W 2 are independent, W χ 2 (ν 1 + ν 2 ) and W 2 χ 2 (ν 2 ), where ν 1 and ν 2 are both positive. Show W 1 χ 2 (ν 1 ). (c) Let X 1,..., X n be random sample from a N(µ, σ 2 ) distribution. Show (n 1)S 2 σ 2 χ 2 (n 1). Hint: n i=1 (X i µ) 2 = n i=1 ( Xi X + X µ ) 2 =... 4

15. Let y = Xβ + ɛ, where X is an n p matrix of known constants, β is a p 1 vector of unknown constants, and ɛ is multivariate normal with mean zero and covariance matrix σ 2 I n. The constant σ 2 > 0 is unknown. (a) The hat matrix H = X(X X) 1 X. i. What are the dimensions (number of rows and columns) of H? ii. Show H is symmetric. iii. Show H is idempotent, meaning HH = H. iv. Show I H is symmetric. v. Show I H is idempotent. (b) What is the distribution of y? (c) The least squares (and maximum likelihood) estimate of β is β = (X X) 1 X y. What is the distribution of β? Show the calculations. (d) Let ŷ = X β. i. Show ŷ = Hy ii. What is the distribution of ŷ? Show the calculation. (e) Let the vector of residuals e = y ŷ. i. Show e = (I H)y ii. What is the distribution of e? Show the calculations. Simplify both the expected value and the covariance matrix. (f) Using Problem 7, show that e and β are independent. (g) The least-squares (and maximum likelihood) estimator β is obtained by minimizing the sum of squares Q = (y Xβ) (y Xβ) over all β R p. i. Show that Q = e e + ( β β) (X X)( β β). Hint: Add and subtract ŷ. ii. Why does this imply that the minimum of Q(β) occurs at β = β? iii. The columns of X are linearly independent. Why does linear independence guarantee that the minimum is unique? You have just minimized a function of p variables without calculus. iv. Show that W 1 = SSE χ 2 (n p), where SSE = e e. Use results from σ 2 earlier parts of this assignment. Start with the distribution of W = 1 (y σ 2 Xβ) (y Xβ). This is the chi-squared random variable that appears in the denominator of all those F and t statistics. 5