Nonparametric Regression and Bonferroni joint confidence intervals. Yang Feng

Similar documents
Simultaneous Inference: An Overview

Formal Statement of Simple Linear Regression Model

Inference in Normal Regression Model. Dr. Frank Wood

Lecture 1 Linear Regression with One Predictor Variable.p2

Multiple Regression. Dr. Frank Wood. Frank Wood, Linear Regression Models Lecture 12, Slide 1

Simple Linear Regression (Part 3)

Final Review. Yang Feng. Yang Feng (Columbia University) Final Review 1 / 58

Diagnostics and Remedial Measures

Inference in Regression Analysis

[4+3+3] Q 1. (a) Describe the normal regression model through origin. Show that the least square estimator of the regression parameter is given by

STAT 540: Data Analysis and Regression

STA 302 H1F / 1001 HF Fall 2007 Test 1 October 24, 2007

Regression Estimation - Least Squares and Maximum Likelihood. Dr. Frank Wood

ST505/S697R: Fall Homework 2 Solution.

PART I. (a) Describe all the assumptions for a normal error regression model with one predictor variable,

Outline. Remedial Measures) Extra Sums of Squares Standardized Version of the Multiple Regression Model

Remedial Measures, Brown-Forsythe test, F test

Lecture 3: Inference in SLR

Bias Variance Trade-off

Chapter 2 Inferences in Simple Linear Regression

Lecture 6 Multiple Linear Regression, cont.

Introduction to Simple Linear Regression

Ch 2: Simple Linear Regression

Measuring the fit of the model - SSR

Need for Several Predictor Variables

Regression Analysis: Basic Concepts

Chapter 2 Inferences in Regression and Correlation Analysis

Simple Linear Regression Analysis

Linear models and their mathematical foundations: Simple linear regression

Linear Algebra Review

Math 3330: Solution to midterm Exam

STA 4210 Practise set 2a

6. Multiple Linear Regression

Model Selection. Frank Wood. December 10, 2009

STAT 360-Linear Models

Chapter 27 Summary Inferences for Regression

Inferences for Regression

BNAD 276 Lecture 10 Simple Linear Regression Model

LINEAR REGRESSION MODELS W4315

Statistical View of Least Squares

Applied Linear Statistical Models. Simultaneous Inference Topics. Simultaneous Estimation of β 0 and β 1 Issues. Simultaneous Inference. Dr.

Lecture 18: Simple Linear Regression

Lecture 14 Simple Linear Regression

Regression Estimation Least Squares and Maximum Likelihood

Inference for Regression

Simple Linear Regression

Contents. 1 Review of Residuals. 2 Detecting Outliers. 3 Influential Observations. 4 Multicollinearity and its Effects

Figure 1: The fitted line using the shipment route-number of ampules data. STAT5044: Regression and ANOVA The Solution of Homework #2 Inyoung Kim

SSR = The sum of squared errors measures how much Y varies around the regression line n. It happily turns out that SSR + SSE = SSTO.

appstats27.notebook April 06, 2017

Lecture 2 Simple Linear Regression STAT 512 Spring 2011 Background Reading KNNL: Chapter 1

STAT 512 MidTerm I (2/21/2013) Spring 2013 INSTRUCTIONS

STAT5044: Regression and Anova. Inyoung Kim

MATH 644: Regression Analysis Methods

Simple and Multiple Linear Regression

STA 108 Applied Linear Models: Regression Analysis Spring Solution for Homework #6

y ˆ i = ˆ " T u i ( i th fitted value or i th fit)

Problems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B

Chapter 5 Friday, May 21st

Regression Models - Introduction

Lecture 10 Multiple Linear Regression

Simple Regression Model Setup Estimation Inference Prediction. Model Diagnostic. Multiple Regression. Model Setup and Estimation.

Unit 6 - Simple linear regression

ST430 Exam 1 with Answers

Inference for Regression Simple Linear Regression

Multiple Linear Regression. Chapter 12

Chapter 2 Multiple Regression I (Part 1)

Applied Regression Modeling: A Business Approach Chapter 3: Multiple Linear Regression Sections

LECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit

1. Simple Linear Regression

Ordinary Least Squares Regression Explained: Vartanian

Simple Linear Regression

Chapter 1: Linear Regression with One Predictor Variable also known as: Simple Linear Regression Bivariate Linear Regression

Concordia University (5+5)Q 1.

Warm-up Using the given data Create a scatterplot Find the regression line

Simple Linear Regression for the Climate Data

Chapter 1. Linear Regression with One Predictor Variable

Simple Linear Regression

STAT 511. Lecture : Simple linear regression Devore: Section Prof. Michael Levine. December 3, Levine STAT 511

Chapter 12: Multiple Linear Regression

Correlation Analysis

Statistics for Managers using Microsoft Excel 6 th Edition

with the usual assumptions about the error term. The two values of X 1 X 2 0 1

Basic Business Statistics 6 th Edition

How the mean changes depends on the other variable. Plots can show what s happening...

Chapter 6 Multiple Regression

Multiple Linear Regression

Ch. 1: Data and Distributions

Regression Models - Introduction

An Introduction to Bayesian Linear Regression

Lecture 11: Simple Linear Regression

High-dimensional regression

Section 3: Simple Linear Regression

Inference in Constrained Linear Regression

Simple linear regression

Parameter estimation: ACVF of AR processes

Correlation and Regression

Remedial Measures for Multiple Linear Regression Models

Linear regression methods

Chapte The McGraw-Hill Companies, Inc. All rights reserved.

Transcription:

Nonparametric Regression and Bonferroni joint confidence intervals Yang Feng

Simultaneous Inferences In chapter 2, we know how to construct confidence interval for β 0 and β 1. If we want a confidence level of 95% of both β 0 and β 1 One could construct a separate confidence interval for β 0 and β 1. BUT, then the probability of both happening is below 95%. How to create a joint confidence interval?

Bonferroni Joint Confidence Intervals Calculation of Bonferroni joint confidence intervals is a general technique We highlight its application in the regression setting Joint confidence intervals for β0 and β 1 Intuition Set each statement confidence level to larger than 1 α so that the family coefficient is at least 1 α BUT how much larger?

Ordinary Confidence Intervals Start with ordinary confidence intervals for β 0 and β 1 b 0 ± t(1 α/2; n 2)s{b 0 } b 1 ± t(1 α/2; n 2)s{b 1 } And ask what the probability that one or both of these intervals is incorrect Remember [ ] 1 s 2 {b 0 } = MSE n + X 2 (Xi X ) 2 s 2 {b 1 } = MSE (Xi X ) 2

General Procedure Let A 1 denote the event that the first confidence interval does not cover β 0, i.e. P(A 1 ) = α Let A 2 denote the event that the second confidence interval does not cover β 1, i.e. P(A 2 ) = α We want to know the probability that both estimates fall in their respective confidence intervals, i.e. P(Ā 1 Ā 2 ) How do we get there from what we know?

Venn Diagram

Bonferroni inequality We can see that P(Ā 1 Ā 2 ) = 1 P(A 2 ) P(A 1 ) + P(A 1 A 2 ) Size of set is equal to area is equal to probability in a Venn diagram. It also is clear that P(A 1 A 2 ) 0 So, P(Ā 1 Ā 2 ) 1 P(A 2 ) P(A 1 ) which is the Bonferroni inequality. In words, in our example P(A1 ) = α is the probability that β 0 is not in A 1 P(A2 ) = α is the probability that β 1 is not in A 2 P(Ā1 Ā 2 ) is the probability that β 0 is in A 1 and β 1 is in A 2 So P(Ā1 Ā2) 1 2α

Using the Bonferroni inequality Forward (less interesting) : If we know that β0 and β 1 are lie within intervals with 95% confidence, the Bonferroni inequality guarantees us a family confidence coefficient (i.e. the probability that both random variables lie within their intervals simultaneously) of at least 90% (if both intervals are correct). This is P(Ā 1 Ā 2 ) 1 2α Backward (more useful): If we know what to specify a family confidence interval of 90%, the Bonferroni procedure instructs us how to adjust the value of α for each interval to achieve the overall family confidence desired

Using the Bonferroni inequality cont. To achieve a 1 α family confidence interval for β 0 and β 1 (for example) using the Bonferroni procedure we know that both individual intervals must shrink. Returning to our confidence intervals for β 0 and β 1 from before b 0 ± t(1 α/2; n 2)s{b 0 } b 1 ± t(1 α/2; n 2)s{b 1 } To achieve a 1 α family confidence interval these intervals must widen to Then b 0 ± t(1 α/4; n 2)s{b 0 } b 1 ± t(1 α/4; n 2)s{b 1 } P(Ā1 Ā2) 1 P(A 2 ) P(A 1 ) = 1 α/4 α/4 = 1 α/2

Using the Bonferroni inequality cont. The Bonferroni procedure is very general. To make joint confidence statements about multiple simultaneous predictions remember that Ŷ h ± t(1 α/2; n 2)s{pred} s 2 {pred} = MSE [1 + 1n + (X h X ) 2 ] i (X i X ) 2 If one is interested in a 1 α confidence statement about g predictions then Bonferroni says that the confidence interval for each individual prediction must get wider (for each h in the g predictions) Ŷ h ± t(1 α/2g; n 2)s{pred} Note: if a sufficiently large number of simultaneous predictions are made, the width of the individual confidence intervals may become so wide that they are no longer useful.

The Toluca Example Say, we want to get a 90 percent confidence interval for β 0 and β 1 simultaneously. Then we require B = t(1.1/4; 23) = t(.975, 23) = 2.069 Then we have the joint confidence interval: b 0 ± B s(b 0 ) and b 1 ± B s(b 1 )

Confidence Band for Regression Line Remember in Chapter 2.5, we get the confidence interval for E{Y h } to be Ŷ h ± t(1 α/2; n 2)s{Ŷh} Now, we want to get a confidence band for the entire regression line E{Y } = β 0 + β 1 X. So called Working-Hotelling 1 α confidence band is Ŷ h ± W s{ŷ h } here W 2 = 2F (1 α; 2, n 2). Same form as before, except the t multiple is replaced with the W multiple.

Example: toluca company Say we want to estimate the boundary value for the band at X h = 30, 65, 100. We have Looking up the table, W 2 = 2F (1 α; 2, n 2) = 2F (.9; 2, 23) = 5.098. R code: w2 = 2 * qf(1-0.1,2,23)

Now we have the confidence band for the three points are

Example confidence band

Compare with Bonferroni Procedure Say we want to simultaneously estimate response for X h = 30, 65, 100. Then the simultaneous confidence intervals are Ŷ h ± t(1 α/(2g); n 2)s{Ŷh} We have B = t(1 α/(2g); n 2) = t(1.1/(2 3), 23) = 2.263, the confidence intervals are

Bonferroni v.s. Working-Hotelling This instance, working-hotelling confidence limits are slighter tighter(better) than bonferroni limits However, in larger families (more X ) to be considered simultaneously, working-hotelling is always tighter, since W stays the same for any number of statements but B becomres larger. The levels of predictor variables are sometimes not known in advance. In such cases, it is better to use Working-Hotelling procedure since the family encompasses all possible levels of X.

Regression through the origin Model Y i = β 1 X i + ɛ i Sometimes it is known that the regression function is linear and that it must go through the origin. β 1 is parameter X i are known constants ɛ i are i.i.d N(0, σ 2 ). The least squares and maximum likelihood estimators for β 1 coincide as before, the estimator is b 1 = Xi Y i X 2 i

Regression through the origin, Cont In regression through the origin there is only one free parameter (β 1 ) so the number of degrees of freedom of the MSE e s 2 2 = MSE = i n 1 = (Yi Ŷ i ) 2 n 1 is increased by one. This is because this is a reduced model in the general linear test sense and because the number of parameters estimated from the data is less by one.

A few notes on regression through the origin e i 0 in general now. Only constraint is X i e i = 0. SSE may exceed the total sum of squares SSTO. In the case of a curvilinear pattern or linear pattern with a intercept away from the origin. Therefore, R 2 = 1 SSE/SSTO may be negative! Generally, it is safer to use the original model opposed with regression-through-the-origin model. Otherwise, it is the wrong model to start with!