Chapter 2 Multiple Regression I (Part 1)
|
|
- Jane Jenkins
- 6 years ago
- Views:
Transcription
1 Chapter 2 Multiple Regression I (Part 1) 1 Regression several predictor variables The response Y depends on several predictor variables X 1,, X p response {}}{ Y predictor variables {}}{ X 1, X 2,, X p Observations (or Design) Thus, generally for individual i, theresponseis:y i obs Y X 1 X 2 X p 1 Y 1 X 11 X 12 X 1p 2 Y 2 X 21 X 22 X 2p n Y 2 X n1 X n2 X np the predictors variables are: X i1,x i2,, X ip 2 Linear regression model with Two predictor variables The linear regression model assumes that for any subject/individual with response Y i and predictor X i1,x i2 satisfies Y i = β 0 + β 1 X i1 + β 2 X i2 + ε i where Eε i =0,orequivalently E(Y i )=β 0 + β 1 X i1 + β 2 X i2 Sometimes, it is also written as, Y = β 0 + β 1 X 1 + β 2 X 2 + ε 1
2 where Eε =0 orequivalently E(Y )=β 0 + β 1 X 1 + β 2 X 2 where β 0,β 1,β 2 are called regression coefficient β 0 is called intercept β 1 is called coefficient of X 1 ; β 2 is called coefficient of X 2 For example: (height in inch) (Expected height of girl) = (Farther s height) + 05(Mother s height) (Expected height of boy) = (Farther s height) + 05(Mother s height) Meaning of the regression coefficients β 1 indicate the change in the mean response EY per unit increase in X 1 when X 2 holds constant β 2 indicate the change in the mean response EY per unit increase in X 2 when X 1 holds constant Note that X 1 and X 2 have some correlation, thus you need to know the difference in statistical and mathematical models [in mathematical model, X 1 and X 2 can be really free the change, but statistical model may not completely free] 3 Linear regression model with p predictor variables The linear regression model assumes that for any subject/individual with response Y i and predictor X i1,, X ip satisfies where Eε i =0,orequivalently Y i = β 0 + β 1 X i1 + + β p X ip + }{{} ε }{{} i predictable unpredictable E(Y i )=β 0 + β 1 X i1 + + β p X ip This means for each individual, the expected value of the response is a functional relationship with the independent variables But the real value has random error ε i from the expected value Sometimes, the model is written as Y = β 0 + β 1 X β p X p + ε 2
3 where Eε =0,orequivalently E(Y )=β 0 + β 1 X β p X p which is called a hyperplane, whereβ 0,β 1,, β p are called regression coefficient Meaning of the regression coefficients β k indicate the change in the mean response EY per unit increase in X k when the other predictors remain constant It is easy to see that we have studied the case p = 1, ie simple linear regression model We usually make the following assumptions (L) Linearity (implied in the model) (I) Independence of Error Terms, thus Cov(ε i,ε j )=0, if i j (N) Normality of Error Terms: ε N(0,σ 2 ) (E) Equal/constant Error Variance: Var{ε i } = σ 2 (F) Fixed design: X i1,, X ip are known and nonrandom There are p+1 coefficients β 0,, β p, one common variance σ 2, they are called parameters of the model 4 Some Examples Here we give some examples that are nonlinear, but can be transformed to linear regression models Qualitative Predictor variables It is understandable that the predictors must be quantitative But we can also consider qualitative predictor, by denoting the predictor using dummy variables For example Y a person s height, X 1 is his/her father s height, X 2 his/her mother s height, S is the gender of the person We can denote the gender by Then our model is X 3 = { 1, if the person is male 0, if the person is male Y = β 0 + β 1 X 1 + β 2 X 2 + β 3 X 3 + ε 3
4 Polynomial regression models, for example Y i = β 0 + β 1 X i + β 2 Xi 2 + ε i, Y i = β 0 + β 1 X i + β 3 Xi β k Xi k + ε i, Y i = β 0 + β 1 X i1 + β 2 X i2 + β 2 X i3 + β 4 Xi1 2 + β 5 X i1 X i2 + β 6 Xi3 3 + ε i, Y i = β 0 + β 1 X i1 + β 2 X i2 + β 2 X i3 + β 4 Xi1 2 + β 5X i1 X i2 + β 6 Xi1 4 + ε i, X i1 X i2 are usually called interaction of X 1 and X 2, how about X i2 X i3? Transformed model (after variable transformation, the model become a linear regression model) Here are some examples (a) For model Y i = a 0 exp(β 1 X i1 + + β p X ip )ξ i,letz i = log(y i ), ε i =log(ξ i )and β 0 =log(a 0 ) Taking logrithm, the model becomes Z i = β 0 + β 1 X i1 + + β p X ip + ε i (b) model Y i = β 0 + β 1 X i1 + β 2 X i2 + β 2 X i3 + β 4 X 2 i1 + β 5X i1 X i2 + β 6 X 3 i3 + ε i, can be written as Y i = β 0 + β 1 X i1 + β 2 X i2 + β 2 X i3 + β 4 Z i4 + β 5 Z i5 + β 6 Z i6 + ε i, where Z i4 = X 2 i1, Z i5 = X i1 X i2 and Z i6 = X 3 i3 5 General linear regression model in matrix terms Again, our general model can be written as Y i = β 0 + β 1 X i1 + + β p X ip + ε i, i =1,, n or Y 1 = β 0 + β 1 X β p X 1p + ε 1, Y 2 = β 0 + β 1 X β p X 2p + ε 2, Y n = β 0 + β 1 X n1 + + β p X np + ε n (with the 5 assumptions) 4
5 Let β = β 0 β 1 β p 1 X 11 X 1,p 1 X 21 X 2,p X =, 1 X n1 X n,p called coefficient vector Y = E = ε 1 ε 2 ε n It is easy to check 1 X 11 X 1,p 1 X 21 X 2,p 1 X n1 X n,p β 0 β 1 β p The regression model can be written as called Design matrix Y 1 Y 2 Y n called random error vector = Y = Xβ + E β 0 + β 1 X 1,1 + + β p X 1,p β 0 + β 1 X 2,1 + + β p X 2,p β 0 + β 1 X n,1 + + β p X n,p called response vector The (L-I-N-E) assumptions can be written as σ σ E{E} = 0, Var{E} = σ 2 E N(0,σ 2 I) = σ2 I 6 Least squares estimation Minimize Q(b 0,, b p )= n (Y i b 0 b 1 X i1 b p X i,p ) 2 i=1 by calculus, we have the following (p+1) Normal equations: (how?) n (Y i b 0 b 1 X i1 b p X i,p ) = 0 i=1 n (Y i b 0 b 1 X i1 b p X i,p )X i1 = 0 i=1 n (Y i b 0 b 1 X i1 b p X i,p )X ip = 0 i=1 5
6 let b =(b 0,b 1,, b p ) Then the Normal equations can be written as X Xb = X Y The solution, ie the estimator of the coefficient vector, is b =(X X) 1 X Y The estimated model is Ŷ = b 0 + b 1 X b p X p Fitted values Ŷ i = b 0 + b 1 X i1 + + b p X ip, i =1,, n (Fitted) residuals e i = Y i Ŷi = Y i (b 0 + b 1 X i1 + + b p X ip ), i =1,, n Estimator of σ 2, denoted by ˆσ 2, MSE = n i=1 e 2 i /{n (p +1)} called Mean squared error why (p+1)? (because there are p+1 constraints, p+1 is the number of (free) coefficients, or more exactly the number of Normal equations) Dwaine Studios example Y -sales, X 1 - number of persons aged 16 or less, X 2 - income 21 observations X X = b = Y = ; X = , , , , , , , , , , , , , X Y = 3, , , 073 3, , , 073 =
7 4 The estimated model is Ŷ = X X obs X 1 X 2 Y Fitted Ŷi residuals e i ˆσ 2 = MSE = 21 i=1 e2 i n p 1 = = Unbias of the estimators of coefficients The estimator of coefficient vector is unbiased, ie E(b) =β and Var(b) =σ 2 (X X) 1 In details E(b k )=β k and Var(b k )=σ 2 c k+1,k+1, k =0, 1,, p 1 where c kk is the (k, k)th entry in (X X) 1 [Proof: Note that EY = Xβ Thus E{b} = E{(X X) 1 X Y} = {(X X) 1 X E{Y} = β and Var(b) =(X X) 1 X Var(Y)X(X X) 1 =(X X) 1 X σ 2 IX(X X) 1 = σ 2 (X X) 1 ] 7
8 8 Fitted values and residuals in matrix form fitted value Ŷ = Ŷ 1 Ŷ 2 Ŷ n = = X(X X) 1 X Y b 0 + b 1 X b p X 1,p b 0 + b 1 X b p X 2,p b 0 + b 1 X n1 + + b p X n,p = Xb fitted residuals e = Y Ŷ =(I X(X X) 1 X )Y Denote X(X X) 1 X by H, wehaveŷ = HY, e =(I H)Y 9 Variance-covariance matrix for residuals e Var{e} = Var{(I H)Y} =(I H)Var{Y}(I H) Var{Y} = Var{E} = σ 2 I (I H) = I H = I H HH = X(X X) 1 X X(X X) 1 X = X(X X) 1 X = H (I H)(I H) =I 2H + HH = I H Var{e} = σ 2 (I H), which can be estimated by ˆσ 2 (I H), where ˆσ 2 = MSE = e e n p 1 = Y (I H)Y (n p 1) Eˆσ 2 = E(MSE)=σ 2 [The proof can be ignored] 8
9 10 Variance-covariance matrix for b Recall b =(X X) 1 X Y, Var{b} =(X X) 1 X Var{Y}X(X X) 1 = σ 2 (X X) 1 where σ 2 can be estimated by ˆσ 2 =MSE In other word, we estimate Var{b} by ˆσ 2 (X X) 1, denoted s(b) =ˆσ 2 (X X) 1 For the above example, MSE = SSE n p 1 = e e s 2 {b} = 12116(X X) 1 = 11 The distribution of estimators If E N(0,σ 2 I) (ie ε i are IID N(0,σ 2 )), then = 2, = , The estimated coefficients b N(β,σ 2 (X X) 1 ) Denote the (i, j)th entry of (X X) 1 by c ij,then b k N(β k,σ 2 c k+1,k+1 ), k =0, 1,, p 1 (where b =(b 0,b 1,, b p ) ) Let s(b k )= MSE c k+1,k+1, called Standard Error (SE) for b k (which can be found in the output of R), then b k β k s(b k ) t(n p 1) t-value t = b k s(b k ) 9
10 12 Confidence interval for β k with 1 α confidence, the Confidence interval for β k is [b k s(b k ) t(1 α/2,n p 1), b k s(b k ) t(1 α/2,n p 1)] For the Dwaine Studios example, the 95% Confidence interval for β 2 is [ , ] = [083, 1790] where quantile (critical value) t(1 α/2,n p 1) = t(0975, 21 3) = 2101 is used 13 test for β k =0 Our hypothesis is H 0 : β k =0, H a : β k 0 under H 0, t = b k β k s(b k ) For significant level α, our criterion is = b k = t(n p 1) s(b k ) If the calculated t >t(1 α/2,n p 1), reject H 0 If the calculated t t(1 α/2,n p 1), accept H 0 Similarly, we can do the test based on the p-value If p-value <α, reject H 0 If p-value α, accept H 0 For the Dwaine Studios example, test H 0 : β 1 =0, H a : β 1 0 with significance level 5%, since t =6868 >t(1 α/2,n p 1) = 2101 we reject H 0 (in other words, β 1 is significantly different from 0) Similarly, H 0 : β 0 = 0 can be accepted H 0 : β 2 = 0 should be rejected 10
11 14 Prediction For any new individual with X new =(x 1,, x p ), the predict mean response is Ŷ new = X b where X =(1,x 1,, x p ) We have EŶnew = EY new Note that if normal errors are assumed, ie ε i are IID N(0,σ 2 ), then Ŷ new N(EY new, X (X X) 1 X σ 2 ) Let s 2 (Ŷnew) =X (X X) 1 X ˆσ 2 = X (X X) 1 X MSE We have Ŷ new EY new t(n p 1) s(ŷnew) With confidence 100(1 α)%, the CI for E(Y new )is [Ŷnew s(ŷnew) t(1 α/2,n p 1), Ŷ new + s(ŷnew) t(1 α/2,n p 1)] What about the prediction interval (PI) for the value Y new? With confidence 100(1 α)%, the PI for Y new is [Ŷnew s(pred) t(1 α/2,n p 1), Ŷ new + s(pred) t(1 α/2,n p 1)] where s 2 (pred) =MSE + s 2 (Ŷnew) =MSE{1+X (X X) 1 X} 15 R code regression=lm(y x 1 + x x p ) summary(regression) Xnew = dataframe(x1=c(), x2=c(),, xp=c()) predict(regression, Xnew, interval = "confidence"/"prediction", level=095) 11
Math 3330: Solution to midterm Exam
Math 3330: Solution to midterm Exam Question 1: (14 marks) Suppose the regression model is y i = β 0 + β 1 x i + ε i, i = 1,, n, where ε i are iid Normal distribution N(0, σ 2 ). a. (2 marks) Compute the
More informationOutline. Remedial Measures) Extra Sums of Squares Standardized Version of the Multiple Regression Model
Outline 1 Multiple Linear Regression (Estimation, Inference, Diagnostics and Remedial Measures) 2 Special Topics for Multiple Regression Extra Sums of Squares Standardized Version of the Multiple Regression
More informationSimple Linear Regression (Part 3)
Chapter 1 Simple Linear Regression (Part 3) 1 Write an Estimated model Statisticians/Econometricians usually write an estimated model together with some inference statistics, the following are some formats
More informationMa 3/103: Lecture 24 Linear Regression I: Estimation
Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the
More informationRegression Models - Introduction
Regression Models - Introduction In regression models there are two types of variables that are studied: A dependent variable, Y, also called response variable. It is modeled as random. An independent
More informationSimple and Multiple Linear Regression
Sta. 113 Chapter 12 and 13 of Devore March 12, 2010 Table of contents 1 Simple Linear Regression 2 Model Simple Linear Regression A simple linear regression model is given by Y = β 0 + β 1 x + ɛ where
More informationSTAT5044: Regression and Anova. Inyoung Kim
STAT5044: Regression and Anova Inyoung Kim 2 / 47 Outline 1 Regression 2 Simple Linear regression 3 Basic concepts in regression 4 How to estimate unknown parameters 5 Properties of Least Squares Estimators:
More informationFinal Review. Yang Feng. Yang Feng (Columbia University) Final Review 1 / 58
Final Review Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Final Review 1 / 58 Outline 1 Multiple Linear Regression (Estimation, Inference) 2 Special Topics for Multiple
More informationChapter 6 Multiple Regression
STAT 525 FALL 2018 Chapter 6 Multiple Regression Professor Min Zhang The Data and Model Still have single response variable Y Now have multiple explanatory variables Examples: Blood Pressure vs Age, Weight,
More information6. Multiple Linear Regression
6. Multiple Linear Regression SLR: 1 predictor X, MLR: more than 1 predictor Example data set: Y i = #points scored by UF football team in game i X i1 = #games won by opponent in their last 10 games X
More informationSimple Linear Regression
Simple Linear Regression In simple linear regression we are concerned about the relationship between two variables, X and Y. There are two components to such a relationship. 1. The strength of the relationship.
More informationNeed for Several Predictor Variables
Multiple regression One of the most widely used tools in statistical analysis Matrix expressions for multiple regression are the same as for simple linear regression Need for Several Predictor Variables
More informationLecture 15 Multiple regression I Chapter 6 Set 2 Least Square Estimation The quadratic form to be minimized is
Lecture 15 Multiple regression I Chapter 6 Set 2 Least Square Estimation The quadratic form to be minimized is Q = (Y i β 0 β 1 X i1 β 2 X i2 β p 1 X i.p 1 ) 2, which in matrix notation is Q = (Y Xβ) (Y
More informationCh 3: Multiple Linear Regression
Ch 3: Multiple Linear Regression 1. Multiple Linear Regression Model Multiple regression model has more than one regressor. For example, we have one response variable and two regressor variables: 1. delivery
More informationTopic 14: Inference in Multiple Regression
Topic 14: Inference in Multiple Regression Outline Review multiple linear regression Inference of regression coefficients Application to book example Inference of mean Application to book example Inference
More informationLecture 10 Multiple Linear Regression
Lecture 10 Multiple Linear Regression STAT 512 Spring 2011 Background Reading KNNL: 6.1-6.5 10-1 Topic Overview Multiple Linear Regression Model 10-2 Data for Multiple Regression Y i is the response variable
More informationChapter 1 Linear Regression with One Predictor
STAT 525 FALL 2018 Chapter 1 Linear Regression with One Predictor Professor Min Zhang Goals of Regression Analysis Serve three purposes Describes an association between X and Y In some applications, the
More informationCh 2: Simple Linear Regression
Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component
More informationSTAT 540: Data Analysis and Regression
STAT 540: Data Analysis and Regression Wen Zhou http://www.stat.colostate.edu/~riczw/ Email: riczw@stat.colostate.edu Department of Statistics Colorado State University Fall 205 W. Zhou (Colorado State
More informationRegression Review. Statistics 149. Spring Copyright c 2006 by Mark E. Irwin
Regression Review Statistics 149 Spring 2006 Copyright c 2006 by Mark E. Irwin Matrix Approach to Regression Linear Model: Y i = β 0 + β 1 X i1 +... + β p X ip + ɛ i ; ɛ i iid N(0, σ 2 ), i = 1,..., n
More informationChapter 14. Linear least squares
Serik Sagitov, Chalmers and GU, March 5, 2018 Chapter 14 Linear least squares 1 Simple linear regression model A linear model for the random response Y = Y (x) to an independent variable X = x For a given
More informationChapter 1 Simple Linear Regression (part 6: matrix version)
Chapter Simple Liear Regressio (part 6: matrix versio) Overview Simple liear regressio model: respose variable Y, a sigle idepedet variable X Y β 0 + β X + ε Multiple liear regressio model: respose Y,
More informationSimple Linear Regression
Simple Linear Regression ST 430/514 Recall: A regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates)
More informationInference in Normal Regression Model. Dr. Frank Wood
Inference in Normal Regression Model Dr. Frank Wood Remember We know that the point estimator of b 1 is b 1 = (Xi X )(Y i Ȳ ) (Xi X ) 2 Last class we derived the sampling distribution of b 1, it being
More informationLecture 14 Simple Linear Regression
Lecture 4 Simple Linear Regression Ordinary Least Squares (OLS) Consider the following simple linear regression model where, for each unit i, Y i is the dependent variable (response). X i is the independent
More informationFormal Statement of Simple Linear Regression Model
Formal Statement of Simple Linear Regression Model Y i = β 0 + β 1 X i + ɛ i Y i value of the response variable in the i th trial β 0 and β 1 are parameters X i is a known constant, the value of the predictor
More informationChapter 5 Matrix Approach to Simple Linear Regression
STAT 525 SPRING 2018 Chapter 5 Matrix Approach to Simple Linear Regression Professor Min Zhang Matrix Collection of elements arranged in rows and columns Elements will be numbers or symbols For example:
More informationSTAT5044: Regression and Anova. Inyoung Kim
STAT5044: Regression and Anova Inyoung Kim 2 / 51 Outline 1 Matrix Expression 2 Linear and quadratic forms 3 Properties of quadratic form 4 Properties of estimates 5 Distributional properties 3 / 51 Matrix
More informationSimple Regression Model Setup Estimation Inference Prediction. Model Diagnostic. Multiple Regression. Model Setup and Estimation.
Statistical Computation Math 475 Jimin Ding Department of Mathematics Washington University in St. Louis www.math.wustl.edu/ jmding/math475/index.html October 10, 2013 Ridge Part IV October 10, 2013 1
More informationRegression Models - Introduction
Regression Models - Introduction In regression models, two types of variables that are studied: A dependent variable, Y, also called response variable. It is modeled as random. An independent variable,
More informationWeighted Least Squares
Weighted Least Squares The standard linear model assumes that Var(ε i ) = σ 2 for i = 1,..., n. As we have seen, however, there are instances where Var(Y X = x i ) = Var(ε i ) = σ2 w i. Here w 1,..., w
More informationAdvanced Econometrics I
Lecture Notes Autumn 2010 Dr. Getinet Haile, University of Mannheim 1. Introduction Introduction & CLRM, Autumn Term 2010 1 What is econometrics? Econometrics = economic statistics economic theory mathematics
More informationStatistics for Engineers Lecture 9 Linear Regression
Statistics for Engineers Lecture 9 Linear Regression Chong Ma Department of Statistics University of South Carolina chongm@email.sc.edu April 17, 2017 Chong Ma (Statistics, USC) STAT 509 Spring 2017 April
More informationLecture 6 Multiple Linear Regression, cont.
Lecture 6 Multiple Linear Regression, cont. BIOST 515 January 22, 2004 BIOST 515, Lecture 6 Testing general linear hypotheses Suppose we are interested in testing linear combinations of the regression
More informationLinear models and their mathematical foundations: Simple linear regression
Linear models and their mathematical foundations: Simple linear regression Steffen Unkel Department of Medical Statistics University Medical Center Göttingen, Germany Winter term 2018/19 1/21 Introduction
More informationMultivariate Regression (Chapter 10)
Multivariate Regression (Chapter 10) This week we ll cover multivariate regression and maybe a bit of canonical correlation. Today we ll mostly review univariate multivariate regression. With multivariate
More informationLinear Regression Model. Badr Missaoui
Linear Regression Model Badr Missaoui Introduction What is this course about? It is a course on applied statistics. It comprises 2 hours lectures each week and 1 hour lab sessions/tutorials. We will focus
More informationInference in Regression Analysis
Inference in Regression Analysis Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 4, Slide 1 Today: Normal Error Regression Model Y i = β 0 + β 1 X i + ǫ i Y i value
More informationSTAT 3A03 Applied Regression With SAS Fall 2017
STAT 3A03 Applied Regression With SAS Fall 2017 Assignment 2 Solution Set Q. 1 I will add subscripts relating to the question part to the parameters and their estimates as well as the errors and residuals.
More information2.4.3 Estimatingσ Coefficient of Determination 2.4. ASSESSING THE MODEL 23
2.4. ASSESSING THE MODEL 23 2.4.3 Estimatingσ 2 Note that the sums of squares are functions of the conditional random variables Y i = (Y X = x i ). Hence, the sums of squares are random variables as well.
More informationLinear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,
Linear Regression In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, y = Xβ + ɛ, where y t = (y 1,..., y n ) is the column vector of target values,
More information2. Regression Review
2. Regression Review 2.1 The Regression Model The general form of the regression model y t = f(x t, β) + ε t where x t = (x t1,, x tp ), β = (β 1,..., β m ). ε t is a random variable, Eε t = 0, Var(ε t
More informationLecture 4 Multiple linear regression
Lecture 4 Multiple linear regression BIOST 515 January 15, 2004 Outline 1 Motivation for the multiple regression model Multiple regression in matrix notation Least squares estimation of model parameters
More informationLecture 6: Linear Regression
Lecture 6: Linear Regression Reading: Sections 3.1-3 STATS 202: Data mining and analysis Jonathan Taylor, 10/5 Slide credits: Sergio Bacallado 1 / 30 Simple linear regression Model: y i = β 0 + β 1 x i
More informationAdvanced Quantitative Methods: ordinary least squares
Advanced Quantitative Methods: Ordinary Least Squares University College Dublin 31 January 2012 1 2 3 4 5 Terminology y is the dependent variable referred to also (by Greene) as a regressand X are the
More informationCategorical Predictor Variables
Categorical Predictor Variables We often wish to use categorical (or qualitative) variables as covariates in a regression model. For binary variables (taking on only 2 values, e.g. sex), it is relatively
More informationChapter 14 Simple Linear Regression (A)
Chapter 14 Simple Linear Regression (A) 1. Characteristics Managerial decisions often are based on the relationship between two or more variables. can be used to develop an equation showing how the variables
More informationApplied Regression Analysis
Applied Regression Analysis Chapter 3 Multiple Linear Regression Hongcheng Li April, 6, 2013 Recall simple linear regression 1 Recall simple linear regression 2 Parameter Estimation 3 Interpretations of
More informationSTAT 100C: Linear models
STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix
More informationSTOR 455 STATISTICAL METHODS I
STOR 455 STATISTICAL METHODS I Jan Hannig Mul9variate Regression Y=X β + ε X is a regression matrix, β is a vector of parameters and ε are independent N(0,σ) Es9mated parameters b=(x X) - 1 X Y Predicted
More informationBusiness Statistics. Tommaso Proietti. Linear Regression. DEF - Università di Roma 'Tor Vergata'
Business Statistics Tommaso Proietti DEF - Università di Roma 'Tor Vergata' Linear Regression Specication Let Y be a univariate quantitative response variable. We model Y as follows: Y = f(x) + ε where
More informationBias Variance Trade-off
Bias Variance Trade-off The mean squared error of an estimator MSE(ˆθ) = E([ˆθ θ] 2 ) Can be re-expressed MSE(ˆθ) = Var(ˆθ) + (B(ˆθ) 2 ) MSE = VAR + BIAS 2 Proof MSE(ˆθ) = E((ˆθ θ) 2 ) = E(([ˆθ E(ˆθ)]
More informationMultivariate Regression
Multivariate Regression The so-called supervised learning problem is the following: we want to approximate the random variable Y with an appropriate function of the random variables X 1,..., X p with the
More informationSTAT420 Midterm Exam. University of Illinois Urbana-Champaign October 19 (Friday), :00 4:15p. SOLUTIONS (Yellow)
STAT40 Midterm Exam University of Illinois Urbana-Champaign October 19 (Friday), 018 3:00 4:15p SOLUTIONS (Yellow) Question 1 (15 points) (10 points) 3 (50 points) extra ( points) Total (77 points) Points
More informationSimple Linear Regression
Simple Linear Regression September 24, 2008 Reading HH 8, GIll 4 Simple Linear Regression p.1/20 Problem Data: Observe pairs (Y i,x i ),i = 1,...n Response or dependent variable Y Predictor or independent
More informationWeighted Least Squares
Weighted Least Squares The standard linear model assumes that Var(ε i ) = σ 2 for i = 1,..., n. As we have seen, however, there are instances where Var(Y X = x i ) = Var(ε i ) = σ2 w i. Here w 1,..., w
More informationSTAT 4385 Topic 03: Simple Linear Regression
STAT 4385 Topic 03: Simple Linear Regression Xiaogang Su, Ph.D. Department of Mathematical Science University of Texas at El Paso xsu@utep.edu Spring, 2017 Outline The Set-Up Exploratory Data Analysis
More informationCorrelation Analysis
Simple Regression Correlation Analysis Correlation analysis is used to measure strength of the association (linear relationship) between two variables Correlation is only concerned with strength of the
More informationChapter 12: Multiple Linear Regression
Chapter 12: Multiple Linear Regression Seungchul Baek Department of Statistics, University of South Carolina STAT 509: Statistics for Engineers 1 / 55 Introduction A regression model can be expressed as
More informationBANA 7046 Data Mining I Lecture 2. Linear Regression, Model Assessment, and Cross-validation 1
BANA 7046 Data Mining I Lecture 2. Linear Regression, Model Assessment, and Cross-validation 1 Shaobo Li University of Cincinnati 1 Partially based on Hastie, et al. (2009) ESL, and James, et al. (2013)
More informationLecture 3: Inference in SLR
Lecture 3: Inference in SLR STAT 51 Spring 011 Background Reading KNNL:.1.6 3-1 Topic Overview This topic will cover: Review of hypothesis testing Inference about 1 Inference about 0 Confidence Intervals
More informationST505/S697R: Fall Homework 2 Solution.
ST505/S69R: Fall 2012. Homework 2 Solution. 1. 1a; problem 1.22 Below is the summary information (edited) from the regression (using R output); code at end of solution as is code and output for SAS. a)
More informationChapter 2 Inferences in Simple Linear Regression
STAT 525 SPRING 2018 Chapter 2 Inferences in Simple Linear Regression Professor Min Zhang Testing for Linear Relationship Term β 1 X i defines linear relationship Will then test H 0 : β 1 = 0 Test requires
More informationMATH 644: Regression Analysis Methods
MATH 644: Regression Analysis Methods FINAL EXAM Fall, 2012 INSTRUCTIONS TO STUDENTS: 1. This test contains SIX questions. It comprises ELEVEN printed pages. 2. Answer ALL questions for a total of 100
More informationSimple Linear Regression
Simple Linear Regression Reading: Hoff Chapter 9 November 4, 2009 Problem Data: Observe pairs (Y i,x i ),i = 1,... n Response or dependent variable Y Predictor or independent variable X GOALS: Exploring
More informationInference for Regression
Inference for Regression Section 9.4 Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Department of Mathematics University of Houston Lecture 13b - 3339 Cathy Poliak, Ph.D. cathy@math.uh.edu
More informationREGRESSION ANALYSIS AND INDICATOR VARIABLES
REGRESSION ANALYSIS AND INDICATOR VARIABLES Thesis Submitted in partial fulfillment of the requirements for the award of degree of Masters of Science in Mathematics and Computing Submitted by Sweety Arora
More informationLecture 9 SLR in Matrix Form
Lecture 9 SLR in Matrix Form STAT 51 Spring 011 Background Reading KNNL: Chapter 5 9-1 Topic Overview Matrix Equations for SLR Don t focus so much on the matrix arithmetic as on the form of the equations.
More informationChapter 3: Multiple Regression. August 14, 2018
Chapter 3: Multiple Regression August 14, 2018 1 The multiple linear regression model The model y = β 0 +β 1 x 1 + +β k x k +ǫ (1) is called a multiple linear regression model with k regressors. The parametersβ
More informationChapter 2 The Simple Linear Regression Model: Specification and Estimation
Chapter The Simple Linear Regression Model: Specification and Estimation Page 1 Chapter Contents.1 An Economic Model. An Econometric Model.3 Estimating the Regression Parameters.4 Assessing the Least Squares
More informationMultiple Linear Regression
Multiple Linear Regression Simple linear regression tries to fit a simple line between two variables Y and X. If X is linearly related to Y this explains some of the variability in Y. In most cases, there
More informationTMA4255 Applied Statistics V2016 (5)
TMA4255 Applied Statistics V2016 (5) Part 2: Regression Simple linear regression [11.1-11.4] Sum of squares [11.5] Anna Marie Holand To be lectured: January 26, 2016 wiki.math.ntnu.no/tma4255/2016v/start
More information1. The OLS Estimator. 1.1 Population model and notation
1. The OLS Estimator OLS stands for Ordinary Least Squares. There are 6 assumptions ordinarily made, and the method of fitting a line through data is by least-squares. OLS is a common estimation methodology
More informationSample Problems. Note: If you find the following statements true, you should briefly prove them. If you find them false, you should correct them.
Sample Problems 1. True or False Note: If you find the following statements true, you should briefly prove them. If you find them false, you should correct them. (a) The sample average of estimated residuals
More informationHomework 2: Simple Linear Regression
STAT 4385 Applied Regression Analysis Homework : Simple Linear Regression (Simple Linear Regression) Thirty (n = 30) College graduates who have recently entered the job market. For each student, the CGPA
More informationChapter 1: Linear Regression with One Predictor Variable also known as: Simple Linear Regression Bivariate Linear Regression
BSTT523: Kutner et al., Chapter 1 1 Chapter 1: Linear Regression with One Predictor Variable also known as: Simple Linear Regression Bivariate Linear Regression Introduction: Functional relation between
More informationMS&E 226: Small Data
MS&E 226: Small Data Lecture 6: Model complexity scores (v3) Ramesh Johari ramesh.johari@stanford.edu Fall 2015 1 / 34 Estimating prediction error 2 / 34 Estimating prediction error We saw how we can estimate
More informationEstimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X.
Estimating σ 2 We can do simple prediction of Y and estimation of the mean of Y at any value of X. To perform inferences about our regression line, we must estimate σ 2, the variance of the error term.
More informationOrdinary Least Squares Regression
Ordinary Least Squares Regression Goals for this unit More on notation and terminology OLS scalar versus matrix derivation Some Preliminaries In this class we will be learning to analyze Cross Section
More informationRegression Estimation - Least Squares and Maximum Likelihood. Dr. Frank Wood
Regression Estimation - Least Squares and Maximum Likelihood Dr. Frank Wood Least Squares Max(min)imization Function to minimize w.r.t. β 0, β 1 Q = n (Y i (β 0 + β 1 X i )) 2 i=1 Minimize this by maximizing
More informationRegression Estimation Least Squares and Maximum Likelihood
Regression Estimation Least Squares and Maximum Likelihood Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 3, Slide 1 Least Squares Max(min)imization Function to minimize
More informationProblems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B
Simple Linear Regression 35 Problems 1 Consider a set of data (x i, y i ), i =1, 2,,n, and the following two regression models: y i = β 0 + β 1 x i + ε, (i =1, 2,,n), Model A y i = γ 0 + γ 1 x i + γ 2
More informationRegression Analysis Chapter 2 Simple Linear Regression
Regression Analysis Chapter 2 Simple Linear Regression Dr. Bisher Mamoun Iqelan biqelan@iugaza.edu.ps Department of Mathematics The Islamic University of Gaza 2010-2011, Semester 2 Dr. Bisher M. Iqelan
More informationSection 4.6 Simple Linear Regression
Section 4.6 Simple Linear Regression Objectives ˆ Basic philosophy of SLR and the regression assumptions ˆ Point & interval estimation of the model parameters, and how to make predictions ˆ Point and interval
More informationUNIVERSITY OF MASSACHUSETTS. Department of Mathematics and Statistics. Basic Exam - Applied Statistics. Tuesday, January 17, 2017
UNIVERSITY OF MASSACHUSETTS Department of Mathematics and Statistics Basic Exam - Applied Statistics Tuesday, January 17, 2017 Work all problems 60 points are needed to pass at the Masters Level and 75
More information1. Simple Linear Regression
1. Simple Linear Regression Suppose that we are interested in the average height of male undergrads at UF. We put each male student s name (population) in a hat and randomly select 100 (sample). Then their
More informationLecture 6: Linear Regression (continued)
Lecture 6: Linear Regression (continued) Reading: Sections 3.1-3.3 STATS 202: Data mining and analysis October 6, 2017 1 / 23 Multiple linear regression Y = β 0 + β 1 X 1 + + β p X p + ε Y ε N (0, σ) i.i.d.
More informationSimple Linear Regression for the Climate Data
Prediction Prediction Interval Temperature 0.2 0.0 0.2 0.4 0.6 0.8 320 340 360 380 CO 2 Simple Linear Regression for the Climate Data What do we do with the data? y i = Temperature of i th Year x i =CO
More informationConcordia University (5+5)Q 1.
(5+5)Q 1. Concordia University Department of Mathematics and Statistics Course Number Section Statistics 360/1 40 Examination Date Time Pages Mid Term Test May 26, 2004 Two Hours 3 Instructor Course Examiner
More informationStatistical View of Least Squares
Basic Ideas Some Examples Least Squares May 22, 2007 Basic Ideas Simple Linear Regression Basic Ideas Some Examples Least Squares Suppose we have two variables x and y Basic Ideas Simple Linear Regression
More informationSTA 302 H1F / 1001 HF Fall 2007 Test 1 October 24, 2007
STA 302 H1F / 1001 HF Fall 2007 Test 1 October 24, 2007 LAST NAME: SOLUTIONS FIRST NAME: STUDENT NUMBER: ENROLLED IN: (circle one) STA 302 STA 1001 INSTRUCTIONS: Time: 90 minutes Aids allowed: calculator.
More informationReview: General Approach to Hypothesis Testing. 1. Define the research question and formulate the appropriate null and alternative hypotheses.
1 Review: Let X 1, X,..., X n denote n independent random variables sampled from some distribution might not be normal!) with mean µ) and standard deviation σ). Then X µ σ n In other words, X is approximately
More informationLINEAR REGRESSION MODELS W4315
LINEAR REGRESSION MODELS W431 HOMEWORK ANSWERS March 9, 2010 Due: 03/04/10 Instructor: Frank Wood 1. (20 points) In order to get a maximum likelihood estimate of the parameters of a Box-Cox transformed
More informationInference for Regression Simple Linear Regression
Inference for Regression Simple Linear Regression IPS Chapter 10.1 2009 W.H. Freeman and Company Objectives (IPS Chapter 10.1) Simple linear regression p Statistical model for linear regression p Estimating
More informationLecture 6: Linear models and Gauss-Markov theorem
Lecture 6: Linear models and Gauss-Markov theorem Linear model setting Results in simple linear regression can be extended to the following general linear model with independently observed response variables
More informationInference for Regression Inference about the Regression Model and Using the Regression Line
Inference for Regression Inference about the Regression Model and Using the Regression Line PBS Chapter 10.1 and 10.2 2009 W.H. Freeman and Company Objectives (PBS Chapter 10.1 and 10.2) Inference about
More informationMAT2377. Rafa l Kulik. Version 2015/November/26. Rafa l Kulik
MAT2377 Rafa l Kulik Version 2015/November/26 Rafa l Kulik Bivariate data and scatterplot Data: Hydrocarbon level (x) and Oxygen level (y): x: 0.99, 1.02, 1.15, 1.29, 1.46, 1.36, 0.87, 1.23, 1.55, 1.40,
More informationApplied Econometrics (QEM)
Applied Econometrics (QEM) The Simple Linear Regression Model based on Prinicples of Econometrics Jakub Mućk Department of Quantitative Economics Jakub Mućk Applied Econometrics (QEM) Meeting #2 The Simple
More informationApplied Regression. Applied Regression. Chapter 2 Simple Linear Regression. Hongcheng Li. April, 6, 2013
Applied Regression Chapter 2 Simple Linear Regression Hongcheng Li April, 6, 2013 Outline 1 Introduction of simple linear regression 2 Scatter plot 3 Simple linear regression model 4 Test of Hypothesis
More informationChapter 1. Linear Regression with One Predictor Variable
Chapter 1. Linear Regression with One Predictor Variable 1.1 Statistical Relation Between Two Variables To motivate statistical relationships, let us consider a mathematical relation between two mathematical
More information