Polynomial Regression Models

Similar documents
LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

Chapter 11: Simple Linear Regression and Correlation

LINEAR REGRESSION ANALYSIS. MODULE VIII Lecture Indicator Variables

is the calculated value of the dependent variable at point i. The best parameters have values that minimize the squares of the errors

Chapter 13: Multiple Regression

x = , so that calculated

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Comparison of Regression Lines

Negative Binomial Regression

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

See Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition)

Nice plotting of proteins II

Economics 130. Lecture 4 Simple Linear Regression Continued

Lecture 2: Prelude to the big shrink

Durban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications

Department of Quantitative Methods & Information Systems. Time Series and Their Components QMIS 320. Chapter 6

Statistics II Final Exam 26/6/18

Linear Regression Analysis: Terminology and Notation

Chapter 8 Indicator Variables

Chapter 15 - Multiple Regression

[The following data appear in Wooldridge Q2.3.] The table below contains the ACT score and college GPA for eight college students.

Tests of Exclusion Restrictions on Regression Coefficients: Formulation and Interpretation

Chapter 15 Student Lecture Notes 15-1

STAT 3008 Applied Regression Analysis

NUMERICAL DIFFERENTIATION

Introduction to Regression

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

NANYANG TECHNOLOGICAL UNIVERSITY SEMESTER I EXAMINATION MTH352/MH3510 Regression Analysis

Least squares cubic splines without B-splines S.K. Lucas

THE ROYAL STATISTICAL SOCIETY 2006 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE

Statistics for Managers Using Microsoft Excel/SPSS Chapter 14 Multiple Regression Models

Lecture 13 APPROXIMATION OF SECOMD ORDER DERIVATIVES

x i1 =1 for all i (the constant ).

Department of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution

where I = (n x n) diagonal identity matrix with diagonal elements = 1 and off-diagonal elements = 0; and σ 2 e = variance of (Y X).

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

Lecture 16 Statistical Analysis in Biomaterials Research (Part II)

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

BAYESIAN CURVE FITTING USING PIECEWISE POLYNOMIALS. Dariusz Biskup

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

( ) [ ( k) ( k) ( x) ( ) ( ) ( ) [ ] ξ [ ] [ ] [ ] ( )( ) i ( ) ( )( ) 2! ( ) = ( ) 3 Interpolation. Polynomial Approximation.

First Year Examination Department of Statistics, University of Florida

STAT 3340 Assignment 1 solutions. 1. Find the equation of the line which passes through the points (1,1) and (4,5).

Module Contact: Dr Susan Long, ECO Copyright of the University of East Anglia Version 1

Here is the rationale: If X and y have a strong positive relationship to one another, then ( x x) will tend to be positive when ( y y)

Correlation and Regression. Correlation 9.1. Correlation. Chapter 9

The Ordinary Least Squares (OLS) Estimator

Chapter 9: Statistical Inference and the Relationship between Two Variables

Y = β 0 + β 1 X 1 + β 2 X β k X k + ε

The Quadratic Trigonometric Bézier Curve with Single Shape Parameter

Introduction to Vapor/Liquid Equilibrium, part 2. Raoult s Law:

Statistics for Business and Economics

Reminder: Nested models. Lecture 9: Interactions, Quadratic terms and Splines. Effect Modification. Model 1

Marginal Effects of Explanatory Variables: Constant or Variable? 1. Constant Marginal Effects of Explanatory Variables: A Starting Point

Modeling and Simulation NETW 707

T E C O L O T E R E S E A R C H, I N C.

Chapter 12 Analysis of Covariance

Statistics MINITAB - Lab 2

ECONOMICS 351*-A Mid-Term Exam -- Fall Term 2000 Page 1 of 13 pages. QUEEN'S UNIVERSITY AT KINGSTON Department of Economics

Psychology 282 Lecture #24 Outline Regression Diagnostics: Outliers

Lecture 6: Introduction to Linear Regression

Module 9. Lecture 6. Duality in Assignment Problems

Linear Approximation with Regularization and Moving Least Squares

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.

Biostatistics 360 F&t Tests and Intervals in Regression 1

e i is a random error

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore

CSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography

However, since P is a symmetric idempotent matrix, of P are either 0 or 1 [Eigen-values

Chapter 5 Multilevel Models

Limited Dependent Variables

Problem Set 9 Solutions

Now we relax this assumption and allow that the error variance depends on the independent variables, i.e., heteroskedasticity

Global Sensitivity. Tuesday 20 th February, 2018

January Examinations 2015

Convexity preserving interpolation by splines of arbitrary degree

Number of cases Number of factors Number of covariates Number of levels of factor i. Value of the dependent variable for case k

ISSN: ISO 9001:2008 Certified International Journal of Engineering and Innovative Technology (IJEIT) Volume 3, Issue 1, July 2013

Statistics Chapter 4

since [1-( 0+ 1x1i+ 2x2 i)] [ 0+ 1x1i+ assumed to be a reasonable approximation

SPANC -- SPlitpole ANalysis Code User Manual

Chapter 3. Two-Variable Regression Model: The Problem of Estimation

Case A. P k = Ni ( 2L i k 1 ) + (# big cells) 10d 2 P k.

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix

n α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

β0 + β1xi. You are interested in estimating the unknown parameters β

Lecture 9: Linear regression: centering, hypothesis testing, multiple covariates, and confounding

Fall 2012 Analysis of Experimental Measurements B. Eisenstein/rev. S. Errede

Lecture 3: Probability Distributions

Chapter Newton s Method

Diagnostics in Poisson Regression. Models - Residual Analysis

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

4.3 Poisson Regression

Salmon: Lectures on partial differential equations. Consider the general linear, second-order PDE in the form. ,x 2

Lecture Notes for STATISTICAL METHODS FOR BUSINESS II BMGT 212. Chapters 14, 15 & 16. Professor Ahmadi, Ph.D. Department of Management

Transcription:

LINEAR REGRESSION ANALYSIS MODULE XII Lecture - 6 Polynomal Regresson Models Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur

Test of sgnfcance To test the sgnfcance of hghest order term, we test the null hypothess H :. α k = Ths hypothess s equvalent to We would use H : β k = n polynomal regresson model. F = reg ( α ) SS ( k) / ( n k ) res SS k n ˆ αk Pk( x) y = = SS ( k) / ( n k ) res ~ F(, n k ) under H. If order of the model s changed to ( k r), we need to compute only r new coeffcents. The remanng coeffcents ˆ α, ˆ α,..., ˆ αk do not change due to orthogonalty property of polynomals. Thus the sequental fttng of the model s computatonally easy. When X are equally spaced, the tables of orthogonal polynomals are avalable and the orthogonal polynomals can be easly constructed.

Frst 7 orthogonal polynomals are as follows: Let d be the spacng between levels of x and values. The tables are avalable. { λ j } be the constants chosen so that polynomals wll have nteger P( x ) = x x P( x ) = λ d ( x x) n P( x ) = λ d ( x x) x x n 7 P( x ) = λ d d 4 ( x x) x x n ( n )( n 9) P4( x ) = λ4 d d 4 56 5 ( x x) 5 x x 4 x x P5( x ) = λ5 ( n 7) ( 5n n 47) ) d 8 d 8 d ( x x) 5 x x 4 x x P6( x ) = λ6 ( n ) ( 5n n 9) ) d 44 d 76 d 6 4 5 ( n )( n 4784 9)( n 5)

An example of the table for n = 5 s as follows: 4 n x P P P P 4 - - - - - -4 5 { P ( )} j x 4 7 = λ The orthogonal polynomals can also be constructed when x s are not equally spaced. Pecewse polynomal (Splnes) Sometmes t s exhbted n the data that a lower order polynomal does not provde a good ft. A possble soluton n such stuaton s to ncrease the order of the polynomal but t may always not work. The hgher order polynomal may not mprove the ft sgnfcantly. Such stuatons can be analyzed through resduals, e.g., the resdual sum of square may not stablze or the resdual plots fal to explan the unexplaned structure. One possble reason for such happenng s that the response functon has dfferent behavor n dfferent ranges of ndependent varables. Ths type of problems can be overcome by fttng an approprate functon n dfferent ranges of explanatory varable. So polynomal wll be ftted nto peces. The splne functon can be used for such fttng of polynomal n peces. 5 6 5

Splnes and knots The pecewse polynomals are called splnes. The jont ponts of such peces are called as knots. If polynomal s of order k, then the splne s a contnuous functon wth (k - ) contnuous dervatves. For ths, the functon values and frst (k - ) dervatves agree at the knots. 5 Cubc splnes: For example, consder a cubc splne wth h knots. Suppose the knots are contnuous frst and second dervatves at these knots. Ths can be expressed as t < t <... < th and cubc splne has where h j oj j= = Ey ( ) = Sx ( ) = β x β ( x t) x t f x t > ( x t ) = f x t. It s assumed that the poston of knots are known. Under ths assumpton, ths model can be ftted usng the usual fttng methods of regresson analyss lke least squares prncpal. In case, the knot postons are unknown, then they can be consdered as unknown parameters whch can be estmated. But n such stuaton, the model becomes non-lnear and methods of non-lnear regresson can be used.

Issue of number and poston of knots It s not so smple to know the number and poston of knots n a gven set of data. It s tred to keep the number of knots as mnmum as possble and each segment should have mnmum four or fve data ponts. There should not be more than one extreme pont and one pont of nflexon n each segment. If such ponts are to be accommodated, then t s suggested to keep the extreme pont n the center of segment and pont of nflexon near the knots. It s also possble to ft the polynomals of dfferent orders n each segment and to mpose dfferent contnuty restrctons at the knots. Suppose t s to be accomplshed n a cubc splne model. If all cubc splne model wthout contnuty restrctons s ( h ) peces of polynomal are cubc, then a 6 where h j oj j= = j= j Ey ( ) = Sx ( ) = β x β ( x t) f x > ( x t ) = f x. If the term β ( ) j j x t s n the model, then j th dervatve of S (x) at t s dscontnuous. If the term ( x t ) j s not n the model, then j th dervatve of S (x) s contnuous at t. β j So the model s ftted better when requred contnuty restrctons are fewer because then more parameters wll be ncluded n the model. If more contnuty restrctons are needed, then t ndcates that the model s not well ftted but the fnally ftted curve wll be smoother. The test of hypothess n multple regresson model can be used to determne the order of polynomal segments and contnuty restrctons.

Example Suppose there s only one knot at t n a cubc splne wthout contnuty restrctons gven by 7 E( y) = S( x) = β β x β x β x β ( x t) β ( x t) β ( x t) β ( x t). The term nvolvng β and are present n the model, so, ts frst dervatve and second dervatve, β β Sx ( ) S'( x) are not necessarly contnuous at t. Next queston arses s how to judge the qualty of ft. Ths can be done by S"( x) test of hypothess as follows: H : β = tests the contnuty of Sx ( ) H : β = β = H : β = β = β = tests the contnuty of Sx ( ) and S'( x) tests the contnuty of Sx ( ), S'( x ) and S"( x). The test H : β = β = β = β = ndcates that cubc splne fts data better than a sngle cubc polynomal over the range of explanatory varable x. Ths approach s not satsfactory f the knots are large n number as ths makes X X ll-condtoned. Ths problem s solved by usng cubc B- splne whch are defned as ( x tj ) B ( x) =,,,..., 4 = h j= 4 ( tj tm) m = 4 m j h 4 Ey ( ) = Sx ( ) = γ B( x) = where γ ' s ( =,,..., h 4) are parameters to be estmated. There are eght more knots - t < t < t < t and t. h < th < th < th 4 Choose t = xmn, th = xmax and other knots arbtrarly.

8 Polynomal models n two or more varables The technques of fttng of polynomal model n one varable can be extended to fttng of polynomal models n two or more varables. A second order polynomal s more used n practce and ts model s specfed by y= β β x β x β x β x β xx ε. Ths s also termed as response surface. The methodology of response surface s used to ft such models and helps n desgnng an experment.