Remedial Measures Wrap-Up and Transformations Box Cox

Similar documents
Remedial Measures Wrap-Up and Transformations-Box Cox

Diagnostics and Remedial Measures

Remedial Measures, Brown-Forsythe test, F test

Multiple Regression. Dr. Frank Wood. Frank Wood, Linear Regression Models Lecture 12, Slide 1

REGRESSION DIAGNOSTICS AND REMEDIAL MEASURES

STAT Checking Model Assumptions

Regression Estimation Least Squares and Maximum Likelihood

Response Surface Methods

3. Diagnostics and Remedial Measures

Formal Statement of Simple Linear Regression Model

Chapter 3. Diagnostics and Remedial Measures

Inference in Regression Analysis

Lecture 7 Remedial Measures

Advanced Regression Topics: Violation of Assumptions

Diagnostics and Transformations Part 2

Introduction to Statistical modeling: handout for Math 489/583

CS 147: Computer Systems Performance Analysis

STAT 571A Advanced Statistical Regression Analysis. Chapter 3 NOTES Diagnostics and Remedial Measures

Box-Cox Transformations

The Model Building Process Part I: Checking Model Assumptions Best Practice

Chapter 12 - Part I: Correlation Analysis

The Model Building Process Part I: Checking Model Assumptions Best Practice (Version 1.1)

Lecture 4: Types of errors. Bayesian regression models. Logistic regression

Math 5305 Notes. Diagnostics and Remedial Measures. Jesse Crawford. Department of Mathematics Tarleton State University

Machine Learning for NLP

Correlation and Regression Theory 1) Multivariate Statistics

Logistic regression. 11 Nov Logistic regression (EPFL) Applied Statistics 11 Nov / 20

Checking model assumptions with regression diagnostics

Topic 23: Diagnostics and Remedies

COS513: FOUNDATIONS OF PROBABILISTIC MODELS LECTURE 9: LINEAR REGRESSION

Outline. Review regression diagnostics Remedial measures Weighted regression Ridge regression Robust regression Bootstrapping

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2

MLR Interactions and variable transformations

Example: Suppose Y has a Poisson distribution with mean

Diagnostics and Remedial Measures: An Overview

Preface Introduction to Statistics and Data Analysis Overview: Statistical Inference, Samples, Populations, and Experimental Design The Role of

THE PRINCIPLES AND PRACTICE OF STATISTICS IN BIOLOGICAL RESEARCH. Robert R. SOKAL and F. James ROHLF. State University of New York at Stony Brook

Statistics Boot Camp. Dr. Stephanie Lane Institute for Defense Analyses DATAWorks 2018

Machine Learning Linear Classification. Prof. Matteo Matteucci

Simple Linear Regression. Part

Lecture 15 Multiple regression I Chapter 6 Set 2 Least Square Estimation The quadratic form to be minimized is

Last updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

* Tuesday 17 January :30-16:30 (2 hours) Recored on ESSE3 General introduction to the course.

Chapter 3. Measuring data

Regression With a Categorical Independent Variable: Mean Comparisons

Modeling Data with Linear Combinations of Basis Functions. Read Chapter 3 in the text by Bishop

Final Review. Yang Feng. Yang Feng (Columbia University) Final Review 1 / 58

Regression. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh.

Machine Learning. Regression-Based Classification & Gaussian Discriminant Analysis. Manfred Huber

Machine Learning and Data Mining. Linear regression. Kalev Kask

Start with review, some new definitions, and pictures on the white board. Assumptions in the Normal Linear Regression Model

Machine Learning Basics: Maximum Likelihood Estimation

DESIGNING EXPERIMENTS AND ANALYZING DATA A Model Comparison Perspective

Linear Regression 1 / 25. Karl Stratos. June 18, 2018

1 The Classic Bivariate Least Squares Model

Gaussian and Linear Discriminant Analysis; Multiclass Classification

Lecture 3. Linear Regression II Bastian Leibe RWTH Aachen

LINEAR REGRESSION MODELS W4315

Lecture 18 Miscellaneous Topics in Multiple Regression

Statistical Machine Learning Hilary Term 2018

Remedial Measures for Multiple Linear Regression Models

A Short Introduction to Curve Fitting and Regression by Brad Morantz

Cox s proportional hazards/regression model - model assessment

Gaussian discriminant analysis Naive Bayes

Lecture 16 Solving GLMs via IRWLS

Linear Models for Regression

Prediction Intervals in the Presence of Outliers

Correlation and Regression

Machine Learning for NLP

Experimental Design and Data Analysis for Biologists

Student Guide: Chapter 1

Need for Several Predictor Variables

Contents. Acknowledgments. xix

Generalized Linear Models. Last time: Background & motivation for moving beyond linear

Regression Estimation - Least Squares and Maximum Likelihood. Dr. Frank Wood

Introduction to Linear regression analysis. Part 2. Model comparisons

Ch 4. Linear Models for Classification

Multicollinearity occurs when two or more predictors in the model are correlated and provide redundant information about the response.

Logistic Regression Review Fall 2012 Recitation. September 25, 2012 TA: Selen Uguroglu

ROBERTO BATTITI, MAURO BRUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Apr 2015

Sampling distribution of GLM regression coefficients

Linear Models for Regression CS534

Biost 518 Applied Biostatistics II. Purpose of Statistics. First Stage of Scientific Investigation. Further Stages of Scientific Investigation

Prentice Hall Stats: Modeling the World 2004 (Bock) Correlated to: National Advanced Placement (AP) Statistics Course Outline (Grades 9-12)

Accelerated Advanced Algebra. Chapter 1 Patterns and Recursion Homework List and Objectives

CSC2515 Winter 2015 Introduction to Machine Learning. Lecture 2: Linear regression

ISQS 5349 Spring 2013 Final Exam

Nonlinear Programming (Hillier, Lieberman Chapter 13) CHEM-E7155 Production Planning and Control

Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation

Six Sigma Black Belt Study Guides

Learning Objectives. c D. Poole and A. Mackworth 2010 Artificial Intelligence, Lecture 7.2, Page 1

Linear Regression In God we trust, all others bring data. William Edwards Deming

Linear Regression 9/23/17. Simple linear regression. Advertising sales: Variance changes based on # of TVs. Advertising sales: Normal error?

Multiple Regression and Model Building Lecture 20 1 May 2006 R. Ryznar

LECTURE 11. Introduction to Econometrics. Autocorrelation

STA 4273H: Sta-s-cal Machine Learning

Support Vector Machines. CAP 5610: Machine Learning Instructor: Guo-Jun QI

Overview. Probabilistic Interpretation of Linear Regression Maximum Likelihood Estimation Bayesian Estimation MAP Estimation

Assumptions in Regression Modeling

Regression Graphics. 1 Introduction. 2 The Central Subspace. R. D. Cook Department of Applied Statistics University of Minnesota St.

Transcription:

Remedial Measures Wrap-Up and Transformations Box Cox Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 1

Last Class Graphical procedures for determining appropriateness of regression fit Normal probability plot Tests to determine Constancy of error variance Appropriateness of linear fit What do we if we determine (through testing or otherwise) that the linear regression fit is not good? Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 2

Overview of Remedial Measures If simple regression model is not appropriate there are two choices 1. Abandon simple regression model and develop and use a more appropriate model 2. Employ some transformation of the data so that the simple regression model is appropriate for the transformed data. Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 3

Fixes For Nonlinearity of regression function Transformation(s) (today) Nonconstancy of error variance Weighted least squares (nice project idea, coming later in class) and transformations Nonindependence of error terms Directly model correlation or use first differences (may skip) Nonnormality of error terms Transformation(s) (today) Omission of important predictor variables Multiple regression coming soon Outlying observations Robust regression (another nice project idea) Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 4

Nonlinearity of regression function Direct approach Modify regression model by altering the nature of the regression function. For instance a quadratic regression function might be used E{Y}=β 0 +β 1 X+β 2 X 2 or an exponential function E{Y}=β 0 β X 1 Such approaches employ a transformation to (approximately) linearize a regression function Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 5

Quick Questions How would you fit such models? How does the exponential regression function relate to regular linear regression? Where did the error terms go? Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 6

Transformations Transformations for Nonlinear Relation Only Appropriate when the distribution of the error terms is reasonably close to a normal distribution In this situation transformation of X should be attempted transformation of Y should not be attempted because it will materially effect the distribution of the error terms Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 7

Prototype Regression Patterns Note constancy error terms Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 8

Experiment Example X : days of training received Y : sales performance (score) Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 9

X = sqrt(x) Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 10

Example Data Transformation Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 11

Graphical Residual Analysis Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 12

Matlab Run matlab_demos\transform_x.m Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 13

Transformations on Y Nonnormality and unequal variances of error terms frequently appear together To remedy these in the normal regression model we need a transformation on Y This is because Shapes and spreads of distributions of Y need to be changed May help linearize a curvilinear regression relation Can be combined with transformation on X Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 14

Prototype Regression Patterns and Y Transformations Note change in error distribution as function of input Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 15

Example Use of logarithmic transformation of Y to linearize regression relations and stabilize error variance Data on age (X) and plasma level of a polyamine (Y) for a portion of the 25 healthy children in a study. Younger children exhibit greater variability than older children. Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 16

Plasma level versus age Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 17

Associated Data Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 18

If we fit a simple linear regression line to the log transformed Y data we obtain And the coefficient of correlation between the ordered residuals and their expected values under normality is.981 (for α =.05 B.6 in the book shows a critical value of.959) Normality of error terms supported, regression model for transformed Y data appropriate Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 19

Box Cox Transforms It can be difficult to graphically determine which transformation of Y is most appropriate for correcting skewness of the distributions of error terms unequal variances nonlinearity of the regression function The Box-Cox procedure automatically identifies a transformation from the family of power transformations on Y Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 20

Box Cox Transforms This family is of the form Examples include Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 21

Box Cox Cont. The normal error regression model with the response variable a member of the family of power transformations becomes This model has an additional parameter that needs to be estimated Maximum likelihood is a way to estimate this parameter Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 22

Box Cox Maximum Likelihood Estimation Before setting up maximum likelihood estimation, the observations are further standardized so that the magnitude of the error sum of squares does not depend on the value of λ The transformation is given by where geometric mean Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 23

Box Cox Maximum Likelihood Estimation Maximize log(l(x,y,σ,λ,b 1,b 0 ))= i (W i (b 1 X i +b 0 )) 2 2σ 2 nlog(σ) w.r.t. λ, σ, b 1, and b 0 How? Take partial derivatives Solve or... gradient ascent methods Show box_cox_demo.m Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 24

Comments on Box Cox The Box-Cox procedure is ordinarily used only to provide a guide for selecting a transformation At times, theoretical or other a priori considerations can be utilized to help in choosing an appropriate transformation It is important to perform residual analysis after the transformation to ensure that the transformation is appropriate When transformed models are employed, b 0 and b 1 obtained via least squares have the least squares property w.r.t. the transformed observations not the original ones. Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 25