Mixed-effects Polynomial Regression Models chapter 5

Similar documents
MIXED-EFFECTS POLYNOMIAL REGRESSION MODELS

Advantages of Mixed-effects Regression Models (MRM; aka multilevel, hierarchical linear, linear mixed models) 1. MRM explicitly models individual

Vectors and Matrices Statistics with Vectors and Matrices

GEE for Longitudinal Data - Chapter 8

A Introduction to Matrix Algebra and the Multivariate Normal Distribution

MS&E 226. In-Class Midterm Examination Solutions Small Data October 20, 2015

Weighted Least Squares

Categorical Predictor Variables

Applied Econometrics (QEM)

Logistic Regression. Will Monroe CS 109. Lecture Notes #22 August 14, 2017

Missing Data in Longitudinal Studies: Mixed-effects Pattern-Mixture and Selection Models

Accelerated Advanced Algebra. Chapter 1 Patterns and Recursion Homework List and Objectives

Implementing contrasts using SAS Proc GLM is a relatively straightforward process. A SAS Proc GLM contrast statement has the following form:

Problems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B

Serial Correlation. Edps/Psych/Stat 587. Carolyn J. Anderson. Fall Department of Educational Psychology

Linear Models in Statistics

Describing Change over Time: Adding Linear Trends

Specifying Latent Curve and Other Growth Models Using Mplus. (Revised )

Describing Nonlinear Change Over Time

Simple linear regression

ANOVA approaches to Repeated Measures. repeated measures MANOVA (chapter 3)

An Introduction to Matrix Algebra

General Principles Within-Cases Factors Only Within and Between. Within Cases ANOVA. Part One

Mixed Models II - Behind the Scenes Report Revised June 11, 2002 by G. Monette

Example of Including Nonlinear Components in Regression

Correlation and Regression Theory 1) Multivariate Statistics

Longitudinal Data Analysis of Health Outcomes

Next is material on matrix rank. Please see the handout

Orthogonal and Non-orthogonal Polynomial Constrasts

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

ECON 616: Lecture Two: Deterministic Trends, Nonstationary Processes

Regression. Estimation of the linear function (straight line) describing the linear component of the joint relationship between two variables X and Y.

Journal of Statistical Software

Assessing the relation between language comprehension and performance in general chemistry. Appendices

Lecture 12 Inference in MLR

Principal Components Analysis (PCA)

This module focuses on the logic of ANOVA with special attention given to variance components and the relationship between ANOVA and regression.

Section 4: Multiple Linear Regression

2. How will we adjust our fitting procedure to compensate for fact that the acceleration differs depending on the direction of motion?

Functions: Polynomial, Rational, Exponential

Principal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis. Chris Funk. Lecture 17

Ma 3/103: Lecture 24 Linear Regression I: Estimation

Regression: Lecture 2

STAT 350: Geometry of Least Squares

Chapter R - Review of Basic Algebraic Concepts (26 topics, no due date)

y Xw 2 2 y Xw λ w 2 2

Statistics 910, #5 1. Regression Methods

CS 195-5: Machine Learning Problem Set 1

Notes on empirical methods

ALGEBRA I CURRICULUM GUIDE Overview and Scope & Sequence

ECE521 lecture 4: 19 January Optimization, MLE, regularization

Statistical Distribution Assumptions of General Linear Models

Review of Multiple Regression

Unit Lesson Topic CCSS

EXST 7015 Fall 2014 Lab 08: Polynomial Regression

Rational and Radical Relationships

Introduce Exploration! Before we go on, notice one more thing. We'll come back to the derivation if we have time.

4 Multiple Linear Regression

Optimal Cone. 1 Grade Levels and Time. 2 Objectives and Topics. 3 Introduction. 4 Procedure and Discussion. Grades:11-12

Learning Goals. 2. To be able to distinguish between a dependent and independent variable.

In the bivariate regression model, the original parameterization is. Y i = β 1 + β 2 X2 + β 2 X2. + β 2 (X 2i X 2 ) + ε i (2)

Unit Lesson Topic CCSS

Section IV MODELS FOR MULTILEVEL DATA

Econometrics Honor s Exam Review Session. Spring 2012 Eunice Han

Machine Learning (CS 567) Lecture 5

Purpose: This lab is an experiment to verify Malus Law for polarized light in both a two and three polarizer system.

MATH Spring 2010 Topics per Section

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN

Application of Item Response Theory Models for Intensive Longitudinal Data

Classification 2: Linear discriminant analysis (continued); logistic regression

Least Squares Estimation

Mixed Models for Longitudinal Binary Outcomes. Don Hedeker Department of Public Health Sciences University of Chicago.

Linear Regression. Chapter 3

LECTURE 6. Introduction to Econometrics. Hypothesis testing & Goodness of fit

Interactions among Continuous Predictors

Optimal design of experiments

Stat 135, Fall 2006 A. Adhikari HOMEWORK 10 SOLUTIONS

Linear Regression. Udacity

Why analyze as ordinal? Mixed Models for Longitudinal Ordinal Data Don Hedeker University of Illinois at Chicago

Time-Invariant Predictors in Longitudinal Models

Introduction to Confirmatory Factor Analysis

Analysis of Repeated Measures and Longitudinal Data in Health Services Research

Confidence Intervals and Sets

Linear Regression Model. Badr Missaoui

College Algebra To learn more about all our offerings Visit Knewton.com

PreCalculus. Curriculum (637 topics additional topics)

Multivariate Statistical Analysis

Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation

Estimation: Problems & Solutions

Algebra. Robert Taggart

An Efficient State Space Approach to Estimate Univariate and Multivariate Multilevel Regression Models

Advanced Econometrics

Constructing SAS Contrast/Estimate Statements S. R. Bowley, University of Guelph 2013

Master s Written Examination - Solution

Lecture 11: Regression Methods I (Linear Regression)

Association studies and regression

ML estimation: Random-intercepts logistic model. and z

Business Calculus

Notes on Discriminant Functions and Optimal Classification

Cointegration Lecture I: Introduction

Transcription:

Mixed-effects Polynomial Regression Models chapter 5 1

Figure 5.1 Various curvilinear models: (a) decelerating positive slope; (b) accelerating positive slope; (c) decelerating negative slope; (d) accelerating negative slope 2

Figure 5.2 More curvilinear models: (a) positive to negative slope (β 0 = 2, β 1 = 8, β 2 = 1.2); (b) inverted U-shaped slope (β 0 = 2, β 1 = 11, β 2 = 2.2); (c) negative to positive slope (β 0 = 14, β 1 = 8, β 2 = 1.2); (d) U-shaped slope (β 0 = 14, β 1 = 11, β 2 = 2.2) 3

Expressing Time with Orthogonal Polynomials Instead of X = Z = 1 1 1 1 1 1 0 1 2 3 4 5 0 1 4 9 16 25 use X = Z = 1 1 1 1 1 1 5 3 1 1 3 5 5 1 4 4 1 5 / 6 / 70 / 84 4

Figure 4.5 Intercept variance changes with coding of time 5

Top 10 reasons to use Orthogonal Polynomials 10 They look complicated, so it seems like you know what you re doing 9 With a name like orthogonal polynomials they have to be good 8 Good for practicing lessons learned from hooked on phonics 7 Decompose more quickly than (orthogonal?) polymers 6 Sound better than lame old releases from Polydor records 5 Less painful than a visit to the orthodonist 4 Unlike ortho, won t kill your lawn weeds 3 Might help you with a polygraph test 2 Great conversation topic for getting rid of unwanted friends 1 Because your instructor will give you an F otherwise 6

Real reasons to use Orthogonal Polynomials for balanced data, and CS structure, estimates of polynomial fixed effects β (e.g., constant and linear) won t change when higher-order polynomial terms (e.g., quadratic and cubic) are added to the model in original scale, it gets increasingly difficult to estimate higher-degree terms (coefficients get smaller and smaller as the scale of X p gets larger and larger) avoids high correlation between estimates, which can cause estimation problems provides comparison of importance of different polynomials intercept (and intercept-related parameters) represents the grand mean of Y 7

Orthogonal Polynomials via the Cholesky decomposition in 4 easy steps! Suppose six equally-spaced timepoints T = 1 1 1 1 1 1 0 1 2 3 4 5 0 1 4 9 16 25 1. Compute T T, which yields a symmetric matrix T T = 6 15 55 15 55 225 55 225 979 8

2. Obtain the Cholesky factor S of T T, and express it in transpose form S = 2.4495 6.1237 22.4537 0 4.1833 20.9165 0 0 6.1101 3. Obtain the inverse (S ) 1 (S ) 1 = 0.4082 0.5976 0.5455 0 0.2390 0.8183 0 0 0.1637 9

4. Multiply T by this inverse (S ) 1 T (S ) 1 = 0.4082 0.5976 0.5455 0.4082 0.3586 0.1091 0.4082 0.1195 0.4364 0.4082 0.1195 0.4364 0.4082 0.3586 0.1091 0.4082 0.5976 0.5455 = 1/ 6 5/ 70 5/ 84 1/ 6 3/ 70 1/ 84 1/ 6 1/ 70 4/ 84 1/ 6 1/ 70 4/ 84 1/ 6 3/ 70 1/ 84 1/ 6 5/ 70 5/ 84 which yields the same orthogonal polynomial values as before 10

Orthogonal Polynomials via SAS TITLE producing orthogonal polynomial matrix ; PROC IML; time = { 1 0 0, 1 1 1, 1 2 4, 1 3 9, 1 4 16, 1 5 25 } ; orthpoly = time INV(ROOT(T(time) time)); PRINT time matrix, time [FORMAT=8.4]; PRINT orthogonalized time matrix, orthpoly [FORMAT=8.4]; 11

Model Representations Consider the matrix representation of the MRM: y i = X i β + Z i υ i + ε i replace X with X(S ) 1 and Z with Z(S ) 1 denote parameters in the orthogonal polynomial metric as γ and θ i for the fixed and random effects, respectively The reparameterized model is given as y i = X i (S ) 1 γ + Z i (S ) 1 θ i + ε i 12

Orthogonal Polynomial analysis of Reisby data Parameter Estimate SE Z p < γ 0 43.24 1.37 31.61.0001 γ 1 9.94 0.86 11.50.0001 γ 2 0.31 0.54 0.58.56 σθ 2 0 111.91 21.60 σ θ0 θ 1 37.99 10.92 σθ 2 1 37.04 8.90 σ θ0 θ 2 10.14 6.19 σ θ1 θ 2 0.82 3.80 σθ 2 2 7.23 3.50 σ 2 10.52 1.11 2 log L = 2207.64 13

log-likelihood value is identical to analysis in raw metric as before, only the constant and linear fixed-effect terms are significant (these terms also dominate in terms of magnitude), thus, the population trend is essentially linear The estimated constant variance (ˆσ θ 2 0 ) is much larger than the estimated linear trend component (ˆσ θ 2 1 ), which is much larger than the estimated quadratic trend component (ˆσ θ 2 2 ) 71.7%, 23.7%, and 4.6%, respectively, of the sum of the estimated individual variance terms. At individual level, there is diminishing heterogeneity as the order of the polynomial increases a strong positive association between the constant and linear terms (ˆσ θ 2 1 θ 0 = 37.99, expressed as a correlation =.59) an individual s linear trend is positively associated with their average depression level 14

Can obtain same estimates of overall and individual trends as previous analysis (in raw metric) β = (S ) 1 γ υ i = (S ) 1 θ i e.g., ˆβ = (S ) 1 ˆγ = 0.4082 0.5976 0.5455 0 0.2390 0.8183 0 0 0.1637 43.24 9.94 0.31 = 23.76 2.63 0.05 15

Figure 5.3 Average linear and individual quadratic trends 16

Figure 5.4 Reisby data: estimated curvilinear trends 17

Figure 5.5 Cubic models of generally positive change across time: (a) y = 1 + 5.5t 1.75t 2 + 0.25t 3 (b) y = 5 4.5t + 3.15t 2 0.40t 3 (c) y = 0 0.5t + 2.30t 2 0.325t 3 (d) y = 1 + 10.5t 3.25t 2 + 0.35t 3 18

Figure 5.6 Extrapolation of cubic models across time: (a) y = 1 + 5.5t 1.75t 2 + 0.25t 3 (b) y = 5 4.5t + 3.15t 2 0.40t 3 (c) y = 0 0.5t + 2.30t 2 0.325t 3 (d) y = 1 + 10.5t 3.25t 2 + 0.35t 3 19

Orthogonal Cubic Trend model - Reisby data Parameter Estimate SE Z p < Intercept γ 0 43.24 1.35 31.95.0001 Linear trend γ 1 9.88 0.84 11.70.0001 Quadratic trend γ 2 0.33 0.55 0.60.55 Cubic trend γ 3 0.62 0.48 1.29.20 σθ 2 0 110.91 21.08 σ θ0 θ 1 34.92 10.25 σθ 2 1 36.25 8.25 σ θ0 θ 2 11.14 6.12 σ θ1 θ 2 0.26 3.69 σθ 2 2 9.24 3.58 σ θ0 θ 3 7.25 5.31 σ θ1 θ 3 4.33 3.27 σ θ2 θ 3 4.03 2.16 σθ 2 3 5.04 2.85 σ 2 8.92 1.14 2 log L = 2196.44; χ 2 5 = 11.2, p <.05 (even without dividing by 2) 20

Relative to the previous model with quadratic trend, estimates for intercept, linear trend, and quadratic trends change very little Fixed effects reaffirm average linear response across time Variance estimates clearly diminish as order of polynomial increases (relative percentages of 68.7, 22.5, 5.7, 3.1) Positive association of constant and linear (higher average corresponds to higher slope) 21

Figure 5.7 Reisby data: observed means (solid circles), estimated means (solid line), and estimated individual trends (dotted) 22

Figure 5.8 Reisby data: observed means (solid circles), estimated means (solid line), and estimated individual trends (dotted) for 10 subjects with largest cubic trend components 23

Observed and Estimated Standard Deviations Week 0 Week 1 Week 2 Week 3 Week 4 Week 5 Observed 4.53 4.70 5.49 6.41 6.97 7.22 Model-Based Estimates Random intercept 5.93 5.93 5.93 5.93 5.93 5.93 Random linear trend 4.98 4.91 5.24 5.92 6.84 7.91 Random quadratic trend 4.58 4.88 5.57 6.19 6.78 7.69 Random cubic trend 4.50 4.73 5.31 6.37 7.06 7.32 fit of the observed standard deviations is incrementally better as the order of the polynomial is increased the cubic trend model does an excellent job of fitting the variation in the observed HDRS scores across time 24

Example 5a: Analysis of Riesby dataset. This handout contrasts quadratic trend models using the raw metric of week versus orthogonal polynomials. (SAS code and output) http://tigger.uic.edu/ hedeker/riesquad.txt 25

Example 5b: Converting data from univariate to multivariate format, and getting the observed correlations and variance-covariance matrix of the repeated measures (SAS code and output) http://www.uic.edu/classes/bstt/bstt513/riescov.txt 26

Example 5c: IML code that illustrates the calculation of estimated means and the variance-covariance matrix (SAS code and output) http://tigger.uic.edu/ hedeker/riesfit.txt 27