ECAS Summer Course. Quantile Regression for Longitudinal Data. Roger Koenker University of Illinois at Urbana-Champaign

Similar documents
Quantile Regression Methods for Reference Growth Charts

Quantile Regression for Longitudinal Data

Nonparametric Quantile Regression

Quantile Regression for Panel/Longitudinal Data

Quantile Regression for Dynamic Panel Data

ECAS Summer Course. Fundamentals of Quantile Regression Roger Koenker University of Illinois at Urbana-Champaign

A DYNAMIC QUANTILE REGRESSION TRANSFORMATION MODEL FOR LONGITUDINAL DATA

Quantile Regression: Inference

Quantile Regression: Inference

Regression based principal component analysis for sparse functional data with applications to screening pubertal growth paths.

Confidence Intervals for Low-dimensional Parameters with High-dimensional Data

Penalty Methods for Bivariate Smoothing and Chicago Land Values

Mixed Effects Multivariate Adaptive Splines Model for the Analysis of Longitudinal and Growth Curve Data. Heping Zhang.

Beyond the Average Man: The Art of Unlikelihood

Supporting Information for Estimating restricted mean. treatment effects with stacked survival models

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract

Integrated Likelihood Estimation in Semiparametric Regression Models. Thomas A. Severini Department of Statistics Northwestern University

Quantile Regression Computation: From Outside, Inside and the Proximal

Motivational Example

Modeling Real Estate Data using Quantile Regression

Stat 579: Generalized Linear Models and Extensions

A Short Introduction to the Lasso Methodology

Asymptotic standard errors of MLE

Reference growth charts for Saudi Arabian children and adolescents

Quantile Processes for Semi and Nonparametric Regression

Lecture 22 Survival Analysis: An Introduction

INFERENCE ON QUANTILE REGRESSION FOR HETEROSCEDASTIC MIXED MODELS

Technical Vignette 5: Understanding intrinsic Gaussian Markov random field spatial models, including intrinsic conditional autoregressive models

Spring 2017 Econ 574 Roger Koenker. Lecture 14 GEE-GMM

Course topics (tentative) The role of random effects

On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models

Reference Growth Charts for Saudi Arabian Children and Adolescents

Variable Selection for Highly Correlated Predictors

A Blockwise Descent Algorithm for Group-penalized Multiresponse and Multinomial Regression

Generalized Elastic Net Regression

Statistics 203: Introduction to Regression and Analysis of Variance Course review

Quantile Regression Computation: From the Inside and the Outside

Modelling the Covariance

Nonparametric Regression. Badr Missaoui

Building a Prognostic Biomarker

Sample Size and Power Considerations for Longitudinal Studies

Single Index Quantile Regression for Heteroscedastic Data

Outline. Mixed models in R using the lme4 package Part 3: Longitudinal data. Sleep deprivation data. Simple longitudinal data

arxiv: v2 [stat.me] 4 Jun 2016

On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models

An Encompassing Test for Non-Nested Quantile Regression Models

Hierarchical Modeling for Univariate Spatial Data

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices

Quantile regression for longitudinal data using the asymmetric Laplace distribution

Gaussian Process Regression

Quantile Regression: Past and Prospects

Supplementary Material to General Functional Concurrent Model

On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models

1 Mixed effect models and longitudinal data analysis

Inversion Base Height. Daggot Pressure Gradient Visibility (miles)

Recap. HW due Thursday by 5 pm Next HW coming on Thursday Logistic regression: Pr(G = k X) linear on the logit scale Linear discriminant analysis:

Bayesian Linear Regression

The Wishart distribution Scaled Wishart. Wishart Priors. Patrick Breheny. March 28. Patrick Breheny BST 701: Bayesian Modeling in Biostatistics 1/11

Recovering Copulae from Conditional Quantiles

A short introduction to INLA and R-INLA

For more information about how to cite these materials visit

Regression, Ridge Regression, Lasso

Chapter 3. Linear Models for Regression

Penalized Empirical Likelihood and Growing Dimensional General Estimating Equations

Fixed Effects, Invariance, and Spatial Variation in Intergenerational Mobility

Issues on quantile autoregression

Bayesian linear regression

Hierarchical Modelling for Univariate Spatial Data

Quantile Regression: 40 Years On

Asymmetric least squares estimation and testing

Regularization Parameter Selection for a Bayesian Multi-Level Group Lasso Regression Model with Application to Imaging Genomics

STAT 512 sp 2018 Summary Sheet

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

Lecture 5 Models and methods for recurrent event data

Bayesian Estimation of Input Output Tables for Russia

Economics 536 Lecture 7. Introduction to Specification Testing in Dynamic Econometric Models

mboost - Componentwise Boosting for Generalised Regression Models

Some properties of Likelihood Ratio Tests in Linear Mixed Models

Practical Bayesian Quantile Regression. Keming Yu University of Plymouth, UK

High-dimensional Ordinary Least-squares Projection for Screening Variables

WU Weiterbildung. Linear Mixed Models

Quantile Regression: An Introductory Overview

Quantile regression 40 years on

FuncICA for time series pattern discovery

CALCULATION METHOD FOR NONLINEAR DYNAMIC LEAST-ABSOLUTE DEVIATIONS ESTIMATOR

Testing Error Correction in Panel data

Introduction to Regression

Spatial Process Estimates as Smoothers: A Review

Lecture 6: Methods for high-dimensional problems

CPSC 540: Machine Learning

Mixed effects modeling with missing data using quantile regression and joint modeling

Statistics 203: Introduction to Regression and Analysis of Variance Penalized models

High-dimensional regression

Part 6: Multivariate Normal and Linear Models

Log Gaussian Cox Processes. Chi Group Meeting February 23, 2016

A Penalized spline approach to functional mixed effects model analysis

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures

Hypothesis Testing Based on the Maximum of Two Statistics from Weighted and Unweighted Estimating Equations

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines

Hierarchical Modelling for Univariate Spatial Data

Transcription:

ECAS Summer Course 1 Quantile Regression for Longitudinal Data Roger Koenker University of Illinois at Urbana-Champaign La Roche-en-Ardennes: September 2005 Part I: Penalty Methods for Random Effects Part II: Estimation of Reference Growth Charts Part II is based on joint work with: Ying Wei, Columbia, Biostatistics Anneli Pere, Children s Hospital, Helsinki and Xuming He, UIUC.

Penalty Methods and Random Effects 2 Classical Models with Independent Random Effects Additive random effects have nice conditional mean functions and nasty conditional quantile functions. Transformations to annihilate fixed effects fail for quantile regression models

Penalty Methods and Random Effects 2 Classical Models with Independent Random Effects Additive random effects have nice conditional mean functions and nasty conditional quantile functions. Transformations to annihilate fixed effects fail for quantile regression models Classical Models with Fixed Effects Fixed effect models are computationally tractable Penalty methods for fixed effect models offer an alternative to classical random effects estimation

A Classical Result 3 Model: or, y ij = x ijβ + α i + u ij j = 1,...m i, i = 1,..., n, y = Xβ + Zα + u. If u and α are independent Gaussian vectors with u N (0, R) and α N (0, Q), the optimal weighted least squares estimator is, ˆβ = (X (R + ZQZ ) 1 X) 1 X (R + ZQZ ) 1 y.

A Classical Result 3 Model: or, y ij = x ijβ + α i + u ij j = 1,...m i, i = 1,..., n, y = Xβ + Zα + u. If u and α are independent Gaussian vectors with u N (0, R) and α N (0, Q), the optimal weighted least squares estimator is, ˆβ = (X (R + ZQZ ) 1 X) 1 X (R + ZQZ ) 1 y. Theorem ˆβ = argmin (α,β) y Xβ Zα 2 R 1 + α 2 Q 1, where x 2 A = x Ax.

Commentary 4 History Henderson (1950) on dairy cows, Goldberger (1962) on BLUP, Lindley and Smith (1972) on Bayesian hierarchical models, and Robinson (1991) for a survey.

Commentary 4 History Henderson (1950) on dairy cows, Goldberger (1962) on BLUP, Lindley and Smith (1972) on Bayesian hierarchical models, and Robinson (1991) for a survey. Interpretation Maximum likelihood estimations for the Gaussian random effects models is identical to penalized least squares estimation of the fixed effects model. Random effects is just shrinkage estimation.

Commentary 4 History Henderson (1950) on dairy cows, Goldberger (1962) on BLUP, Lindley and Smith (1972) on Bayesian hierarchical models, and Robinson (1991) for a survey. Interpretation Maximum likelihood estimations for the Gaussian random effects models is identical to penalized least squares estimation of the fixed effects model. Random effects is just shrinkage estimation. Ideology Our focus is on design of effective algorithms for data analysis. Models are useful as an interpretative device, but they shouldn t be allowed to become sacred objects.

Quantile Regression with Fixed Effects 5 Model: Q yij (τ x ij ) = α i + x ijβ(τ) j = 1,...m i, i = 1,..., n.

Quantile Regression with Fixed Effects 5 Model: Q yij (τ x ij ) = α i + x ijβ(τ) j = 1,...m i, i = 1,..., n. Estimator: With (the inevitable) ρ τ (u) = u(τ I(u < 0)), min (α,β) q k=1 n j=1 m i i=1 w k ρ τk (y ij α i x ijβ(τ k ))

Quantile Regression with Fixed Effects 5 Model: Q yij (τ x ij ) = α i + x ijβ(τ) j = 1,...m i, i = 1,..., n. Estimator: With (the inevitable) ρ τ (u) = u(τ I(u < 0)), min (α,β) q k=1 n j=1 m i i=1 w k ρ τk (y ij α i x ijβ(τ k )) Implementation: Linear programming formulation Interior point algorithms exploiting sparse linear algebra

Quantile Regression with Penalized Fixed Effects 6 Model: (as before) Q yij (τ x ij ) = α i + x ijβ(τ) j = 1,...m i, i = 1,..., n. Estimator: min (α,β) q k=1 Limiting Forms: n j=1 m i i=1 w k ρ τk (y ij α i x ijβ(τ k )) + λ q α i i=1 as λ 0 we obtain fixed effects, as λ we force all α i to 0,

L 1 versus L 2 Shrinkage 7 α 2 0 2 4 6 8 10 α 4 2 0 2 4 6 8 0 1 2 3 4 5 λ 0 1 2 3 4 5 λ Figure 1: Shrinkage of the fixed effect parameter estimates, ˆα i. The left panel illustrates an example of the l 1 shrinkage effect. The right panel illustrates an example of the l 2 shrinkage effect. There are m = 5 observations on each of n = 50 cross-sectional units.

Sparsity 8 row 800 600 400 200 10 20 30 40 50 column Figure 2: Sparsity of the design matrix for the penalized fixed effects quantile regression estimator. Grey areas indicate portions of the matrix that are nonzero, white space indicates zeros. Here, n = 50 and m i = 5 for all i. Three quantiles are being simultaneously estimated with the fixed effects restricted to be identical across quantiles.

Regularity Conditions 9 A1. The y ij are independent with conditional distribution functions, F ij, given x ij, and differentiable conditional densities, 0 < f ij <, with bounded derivatives f ij, at ξ ij(τ), for j = 1,..., m, i = 1,..., n, A2. Let ω = τ(1 τ) and denote Φ = diag(f ij (ξ ij (τ))). The limiting forms of the following matrices are positive definite: D 0 = lim m n ω m ( ( D 1 = lim m 1 m n A3. max 1 i n x ij < M. 1 j m Z Z X Z/ n Z ΦZ X ΦZ/ n Z X/ n X X/n ), Z ΦX/ n X ΦX/n ).

A Theorem 10 Theorem 1. Under conditions A1-3, with n a /m 0 for some a > 0, the ˆδ 1 component of the minimizer, ˆδ, converges in distribution to a Gaussian random vector with mean zero and covariance matrix given by the lower p by p block of the matrix D1 1 D 0D1 1.

More Regularity Conditions 11 B1. The y ij are independent with conditional distribution functions, F ij, given x ij, and differentiable conditional densities, 0 < f ij <, with bounded derivatives f ij, at ξ ij(τ), for j = 1,..., m, i = 1,..., n, B2. Let Ω denote the q by q matrix with typical element τ k τ l τ k τ l and Φ k = diag(f ij (ξ ij (τ k ))). The limiting forms of the following matrices are positive definite: ( D 0 = lim m 1 m n D 1 = lim m 1 m n w ΩwZ Z W Ωw X Z/ n w ΩW Z X/ n W ΩW X X/n ), wk Z Φ k Z w 1 Z Φ 1 X/ n w q Z Φ q X/ n w 1 X Φ 1 Z/ n w 1 X Φ 1 X/n 0..... w q X Φ q Z/ n 0 w q X Φ q X/n.

B3. max 1 i n x ij < M. 1 j m 12

Another Theorem 13 Theorem 2. Under conditions B1-3, provided that λ m / m λ 0, and n a /m 0 for some a > 0, the first component ˆδ 1 minimizing V mn has the same limiting distribution as the as the first component of the minimizer of, V 0 (δ) = δ B + 1 2 δ D 1 δ + λ 0 δ s where B is a zero mean Gaussian vector with covariance matrix D 0, and s = (s 0 0 pq) and s 0 = (sgn(α i )).

Monte-Carlo Models 14 We consider two very simple models: and where y ij = α i + x ij β + u ij (1) y ij = α i + x ij β + (1 + x ij γ)u ij. (2) x ij = δ i + v ij (3) with γ i and v ij independent and identically distributed over i and i, j respectively, then the interclass correlation coefficient, ρ x = σ 2 δ/(σ 2 δ + σ 2 u),

Monte Carlo: Location Shift Model 15 LS PLS LSFE QR PQR QRFE N Bias 0.0031 0.0048 0.0056 0.0048 0.0067 0.0047 RMSE 0.0847 0.0604 0.0668 0.0977 0.0781 0.0815 t 3 Bias -0.0062-0.0054-0.0051-0.0063-0.0101-0.0082 RMSE 0.1377 0.1031 0.1143 0.1274 0.0881 0.0921 χ 2 3 Bias -0.0068 0.0002 0.0032-0.0052 0.0063 0.0072 RMSE 0.2155 0.1503 0.1650 0.2362 0.1506 0.1513

Monte Carlo: Location-Scale Shift Model 16 LS PLS LSFE QR PQR QRFE N Bias 0.0000 0.0010 0.0012-0.0020-0.0021-0.0022 RMSE 0.0559 0.0501 0.0542 0.0638 0.0526 0.0556 t 3 Bias -0.0045 0.0000 0.0008-0.0044-0.0015 0.0021 RMSE 0.0948 0.0806 0.0870 0.0758 0.0620 0.0693 χ 2 3 Bias 0.0617 0.0609 0.0608 0.0317-0.0055-0.0128 RMSE 0.1608 0.1292 0.1368 0.1627 0.1042 0.1092

Conclusions 17 Classical random effects in Gaussian models is just penalized least squares estimation in which fixed effects are shrunken toward their prior mean of zero. Other penalties may be useful in other settings.

Conclusions 17 Classical random effects in Gaussian models is just penalized least squares estimation in which fixed effects are shrunken toward their prior mean of zero. Other penalties may be useful in other settings. L 1 penalty methods have attractive properties in a variety of model selection, image analysis and function estimation problems. Interior point algorithms exploiting inherently sparse linear algebra facilate efficient computations.

Conclusions 17 Classical random effects in Gaussian models is just penalized least squares estimation in which fixed effects are shrunken toward their prior mean of zero. Other penalties may be useful in other settings. L 1 penalty methods have attractive properties in a variety of model selection, image analysis and function estimation problems. Interior point algorithms exploiting inherently sparse linear algebra facilate efficient computations. There is more to longitudinal data analysis than the estimation of location shift effects. Many interesting questions can only be addressed with broader models that incorporate distributional effects for treatments and other covariates.

Quantile Regression Methods for Reference Growth Charts 18 Roger Koenker University of Illinois at Urbana-Champaign Based on joint work with: Ying Wei, Columbia, Biostatistics Anneli Pere, Children s Hospital, Helsinki Xuming He, UIUC.

Quetelet s (1871) Growth Chart 19

Penalized Maximum Likelihood Estimation 20 References: Cole (1988), Cole and Green (1992), and Carey(2002) Data: {Y i (t i,j ) : j = 1,..., J i, i = 1,..., n.} Model: Z(t) = (Y (t)/µ(t))λ(t) 1 λ(t)σ(t) N (0, 1) Estimation: max l(λ, µ, σ) ν λ (λ (t)) 2 dt ν µ (µ (t)) 2 dt ν σ (σ (t)) 2 dt, l(λ, µ, σ) = n [λ(t i ) log(y (t i )/µ(t i )) log σ(t i ) 1 2 Z2 (t i )], i=1

Quantiles as Argmins 21 The τ th quantile of a random variable Y having distribution function F is: α(τ) = argmin ρ τ (y α)df (y) where ρ τ (u) = u (τ I(u < 0)).

Quantiles as Argmins 21 The τ th quantile of a random variable Y having distribution function F is: α(τ) = argmin ρ τ (y α)df (y) where ρ τ (u) = u (τ I(u < 0)). The τth sample quantile is thus: ˆα(τ) = argmin ρ τ (y α)df n (y) n = argmin n 1 ρ τ (y i α) i=1

Quantile Regression 22 The τth conditional quantile function of Y X = x is g(τ x) = argmin g G ρ τ (y g(x))df A natural estimator of g(τ x) is ĝ(τ x) = argmin g G n i=1 ρ τ (y i g(x i )) with G chosen as a finite dimensional linear space, g(x) = p ϕ j (x)β j. j=1

Choice of Basis 23 There are many possible choices for the basis expansion {ϕ j }. We opt for the (very conventional) cubic B-spline functions: 0 5 10 15 20 Age In R these quantile regression models can be estimated with the command. fit <- rq(y bs(x,knots=knots),tau = 1:9/10) Similar functionality in SAS is coming real soon now.

Data 24 Longitudinal measurements on height for 2514 Finnish children,

Data 24 Longitudinal measurements on height for 2514 Finnish children, 1143 boys, 1162 girls all healthy, full-term, singleton births,

Data 24 Longitudinal measurements on height for 2514 Finnish children, 1143 boys, 1162 girls all healthy, full-term, singleton births, About 20 measurements per child,

Data 24 Longitudinal measurements on height for 2514 Finnish children, 1143 boys, 1162 girls all healthy, full-term, singleton births, About 20 measurements per child, Two cohorts: 1096 born between 1959-61, 1209 born between 1968-72

Data 24 Longitudinal measurements on height for 2514 Finnish children, 1143 boys, 1162 girls all healthy, full-term, singleton births, About 20 measurements per child, Two cohorts: 1096 born between 1959-61, 1209 born between 1968-72 Sample constitutes 0.5 percent of Finns born in these periods.

25 100 90 Unconditional Reference Quantiles Boys 0 2.5 Years 0.97 0.9 0.75 0.5 0.25 0.1 0.03 Box Cox Parameter Functions λ(t) 2 1.5 1 0.5 Height (cm) 80 70 µ(t) 90 80 70 60 60 50 LMS edf = (7,10,7) LMS edf = (22,25,22) QR edf = 16 σ(t) 0.046 0.044 0.042 0.04 0.038 0.036 0.034 0 0.5 1 1.5 2 2.5 Age (years) 0 0.5 1 1.5 2 2.5 Age (years)

26

27 180 160 Unconditional Reference Quantiles Boys 2 18 Years 0.97 0.9 0.75 0.5 0.25 0.1 0.03 Box Cox Parameter Functions λ(t) 3 2.5 2 1.5 1 0.5 0 µ(t) Height (cm) 140 160 140 120 120 LMS edf = (7,10,7) σ(t) 100 0.055 100 LMS edf = (22,25,22) QR edf = 16 0.05 0.045 0.04 80 5 10 15 Age (years) 5 10 15 Age (years)

28

29 100 90 Unconditional Reference Quantiles Girls 0 2.5 Years 0.97 0.9 0.75 0.5 0.25 0.1 0.03 Box Cox Parameter Functions λ(t) 1.5 1 0.5 0 80 µ(t) 90 Height (cm) 70 80 70 60 60 LMS edf = (7,10,7) σ(t) 0.044 LMS edf = (22,25,22) 0.042 0.04 50 QR edf = 16 0.038 0.036 0.034 0 0.5 1 1.5 2 2.5 Age (years) 0 0.5 1 1.5 2 2.5 Age (years)

30

31 Unconditional Reference Quantiles Girls 2 18 Years 0.97 0.9 0.75 Box Cox Parameter Functions λ(t) 3 160 0.5 0.25 0.1 0.03 2 1 0 Height (cm) 140 120 µ(t) 160 140 120 100 LMS edf = (7,10,7) σ(t) 100 80 LMS edf = (22,25,22) QR edf = 16 0.046 0.044 0.042 0.04 0.038 0.036 5 10 15 Age (years) 2 4 6 8 10 12 14 16 Age (years)

32

33 Growth Velocity Curves Boys 0 2.5 years Girls 0 2.5 years τ Boys 2 18 years Girls 2 18 years 40 30 20 LMS (7,10,7) LMS (22,25,22) QR 0.1 8 6 4 2 10 0 40 8 30 20 0.25 6 4 2 Velocity (cm/year) 10 40 30 20 0.5 0 8 6 4 2 10 0 40 8 30 20 0.75 6 4 2 10 0 40 8 30 20 0.9 6 4 2 10 0 0.5 1 1.5 2 0.5 1 1.5 2 5 10 15 Age (years) 5 10 15

34

35 Estimated Age Specific Density Functions Age = 0.5 Age = 1 Age = 1.5 N= 235 N= 204 N= 67 0.15 Boys 0.10 0.05 Data QR LMS N= 215 N= 191 N= 68 0.00 0.15 0.10 Girls 0.05 0.00 65 70 75 70 75 80 85 75 80 85 90 Height (cm)

36

37 Height Density at Age 1 0.15 Boys 0.10 0.05 Pooled Born in 1960 Born in 1970 0.00 0.20 Girls 0.15 0.10 0.05 0.00 70 72 74 76 78 80 82 84 Height(cm)

Conditioning on Prior Growth 38 It is often important to condition not only on age, but also on prior growth and possibly on other covariates. Autoregressive models are natural, but complicated due to the irregular spacing of typical longitudinal measurements. Data: {Y i (t i,j ) : j = 1,..., J i, i = 1,..., n.} Model: Q Yi (t i,j )(τ t i,j, Y i (t i,j 1 ), x i ) = g τ (t i,j ) + [α(τ) + β(τ)(t i,j t i,j 1 )]Y i (t i,j 1 ) + x i γ(τ).

AR Components of the Infant Conditional Growth Model 39 τ Boys Girls ˆα(τ) ˆβ(τ) ˆγ(τ) ˆα(τ) ˆβ(τ) ˆγ(τ) 0.03 0.845 0.147 0.024 0.809 0.135 (0.020) (0.011) (0.011) (0.024) (0.011) 0.1 0.787 (0.020) 0.25 0.725 (0.019) 0.5 0.635 (0.025) 0.75 0.483 (0.029) 0.9 0.422 (0.024) 0.97 0.383 (0.024) 0.159 (0.007) 0.170 (0.006) 0.173 (0.009) 0.187 (0.009) 0.213 (0.016) 0.214 (0.016) 0.036 (0.007) 0.051 (0.009) 0.060 (0.013) 0.063 (0.017) 0.070 (0.017) 0.077 (0.018) 0.757 (0.022) 0.685 (0.021) 0.612 (0.027) 0.457 (0.027) 0.411 (0.030) 0.400 (0.038) 0.153 (0.007) 0.163 (0.006) 0.175 (0.008) 0.183 (0.012) 0.201 (0.015) 0.232 (0.024) 0.042 (0.010) 0.054 (0.009) 0.061 (0.008) 0.070 (0.009) 0.094 (0.015) 0.100 (0.018) 0.086 (0.027)

AR Components of the Childrens Conditional Growth Model 40 τ Boys Girls ˆα(τ) ˆβ(τ) ˆγ(τ) ˆα(τ) ˆβ(τ) ˆγ(τ) 0.03 0.976 0.036 0.011 0.993 0.033 (0.010) (0.002) (0.013) (0.012) (0.002) 0.1 0.980 (0.005) 0.25 0.978 (0.006) 0.5 0.984 (0.004) 0.75 0.990 (0.004) 0.9 0.987 (0.009) 0.97 0.980 (0.014) 0.039 (0.001) 0.042 (0.001) 0.045 (0.001) 0.047 (0.001) 0.049 (0.001) 0.050 (0.002) 0.022 (0.007) 0.021 (0.006) 0.019 (0.004) 0.014 (0.006) 0.012 (0.009) 0.023 (0.015) 0.989 (0.006) 0.986 (0.005) 0.984 (0.007) 0.985 (0.007) 0.984 (0.008) 0.982 (0.013) 0.039 (0.001) 0.042 (0.001) 0.045 (0.001) 0.050 (0.001) 0.052 (0.001) 0.053 (0.001) 0.006 (0.015) 0.008 (0.007) 0.019 (0.006) 0.022 (0.006) 0.016 (0.006) 0.002 (0.012) 0.021 (0.018)

Transformation to Normality 41 A presumed advantage of univariate (age-specific) transformation to normality is that once observations are transformed to univariate Z-scores they are automatically prepared to longitudinal autoregression: Z t = α 0 + α 1 Z t 1 + U t Premise: Marginal Normality Joint Normality

Transformation to Normality 41 A presumed advantage of univariate (age-specific) transformation to normality is that once observations are transformed to univariate Z-scores they are automatically prepared to longitudinal autoregression: Z t = α 0 + α 1 Z t 1 + U t Premise: Marginal Normality Joint Normality Of course we know it isn t true, but we also think we know that counterexamples are pathological, and don t occur in nature.

42 3 2 1 0 1 2 3 75 80 85 90 Boys age 6.5 Theoretical Quantiles Sample Quantiles 3 2 1 0 1 2 3 75 80 85 90 Boys age 6 Theoretical Quantiles Sample Quantiles 3 2 1 0 1 2 3 1.5 0.5 0.5 1.5 Boys AR(1) residuals Theoretical Quantiles Sample Quantiles 3 2 1 0 1 2 3 80 85 90 95 100 Girls age 6.5 Theoretical Quantiles Sample Quantiles 3 2 1 0 1 2 3 80 85 90 95 100 Girls age 6 Theoretical Quantiles Sample Quantiles 3 2 1 0 1 2 3 2.0 1.0 0.0 1.0 Girls AR(1) residuals Theoretical Quantiles Sample Quantiles

43

44 Height (cm) 110 120 130 140 150 Subject 89 Subject 225 0.97 0.9 0.75 0.5 0.25 0.1 0.03 6 7 8 9 10 Age (years)

45 1 Subject 89 Subject 225 1 Quantile 0.8 0.6 0.4 Unconditional Conditional Observed 0.8 0.6 0.4 0.2 0.2 0 120 125 130 135 140 145 Height (cm) 120 125 130 135 140 Height (cm) 0

Conclusions 46 Nonparametric quantile regression using B-splines offers a reasonable alternative to parametric methods for constructing reference growth charts.

Conclusions 46 Nonparametric quantile regression using B-splines offers a reasonable alternative to parametric methods for constructing reference growth charts. The flexibility of quantile regression methods exposes features of the data that are not easily observable with conventional parametric methods. Even for height data.

Conclusions 46 Nonparametric quantile regression using B-splines offers a reasonable alternative to parametric methods for constructing reference growth charts. The flexibility of quantile regression methods exposes features of the data that are not easily observable with conventional parametric methods. Even for height data. Longitudinal data can be easily accomodated into the quantile regression framework by adding covariates, including the use of autoregressive effects for unequally spaced measurements.