Supplemental materials to Reduced Rank Mixed Effects Models for Spatially Correlated Hierarchical Functional Data
|
|
- Darlene Simon
- 5 years ago
- Views:
Transcription
1 Supplemental materials to Reduced Rank Mixed Effects Models for Spatially Correlated Hierarchical Functional Data Lan Zhou, Jianhua Z. Huang, Josue G. Martinez, Arnab Maity, Veerabhadran Baladandayuthapani and Raymond J. Carroll 1 Outline Section details the computational methods for implementing the proposed methodology, including steps of the EM algorithm and techniques for avoiding storage and inverting large matrices. Section 3 provides additional simulation results. Section 4 contains residuals plots after model fitting for the real data example. Computational methods.1 Computation of the Conditional Moments in the E-step of the EM Algorithm In the following, the conditional expectations are calculated given the current parameter values; for simplicity of presentation, the dependence on the current parameter values is suppressed in our notation. To calculate the conditional moments, we use some standard results for multivariate normal distributions. The covariances between α ab, β ab and Y ab are cov(α ab, Y ab ) = D α,a Γ T ξ B T ab and cov(β ab, Y ab ) = V ab Γ T η,abb T ab. Since cov(α ab, β ab ) = 0, we have cov(y ab ) = B ab Γ ξ D α,a Γ T ξ B T ab + B abγ η,ab V ab Γ T η,abb T ab + σ I ab, where I ab is the identity matrix of rank N ab. The conditional distribution of (α ab, β ab ) given Y ab is normal and is denoted as ( α ab β ab The conditional means are ) ( ) ( ) } m α,ab Σ αα,ab, Σ αβ,ab N, Σ ab =. (1) m β,ab Σ T αβ,ab, Σ ββ,ab m α,ab = E(α ab Y ab ) = D α,a Γ T ξ B T abcov(y ab ) 1 (Y ab B ab γ µ,a ) () and m β,ab = E(β ab Y ab ) = V ab Γ T η,abb T abcov(y ab ) 1 (Y ab B ab γ µ,a ). (3) 1
2 The conditional covariance matrix is ( ) } ( ) ( ) α ab cov β Y D α,a 0 D α,a Γ T ξ B T ab ab = cov(y ab 0 V ab V ab Γ T η,abb T ab )} 1 (B ab Γ ξ D α,a, B ab Γ η,ab V ab ). ab Therefore, we have Σ αα,ab = D α,a D α,a Γ T ξ B T abcov(y ab )} 1 B ab Γ ξ D α,a, (4) Σ ββ,ab = V ab V ab Γ T η,abb T abcov(y ab )} 1 B ab Γ η,ab V ab, (5) and Σ αβ,ab = D α,a Γ T ξ B T abcov(y ab )} 1 B ab Γ η,ab V ab. (6) The desired predictions required by the EM algorithm are α ab = E(α ab Y ab ) = m α,ab, βab = E(β ab Y ab ) = m β,ab, α ab α T ab = E(α abα T ab Y ab ) = α ab α T ab + Σ αα,ab β ab β T ab = E(β ab β T ab Y ab ) = β ab βt ab + Σ ββ,ab, α ab β T ab = E(α ab β T ab Y ab ) = α ab βt ab + Σ αβ,ab. (7) The first two equalities also give expressions of the best linear unbiased predictors (BLUP) of the random effects α ab and β ab. Computation of V ab. It is convenient to re-group the elements of β ab according to the principal components. Denote β ab,j = (β ab1j,..., β abcab j) T, j = 1,..., K η, and β ab = (β T ab,1,..., β T ab,k η ) T. Let P abj be a C ab by C ab matrix with elements P abj,cc = ρ( x abc x abc ; θ aj ), c, c = 1,..., C ab. Note that P abj = P T abj. Then Σ abj = cov(β ab,j ) = σβ,aj P abj. Since cov(β ab,j, β ab,j ) = 0 for j j, the covariance matrix of β ab is block diagonal Ṽ ab = cov( β ab ) = diag(σ ab1,... Σ abkη ). Note that β ab is just a reordering of β ab, that is, β ab = O β ab for a permutation matrix O. It follows the covariance matrix of β ab is V ab = OṼabO T. Circumventing inversion of cov(y ab ). We suppress the subscripts of D α,a and V ab in the following discussion. Denote E = B ab Γ ξ, F = B ab Γ η,ab, and S = cov(y ab ) = EDE T + FVF T + σ I. To calculate the conditional moments given in () (6), we need to compute DE T S 1 in (), VF T S 1 in (3), D DE T S 1 ED in (4), V VF T S 1 F T V in (6), and DE T S 1 FV in (5). Denote J = (EDE T + σ I) 1 and K = (FVF T + σ I) 1. Repeatedly using the identities (A 1 + C T B 1 C) 1 = A AC T (CAC T + B) 1 CA, (8) (A 1 + C T B 1 C) 1 C T B 1 = AC T (CAC T + B) 1, (9)
3 we obtain J = σ I σ E(E T E + σ D 1 ) 1 E T, (10) K = σ I σ F(F T F + σ V 1 ) 1 F T, (11) and DE T S 1 = (D 1 + E T KE) 1 E T K, VF T S 1 = (V 1 + F T JF) 1 F T J, D DE T S 1 ED = (D 1 + E T KE) 1, V VF T S 1 FV = (V 1 + F T JF) 1, DE T S 1 FV = DE T JF(V 1 + F T JF) 1. The matrices that need to be inverted here are of the same size as D or V, which is K ξ K ξ or (C ab K η ) (C ab K η ), much smaller than the size of cov(y ab ).. Updating Formula in the M-step of the EM Algorithm In the updating formulas given below, the parameters appear on the right hand side of equations are all fixed at their values obtained from the previous iteration of the algorithm. 1. Update the estimate of σ. Let N = A Ba a=1 N ab. The updating formula is σ = 1 A E(ɛ T N abɛ ab Y ab ), a=1 where ɛ ab = Y ab B ab γ µ,a B ab Γ ξ α ab B ab Γ η,ab β ab. Using (7), we obtain that E(ɛ T ab ɛ ab Y ab ) can be expressed as (Y ab B ab γ µ,a B ab Γ ξ α ab B ab Γ η,ab βab ) T (Y ab B ab γ µ,a B ab Γ ξ α ab B ab Γ η,ab βab ) D α,a is + trace(b ab Γ ξ Σ αα,ab Γ T ξ B T ab + B ab Γ η,ab Σ ββ,ab Γ T η,abb T ab + B ab Γ ξ Σ αβ,ab Γ T η,abb T ab).. Update D α,a = diag(σα,aj) and D β,a = diag(σβ,aj ). The updating formula for 1 D α,a = diag B a } α ab α T ab. (1) To update D β,a, we need to minimize the conditional expectation of K η j=1 log( V ab )+β T abv 1 ab β ab = K η j=1 C ab log(σ β,aj)+log P abj +σ β,aj βt ab,jp 1 ab,j β ab,j }. 3
4 Thus the updating formulas are ( σ β,aj 1 = trace Ba C ab P 1 ab,j β ab,j β T ab,j ), j = 1,..., K η. (13) If the variances of random effects do not depend on the treatment level a, we simply need to average over a in (1) and (13). 3. Update γ µ,a, a = 1,..., A. The updating formula are Ba } 1 Ba γ µ,a = B T abb ab + σ λ µ b (t)b (t) T B T ab(y ab B ab Γ ξ α ab B ab Γ η,ab βab ). 4. Update the columns of Γ ξ sequentially. For j = 1,..., K ξ, the updating formula are A γ ξ,j = α abj BT abb ab + σ λ ξ a=1 A a=1 B T ab b (t)b (t) T } 1 ( (Y ab B ab γ µ,a ) α abj B ab l j ) γ ξ,l α abj α abl B ab Γ η,ab α abj β ab }. 5. Update the columns of Γ η sequentially. For j = 1,..., K η, the updating formula are A C ab γ η,j = β abcj BT abcb abc + σ λ η,a a=1 c=1 A C ab a=1 c=1 B T abc b (t)b (t) T } 1 (Y abc B abc γ µ,a ) β abcj B abc Γ ξ β abcj α ab B abc ( l j γ η,l β abcj β abcl )}. 6. Orthogonalizing Γ ξ and Γ η. The matrix Γ ξ and Γ η obtained in Steps 4 and 5 need not have orthonormal columns. We orthogonalize them in this step. Compute the QR decomposition Γ ξ = Q ξ R ξ where Q ξ has orthonormal columns. Let Γ ξ Q ξ. Orthogonalize Γ η similarly. 7. Update the correlation parameter θ aj. Given the current estimates of other parameters, we minimize for each a and j the conditional expectation of which is log P 1 abj + 1 } β T σ ab,jp 1 β,aj abj β ab,j, log P 1 abj + 1 trace σβ,aj 4 ( P 1 abj β ab,j β T ab,j) }. (14)
5 The minimizer can be found by a gradient-based method. Denote the components of θ aj as θ ajk s. The gradient with respect to θ aj is a vector with elements ( trace P 1 abj P ) abj 1 ( trace P 1 P abj θ ajk σβ,aj abj P 1 abjβ θ ab,j βab,j) } T. ajk Alternatively, a direct search algorithm that does not require the calculation of the gradient, such as the downhill simplex method of Nelder and Mead (1965), can be applied to minimize (14). If the correlation parameters do not depend on j, we consider for each a a new objective function by summing (14) over j. If the correlation parameters do not depend on a, we need only sum (14) over a for each j to define the new objective function. Note that the orthogonalization in Step 6 is used to ensure identifiability. This approach was adapted from Zhou et al. (008) and worked well in our simulation study and real data analysis. An alternative approach is direct optimization within a restricted space (Peng and Paul, 009) but its implementation is beyond the scope of this paper..3 Computing the Observed Data Log Likelihood When we do crossvalidation, we need to compute the log likelihood of observed data, which depends on the determinant and inverse of the possibly very large matrix S, defined in Section.1. This section gives a method for computation of S and S 1 without constructing the matrix S. Using the identity A + BC T = A I + C T A 1 B, (15) we obtain S = σ I + EDE T + FVF T = σ I + EDE T I + V 1/ F T (σ I + EDE T ) 1 FV 1/. The two factors on the right hand side of the above equation can be computed as follows. Using the identity (15), we have σ I + EDE T = σ N I + σ D 1/ E T ED 1/ = σ N D D 1 + σ E T E. Using (σ I + EDE T ) 1 = σ I σ E(E T E + σ D 1 ) 1 E T, we obtain I + V 1/ F T (σ I + EDE T ) 1 FV 1/ = V V 1 + σ F T F σ F T E(E T E + σ D 1 ) 1 E T F. 5
6 Using the definition of J in Section.1 and the identity (8), we have S 1 = (J 1 + FVF T ) 1 = J JF(F T JF + V 1 ) 1 F T J, where J can be computed using (10)..4 Calculating the partial derivatives of P abj Consider the Matérn family of autocorrelation functions ( ) ρ(d; φ, ν) = 1 ν dν 1/ ν ( ) dν 1/ K ν, Γ(ν) φ φ φ > 0, ν > 0, where K ν ( ) is the modified Bessel function of order ν taking the form with K ν (u) = I ν (u) = m=0 π sin(νπ) I ν(u) I ν (u)} ( ) m+ν 1 u. m!γ(m + ν + 1) Denote u = xν 1/ /φ, then the partial derivatives of ρ(x; φ, ν) are ρ(x; φ, ν) φ = φγ(ν)( 1 u ) ν[ukν 1 (u) + K ν+1 (u)} νk ν (u)] and ρ(x; φ, ν) ν = ( u ) ψ(ν) π cot(νπ) + log + 1 } ρ(x; φ, ν) π ) [ ν u + ν I ν+1(u) I ν+1 (u)} u Γ(ν) sin(νπ)( ( u ) log m=0 m=0 } I ν (u) + I ν (u)} 1 u ) m+νψ(m + ν + 1) m!γ(m + ν + 1)( 1 u ) ] m νψ(m ν + 1), m!γ(m ν + 1)( where ψ(z) = C l=0 ( 1 1 ) with C = lim z+l 1+l k k 1 l=1 ln(k)}.557, and k Γ( ) is the Gamma function. 6
7 3 Additional simulation results Two other setups from the Bayesian hierarchical model in Baladandayuthapani, et al. (008) were also considered but the software of the Bayesian method encountered some serious numerical problems and could only run on a small proportion of simulated data sets. We ran simulation 100 times for each setup. The results of our methods are based on three unit and subunit level principal components. Our methods successfully ran on all data sets. and Setup 6. Same as Setup 5 except that Σ 1 = Σ = Σ 31 = Σ 3 = For this setup, the Bayesian method only ran on 10 out of 100 simulated data sets. Setup 7. Same as Setup 6 except that σ = σ 3 = 0.4. For this setup, the Bayesian method only ran on 3 out of 100 simulated data sets. We used the measures defined at the beginning of Section 5 of the paper to assess/compare performance of the two methods. For our reduced rank methods, the measures were computed using all 100 data sets as well as only those data sets that the Bayesian method ran. The results are summarized in Table 1. Our reduced rank method is comparable to the Bayesian method for estimating the mean functions and predicting the random effects. Next we report some simulation results on parameter estimation for setups following our model. Note that comparison with the Bayesian method is not available since two methods use different parameters. Table shows the summary statistics of correlation parameter estimates for the first three simulation setups where the data were generated from our model. The parameter estimates are reasonably unbiased. Figure 1 presents the pointwise sample mean of the estimated treatment effects, unit level and sub-unit level PCs, along with 90% pointwise coverage intervals for simulation Setup. It is not surprising that sub-unit level PCs are better estimated than the unit level PCs since there are more sub-unit level data than the unit level data.. 7
8 Table 1: Comparison of two methods based on 100 simulation runs. Mean (SD) of the integrated absolute errors of estimating the mean functions and predicting the unit and subunit level random effects. Numbers shown are the actual numbers multiplied by 10. Reduced rank refers to our method; Bayesian refers to the Bayesian method of Baladandayuthapani, et al. (008). 100 sims, 10 sims and 3 sims indicate the number of data sets used in the calculation. Setup Method Mean Unit Subunit 6 7 Reduced rank (100 sims) (1.45) (1.088) (0.481) Reduced rank (10 sims) (0.963) (1.044) (0.305) Bayesian (10 sims) (0.963) 6.55 (0.899) (0.93) Reduced rank (100 sims) (1.515) (1.187) (1.064) Reduced rank (3 sims) (1.34) 7.86 (0.890) (0.845) Bayesian (3 sims) (1.344) 7.84 (0.99) (0.567) Table : Sample mean (SD) of correlation parameter estimates, based on 100 simulation runs. PC 1 PC Setup φ = 8 ν = 0.1 φ = 4 ν = (4.98) (0.00) (5.67) (0.018) 4.6 (1.306) (0.040) (.369) (0.019) (1.19) (0.036) 4 Colon Carcinogenesis Data Figures and 3 contains respectively residual plots after fitting our reduced rank model and the Bayesian model of Baladandayuthapani, et al. (008). It is clear that our model fits the data better. After fitting the Bayesian model, there were still obvious patterns left in the residual plots for treatment groups FO+B and FO B. References Baladandayuthapani, V., Mallick, B., Turner, N., Hong, M., Chapkin, R., Lupton, J. and Carroll, R. J. (008). Bayesian hierarchical spatially correlated functional data analysis with application to colon carcinogenesis. Biometrics 64, Nelder, J.A. & Mead, R. (1965). Computer Journal, 7, A simplex method for function minimization. Peng, J. and Paul, D. (009). A geometric approach to maximum likelihood estimation of the functional principal components from sparse longitudinal data. Journal of Computational and Graphical Statistics, to appear. 8
9 Figure 1: Simulation Setup. The pointwise sample mean of the estimated treatment effects, unit level and sub-unit level PCs, along with 90% pointwise coverage intervals. Group 1 Unit PC 1 Sub unit PC Group Unit PC Sub unit PC t Zhou, L., Huang, J.Z. and Carroll, J.R. (008). Joint modelling of paired sparse functional data using principal components. Biometrika, 95,
10 Figure : Residual plots after fitting the reduced rank model. Separate plots are drawn for the four diet groups. CO is Corn Oil, FO is Fish Oil, +B and B represent with or without (±) Butyrate supplement. Diet CO+B Diet CO B Residual Diet FO+B Diet FO B Fitted p7 10
11 Figure 3: Residual plots after fitting the Bayesian model of Baladandayuthapani, et al. (008). Separate plots are drawn for the four diet groups. CO is Corn Oil, FO is Fish Oil, +B and B represent with or without Butyrate supplement. Diet CO+B Diet CO B Residual Diet FO+B Diet FO B 3 Fitted p7 11
Fast Methods for Spatially Correlated Multilevel Functional Data
Fast Methods for Spatially Correlated Multilevel Functional Data Ana-Maria Staicu, Department of Statistics, North Carolina State University, 23 Stinson Drive Raleigh, NC 27695-8203, USA email: staicu@stat.ncsu.edu,
More informationFunctional Latent Feature Models. With Single-Index Interaction
Generalized With Single-Index Interaction Department of Statistics Center for Statistical Bioinformatics Institute for Applied Mathematics and Computational Science Texas A&M University Naisyin Wang and
More informationWavelet-Based Nonparametric Modeling of Hierarchical Functions in Colon Carcinogenesis
Wavelet-Based Nonparametric Modeling of Hierarchical Functions in Colon Carcinogenesis Jeffrey S. Morris University of Texas, MD Anderson Cancer Center Joint wor with Marina Vannucci, Philip J. Brown,
More informationBayesian Hierarchical Spatially Correlated Functional Data Analysis with Application to Colon Carcinogenesis. Abstract
Bayesian Hierarchical Spatially Correlated Functional Data Analysis with Application to Colon Carcinogenesis Veerabhadran Baladandayuthapani 1, Bani K. Mallick 2, Mee Young Hong 3 4, Joanne R. Lupton 4,
More informationUsing Estimating Equations for Spatially Correlated A
Using Estimating Equations for Spatially Correlated Areal Data December 8, 2009 Introduction GEEs Spatial Estimating Equations Implementation Simulation Conclusion Typical Problem Assess the relationship
More informationConcentration Ellipsoids
Concentration Ellipsoids ECE275A Lecture Supplement Fall 2008 Kenneth Kreutz Delgado Electrical and Computer Engineering Jacobs School of Engineering University of California, San Diego VERSION LSECE275CE
More informationKriging models with Gaussian processes - covariance function estimation and impact of spatial sampling
Kriging models with Gaussian processes - covariance function estimation and impact of spatial sampling François Bachoc former PhD advisor: Josselin Garnier former CEA advisor: Jean-Marc Martinez Department
More informationHypothesis testing in multilevel models with block circular covariance structures
1/ 25 Hypothesis testing in multilevel models with block circular covariance structures Yuli Liang 1, Dietrich von Rosen 2,3 and Tatjana von Rosen 1 1 Department of Statistics, Stockholm University 2 Department
More informationPh.D. Qualifying Exam Friday Saturday, January 6 7, 2017
Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a
More informationState-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Financial Econometrics / 49
State-space Model Eduardo Rossi University of Pavia November 2013 Rossi State-space Model Financial Econometrics - 2013 1 / 49 Outline 1 Introduction 2 The Kalman filter 3 Forecast errors 4 State smoothing
More informationarxiv: v1 [math.st] 22 Dec 2018
Optimal Designs for Prediction in Two Treatment Groups Rom Coefficient Regression Models Maryna Prus Otto-von-Guericke University Magdeburg, Institute for Mathematical Stochastics, PF 4, D-396 Magdeburg,
More information26. Filtering. ECE 830, Spring 2014
26. Filtering ECE 830, Spring 2014 1 / 26 Wiener Filtering Wiener filtering is the application of LMMSE estimation to recovery of a signal in additive noise under wide sense sationarity assumptions. Problem
More informationMA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1
MA 575 Linear Models: Cedric E Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1 1 Within-group Correlation Let us recall the simple two-level hierarchical
More informationCourse topics (tentative) The role of random effects
Course topics (tentative) random effects linear mixed models analysis of variance frequentist likelihood-based inference (MLE and REML) prediction Bayesian inference The role of random effects Rasmus Waagepetersen
More information. a m1 a mn. a 1 a 2 a = a n
Biostat 140655, 2008: Matrix Algebra Review 1 Definition: An m n matrix, A m n, is a rectangular array of real numbers with m rows and n columns Element in the i th row and the j th column is denoted by
More informationNext is material on matrix rank. Please see the handout
B90.330 / C.005 NOTES for Wednesday 0.APR.7 Suppose that the model is β + ε, but ε does not have the desired variance matrix. Say that ε is normal, but Var(ε) σ W. The form of W is W w 0 0 0 0 0 0 w 0
More informationCovariance function estimation in Gaussian process regression
Covariance function estimation in Gaussian process regression François Bachoc Department of Statistics and Operations Research, University of Vienna WU Research Seminar - May 2015 François Bachoc Gaussian
More informationfor Sparsely Observed Functional Data University of Glasgow
Spatial Correlation Estimation for Sparsely Observed Functional Data Workshop on High dimensional and dependent functional data, Bristol, Sep 2012 Surajit Ray University of Glasgow Contributors t Chong
More informationState-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Fin. Econometrics / 53
State-space Model Eduardo Rossi University of Pavia November 2014 Rossi State-space Model Fin. Econometrics - 2014 1 / 53 Outline 1 Motivation 2 Introduction 3 The Kalman filter 4 Forecast errors 5 State
More informationGeostatistical Modeling for Large Data Sets: Low-rank methods
Geostatistical Modeling for Large Data Sets: Low-rank methods Whitney Huang, Kelly-Ann Dixon Hamil, and Zizhuang Wu Department of Statistics Purdue University February 22, 2016 Outline Motivation Low-rank
More informationMultivariate Statistical Analysis
Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 4 for Applied Multivariate Analysis Outline 1 Eigen values and eigen vectors Characteristic equation Some properties of eigendecompositions
More informationFast Methods for Spatially Correlated Multilevel Functional Data
Biostatistics (2007), 8, 1, pp. 1 29 doi:10.1093/biostatistics/kxl000 Advance Access publication on May 15, 2006 Fast Methods for Spatially Correlated Multilevel Functional Data ANA-MARIA STAICU Department
More informationReflections and Rotations in R 3
Reflections and Rotations in R 3 P. J. Ryan May 29, 21 Rotations as Compositions of Reflections Recall that the reflection in the hyperplane H through the origin in R n is given by f(x) = x 2 ξ, x ξ (1)
More informationAn Introduction to Bayesian Linear Regression
An Introduction to Bayesian Linear Regression APPM 5720: Bayesian Computation Fall 2018 A SIMPLE LINEAR MODEL Suppose that we observe explanatory variables x 1, x 2,..., x n and dependent variables y 1,
More informationMultivariate Distributions
IEOR E4602: Quantitative Risk Management Spring 2016 c 2016 by Martin Haugh Multivariate Distributions We will study multivariate distributions in these notes, focusing 1 in particular on multivariate
More informationPrediction. is a weighted least squares estimate since it minimizes. Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark
Prediction Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark March 22, 2017 WLS and BLUE (prelude to BLUP) Suppose that Y has mean β and known covariance matrix V (but Y need not
More informationInformation geometry for bivariate distribution control
Information geometry for bivariate distribution control C.T.J.Dodson + Hong Wang Mathematics + Control Systems Centre, University of Manchester Institute of Science and Technology Optimal control of stochastic
More informationWLS and BLUE (prelude to BLUP) Prediction
WLS and BLUE (prelude to BLUP) Prediction Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark April 21, 2018 Suppose that Y has mean X β and known covariance matrix V (but Y need
More informationBare minimum on matrix algebra. Psychology 588: Covariance structure and factor models
Bare minimum on matrix algebra Psychology 588: Covariance structure and factor models Matrix multiplication 2 Consider three notations for linear combinations y11 y1 m x11 x 1p b11 b 1m y y x x b b n1
More informationMultilevel Cross-dependent Binary Longitudinal Data
Multilevel Cross-dependent Binary Longitudinal Data Nicoleta Serban 1 H. Milton Stewart School of Industrial Systems and Engineering Georgia Institute of Technology nserban@isye.gatech.edu Ana-Maria Staicu
More informationParameter Estimation in the Spatio-Temporal Mixed Effects Model Analysis of Massive Spatio-Temporal Data Sets
Parameter Estimation in the Spatio-Temporal Mixed Effects Model Analysis of Massive Spatio-Temporal Data Sets Matthias Katzfuß Advisor: Dr. Noel Cressie Department of Statistics The Ohio State University
More information1 Cricket chirps: an example
Notes for 2016-09-26 1 Cricket chirps: an example Did you know that you can estimate the temperature by listening to the rate of chirps? The data set in Table 1 1. represents measurements of the number
More informationPrincipal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis. Chris Funk. Lecture 17
Principal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis Chris Funk Lecture 17 Outline Filters and Rotations Generating co-varying random fields Translating co-varying fields into
More informationRegularized Discriminant Analysis and Reduced-Rank LDA
Regularized Discriminant Analysis and Reduced-Rank LDA Department of Statistics The Pennsylvania State University Email: jiali@stat.psu.edu Regularized Discriminant Analysis A compromise between LDA and
More informationOn Gaussian Process Models for High-Dimensional Geostatistical Datasets
On Gaussian Process Models for High-Dimensional Geostatistical Datasets Sudipto Banerjee Joint work with Abhirup Datta, Andrew O. Finley and Alan E. Gelfand University of California, Los Angeles, USA May
More informationLecture 4 Orthonormal vectors and QR factorization
Orthonormal vectors and QR factorization 4 1 Lecture 4 Orthonormal vectors and QR factorization EE263 Autumn 2004 orthonormal vectors Gram-Schmidt procedure, QR factorization orthogonal decomposition induced
More informationKneib, Fahrmeir: Supplement to "Structured additive regression for categorical space-time data: A mixed model approach"
Kneib, Fahrmeir: Supplement to "Structured additive regression for categorical space-time data: A mixed model approach" Sonderforschungsbereich 386, Paper 43 (25) Online unter: http://epub.ub.uni-muenchen.de/
More informationBasic Concepts in Matrix Algebra
Basic Concepts in Matrix Algebra An column array of p elements is called a vector of dimension p and is written as x p 1 = x 1 x 2. x p. The transpose of the column vector x p 1 is row vector x = [x 1
More informationChapter 5 Prediction of Random Variables
Chapter 5 Prediction of Random Variables C R Henderson 1984 - Guelph We have discussed estimation of β, regarded as fixed Now we shall consider a rather different problem, prediction of random variables,
More informationAlgebra C Numerical Linear Algebra Sample Exam Problems
Algebra C Numerical Linear Algebra Sample Exam Problems Notation. Denote by V a finite-dimensional Hilbert space with inner product (, ) and corresponding norm. The abbreviation SPD is used for symmetric
More informationGaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012
Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature
More informationLinear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,
Linear Regression In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, y = Xβ + ɛ, where y t = (y 1,..., y n ) is the column vector of target values,
More informationA Cautionary Note on Generalized Linear Models for Covariance of Unbalanced Longitudinal Data
A Cautionary Note on Generalized Linear Models for Covariance of Unbalanced Longitudinal Data Jianhua Z. Huang a, Min Chen b, Mehdi Maadooliat a, Mohsen Pourahmadi a a Department of Statistics, Texas A&M
More informationHierarchical Modeling for Univariate Spatial Data
Hierarchical Modeling for Univariate Spatial Data Geography 890, Hierarchical Bayesian Models for Environmental Spatial Data Analysis February 15, 2011 1 Spatial Domain 2 Geography 890 Spatial Domain This
More informationLecture 4: Types of errors. Bayesian regression models. Logistic regression
Lecture 4: Types of errors. Bayesian regression models. Logistic regression A Bayesian interpretation of regularization Bayesian vs maximum likelihood fitting more generally COMP-652 and ECSE-68, Lecture
More informationA General Framework for Variable Selection in Linear Mixed Models with Applications to Genetic Studies with Structured Populations
A General Framework for Variable Selection in Linear Mixed Models with Applications to Genetic Studies with Structured Populations Joint work with Karim Oualkacha (UQÀM), Yi Yang (McGill), Celia Greenwood
More informationGeneralized Functional Linear Models with Semiparametric Single-Index Interactions
Generalized Functional Linear Models with Semiparametric Single-Index Interactions Yehua Li Department of Statistics, University of Georgia, Athens, GA 30602, yehuali@uga.edu Naisyin Wang Department of
More informationHierarchical Functional Data with Mixed Continuous and Binary Measurements
Biometrics 70, 802 811 December 2014 DOI: 10.1111/biom.12211 Hierarchical Functional Data with Mixed Continuous and Binary Measurements Haocheng Li, 1, * John Staudenmayer, 2 and Raymond J. Carroll 1 1
More informationsparse and low-rank tensor recovery Cubic-Sketching
Sparse and Low-Ran Tensor Recovery via Cubic-Setching Guang Cheng Department of Statistics Purdue University www.science.purdue.edu/bigdata CCAM@Purdue Math Oct. 27, 2017 Joint wor with Botao Hao and Anru
More informationHierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets
Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets Abhirup Datta 1 Sudipto Banerjee 1 Andrew O. Finley 2 Alan E. Gelfand 3 1 University of Minnesota, Minneapolis,
More informationRestricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model
Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model Xiuming Zhang zhangxiuming@u.nus.edu A*STAR-NUS Clinical Imaging Research Center October, 015 Summary This report derives
More informationSTAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method.
STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method. Rebecca Barter May 5, 2015 Linear Regression Review Linear Regression Review
More informationPreface Introduction to Statistics and Data Analysis Overview: Statistical Inference, Samples, Populations, and Experimental Design The Role of
Preface Introduction to Statistics and Data Analysis Overview: Statistical Inference, Samples, Populations, and Experimental Design The Role of Probability Sampling Procedures Collection of Data Measures
More informationLinear Methods for Regression. Lijun Zhang
Linear Methods for Regression Lijun Zhang zlj@nju.edu.cn http://cs.nju.edu.cn/zlj Outline Introduction Linear Regression Models and Least Squares Subset Selection Shrinkage Methods Methods Using Derived
More informationStatistics 910, #5 1. Regression Methods
Statistics 910, #5 1 Overview Regression Methods 1. Idea: effects of dependence 2. Examples of estimation (in R) 3. Review of regression 4. Comparisons and relative efficiencies Idea Decomposition Well-known
More informationNearest Neighbor Gaussian Processes for Large Spatial Data
Nearest Neighbor Gaussian Processes for Large Spatial Data Abhi Datta 1, Sudipto Banerjee 2 and Andrew O. Finley 3 July 31, 2017 1 Department of Biostatistics, Bloomberg School of Public Health, Johns
More informationEM Algorithm II. September 11, 2018
EM Algorithm II September 11, 2018 Review EM 1/27 (Y obs, Y mis ) f (y obs, y mis θ), we observe Y obs but not Y mis Complete-data log likelihood: l C (θ Y obs, Y mis ) = log { f (Y obs, Y mis θ) Observed-data
More informationTesting Some Covariance Structures under a Growth Curve Model in High Dimension
Department of Mathematics Testing Some Covariance Structures under a Growth Curve Model in High Dimension Muni S. Srivastava and Martin Singull LiTH-MAT-R--2015/03--SE Department of Mathematics Linköping
More informationMatrices and systems of linear equations
Matrices and systems of linear equations Samy Tindel Purdue University Differential equations and linear algebra - MA 262 Taken from Differential equations and linear algebra by Goode and Annin Samy T.
More informationEconomics Department LSE. Econometrics: Timeseries EXERCISE 1: SERIAL CORRELATION (ANALYTICAL)
Economics Department LSE EC402 Lent 2015 Danny Quah TW1.10.01A x7535 : Timeseries EXERCISE 1: SERIAL CORRELATION (ANALYTICAL) 1. Suppose ɛ is w.n. (0, σ 2 ), ρ < 1, and W t = ρw t 1 + ɛ t, for t = 1, 2,....
More informationX t = a t + r t, (7.1)
Chapter 7 State Space Models 71 Introduction State Space models, developed over the past 10 20 years, are alternative models for time series They include both the ARIMA models of Chapters 3 6 and the Classical
More informationMixed models in R using the lme4 package Part 4: Theory of linear mixed models
Mixed models in R using the lme4 package Part 4: Theory of linear mixed models Douglas Bates 8 th International Amsterdam Conference on Multilevel Analysis 2011-03-16 Douglas Bates
More informationInverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1
Inverse of a Square Matrix For an N N square matrix A, the inverse of A, 1 A, exists if and only if A is of full rank, i.e., if and only if no column of A is a linear combination 1 of the others. A is
More informationMaximum Likelihood Estimation
Maximum Likelihood Estimation Merlise Clyde STA721 Linear Models Duke University August 31, 2017 Outline Topics Likelihood Function Projections Maximum Likelihood Estimates Readings: Christensen Chapter
More informationPQL Estimation Biases in Generalized Linear Mixed Models
PQL Estimation Biases in Generalized Linear Mixed Models Woncheol Jang Johan Lim March 18, 2006 Abstract The penalized quasi-likelihood (PQL) approach is the most common estimation procedure for the generalized
More informationFunctional SVD for Big Data
Functional SVD for Big Data Pan Chao April 23, 2014 Pan Chao Functional SVD for Big Data April 23, 2014 1 / 24 Outline 1 One-Way Functional SVD a) Interpretation b) Robustness c) CV/GCV 2 Two-Way Problem
More information1 Bayesian Linear Regression (BLR)
Statistical Techniques in Robotics (STR, S15) Lecture#10 (Wednesday, February 11) Lecturer: Byron Boots Gaussian Properties, Bayesian Linear Regression 1 Bayesian Linear Regression (BLR) In linear regression,
More informationFor more information about how to cite these materials visit
Author(s): Kerby Shedden, Ph.D., 2010 License: Unless otherwise noted, this material is made available under the terms of the Creative Commons Attribution Share Alike 3.0 License: http://creativecommons.org/licenses/by-sa/3.0/
More informationPoint-Referenced Data Models
Point-Referenced Data Models Jamie Monogan University of Georgia Spring 2013 Jamie Monogan (UGA) Point-Referenced Data Models Spring 2013 1 / 19 Objectives By the end of these meetings, participants should
More informationLecture 20: Linear model, the LSE, and UMVUE
Lecture 20: Linear model, the LSE, and UMVUE Linear Models One of the most useful statistical models is X i = β τ Z i + ε i, i = 1,...,n, where X i is the ith observation and is often called the ith response;
More informationSTAT 512 sp 2018 Summary Sheet
STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}
More informationCourse Notes: Week 1
Course Notes: Week 1 Math 270C: Applied Numerical Linear Algebra 1 Lecture 1: Introduction (3/28/11) We will focus on iterative methods for solving linear systems of equations (and some discussion of eigenvalues
More informationRandom vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables.
Random vectors Recall that a random vector X = X X 2 is made up of, say, k random variables X k A random vector has a joint distribution, eg a density f(x), that gives probabilities P(X A) = f(x)dx Just
More informationPart 6: Multivariate Normal and Linear Models
Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of
More informationSparse Bayesian Logistic Regression with Hierarchical Prior and Variational Inference
Sparse Bayesian Logistic Regression with Hierarchical Prior and Variational Inference Shunsuke Horii Waseda University s.horii@aoni.waseda.jp Abstract In this paper, we present a hierarchical model which
More informationCS281 Section 4: Factor Analysis and PCA
CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we
More information[y i α βx i ] 2 (2) Q = i=1
Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation
More informationComputationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models
Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models Tihomir Asparouhov 1, Bengt Muthen 2 Muthen & Muthen 1 UCLA 2 Abstract Multilevel analysis often leads to modeling
More informationMultivariate Linear Regression Models
Multivariate Linear Regression Models Regression analysis is used to predict the value of one or more responses from a set of predictors. It can also be used to estimate the linear association between
More informationParameter Estimation
Parameter Estimation Tuesday 9 th May, 07 4:30 Consider a system whose response can be modeled by R = M (Θ) where Θ is a vector of m parameters. We take a series of measurements, D (t) where t represents
More informationOn construction of constrained optimum designs
On construction of constrained optimum designs Institute of Control and Computation Engineering University of Zielona Góra, Poland DEMA2008, Cambridge, 15 August 2008 Numerical algorithms to construct
More information2. Matrix Algebra and Random Vectors
2. Matrix Algebra and Random Vectors 2.1 Introduction Multivariate data can be conveniently display as array of numbers. In general, a rectangular array of numbers with, for instance, n rows and p columns
More informationPh.D. Qualifying Exam Friday Saturday, January 3 4, 2014
Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Put your solution to each problem on a separate sheet of paper. Problem 1. (5166) Assume that two random samples {x i } and {y i } are independently
More informationBasics of Point-Referenced Data Models
Basics of Point-Referenced Data Models Basic tool is a spatial process, {Y (s), s D}, where D R r Chapter 2: Basics of Point-Referenced Data Models p. 1/45 Basics of Point-Referenced Data Models Basic
More informationMatrix Approach to Simple Linear Regression: An Overview
Matrix Approach to Simple Linear Regression: An Overview Aspects of matrices that you should know: Definition of a matrix Addition/subtraction/multiplication of matrices Symmetric/diagonal/identity matrix
More informationFixed Effects, Invariance, and Spatial Variation in Intergenerational Mobility
American Economic Review: Papers & Proceedings 2016, 106(5): 400 404 http://dx.doi.org/10.1257/aer.p20161082 Fixed Effects, Invariance, and Spatial Variation in Intergenerational Mobility By Gary Chamberlain*
More informationBayesian Linear Models
Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Department of Forestry & Department of Geography, Michigan State University, Lansing Michigan, U.S.A. 2 Biostatistics, School of Public
More informatione-companion ONLY AVAILABLE IN ELECTRONIC FORM Electronic Companion Stochastic Kriging for Simulation Metamodeling
OPERATIONS RESEARCH doi 10.187/opre.1090.0754ec e-companion ONLY AVAILABLE IN ELECTRONIC FORM informs 009 INFORMS Electronic Companion Stochastic Kriging for Simulation Metamodeling by Bruce Ankenman,
More informationSlice Sampling with Adaptive Multivariate Steps: The Shrinking-Rank Method
Slice Sampling with Adaptive Multivariate Steps: The Shrinking-Rank Method Madeleine B. Thompson Radford M. Neal Abstract The shrinking rank method is a variation of slice sampling that is efficient at
More informationGARCH Models Estimation and Inference. Eduardo Rossi University of Pavia
GARCH Models Estimation and Inference Eduardo Rossi University of Pavia Likelihood function The procedure most often used in estimating θ 0 in ARCH models involves the maximization of a likelihood function
More informationLecture 11: Regression Methods I (Linear Regression)
Lecture 11: Regression Methods I (Linear Regression) Fall, 2017 1 / 40 Outline Linear Model Introduction 1 Regression: Supervised Learning with Continuous Responses 2 Linear Models and Multiple Linear
More informationMIXED MODELS THE GENERAL MIXED MODEL
MIXED MODELS This chapter introduces best linear unbiased prediction (BLUP), a general method for predicting random effects, while Chapter 27 is concerned with the estimation of variances by restricted
More informationFE670 Algorithmic Trading Strategies. Stevens Institute of Technology
FE670 Algorithmic Trading Strategies Lecture 3. Factor Models and Their Estimation Steve Yang Stevens Institute of Technology 09/12/2012 Outline 1 The Notion of Factors 2 Factor Analysis via Maximum Likelihood
More informationVector autoregressions, VAR
1 / 45 Vector autoregressions, VAR Chapter 2 Financial Econometrics Michael Hauser WS17/18 2 / 45 Content Cross-correlations VAR model in standard/reduced form Properties of VAR(1), VAR(p) Structural VAR,
More informationSupplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements
Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements Jeffrey N. Rouder Francis Tuerlinckx Paul L. Speckman Jun Lu & Pablo Gomez May 4 008 1 The Weibull regression model
More informationSTAT 100C: Linear models
STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix
More informationVariational Principal Components
Variational Principal Components Christopher M. Bishop Microsoft Research 7 J. J. Thomson Avenue, Cambridge, CB3 0FB, U.K. cmbishop@microsoft.com http://research.microsoft.com/ cmbishop In Proceedings
More informationModeling Longitudinal Count Data with Excess Zeros and Time-Dependent Covariates: Application to Drug Use
Modeling Longitudinal Count Data with Excess Zeros and : Application to Drug Use University of Northern Colorado November 17, 2014 Presentation Outline I and Data Issues II Correlated Count Regression
More informationTime-Varying Parameters
Kalman Filter and state-space models: time-varying parameter models; models with unobservable variables; basic tool: Kalman filter; implementation is task-specific. y t = x t β t + e t (1) β t = µ + Fβ
More informationOn Reparametrization and the Gibbs Sampler
On Reparametrization and the Gibbs Sampler Jorge Carlos Román Department of Mathematics Vanderbilt University James P. Hobert Department of Statistics University of Florida March 2014 Brett Presnell Department
More information