Supplemental materials to Reduced Rank Mixed Effects Models for Spatially Correlated Hierarchical Functional Data

Size: px
Start display at page:

Download "Supplemental materials to Reduced Rank Mixed Effects Models for Spatially Correlated Hierarchical Functional Data"

Transcription

1 Supplemental materials to Reduced Rank Mixed Effects Models for Spatially Correlated Hierarchical Functional Data Lan Zhou, Jianhua Z. Huang, Josue G. Martinez, Arnab Maity, Veerabhadran Baladandayuthapani and Raymond J. Carroll 1 Outline Section details the computational methods for implementing the proposed methodology, including steps of the EM algorithm and techniques for avoiding storage and inverting large matrices. Section 3 provides additional simulation results. Section 4 contains residuals plots after model fitting for the real data example. Computational methods.1 Computation of the Conditional Moments in the E-step of the EM Algorithm In the following, the conditional expectations are calculated given the current parameter values; for simplicity of presentation, the dependence on the current parameter values is suppressed in our notation. To calculate the conditional moments, we use some standard results for multivariate normal distributions. The covariances between α ab, β ab and Y ab are cov(α ab, Y ab ) = D α,a Γ T ξ B T ab and cov(β ab, Y ab ) = V ab Γ T η,abb T ab. Since cov(α ab, β ab ) = 0, we have cov(y ab ) = B ab Γ ξ D α,a Γ T ξ B T ab + B abγ η,ab V ab Γ T η,abb T ab + σ I ab, where I ab is the identity matrix of rank N ab. The conditional distribution of (α ab, β ab ) given Y ab is normal and is denoted as ( α ab β ab The conditional means are ) ( ) ( ) } m α,ab Σ αα,ab, Σ αβ,ab N, Σ ab =. (1) m β,ab Σ T αβ,ab, Σ ββ,ab m α,ab = E(α ab Y ab ) = D α,a Γ T ξ B T abcov(y ab ) 1 (Y ab B ab γ µ,a ) () and m β,ab = E(β ab Y ab ) = V ab Γ T η,abb T abcov(y ab ) 1 (Y ab B ab γ µ,a ). (3) 1

2 The conditional covariance matrix is ( ) } ( ) ( ) α ab cov β Y D α,a 0 D α,a Γ T ξ B T ab ab = cov(y ab 0 V ab V ab Γ T η,abb T ab )} 1 (B ab Γ ξ D α,a, B ab Γ η,ab V ab ). ab Therefore, we have Σ αα,ab = D α,a D α,a Γ T ξ B T abcov(y ab )} 1 B ab Γ ξ D α,a, (4) Σ ββ,ab = V ab V ab Γ T η,abb T abcov(y ab )} 1 B ab Γ η,ab V ab, (5) and Σ αβ,ab = D α,a Γ T ξ B T abcov(y ab )} 1 B ab Γ η,ab V ab. (6) The desired predictions required by the EM algorithm are α ab = E(α ab Y ab ) = m α,ab, βab = E(β ab Y ab ) = m β,ab, α ab α T ab = E(α abα T ab Y ab ) = α ab α T ab + Σ αα,ab β ab β T ab = E(β ab β T ab Y ab ) = β ab βt ab + Σ ββ,ab, α ab β T ab = E(α ab β T ab Y ab ) = α ab βt ab + Σ αβ,ab. (7) The first two equalities also give expressions of the best linear unbiased predictors (BLUP) of the random effects α ab and β ab. Computation of V ab. It is convenient to re-group the elements of β ab according to the principal components. Denote β ab,j = (β ab1j,..., β abcab j) T, j = 1,..., K η, and β ab = (β T ab,1,..., β T ab,k η ) T. Let P abj be a C ab by C ab matrix with elements P abj,cc = ρ( x abc x abc ; θ aj ), c, c = 1,..., C ab. Note that P abj = P T abj. Then Σ abj = cov(β ab,j ) = σβ,aj P abj. Since cov(β ab,j, β ab,j ) = 0 for j j, the covariance matrix of β ab is block diagonal Ṽ ab = cov( β ab ) = diag(σ ab1,... Σ abkη ). Note that β ab is just a reordering of β ab, that is, β ab = O β ab for a permutation matrix O. It follows the covariance matrix of β ab is V ab = OṼabO T. Circumventing inversion of cov(y ab ). We suppress the subscripts of D α,a and V ab in the following discussion. Denote E = B ab Γ ξ, F = B ab Γ η,ab, and S = cov(y ab ) = EDE T + FVF T + σ I. To calculate the conditional moments given in () (6), we need to compute DE T S 1 in (), VF T S 1 in (3), D DE T S 1 ED in (4), V VF T S 1 F T V in (6), and DE T S 1 FV in (5). Denote J = (EDE T + σ I) 1 and K = (FVF T + σ I) 1. Repeatedly using the identities (A 1 + C T B 1 C) 1 = A AC T (CAC T + B) 1 CA, (8) (A 1 + C T B 1 C) 1 C T B 1 = AC T (CAC T + B) 1, (9)

3 we obtain J = σ I σ E(E T E + σ D 1 ) 1 E T, (10) K = σ I σ F(F T F + σ V 1 ) 1 F T, (11) and DE T S 1 = (D 1 + E T KE) 1 E T K, VF T S 1 = (V 1 + F T JF) 1 F T J, D DE T S 1 ED = (D 1 + E T KE) 1, V VF T S 1 FV = (V 1 + F T JF) 1, DE T S 1 FV = DE T JF(V 1 + F T JF) 1. The matrices that need to be inverted here are of the same size as D or V, which is K ξ K ξ or (C ab K η ) (C ab K η ), much smaller than the size of cov(y ab ).. Updating Formula in the M-step of the EM Algorithm In the updating formulas given below, the parameters appear on the right hand side of equations are all fixed at their values obtained from the previous iteration of the algorithm. 1. Update the estimate of σ. Let N = A Ba a=1 N ab. The updating formula is σ = 1 A E(ɛ T N abɛ ab Y ab ), a=1 where ɛ ab = Y ab B ab γ µ,a B ab Γ ξ α ab B ab Γ η,ab β ab. Using (7), we obtain that E(ɛ T ab ɛ ab Y ab ) can be expressed as (Y ab B ab γ µ,a B ab Γ ξ α ab B ab Γ η,ab βab ) T (Y ab B ab γ µ,a B ab Γ ξ α ab B ab Γ η,ab βab ) D α,a is + trace(b ab Γ ξ Σ αα,ab Γ T ξ B T ab + B ab Γ η,ab Σ ββ,ab Γ T η,abb T ab + B ab Γ ξ Σ αβ,ab Γ T η,abb T ab).. Update D α,a = diag(σα,aj) and D β,a = diag(σβ,aj ). The updating formula for 1 D α,a = diag B a } α ab α T ab. (1) To update D β,a, we need to minimize the conditional expectation of K η j=1 log( V ab )+β T abv 1 ab β ab = K η j=1 C ab log(σ β,aj)+log P abj +σ β,aj βt ab,jp 1 ab,j β ab,j }. 3

4 Thus the updating formulas are ( σ β,aj 1 = trace Ba C ab P 1 ab,j β ab,j β T ab,j ), j = 1,..., K η. (13) If the variances of random effects do not depend on the treatment level a, we simply need to average over a in (1) and (13). 3. Update γ µ,a, a = 1,..., A. The updating formula are Ba } 1 Ba γ µ,a = B T abb ab + σ λ µ b (t)b (t) T B T ab(y ab B ab Γ ξ α ab B ab Γ η,ab βab ). 4. Update the columns of Γ ξ sequentially. For j = 1,..., K ξ, the updating formula are A γ ξ,j = α abj BT abb ab + σ λ ξ a=1 A a=1 B T ab b (t)b (t) T } 1 ( (Y ab B ab γ µ,a ) α abj B ab l j ) γ ξ,l α abj α abl B ab Γ η,ab α abj β ab }. 5. Update the columns of Γ η sequentially. For j = 1,..., K η, the updating formula are A C ab γ η,j = β abcj BT abcb abc + σ λ η,a a=1 c=1 A C ab a=1 c=1 B T abc b (t)b (t) T } 1 (Y abc B abc γ µ,a ) β abcj B abc Γ ξ β abcj α ab B abc ( l j γ η,l β abcj β abcl )}. 6. Orthogonalizing Γ ξ and Γ η. The matrix Γ ξ and Γ η obtained in Steps 4 and 5 need not have orthonormal columns. We orthogonalize them in this step. Compute the QR decomposition Γ ξ = Q ξ R ξ where Q ξ has orthonormal columns. Let Γ ξ Q ξ. Orthogonalize Γ η similarly. 7. Update the correlation parameter θ aj. Given the current estimates of other parameters, we minimize for each a and j the conditional expectation of which is log P 1 abj + 1 } β T σ ab,jp 1 β,aj abj β ab,j, log P 1 abj + 1 trace σβ,aj 4 ( P 1 abj β ab,j β T ab,j) }. (14)

5 The minimizer can be found by a gradient-based method. Denote the components of θ aj as θ ajk s. The gradient with respect to θ aj is a vector with elements ( trace P 1 abj P ) abj 1 ( trace P 1 P abj θ ajk σβ,aj abj P 1 abjβ θ ab,j βab,j) } T. ajk Alternatively, a direct search algorithm that does not require the calculation of the gradient, such as the downhill simplex method of Nelder and Mead (1965), can be applied to minimize (14). If the correlation parameters do not depend on j, we consider for each a a new objective function by summing (14) over j. If the correlation parameters do not depend on a, we need only sum (14) over a for each j to define the new objective function. Note that the orthogonalization in Step 6 is used to ensure identifiability. This approach was adapted from Zhou et al. (008) and worked well in our simulation study and real data analysis. An alternative approach is direct optimization within a restricted space (Peng and Paul, 009) but its implementation is beyond the scope of this paper..3 Computing the Observed Data Log Likelihood When we do crossvalidation, we need to compute the log likelihood of observed data, which depends on the determinant and inverse of the possibly very large matrix S, defined in Section.1. This section gives a method for computation of S and S 1 without constructing the matrix S. Using the identity A + BC T = A I + C T A 1 B, (15) we obtain S = σ I + EDE T + FVF T = σ I + EDE T I + V 1/ F T (σ I + EDE T ) 1 FV 1/. The two factors on the right hand side of the above equation can be computed as follows. Using the identity (15), we have σ I + EDE T = σ N I + σ D 1/ E T ED 1/ = σ N D D 1 + σ E T E. Using (σ I + EDE T ) 1 = σ I σ E(E T E + σ D 1 ) 1 E T, we obtain I + V 1/ F T (σ I + EDE T ) 1 FV 1/ = V V 1 + σ F T F σ F T E(E T E + σ D 1 ) 1 E T F. 5

6 Using the definition of J in Section.1 and the identity (8), we have S 1 = (J 1 + FVF T ) 1 = J JF(F T JF + V 1 ) 1 F T J, where J can be computed using (10)..4 Calculating the partial derivatives of P abj Consider the Matérn family of autocorrelation functions ( ) ρ(d; φ, ν) = 1 ν dν 1/ ν ( ) dν 1/ K ν, Γ(ν) φ φ φ > 0, ν > 0, where K ν ( ) is the modified Bessel function of order ν taking the form with K ν (u) = I ν (u) = m=0 π sin(νπ) I ν(u) I ν (u)} ( ) m+ν 1 u. m!γ(m + ν + 1) Denote u = xν 1/ /φ, then the partial derivatives of ρ(x; φ, ν) are ρ(x; φ, ν) φ = φγ(ν)( 1 u ) ν[ukν 1 (u) + K ν+1 (u)} νk ν (u)] and ρ(x; φ, ν) ν = ( u ) ψ(ν) π cot(νπ) + log + 1 } ρ(x; φ, ν) π ) [ ν u + ν I ν+1(u) I ν+1 (u)} u Γ(ν) sin(νπ)( ( u ) log m=0 m=0 } I ν (u) + I ν (u)} 1 u ) m+νψ(m + ν + 1) m!γ(m + ν + 1)( 1 u ) ] m νψ(m ν + 1), m!γ(m ν + 1)( where ψ(z) = C l=0 ( 1 1 ) with C = lim z+l 1+l k k 1 l=1 ln(k)}.557, and k Γ( ) is the Gamma function. 6

7 3 Additional simulation results Two other setups from the Bayesian hierarchical model in Baladandayuthapani, et al. (008) were also considered but the software of the Bayesian method encountered some serious numerical problems and could only run on a small proportion of simulated data sets. We ran simulation 100 times for each setup. The results of our methods are based on three unit and subunit level principal components. Our methods successfully ran on all data sets. and Setup 6. Same as Setup 5 except that Σ 1 = Σ = Σ 31 = Σ 3 = For this setup, the Bayesian method only ran on 10 out of 100 simulated data sets. Setup 7. Same as Setup 6 except that σ = σ 3 = 0.4. For this setup, the Bayesian method only ran on 3 out of 100 simulated data sets. We used the measures defined at the beginning of Section 5 of the paper to assess/compare performance of the two methods. For our reduced rank methods, the measures were computed using all 100 data sets as well as only those data sets that the Bayesian method ran. The results are summarized in Table 1. Our reduced rank method is comparable to the Bayesian method for estimating the mean functions and predicting the random effects. Next we report some simulation results on parameter estimation for setups following our model. Note that comparison with the Bayesian method is not available since two methods use different parameters. Table shows the summary statistics of correlation parameter estimates for the first three simulation setups where the data were generated from our model. The parameter estimates are reasonably unbiased. Figure 1 presents the pointwise sample mean of the estimated treatment effects, unit level and sub-unit level PCs, along with 90% pointwise coverage intervals for simulation Setup. It is not surprising that sub-unit level PCs are better estimated than the unit level PCs since there are more sub-unit level data than the unit level data.. 7

8 Table 1: Comparison of two methods based on 100 simulation runs. Mean (SD) of the integrated absolute errors of estimating the mean functions and predicting the unit and subunit level random effects. Numbers shown are the actual numbers multiplied by 10. Reduced rank refers to our method; Bayesian refers to the Bayesian method of Baladandayuthapani, et al. (008). 100 sims, 10 sims and 3 sims indicate the number of data sets used in the calculation. Setup Method Mean Unit Subunit 6 7 Reduced rank (100 sims) (1.45) (1.088) (0.481) Reduced rank (10 sims) (0.963) (1.044) (0.305) Bayesian (10 sims) (0.963) 6.55 (0.899) (0.93) Reduced rank (100 sims) (1.515) (1.187) (1.064) Reduced rank (3 sims) (1.34) 7.86 (0.890) (0.845) Bayesian (3 sims) (1.344) 7.84 (0.99) (0.567) Table : Sample mean (SD) of correlation parameter estimates, based on 100 simulation runs. PC 1 PC Setup φ = 8 ν = 0.1 φ = 4 ν = (4.98) (0.00) (5.67) (0.018) 4.6 (1.306) (0.040) (.369) (0.019) (1.19) (0.036) 4 Colon Carcinogenesis Data Figures and 3 contains respectively residual plots after fitting our reduced rank model and the Bayesian model of Baladandayuthapani, et al. (008). It is clear that our model fits the data better. After fitting the Bayesian model, there were still obvious patterns left in the residual plots for treatment groups FO+B and FO B. References Baladandayuthapani, V., Mallick, B., Turner, N., Hong, M., Chapkin, R., Lupton, J. and Carroll, R. J. (008). Bayesian hierarchical spatially correlated functional data analysis with application to colon carcinogenesis. Biometrics 64, Nelder, J.A. & Mead, R. (1965). Computer Journal, 7, A simplex method for function minimization. Peng, J. and Paul, D. (009). A geometric approach to maximum likelihood estimation of the functional principal components from sparse longitudinal data. Journal of Computational and Graphical Statistics, to appear. 8

9 Figure 1: Simulation Setup. The pointwise sample mean of the estimated treatment effects, unit level and sub-unit level PCs, along with 90% pointwise coverage intervals. Group 1 Unit PC 1 Sub unit PC Group Unit PC Sub unit PC t Zhou, L., Huang, J.Z. and Carroll, J.R. (008). Joint modelling of paired sparse functional data using principal components. Biometrika, 95,

10 Figure : Residual plots after fitting the reduced rank model. Separate plots are drawn for the four diet groups. CO is Corn Oil, FO is Fish Oil, +B and B represent with or without (±) Butyrate supplement. Diet CO+B Diet CO B Residual Diet FO+B Diet FO B Fitted p7 10

11 Figure 3: Residual plots after fitting the Bayesian model of Baladandayuthapani, et al. (008). Separate plots are drawn for the four diet groups. CO is Corn Oil, FO is Fish Oil, +B and B represent with or without Butyrate supplement. Diet CO+B Diet CO B Residual Diet FO+B Diet FO B 3 Fitted p7 11

Fast Methods for Spatially Correlated Multilevel Functional Data

Fast Methods for Spatially Correlated Multilevel Functional Data Fast Methods for Spatially Correlated Multilevel Functional Data Ana-Maria Staicu, Department of Statistics, North Carolina State University, 23 Stinson Drive Raleigh, NC 27695-8203, USA email: staicu@stat.ncsu.edu,

More information

Functional Latent Feature Models. With Single-Index Interaction

Functional Latent Feature Models. With Single-Index Interaction Generalized With Single-Index Interaction Department of Statistics Center for Statistical Bioinformatics Institute for Applied Mathematics and Computational Science Texas A&M University Naisyin Wang and

More information

Wavelet-Based Nonparametric Modeling of Hierarchical Functions in Colon Carcinogenesis

Wavelet-Based Nonparametric Modeling of Hierarchical Functions in Colon Carcinogenesis Wavelet-Based Nonparametric Modeling of Hierarchical Functions in Colon Carcinogenesis Jeffrey S. Morris University of Texas, MD Anderson Cancer Center Joint wor with Marina Vannucci, Philip J. Brown,

More information

Bayesian Hierarchical Spatially Correlated Functional Data Analysis with Application to Colon Carcinogenesis. Abstract

Bayesian Hierarchical Spatially Correlated Functional Data Analysis with Application to Colon Carcinogenesis. Abstract Bayesian Hierarchical Spatially Correlated Functional Data Analysis with Application to Colon Carcinogenesis Veerabhadran Baladandayuthapani 1, Bani K. Mallick 2, Mee Young Hong 3 4, Joanne R. Lupton 4,

More information

Using Estimating Equations for Spatially Correlated A

Using Estimating Equations for Spatially Correlated A Using Estimating Equations for Spatially Correlated Areal Data December 8, 2009 Introduction GEEs Spatial Estimating Equations Implementation Simulation Conclusion Typical Problem Assess the relationship

More information

Concentration Ellipsoids

Concentration Ellipsoids Concentration Ellipsoids ECE275A Lecture Supplement Fall 2008 Kenneth Kreutz Delgado Electrical and Computer Engineering Jacobs School of Engineering University of California, San Diego VERSION LSECE275CE

More information

Kriging models with Gaussian processes - covariance function estimation and impact of spatial sampling

Kriging models with Gaussian processes - covariance function estimation and impact of spatial sampling Kriging models with Gaussian processes - covariance function estimation and impact of spatial sampling François Bachoc former PhD advisor: Josselin Garnier former CEA advisor: Jean-Marc Martinez Department

More information

Hypothesis testing in multilevel models with block circular covariance structures

Hypothesis testing in multilevel models with block circular covariance structures 1/ 25 Hypothesis testing in multilevel models with block circular covariance structures Yuli Liang 1, Dietrich von Rosen 2,3 and Tatjana von Rosen 1 1 Department of Statistics, Stockholm University 2 Department

More information

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a

More information

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Financial Econometrics / 49

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Financial Econometrics / 49 State-space Model Eduardo Rossi University of Pavia November 2013 Rossi State-space Model Financial Econometrics - 2013 1 / 49 Outline 1 Introduction 2 The Kalman filter 3 Forecast errors 4 State smoothing

More information

arxiv: v1 [math.st] 22 Dec 2018

arxiv: v1 [math.st] 22 Dec 2018 Optimal Designs for Prediction in Two Treatment Groups Rom Coefficient Regression Models Maryna Prus Otto-von-Guericke University Magdeburg, Institute for Mathematical Stochastics, PF 4, D-396 Magdeburg,

More information

26. Filtering. ECE 830, Spring 2014

26. Filtering. ECE 830, Spring 2014 26. Filtering ECE 830, Spring 2014 1 / 26 Wiener Filtering Wiener filtering is the application of LMMSE estimation to recovery of a signal in additive noise under wide sense sationarity assumptions. Problem

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1 MA 575 Linear Models: Cedric E Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1 1 Within-group Correlation Let us recall the simple two-level hierarchical

More information

Course topics (tentative) The role of random effects

Course topics (tentative) The role of random effects Course topics (tentative) random effects linear mixed models analysis of variance frequentist likelihood-based inference (MLE and REML) prediction Bayesian inference The role of random effects Rasmus Waagepetersen

More information

. a m1 a mn. a 1 a 2 a = a n

. a m1 a mn. a 1 a 2 a = a n Biostat 140655, 2008: Matrix Algebra Review 1 Definition: An m n matrix, A m n, is a rectangular array of real numbers with m rows and n columns Element in the i th row and the j th column is denoted by

More information

Next is material on matrix rank. Please see the handout

Next is material on matrix rank. Please see the handout B90.330 / C.005 NOTES for Wednesday 0.APR.7 Suppose that the model is β + ε, but ε does not have the desired variance matrix. Say that ε is normal, but Var(ε) σ W. The form of W is W w 0 0 0 0 0 0 w 0

More information

Covariance function estimation in Gaussian process regression

Covariance function estimation in Gaussian process regression Covariance function estimation in Gaussian process regression François Bachoc Department of Statistics and Operations Research, University of Vienna WU Research Seminar - May 2015 François Bachoc Gaussian

More information

for Sparsely Observed Functional Data University of Glasgow

for Sparsely Observed Functional Data University of Glasgow Spatial Correlation Estimation for Sparsely Observed Functional Data Workshop on High dimensional and dependent functional data, Bristol, Sep 2012 Surajit Ray University of Glasgow Contributors t Chong

More information

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Fin. Econometrics / 53

State-space Model. Eduardo Rossi University of Pavia. November Rossi State-space Model Fin. Econometrics / 53 State-space Model Eduardo Rossi University of Pavia November 2014 Rossi State-space Model Fin. Econometrics - 2014 1 / 53 Outline 1 Motivation 2 Introduction 3 The Kalman filter 4 Forecast errors 5 State

More information

Geostatistical Modeling for Large Data Sets: Low-rank methods

Geostatistical Modeling for Large Data Sets: Low-rank methods Geostatistical Modeling for Large Data Sets: Low-rank methods Whitney Huang, Kelly-Ann Dixon Hamil, and Zizhuang Wu Department of Statistics Purdue University February 22, 2016 Outline Motivation Low-rank

More information

Multivariate Statistical Analysis

Multivariate Statistical Analysis Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 4 for Applied Multivariate Analysis Outline 1 Eigen values and eigen vectors Characteristic equation Some properties of eigendecompositions

More information

Fast Methods for Spatially Correlated Multilevel Functional Data

Fast Methods for Spatially Correlated Multilevel Functional Data Biostatistics (2007), 8, 1, pp. 1 29 doi:10.1093/biostatistics/kxl000 Advance Access publication on May 15, 2006 Fast Methods for Spatially Correlated Multilevel Functional Data ANA-MARIA STAICU Department

More information

Reflections and Rotations in R 3

Reflections and Rotations in R 3 Reflections and Rotations in R 3 P. J. Ryan May 29, 21 Rotations as Compositions of Reflections Recall that the reflection in the hyperplane H through the origin in R n is given by f(x) = x 2 ξ, x ξ (1)

More information

An Introduction to Bayesian Linear Regression

An Introduction to Bayesian Linear Regression An Introduction to Bayesian Linear Regression APPM 5720: Bayesian Computation Fall 2018 A SIMPLE LINEAR MODEL Suppose that we observe explanatory variables x 1, x 2,..., x n and dependent variables y 1,

More information

Multivariate Distributions

Multivariate Distributions IEOR E4602: Quantitative Risk Management Spring 2016 c 2016 by Martin Haugh Multivariate Distributions We will study multivariate distributions in these notes, focusing 1 in particular on multivariate

More information

Prediction. is a weighted least squares estimate since it minimizes. Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark

Prediction. is a weighted least squares estimate since it minimizes. Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark Prediction Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark March 22, 2017 WLS and BLUE (prelude to BLUP) Suppose that Y has mean β and known covariance matrix V (but Y need not

More information

Information geometry for bivariate distribution control

Information geometry for bivariate distribution control Information geometry for bivariate distribution control C.T.J.Dodson + Hong Wang Mathematics + Control Systems Centre, University of Manchester Institute of Science and Technology Optimal control of stochastic

More information

WLS and BLUE (prelude to BLUP) Prediction

WLS and BLUE (prelude to BLUP) Prediction WLS and BLUE (prelude to BLUP) Prediction Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark April 21, 2018 Suppose that Y has mean X β and known covariance matrix V (but Y need

More information

Bare minimum on matrix algebra. Psychology 588: Covariance structure and factor models

Bare minimum on matrix algebra. Psychology 588: Covariance structure and factor models Bare minimum on matrix algebra Psychology 588: Covariance structure and factor models Matrix multiplication 2 Consider three notations for linear combinations y11 y1 m x11 x 1p b11 b 1m y y x x b b n1

More information

Multilevel Cross-dependent Binary Longitudinal Data

Multilevel Cross-dependent Binary Longitudinal Data Multilevel Cross-dependent Binary Longitudinal Data Nicoleta Serban 1 H. Milton Stewart School of Industrial Systems and Engineering Georgia Institute of Technology nserban@isye.gatech.edu Ana-Maria Staicu

More information

Parameter Estimation in the Spatio-Temporal Mixed Effects Model Analysis of Massive Spatio-Temporal Data Sets

Parameter Estimation in the Spatio-Temporal Mixed Effects Model Analysis of Massive Spatio-Temporal Data Sets Parameter Estimation in the Spatio-Temporal Mixed Effects Model Analysis of Massive Spatio-Temporal Data Sets Matthias Katzfuß Advisor: Dr. Noel Cressie Department of Statistics The Ohio State University

More information

1 Cricket chirps: an example

1 Cricket chirps: an example Notes for 2016-09-26 1 Cricket chirps: an example Did you know that you can estimate the temperature by listening to the rate of chirps? The data set in Table 1 1. represents measurements of the number

More information

Principal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis. Chris Funk. Lecture 17

Principal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis. Chris Funk. Lecture 17 Principal Component Analysis-I Geog 210C Introduction to Spatial Data Analysis Chris Funk Lecture 17 Outline Filters and Rotations Generating co-varying random fields Translating co-varying fields into

More information

Regularized Discriminant Analysis and Reduced-Rank LDA

Regularized Discriminant Analysis and Reduced-Rank LDA Regularized Discriminant Analysis and Reduced-Rank LDA Department of Statistics The Pennsylvania State University Email: jiali@stat.psu.edu Regularized Discriminant Analysis A compromise between LDA and

More information

On Gaussian Process Models for High-Dimensional Geostatistical Datasets

On Gaussian Process Models for High-Dimensional Geostatistical Datasets On Gaussian Process Models for High-Dimensional Geostatistical Datasets Sudipto Banerjee Joint work with Abhirup Datta, Andrew O. Finley and Alan E. Gelfand University of California, Los Angeles, USA May

More information

Lecture 4 Orthonormal vectors and QR factorization

Lecture 4 Orthonormal vectors and QR factorization Orthonormal vectors and QR factorization 4 1 Lecture 4 Orthonormal vectors and QR factorization EE263 Autumn 2004 orthonormal vectors Gram-Schmidt procedure, QR factorization orthogonal decomposition induced

More information

Kneib, Fahrmeir: Supplement to "Structured additive regression for categorical space-time data: A mixed model approach"

Kneib, Fahrmeir: Supplement to Structured additive regression for categorical space-time data: A mixed model approach Kneib, Fahrmeir: Supplement to "Structured additive regression for categorical space-time data: A mixed model approach" Sonderforschungsbereich 386, Paper 43 (25) Online unter: http://epub.ub.uni-muenchen.de/

More information

Basic Concepts in Matrix Algebra

Basic Concepts in Matrix Algebra Basic Concepts in Matrix Algebra An column array of p elements is called a vector of dimension p and is written as x p 1 = x 1 x 2. x p. The transpose of the column vector x p 1 is row vector x = [x 1

More information

Chapter 5 Prediction of Random Variables

Chapter 5 Prediction of Random Variables Chapter 5 Prediction of Random Variables C R Henderson 1984 - Guelph We have discussed estimation of β, regarded as fixed Now we shall consider a rather different problem, prediction of random variables,

More information

Algebra C Numerical Linear Algebra Sample Exam Problems

Algebra C Numerical Linear Algebra Sample Exam Problems Algebra C Numerical Linear Algebra Sample Exam Problems Notation. Denote by V a finite-dimensional Hilbert space with inner product (, ) and corresponding norm. The abbreviation SPD is used for symmetric

More information

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature

More information

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, Linear Regression In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, y = Xβ + ɛ, where y t = (y 1,..., y n ) is the column vector of target values,

More information

A Cautionary Note on Generalized Linear Models for Covariance of Unbalanced Longitudinal Data

A Cautionary Note on Generalized Linear Models for Covariance of Unbalanced Longitudinal Data A Cautionary Note on Generalized Linear Models for Covariance of Unbalanced Longitudinal Data Jianhua Z. Huang a, Min Chen b, Mehdi Maadooliat a, Mohsen Pourahmadi a a Department of Statistics, Texas A&M

More information

Hierarchical Modeling for Univariate Spatial Data

Hierarchical Modeling for Univariate Spatial Data Hierarchical Modeling for Univariate Spatial Data Geography 890, Hierarchical Bayesian Models for Environmental Spatial Data Analysis February 15, 2011 1 Spatial Domain 2 Geography 890 Spatial Domain This

More information

Lecture 4: Types of errors. Bayesian regression models. Logistic regression

Lecture 4: Types of errors. Bayesian regression models. Logistic regression Lecture 4: Types of errors. Bayesian regression models. Logistic regression A Bayesian interpretation of regularization Bayesian vs maximum likelihood fitting more generally COMP-652 and ECSE-68, Lecture

More information

A General Framework for Variable Selection in Linear Mixed Models with Applications to Genetic Studies with Structured Populations

A General Framework for Variable Selection in Linear Mixed Models with Applications to Genetic Studies with Structured Populations A General Framework for Variable Selection in Linear Mixed Models with Applications to Genetic Studies with Structured Populations Joint work with Karim Oualkacha (UQÀM), Yi Yang (McGill), Celia Greenwood

More information

Generalized Functional Linear Models with Semiparametric Single-Index Interactions

Generalized Functional Linear Models with Semiparametric Single-Index Interactions Generalized Functional Linear Models with Semiparametric Single-Index Interactions Yehua Li Department of Statistics, University of Georgia, Athens, GA 30602, yehuali@uga.edu Naisyin Wang Department of

More information

Hierarchical Functional Data with Mixed Continuous and Binary Measurements

Hierarchical Functional Data with Mixed Continuous and Binary Measurements Biometrics 70, 802 811 December 2014 DOI: 10.1111/biom.12211 Hierarchical Functional Data with Mixed Continuous and Binary Measurements Haocheng Li, 1, * John Staudenmayer, 2 and Raymond J. Carroll 1 1

More information

sparse and low-rank tensor recovery Cubic-Sketching

sparse and low-rank tensor recovery Cubic-Sketching Sparse and Low-Ran Tensor Recovery via Cubic-Setching Guang Cheng Department of Statistics Purdue University www.science.purdue.edu/bigdata CCAM@Purdue Math Oct. 27, 2017 Joint wor with Botao Hao and Anru

More information

Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets

Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets Hierarchical Nearest-Neighbor Gaussian Process Models for Large Geo-statistical Datasets Abhirup Datta 1 Sudipto Banerjee 1 Andrew O. Finley 2 Alan E. Gelfand 3 1 University of Minnesota, Minneapolis,

More information

Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model

Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model Xiuming Zhang zhangxiuming@u.nus.edu A*STAR-NUS Clinical Imaging Research Center October, 015 Summary This report derives

More information

STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method.

STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method. STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method. Rebecca Barter May 5, 2015 Linear Regression Review Linear Regression Review

More information

Preface Introduction to Statistics and Data Analysis Overview: Statistical Inference, Samples, Populations, and Experimental Design The Role of

Preface Introduction to Statistics and Data Analysis Overview: Statistical Inference, Samples, Populations, and Experimental Design The Role of Preface Introduction to Statistics and Data Analysis Overview: Statistical Inference, Samples, Populations, and Experimental Design The Role of Probability Sampling Procedures Collection of Data Measures

More information

Linear Methods for Regression. Lijun Zhang

Linear Methods for Regression. Lijun Zhang Linear Methods for Regression Lijun Zhang zlj@nju.edu.cn http://cs.nju.edu.cn/zlj Outline Introduction Linear Regression Models and Least Squares Subset Selection Shrinkage Methods Methods Using Derived

More information

Statistics 910, #5 1. Regression Methods

Statistics 910, #5 1. Regression Methods Statistics 910, #5 1 Overview Regression Methods 1. Idea: effects of dependence 2. Examples of estimation (in R) 3. Review of regression 4. Comparisons and relative efficiencies Idea Decomposition Well-known

More information

Nearest Neighbor Gaussian Processes for Large Spatial Data

Nearest Neighbor Gaussian Processes for Large Spatial Data Nearest Neighbor Gaussian Processes for Large Spatial Data Abhi Datta 1, Sudipto Banerjee 2 and Andrew O. Finley 3 July 31, 2017 1 Department of Biostatistics, Bloomberg School of Public Health, Johns

More information

EM Algorithm II. September 11, 2018

EM Algorithm II. September 11, 2018 EM Algorithm II September 11, 2018 Review EM 1/27 (Y obs, Y mis ) f (y obs, y mis θ), we observe Y obs but not Y mis Complete-data log likelihood: l C (θ Y obs, Y mis ) = log { f (Y obs, Y mis θ) Observed-data

More information

Testing Some Covariance Structures under a Growth Curve Model in High Dimension

Testing Some Covariance Structures under a Growth Curve Model in High Dimension Department of Mathematics Testing Some Covariance Structures under a Growth Curve Model in High Dimension Muni S. Srivastava and Martin Singull LiTH-MAT-R--2015/03--SE Department of Mathematics Linköping

More information

Matrices and systems of linear equations

Matrices and systems of linear equations Matrices and systems of linear equations Samy Tindel Purdue University Differential equations and linear algebra - MA 262 Taken from Differential equations and linear algebra by Goode and Annin Samy T.

More information

Economics Department LSE. Econometrics: Timeseries EXERCISE 1: SERIAL CORRELATION (ANALYTICAL)

Economics Department LSE. Econometrics: Timeseries EXERCISE 1: SERIAL CORRELATION (ANALYTICAL) Economics Department LSE EC402 Lent 2015 Danny Quah TW1.10.01A x7535 : Timeseries EXERCISE 1: SERIAL CORRELATION (ANALYTICAL) 1. Suppose ɛ is w.n. (0, σ 2 ), ρ < 1, and W t = ρw t 1 + ɛ t, for t = 1, 2,....

More information

X t = a t + r t, (7.1)

X t = a t + r t, (7.1) Chapter 7 State Space Models 71 Introduction State Space models, developed over the past 10 20 years, are alternative models for time series They include both the ARIMA models of Chapters 3 6 and the Classical

More information

Mixed models in R using the lme4 package Part 4: Theory of linear mixed models

Mixed models in R using the lme4 package Part 4: Theory of linear mixed models Mixed models in R using the lme4 package Part 4: Theory of linear mixed models Douglas Bates 8 th International Amsterdam Conference on Multilevel Analysis 2011-03-16 Douglas Bates

More information

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1 Inverse of a Square Matrix For an N N square matrix A, the inverse of A, 1 A, exists if and only if A is of full rank, i.e., if and only if no column of A is a linear combination 1 of the others. A is

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Maximum Likelihood Estimation Merlise Clyde STA721 Linear Models Duke University August 31, 2017 Outline Topics Likelihood Function Projections Maximum Likelihood Estimates Readings: Christensen Chapter

More information

PQL Estimation Biases in Generalized Linear Mixed Models

PQL Estimation Biases in Generalized Linear Mixed Models PQL Estimation Biases in Generalized Linear Mixed Models Woncheol Jang Johan Lim March 18, 2006 Abstract The penalized quasi-likelihood (PQL) approach is the most common estimation procedure for the generalized

More information

Functional SVD for Big Data

Functional SVD for Big Data Functional SVD for Big Data Pan Chao April 23, 2014 Pan Chao Functional SVD for Big Data April 23, 2014 1 / 24 Outline 1 One-Way Functional SVD a) Interpretation b) Robustness c) CV/GCV 2 Two-Way Problem

More information

1 Bayesian Linear Regression (BLR)

1 Bayesian Linear Regression (BLR) Statistical Techniques in Robotics (STR, S15) Lecture#10 (Wednesday, February 11) Lecturer: Byron Boots Gaussian Properties, Bayesian Linear Regression 1 Bayesian Linear Regression (BLR) In linear regression,

More information

For more information about how to cite these materials visit

For more information about how to cite these materials visit Author(s): Kerby Shedden, Ph.D., 2010 License: Unless otherwise noted, this material is made available under the terms of the Creative Commons Attribution Share Alike 3.0 License: http://creativecommons.org/licenses/by-sa/3.0/

More information

Point-Referenced Data Models

Point-Referenced Data Models Point-Referenced Data Models Jamie Monogan University of Georgia Spring 2013 Jamie Monogan (UGA) Point-Referenced Data Models Spring 2013 1 / 19 Objectives By the end of these meetings, participants should

More information

Lecture 20: Linear model, the LSE, and UMVUE

Lecture 20: Linear model, the LSE, and UMVUE Lecture 20: Linear model, the LSE, and UMVUE Linear Models One of the most useful statistical models is X i = β τ Z i + ε i, i = 1,...,n, where X i is the ith observation and is often called the ith response;

More information

STAT 512 sp 2018 Summary Sheet

STAT 512 sp 2018 Summary Sheet STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}

More information

Course Notes: Week 1

Course Notes: Week 1 Course Notes: Week 1 Math 270C: Applied Numerical Linear Algebra 1 Lecture 1: Introduction (3/28/11) We will focus on iterative methods for solving linear systems of equations (and some discussion of eigenvalues

More information

Random vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables.

Random vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables. Random vectors Recall that a random vector X = X X 2 is made up of, say, k random variables X k A random vector has a joint distribution, eg a density f(x), that gives probabilities P(X A) = f(x)dx Just

More information

Part 6: Multivariate Normal and Linear Models

Part 6: Multivariate Normal and Linear Models Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of

More information

Sparse Bayesian Logistic Regression with Hierarchical Prior and Variational Inference

Sparse Bayesian Logistic Regression with Hierarchical Prior and Variational Inference Sparse Bayesian Logistic Regression with Hierarchical Prior and Variational Inference Shunsuke Horii Waseda University s.horii@aoni.waseda.jp Abstract In this paper, we present a hierarchical model which

More information

CS281 Section 4: Factor Analysis and PCA

CS281 Section 4: Factor Analysis and PCA CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we

More information

[y i α βx i ] 2 (2) Q = i=1

[y i α βx i ] 2 (2) Q = i=1 Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation

More information

Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models

Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models Tihomir Asparouhov 1, Bengt Muthen 2 Muthen & Muthen 1 UCLA 2 Abstract Multilevel analysis often leads to modeling

More information

Multivariate Linear Regression Models

Multivariate Linear Regression Models Multivariate Linear Regression Models Regression analysis is used to predict the value of one or more responses from a set of predictors. It can also be used to estimate the linear association between

More information

Parameter Estimation

Parameter Estimation Parameter Estimation Tuesday 9 th May, 07 4:30 Consider a system whose response can be modeled by R = M (Θ) where Θ is a vector of m parameters. We take a series of measurements, D (t) where t represents

More information

On construction of constrained optimum designs

On construction of constrained optimum designs On construction of constrained optimum designs Institute of Control and Computation Engineering University of Zielona Góra, Poland DEMA2008, Cambridge, 15 August 2008 Numerical algorithms to construct

More information

2. Matrix Algebra and Random Vectors

2. Matrix Algebra and Random Vectors 2. Matrix Algebra and Random Vectors 2.1 Introduction Multivariate data can be conveniently display as array of numbers. In general, a rectangular array of numbers with, for instance, n rows and p columns

More information

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Put your solution to each problem on a separate sheet of paper. Problem 1. (5166) Assume that two random samples {x i } and {y i } are independently

More information

Basics of Point-Referenced Data Models

Basics of Point-Referenced Data Models Basics of Point-Referenced Data Models Basic tool is a spatial process, {Y (s), s D}, where D R r Chapter 2: Basics of Point-Referenced Data Models p. 1/45 Basics of Point-Referenced Data Models Basic

More information

Matrix Approach to Simple Linear Regression: An Overview

Matrix Approach to Simple Linear Regression: An Overview Matrix Approach to Simple Linear Regression: An Overview Aspects of matrices that you should know: Definition of a matrix Addition/subtraction/multiplication of matrices Symmetric/diagonal/identity matrix

More information

Fixed Effects, Invariance, and Spatial Variation in Intergenerational Mobility

Fixed Effects, Invariance, and Spatial Variation in Intergenerational Mobility American Economic Review: Papers & Proceedings 2016, 106(5): 400 404 http://dx.doi.org/10.1257/aer.p20161082 Fixed Effects, Invariance, and Spatial Variation in Intergenerational Mobility By Gary Chamberlain*

More information

Bayesian Linear Models

Bayesian Linear Models Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Department of Forestry & Department of Geography, Michigan State University, Lansing Michigan, U.S.A. 2 Biostatistics, School of Public

More information

e-companion ONLY AVAILABLE IN ELECTRONIC FORM Electronic Companion Stochastic Kriging for Simulation Metamodeling

e-companion ONLY AVAILABLE IN ELECTRONIC FORM Electronic Companion Stochastic Kriging for Simulation Metamodeling OPERATIONS RESEARCH doi 10.187/opre.1090.0754ec e-companion ONLY AVAILABLE IN ELECTRONIC FORM informs 009 INFORMS Electronic Companion Stochastic Kriging for Simulation Metamodeling by Bruce Ankenman,

More information

Slice Sampling with Adaptive Multivariate Steps: The Shrinking-Rank Method

Slice Sampling with Adaptive Multivariate Steps: The Shrinking-Rank Method Slice Sampling with Adaptive Multivariate Steps: The Shrinking-Rank Method Madeleine B. Thompson Radford M. Neal Abstract The shrinking rank method is a variation of slice sampling that is efficient at

More information

GARCH Models Estimation and Inference. Eduardo Rossi University of Pavia

GARCH Models Estimation and Inference. Eduardo Rossi University of Pavia GARCH Models Estimation and Inference Eduardo Rossi University of Pavia Likelihood function The procedure most often used in estimating θ 0 in ARCH models involves the maximization of a likelihood function

More information

Lecture 11: Regression Methods I (Linear Regression)

Lecture 11: Regression Methods I (Linear Regression) Lecture 11: Regression Methods I (Linear Regression) Fall, 2017 1 / 40 Outline Linear Model Introduction 1 Regression: Supervised Learning with Continuous Responses 2 Linear Models and Multiple Linear

More information

MIXED MODELS THE GENERAL MIXED MODEL

MIXED MODELS THE GENERAL MIXED MODEL MIXED MODELS This chapter introduces best linear unbiased prediction (BLUP), a general method for predicting random effects, while Chapter 27 is concerned with the estimation of variances by restricted

More information

FE670 Algorithmic Trading Strategies. Stevens Institute of Technology

FE670 Algorithmic Trading Strategies. Stevens Institute of Technology FE670 Algorithmic Trading Strategies Lecture 3. Factor Models and Their Estimation Steve Yang Stevens Institute of Technology 09/12/2012 Outline 1 The Notion of Factors 2 Factor Analysis via Maximum Likelihood

More information

Vector autoregressions, VAR

Vector autoregressions, VAR 1 / 45 Vector autoregressions, VAR Chapter 2 Financial Econometrics Michael Hauser WS17/18 2 / 45 Content Cross-correlations VAR model in standard/reduced form Properties of VAR(1), VAR(p) Structural VAR,

More information

Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements

Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements Jeffrey N. Rouder Francis Tuerlinckx Paul L. Speckman Jun Lu & Pablo Gomez May 4 008 1 The Weibull regression model

More information

STAT 100C: Linear models

STAT 100C: Linear models STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix

More information

Variational Principal Components

Variational Principal Components Variational Principal Components Christopher M. Bishop Microsoft Research 7 J. J. Thomson Avenue, Cambridge, CB3 0FB, U.K. cmbishop@microsoft.com http://research.microsoft.com/ cmbishop In Proceedings

More information

Modeling Longitudinal Count Data with Excess Zeros and Time-Dependent Covariates: Application to Drug Use

Modeling Longitudinal Count Data with Excess Zeros and Time-Dependent Covariates: Application to Drug Use Modeling Longitudinal Count Data with Excess Zeros and : Application to Drug Use University of Northern Colorado November 17, 2014 Presentation Outline I and Data Issues II Correlated Count Regression

More information

Time-Varying Parameters

Time-Varying Parameters Kalman Filter and state-space models: time-varying parameter models; models with unobservable variables; basic tool: Kalman filter; implementation is task-specific. y t = x t β t + e t (1) β t = µ + Fβ

More information

On Reparametrization and the Gibbs Sampler

On Reparametrization and the Gibbs Sampler On Reparametrization and the Gibbs Sampler Jorge Carlos Román Department of Mathematics Vanderbilt University James P. Hobert Department of Statistics University of Florida March 2014 Brett Presnell Department

More information