Wavelet-Based Functional Mixed Models

Size: px
Start display at page:

Download "Wavelet-Based Functional Mixed Models"

Transcription

1 3//008 Wavelet-Based Functional Mied Models Jeffrey S. Morris Department of Biostatistics The University of Teas MD Anderson Cancer Center Rice University /7/008

2 3//008 Linear Mied Models: General Linear Model Suppose we have N data points y i, i,, N that we would like to eplain by a set of p eplanatory variables ij ; i,, N; j,, p. The ij may be continuous variables, or discrete (dummy) variables. A General Linear Model is given by: p y b + i ij j i j In many applications, we assume the errors are independent and identically distributed, so e i ~N(0,σ e ) e

3 3// Linear Mied Models: General Linear Model This form is very general, with different specifications of ij corresponding to various classical statistical methods Continuous: Linear regression Dummy variables ( groups): t-test Dummy variables (> groups): ANOVA Combination of continuous and dummy variables: Analysis of Covariance Note: b j are partial effects; i.e. describe the effect of predictor ij after adjusting for all other effects in the model

4 Linear Mied Models: General Linear Model This model can be written in matri form: y Xb + e e ~ N (0, σ e I N ) Where y is an N response vector, X is an N p design matri, and b is a p vector of fied effects Typical goal: estimate and perform inference on fied effect vector b The form of the design matri, X, indicates the type of modeling being done 3//

5 3// Linear Mied Models: Eamples Simple linear regression To perform a simple linear regression of Y on, specify: 0, b b b X N M M b 0 Intercept b slope

6 3// Linear Mied Models: Eamples Polynomial Regression: To perform a polynomial regression of Y on, specify: p p N N N p p b b b b b X M L M O M M M L L 0, ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( ) ( b 0 Intercept, b j polynomial level j coefficient Note: Centering reduces collinearity

7 3// Linear Mied Models: Eamples Multiple linear regression To perform multiple regression of y on {,,, p }: p pn N N p p b b b b X M L M O M M M L L 0, b 0 Intercept b, b,, b p partial slopes for factors,, p

8 3// Linear Mied Models: Eamples T-tests: Define dummy variable (categorical factor) if group 0, - if group 0, b b b X N M M b 0 Intercept (overall mean of y) b Main effect for group (0 if no group effect) Note: F-test for H 0 :b 0 is equivalent to t What if there are > groups we want to compare? - way ANOVA

9 3// Linear Mied Models: Eamples ANOVA (Cell Means model): (g groups) Let ij if subject i in group j, and no intercept g Ng N N g g b b b b X M L M O M M L L, b j : overall mean for group j b j b j implies groups j and j share common mean Contrasts can be used to set up comparisons of interest

10 3// Linear Mied Models: Eamples A p vector c{c, c,,c g } specifies a contrast g j j j j c b c g j 0 if, L Specifies a linear combination of fied effects that indicate a certain comparison of interest. E.g. Can be used with cell means model to test difference between groups j and j : c j, c j - Given set of l contrasts of interest, can construct l g contrast matri C Main Effect(ANOVA): L M M O M M L L C

11 3//008 Linear Mied Models: Eamples Analysis of Covariance (ANOCOVA) Cell Means model: (g groups, p covariates) Let ij if subject i in group j, j,, g Let ij covariate (j-g+) for subject i, jg+, g+p p g g g p g N g N Ng N p g g g p g g g b b b b b X M M L L M O M M O M L L L L ) ( ) ( ) ( ) ( ) ( ) (, b j, j,, g: separate intercepts by group j b j, jg+,, g+p: partial slopes for p covariates

12 3//008 Linear Mied Models: Eamples What if we have multiple categorical factors we would like to simultaneously consider? -way ANOVA: Includes main effects for factors and, plus their -way interaction Interaction: The effect of factor is different for different levels of factor (and vice versa) Again, different parameterizations can be used Overparameterized model: (messy!) Main effect/interaction model Cell means model, with contrasts Can also interact continuous factors with categorical factors e.g. growth curves with intercepts and slopes varying by group

13 3// Linear Mied Models: Eamples Cell Means model: (gg g groups) Let ij if subject i in group j, and no intercept g Ng N N g g b b b b X M L M O M M L L, b j : overall mean for group j b j b j implies groups j and j share common mean Contrasts can be used to set up comparisons of interest, e.g. main effect and interactions

14 3// Linear Mied Models: Eamples C [ ] C Suppose factor has levels, and factor has 3 levels, with b[b, b, b 3, b, b, b 3 ] C L C b: Main effect for factor L C b: Main effect for factor L C b: -way interaction effect

15 Linear Mied Models: Eamples General Linear Model: y Xb + e e ~ N (0, σ e I N ) What if errors are not iid? Subsampling (multiple randomly sampled observations per subject) Longitudinal Data (multiple observations per subject over time, so observations are ordered) Solutions: Assume parametric error structure on e~n(0,r) Include random effect terms in the model. 3//

16 3// Linear Mied Models: Eamples Simple Random Effects model: Suppose we observe r repeated measurements of response y for each of n subjects (total of Nrn obs.) Let y[y, y,, y r, y,, y r,, y nr ] Let X be a design matri for the fied effects b Define an N n random effects design matri Z to have elements z ij if obs i from subject j, 0 otherwise. A random effects model can be written as: y Xb + Zu + e u e ~ ~ N(0, σ N(0, σ u e I I n r ) )

17 3// Linear Mied Models: Eamples y Xb + Zu + e u e ~ ~ N N (0, σ (0, σ u e I I n r ) ) u k random effect for subject k: E(y subject k)xb+u k σ u : describes subject-to-subject variability σ e : describes variability between observations within subject If we integrate the random effects u out of the model, thus marginalizing over u, y Xb + e e ~ N (0, Σ) Where Σσ u ZZ + σ e I N has block compound symmetry structure

18 3// Linear Mied Models: Eamples Block Compound Symmetry blocks blocks n r r e e e e e e e e e e e e e e e e e e e e e e e e e e e Σ L M O L M M L σ ρσ ρσ ρσ σ ρσ ρσ ρσ σ σ ρσ ρσ ρσ σ ρσ ρσ ρσ σ σ ρσ ρσ ρσ σ ρσ ρσ ρσ σ n diagonal blocks of size r r, each having compound symmetry Note: σ e σ e + σ u is the total variability and ρ σ u /σ e + σ u is the intraclass correlation

19 3// Linear Mied Models: Eamples General Linear Mied Model { { } { } { + + N m m N p p N N e u Z b X y } { ) 0, ( ~ ) 0, ( ~ N N m m R N e P N u Var(y)ΣZPZ +R Most common applications use independent random effects and residual errors, Pσ u I m and Rσ e I N Other possible covariance structures can also be used for P and R, as needed best linear unbiased predictor (BLUP) ˆ) ( ' ˆ generalized least squares estimator ' ) ' ( ˆ Σ Σ Σ Xb y PZ u y X X X b

20 3// Linear Mied Models: Eamples Fitting Linear Mied Models Standard software packages are available for fitting linear mied models PROC MIXED (SAS) lme/nlme (S+/R) Based on EM or Newton-Raphson algorithms MLE for fied effect functions (GLS) REML for covariance parameters Straightforward to perform frequentist inference on fied effect functions, or linear combinations (contrasts) F tests Trickier to test whether VC0 but recent work in this area Can also be fit using Bayesian hierarchical model If standard improper priors (uninformative) are specified for fied effects b and covariance parameters, it can be shown that the posterior mode Bayesian estimators are equivalent to MLE frequentist estimators, and the confidence/credible bands are similar

21 Functional Data Analysis Functional Data Analysis (Ramsay&Silverman 997) Methods for data where ideal units of observation are curves Sometimes distinguish between methods for sparsely sampled functional data (few obs/curve, longitudinal) and functions sampled on fine grid (many obs/curve, functional) Our main concern here is on functional data on fine grid Challenge of Functional Data Analysis: Must simultaneously consider regularization and replication Regularization: take advantage of functional structure to borrow strength from adjacent observations within curve Replication: combining information across sample curves to make inferences on population from which they came 3//008

22 3//008 Functional Mied Model (FMM) Idea: Relate functional response to set of scalar predictors through functional coefficients, while adjusting for possible correlation between functions induced by design. Suppose we observe a sample of N curves, Y i (t), i,, N, on a closed interval T Y { i ( t) X ij B j ( t) + Zik U k ( t) + E ( t) E { { i 3 response functions p j fied effect function m k functions random effect U functions residual error k i ( t) ~ GP(0, Q) ( t) ~ GP(0, S) B j (t) summarizes partial effect of X j on Y(t) Q(t,t ) and S(t,t ) are covariance surfaces on T T describing the form of the function-function deviations

23 Discrete Version of FMM Suppose each observed curve is sampled on a common equally-spaced grid of length T. Y{ N N p } X { B + } Z U{ + { E T p T m T N T Rows of B contain fied effect functions on grid Q and S are within-curve covariance matrices (T T) approimating surfaces on the grid N For irregular functional data, Q and S typically contain many nonstationarities, yet their dimension is too high to leave unstructured 3// m U E k i ~ ~ MVN (0, Q) MVN (0, S)

24 3// FDA: Functional Mied Models Growth curve eample: Goal: Assess effect of gene A on tumor growth N0 mice, ½ from knockout and ½ from wild type groups Xenograft model: microscopic tumor implanted Tumor measurements at T6 time points t{0,3,6,9,,5} Y i (t)log(tumor size) profile for animal i Y i (t) X ijb j(t) + j Cell Means Model Used X i if animal i from knockout group, 0 otherwise X i if animal i from wild type group, 0 otherwise Test: H 0 :B (t)b (t) E (t) i E i { 0, S( t, t )} ( t) ~ N

25 3// FDA: Functional Mied Models Growth curves are (log) linear, so we can represent functions with parametric form (lines) No intercepts (tumor size0 at time0 by design) B j (t)b j t; E i (t)u i t+e i ; u i ~N(0,σ u ); e i ~N(0,σ e I 6 ) S t t σ u +σ e I 6, e.g. if σ u 0., σ e : S ρ S

26 3// FDA: Functional Mied Models Mied model in matri form: Let y{y k, k,,nt}, X k if obs k from knockout animal, X k if obs k from wild type and Z kl if obs k from animal l y Xb + Zu + e u ( ) ( 0, I e ~ N 0, I ) ~ N σ σ u N e NT X X X X M NT t t t NT X X X M NT t t t NT b C b b [ ] u u u M u m

27 3// FDA: Functional Mied Models What if there is no acceptable parametric representation for the observed functions? Nonparametric representations for functions: generalizations of nonparametric regression to multiple function case. Much recent work in this area. Limitations of previous work: Most focus on smoothing methods like kernels and splines Much intended for functional data on sparse, not fine, grids Most models not fully general FMM (arbitrary X matri, nested random effects, general representations for Q or S) Many focus on estimation, not inference Most of this work works with models that are special cases of the functional mied model.

28 Applications: SELDI Organ-by-Cell-Line 6 mice had of cancer cell lines injected into of organs (lung or brain) Cell lines: A375P: human melanoma, low metastatic potential PC3MM: human prostate, highly metastatic Blood serum etracted and placed on SELDI chip Run at different laser intensities (low/ high) Total of 3 spectra (observed functions), per mouse Observations on equally-spaced grid of //

29 Applications: SELDI Organ-by-Cell-Line Goal: Find proteins differentially epressed by: Host organ site (lung/brain) Donor cell line (A375P/PC3MM) Organ-by-cell line interaction Combine information across laser intensities: Requires us to include in modeling: Functional laser intensity effect Random effect functions to account for correlation between spectra from same mouse 3//

30 3// Applications: SELDI Organ-by-Cell-Line MALDI-TOF Spectrum: observed function g(t) intensity of spectrum at m/z value t Intensity at peak (roughly) estimates the abundance of some protein with molecular weight of t Daltons

31 3// FDA: Functional Mied Model Let Y i (t) be the SELDI spectrum i { Y ( t) } log i B0 ( t) + XijBj( t) + ZikUk( t) + Ei ( t) X i for lung, - brain. X i for A375P, - for PC3MM X i3 X X 4 j 6 k X i4 for low laser intensity, - high. B 0 (t) overall mean spectrum B (t) organ main effect function B (t) cell-line main effect B 4 (t) laser intensity effect function B 3 (t) org cell-line int function Z ik if spectrum i is from mouse k (k,, 6) U k (t) is random effect function for mouse k.

32 3// Functional Mied Model (FMM) Idea: Relate functional response to set of scalar predictors through functional coefficients, while adjusting for possible correlation between functions induced by design. Suppose we observe a sample of N curves, Y i (t), i,, N, on a closed interval T Y { i ( t) X ij B j ( t) + Zik U k ( t) + E ( t) E { { i 3 response functions p j fied effect function m k functions random effect U functions residual error k i ( t) ~ GP(0, Q) ( t) ~ GP(0, S) B j (t) summarizes partial effect of X j on Y(t) Q(t,t ) and S(t,t ) are covariance surfaces on T T describing the form of the function-function deviations

33 3// Wavelet-Based Functional Mied Models We would like to fit functional mied models to data like in our motivating eamples (high dim., spiky) Need to find way to build fleibility into Q and S, yet be parsimonious enough to fit to large data sets. Need to find a way to accomplish adaptive regularization for B j (t) and U j (t) Need to find way to compute this model that is computationally efficient enough to apply to etremely large data sets (00 s-000 s of curves on grids of 0,000 or more) Want to be able to perform inference, not just estimation Key: Use of wavelet space instead of data space to do our modeling.

34 Wavelet-Based Functional Mied Models Idea: Represent functions by their wavelet epansion Basis function approach: y(t)σd jk ψ jk (t) Benefits of Using Wavelet Bases. Compact support allows efficient representations of local features. Whitening property allows parsimonious yet fleible representations of Q and S 3. Decomposes function in both frequency and time domains Enables mechanism for adaptive regularization of functions 4. Orthonormal transformation has linear representation, and special structure allows fast calculation of coefficients. 3//

35 3// WFMM: General Approach. Project observed functions Y into wavelet space.. Fit FMM in wavelet space. (Use MCMC to get posterior samples) 3. Project wavelet-space estimates (posterior samples) back to data space.

36 3// WFMM: General Approach. Project observed functions Y into wavelet space.. Fit FMM in wavelet space (Use MCMC to get posterior samples) 3. Project wavelet-space estimates (posterior samples) back to data space.

37 3// WFMM: Projecting to Wavelet Space. Project observed functions Y to wavelet space After choice of () wavelet basis, () boundary correction method, and (3) number of decomposition levels J, apply DWT to rows of Y to get wavelet coefficients corresponding to each observed function { D Y{ W{ ' N T N T T T Projects the observed curves into the space spanned by the wavelet bases (rotate your coordinate aes according to wavelet bases).

38 3// WFMM: Projecting to Wavelet Space The matri D that results contains the observed wavelet coefficients for our data set { D N T K coeff K coeff K coeff K J } } } } c, d, d, L d, J c, d, d, L d, J M M M O M cn, d N, d N, L d N, J coeff With T J, Haar wavelets, and full decomposition, we have K j j-. Otherwise not quite true because of adjustments for T J and boundary correction from other wavelet bases with wider support

39 3// WFMM: General Approach. Project observed functions Y into wavelet space.. Fit FMM in wavelet space (Use MCMC to get posterior samples) 3. Project wavelet-space estimates (posterior samples) back to data space.

40 3// WFMM: Wavelet Space Model Y{ } X N p { B + N } Z m U{ + { E N T p T m T N T Wavelet Representations YDW BBW UUW EEW

41 3// WFMM: Wavelet Space Model { DW } X N p B{ W + N } Z m U3 W + 3 E W N T p T m T N T Wavelet Representations YDW BBW UUW EEW

42 3// WFMM: Wavelet Space Model { DW W N T ' } X N p B{ W W p T ' + N } Z m U3 W W m T ' + 3 E W W N T ' Wavelet Representations YDW BBW UUW EEW WW I

43 3// WFMM: Wavelet Space Model { D } X N p { B + N } Z m U{ + { E N T p T m T N T Wavelet Representations YW D BW B UW U EW E WW I

44 3// Wavelet Space FMM D : empirical wavelet coefficients for observed curves Row i contains wavelet coefficients for observed curve i Each column double-indeed by wavelet scale j and location k { D } X N p { B + N } Z m U{ + { E N T p T m T N T U E k i ~ MVN(0, Q ~ MVN(0, S ) ) B BW & U UW : Rows contain wavelet coefficients for the fied and random effect functions, respectively E EW is the matri of wavelet-space residuals

45 3// WFMM: Covariance Assumptions We assume the between-wavelet covariance matrices Q and S are diagonal (Qdiag{q jk },Sdiag{s jk }). Wavelet coefficients within given function modeled as independent Heuristically justified by whitening property of DWT Common working assumption in wavelet regression settings Is parsimonious in wavelet space (T parameters), yet leads to fleible class of nonstationary covariance structures in data space Key: We allow variances to vary by scale j and location k Other alternatives possible (parent-sib correlation, block thresholding), but add more computational compleity. We have found independence to be very fleible: limits form of covariance matrices, but enough parameters and degrees of freedom to capture key nonstationary features of these matrices that are typical of our data

46 WFMM: Covariance Assumptions Whitening: autocorrelation of wavelet coefficients tends to die away rapidly across k; d jk approimately uncorrelated (Johnstone and Silverman 997) 3//

47 WFMM: Covariance Assumptions Whitening: autocorrelation of wavelet coefficients tends to die away rapidly across k; d jk approimately uncorrelated (Johnstone and Silverman 997) 3//

48 3// WFMM: Covariance Assumptions True mean: line plus peak True variance: increasing in t, with etra var at peak True autocorrelation: Strong autocorrelation (0.9) at left, weak autocorrelation (0.) right, etra at peak

49 3// WFMM: Covariance Assumptions Independence in wavelet space accommodates varying degrees of autocorrelation in data space Allowing variance components to vary across scale j and location k accommodates nonstationarities

50 3// WFMM: Covariance Assumptions Most wavelet regression methods (Fan and Lin 998, Morris, et al. 003, Abramovich and Angelini 007, Antoniadis and Sapatinas 007) only inde variances by scale j, but not location k. Not fleible enough to capture nonstatioinary covariance features Unnecessary restriction in multiple function case, since replicate functions allow estimation of separate VC for each (j,k)

51 3// WFMM: Wavelet-Space Model Key benefit of independence assumption: With Q,S diagonal, the columns of the WFMM are now unlinked, and can be modeled separately. A set of T regular linear mied model instead of one HUGE functional mied model. Key to computational feasibility (run time AND memory management) d { jk N p } X B { jk + N } Z m u + e { jk { jk N p m N u jk ~ N (0, q jk ) e jk ~ N (0, s jk )

52 3// WFMM: Prior Assumptions Another benefit of wavelet-space modeling is that it provides us a natural mechanism for inducing adaptive regularization for our estimates of the B a (t) To induce nonlinear shrinkage, we assume a spikeslab prior on B ajk : B ajk γ ajk aj ajk N( 0, τ ) + ( γ ) δ Bernoulli( π ajk aj γ Nonlinearly shrinks B ajk towards 0, leading to adaptively regularized estimates of B a (t). τ aj & π aj are regularization parameters ) 0

53 3// WFMM: Prior Assumptions How to choose smoothing parameters? Not intuitive. Antoniadis, Sapatinas, and Silverman (998) suggest a parameterization in which π j decrease eponentially in j, and relate them theoretically to the underlying Besov space Clyde and George (000) estimated these from data using local MLE, in empirical Bayes fashion. We follow. Involves iterative (EM) algorithm to estimate the proportion of large wavelet coefficients at each level j (π ij ), and their associated variances (τ ij ), with lower bound on τ ij to avoid ecessive linear shrinkage Similar in principal to idea of estimating smoothing parameters from data in nonparametric regression

54 WFMM: Prior Assumptions To complete the specification of the Bayesian model, we need priors on the variance components. When PRI, we only need to worry about q jk and s jk We use diffuse conjugate Inverse Gamma priors Center on starting value estimates (MLE s), and set scale parameter such that prior has /000 information of data These empirical Bayes priors are by default automatically calculated by the code, which is important since it would not be feasible to set these by hand since there are (H+)T variance components. 3//

55 3// WFMM: Model Fitting MCMC to obtain posterior samples of model quantities Work with marginal likelihood; U integ. out; Let Ω be a vector containing ALL covariance parameters (i.e. P,R,Q,S). MCMC Steps. Sample from f(b D,Ω): Miture of Normals and point masses at 0 for each i,j,k.. Sample from f(ω D,B): Metropolis-Hastings steps for each j,k 3. If desired, sample from f(u D,B,Ω): Multivariate Normals

56 3// B Bˆ ajk LS aj α ijk N { X WFMM: Model Fitting Updating fied effects, B : Gibbs step (conjugate spike-slab miture) for each a,, p, j,, J, k,, Kj { } Bˆ ˆ ajk, MLELSaj, Vajk LSaj + ( αijk ) I0 ' Σ T aj Taj Linear Shrinkage X } X α ijk ' Σ ajk, MLE a jk a a jk jk ( a) ( a) jk, MLE ajk ajk, MLE a jk a ( d V Bˆ Var( ) ( X ' Σ X ) Σ ZPZ' q + jk X Bˆ { } d 4 ajk 44 ajk 3 Pr γ 4 aj jk Nonlinear Shrinkage aj T τ / V ajk Rs ) jk

57 3// WFMM: Model Fitting α ajk O Pr ajk ajk O + ( ) ajk γ d, O Posterior Odds ajk ajk π ζ ajk Taj ep π aj Taj + 3 aj ( O + ) ajk T { aj Posterior Odds Prior Odds Bayes Factor ζ ajk ˆ ajk, MLE B / V ajk

58 3// Wavelets: Wavelet Regression σ, τ 0, π0.0 ˆ θ ˆ θ ˆ θ jk,shrinkage jk,linear jk,nonlinear d d jk d jk SF jk P j P jk jk SF j

59 3// Wavelets: Wavelet Regression σ, τ 5, π0.0 ˆ θ ˆ θ ˆ θ jk,shrinkage jk,linear jk,nonlinear d d jk d jk SF jk P j P jk jk SF j

60 3// Wavelets: Wavelet Regression σ, τ, π0.0 ˆ θ ˆ θ ˆ θ jk,shrinkage jk,linear jk,nonlinear d d jk d jk SF jk P j P jk jk SF j

61 WFMM: Model Fitting Updating of variance components Ω: Could use Gibbs step for q jk and s jk : f(ω D,B,U) Problem: As q jk 0, U jk 0, get stuck in local mode ( black hole phenomenon ) Solution: Marginalize model by integrating out U, then update f(ω D,B), then f(u D,B,Ω) Equivalent to updating (Ω,U ) simultaneously, since f(u,ω D,B) f(ω D,B) f(u D,B,Ω) Another advantage: If inference on random effects not needed, then there is no need to update them! Disadvantage: Can t update q jk and s jk with simple Gibbs steps, since f(ω D,B) is not a known distributional form ; instead must use Metropolis-Hastings step 3//

62 WFMM: Model Fitting Metropolis-Hastings for q jk, s jk Use random-walk proposal: Proposed value for new MCMC iteration q jk () is centered at value at previous iteration q jk (0), as N(q jk (0),propvar jk ) Key is appropriate choice of propvar jk ; if too big, sampler won t move, if too small, will not eplore posterior (get stuck in local modes) Typically found by trial-and-error; not feasible here! We estimate propvar jk as.5var(q jk,mle ), which should be slightly more than variance of the true complete conditional distribution; works well 3//

63 3// WFMM: Model Fitting Updating of random effects, U : Gibbs step, Normal f ( u d, B, Ω) jk jk jk Normal( M jk, V jk ) V jk { ( ) } Ψ + q P jk jk Ψ jk { '( ) } Z s R Z jk M jk V jkψ jk uˆ jk, MLE 3 Linear Shrinkage ˆ jk, MLE ΨjkZ'( s jk R) ( d jk XB jk u ) When PRI, this step is very quick and trivial

64 3// WFMM: General Approach. Project observed functions Y into wavelet space.. Fit FMM in wavelet space (Use MCMC to get posterior samples) 3. Project wavelet-space estimates (posterior samples) back to data space.

65 3// WFMM: Back to Data Space 3. Project wavelet-space estimates (posterior samples) back to data space. Apply IDWT to posterior samples of B to get posterior samples of fied effect functions B j (t) for i,, p, on grid t. BBW Posterior samples of U k (t), Q, and S are also available, if desired.

66 3// WFMM: Back to Data Space For each fied effect function a,, p, and MCMC sample g,, G, construct: B [ ] ( g ) ( g ) ( ) B B L B ( g ) g a a a ajk Then apply the pyramid-based IDWT to this vector to obtain a posterior sample of B (g) a (t) on the grid t, since B a B a W In the same way, posterior samples of U(t) can be obtained by applying the IDWT to U (g), if desired. You can also compute posterior samples for any contrast involving the fied effect functions, CB J

67 3// WFMM: Back to Data Space We can also obtain posterior samples of Q or S from the wavelet space versions Q (g) diag{q jk (g) } and S (g) diag{s jk (g) } using the results of Vannucci and Corradi (999), based on the fact that QWQ W Can be done efficiently, in O(T ), by applying the -d IDWT to Q (g) and S (g) This gives us estimates of the within-curve covariance surfaces, The diagonal of these give us estimates of the variance function

68 3// WFMM: Inference Given posterior samples of all model quantities, we can perform desired Bayesian inference or prediction:. Pointwise posterior credible intervals for functional effects or contrasts. Posterior probabilities for pointwise hypotheses of interest 3. Bayes Factors for functional testing 4. Can account for multiple testing in identifying significant regions of curves by controlling the epected Bayesian FDR 5. Can compute posterior predictive distributions, which can be used for model-checking or classification.

69 3// Pointwise Inference: WFMM: Inference Given posterior samples of B a (t), can construct posterior mean (shrinkage) estimates for B a (t), or any contrast of fied effect functions CB Can compute quantiles of posterior distribution for B a (t) for each t, which can be used for pointwise credible intervals. Pr{Q j,α/ (t)<b a (t)< Q j,-α/ (t) D}-α Can compute probability of having a certain minimum effect size: p a (t)pr{ B a (t) >δ D}, e.g. log: δ -fold Takes both Statistical and Clinical significance into account

70 3// WFMM: Properties Draws of spectra from posterior predictive distribution yield data that looks like real MALDI data (3 rd column), indicating reasonable model fit.

71 3// WFMM: Properties Posterior samples/estimates of fied effect functions B i (t) adaptively regularized as a result of shrinkage prior applied to wavelet coefficients..5 (a) School E log(met) Able to preserve dominant spikes in mean curves, if present am pm 3pm 5pm 7pm Tim e of Day

72 WFMM: Properties 3//

73 WFMM: Properties Posterior samples/estimates of random effect functions U j (t) also appear to be adaptively regularized Able to preserve spikes in random effect functions, as well Important for estimation of random effect functions AND for valid inference on fied effect functions. 3//

74 3// WFMM: Properties How can the random effects be adaptively regularized when they have a Gaussian prior (with linear shrink)? U jk ~N(0,q jk ) However, note that each wavelet coefficient has its own variance component q jk, which determines the amount of shrinkage from u jk,mle towards 0 (smaller q relative to s, more shrinkage towards 0) Since (j,k) corresponding to strong signals in the random effect functions will tend to be larger, q jk will also tend to be larger, leading to less shrinkage. Unimportant (j,k) will tend to have small q jk, and thus lead to more shrinkage Although linear for each wavelet coefficient, the shrinkage is nonlinear when viewed for the entire set of wavelet coefficients together. This same dynamic allows very smooth random effect estimates when the data supports this notion.

75 3// WFMM: Properties While adaptive to irregularity, this framework can also yield relatively smooth effect functions when the data supports smooth representations.

76 WFMM: Inference Multiple testing problem is inherent in pointwise inference: T positions, T tests One Solution: perform functional test of H 0 :B a (t)0 Popular approach for some researchers (FANOVA) Not as interesting to me: key is for what t we have B a (t) differ from zero, and by how much. We can do functional testing in WFMM using Bayes Factors Another Solution: Apply some sort of multiple testing adjustment to the inference. Here we will look at applying the False Discovery Rate (FDR) ideas popular in microarrays to this pointwise functional setting. 3//

77 3// Adjusting for Multiple testing Classic approach: Bonferroni T tests, use α/t as sig. level WFMM: Inference Controls eperiment-wise error rate H 0 Pr(at least one false positive)< α Too conservative for some fields Decision (e.g. microarrays), so alternative criterion devised H False Discovery Rate (FDR): Benjamini and Hochberg (995) Controls proportion of discoveries that are false to be no more than α A C Truth H 0 H FDR C B D C + D

78 WFMM: Inference Many procedures have been developed for controlling FDR Benjamini and Hochberg (995), Yekutieli and Benjamin (999), Benjamini and Liu (999), Storey (00), Storey (003), Genovese and Wasserman (00), Ishwaran and Rao (003), Pounds and Morris (003), Efron (004), Newton (004), Pounds and Cheng (004) Most methods follow one of three approaches:. Fit miture model on tests stats (known distn under H 0 ). Fit miture model on p-values {U(0,) under H 0 } 3. Fit Bayesian miture model, with point mass prior on H 0 Methods can either set limit on global FDR (α), or on local FDR for each statistic i (Storey 003) q i min Pr(H 0 true for gene i Gene i in RR) Note: p-value is Pr(Gene i in RR H 0 true for statistic i) 3//

79 3// WFMM: Inference Global: αc/(c+d) Local: q(pv)c/(c+d) D d B C c A pv

80 3// WFMM: Inference Bayesian FDR from posterior probabilities Posterior probs of effect sizes: p i Pr{ B i <δ Y}, i,, T If we define a discovery to be effect size at least δ (H 0i : B i < δ), then p i is a Bayesian local FDR To control the global epected Bayesian FDR. Sort in ascending order of p i {p (i), i,, T}. Identify cutpoint ϕ α on posterior probabilities that controls epected Bayesian FDR to be α i ϕ α p (λ), where λ mai: pi i α i 3. Flag the set of statistics {i: p i ϕ α } as significant (According to model, we epect only α to be false pos.)

81 3// WFMM: Inference Notes This approach takes both statistical and practical significance into account when declaring differences significant. Given choice of cutpoint on posterior probabilities to flag significant regions, we can also estimate false negative rate (FNR), sensitivity, and specificity for declaring flagged regions significantly different.

82 WFMM: Inference FDR for pointwise functional inference: Substituting Lebesgue measure for counting measures, we can apply FDR ideas to pointwise inference: From MCMC samples, we have the quantities: p al Pr{ B a (t l ) >δ Y} for l,, T Using H 0t : B a (t) <δ, p al is a Bayesian local FDR for fied effect function a at position t l within the curve. As described above, we can find a cutpoint ϕ α on the p al that controls the global average Bayesian FDR<α α > t : t flagged falsely Ba ( t) < δ t : t flagged as "significant" 3//

83 3// Applications: SELDI Organ-by-Cell-Line Let Y i (t) be the SELDI spectrum i { Y ( t) } 5 log i XijBj( t) + ZikUk ( t) + Ei ( t) j X i if lung/a375p, 0 ow X i for brain/a375p, 0 ow X i3 if lung/pcmm X i4 for brain/pcmm X i5 if high laser intensity, - if low 6 k B j (t) overall mean spectrum for treatment group j,,3,4 B 5 (t) laser intensity effect function Z ik if spectrum i is from mouse k (k,, 6) U k (t) is random effect function for mouse k.

84 Applications: SELDI Organ-by-Cell-Line 3//

85 Applications: SELDI Organ-by-Cell-Line 3//

86 Applications: SELDI Organ-by-Cell-Line 3//

87 3// Applications: SELDI Organ-by-Cell-Line Using α0.05, δ (-fold epression on log scale), we flag a number of spectral regions.

88 Applications: SELDI Organ-by-Cell-Line 3//

89 Applications: SELDI Organ-by-Cell-Line 3900 D (~00-fold) (CGRP-II): dilates blood vessels in brain 760 D (~5-fold) (neurogranin): active in synaptic modeling in brain (Not detected as peak) 3//

90 Applications: SELDI Organ-by-Cell-Line 3//

91 3// Applications: SELDI Organ-by-Cell-Line Inclusion of nonparametric functional laser intensity effect is able to adjust for systematic differences in the and y aes between laser intensity scans

92 3// Conclusion Presented unified modeling approach for FDA Adaptive enough to handle irregularities in both mean structures and random effects (covariances) Method based on mied models; is FLEXIBLE Accommodates a wide range of eperimental designs Addresses large number of research questions Posterior samples allow Bayesian inference and prediction Posterior credible intervals; pointwise or joint Predictive distributions for future sampled curves Predictive probabilities for classification of new curves Bayesian functional inference can be done via Bayes Factors Since a unified modeling approach is used, all sources of variability in the model propagated throughout inference.

93 3// Conclusion Approach is Bayesian. The only informative priors to elicit are regularization parameters, which can be estimated from data using empirical Bayes. Developed general-use code reasonably fast and straightforward to use minimum information to specify is Y, X, Z matrices. Can deal with missing data, i.e. partially observed functions Method has been generalized to higher dimensional functions, e.g. image data, space/time data The Gaussian/independence assumptions can be relaed to robustify modeling

94 3// Acknowledgements Work presented here is from 3 papers. Wavelet-Based Functional Mied Models (006) Jeffrey S. Morris and Raymond J. Carroll, JRSS-B, 68(): Using Wavelet-Based Functional Mied Models to Characterize Population Heterogeneity in Accelerometer Profiles: A Case Study (006) Jeffrey S. Morris, Cassandra Arroyo, Brent Coull, Louise Ryan, Richard Herrick, and Steve Gortmaker, JASA, 0(4): Bayesian Analysis of Mass Spectrometry Proteomics Data using Wavelet Based Functional Mied Models (007) Jeffrey S. Morris, Philip J. Brown, Richard Herrick, Keith A. Baggerly, and Kevin R. Coombes, Biometrics, online. Supported by NIH Grant R0 CA07304 Computer code/papers on web at

95 3// Acknowledgements Statistical Collaborators Raymond J. Carroll Philip J. Brown Louise M. Ryan Keith A. Baggerly Kevin R. Coombes Marina Vannucci Brent A. Coull Elizabeth J. Malloy Cassandra Arroyo Veera Baladandayuthapani Richard C. Herrick (computing) Biomedical Collaborators Joanne Lupton Meeyoung Hong Nancy Turner Steve Gortmaker Stanley Hamilton James Abbruzzesse Ryuji Kobayashi John Koomen Nancy Shih Howard Gutstein Josh Fidler Donghui Li

Bayesian Wavelet-Based Functional Mixed Models

Bayesian Wavelet-Based Functional Mixed Models Bayesian Wavelet-Based Functional Mixed Models Jeffrey S. Morris U M.D. Anderson Cancer Center Raymond J. Carroll exas A&M University Functional Data Analysis Functional Data: Ideal units of observation:

More information

Bayesian Analysis of Mass Spectrometry Data Using Wavelet-Based Functional Mixed Models

Bayesian Analysis of Mass Spectrometry Data Using Wavelet-Based Functional Mixed Models Bayesian Analysis of Mass Spectrometry Data Using Wavelet-Based Functional Mixed Models Jeffrey S. Morris UT MD Anderson Cancer Center joint work with Philip J. Brown, Kevin R. Coombes, Keith A. Baggerly

More information

Wavelet-Based Nonparametric Modeling of Hierarchical Functions in Colon Carcinogenesis

Wavelet-Based Nonparametric Modeling of Hierarchical Functions in Colon Carcinogenesis Wavelet-Based Nonparametric Modeling of Hierarchical Functions in Colon Carcinogenesis Jeffrey S. Morris University of Texas, MD Anderson Cancer Center Joint wor with Marina Vannucci, Philip J. Brown,

More information

Functional Data Analysis for Accelerometer Data

Functional Data Analysis for Accelerometer Data Functional Data Analysis for Accelerometer Data Jeffrey S. Morris University of Texas M.D. Anderson Cancer Center, Houston, TX Joint work with: Louise Ryan, Cassandra Arroyo, Brent Coull, and Steve Gortmaker

More information

Aspects of Feature Selection in Mass Spec Proteomic Functional Data

Aspects of Feature Selection in Mass Spec Proteomic Functional Data Aspects of Feature Selection in Mass Spec Proteomic Functional Data Phil Brown University of Kent Newton Institute 11th December 2006 Based on work with Jeff Morris, and others at MD Anderson, Houston

More information

Wavelet-based functional mixed models

Wavelet-based functional mixed models J. R. Statist. Soc. B (26) 68, Part 2, pp. 79 99 Wavelet-based functional mixed models Jeffrey S. Morris University of Texas MD Anderson Cancer Center, Houston, USA and Raymond J. Carroll Texas A&M University,

More information

Default Priors and Effcient Posterior Computation in Bayesian

Default Priors and Effcient Posterior Computation in Bayesian Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature

More information

Part 6: Multivariate Normal and Linear Models

Part 6: Multivariate Normal and Linear Models Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of

More information

Part 8: GLMs and Hierarchical LMs and GLMs

Part 8: GLMs and Hierarchical LMs and GLMs Part 8: GLMs and Hierarchical LMs and GLMs 1 Example: Song sparrow reproductive success Arcese et al., (1992) provide data on a sample from a population of 52 female song sparrows studied over the course

More information

Wavelet-Based Preprocessing Methods for Mass Spectrometry Data

Wavelet-Based Preprocessing Methods for Mass Spectrometry Data Wavelet-Based Preprocessing Methods for Mass Spectrometry Data Jeffrey S. Morris Department of Biostatistics and Applied Mathematics UT M.D. Anderson Cancer Center Overview Background and Motivation Preprocessing

More information

Stat 5101 Lecture Notes

Stat 5101 Lecture Notes Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random

More information

Bayesian Linear Regression

Bayesian Linear Regression Bayesian Linear Regression Sudipto Banerjee 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. September 15, 2010 1 Linear regression models: a Bayesian perspective

More information

A Practitioner s Guide to Generalized Linear Models

A Practitioner s Guide to Generalized Linear Models A Practitioners Guide to Generalized Linear Models Background The classical linear models and most of the minimum bias procedures are special cases of generalized linear models (GLMs). GLMs are more technically

More information

Partial factor modeling: predictor-dependent shrinkage for linear regression

Partial factor modeling: predictor-dependent shrinkage for linear regression modeling: predictor-dependent shrinkage for linear Richard Hahn, Carlos Carvalho and Sayan Mukherjee JASA 2013 Review by Esther Salazar Duke University December, 2013 Factor framework The factor framework

More information

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 Lecture 2: Linear Models Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector

More information

Bayes methods for categorical data. April 25, 2017

Bayes methods for categorical data. April 25, 2017 Bayes methods for categorical data April 25, 2017 Motivation for joint probability models Increasing interest in high-dimensional data in broad applications Focus may be on prediction, variable selection,

More information

Linear Mixed Models. One-way layout REML. Likelihood. Another perspective. Relationship to classical ideas. Drawbacks.

Linear Mixed Models. One-way layout REML. Likelihood. Another perspective. Relationship to classical ideas. Drawbacks. Linear Mixed Models One-way layout Y = Xβ + Zb + ɛ where X and Z are specified design matrices, β is a vector of fixed effect coefficients, b and ɛ are random, mean zero, Gaussian if needed. Usually think

More information

BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA

BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA Part 3: Functional Data & Wavelets Marina Vannucci Rice University, USA PASI-CIMAT 4/28-3/2 Marina Vannucci (Rice University,

More information

Controlling the False Discovery Rate: Understanding and Extending the Benjamini-Hochberg Method

Controlling the False Discovery Rate: Understanding and Extending the Benjamini-Hochberg Method Controlling the False Discovery Rate: Understanding and Extending the Benjamini-Hochberg Method Christopher R. Genovese Department of Statistics Carnegie Mellon University joint work with Larry Wasserman

More information

Biostatistics Workshop Longitudinal Data Analysis. Session 4 GARRETT FITZMAURICE

Biostatistics Workshop Longitudinal Data Analysis. Session 4 GARRETT FITZMAURICE Biostatistics Workshop 2008 Longitudinal Data Analysis Session 4 GARRETT FITZMAURICE Harvard University 1 LINEAR MIXED EFFECTS MODELS Motivating Example: Influence of Menarche on Changes in Body Fat Prospective

More information

Computation fundamentals of discrete GMRF representations of continuous domain spatial models

Computation fundamentals of discrete GMRF representations of continuous domain spatial models Computation fundamentals of discrete GMRF representations of continuous domain spatial models Finn Lindgren September 23 2015 v0.2.2 Abstract The fundamental formulas and algorithms for Bayesian spatial

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Multivariate Survival Analysis

Multivariate Survival Analysis Multivariate Survival Analysis Previously we have assumed that either (X i, δ i ) or (X i, δ i, Z i ), i = 1,..., n, are i.i.d.. This may not always be the case. Multivariate survival data can arise in

More information

Expression Data Exploration: Association, Patterns, Factors & Regression Modelling

Expression Data Exploration: Association, Patterns, Factors & Regression Modelling Expression Data Exploration: Association, Patterns, Factors & Regression Modelling Exploring gene expression data Scale factors, median chip correlation on gene subsets for crude data quality investigation

More information

Spatial smoothing using Gaussian processes

Spatial smoothing using Gaussian processes Spatial smoothing using Gaussian processes Chris Paciorek paciorek@hsph.harvard.edu August 5, 2004 1 OUTLINE Spatial smoothing and Gaussian processes Covariance modelling Nonstationary covariance modelling

More information

Nonparametric Bayes tensor factorizations for big data

Nonparametric Bayes tensor factorizations for big data Nonparametric Bayes tensor factorizations for big data David Dunson Department of Statistical Science, Duke University Funded from NIH R01-ES017240, R01-ES017436 & DARPA N66001-09-C-2082 Motivation Conditional

More information

BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA

BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA Intro: Course Outline and Brief Intro to Marina Vannucci Rice University, USA PASI-CIMAT 04/28-30/2010 Marina Vannucci

More information

Lecture 3: Pattern Classification. Pattern classification

Lecture 3: Pattern Classification. Pattern classification EE E68: Speech & Audio Processing & Recognition Lecture 3: Pattern Classification 3 4 5 The problem of classification Linear and nonlinear classifiers Probabilistic classification Gaussians, mitures and

More information

WU Weiterbildung. Linear Mixed Models

WU Weiterbildung. Linear Mixed Models Linear Mixed Effects Models WU Weiterbildung SLIDE 1 Outline 1 Estimation: ML vs. REML 2 Special Models On Two Levels Mixed ANOVA Or Random ANOVA Random Intercept Model Random Coefficients Model Intercept-and-Slopes-as-Outcomes

More information

Bayesian Multivariate Logistic Regression

Bayesian Multivariate Logistic Regression Bayesian Multivariate Logistic Regression Sean M. O Brien and David B. Dunson Biostatistics Branch National Institute of Environmental Health Sciences Research Triangle Park, NC 1 Goals Brief review of

More information

Robust Classification of Functional and Quantitative Image Data Using. Functional Mixed Models

Robust Classification of Functional and Quantitative Image Data Using. Functional Mixed Models Biometrics 00, 1?? July 2011 DOI: 10.1111/j.1541-0420.2005.00454.x Robust Classification of Functional and Quantitative Image Data Using Functional Mixed Models Hongxiao Zhu 1, Philip J. Brown 2, and Jeffrey

More information

Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood

Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood Jonathan Gruhl March 18, 2010 1 Introduction Researchers commonly apply item response theory (IRT) models to binary and ordinal

More information

University of Texas, MD Anderson Cancer Center

University of Texas, MD Anderson Cancer Center University of Texas, MD Anderson Cancer Center UT MD Anderson Cancer Center Department of Biostatistics Working Paper Series Year 2010 Paper 56 Automated Analysis of Quantitative Image Data Using Isomorphic

More information

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007 MIT OpenCourseWare http://ocw.mit.edu HST.582J / 6.555J / 16.456J Biomedical Signal and Image Processing Spring 2007 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 EPSY 905: Intro to Bayesian and MCMC Today s Class An

More information

A Large-Sample Approach to Controlling the False Discovery Rate

A Large-Sample Approach to Controlling the False Discovery Rate A Large-Sample Approach to Controlling the False Discovery Rate Christopher R. Genovese Department of Statistics Carnegie Mellon University Larry Wasserman Department of Statistics Carnegie Mellon University

More information

MACHINE LEARNING ADVANCED MACHINE LEARNING

MACHINE LEARNING ADVANCED MACHINE LEARNING MACHINE LEARNING ADVANCED MACHINE LEARNING Recap of Important Notions on Estimation of Probability Density Functions 2 2 MACHINE LEARNING Overview Definition pdf Definition joint, condition, marginal,

More information

Modelling geoadditive survival data

Modelling geoadditive survival data Modelling geoadditive survival data Thomas Kneib & Ludwig Fahrmeir Department of Statistics, Ludwig-Maximilians-University Munich 1. Leukemia survival data 2. Structured hazard regression 3. Mixed model

More information

Multivariate Bayes Wavelet Shrinkage and Applications

Multivariate Bayes Wavelet Shrinkage and Applications Journal of Applied Statistics Vol. 32, No. 5, 529 542, July 2005 Multivariate Bayes Wavelet Shrinkage and Applications GABRIEL HUERTA Department of Mathematics and Statistics, University of New Mexico

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Contents. Part I: Fundamentals of Bayesian Inference 1

Contents. Part I: Fundamentals of Bayesian Inference 1 Contents Preface xiii Part I: Fundamentals of Bayesian Inference 1 1 Probability and inference 3 1.1 The three steps of Bayesian data analysis 3 1.2 General notation for statistical inference 4 1.3 Bayesian

More information

A GENERAL DECISION THEORETIC FORMULATION OF PROCEDURES CONTROLLING FDR AND FNR FROM A BAYESIAN PERSPECTIVE

A GENERAL DECISION THEORETIC FORMULATION OF PROCEDURES CONTROLLING FDR AND FNR FROM A BAYESIAN PERSPECTIVE A GENERAL DECISION THEORETIC FORMULATION OF PROCEDURES CONTROLLING FDR AND FNR FROM A BAYESIAN PERSPECTIVE Sanat K. Sarkar 1, Tianhui Zhou and Debashis Ghosh Temple University, Wyeth Pharmaceuticals and

More information

Introduction to Within-Person Analysis and RM ANOVA

Introduction to Within-Person Analysis and RM ANOVA Introduction to Within-Person Analysis and RM ANOVA Today s Class: From between-person to within-person ANOVAs for longitudinal data Variance model comparisons using 2 LL CLP 944: Lecture 3 1 The Two Sides

More information

Hierarchical Modelling for Univariate Spatial Data

Hierarchical Modelling for Univariate Spatial Data Spatial omain Hierarchical Modelling for Univariate Spatial ata Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A.

More information

A. Motivation To motivate the analysis of variance framework, we consider the following example.

A. Motivation To motivate the analysis of variance framework, we consider the following example. 9.07 ntroduction to Statistics for Brain and Cognitive Sciences Emery N. Brown Lecture 14: Analysis of Variance. Objectives Understand analysis of variance as a special case of the linear model. Understand

More information

Bayesian Regression Linear and Logistic Regression

Bayesian Regression Linear and Logistic Regression When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we

More information

Bayesian non-parametric model to longitudinally predict churn

Bayesian non-parametric model to longitudinally predict churn Bayesian non-parametric model to longitudinally predict churn Bruno Scarpa Università di Padova Conference of European Statistics Stakeholders Methodologists, Producers and Users of European Statistics

More information

ebay/google short course: Problem set 2

ebay/google short course: Problem set 2 18 Jan 013 ebay/google short course: Problem set 1. (the Echange Parado) You are playing the following game against an opponent, with a referee also taking part. The referee has two envelopes (numbered

More information

State Space and Hidden Markov Models

State Space and Hidden Markov Models State Space and Hidden Markov Models Kunsch H.R. State Space and Hidden Markov Models. ETH- Zurich Zurich; Aliaksandr Hubin Oslo 2014 Contents 1. Introduction 2. Markov Chains 3. Hidden Markov and State

More information

Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling

Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling Jae-Kwang Kim 1 Iowa State University June 26, 2013 1 Joint work with Shu Yang Introduction 1 Introduction

More information

Part 7: Hierarchical Modeling

Part 7: Hierarchical Modeling Part 7: Hierarchical Modeling!1 Nested data It is common for data to be nested: i.e., observations on subjects are organized by a hierarchy Such data are often called hierarchical or multilevel For example,

More information

Simultaneous inference for multiple testing and clustering via a Dirichlet process mixture model

Simultaneous inference for multiple testing and clustering via a Dirichlet process mixture model Simultaneous inference for multiple testing and clustering via a Dirichlet process mixture model David B Dahl 1, Qianxing Mo 2 and Marina Vannucci 3 1 Texas A&M University, US 2 Memorial Sloan-Kettering

More information

Bayesian shrinkage approach in variable selection for mixed

Bayesian shrinkage approach in variable selection for mixed Bayesian shrinkage approach in variable selection for mixed effects s GGI Statistics Conference, Florence, 2015 Bayesian Variable Selection June 22-26, 2015 Outline 1 Introduction 2 3 4 Outline Introduction

More information

Biost 518 Applied Biostatistics II. Purpose of Statistics. First Stage of Scientific Investigation. Further Stages of Scientific Investigation

Biost 518 Applied Biostatistics II. Purpose of Statistics. First Stage of Scientific Investigation. Further Stages of Scientific Investigation Biost 58 Applied Biostatistics II Scott S. Emerson, M.D., Ph.D. Professor of Biostatistics University of Washington Lecture 5: Review Purpose of Statistics Statistics is about science (Science in the broadest

More information

Bayesian inference for sample surveys. Roderick Little Module 2: Bayesian models for simple random samples

Bayesian inference for sample surveys. Roderick Little Module 2: Bayesian models for simple random samples Bayesian inference for sample surveys Roderick Little Module : Bayesian models for simple random samples Superpopulation Modeling: Estimating parameters Various principles: least squares, method of moments,

More information

GS Analysis of Microarray Data

GS Analysis of Microarray Data GS01 0163 Analysis of Microarray Data Keith Baggerly and Kevin Coombes Section of Bioinformatics Department of Biostatistics and Applied Mathematics UT M. D. Anderson Cancer Center kabagg@mdanderson.org

More information

1 Data Arrays and Decompositions

1 Data Arrays and Decompositions 1 Data Arrays and Decompositions 1.1 Variance Matrices and Eigenstructure Consider a p p positive definite and symmetric matrix V - a model parameter or a sample variance matrix. The eigenstructure is

More information

Lecture Notes based on Koop (2003) Bayesian Econometrics

Lecture Notes based on Koop (2003) Bayesian Econometrics Lecture Notes based on Koop (2003) Bayesian Econometrics A.Colin Cameron University of California - Davis November 15, 2005 1. CH.1: Introduction The concepts below are the essential concepts used throughout

More information

Bayesian Linear Models

Bayesian Linear Models Eric F. Lock UMN Division of Biostatistics, SPH elock@umn.edu 03/07/2018 Linear model For observations y 1,..., y n, the basic linear model is y i = x 1i β 1 +... + x pi β p + ɛ i, x 1i,..., x pi are predictors

More information

Bayesian Inference on Joint Mixture Models for Survival-Longitudinal Data with Multiple Features. Yangxin Huang

Bayesian Inference on Joint Mixture Models for Survival-Longitudinal Data with Multiple Features. Yangxin Huang Bayesian Inference on Joint Mixture Models for Survival-Longitudinal Data with Multiple Features Yangxin Huang Department of Epidemiology and Biostatistics, COPH, USF, Tampa, FL yhuang@health.usf.edu January

More information

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

Bayesian Inference. Chapter 4: Regression and Hierarchical Models Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Advanced Statistics and Data Mining Summer School

More information

A short introduction to INLA and R-INLA

A short introduction to INLA and R-INLA A short introduction to INLA and R-INLA Integrated Nested Laplace Approximation Thomas Opitz, BioSP, INRA Avignon Workshop: Theory and practice of INLA and SPDE November 7, 2018 2/21 Plan for this talk

More information

STA 294: Stochastic Processes & Bayesian Nonparametrics

STA 294: Stochastic Processes & Bayesian Nonparametrics MARKOV CHAINS AND CONVERGENCE CONCEPTS Markov chains are among the simplest stochastic processes, just one step beyond iid sequences of random variables. Traditionally they ve been used in modelling a

More information

Petr Volf. Model for Difference of Two Series of Poisson-like Count Data

Petr Volf. Model for Difference of Two Series of Poisson-like Count Data Petr Volf Institute of Information Theory and Automation Academy of Sciences of the Czech Republic Pod vodárenskou věží 4, 182 8 Praha 8 e-mail: volf@utia.cas.cz Model for Difference of Two Series of Poisson-like

More information

CTDL-Positive Stable Frailty Model

CTDL-Positive Stable Frailty Model CTDL-Positive Stable Frailty Model M. Blagojevic 1, G. MacKenzie 2 1 Department of Mathematics, Keele University, Staffordshire ST5 5BG,UK and 2 Centre of Biostatistics, University of Limerick, Ireland

More information

Correlation and simple linear regression S5

Correlation and simple linear regression S5 Basic medical statistics for clinical and eperimental research Correlation and simple linear regression S5 Katarzyna Jóźwiak k.jozwiak@nki.nl November 15, 2017 1/41 Introduction Eample: Brain size and

More information

Empirical Bayes Quantile-Prediction aka E-B Prediction under Check-loss;

Empirical Bayes Quantile-Prediction aka E-B Prediction under Check-loss; BFF4, May 2, 2017 Empirical Bayes Quantile-Prediction aka E-B Prediction under Check-loss; Lawrence D. Brown Wharton School, Univ. of Pennsylvania Joint work with Gourab Mukherjee and Paat Rusmevichientong

More information

Spatial inference. Spatial inference. Accounting for spatial correlation. Multivariate normal distributions

Spatial inference. Spatial inference. Accounting for spatial correlation. Multivariate normal distributions Spatial inference I will start with a simple model, using species diversity data Strong spatial dependence, Î = 0.79 what is the mean diversity? How precise is our estimate? Sampling discussion: The 64

More information

Kneib, Fahrmeir: Supplement to "Structured additive regression for categorical space-time data: A mixed model approach"

Kneib, Fahrmeir: Supplement to Structured additive regression for categorical space-time data: A mixed model approach Kneib, Fahrmeir: Supplement to "Structured additive regression for categorical space-time data: A mixed model approach" Sonderforschungsbereich 386, Paper 43 (25) Online unter: http://epub.ub.uni-muenchen.de/

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 25: Markov Chain Monte Carlo (MCMC) Course Review and Advanced Topics Many figures courtesy Kevin

More information

False discovery rate and related concepts in multiple comparisons problems, with applications to microarray data

False discovery rate and related concepts in multiple comparisons problems, with applications to microarray data False discovery rate and related concepts in multiple comparisons problems, with applications to microarray data Ståle Nygård Trial Lecture Dec 19, 2008 1 / 35 Lecture outline Motivation for not using

More information

Bayesian Linear Models

Bayesian Linear Models Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Department of Forestry & Department of Geography, Michigan State University, Lansing Michigan, U.S.A. 2 Biostatistics, School of Public

More information

A general mixed model approach for spatio-temporal regression data

A general mixed model approach for spatio-temporal regression data A general mixed model approach for spatio-temporal regression data Thomas Kneib, Ludwig Fahrmeir & Stefan Lang Department of Statistics, Ludwig-Maximilians-University Munich 1. Spatio-temporal regression

More information

Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands

Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands Elizabeth C. Mannshardt-Shamseldin Advisor: Richard L. Smith Duke University Department

More information

Regularization in Cox Frailty Models

Regularization in Cox Frailty Models Regularization in Cox Frailty Models Andreas Groll 1, Trevor Hastie 2, Gerhard Tutz 3 1 Ludwig-Maximilians-Universität Munich, Department of Mathematics, Theresienstraße 39, 80333 Munich, Germany 2 University

More information

Lecture 3: Latent Variables Models and Learning with the EM Algorithm. Sam Roweis. Tuesday July25, 2006 Machine Learning Summer School, Taiwan

Lecture 3: Latent Variables Models and Learning with the EM Algorithm. Sam Roweis. Tuesday July25, 2006 Machine Learning Summer School, Taiwan Lecture 3: Latent Variables Models and Learning with the EM Algorithm Sam Roweis Tuesday July25, 2006 Machine Learning Summer School, Taiwan Latent Variable Models What to do when a variable z is always

More information

Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D.

Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D. Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D. Ruppert A. EMPIRICAL ESTIMATE OF THE KERNEL MIXTURE Here we

More information

Wavelet-based clustering for mixed-effects functional models in high dimension.

Wavelet-based clustering for mixed-effects functional models in high dimension. Wavelet-based clustering for mixed-effects functional models in high dimension. Franck Picard, LBBE - Lyon Madison Giacofci LJK (Grenoble) Sophie Lambert-Lacroix (TIMC - Grenoble) Guillemette Marot (Univ.

More information

Supplementary Materials for Molecular QTL Discovery Incorporating Genomic Annotations using Bayesian False Discovery Rate Control

Supplementary Materials for Molecular QTL Discovery Incorporating Genomic Annotations using Bayesian False Discovery Rate Control Supplementary Materials for Molecular QTL Discovery Incorporating Genomic Annotations using Bayesian False Discovery Rate Control Xiaoquan Wen Department of Biostatistics, University of Michigan A Model

More information

Experimental Design and Data Analysis for Biologists

Experimental Design and Data Analysis for Biologists Experimental Design and Data Analysis for Biologists Gerry P. Quinn Monash University Michael J. Keough University of Melbourne CAMBRIDGE UNIVERSITY PRESS Contents Preface page xv I I Introduction 1 1.1

More information

Some Curiosities Arising in Objective Bayesian Analysis

Some Curiosities Arising in Objective Bayesian Analysis . Some Curiosities Arising in Objective Bayesian Analysis Jim Berger Duke University Statistical and Applied Mathematical Institute Yale University May 15, 2009 1 Three vignettes related to John s work

More information

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

Bayesian Inference. Chapter 4: Regression and Hierarchical Models Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Master in Business Administration and Quantitative

More information

Comparing Non-informative Priors for Estimation and Prediction in Spatial Models

Comparing Non-informative Priors for Estimation and Prediction in Spatial Models Environmentrics 00, 1 12 DOI: 10.1002/env.XXXX Comparing Non-informative Priors for Estimation and Prediction in Spatial Models Regina Wu a and Cari G. Kaufman a Summary: Fitting a Bayesian model to spatial

More information

Bayesian variable selection via. Penalized credible regions. Brian Reich, NCSU. Joint work with. Howard Bondell and Ander Wilson

Bayesian variable selection via. Penalized credible regions. Brian Reich, NCSU. Joint work with. Howard Bondell and Ander Wilson Bayesian variable selection via penalized credible regions Brian Reich, NC State Joint work with Howard Bondell and Ander Wilson Brian Reich, NCSU Penalized credible regions 1 Motivation big p, small n

More information

BNP survival regression with variable dimension covariate vector

BNP survival regression with variable dimension covariate vector BNP survival regression with variable dimension covariate vector Peter Müller, UT Austin Survival 0.0 0.2 0.4 0.6 0.8 1.0 Truth Estimate Survival 0.0 0.2 0.4 0.6 0.8 1.0 Truth Estimate BC, BRAF TT (left),

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1 Inverse of a Square Matrix For an N N square matrix A, the inverse of A, 1 A, exists if and only if A is of full rank, i.e., if and only if no column of A is a linear combination 1 of the others. A is

More information

Bayesian linear regression

Bayesian linear regression Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding

More information

MACHINE LEARNING ADVANCED MACHINE LEARNING

MACHINE LEARNING ADVANCED MACHINE LEARNING MACHINE LEARNING ADVANCED MACHINE LEARNING Recap of Important Notions on Estimation of Probability Density Functions 22 MACHINE LEARNING Discrete Probabilities Consider two variables and y taking discrete

More information

Linear Models A linear model is defined by the expression

Linear Models A linear model is defined by the expression Linear Models A linear model is defined by the expression x = F β + ɛ. where x = (x 1, x 2,..., x n ) is vector of size n usually known as the response vector. β = (β 1, β 2,..., β p ) is the transpose

More information

6. Fractional Imputation in Survey Sampling

6. Fractional Imputation in Survey Sampling 6. Fractional Imputation in Survey Sampling 1 Introduction Consider a finite population of N units identified by a set of indices U = {1, 2,, N} with N known. Associated with each unit i in the population

More information

A Fully Nonparametric Modeling Approach to. BNP Binary Regression

A Fully Nonparametric Modeling Approach to. BNP Binary Regression A Fully Nonparametric Modeling Approach to Binary Regression Maria Department of Applied Mathematics and Statistics University of California, Santa Cruz SBIES, April 27-28, 2012 Outline 1 2 3 Simulation

More information

Bayesian Linear Models

Bayesian Linear Models Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department

More information

False Discovery Control in Spatial Multiple Testing

False Discovery Control in Spatial Multiple Testing False Discovery Control in Spatial Multiple Testing WSun 1,BReich 2,TCai 3, M Guindani 4, and A. Schwartzman 2 WNAR, June, 2012 1 University of Southern California 2 North Carolina State University 3 University

More information

Course topics (tentative) The role of random effects

Course topics (tentative) The role of random effects Course topics (tentative) random effects linear mixed models analysis of variance frequentist likelihood-based inference (MLE and REML) prediction Bayesian inference The role of random effects Rasmus Waagepetersen

More information

Hierarchical Modelling for Univariate Spatial Data

Hierarchical Modelling for Univariate Spatial Data Hierarchical Modelling for Univariate Spatial Data Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department

More information

Frequentist Accuracy of Bayesian Estimates

Frequentist Accuracy of Bayesian Estimates Frequentist Accuracy of Bayesian Estimates Bradley Efron Stanford University Bayesian Inference Parameter: µ Ω Observed data: x Prior: π(µ) Probability distributions: Parameter of interest: { fµ (x), µ

More information

2 Statistical Estimation: Basic Concepts

2 Statistical Estimation: Basic Concepts Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:

More information

Bayesian Linear Models

Bayesian Linear Models Bayesian Linear Models Sudipto Banerjee September 03 05, 2017 Department of Biostatistics, Fielding School of Public Health, University of California, Los Angeles Linear Regression Linear regression is,

More information

The linear model is the most fundamental of all serious statistical models encompassing:

The linear model is the most fundamental of all serious statistical models encompassing: Linear Regression Models: A Bayesian perspective Ingredients of a linear model include an n 1 response vector y = (y 1,..., y n ) T and an n p design matrix (e.g. including regressors) X = [x 1,..., x

More information