Data Analysis I: Single Subject

Similar documents
The General Linear Model (GLM)

The General Linear Model (GLM)

The General Linear Model. Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London

Modelling temporal structure (in noise and signal)

Event-related fmri. Christian Ruff. Laboratory for Social and Neural Systems Research Department of Economics University of Zurich

Overview of SPM. Overview. Making the group inferences we want. Non-sphericity Beyond Ordinary Least Squares. Model estimation A word on power

What is NIRS? First-Level Statistical Models 5/18/18

Optimization of Designs for fmri

Contents. Introduction The General Linear Model. General Linear Linear Model Model. The General Linear Model, Part I. «Take home» message

Jean-Baptiste Poline

Statistical Inference

NeuroImage. Modeling the hemodynamic response function in fmri: Efficiency, bias and mis-modeling

Extracting fmri features

Statistical Analysis Aspects of Resting State Functional Connectivity

Statistical Inference

Signal Processing for Functional Brain Imaging: General Linear Model (2)

Neuroimaging for Machine Learners Validation and inference

General linear model: basic

Beyond Univariate Analyses: Multivariate Modeling of Functional Neuroimaging Data

1st level analysis Basis functions, parametric modulation and correlated regressors

Mixed effects and Group Modeling for fmri data

MIXED EFFECTS MODELS FOR TIME SERIES

HST 583 FUNCTIONAL MAGNETIC RESONANCE IMAGING DATA ANALYSIS AND ACQUISITION A REVIEW OF STATISTICS FOR FMRI DATA ANALYSIS

Economics 308: Econometrics Professor Moody

Experimental design of fmri studies & Resting-State fmri

Experimental Design. Rik Henson. With thanks to: Karl Friston, Andrew Holmes

The ASL signal. Parenchy mal signal. Venous signal. Arterial signal. Input Function (Label) Dispersion: (t e -kt ) Relaxation: (e -t/t1a )

Effective Connectivity & Dynamic Causal Modelling

Experimental design of fmri studies

Overview of Spatial Statistics with Applications to fmri

Variance. Standard deviation VAR = = value. Unbiased SD = SD = 10/23/2011. Functional Connectivity Correlation and Regression.

Experimental design of fmri studies

Multiple Regression Analysis

Model-free Functional Data Analysis

Granger Mediation Analysis of Functional Magnetic Resonance Imaging Time Series

Lab 07 Introduction to Econometrics

Contents. design. Experimental design Introduction & recap Experimental design «Take home» message. N εˆ. DISCOS SPM course, CRC, Liège, 2009

Commentary on the statistical properties of noise and its implication on general linear models in functional near-infrared spectroscopy

Econometrics Summary Algebraic and Statistical Preliminaries

Experimental design of fmri studies

Analysis of fmri Timeseries:

AUTOCORRELATION. Phung Thanh Binh

INTRODUCTION TO BASIC LINEAR REGRESSION MODEL

Dynamic Causal Modelling for fmri

Overview. Experimental Design. A categorical analysis. A taxonomy of design. A taxonomy of design. A taxonomy of design. 1. A Taxonomy of Designs

LECTURE 11. Introduction to Econometrics. Autocorrelation

FinQuiz Notes

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Covers Chapter 10-12, some of 16, some of 18 in Wooldridge. Regression Analysis with Time Series Data

Group analysis. Jean Daunizeau Wellcome Trust Centre for Neuroimaging University College London. SPM Course Edinburgh, April 2010

Group Analysis. Lexicon. Hierarchical models Mixed effect models Random effect (RFX) models Components of variance

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

ISyE 691 Data mining and analytics

Classic Time Series Analysis

Making sense of Econometrics: Basics

Regression Models - Introduction

Lectures 5 & 6: Hypothesis Testing

Simple Linear Regression: The Model

Statistics: A review. Why statistics?

Bayesian modelling of fmri time series

Detecting fmri activation allowing for unknown latency of the hemodynamic response

MULTICOLLINEARITY AND VARIANCE INFLATION FACTORS. F. Chiaromonte 1

Experimental design of fmri studies

FIL. Event-related. fmri. Rik Henson. With thanks to: Karl Friston, Oliver Josephs

The General Linear Model. Monday, Lecture 2 Jeanette Mumford University of Wisconsin - Madison

Statistical Analysis of fmrl Data

Review of Econometrics

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Job Training Partnership Act (JTPA)

Contrasts and Classical Inference

1. How can you tell if there is serial correlation? 2. AR to model serial correlation. 3. Ignoring serial correlation. 4. GLS. 5. Projects.

Lecture 2: Univariate Time Series

1. The Multivariate Classical Linear Regression Model

11.1 Gujarati(2003): Chapter 12

Ch 2: Simple Linear Regression

Econometrics Honor s Exam Review Session. Spring 2012 Eunice Han

Lecture 3: Multiple Regression

The General Linear Model Ivo Dinov

Functional Causal Mediation Analysis with an Application to Brain Connectivity. Martin Lindquist Department of Biostatistics Johns Hopkins University

Statistical Analysis of Functional ASL Images

Bayesian inference J. Daunizeau

LECTURE 2 LINEAR REGRESSION MODEL AND OLS

Machine learning strategies for fmri analysis

Simple Linear Regression

ECONOMETRICS HONOR S EXAM REVIEW SESSION

Heteroscedasticity and Autocorrelation

Matematické Metody v Ekonometrii 7.

Wellcome Trust Centre for Neuroimaging, UCL, UK.

covariance function, 174 probability structure of; Yule-Walker equations, 174 Moving average process, fluctuations, 5-6, 175 probability structure of

E 4160 Autumn term Lecture 9: Deterministic trends vs integrated series; Spurious regression; Dickey-Fuller distribution and test

1. The OLS Estimator. 1.1 Population model and notation

Lecture 4: Multivariate Regression, Part 2

Introduction to Econometrics

11. Further Issues in Using OLS with TS Data

MS&E 226. In-Class Midterm Examination Solutions Small Data October 20, 2015

Föreläsning /31

Introduction to Econometrics. Heteroskedasticity

Linear Regression. Junhui Qian. October 27, 2014

Reliability of inference (1 of 2 lectures)

Contents. Data. Introduction & recap Variance components Hierarchical model RFX and summary statistics Variance/covariance matrix «Take home» message

Transcription:

Data Analysis I: Single Subject

ON OFF

he General Linear Model (GLM) y= X fmri Signal = Design Matrix our data = what we CAN explain x β x Betas + + how much x of it we CAN + explain ε Residuals what we CANNO explain If a voxel cares about the task, then you should be able to predict its time-course from the onset and offset of the task. β captures how much of 'y' your task/design is predicting.

he General Linear Model (GLM) y= X fmri Signal = Design Matrix our data = what we CAN explain x β x Betas + + how much x of it we CAN + explain ε Residuals what we CANNO explain

he General Linear Model (GLM) yi = β0+β1x1+ εi yi Reaction time X1 Age β0, β1 are the model parameters εi is the random error; how observation deviates from population mean

he General Linear Model (GLM) yi = β0+β1x1+ εi Fixed: β0+β1x1 Mean of Yi (E[Yi]) Random: εi he variability of Y i (needs to have specific properties)

he General Linear Model (GLM) AGE Source: J Mumford

Statistical Analysis Conceptually the analysis is divided into 2 steps: 1. Estimation: the magnitude of the effect of interest (β) 2. Inference: assess significance of the estimated parameter

Ordinary Least Squares (OLS) e.g. What is the relationship tying years of MALB experience and coding efficiency?

Ordinary Least Squares (OLS) e.g. What is the relationship tying years of MALB experience and coding efficiency? Minimize error: Actual (i.e. Y) n 2 i n 2 e = Y i Y i i=1 n Predicted (i.e. Y-hat) i=1 2 Y i 1 X 1 i=1

Ordinary Least Squares (OLS) e.g. What is the relationship tying years of MALB experience and coding efficiency? Minimize error: Actual (i.e. Y) n 2 i n 2 e = Y i Y i i=1 n Predicted (i.e. Y-hat) i=1 2 Y i 1 X 1 i=1

Least Squares Difference Minimize error n n i=1 i=1 2 2 e = Y i Y i = Y i i X i i=1 2 i n Y i Y X X i 1= 2 Y i Y 0 =Y 1 X

Massive Univariate Approach y = 0 1 X 1

Massive Univariate Approach y = 0 1 X 1 2 X 2... p X p

Voxel-wise GLM y 1= 0 1 X 1,1 2 X 2,1 p X p,1 1 y 2= 0 1 X 1,2 2 X 2,2 p X p, 2 2 y 3= 0 1 X 1,3 2 X 2,3 p X p,3 3 y n= 0 1 X 1, n 2 X 2, n p X p, n n

Voxel-wise GLM

Voxel-wise GLM

Voxel-wise GLM

Voxel-wise GLM

Voxel-wise GLM

Ordinary Least Squares (OLS) Y = X Y = X X Y = X X X X X Y = 1

Ordinary Least Squares (OLS) Y = X = X X 1 X Y 2 1 Var = X X 1 X X X Y t= = ~ N p 2 1 Var X X

Ordinary Least Squares (OLS) Effect of interest Nature 1 X X X Y t= = ~ N p 2 1 X X Var Subject, scanner,... You Residuals Error (i.e., what we don't know) Design Matrix

Gauss Markov Assumptions Under the following assumptions, the OLS estimator is the Best Linear Unbiased Estimator (BLUE).

Source: J Mumford

Gauss Markov Assumptions Under the following assumptions, the OLS estimator is the Best Linear Unbiased Estimator (BLUE). i.i.d. 2 ~ N 0, I E i =0 Error has mean 0 Error is Homoschedastic Var i = 2 Errors have same variance Errors are not Autocorrelated Cov i, j =0 Errors are independent of each other

A simple experiment: the FFA Ss: 1 participant Design: 16s Faces, 16s Houses (block design) ake Voxel 15,19,18, in right posterior temporal cortex. Does this voxel care about faces?

Model I: Boxcar Regressor Y = X *β

Model I: Boxcar Regressor

Model I: Boxcar Regressor 1 X X X Y t df = 2 1 X X 150.99 t 159 = =4.66 202.44 0.16

Model II: Convolved Boxcar Regressor Gaussian 1 X X X Y t df = 2 1 X X 267.02 t 159 = =10.01 166.70 0.16

Model II: Convolved Boxcar Regressor Gamma 1 X X X Y t df = 2 1 X X 339.94 t 159 = =12.77 166.31 0.16

Model II: Convolved Boxcar Regressor Double Gamma 1 X X X Y t df = 2 1 X X 344.74 t 159 = =12.77 168.67 0.16

What Convolution? Linear ime Invariant Basis Functions: Canonical HRF Canonical HRF + temp & disp derivs Informed basis set FIR/Smooth FIR Non Linear: Inverse Logit Balloon model

Canonical HRF Stimulus HRF (DGamma)

HRF Variability Across Subjects Handwerker et al 2004 Aguirre et Al 1998

HRF Variability Within Subjects same subject, same session same subject, different session Aguirre et al 1998

HRF Variability Across Brain Area Handwerker et al 2004 Handwerker et al 2004

Basis Functions I: HRF + Derivatives Canonical HRF emporal derivative Dispersion derivative

Basis Functions II: Fourier Basis Set f(t) h1(t) h2(t) h3(t)

Basis Functions III: Informed Basis Functions (FLOBS)

Basis Functions IV: FIR

Non Linear Approaches Model all 3 parameters (,W,H)

Slide from M Lindquist Inverse Logit Model Superposition of three inverse logit (sigmoid) functions. x 1 L x = 1 e Each function has three variable parameters dictating the amplitude, position and slope. h t θ =α 1 L t 1 / D 1 α 2 L t 2 / D2 α 3 L t 3 / D3 Lindquist & Wager (2007)

Slide from M Lindquist Flexibility By shifting the position of the second IL function one can model differences in duration. By shifting the position of all three IL functions one can model differences in onset.

Balloon Model BOLD response, % positive BOLD response 3 2 1 initial overshoot dip post stimulus undershoot 0 stimulus time

What Convolution? - Recap he problem: Bias/Variance trade off Assume known shape v estimate X-Matrix mis-specification: Incorrect HRF will increase σ2

What Convolution? - Recap Canonical HRF: Pros: 1 parameter only, easy to interpret; easy to integrate at the second level. Cons: Bias in favour of a specific filter Basis Sets Functions: Pros: Flexible, can capture greater variability, unexpected and subject specific differences. Cons: many parameters to estimate (degrading dfs); difficult to interpret; how to aggregate at the 2nd level? Non-Linear Systems: Potentially very flexible and may better capture parameters and vasculature Computationally very demanding, parameters not obvious to interpret and aggregate across subjects (e.g., balloon model)

X-Matrix Assumptions Deterministic & known No regressor is a linear transformation of one (or more) regressors

Multicollinearity No column of the X matrix can be obtained by linear combination of one or more other columns. he issue is not whether there is/isn't multicollinearity, but rather, how much is there. wo main consequences: Decreased efficiency of the design degrades power. Parameter estimates can vary wildly, even in the wrong direction!

Source: J Mumford

Source: J Mumford

Source: J Mumford

Source: J Mumford

Source: J Mumford

Contrasts

Contrasts Cond A D A Cond B D B

Contrasts t= Var 1 2 3 = 4. p Cond A D A Cond B D B

Contrasts t= Var 1 2 3 = 4. p Cond A D A Cond B D B

Contrasts t= Var 1 2 3 = 4. p Cond A D A Cond B D B

Contrasts t= Var 1 2 3 = 4. p Cond A D A Cond B D B

Contrasts c t= Var c 1 2 3 = 4. p c1 c2 c3 c= c4. cp Cond A D A Cond B D B

Contrasts c t= Var c 1 2 3 = 4. p 1 0 0 c= 0. 0 Cond A D A Cond B D B

Contrasts c t= Var c 1 2 3 = 4. p 0 0 1 c= 0. 0 Cond A D A Cond B D B

Contrasts c t= Var c 1 2 3 = 4. p 1 0 1 c= 0. 0 Cond A D A Cond B D B

Contrasts c t= Var c 1 2 3 = 4. p 1 0 1 c= 0. 0 Cond A D A Cond B D B

Contrasts c t= Var c

Model I: Boxcar Regressor 1 X X X Y t df = 2 1 X X 150.99 t 159 = =4.66 202.44 0.16

Model II: Convolved Boxcar Regressor Double Gamma 1 X X X Y t df = 2 1 X X 344.74 t 159 = =12.77 168.67 0.16

Gauss Markov Assumptions Under the following assumptions, the OLS estimator is the Best Linear Unbiased Estimator (BLUE). E i =0 Error has mean 0 Error is Homoschedastic i.i.d. 2 ~ N 0, I Var i = 2 Errors have same variance Errors are not Autocorrelated Cov i, j =0 Errors are independent of each other

emporal Structure If residuals are correlated, estimator is still unbiased but the variance estimate is not; and df are less than nominal. Corr(εt,εt-1) > 0 Corr(εt,εt-1) < 0

emporal Structure If residuals are correlated, estimator is still unbiased but the variance estimate is not; and df are less than nominal. his can affect your -test either way: i. Positive autocorrelation: underestimate of variance, -tests artificially inflated (actual α up to 0.16 for a nominal of 0.05). ii.negative autocorrelation: overestimate of variance, -test artificially reduced. wo strategies: i. High Pass filtering ii.autocorrelation correction ( pre-whitening ; colouring )

Low frequency noise Plenty low frequency noise: scanner drift, physiological (e.g., basal metabolism), psychological (e.g., learning, alertness changes), left-over from motion correction,... power spectrum noise signal power spectrum

HP Filtering Strategy I: SPM Model low drifts to soak up their variance (using a discrete cosine transform basis set).

HP Filtering Strategy II: FSL Remove low drifts from the signal: i. Fit a Gaussian-weighted running line

HP Filtering Strategy II: FSL Remove low drifts from the signal: i. Fit a Gaussian-weighted running line ii.subtract from data (red is pre-hpf, green is post-hpf)

Model II: Convolved Boxcar Regressor Double Gamma 1 X X X Y t df = 2 1 X X 344.74 t 159 = =12.77 168.67 0.16

Model III: High Pass Filter Double Gamma 1 X X X Y t df = 2 1 X X 349 t 159 = =11.05 154.08 0.21

Autocorrelation GM Assumption: 2 i.i.d. ~ N 0, I Errors are independent of each other Cov i, j =0 fmri Data: i.i.d. 2 2 ~ N 0, V Cov i, j = V Y = X V Errors are NO independent of each other

Autocorrelation GM Assumption: i.i.d. 2 ~ N 0, I Cov i, j =0 cov(ε1,ε2) cov(ε2,ε1) cov(εn,ε1) cov(ε2,εn) Errors are independent of each other cov(ε1,εn) cov(ε2,εn)

Autocorrelation: Pre-whitening FSL If only we knew the structure of the serial correlation in the data (V) we could get rid of it: Given V, there exist a square (non singular) matrix K such that: i. V = KK ii. e = Kε (where ε is a 'well behaved' residual ~N(0,σ2I)) So, we can now clean-up (i.e., whiten ) our GLM 1 K Y =K 1 1 1 X K 2 1 1 Cov K = K V K 2 1 1 = K KK K 2 = I

Autocorrelation: Pre-whitening FSL If only we knew the structure of the serial correlation in the data (V) we could get rid of it; but we don't know it. 1. Estimate the autocorrelation structure (for every voxel): Lag 1 [e01 e02 e03 e04 e05 e06 e07 e08 e09 e10 e11 e12] [e01 e02 e03 e04 e05 e06 e07 e08 e09 e10 e11 e12]

Autocorrelation: Pre-whitening FSL If only we knew the structure of the serial correlation in the data (V) we could get rid of it; but we don't know it. 1. Estimate the autocorrelation structure (for every voxel): Lag 2 [e01 e02 e03 e04 e05 e06 e07 e08 e09 e10 e11 e12] [e01 e02 e03 e04 e05 e06 e07 e08 e09 e10 e11 e12]

Autocorrelation: Pre-whitening FSL If only we knew the structure of the serial correlation in the data (V) we could get rid of it; but we don't know it. 1. Estimate the autocorrelation structure (for every voxel): Lag 3 [e01 e02 e03 e04 e05 e06 e07 e08 e09 e10 e11 e12] [e01 e02 e03 e04 e05 e06 e07 e08 e09 e10 e11 e

Autocorrelation: Pre-whitening FSL If only we knew the structure of the serial correlation in the data (V) we could get rid of it; but we don't know it. 1. Estimate the autocorrelation structure (for every voxel): Lag 8 [e01 e02 e03 e04 e05 e06 e07 e08 e09 e10 e11 e12] [e01 e02 e03 e04 e05 e06

Autocorrelation: Pre-whitening FSL If only we knew the structure of the serial correlation in the data (V) we could get rid of it; but we don't know it. 1. Estimate the autocorrelation structure (for every voxel): Lag 9 [e01 e02 e03 e04 e05 e06 e07 e08 e09 e10 e11 e12] [e01 e02 e03 e04 e05 e

Autocorrelation: Pre-whitening FSL If only we knew the structure of the serial correlation in the data (V) we could get rid of it; but we don't know it. 1. Estimate the autocorrelation structure (for every voxel) 2. For each lag, smooth estimates spatially ('regularization')

Autocorrelation: Pre-whitening FSL If only we knew the structure of the serial correlation in the data (V) we could get rid of it; but we don't know it. 1. Estimate the autocorrelation structure (for every voxel) 2. For each lag, smooth estimates spatially ('regularization') 3. Smooth (temporally) the correlation estimate, for each voxel ('ukey taper'). Woolrich et al, 2001 NeuroImage

Autocorrelation: Pre-whitening FSL If only we knew the structure of the serial correlation in the data (V) we could get rid of it; but we don't know it. 1. Estimate the autocorrelation structure (for every voxel) 2. For each lag, smooth estimates spatially ('regularization') 3. Smooth (temporally) the correlation estimate, for each voxel ('ukey taper'). Raw (rxx) Woolrich et al, 2001 NeuroImage Smooth (ρxx)

Autocorrelation: Pre-whitening FSL If only we knew the structure of the serial correlation in the data (V) we could get rid of it; but we don't know it. Now that we have an estimate of the structure, we can run a second regression in which we whiten the residuals using the estimated V. he estimated β for the new model are BLUE (under the assumption that V was appropriately modelled)

Autocorrelation: pre-whitening SPM Global estimation of the correlations: the correlation of the time series is averaged over voxels. Imposed AR(1) model with correlation 0.2 plus white noise.

Model III: High Pass Filter Double Gamma 1 X X X Y t df = 2 1 X X 349 t 159 = =10.78 154.08 0.21

Model IV: Whitening Double Gamma 1 X X X Y t df = 2 1 X X 273.36 t 159 = =10.00 130.20 0.21

Whitening algorithms performance Lenoski 08 IEEE

Autocorrelation: Coloring i.i.d. 2 2 ~ N 0, V Cov i, j = V Y = X V Errors are NO independent of each other Impose a known correlation structure S If indeed SVS ~ SS then the error is N(0, σ2s) So: SY = SXβ + Sε Estimates are unbiased but no longer efficient (according to how well V is swamped )

Coloring v. Whitening Design Power Spectrum Coloring Pre-whitening

omorrow Setting up group fmri models Email Jeanette your group model challenges and she may choose it to include in her workshop jeanette.mumford@gmail.com