Estimating representational dissimilarity measures

Similar documents
Representational similarity analysis. Nikolaus Kriegeskorte MRC Cognition and Brain Sciences Unit Cambridge, UK

Introduction to representational similarity analysis

Advanced topics in RSA. Nikolaus Kriegeskorte MRC Cognition and Brain Sciences Unit Cambridge, UK

Extracting fmri features

The General Linear Model (GLM)

Bayesian Decision and Bayesian Learning

Neuroimaging for Machine Learners Validation and inference

Application: Can we tell what people are looking at from their brain activity (in real time)? Gaussian Spatial Smooth

The General Linear Model (GLM)

Estimation of Parameters

The General Linear Model. Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London

Event-related fmri. Christian Ruff. Laboratory for Social and Neural Systems Research Department of Economics University of Zurich

Searchlight-based multi-voxel pattern analysis of fmri by cross-validated MANOVA

Multivariate Statistical Analysis

Statistical Inference

Peak Detection for Images

University of Cambridge Engineering Part IIB Module 4F10: Statistical Pattern Processing Handout 2: Multivariate Gaussians

Introduction to Machine Learning

The exam is closed book, closed notes except your one-page (two sides) or two-page (one side) crib sheet.

CHAPTER 2. Types of Effect size indices: An Overview of the Literature

Holdout and Cross-Validation Methods Overfitting Avoidance

Introduction to Machine Learning

Data Analysis I: Single Subject

Minwise hashing for large-scale regression and classification with sparse data

Detection of Anomalies in Texture Images using Multi-Resolution Features

6348 Final, Fall 14. Closed book, closed notes, no electronic devices. Points (out of 200) in parentheses.

LDA, QDA, Naive Bayes

Overview of SPM. Overview. Making the group inferences we want. Non-sphericity Beyond Ordinary Least Squares. Model estimation A word on power

Statistical Inference

University of Cambridge Engineering Part IIB Module 4F10: Statistical Pattern Processing Handout 2: Multivariate Gaussians

Research Design - - Topic 15a Introduction to Multivariate Analyses 2009 R.C. Gardner, Ph.D.

Supervised Learning: Linear Methods (1/2) Applied Multivariate Statistics Spring 2012

CMU-Q Lecture 24:

Chapter 3: Maximum-Likelihood & Bayesian Parameter Estimation (part 1)

Multivariate Regression Generalized Likelihood Ratio Tests for FMRI Activation

Statistical Machine Learning

Problem Set 2. MAS 622J/1.126J: Pattern Recognition and Analysis. Due: 5:00 p.m. on September 30

Non-linear Measure Based Process Monitoring and Fault Diagnosis

Advanced Econometrics

Eigenfaces. Face Recognition Using Principal Components Analysis

Outline. Supervised Learning. Hong Chang. Institute of Computing Technology, Chinese Academy of Sciences. Machine Learning Methods (Fall 2012)

F & B Approaches to a simple model

Machine Learning Basics

A. Motivation To motivate the analysis of variance framework, we consider the following example.

Support Vector Machine. Industrial AI Lab.

EEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1

Interpreting Regression Results

Statistical Analysis of fmrl Data

MS&E 226. In-Class Midterm Examination Solutions Small Data October 20, 2015

Signal Modeling Techniques in Speech Recognition. Hassan A. Kingravi

Classifier Performance. Assessment and Improvement

6.867 Machine learning

Lecture 9: Classification, LDA

Recipes for the Linear Analysis of EEG and applications

Lecture 9: Classification, LDA

High Dimensional Discriminant Analysis

Continuous Univariate Distributions

Decoding conceptual representations

6-1. Canonical Correlation Analysis

Mathematical Tools for Neuroscience (NEU 314) Princeton University, Spring 2016 Jonathan Pillow. Homework 8: Logistic Regression & Information Theory

Solutions. Part I Logistic regression backpropagation with a single training example

DISCRIMINANT ANALYSIS. 1. Introduction

An investigation into the use of maximum posterior probability estimates for assessing the accuracy of large maps

DEPARTMENT OF COMPUTER SCIENCE Autumn Semester MACHINE LEARNING AND ADAPTIVE INTELLIGENCE

Motivating the Covariance Matrix

Pattern Recognition and Machine Learning

Lecture 9: Classification, LDA

Y (Nominal/Categorical) 1. Metric (interval/ratio) data for 2+ IVs, and categorical (nominal) data for a single DV

Homework #2 Due date: 2/19/2013 (Tuesday) Translate the slice Two Main Paths in Lecture Part II: Neurophysiology and BOLD to words.

Manual: R package HTSmix

Support Vector Regression (SVR) Descriptions of SVR in this discussion follow that in Refs. (2, 6, 7, 8, 9). The literature

COMPARISON OF GMM WITH SECOND-ORDER LEAST SQUARES ESTIMATION IN NONLINEAR MODELS. Abstract

FMRI Neurologic Synchrony Measures for Alzheimer s Patients With Monte Carlo Critical Values

Principal component analysis

26:010:557 / 26:620:557 Social Science Research Methods

Linear Regression. Machine Learning CSE546 Kevin Jamieson University of Washington. Oct 5, Kevin Jamieson 1

Multi-Category Classification by Soft-Max Combination of Binary Classifiers

CSE446: non-parametric methods Spring 2017

CSC2541 Lecture 5 Natural Gradient

Dynamic Data Modeling, Recognition, and Synthesis. Rui Zhao Thesis Defense Advisor: Professor Qiang Ji

Machine learning strategies for fmri analysis

Outline Introduction OLS Design of experiments Regression. Metamodeling. ME598/494 Lecture. Max Yi Ren

Support Vector Machine. Industrial AI Lab. Prof. Seungchul Lee

Statistical Learning with the Lasso, spring The Lasso

Interpreting Regression Results -Part II

The Bayes classifier

Class 4: Classification. Quaid Morris February 11 th, 2011 ML4Bio

Active and Semi-supervised Kernel Classification

Resampling Based Design, Evaluation and Interpretation of Neuroimaging Models Lars Kai Hansen DTU Compute Technical University of Denmark

CS7267 MACHINE LEARNING

Application of Raman Spectroscopy for Detection of Aflatoxins and Fumonisins in Ground Maize Samples

Is cross-validation valid for small-sample microarray classification?

Discriminant analysis and supervised classification

Gaussian Process Regression with K-means Clustering for Very Short-Term Load Forecasting of Individual Buildings at Stanford

Confidence Estimation Methods for Neural Networks: A Practical Comparison

High Dimensional Discriminant Analysis

ECON Program Evaluation, Binary Dependent Variable, Misc.

Part I. Linear regression & LASSO. Linear Regression. Linear Regression. Week 10 Based in part on slides from textbook, slides of Susan Holmes

Piotr Majer Risk Patterns and Correlated Brain Activities

Incorporating detractors into SVM classification

Transcription:

Estimating representational dissimilarity measures lexander Walther <awalthermail@gmail.com> MRC Cognition and rain Sciences Unit University of Cambridge Institute of Cognitive Neuroscience University College London

Overview Distance measures for Representational Similarity nalysis fmri pattern noise normalization Crossvalidated dissimilarity measures Distance measures vs. pattern classifiers

What does Representational Similarity nalysis measure? General linear model True Regression Coefficients Ordinary least squares ˆ = (X T X) X T Y Estimated regression coefficients True activity patterns U Dissimilarity measure

The Euclidean distance in RS fmri activity patterns U u 2 = u 2 2, + u,2 u 2 = u 2 2, + u,2.8.5!.3.6! u u u.6,.3 u,2 u,2 d(,) u,2.5 u,2 Voxel u,.8 Representational dissimilarity matrix d d Voxel 2 u, u, Squared Euclidean distance d(,) 2 = (u u ) 2 + (u u ) 2,,,2,2 = (u u )(u u ) T

Euclidean distance & Pearson correlation distance fmri activity patterns U C 62.8.5.2! u.3.6.5! u u C Voxel 3 d(,) u u Voxel 2

Euclidean distance & Pearson correlation distance fmri activity patterns U.8.5.2! u.3.6.5! u Voxel 3 Voxel 2

Euclidean distance & Pearson correlation distance fmri activity patterns U C 62.8.5.2! u.3.6.5! u u C u * = u u Voxel 3 u * = u u Voxel 2

Euclidean distance & Pearson correlation distance fmri activity patterns U C 62.8.5.2! u.3.6.5! u u C u * = u u σ u Voxel 3 u * = u u σ u r = cos(u * u * ) u * Correlation distance d(,) = r α u * Voxel 2

Euclidean distance & Pearson correlation distance Euclidean distance is invariant to baseline shifts Correlation distance is invariant to scaling differences

Reliability of dissimilarity measures Split-half reliability analysis Stimulus 2 3 4 5 Stimulus 2 3 4 5 r Split Split 2 high low SVM LD Euclidean Correlation LDC LDt RDM split-half reliability [Pearson correlation].5 High SNR.7.35 Low SNR

Overview Distance measures for Representational Similarity nalysis fmri pattern noise normalization Crossvalidated dissimilarity measures Distance measures vs. pattern classifiers

Univariate & multivariate noise normalization General linear model.8.5!.3.6! Noise normalizations Univariate ˆb p * = ˆb p σ εp for p = P Voxel ˆΣ ε = T εt ε Multivariate û = ˆb ˆΣ 2 Voxel

Reliability of dissimilarity measures Split-half reliability analysis Stimulus 2 3 4 5 Stimulus 2 3 4 5 r Split Split 2 high low SVM LD Euclidean Correlation LDC LDt RDM split-half reliability [Pearson correlation].5 High SNR.7.35 Low SNR

From raw to noise normalized dissimilarity measures fmri beta weights ˆ.8.5!.3.6! ˆb ˆb Squared Euclidean distance d(,) 2 = ( ˆb ˆb )( ˆb ˆb ) T Multivariate noise normalization û = ˆb ˆΣ 2 Squared Mahalanobis distance d(,) 2 = (û û )(û û ) T Representational dissimilarity matrix d d

Overview Distance measures for Representational Similarity nalysis fmri pattern noise normalization Crossvalidated dissimilarity measures Distance measures vs. pattern classifiers

Conventional distances are positively biased

Crossvalidated distance estimates Run Run 2 Û ().8.5!.3.6! û () û () Û (2).7.6!.5.4! û (2) û (2) û () = u () + ε () û (2) = u (2) + ε (2) û () = u () + ε () û (2) = u (2) + ε (2) Crossvalidated distance d(,) = (û () û () )(û (2) û (2) ) T Training Testing

Conventional distances are positively biased

Conventional distances are positively biased Run Training ˆd(û,û ) = û () û () Run 2 Testing ( )( û (2) (2) û ) T Estimated activity patterns Training Testing True activity patterns û = u + ε û 2 = u 2 + ε 2 Estimated error = û ()û (2) T û ()û (2) T û ()û (2) T + û ()û (2) T Crossvalidation = ( u () + ε () )( u (2) + ε (2) ) T ( u () + ε () )( u (2) + ε (2) ) T ( u () + ε () )( u (2) + ε (2) ) T... = + ( u () + ε () )( u (2) + ε (2) ) T = u () u (2) T + u () ε (2)T + ε () u (2) T + ε () ε (2)T u () u (2) T u () ε (2)T ε () u (2) T ε () ε (2)T... = u () u (2) T u () ε (2)T... ε () u (2) T ε () ε (2)T + u () u (2) T + ε (2) u () T + ε () u (2) T + ε () ε (2)T E ( ˆd(û,û )) = u () (2) u T u () (2) u T u () (2) u T + u () (2) u T = ( u () () u )( u (2) (2) u ) T = d(u,u )

Crossvalidation schemes Crossvalidated distance ˆd(,) = (û () û () )(û (2) û (2) ) T Training Testing ˆ δ () ˆ δ (2) Split-half crossvalidation Leave-one-run-out crossvalidation Run ˆ δ () ˆ δ () ˆ δ () ˆ δ () ˆ δ () ˆ δ () ˆ δ () ˆ δ () ˆ δ () ˆ δ () Run 2 Run 3 ˆ δ (2) ˆ δ (3) ˆ δ (2) ˆ δ (3) ˆ δ (2) ˆ δ (3) = ˆ δ (2) ˆ δ (3) ˆ δ (2) ˆ δ (3) ˆ δ (2) ˆ δ (3) ˆ δ (2) ˆ δ (3) = ˆ δ (2) ˆ δ (3) ˆ δ (2) ˆ δ (3) ˆ δ (2) ˆ δ (3) Run 4 ˆ δ (4) ˆ δ (4) ˆ δ (4) ˆ δ (4) ˆ δ (4) ˆ δ (4) ˆ δ (4) ˆ δ (4) ˆ δ (4) ˆ δ (4) ˆd(,) = δ ˆ () + ˆ 3 2 δ (2)! δ ˆ (3) + ˆ 2 δ (4) T 4 ( δ ˆ () ) δ ˆ (2) + δ ˆ(3) + δ ˆ(4)! 3 T 6 ( δ ˆ () δ ˆ (2) T )!

Reliability of dissimilarity measures Split-half reliability analysis Stimulus 2 3 4 5 Stimulus 2 3 4 5 r Split Split 2 high low SVM LD Euclidean Correlation LDC LDt RDM split-half reliability [Pearson correlation].5 High SNR.7.35 Low SNR

From raw to noise normalized dissimilarity measures fmri beta weights ˆ.8.5!.3.6! ˆb ˆb Squared Euclidean distance d(,) = ( ˆb ˆb )( ˆb ˆb ) T Multivariate noise normalization û = ˆb ˆΣ 2 Squared Mahalanobis distance d(,) = (û û )(û û ) T Crossvalidation Representational dissimilarity matrix d d Crossvalidated Mahalanobis distance estimate ( Linear discriminant contrast, LDC) ˆd(,) = (û () u () )(u (2) u (2) ) T

From conventional to crossvalidated dissimilarity measures C! 62 fmri beta weights ˆ.8.5!.3.6!.5.!! C 62 u C ˆb ˆb ˆb C Crossvalidated Mahalanobis distance estimate ( Linear discriminant contrast, LDC) ˆd(i, j) = (û i () û j () )(û i (2) û j (2) ) T Representational dissimilarity matrix (FF) faces bodies inanimate Preserves ratio between distances faces inanimate bodies t test (FDR=.5) 6 3 dissimilarity [Linear discriminent contrast]

From crossvalidated Mahalanobis to Linear Discriminant t fmri beta weights ˆ.8.5!.3.6! ˆb ˆb Crossvalidated Mahalanobis distance estimate ( Linear discriminant contrast, LDC) ˆd(,) = (û () û () )(û (2) û (2) ) T Training w Testing Normalize by standard error SE LDC σ ε 2 = diag(w ˆΣ b w T ) SE LDC = σ ε 2 ( c X T ) X c T Representational dissimilarity matrix t t Linear discriminant t value, LDt ( ) ˆt(,) = (û () û () )(û (2) û (2) ) T SE LDC

Reliability of dissimilarity measures Split-half reliability analysis Stimulus 2 3 4 5 Stimulus 2 3 4 5 r Split Split 2 high low SVM LD Euclidean Correlation LDC LDt RDM split-half reliability [Pearson correlation].5 High SNR.7.35 Low SNR

Overview Distance measures for Representational Similarity nalysis fmri pattern noise normalization Crossvalidated dissimilarity measures Distance measures vs. pattern classifiers

Distances vs. linear classifiers Split-half reliability analysis Stimulus 2 3 4 5 Stimulus 2 3 4 5 r Split Split 2 high low Euclidean Correlation LDC LDt SVM LD RDM split-half reliability [Pearson correlation] High SNR.7.5.35 Multivariate noise normalization Low SNR

Walther et al (in review) Key insights Euclidean and correlation distance measure dissimilarity differently and are invariant to different pattern transformations 2 Multivariate noise normalization significantly enhances the reliability of dissimilarity measures 3 Crossvalidated dissimilarity estimates are unbiased and ratio scale (with interpretable zero point) 4 Pair-wise classification accuracy is a quantized measure of representational distinctness and significantly less reliable than continuous distance measures

Thanks!