Lecture'12:' SSMs;'Independent'Component'Analysis;' Canonical'Correla;on'Analysis'

Similar documents
Independent Component Analysis

Independent Component Analysis

STATS 306B: Unsupervised Learning Spring Lecture 12 May 7

Gatsby Theoretical Neuroscience Lectures: Non-Gaussian statistics and natural images Parts I-II

CIFAR Lectures: Non-Gaussian statistics and natural images

Independent Component Analysis

ICA. Independent Component Analysis. Zakariás Mátyás

Probabilistic Graphical Models

Introduction to Independent Component Analysis. Jingmei Lu and Xixi Lu. Abstract

Independent Component Analysis and Its Applications. By Qing Xue, 10/15/2004

Independent Component Analysis. PhD Seminar Jörgen Ungh

Natural Image Statistics

Estimation of linear non-gaussian acyclic models for latent factors

1 Introduction Independent component analysis (ICA) [10] is a statistical technique whose main applications are blind source separation, blind deconvo

Fundamentals of Principal Component Analysis (PCA), Independent Component Analysis (ICA), and Independent Vector Analysis (IVA)

TWO METHODS FOR ESTIMATING OVERCOMPLETE INDEPENDENT COMPONENT BASES. Mika Inki and Aapo Hyvärinen

Independent Component Analysis. Contents

Unsupervised learning: beyond simple clustering and PCA

From independent component analysis to score matching

One-unit Learning Rules for Independent Component Analysis

Independent Component Analysis

Emergence of Phase- and Shift-Invariant Features by Decomposition of Natural Images into Independent Feature Subspaces

Independent Component Analysis (ICA) Bhaskar D Rao University of California, San Diego

Independent Components Analysis

Blind Machine Separation Te-Won Lee

Separation of Different Voices in Speech using Fast Ica Algorithm

ICA [6] ICA) [7, 8] ICA ICA ICA [9, 10] J-F. Cardoso. [13] Matlab ICA. Comon[3], Amari & Cardoso[4] ICA ICA

Independent Component Analysis

Dimension Reduction (PCA, ICA, CCA, FLD,

Independent Component Analysis and Blind Source Separation

Efficient Coding. Odelia Schwartz 2017

Temporal Coherence, Natural Image Sequences, and the Visual Cortex

Independent component analysis: algorithms and applications

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

THE functional role of simple and complex cells has

An Introduction to Independent Components Analysis (ICA)

c Springer, Reprinted with permission.

Non-Euclidean Independent Component Analysis and Oja's Learning

Independent Component Analysis (ICA)

Independent Component Analysis for Blind Source Separation

Massoud BABAIE-ZADEH. Blind Source Separation (BSS) and Independent Componen Analysis (ICA) p.1/39

Different Estimation Methods for the Basic Independent Component Analysis Model

Donghoh Kim & Se-Kang Kim

Blind separation of sources that have spatiotemporal variance dependencies

Principal Component Analysis vs. Independent Component Analysis for Damage Detection

PROPERTIES OF THE EMPIRICAL CHARACTERISTIC FUNCTION AND ITS APPLICATION TO TESTING FOR INDEPENDENCE. Noboru Murata

Advanced Introduction to Machine Learning CMU-10715

Dimensionality Reduction. CS57300 Data Mining Fall Instructor: Bruno Ribeiro

Introduction to Machine Learning

Linear Factor Models. Sargur N. Srihari

Independent Component Analysis

Simple-Cell-Like Receptive Fields Maximize Temporal Coherence in Natural Video

Independent Component Analysis and FastICA. Copyright Changwei Xiong June last update: July 7, 2016

Comparative Analysis of ICA Based Features

INDEPENDENT COMPONENT ANALYSIS

Independent Component Analysis (ICA)

Single Channel Signal Separation Using MAP-based Subspace Decomposition

MTTS1 Dimensionality Reduction and Visualization Spring 2014 Jaakko Peltonen

ARTEFACT DETECTION IN ASTROPHYSICAL IMAGE DATA USING INDEPENDENT COMPONENT ANALYSIS. Maria Funaro, Erkki Oja, and Harri Valpola

Independent Component Analysis (ICA)

Artificial Intelligence Module 2. Feature Selection. Andrea Torsello

Tutorial on Blind Source Separation and Independent Component Analysis

Kernel Independent Component Analysis

Independent Component Analysis (ICA)

Recursive Generalized Eigendecomposition for Independent Component Analysis

Independent Component Analysis of Incomplete Data

Multidimensional scaling (MDS)

EXTENSIONS OF ICA AS MODELS OF NATURAL IMAGES AND VISUAL PROCESSING. Aapo Hyvärinen, Patrik O. Hoyer and Jarmo Hurri

ADAPTIVE LATERAL INHIBITION FOR NON-NEGATIVE ICA. Mark Plumbley

Learning features by contrasting natural images with noise

Nonlinear reverse-correlation with synthesized naturalistic noise

Machine Learning (BSMC-GA 4439) Wenke Liu

Optimization and Testing in Linear. Non-Gaussian Component Analysis

Using Kernel PCA for Initialisation of Variational Bayesian Nonlinear Blind Source Separation Method

Speed and Accuracy Enhancement of Linear ICA Techniques Using Rational Nonlinear Functions

ACENTRAL problem in neural-network research, as well

Apprentissage non supervisée

Autoregressive Independent Process Analysis with Missing Observations

NONLINEAR INDEPENDENT FACTOR ANALYSIS BY HIERARCHICAL MODELS

A Constrained EM Algorithm for Independent Component Analysis

Simple-Cell-Like Receptive Fields Maximize Temporal Coherence in Natural Video

Principal Component Analysis CS498

Estimating Overcomplete Independent Component Bases for Image Windows

Separation of the EEG Signal using Improved FastICA Based on Kurtosis Contrast Function

Package ProDenICA. February 19, 2015

Maximum variance formulation

Independent Component Analysis and Its Application on Accelerator Physics

Finding a causal ordering via independent component analysis

Discovery of Linear Acyclic Models Using Independent Component Analysis

MULTI-VARIATE/MODALITY IMAGE ANALYSIS

An Improved Cumulant Based Method for Independent Component Analysis

LECTURE :ICA. Rita Osadchy. Based on Lecture Notes by A. Ng

Independent Component Analysis

New Machine Learning Methods for Neuroimaging

Distinguishing Causes from Effects using Nonlinear Acyclic Causal Models

Hierarchical Sparse Bayesian Learning. Pierre Garrigues UC Berkeley

Mining Big Data Using Parsimonious Factor and Shrinkage Methods

Chapter 15 - BLIND SOURCE SEPARATION:

Basics of Multivariate Modelling and Data Analysis

Transcription:

Lecture'12:' SSMs;'Independent'Component'Analysis;' Canonical'Correla;on'Analysis' Lester'Mackey' May'7,'2014' ' Stats'306B:'Unsupervised'Learning'

Beyond'linearity'in'state'space'modeling' Credit:'Alex'Simma' Weather'forecas,ng''''''' Pose'es,ma,on' Dynamics implicitly determined by geophysical simulations Observed image is a complex function of the 3D pose, other nearby objects & clutter, lighting conditions, camera calibration, etc.

Approximate'nonlinear'filters' Credit:'Alex'Simma' Typically'cannot'directly'represent'these' con,nuous'filtering'distribu,ons'or'determine'a' closed'form'for'the'predic;on'integral' A'wide'range'of'approximate'nonlinear'filters'have' thus'been'proposed,'including'! Histogramfilters! Extended&unscentedKalmanfilters! Par7clefilters!

Approximate'nonlinear'filtering'methods' Credit:'Alex'Simma' Histogram Filter:! Evaluateonfixeddiscre7za7ongrid! Replacescon7nuouslatentvariablebydiscretevariable! Onlyfeasibleinlowdimensions! Expensiveorinaccurate Extended/Unscented Kalman Filter:! ApproximateposteriorasGaussianvialineariza7on, quadrature,! Inaccurateformul7modalposteriordistribu7ons Particle Filter:! Approximatefilteringdistribu7onwithfixed setofpar7cles(deltamasses)! Par7cleloca7onsnotfixed;dynamically evaluatestateswithhighestprobability! MonteCarloapproxima7on

Independent'component'analysis'(ICA)' Courtesy:'Rob'Tibshirani' X = AS where X is a random p-vector representing multivariate input measurements. S is a latent source p-vector whose components are independently distributed random variables. A is p p mixing matrix. Given realizations x 1,x 2,...,x N of X, thegoalsoficaareto Estimate A Estimate the source distributions S j f Sj,j=1,...,p.

The'cocktail'party'problem' Courtesy:'Rob'Tibshirani' In a room there are p independent sources of sound, and p microphones placed around the room hear different mixtures. Source Signals Measured Signals PCA Solution ICA Solution Here each of the x ij = x j (t i )andrecoveredsourcesarea time-series sampled uniformly at times t i.

Independent'vs.'uncorrelated' Courtesy:'Rob'Tibshirani' WoLOG can assume that E(S) = 0 and Cov(S) = I, and hence Cov(X) =Cov(AS) =AA T. Suppose X = AS with S unit variance, uncorrelated Let R be any orthogonal p p matrix. Then and Cov(S )=I X = AS = AR T RS = A S It is not enough to find uncorrelated variables, as they are not unique under rotations. Hence methods based on second order moments, like principal components and Gaussian factor analysis, cannot recover A. ICA uses independence, and non-gaussianity of S, torecovera e.g. higher order moments.

Independent'vs.'uncorrelated' Source S Data X PCA Solution ICA Solution Principal components are uncorrelated linear combinations of X, chosen to successively maximize variance. Independent components are also uncorrelated linear combinations of X, chosen to be as independent as possible. Courtesy:'Rob'Tibshirani'

ICA'representa;on'of'natural'images' Courtesy:'Rob'Tibshirani' Pixel blocks are treated as vectors, and then the collection of such vectors for an image forms an image database. ICA can lead to a sparse coding for the image, using a natural basis. see http://www.cis.hut.fi/projects/ica/imageica/(patrik Hoyer and Aapo Hyvärinen, Helsinki University of Technology)

ICA'and'electroencephalographic'(EEG)'data' 15'seconds'of'EEG'data'' 9'(of'100)'scalp'channels' 9'ICA'components'' Nearby'electrodes'record' nearly'iden;cal'mixtures' of'brain'and'nonxbrain' ac;vity;'ica'components' are'temporally'dis;nct.'' Pa;ent'blinking'(IC1/3)' IC12:'cardiac'pulse'' IC4/7'ac;vity'a[er'string' of'correct'responses' Colored'scalps'represent' ICA'unmixing'coefficients' a j 'as'heatmap,'showing' brain'or'scalp'loca;on'of' the'source.''

Approaches'to'ICA' Courtesy:'Rob'Tibshirani' ICA literature is HUGE. Recent book by Hyvärinen, Karhunen & Oja (Wiley, 2001) is a great source for learning about ICA, and some good computational tricks. Mutual Information and Entropy, maximizing non-gaussianity FastICA (HKO 2001), Infomax (Bell and Sejnowski, 1995) Likelihood methods ProdDenICA (Hastie +Tibshirani)- later Nonlinear decorrelation Y 1 independent Y 2 iff max g,f Corr[f(Y 1 ),g(y 2 )] = 0 (Hérault-Jutten, 1984), KernelICA (Bach and Jordan, 2001) Tensorial moment methods

Blackboard'discussion' " See'lecture'notes'