DIFFUSION NETWORKS, PRODUCT OF EXPERTS, AND FACTOR ANALYSIS. Tim K. Marks & Javier R. Movellan

Similar documents
Unsupervised Discovery of Nonlinear Structure Using Contrastive Backpropagation

Lecture 16 Deep Neural Generative Models

Probabilistic Modelling and Reasoning: Assignment Modelling the skills of Go players. Instructor: Dr. Amos Storkey Teaching Assistant: Matt Graham

CSC321 Lecture 20: Autoencoders

Hierarchy. Will Penny. 24th March Hierarchy. Will Penny. Linear Models. Convergence. Nonlinear Models. References

Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, Spis treści

Eigenfaces. Face Recognition Using Principal Components Analysis

Static and Dynamic Modelling of Materials Forging

arxiv: v1 [stat.ml] 10 Apr 2017

Lecture 7: Con3nuous Latent Variable Models

STA 414/2104: Lecture 8

Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation)

Pattern Recognition and Machine Learning

Tutorial on Principal Component Analysis

Contrastive Divergence

A graph contains a set of nodes (vertices) connected by links (edges or arcs)

Deriving Principal Component Analysis (PCA)

Expectation Maximization

Equivalence of Backpropagation and Contrastive Hebbian Learning in a Layered Network

STA 414/2104: Lecture 8

Viewpoint invariant face recognition using independent component analysis and attractor networks

Gaussian Process Approximations of Stochastic Differential Equations

Nonlinear reverse-correlation with synthesized naturalistic noise

Large-Scale Feature Learning with Spike-and-Slab Sparse Coding

ADAPTIVE LATERAL INHIBITION FOR NON-NEGATIVE ICA. Mark Plumbley

Basic Principles of Unsupervised and Unsupervised

STA 414/2104: Machine Learning

How to do backpropagation in a brain

Deep unsupervised learning

Eigenface-based facial recognition

Machine learning for pervasive systems Classification in high-dimensional spaces

Implementation of a Restricted Boltzmann Machine in a Spiking Neural Network

Speaker Representation and Verification Part II. by Vasileios Vasilakakis

Learning Energy-Based Models of High-Dimensional Data

A NEW VIEW OF ICA. G.E. Hinton, M. Welling, Y.W. Teh. S. K. Osindero

Introduction to Machine Learning

Higher Order Statistics

Unsupervised Neural Nets

Knowledge Extraction from DBNs for Images

Recursive Generalized Eigendecomposition for Independent Component Analysis

c Springer, Reprinted with permission.

The connection of dropout and Bayesian statistics

Nonparametric Bayesian Methods (Gaussian Processes)

Sum-Product Networks: A New Deep Architecture

Chapter 11. Stochastic Methods Rooted in Statistical Mechanics

Supplement to: Guidelines for constructing a confidence interval for the intra-class correlation coefficient (ICC)

Techniques for Dimensionality Reduction. PCA and Other Matrix Factorization Methods

CS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision)

Single Channel Signal Separation Using MAP-based Subspace Decomposition

Deep Learning Srihari. Deep Belief Nets. Sargur N. Srihari

Emergence of Phase- and Shift-Invariant Features by Decomposition of Natural Images into Independent Feature Subspaces

Connections between score matching, contrastive divergence, and pseudolikelihood for continuous-valued variables. Revised submission to IEEE TNN

Convolution and Pooling as an Infinitely Strong Prior

Unsupervised Learning

CS 4495 Computer Vision Principle Component Analysis

Principal Component Analysis (PCA)

Principal Component Analysis

UNSUPERVISED LEARNING

Dimensionality Reduction. CS57300 Data Mining Fall Instructor: Bruno Ribeiro

Reconnaissance d objetsd et vision artificielle

below, kernel PCA Eigenvectors, and linear combinations thereof. For the cases where the pre-image does exist, we can provide a means of constructing

Nonlinear Dimensionality Reduction

The wake-sleep algorithm for unsupervised neural networks

Deep Belief Networks are compact universal approximators

1 Hoeffding s Inequality

Deep Learning. What Is Deep Learning? The Rise of Deep Learning. Long History (in Hind Sight)

Bayesian Mixtures of Bernoulli Distributions

Class Mean Vector Component and Discriminant Analysis for Kernel Subspace Learning

Replicated Softmax: an Undirected Topic Model. Stephen Turner

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

Machine Learning (BSMC-GA 4439) Wenke Liu

Density Propagation for Continuous Temporal Chains Generative and Discriminative Models

CSC2535: Computation in Neural Networks Lecture 7: Variational Bayesian Learning & Model Selection

Greedy Layer-Wise Training of Deep Networks

TUTORIAL PART 1 Unsupervised Learning

Modeling human function learning with Gaussian processes

Natural Gradient Learning for Over- and Under-Complete Bases in ICA

STA 4273H: Sta-s-cal Machine Learning

Image Analysis. PCA and Eigenfaces

Cheng Soon Ong & Christian Walder. Canberra February June 2018

A732: Exercise #7 Maximum Likelihood

Lecture 24: Principal Component Analysis. Aykut Erdem May 2016 Hacettepe University

ECE 521. Lecture 11 (not on midterm material) 13 February K-means clustering, Dimensionality reduction

Reading Group on Deep Learning Session 4 Unsupervised Neural Networks

Dimensionality Reduction

System 1 (last lecture) : limited to rigidly structured shapes. System 2 : recognition of a class of varying shapes. Need to:

Kyle Reing University of Southern California April 18, 2018

Lecture: Face Recognition

STA414/2104. Lecture 11: Gaussian Processes. Department of Statistics

Instantaneous Measurement of Nonlocal Variables

Learning and Memory in Neural Networks

Variational Principal Components

Face Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition

Approximate Inference Part 1 of 2

Regression with Input-Dependent Noise: A Bayesian Treatment

Blind Machine Separation Te-Won Lee

Large-scale Collaborative Prediction Using a Nonparametric Random Effects Model

CS281 Section 4: Factor Analysis and PCA

Lecture 10: Dimension Reduction Techniques

COMP 551 Applied Machine Learning Lecture 13: Dimension reduction and feature selection

Transcription:

DIFFUSION NETWORKS PRODUCT OF EXPERTS AND FACTOR ANALYSIS Tim K Marks & Javier R Movellan University of California San Diego 9500 Gilman Drive La Jolla CA 92093-0515 ABSTRACT Hinton (in press) recently proposed a learning algorithm called contrastive divergence learning for a class of proailistic models called product of experts (PoE) Whereas in standard mixture models the eliefs of individual experts are averaged in PoEs the eliefs are multiplied together and then renormalized One advantage of this approach is that the comined eliefs can e much sharper than the individual eliefs of each expert It has een shown that a restricted version of the Boltzmann machine in which there are no lateral connections etween hidden units or etween oservation units is a PoE In this paper we generalize these results to diffusion networks a continuous-time continuous-state version of the Boltzmann machine We show that when the unit activation functions are linear this PoE architecture is equivalent to a factor analyzer This result suggests novel non-linear generalizations of factor analysis and independent component analysis that could e implemented using interactive neural circuitry 1 INTRODUCTION Hinton (in press) recently proposed a fast learning algorithm known as contrastive divergence learning for a class of models called product of experts (PoE) In standard mixture models the eliefs of individual experts are averaged ut in PoEs the eliefs are multiplied together and then renormalized ie (1) is the proaility of the oserved vector under expert and is the proaility assiged to when the opinions of all the experts are comined One advantage of this approach is that the comined eliefs are sharper than the individual eliefs of each expert potentially avoiding the prolems of standard mixture models when applied to high-dimensional prolems It can e shown that a restricted version of the Boltzmann machine in which hidden units and oservale units form a ipartite graph is in fact a PoE (Hinton in press) Indeed Hinton (in press) recently trained these restricted Boltzmann machines (RBMs) using contrastive divergence learning with very good results One prolem with Boltzmann machines is that they use inarystate units a representation which may not e optimal for continuous data such as images and sounds This paper concerns the diffusion network a continuoustime continuous-state version of the Boltzmann machine Like Boltzmman machines diffusion networks have idirectional connections which in this paper we assume to e symmetric If a diffusion network is given the same connectivity structure as an RBM the result will also e a PoE model The main result presented here is that when the activation functions are linear these restricted diffusion networks model the same class of oservale distriutions as factor analysis This is somewhat surprising given the fact that diffusion networks are feedack models while factor analyzers are feedforward models Most importantly the result suggests novel non-linear generalizations of factor analysis and independent component analysis that could e implemented using interactive circuitry We show results of simulations in which we trained restricted diffusion networks using contrastive divergence 2 DIFFUSION NETWORKS A diffusion network is a neural network specified y a stochastic differential equation (2) is a random vector descriing the state of the system at time and is a standard Brownian motion process The term is a constant known as the dispersion that controls the amount of noise in the system The function known as the drift represents the deterministic kernel of the system If the drift is the gradient of an energy function!#"%$'&)( and the energy function satisfies certain conditions (Gidas 1986) then it can e shown that has a limit proaility density * +-0/1 2"%3 ( 4 5 768 which is independent of the initial conditions 481

[ [ [ x a 21 Linear diffusions If we let 9;: 3 and the energy is a quadratic function ( 4!< =>5 3)?4@1A A is a symmetric positive definite matrix then the limit distriution is B +-0/1 2" = 3 @ A DC (3) This distriution is Gaussian with zero mean and covariance matrix AFEG Taking the gradient of the energy function we get 4 *#"%$ & ( H#"IA) (4) and thus the activation dynamics are linear: #"IA J (5) These dynamics have the following neural network interpretation: Let AKL "NMO M represents a matrix of synaptic conductances and is a diagonal matrix of transmemrane conductances )P (current leakage terms) We then get the following form QP *<RTS QP U" )PVQP?W B ) (6) QP S KX M P (7) is interpreted as the net input current to neuron Y and QP as the activation of that neuron 3 FACTOR ANALYSIS Factor analysis is a proailistic model of the form [ N\ _^ (8) is an `1a dimensional random vector representing oservale data; is an `1 dimensional Gaussian vector representing hidden independent sources c dfe!lk the identity matrix; and ^ is a Gaussian vec- tor independent of representing noise in the oservation generation process c m^o Qne m^o oqp a diagonal matrix If is logistic instead of Gaussian and in the limit as psrte Equation (8) is the standard model underlying independent component analysis (ICA) (Bell & Sejnowski 1995; Attias 1999) Note that in factor analysis as well as in ICA the oservations are generated in a feedforward manner from the hidden sources and the noise process It follows from Equation 8 that and have covariances vu dw yx Taking the inverse of this lock matrix we get \ \z\z@{_p \z@ k (9) p EG "Op E7G \ E7G yx "I\z@ p EG ko\z@up EG \ (10) 4 RESTRICTED DIFFUSION NETWORKS In this section we discuss a suclass of diffusion networks that have a PoE architecture To egin with we divide the state vector of a diffusion network into oservale and hidden units @N} [ @!~ @ It is easy to show that if the connectivity matrix is restricted so that there are no lateral connections from hidden units to hidden units and no lateral connections from oservale units to oservale units ie if "M A' a "M a @ (11) a and are diagonal then the limit distriution of is a PoE Hereafter we refer to a diffusion network with no hidden-hidden or oservale-oservale connections as a restricted diffusion network (RDN) Suppose we are given an aritrary linear RDN The limit distriution B [ has Gaussian marginal distriutions B [ and B whose covariances can e found y inverting the lock matrix in Equation 11: # " M a @ ae7g M a EG (12) [ * a " M a EG a E7G (13) Let ƒ a represent the set containing every distriution on U that is the limit distriution of the oservale units of any linear RDN with ` hidden units In some cases we can think of the hidden units of a network as the network s internal representation of the world Some authors such as Barlow (1994) have postulated that one of the organizing principles of early processing in the rain is for these internal representations to e as independent as possile Thus it is of interest to study the set of linear RDNs whose hidden units are independent ie for which is diagonal One might wonder for example whether restricting to e the identity matrix imposes any constraints on [ Let ƒ Pa represent the set containing every distriution on ˆ that is the limit distriution of the oservale units of any linear RDN with ` hidden units that are independent ˆ k We now show that forcing the hidden units to e independent does not constrain the class of oservale distriutions that can e represented y a linear RDN Theorem 1 : ƒ Pa ƒ a Proof : Since ƒ Pa is a suset of ƒ a we just need to show that any distriution in ƒ a is also in ƒ P a Let [ e the limit distriution of an aritrary linear RDN with parameters M a and a We will show there exists a new linear RDN with limit distriution [ˆŠ Š and parameters M Š a Š and a such that Š Nk and [ Š [ 482

First we take the eigenvalue decomposition k " E7GŒ 6 a E7G a M a EGŒ 6 N7 H @ (14) is a unitary matrix and is diagonal Equation 14 can e rewritten a E7G a M a " GŒ 6 Š M Š We define N EG Ž F a M a 7 H @ GŒ 6 E7GŒ 6 C (15) 7 E7GŒ 6 C (16) By Equation 12 (and after some derivations) we find that Š Š " M Š lk C a @ ae7g M Š a EG By Equation 13 (and after some derivations) we find that [ Š a " M Š a Š DE7G M Š a @ EG a " M a EG M%@ a E7G [ DC (17) (18) Theorem 1 will help to elucidate the relationship etween factor analysis and linear RDNs We will show that the class of distriutions over oservale variales that can e generated y factor analysis models with ` hidden sources is the same as the class of distriutions over oservale units that can e generated y linear RDNs with `1 hidden units 5 FACTOR ANALYSIS AND DIFFUSION NETS In this section we examine the relationship etween feedforward factor analysis models and feedack diffusion networks Let a represent the class of proaility distriutions over oservale units that can e generated y factor analysis models with ` hidden units In Theorem 2 we will show that this class of distriutions is equivalent to ƒ a the class of distriutions generated y linear RDNs To do so we first need to prove a Lemma regarding the suclass of factor analysis models for which the lower right lock of the matrix in Equation 10 kb \z@up E7G \ is diagonal We define Pa as the class of proaility distriutions on that can e generated y such restricted factor analysis models Lemma : Pa a Proof : Since Pa is a suset of a we just need to show that any distriution in a is also in P a Let [ e the distriution generated y an aritrary factor analysis model with parameters \ and p We will show that there exists a new factor analysis model with joint distriution [ˆŠ and parameters \ Š and p such that [ˆŠ H [ and such that kol \ Š @ p EG \ Š is diagonal First we take the eigenvalue decomposition \ @ p E7G \IK { @ (19) S is a unitary matrix and the eigenvalue matrix is diagonal Now define a rotation š y š l Ž F \ Š N\ š C (20) Then V\ Š @ p EG \ Š N (21) which is diagonal Thus k < V\ [ˆŠ Š 2@ p E7G \ is diagonal Furthermore from the equation 9\ Š ^ we can derive [ˆŠ l\ Š \ Š 2@œ pkn\z\@{ p [ DC Now we are ready to prove the main result of this paper: the fact that factor analysis models are equivalent to linear RDNs Theorem 2 : ƒ a a Proof : [ž : 'ŸQJƒ Ÿ Let B [ e the distriution over oservale units generated y any factor analysis model By the Lemma there exists a factor analysis model with distriution @ *B [ @*~ @B and parameters \j~dp such that its marginal distriution over the oservale units equals the given distriution B [ and for which k \@ p E7G \ is diagonal Let @ [ @ ~ @ e the limit distriution of the linear diffusion network with parameters M a a given y setting the connection matrix of Equation 11 equal to the inverse covariance matrix of Equation 10: A' EG C (22) Then 9p EG and a <k \z@ p E7G \ are oth diagonal so this diffusion net is a RDN By Equation 3 the limit distriution of is Gaussian with covariance la E7G from which it follows that ˆ In particular [ [ Thus [ ˆlB [ is the limit distriution of a linear RDN R W ƒ ŸQ Ÿ Let [ e the limit distriution over oservale units of any linear RDN By Theorem 1 there exists a linear RDN with distriution B @B JB [ @*~ @B and parameters M a a such that its marginal distriution over the oservale units equals the given distriution B [ and for which lk Consider the factor analysis model with parameters p a E7G ~ˆ\ aeg M a C Using reasoning similar to that used in the first part of this proof one can verify that this factor analysis model generates a distriution over oservale units that is equal to B [ 483

6 SIMULATIONS RDNs have continuous states rather than inary states so they can represent pixel intensities in a more natural manner than restricted Boltzmann machines (RBMs) can We used contrastive divergence learning rules to train a linear RDN with 10 hidden units in an unsupervised manner on a dataase of 48 face images (16 people in three different lighting conditions) from the MIT Face Dataase (Turk & Pentland 1991) The linear RDN had 3600 oservale units ( e ª e pixels) Each expert in the product of experts consists of one hidden unit and the learned connection weights etween that hidden unit and all of the oservale units These learned connection weights can e thought of as the receptive field of that hidden unit Figure 1 shows the receptive fields of the 10 hidden units after training Because of Fig 2: Eigenfaces calculated from the same dataase of 48 face images the unoccluded oservale units to the values «and let the network settle to equilirium The equilirium distriution of the network will then e the posterior distriution B [% ˆ [ «N «Using the linear RDN that was trained on face images we reconstructed occluded versions of the images in the training set y taking the mean of this posterior distriution Figure 3 shows the results of reconstructing two occluded face images Fig 1: Receptive fields of the 10 hidden units of a linear RDN that was trained on a dataase of 48 face images y minimizing contrastive divergence the correspondence etween linear RDNs and factor analysis that was proven in previous sections the receptive field of a hidden unit in the linear RDN should correspond to a column of the factor loading matrix \ in Equation 8 (after the column is normalized y premultiplying y p EG the inverse of the variance of the noise in each oservale dimension) Because of the relationship etween factor analysis and principal component analysis one would expect the receptive fields of the hidden units of the linear RDN to resemle the eigenfaces (Turk & Pentland 1991) which are the eigenvectors of the pixelwise covariance matrix of the same dataase We calculated the eigenfaces of the dataase The eigenfaces corresponding to the 10 largest eigenvalues are shown in Figure 2 The qualitative similarity etween the linear RDN receptive fields in Figure 1 and the eigenfaces in Figure 2 is readily apparent Partially occluded [ images correspond to splitting the oservale units into those [% [O«with known values and those with unknown values Given an image with known values «reconstructing the values of the occluded units involves finding the posterior distriution [O [O««The feedack architecture of diffusion networks provides a natural way to find such a posterior distriution: clamp Fig 3: Reconstruction of two occluded images Left: Image from the training set for the linear RDN Center: The oservale units corresponding to the visile pixels were clamped while the oservale units corresponding to the occluded pixels were allowed to run free Right: Reconstructed image shown is the mean of the equilirium distriution 7 DISCUSSION We proved the rather surprising result that a feedack model the linear RDN spans the same family of distriutions as factor analysis a feedworward model Furthermore we have shown that for this class of feedack models restricting the marginal distriution of the hidden units to e independent does not restrict the distriutions over the oservale units that they can generate Linear RDNs can e trained using the contrastive divergence learning method of Hinton (in press) which is relatively fast A main advan- 484

tage of feedack networks over feedforward models is that one can easily solve inference prolems such as pattern completion for occluded images using the same architecture and algorithm that are used for generation Most importantly diffusion networks can also e defined with nonlinear activation functions (Movellan Mineiro & Williams In Press) and we have egun to explore nonlinear extensions of the linear case outlined here Extensions of this work to nonlinear diffusion networks open the door to possiilities of new ICA-like algorithms ased on feedack rather than feedforward architectures References Attias H (1999) Independent Factor Analysis Neural Computation 11(4) 803 851 Barlow H (1994) What is the computational goal of the neocortex? In C Koch (Ed) Large scale neuronal theories of the rain pages 1 22 Camridge MA: MIT Press Bell A & Sejnowski T (1995) An informationmaximisation approach to lind separation and lind deconvolution Neural Computation 7 1129 1159 Gidas B (1986) Metropolis-type monte carlo simulation algorithms and simulated annealing In J L Snell (Ed) Topics in contermporary proaility and its applications pages 159 232 Boca Raton: CRC Press Hinton G E (in press) Training Products of Experts y Minimizing Contrastive Divergence Neural Computation Movellan J R Mineiro P & Williams R J (In Press) Partially Oservale SDE Models for Image sequence recognition tasks In T Leen T G Dietterich & V Tresp (Eds) Advances in Neural Information Processing Systems numer 13 pages 880 886 Camridge Massachusetts: MIT Press Turk M & Pentland A (1991) Eigenfaces for recognition Journal of Cognitive Neuroscience 3(1) 71 86 485