Kartik Audhkhasi, Osonde Osoba, Bart Kosko

Similar documents
Speaker Representation and Verification Part II. by Vasileios Vasilakakis

NOISE BENEFITS IN CONVOLUTIONAL NEURAL NETWORKS. Kartik Audhkhasi, Osonde Osoba, Bart Kosko

Lecture 16 Deep Neural Generative Models

A graph contains a set of nodes (vertices) connected by links (edges or arcs)

Deep Neural Networks

Restricted Boltzmann Machines

Deep unsupervised learning

Variational Autoencoder

Bidirectional Representation and Backpropagation Learning

Optimal Mean-Square Noise Benefits in Quantizer-Array Linear Estimation Ashok Patel and Bart Kosko

UNSUPERVISED LEARNING

The Noisy Expectation-Maximization Algorithm for Multiplicative Noise Injection

How to do backpropagation in a brain

Knowledge Extraction from DBNs for Images

Deep Learning. What Is Deep Learning? The Rise of Deep Learning. Long History (in Hind Sight)

Does the Wake-sleep Algorithm Produce Good Density Estimators?

Stochastic Gradient Estimate Variance in Contrastive Divergence and Persistent Contrastive Divergence

Pattern Recognition and Machine Learning

Artificial Neural Networks Examination, March 2004

Unsupervised Learning

Reading Group on Deep Learning Session 1

The Origin of Deep Learning. Lili Mou Jan, 2015

U-Likelihood and U-Updating Algorithms: Statistical Inference in Latent Variable Models

Deep Learning. What Is Deep Learning? The Rise of Deep Learning. Long History (in Hind Sight)

Weighted Finite-State Transducers in Computational Biology

Reading Group on Deep Learning Session 4 Unsupervised Neural Networks

Data Mining Part 5. Prediction

Chapter 11. Stochastic Methods Rooted in Statistical Mechanics

Greedy Layer-Wise Training of Deep Networks

Machine Learning Lecture 5

Contrastive Divergence

4. Multilayer Perceptrons

Neural Networks and the Back-propagation Algorithm

Basic Principles of Unsupervised and Unsupervised

Lecture 3: Pattern Classification

Deep Generative Models. (Unsupervised Learning)

Parallel Tempering is Efficient for Learning Restricted Boltzmann Machines

The Noisy Expectation-Maximization Algorithm for Multiplicative Noise Injection

Artificial Neural Networks

arxiv: v1 [cs.ne] 6 May 2014

COMP9444 Neural Networks and Deep Learning 11. Boltzmann Machines. COMP9444 c Alan Blair, 2017

arxiv: v2 [cs.ne] 22 Feb 2013

Lecture 5: Logistic Regression. Neural Networks

Latent Variable Models and EM algorithm

Using Noise to Speed Up Video Classification with Recurrent Backpropagation

A Note on the Expectation-Maximization (EM) Algorithm

ECE521 Lectures 9 Fully Connected Neural Networks

Lecture 7 Artificial neural networks: Supervised learning

Connections between score matching, contrastive divergence, and pseudolikelihood for continuous-valued variables. Revised submission to IEEE TNN

Deep Learning & Neural Networks Lecture 2

Mark Gales October y (x) x 1. x 2 y (x) Inputs. Outputs. x d. y (x) Second Output layer layer. layer.

Linear & nonlinear classifiers

Clustering by Mixture Models. General background on clustering Example method: k-means Mixture model based clustering Model estimation

Measuring the Usefulness of Hidden Units in Boltzmann Machines with Mutual Information

Introduction to Neural Networks

Variational Inference (11/04/13)

Learning and Memory in Neural Networks

Density estimation. Computing, and avoiding, partition functions. Iain Murray

Speech Recognition Lecture 8: Expectation-Maximization Algorithm, Hidden Markov Models.

A Variance Modeling Framework Based on Variational Autoencoders for Speech Enhancement

Machine Learning. Neural Networks

Energy Based Models. Stefano Ermon, Aditya Grover. Stanford University. Lecture 13

Artificial Neural Networks

Introduction to Neural Networks

Artificial Neural Networks Examination, June 2005

Linear Models for Classification

Clustering K-means. Clustering images. Machine Learning CSE546 Carlos Guestrin University of Washington. November 4, 2014.

Unsupervised Learning

Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, Spis treści

PATTERN CLASSIFICATION

The connection of dropout and Bayesian statistics

Plan. Perceptron Linear discriminant. Associative memories Hopfield networks Chaotic networks. Multilayer perceptron Backpropagation

A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models

STA 414/2104: Lecture 8

Need for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels

Using Expectation-Maximization for Reinforcement Learning

STA 414/2104: Lecture 8

Neural networks. Chapter 20. Chapter 20 1

Expectation Propagation Algorithm

Deep Feedforward Networks

Sequence labeling. Taking collective a set of interrelated instances x 1,, x T and jointly labeling them

From perceptrons to word embeddings. Simon Šuster University of Groningen

Lecture 4: Perceptrons and Multilayer Perceptrons

COM336: Neural Computing

Logistic Regression Review Fall 2012 Recitation. September 25, 2012 TA: Selen Uguroglu

Learning Deep Architectures for AI. Part II - Vijay Chakilam

STA 4273H: Statistical Machine Learning

Implementation of a Restricted Boltzmann Machine in a Spiking Neural Network

Deep Boltzmann Machines

Course 395: Machine Learning - Lectures

ECE 521. Lecture 11 (not on midterm material) 13 February K-means clustering, Dimensionality reduction

Index. Santanu Pattanayak 2017 S. Pattanayak, Pro Deep Learning with TensorFlow,

Bayesian Inference Course, WTCN, UCL, March 2013

Recent Advances in Bayesian Inference Techniques

Generative models for missing value completion

Neural Networks. Bishop PRML Ch. 5. Alireza Ghane. Feed-forward Networks Network Training Error Backpropagation Applications

13: Variational inference II

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013

Artificial Neural Networks. MGS Lecture 2

Transcription:

To appear: 2013 International Joint Conference on Neural Networks (IJCNN-2013 Noise Benefits in Backpropagation and Deep Bidirectional Pre-training Kartik Audhkhasi, Osonde Osoba, Bart Kosko Abstract We prove that noise can speed convergence in the backpropagation algorithm. The proof consists of two separate results. The first result proves that the backpropagation algorithm is a special case of the generalized Expectation- Maximization (EM algorithm for iterative maximum likelihood estimation. The second result uses the recent EM noise benefit to derive a sufficient condition for backpropagation training. The noise adds directly to the training data. A noise benefit also applies to the deep bidirectional pre-training of the neural network as well as to the backpropagation training of the network. The geometry of the noise benefit depends on the probability structure of the neurons at each layer. Logistic sigmoidal neurons produce a forbidden noise region that lies below a hyperplane. Then all noise on or above the hyperplane can only speed convergence of the neural network. The forbidden noise region is a sphere if the neurons have a Gaussian signal or activation function. These noise benefits all follow from the general noise benefit of the EM algorithm. Monte Carlo sample means estimate the population expectations in the EM algorithm. We demonstrate the noise benefits using MNIST digit classification. Index Terms Backpropagation, bidirectional associative memory, noise benefit, neural network, stochastic resonance, Expectation-Maximization algorithm I. NOISE BENEFITS IN BACKPROPAGATION We prove for the first time that noise can speed convergence in the popular backpropagation gradient-descent algorithm for training feedforward multilayer-perceptron neural networks [1], [2]. The proof casts backpropagation in terms of maximum likelihood estimation [3] and then shows that the iterative backpropagation algorithm is a special case of the general Expectation-Maximization (EM algorithm. We then invoke the new noisy EM (NEM theorem [4] [6] that gives a sufficient condition for speeding convergence in the EM algorithm. Then the NEM result speeds convergence in the backpropagation algorithm. Figure 1 shows the noise benefit for squared-error training. Figure 4 shows a similar figure for cross-entropy training. The NEM version displays a 5.3% median decrease in squared error and a 4.2% median decrease in cross entropy per iteration compared with noiseless backpropagation training. We also observed that noise improves the recognition accuracy in both cases. There was little difference between the backpropagation and EM convergence rates. We further show that a related NEM result holds for the pre-training of the individual layers of neurons in the multilayer perceptron. These so-called restricted Boltzmann Kartik Audhkhasi, Osonde Osoba,and Bart Kosko are with the Signal and Image Processing Institute, Electrical Engineering Department, University of Southern California, Los Angeles, CA (email: kosko@sipi.usc.edu machine (RBM [7] [9] layers are in fact simple bidirectional associative memories (BAMs [1], [10], [11] that undergo synchronous updating of the neurons. They are BAMs because the neurons in contiguous layers use a connection matrix M in the forward pass and the corresponding matrix transpose M T in the backward pass and because the neurons have no within-layer connections. The general BAM convergence theorem [1], [10], [11] guarantees that all such rectangular matrices M are bidirectionally stable for synchronous or asynchronous neuron updates and for quite general neuronal activation nonlinearities because the RBM energy function is a Lyapunov function for the BAM network. This gives almost immediate convergence to a BAM fixed point after only a small number of synchronous backand-forth updates when both layers use logistic neurons. The NEM Theorem gives a type of forbidden condition [4] [6], [12], [13] that ensures a noise speed up so long as the noise lies outside of a specified region in the noise state space. Figures 2 and 3 show that the noise must lie outside such regions to speed convergence. The neuron probability distribution function (pdf and network connection or synaptic weights control the geometry of this forbidden region. Logistic neurons give the forbidden region as a half-space while Gaussian neurons give it as a sphere. Theorems 3 and 4 give the sufficient conditions for a noise benefit in the popular cases of neural networks with logistic and Gaussian output neurons. Theorems 5 and 6 give similar sufficient conditions for Bernoulli-Bernoulli and Gaussian-Bernoulli BAMs. This is a type of stochastic resonance effect where a small amount of noise improves the performance of a nonlinear system while too much noise harms the system [12] [21]. Some prior research has found an approximate regularizing effect of adding white noise to backpropagation [22] [25]. We instead add non-white noise that satisfies a simple geometric condition that depends on both the network parameters and the output activations. The next section casts the backpropagation algorithm as ML estimation. Section III presents the EM algorithm for neural network training and proves that it reduces to the backpropagation algorithm. Section IV reviews the NEM theorem. Section V proves the noise benefit sufficient conditions for a neural network. Section VI reviews RBMs or BAMs. Section VII derives sufficient conditions for a noise benefit in ML training of Bernoulli-Bernoulli and Gaussian-Bernoulli RBMs. Section VIII presents simulation results. 1

x 10 5 5.6 Training set squared error 5.4 5.2 5 4.8 4.6 4.4 BP NEM BP 4.2 4 2 3 4 5 6 7 8 9 10 Iteration number Fig. 1. This figure shows the training-set squared error for backpropagation and NEM-backpropagation (NEM-BP training of an auto-encoder neural network on the MNIST digit classification data set. There is a 5.3% median decrease in the squared error per iteration for NEM-BP when compared with backpropagation training. We added annealed independent and identicallydistributed (i.i.d. Gaussian noise to the target variables. The noise had mean a t t and a variance that decayed with the training epochs as 0.1,0.1/2, 0.1/3,... where a t is the vector of activations of the output layer and t is the vector of target values. The network used three logistic (sigmoidal hidden layers with 20 neurons each. The output layer used 784 logistic neurons. Fig. 2. Noise benefit region for a neural network with Bernoulli (logistic output neurons: Noise speeds up maximum-likelihood parameter estimation of the neural network with Bernoulli output neurons if the noise lies above a hyperplane that passes through the origin of the noise space. The activation signal a t of the output layer controls the normal to the hyperplane. The hyperplane changes as learning proceeds because the parameters and hidden layer neuron activations change. We used independent and identically distributed (i.i.d. Gaussian noise with mean 0, variance 3, and (3, 1, 1 as the normal to the hyperplane. II. BACKPROPAGATION AS MAXIMUM LIKELIHOOD ESTIMATION We show that the backpropagation algorithm performs ML estimation of a neural network s parameters. We use a 3-layer neural network for notational convenience. All results in this paper extend to deep networks with more hidden layers. x denotes the neuron values at the input layer that consists of I neurons. a h is the vector of hidden neuron sigmoidal activations whose j th element is a h 1 ( j ( 1 + exp σ w ji x i I w jix i where w ji is the weight of the link that connects the i th visible and j th hidden neuron. y denotes the K-valued target variable and t is its 1-in-K encoding. t k is the k th output neuron s value with activation ( J exp a t k j1 u kja h j ( K k exp J (2 11 j1 u k 1ja h j (1 p(y k x, Θ (3 where u kj is the weight of the link that connects the j th hidden and k th target neuron. a t k depends on input x and parameter matrices U and W. Backpropagation minimizes the following cross entropy: E t k log(a t k. (4 The cross-entropy equals the negative conditional loglikelihood L of the target y given the inputs x because [ K E log ] [ K (a t k t k log ] p(y k x, Θ t k (5 log p(y x, Θ L. (6 Backpropagation updates the network parameters Θ using gradient ascent to maximize the log likelihood log p(y x, Θ. The partial derivative of this log-likelihood with respect to u kj is and with respect to w ji is L u kj (t k a t ka h j (7 L w ji a h j (1 ah j x i (t k a t k u kj. (8 So (7 and (8 give the partial derivatives that perform gradient ascent on the log-likelihood L. A linear signal function often replaces the Gibbs function at the output layer for regression: a t k u kj a h j. (9 j1 The target values t of the output neuron layer can assume any real values for regression. Then backpropagation minimizes

Training set squared error 11500 11000 10500 10000 9500 9000 8500 8000 7500 BP NEM BP Fig. 3. Noise benefit region for a neural network with Gaussian output neurons: Noise speeds up maximum-likelihood parameter estimation of the neural network with Gaussian output neurons if the noise lies inside a hypersphere. The activation signal a t of the output layer and the target signal t control the center and radius of this hypersphere. This hypersphere changes as learning proceeds because the parameters and hidden-layer neuron activations change. We used i.i.d. Gaussian noise with mean 0, variance 3, and center t a t (1, 1,1. the following squared-error function: E 1 2 (t k a t k 2. (10 We assume that t is Gaussian with mean a t and identity covariance matrix I. Then backpropagation maximizes the following log-likelihood function: for N(t;a t,i L log p(t x, Θ log N(t;a t,i (11 1 exp 1 (2π d/2 2 (t k a t k 2. (12 Thus the gradient partial derivatives of this log-likelihood function are the same as those for the K-class classification case in (7 and (8. III. EM ALGORITHM FOR NEURAL NETWORK ML ESTIMATION Both backpropagation and the EM algorithm find the ML estimate of a neural network s parameters. The next theorem shows that backpropagation is a special case of the generalized EM algorithm. Theorem 1. Backpropagation is the GEM Algorithm The backpropagation update equation for a differentiable likelihood function p(y x, Θ at epoch n Θ n+1 Θ n + η Θ log p(y x, Θ (13 equals the GEM update equation at epoch n Θ n+1 Θ n + η Θ Q(Θ Θ n ΘΘ n ΘΘ n (14 7000 2 3 4 5 6 7 8 9 10 Iteration number Fig. 4. This figure shows the training-set cross entropy for backpropagation and NEM-BP training of a 10-class classification neural network on the MNIST digit classification data set. There is a 4.2% median decrease in the cross entropy per iteration for NEM-BP when compared with backpropagation training. We added annealed i.i.d. Gaussian noise to the target variables. The noise had mean 0 and a variance that decayed with training epochs as 0.2, 0.2/2, 0.2/3,.... The network used three logistic (sigmoidal hidden layers with 20 neurons each. The output layer used 10 neurons with the Gibbs activation function in (2. where GEM uses the differentiable Q-function Q(Θ Θ n E p(h x,y,θn log p(y,h x, Θ. (15 Proof: We know that [3], [26] log p(y x, Θ Q(Θ Θ n + H(Θ Θ n (16 if H(Θ Θ n is the following cross entropy [27]: H(Θ Θ n lnp(h x, y, Θ dp(h x, y, Θ n. (17 Hence H(Θ Θ n log p(y x, Θ Q(Θ Θ n. (18 Now expand the Kullback-Leibler divergence [28]: ( D KL (Θ n p(h x, y, Θ n Θ ln dp(h x, y, Θ n p(h x, y, Θ (19 lnp(h x, y, Θ n dp(h x, y, Θ n lnp(h x, y, Θ dp(h x, y, Θ n (20 H(Θ n Θ n + H(Θ Θ n. (21 So H(Θ Θ n H(Θ n Θ n for all Θ because D KL (Θ n Θ 0 [28]. Thus Θ n minimizes H(Θ Θ n and hence Θ H(Θ Θ n 0 at Θ Θ n. Putting this in (18 gives Θ log p(y x, Θ ΘQ(Θ Θ n. (22 ΘΘ n ΘΘ n Hence the backpropagation and GEM update equations are identical.

The GEM algorithm uses a probabilistic description of the hidden layer neurons. We assume that the hidden layer neurons are Bernoulli random variables. So their activation is the following conditional probability: a h j p(h j 1 x, Θ. (23 We can now formulate an EM algorithm for ML estimation of a feedforward neural network s parameters. The E-step computes the Q-function in (15. Computing the expectation in (15 requires 2 J values of p(h x, y, Θ n. This is computationally-intensive for large values of J. So we use Monte Carlo sampling to approximate the above Q-function. The strong law of large numbers ensures that this Monte Carlo approximation converges almost surely to the true Q- function. Bayes theorem gives p(h x, y, Θ n as p(h x, y, Θ n p(h x, Θn p(y h, Θ n h p(h x, Θn p(y h, Θ n. (24 We can sample more easily from p(h x, Θ n than from p(h x, y, Θ n because the h j terms are independent given x. We replace p(h x, Θ n by its Monte Carlo approximation using M independent and identically-distributed (i.i.d. samples: p(h x, Θ n 1 M δ K (h h m (25 where δ K is the J-dimensional Kronecker delta function. The Monte Carlo approximation of the hidden data conditional PDF becomes M p(h x, y, Θ n δ K(h h m p(y h, Θ n M h m δ 11 K(h h m1 p(y h, Θ n (26 M δ K(h h m p(y h m, Θ n M (27 m 11, Θ p(y hm1 n δ K (h h m γ m (28 where γ m p(y h m, Θ n M m 11 p(y hm1, Θ n (29 is the importance of h m. (28 gives an importance-sampled approximation of p(h x, y, Θ n where each sample h m has weight γ m. We can now approximate the Q-function as Q(Θ Θ n h γ m δ K (h h m log p(y,h x, Θ (30 γ m log p(y,h m x, Θ (31 γ m[ ] log p(h m x, Θ + log p(y h m, Θ (32 where log p(h m x, Θ j1 [ ] h m j log a h j + (1 h m j log(1 a h j (33 for sigmoidal hidden layer neurons. Gibbs activation neurons at the output layer give log p(y h m, Θ t k log a mt k (34 where a h j is as in (1 and a mt k ( J j1 u kja mh j exp ( K k exp J 11 j1 u k 1ja mh j. (35 Gaussian output layer neurons give log p(y h m, Θ 1 (t k a mt k 2 2. (36 The Q-function in (32 equals a sum of log-likelihood functions for two 2-layer neural networks between the visible-hidden and hidden-output layers. The M-step maximizes this Q-function by gradient ascent. So it is equivalent to two distinct backpropagation steps performed on these two 2-layer neural networks. IV. THE NOISY EXPECTATION-MAXIMIZATION THEOREM The Noisy Expectation-Maximization (NEM algorithm [4], [5] modifies the EM scheme and achieves faster convergence times on average. The NEM algorithm injects additive noise into the data at each EM iteration. The noise decays with the iteration count to guarantee convergence to the optimal parameters of the original data model. The additive noise must also satisfy the NEM condition below that guarantees that the NEM parameter estimates will climb faster up the likelihood surface on average. A. NEM Theorem The NEM Theorem [4], [5] states a general sufficient condition when noise speeds up the EM algorithm s convergence to a local optimum. The NEM Theorem uses the following notation. The noise random variable N has pdf p(n x. So the noise N can depend on the data x. h are the latent variables in the model. Θ (n is a sequence of EM estimates for Θ. Θ lim n Θ (n is the converged EM estimate for Θ. Define the noisy Q function Q N (Θ Θ (n E h x,θk [lnp(x + N,h θ]. Assume that the differential entropy of all random variables is finite and that the additive noise keeps the data in the likelihood function s support. Then we can state the NEM theorem [4], [5]. Theorem 2. Noisy Expectation Maximization (NEM The EM estimation iteration noise benefit Q(Θ Θ Q(Θ (n Θ Q(Θ Θ Q N (Θ (n Θ (37

or equivalently Q N (Θ (n Θ Q(Θ (n Θ (38 holds on average if the following positivity condition holds: [ ( ] p(x + N,h Θk E x,h,n Θ ln 0. (39 p(x,h Θ k The NEM Theorem states that each iteration of a suitably noisy EM algorithm gives higher likelihood estimates on average than do the regular EM s estimates. So the NEM algorithm converges faster than EM if we can identify the data model. The faster NEM convergence occurs both because the likelihood function has an upper bound and because the NEM algorithm takes larger average steps up the likelihood surface. Maximum A Posteriori (MAP estimation for missing information problems can use a modified version of the EM algorithm. The MAP version modifies the Q-function by adding a log prior term G(Θ lnp(θ [29], [30]: Q(Θ Θ t E h x,θt [lnp(x,h Θ] + G(Θ. (40 The MAP version of the NEM algorithm applies a similar modification to the Q N -function: Q N (Θ Θ t E h x,θt [lnp(x + N,h Θ] + G(Θ. (41 Many latent-variable models are not identifiable [31] and thus need not have global optima. These models include Gaussian mixture models [32], hidden Markov models [33], and neural networks. The EM and NEM algorithms converge to local optima in these cases. The additive noise in the NEM algorithm helps the NEM estimates search other nearby local optima. The NEM Theorem still guarantees that NEM estimates have higher likelihood on average than EM estimates for non-identifiable models. V. NOISE BENEFITS IN NEURAL NETWORK ML ESTIMATION Consider adding noise n to the 1-in-K encoding t of the target variable y. We first present the noise benefit sufficient condition for Gibbs activation output neurons used in K-class classification. Theorem 3. Forbidden Hyperplane Noise Benefit Condition The NEM positivity condition holds for ML training of feedforward neural network with Gibbs activation output neurons if n T log(a t 0. (42 E t,h,n x,θ Proof: We add noise to the target 1-in-K encoding t. The likelihood ratio in the NEM sufficient condition becomes p(t + n,h x, Θ p(t + n h, Θp(h x, Θ (43 p(t,h x, Θ p(t h, Θp(h x, Θ p(t + n h, Θ (44 p(t h, Θ K (a t k t k+n k K (a t (a t k t k n k. (45 k So the NEM positivity condition becomes K E t,h,n x,θ log (a t k n k 0. (46 This condition is equivalent to E t,h,n x,θ K n k log(a t k 0. (47 We can rewrite this positivity condition as the following matrix inequality: E t,h,n x,θ n T log(a t 0 (48 where log(a t is the vector of output neuron log-activations. The above sufficient condition requires that the noise n lie above a hyperplane with normal log(a t. The next theorem gives a sufficient condition for a noise benefit in the case of Gaussian output neurons. Theorem 4. Forbidden Sphere Noise Benefit Condition The NEM positivity condition holds for ML training of a feedforward neural network with Gaussian output neurons if n E t,h,n,x,θ a t + t 2 a t t 2 0 (49 where. is the L 2 vector norm. Proof: We add noise n to the K output neuron values t. The log-likelihood in the NEM sufficient condition becomes p(t + n,h x, Θ p(t,h x, Θ p(t + n h, Θp(h x, Θ p(t h, Θp(h x, Θ N(t + n;at,i N(t;a t,i ( 1 [ t exp a t 2 t + n a t 2] 2 So the NEM sufficient condition becomes (50 (51. (52 n E t,h,n,xθ a t + t 2 a t t 2 0. (53 The above sufficient condition defines a forbidden noise region outside a sphere with center t a t and radius t a t. All noise inside this sphere speeds convergence of ML estimation in the neural network on average. This section presented sufficient conditions for a noise benefit in training a neural network that uses the EM algorithm. We now discuss noise benefit in pre-training or initialization of the parameters of a neural network using RBMs or the equivalent BAMs. VI. TRAINING BAMS OR RESTRICTED BOLTZMANN MACHINES Restricted Boltzmann Machines [7], [8] are a special type of bidirectional associative memory (BAM [1], [10], [11]. So they enjoy rapid convergence to a bidirectional fixed point for synchronous updating of the neurons. A BAM is a two-layer heteroassociative network that uses the synaptic

connection matrix M on the forward pass of the neuronal signals from the lower layer to the higher layer but also uses the transpose matrix M T on the backward from the higher layer to the lower layer. The lower layer is visible during training of deep neural networks [8] while the higher field is hidden. The general BAM Theorem ensures that any such matrix M is bidirectionally stable for threshold neurons as well for most continuous neurons. Logistic neurons satisfy the BAM Theorem because logistic signal functions are bounded and monotone decreasing. The following results use the term RBM instead of BAM for simplicity. Consider an RBM with I visible neurons and J hidden neurons. Let x i and h j denote the values of the i th visible and j th hidden neuron. Let E(x,h; Θ be the energy function for the network. Then the joint probability density function of x and h is the Gibbs distribution: ( exp E(x,h; Θ p(x,h; Θ (54 Z(Θ where Z(Θ exp( E(x,h; Θ. (55 x h Integrals replace sums for continuous variables in the above partition function Z(Θ. The Gibbs energy function E(v, h; Θ depends on the type of RBM. A Bernoulli(visible-Bernoulli(hidden RBM has logistic conditional PDFs at the hidden and visible layers and has the following BAM energy or Lyapunov function [1], [10], [11]: E(x,h; Θ j1 w ij x i h j b i x i a j h j j1 (56 where w ij is the weight of the connection between the i th visible and j th hidden neuron, b i is the bias for the i th visible neuron, and a j is the bias for the j th hidden neuron. A Gaussian(visible-Bernoulli(hidden RBM has Gaussian conditional PDFs at the visible layer, logistic conditional PDFs at the hidden layer, and the energy function [8], [34] E(x,h; Θ j1 w ij x i h j + 1 2 (x i b i 2 a j h j. (57 j1 The neural network uses an RBM or BAM as a building block. The systems finds ML estimates of each RBM s parameters and then stacks up the resulting RBMs on top of each other. Then backpropagation trains the neural network. The next section reviews ML training for an RBM. A. ML Training for RBMs using Contrastive Divergence The maximum likelihood (ML estimate of the parameters Θ for a RBM is Θ argmax log p(x; Θ. (58 Θ Gradient ascent can iteratively solve this optimization problem. We estimate w ij in the quadratic forms in (53 and (54 because the terms are the same for a Bernoulli-Bernoulli and Gaussian-Bernoulli RBM. The gradient of log p(x; Θ with respect to w ij is log p(x; Θ w ij E p(h x,θ x i h j E p(x,h Θ x i h j. (59 So the update rule for w ij at iteration (n + 1 becomes w (n+1 ij w (n ij ( + η E p(h x,θ (n x i h j E p(x,h Θ (n x i h j (60 where η > 0 is the learning rate. We can easily compute p(h x, Θ (n for the RBM because there are no connections between any two hidden or two visible neurons. This gives the expectation E p(h x,θ (n x i h j. But we cannot so easily compute p(x,h Θ (n due to the partition function Z(Θ in (55. Constrastive divergence (CD [8] approximates Z(Θ through activations that derive from a forward and a backward pass in the RBM. B. ML Training for RBMs using the EM algorithm The EM algorithm provides an iterative method for learning RBM parameters. Consider an RBM with I visible neurons and J hidden neurons. EM maximizes a simpler lower bound on log p(x; Θ because the log-likelihood can be intractable to compute. This lower bound at Θ Θ (n is Q(Θ Θ (n E h x,θ (nlogp(x,h; Θ (61 E h x,θ (n E(x,h; Θ log Z(Θ. (62 A generalized EM (GEM algorithm uses gradient descent to iteratively maximize the above Q-function. The gradient with respect to w ij is Q ( Θ Θ (n E h x,θ(n E(x,h; Θ log Z(Θ w ij w ij (63 E(x,h; Θ log Z(Θ E h x,θ (n (64 w ij w ij E h x,θ (n x i h j 1 Z(Θ. (65 Z(Θ w ij

But the partition funtion term expands as 1 Z(Θ 1 x h exp( E(x,h; Θ Z(Θ w ij Z(Θ w ij (66 1 exp( E(x,h; Θ (67 Z(Θ w x ij h 1 exp( E(x,h; Θx i h j (68 Z(Θ x h exp( E(x,h; Θ x i h j (69 Z(Θ x h p(x,h Θx i h j E x,h Θ x i h j. (70 x h So the partial derivative of the Q-function becomes Q ( Θ Θ (n E h x,θ (n x i h j E x,h Θ x i h j w ij (71 E h x,θ (nx i h j E x,h Θ x i h j. (72 This leads to the key GEM gradient ascent equation: ( w (n+1 ij w (n ij + η E p(h x,θ (n x i h j E p(x,h Θ x (n i h j. (73 The above update equation is the same as the contrastive divergence update equation in (60. The next section shows that this equivalence between CD and GEM lets us to derive a NEM sufficient condition for RBM training. VII. NOISE BENEFITS IN RBM ML ESTIMATION Consider now addition of noise n to the input data x. A NEM noise benefit exists in the RBM if p(x + n,h; Θ E x,h,n Θ log 0. (74 p(x,h; Θ The noisy complete data likelihood is p(x + n,h; Θ exp( E(x + n,h; Θ Z(Θ. (75 So a NEM noise benefit for an RBM occurs if exp( E(x + n,h; Θ E x,h,n Θ log 0. (76 exp( E(x, h; Θ This is equivalent to the RBM noise benefit inequality: E(x + n,h; Θ + E(x,h; Θ 0. (77 E x,h,n Θ The theorem below states a simple hyperplane separation condition that guarantees a noise benefit in the Bernoulli- Bernoulli RBM. Theorem 5. Forbidden Hyperplane Noise Benefit Condition The NEM positivity condition holds for Bernoulli-Bernoulli RBM training if E x,h,n Θ n T (Wh + b 0. (78 Proof: The noise benefit for a Bernoulli(visible- Bernoulli(hidden RBM results if we apply the energy function from (56 to the expectation in (77 to get E x,h,n Θ j1 w ij n i h j + The term in brackets is equivalent to j1 w ij n i h j + n i b i 0. (79 n i b i n T (Wh + b. (80 So the noise benefit sufficient condition becomes n T (Wh + b 0. (81 E x,h,n Θ The above sufficient condition is similar to the hyperplane condition for neural network training in Theorem 3. All noise above a hyperplane based on the RBM s parameters gives a noise benefit. The next theorem states a spherical separation condition that guarantees a noise benefit in the Bernoulli- Bernoulli RBM. Theorem 6. Forbidden Sphere Noise Benefit Condition The NEM positivity condition holds for Gaussian-Bernoulli RBM training if 1 E x,h,n Θ 2 n 2 n T (Wh + b x 0. (82 Proof: Putting the energy function in (57 into (77 gives the noise benefit condition for a Gaussian(visible- Bernoulli(hidden: E x,h,n Θ j1 w ij n i h j + n i x i 0. n i b i 1 2 n 2 i (83 The term in brackets equals the following matrix expression: j1 w ij n i h j + n i b i 1 2 n 2 i n i x i n T (Wh + b x 1 2 n 2. (84 So the noise benefit sufficient condition becomes 1 E x,h,n Θ 2 n 2 n T (Wh + b x 0. (85 The above condition bisects the noise space. But the bisecting surface for (82 is a hypersphere. This condition is also similar to the noise benefit condition for neural network ML training in Theorem 4.

VIII. SIMULATION RESULTS We modified the Matlab code available in [35] to inject noise during EM-backpropagation training of a neural network. We used 10,000 training instances from the training set of the MNIST digit classification data set. Each image in the data set had 28 28 pixels with each pixel value lying between 0 and 1. We fed each pixel into the input neuron of a neural network. We used a 5-layer neural network with 20 neurons in each of the three hidden layers and 10 neurons in the output layer for classifying the 10 digits. We also trained an auto-encoder neural network with 20 neurons in each of the three hidden layers and 784 neurons in the output layer for estimating the pixels of a digit s image. The output layer used the Gibbs activation function for the 10-class classification network and logistic activation function for the auto-encoder. We used logistic activation functions in all other layers. Simulations used 10 Monte Carlo samples for approximating the Q-function in the 10- class classification network and 5 Monte Carlo samples for the auto-encoder. Figure 1 shows the training-set squared error for the auto-encoder neural network for backpropagation and NEM-backpropagation when we added annealed Gaussian noise with mean a t t and variance 0.1 epoch 1. Figure 4 shows the training-set cross entropy for the two cases when we added annealed Gaussian noise with mean 0 and variance 0.2 epoch 1. We used 10 Monte Carlo samples to approximate the Q-function. We observed a 5.3% median decrease in squared error and 4.2% median decrease in cross entropy per iteration for the NEM-backpropagation algorithm compared to standard backpropagation. IX. CONCLUSIONS The backpropagation algorithm is a special case of the generalized EM algorithm. So proper noise injection speeds backpropagation convergence because it speeds EM convergence. These sufficient conditions use the recent noisy EM (NEM theorem. Similar sufficient conditions hold for a noise benefit in pre-training neural networks based on the NEM theorem. Noise-injection simulations on the MNIST digit recognition data set reduced both network squared error and cross entropy. REFERENCES [1] B. Kosko, Neural networks and fuzzy systems: A dynamical systems approach to machine intelligence. Prentice Hall, 1991. [2] S. Haykin, Neural networks: A comprehensive foundation. Prentice Hall, 1998. [3] C. M. Bishop, Pattern recognition and machine learning. Springer New York, 2006, vol. 4, no. 4. [4] O. Osoba, S. Mitaim, and B. Kosko, Noise Benefits in the Expectation-Maximization Algorithm: NEM theorems and Models, in The International Joint Conference on Neural Networks (IJCNN. IEEE, 2011, pp. 3178 3183. [5], The Noisy Expectation-Maximization Algorithm, in review, 2013. [6] O. Osoba and B. Kosko, Noise-Enhanced Clustering and Competitive Learning Algorithms, Neural Networks, Jan. 2013. [7] P. Smolensky, Information processing in dynamical systems: Foundations of harmony theory, 1986. [8] G. Hinton, S. Osindero, and Y. Teh, A fast learning algorithm for deep belief nets, Neural computation, vol. 18, no. 7, pp. 1527 1554, 2006. [9] G. Hinton, L. Deng, D. Yu, G. Dahl, A.-R. Mohamed, N. Jaitly, A. Senior, V. Vanhoucke, P. Nguyen, T. Sainath et al., Deep neural networks for acoustic modeling in speech recognition, IEEE Signal Processing Magazine, 2012. [10] B. Kosko, Bidirectional associative memories, Systems, Man and Cybernetics, IEEE Transactions on, vol. 18, no. 1, pp. 49 60, 1988. [11], Adaptive bidirectional associative memories, Applied optics, vol. 26, no. 23, pp. 4947 4960, 1987. [12] A. Patel and B. Kosko, Error-probability noise benefits in threshold neural signal detection, Neural Networks, vol. 22, no. 5, pp. 697 706, 2009. [13], Stochastic Resonance in Continuous and Spiking Neurons with Levy Noise, IEEE Transactions on Neural Networks, vol. 19, no. 12, pp. 1993 2008, December 2008. [14] B. Kosko, Noise. Viking, 2006. [15] A. Patel and B. Kosko, Levy Noise Benefits in Neural Signal Detection, in Acoustics, Speech and Signal Processing, 2007. ICASSP 2007. IEEE International Conference on, vol. 3, 2007, pp. III 1413 III 1416. [16] M. Wilde and B. Kosko, Quantum forbidden-interval theorems for stochastic resonance, Journal of Physical A: Mathematical Theory, vol. 42, no. 46, 2009. [17] A. Patel and B. Kosko, Optimal Mean-Square Noise Benefits in Quantizer-Array Linear Estimation, IEEE Signal Processing Letters, vol. 17, no. 12, pp. 1005 1009, Dec. 2010. [18], Noise Benefits in Quantizer-Array Correlation Detection and Watermark Decoding, IEEE Transactions on Signal Processing, vol. 59, no. 2, pp. 488 505, Feb. 2011. [19] B. Franzke and B. Kosko, Noise Can Speed Convergence in Markov Chains, Physical Review E, vol. 84, no. 4, p. 041112, 2011. [20] A. Bulsara, R. Boss, and E. Jacobs, Noise effects in an electronic model of a single neuron, Biological cybernetics, vol. 61, no. 3, pp. 211 222, 1989. [21] L. Gammaitoni, P. Hänggi, P. Jung, and F. Marchesoni, Stochastic resonance, Reviews of Modern Physics, vol. 70, no. 1, p. 223, 1998. [22] C. M. Bishop, Training with noise is equivalent to Tikhonov regularization, Neural computation, vol. 7, no. 1, pp. 108 116, 1995. [23] Y. Hayakawa, A. Marumoto, and Y. Sawada, Effects of the chaotic noise on the performance of a neural network model for optimization problems, Physical review E, vol. 51, no. 4, pp. 2693 2696, 1995. [24] K. Matsuoka, Noise injection into inputs in back-propagation learning, Systems, Man and Cybernetics, IEEE Transactions on, vol. 22, no. 3, pp. 436 440, 1992. [25] G. An, The effects of adding noise during backpropagation training on a generalization performance, Neural Computation, vol. 8, no. 3, pp. 643 674, 1996. [26] D. Oakes, Direct calculation of the information matrix via the EM, Journal of the Royal Statistical Society: Series B (Statistical Methodology, vol. 61, no. 2, pp. 479 482, 1999. [27] T. M. Cover and J. A. Thomas, Elements of information theory. Wileyinterscience, 2012. [28] S. Kullback and R. A. Leibler, On information and sufficiency, The Annals of Mathematical Statistics, vol. 22, no. 1, pp. 79 86, 1951. [29] A. P. Dempster, N. M. Laird, and D. B. Rubin, Maximum likelihood from incomplete data via the EM algorithm, Journal of the Royal Statistical Society. Series B (Methodological, pp. 1 38, 1977. [30] G. J. McLachlan and T. Krishnan, The EM algorithm and extensions. Wiley-Interscience, 2007, vol. 382. [31] H. Teicher, Identifiability of finite mixtures, The Annals of Mathematical Statistics, vol. 34, no. 4, pp. 1265 1269, 1963. [32] G. McLachlan and D. Peel, Finite mixture models. Wiley-Interscience, 2000. [33] L. R. Rabiner, A tutorial on hidden Markov models and selected applications in speech recognition, Proceedings of the IEEE, vol. 77, no. 2, pp. 257 286, 1989. [34] G. E. Hinton and R. R. Salakhutdinov, Reducing the dimensionality of data with neural networks, Science, vol. 313, no. 5786, pp. 504 507, 2006. [35] G. Hinton, Training a deep autoencoder or a classifier on MNIST digits, http://www.cs.toronto.edu/ hinton/matlabforsciencepaper.html, [Online; accessed 20-Feb-2013].