Abstract. In this paper we propose recurrent neural networks with feedback into the input

Similar documents
Samy Bengioy, Yoshua Bengioz. y INRS-Telecommunications, 16, Place du Commerce, Ile-des-Soeurs, Qc, H3E 1H6, CANADA

Modeling High-Dimensional Discrete Data with Multi-Layer Neural Networks

a subset of these N input variables. A naive method is to train a new neural network on this subset to determine this performance. Instead of the comp

Lecture 5: Logistic Regression. Neural Networks

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

LSTM CAN SOLVE HARD. Jurgen Schmidhuber Lugano, Switzerland. Abstract. guessing than by the proposed algorithms.

Direct Method for Training Feed-forward Neural Networks using Batch Extended Kalman Filter for Multi- Step-Ahead Predictions

Recurrent Neural Net Learning and Vanishing Gradient. International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems 6(2):107{116, 1998

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92

Adaptive Boosting of Neural Networks for Character Recognition

Experts. Lei Xu. Dept. of Computer Science, The Chinese University of Hong Kong. Dept. of Computer Science. Toronto, M5S 1A4, Canada.

Lecture 7 Artificial neural networks: Supervised learning

From perceptrons to word embeddings. Simon Šuster University of Groningen

Christian Mohr

Introduction to Neural Networks

Deep Feedforward Networks

memory networks, have been proposed by Hopeld (1982), Lapedes and Farber (1986), Almeida (1987), Pineda (1988), and Rohwer and Forrest (1987). Other r

In Advances in Neural Information Processing Systems 6. J. D. Cowan, G. Tesauro and. Convergence of Indirect Adaptive. Andrew G.

EEE 241: Linear Systems

DEEP LEARNING AND NEURAL NETWORKS: BACKGROUND AND HISTORY

Stable Adaptive Momentum for Rapid Online Learning in Nonlinear Systems

THE presence of missing values in a dataset often makes

Large-Scale Feature Learning with Spike-and-Slab Sparse Coding

Introduction Neural Networks - Architecture Network Training Small Example - ZIP Codes Summary. Neural Networks - I. Henrik I Christensen

Need for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels

Training Neural Networks with Deficient Data

Machine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6

Other Topologies. Y. LeCun: Machine Learning and Pattern Recognition p. 5/3

Artificial Neural Network

Early Brain Damage. Volker Tresp, Ralph Neuneier and Hans Georg Zimmermann Siemens AG, Corporate Technologies Otto-Hahn-Ring München, Germany

Clustering. Professor Ameet Talwalkar. Professor Ameet Talwalkar CS260 Machine Learning Algorithms March 8, / 26

Learning Gaussian Process Models from Uncertain Data

Deep learning / Ian Goodfellow, Yoshua Bengio and Aaron Courville. - Cambridge, MA ; London, Spis treści

Gaussian Processes for Regression. Carl Edward Rasmussen. Department of Computer Science. Toronto, ONT, M5S 1A4, Canada.

Speaker Representation and Verification Part II. by Vasileios Vasilakakis

4. Multilayer Perceptrons

Equivalence of Backpropagation and Contrastive Hebbian Learning in a Layered Network

Computational statistics

Deep Neural Networks

Learning Deep Architectures for AI. Part II - Vijay Chakilam

The wake-sleep algorithm for unsupervised neural networks

Neural Networks and Ensemble Methods for Classification

Credit Assignment: Beyond Backpropagation

Variable Selection in Data Mining Project

A Robust PCA by LMSER Learning with Iterative Error. Bai-ling Zhang Irwin King Lei Xu.

Need for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels

Unsupervised Discovery of Nonlinear Structure Using Contrastive Backpropagation

Neural Networks with Applications to Vision and Language. Feedforward Networks. Marco Kuhlmann

STA 414/2104: Lecture 8

Deep Belief Networks are compact universal approximators

Machine Learning. Neural Networks

Fast pruning using principal components

MODULAR ECHO STATE NEURAL NETWORKS IN TIME SERIES PREDICTION

STA 414/2104: Lecture 8

Pattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore

POWER SYSTEM DYNAMIC SECURITY ASSESSMENT CLASSICAL TO MODERN APPROACH

Lecture 17: Neural Networks and Deep Learning

MACHINE LEARNING AND PATTERN RECOGNITION Fall 2005, Lecture 4 Gradient-Based Learning III: Architectures Yann LeCun

Aijun An and Nick Cercone. Department of Computer Science, University of Waterloo. methods in a context of learning classication rules.

Reinforcement Learning, Neural Networks and PI Control Applied to a Heating Coil

A Novel Activity Detection Method

Temporal Backpropagation for FIR Neural Networks

Effect of number of hidden neurons on learning in large-scale layered neural networks

Greedy Layer-Wise Training of Deep Networks

Second-order Learning Algorithm with Squared Penalty Term

Network Structuring And Training Using Rule-based Knowledge

Pattern Recognition and Machine Learning

Learning in Boltzmann Trees. Lawrence Saul and Michael Jordan. Massachusetts Institute of Technology. Cambridge, MA January 31, 1995.

output dimension input dimension Gaussian evidence Gaussian Gaussian evidence evidence from t +1 inputs and outputs at time t x t+2 x t-1 x t+1

Data Mining Part 5. Prediction

Course 10. Kernel methods. Classical and deep neural networks.

Feed-forward Networks Network Training Error Backpropagation Applications. Neural Networks. Oliver Schulte - CMPT 726. Bishop PRML Ch.

UNSUPERVISED LEARNING

Probabilistic Time Series Classification

Modelling Time Series with Neural Networks. Volker Tresp Summer 2017

Deep Learning. What Is Deep Learning? The Rise of Deep Learning. Long History (in Hind Sight)

Machine Learning and Data Mining. Multi-layer Perceptrons & Neural Networks: Basics. Prof. Alexander Ihler

ADAPTIVE NEURO-FUZZY INFERENCE SYSTEMS

Multi-Layer Boosting for Pattern Recognition

DETECTING PROCESS STATE CHANGES BY NONLINEAR BLIND SOURCE SEPARATION. Alexandre Iline, Harri Valpola and Erkki Oja

Chapter 11. Stochastic Methods Rooted in Statistical Mechanics

The application of neural networks to the paper-making industry

CSE 352 (AI) LECTURE NOTES Professor Anita Wasilewska. NEURAL NETWORKS Learning

Phase-Space learning for recurrent networks. and. Abstract. We study the problem of learning nonstatic attractors in recurrent networks.

SPSS, University of Texas at Arlington. Topics in Machine Learning-EE 5359 Neural Networks

EE04 804(B) Soft Computing Ver. 1.2 Class 2. Neural Networks - I Feb 23, Sasidharan Sreedharan

Deep Learning Recurrent Networks 2/28/2018

Deep Learning Architecture for Univariate Time Series Forecasting

I D I A P. Online Policy Adaptation for Ensemble Classifiers R E S E A R C H R E P O R T. Samy Bengio b. Christos Dimitrakakis a IDIAP RR 03-69

Artificial Neural Networks. Edward Gatt

Fast Exact Multiplication by the Hessian

STA 414/2104: Machine Learning

Neural networks. Chapter 20. Chapter 20 1

PATTERN CLASSIFICATION

Sum-Product Networks: A New Deep Architecture

100 inference steps doesn't seem like enough. Many neuron-like threshold switching units. Many weighted interconnections among units

WHY ARE DEEP NETS REVERSIBLE: A SIMPLE THEORY,

Neural Networks and the Back-propagation Algorithm

CSE446: Neural Networks Spring Many slides are adapted from Carlos Guestrin and Luke Zettlemoyer

Mark Gales October y (x) x 1. x 2 y (x) Inputs. Outputs. x d. y (x) Second Output layer layer. layer.

Transcription:

Recurrent Neural Networks for Missing or Asynchronous Data Yoshua Bengio Dept. Informatique et Recherche Operationnelle Universite de Montreal Montreal, Qc H3C-3J7 bengioy@iro.umontreal.ca Francois Gingras Dept. Informatique et Recherche Operationnelle Universite demontreal Montreal, Qc H3C-3J7 gingras@iro.umontreal.ca Abstract In this paper we propose recurrent neural networks with feedback into the input units for handling two types of data analysis problems. On the one hand, this scheme can be used for static data when some of the input variables are missing. On the other hand, it can also be used for sequential data, when some of the input variables are missing or are available at dierent frequencies. Unlike inthe case of probabilistic models (e.g. Gaussian) of the missing variables, the network does not attempt to model the distribution of the missing variables given the observed variables. Instead it is a more \discriminant" approach that lls in the missing variables for the sole purpose of minimizing a learning criterion (e.g., to minimize an output error). Introduction Learning from examples implies discovering certain relations between variables of interest. The most general form of learning requires to essentially capture the joint distribution between these variables. However, for many specic problems, we are only interested in predicting the value of certain variables when the others (or some of the others) are given. A distinction is therefore made between input variables and output variables. Such a task requires less information (and less parameters, in the case of a parameterized model) than that of estimating the full joint distribution. For example in the case of classication problems, a traditional statistical approach is based on estimating the conditional distribution of the inputs for each class, as well as the class prior probabilities (thus yielding the full joint distribution of inputs and classes). A more discriminant approach concentrates on estimating the class boundaries (and therefore requires less parameters), as for example with a feedforward neural network trained to estimate the output class probabilities given the observed variables. However, for many learning problems, only some of the input variables are given for each particular training case, and the missing variables dier from case to case. The simplest way to deal with this missing data problem consists in replacing the missing values by their unconditional mean. It can be used with \discriminant" training algorithms such as those used with feedforward neural networks. However, in some problems, one can obtain better results by taking advantage of the dependencies between the input variables. A simple idea therefore consists also, AT&T Bell Labs, Holmdel, NJ 07733

() (4) (2) 4 (3) (3) 2 (90) (6) Figure : Architectures of the recurrent networks in the experiments. On the left a 90-3-4 architecture for static data with missing values, on the right a 6-3-2- architecture with multiple time-scales for asynchronous sequential data. Small squares represent a unit delay. The number of units in each layer is inside the rectangles. The time scale at which eachlayer operates is on the right of each rectangle. in replacing the missing input variables by theirconditional expected value, when the observed input variables are given. An even better scheme is to compute the expected output given the observed inputs, e.g. with a mixture of Gaussian. Unfortunately, this amounts to estimating the full joint distribution of all the variables. For example, with n i inputs, capturing the possible eect of each observed variable on each missing variable would require O(n 2 i ) parameters (at least one parameter to capture some co-occurrence statistic on each pair of input variables). Many related approaches have been proposed to deal with missing inputs using a Gaussian (or Gaussian mixture) model (Ahmad and Tresp, 993 Tresp, Ahmad and Neuneier, 994 Ghahramani and Jordan, 994). In the experiments presented here, the proposed recurrent network is compared with a Gaussian mixture model trained with EM to handle missing values (Ghahramani and Jordan, 994). The approach proposed in section 2 is more economical than the traditional Gaussian-based approaches for two reasons. Firstly, we take advantage of hidden units in a recurrent network, which might be less numerous than the inputs. The number of parameters depends on the product of the number of hidden units and the number of inputs. The hidden units only need to capture the dependencies between input variables which have some dependencies, and which are useful to reducing the output error. The second advantage is indeed that training is based on optimizing the desired criterion (e.g., reducing an output error), rather than predicting as well as possible the values of the missing inputs. The recurrent network is allowed to relax for a few iterations (typically as few as 4 or 5) in order to ll-in some values for the missing inputs and produce an output. In section 3 we present experimental results with this approach, comparing the results with those obtained with a feedforward network. In section 4 we propose an extension of this scheme to sequential data. In this case, the network is not relaxing: inputs keep changing with time and the network maps an input sequence (with possibly missing values) to an output sequence. The main advantage of this extension is that it allows to deal with sequential data in which the variables occur at dierent frequencies. This type of problem is frequent for example with economic or nancial data. An experiment with asynchronous data is presented in section 5. 2 Relaxing Recurrent Network for Missing Inputs Networks with feedback such as those proposed in (Almeida, 987 Pineda, 989) can be applied to learning a static input/output mapping when some of the inputs are missing. In both cases, however,onehastowait for the network to relax either to a xed point (assuming it does nd one) or to a \stable distribution" (in the case of the Boltzmann machine). In the case of xedpoint recurrent networks, the training algorithm assumes that a xed point has been reached. The gradient with respect to the weights is then computed in order to move the xed point to a more desirable position. The approach we have preferred here avoids such an assumption.

Instead it uses a more explicit optimization of the whole behavior of the network as it unfolds in time, lls-in the missing inputs and produces an output. The network is trained to minimize some function of its output by back-propagation through time. Computation of Outputs Given Observed Inputs Given: input vector u =[u u 2 ::: u ni ] Result: output vector y =[y y 2 ::: y no ]. Initialize for t =0: For i =:::n u x 0 i 0 For i =:::n i,ifu i is missing then x 0 I(i) Else x 0 I(i) u i. 2. Loop over time: For t =tot For i =:::n u E(i), If i = I(k) is an input unit and u k is not missing then x t i u k P Else x t i ( ; )x t; i + f( l2s i w l x t;dl p l ) where S i is a set of links from unit p l to unit i, each with weight w l and a discrete delay d l (but terms for which t ; d l < 0were not considered). 3. Collect outputs by averaging at the end of the sequence: P T t= v t x t O(i) y i Back-Propagation The back-propagation computation requires an extra set of variables _x t and _w, which will contain respectively @x t and after this computation. @w Given: output gradient vector @y Result: input gradient and parameter gradient. @u @w. Initialize unit gradients using outside gradient: Initialize _x t i = 0 for all t and i. For i =:::n o, initialize _x t O(i) v t @y i 2. Backward loop over time: For t = T to For i = n u ::: If i = I(k) is an input unit and u k is not missing then no backward propagation Else For l 2 S i If t ; d l > 0 _x t;dl p l _x t;dl p l +(; )_x t;dl+ + w l _x t i f 0 ( P l2s i w l x t;dl p l ) P _w l _w l + f 0 ( l2s w i l x t;dl p l )x t;dl p l 3. Collect input gradients: For i =:::n i, If u i is missing, then @u i 0 Else @u i Pt _x t I(i) The observed inputs are clamped for the whole duration of the sequence. The missing units corresponding to missing inputs are initialized to their unconditional expectation and their value is then updated using the feedback links for the rest of the sequence (just as if they were hidden units). To help stability ofthenetwork and P prevent it from nding periodic solutions (in which the outputs have a correct output only periodically), output supervision is given for several time steps. A xed vector v, with v t > 0and t v t = species a weighing scheme that distributes

the responsibility for producing the correct output among dierent time steps. Its purpose is to encourage the network to develop stable dynamics which gradually converge toward the correct output (thus the weights v t were chosen to gradually increase with t). The neuron transfer function was a hyperbolic tangent in our experiments. The inertial term weighted by (in step 3 of the forward propagation algorithm below) was used to help the network nd stable solutions. The parameter was xed by hand. In the experiments described below, a value of 0.7 was used, but near values yielded similar results. This module can therefore be combined within a hybrid system composed of several modules by propagating gradient through the combined system (as in (Bottou and Gallinari, 99)). For example, as in Figure 2, there might be another module taking as input the recurrent network's output. In this case the recurrent network can be seen as a feature extractor that accepts data with missing values in input and computes a set of features that are never missing. In another example of hybrid system the non-missing values in input of the recurrent network are computed by another, upstream module (such as the preprocessing normalization used in our experiments), and the recurrent network would provide gradients to this upstream module (for example to better tune its normalization parameters). 3 Experiments with Static Data A network with three layers (inputs, hidden, outputs) was trained to classify data with missing values from the audiology database. This database was made public thanks to Jergen and Quinlan, was used by (Bareiss and Porter, 987), and was obtained from the UCI Repository of machine learning databases (ftp.ics.uci.edu:pub/machine-learning-databases). The original database has 226 patterns, with 69 attributes, and 24 classes. Unfortunately, most of the classes have only exemplar. Hence we decided to cluster the classes into four groups. To do so, the average pattern for each of the 24 classes was computed, and the K-Means clustering algorithm was then applied on those 24 prototypical class \patterns", to yield the 4 \superclasses" used in our experiments. The multi-valued input symbolic attributes (with more than 2 possible values) where coded with a \one-out-of-n" scheme, using n inputs (all zeros except the one corresponding to the attribute value). Note that a missing value was represented with a special numeric value recognized by the neural network module. The inputs which were constant over the training set were then removed. The remaining 90 inputs were nally standardized (by computing mean and standard deviation) and transformed by a saturating non-linearity (a scaled hyperbolic tangent). The output class is coded with a \one-out-of-4" scheme, and the recognized class is the one for which the corresponding output has the largest value. The architecture of the network is depicted in Figure (left). The length of each relaxing sequence in the experiments was 5. Higher values would not bring any measurable improvements, whereas for shorter sequences performance would degrade. The number of hidden units was varied, with the best generalization performance obtained using 3 hidden units. The recurrent network was compared with feedforward networks as well as with a mixture of Gaussians. For the feedforward networks, the missing input values were replaced by their unconditional expected value. They were trained to minimize the same criterion as the recurrent networks, i.e., the sum of squared dierences between network output and desired output. Several feedforward neural networks with varying numbers of hidden units were trained. The best generalization was obtained with 5 hidden units. Experiments were also performed with no hidden units and two hidden layers (see Table ). We found that the recurrent network not only generalized better but also learned much faster (although each pattern required 5 times more work because of the relaxation), as depicted in Figure 3. The recurrent network was also compared with an approach based on a Gaussian and Gaussian mixture model of the data. We used the algorithm described in (Ghahramani and Jordan, 994) for supervised leaning from incomplete data with the EM algorithm. The whole joint input/output distribution is modeled using a mixture model with Gaussians (for the inputs) and multinomial (outputs) components: X P (X = x C = c) = (2) n=2 j j j expf; =2 2 (x ; j) 0 ; j (x ; j )g j where x is the input vector, c the output class, and P (! j ) the prior probability of component j of the mixture. The jd are the multinomial parameters j and j are the Gaussian mean vector P (! j ) jd

cost down stream static module recurrent network... Upstream normalization module Figure 2: Example of hybrid modular system, using the recurrent network (middle) to extract features from patterns which mayhave missing values. Itcanbecombined with upstream modules (e.g., a normalizing preprocessor, right) and downstream modules (e.g., a static classier, left). Dotted arrows show the backward ow of gradients. 50 training set 50 test set 45 45 40 40 35 35 30 30 feedforward % error 25 % error 25 20 20 5 5 recurrent 0 recurrent 0 5 feedforward 5 0 0 0 20 30 40 training epoch 0 0 0 20 30 40 training epoch Figure 3: Evolution of training and test error for the recurrent network and for the best of the feedforward networks (90-5-4): average classication error w.r.t. training epoch, (with standard deviation error bars, computed over 0 trials). and covariance matrix for component j. Maximum likelihood training is applied as explained in (Ghahramani and Jordan, 994), taking missing values into account (as additional missing variables of the EM algorithm). For each architecture in Table, 0 training trials were run with a dierent subset of 200 training and 26 test patterns (and dierent initial weights for the neural networks). The recurrent network was clearly superior to the other architectures, probably for the reasons discussed in the conclusion. In addition, we have shown graphically the rate of convergence during training of the best feedforward network (90-5-4) as well as the best recurrent network (90-3-4), in Figure 3. Clearly, the recurrent network not only performs better at the end of training but also learns much faster. 4 Recurrent Network for Asynchronous Sequential Data An important problem with many sequential data analysis problems such as those encountered in nancial data sets is that dierent variables are known at dierent frequencies, at dierent times (phase), or are sometimes missing. For example, some variables are given daily, weekly, monthly, quarterly, or yearly. Furthermore, some variables may not even be given for some of the periods or the precise timing may change (for example the date at which a company reports nancial performance my vary). Therefore, we propose to extend the algorithm presented above for static data with missing values to the general case of sequential data with missing values or asynchronous variables. For time steps at which a low-frequency variable is not given, a missing value is assumed in input. Again, the feedback links from the hidden and output units to the input units allow the network

Table : Comparative performances of recurrent network, feedforward network, and Gaussian mixture density model on audiology data. The average percentage of classication error is shown after training, for both training and test sets, and the standard deviation in parenthesis, for 0 trials. Training set error Test set error 90-3-4 Recurrent net 0.3(0.6) 2.7(2.6) 90-6-4 Recurrent net 90-25-4 Feedforward net 0(0) 0.5(.6) 3.8(4) 5(7.3) 90-5-4 Feedforward net 0.8(0.4) 3.8(7) 90-0-6-4 Feedforward net (0.9) 6(5.3) 90-6-4 Feedforward net 6(4.9) 29(8.9) 90-2-4 Feedforward net 90-4 Feedforward net 8.5() 22() 27(0) 33(8) Gaussian 35(.6) 38(9.3) 4 Gaussians Mixture 8 Gaussians Mixture 36(.5) 36(2.) 38(9.2) 38(9.3) to \complete" the missing data. The main dierences with the static case are that the inputs and outputs vary with t (we useu t and y t at each time step instead of u and y). The training algorithm is otherwise the same. 5 Experiments with Asynchronous Data To evaluate the algorithm, we have used a recurrent network with random weights, and feedback links on the input units to generate articial data. The generating network has 6 inputs, 3 hidden and outputs. The hidden layer is connected to the input layer ( delay). The hidden layer receives inputs with delays 0 and from the input layer and with delay from itself. The output layer receives inputs from the hidden layer. At the initial time step as well as at 5% of the time steps (chosen randomly), the input units were clamped with random values to introduce some further variability. The missing values were then completed by the recurrent network. To generate asynchronous data, half of the inputs were then hidden with missing values 4 out of every 5 time steps. 00 training sequences and 50 test sequences were generated. The learning problem is therefore a sequence regression problem with missing and asynchronous input variables. Preliminary comparative experiments show a clear advantage to completing the missing values (due to the the dierent frequencies of the input variables) with the recurrent network, as shown in Figure 4. The recognition recurrent network is shown on the right of Figure. It has multiple time scales (implemented with subsampling and oversampling, as in TDNNs (Lang, Waibel and Hinton, 990) and reverse-tdnns (Simard and LeCun, 992)), to facilitate the learning of such asynchronous data. The static network is a time-delay neural network with 6 input, 8 hidden, and output unit, and connections with delays 0, 2, and 4 from the input to hidden and hidden to output units. The \missing values" for slow-varying variables were replaced by the last observed value in the sequence. Experiments with 4 and 6 hidden units yielded similar results. 6 Conclusion When there are dependencies between input variables, and the output prediction can be improved by taking them into account, we have seen that a recurrent network with input feedback can perform signicantly better than a simpler approach that replaces missing values by their unconditional expectation. According to us, this explains the signicant improvement brought by using the recurrent network instead of a feedforward network in the experiments. On the other hand, the large number of input variables (n i = 90, in the experiments) most likely explains the poor performance of the mixture of Gaussian model in comparison to both the static networks and the recurrent network. The Gaussian model requires estimating O(n 2 i ) parameters and inverting large covariance matrices. The approach to handling missing values presented here can also be extended to sequential data with missing or asynchronous variables. As our experiments suggest, for such problems, using recurrence and multiple time scales yields better performance than static or time-delay networks for which the missing values are lled using a heuristic.

0.8 0.6 0.4 test set mse 0.2 0. time delay network 0.08 0.06 recurrent network 0.04 0 2 4 6 8 0 2 4 6 8 20 training epoch Figure 4: Test set mean squared error on the asynchronous data. Top: static network with time delays. Bottom: recurrent network with feedback to input values to complete missing data. References Ahmad, S. and Tresp, V. (993). Some solutions to the missing feature problem in vision. In Hanson, S. J., Cowan, J. D., and Giles, C. L., editors, Advances in Neural Information Processing Systems 5, San Mateo, CA. Morgan Kaufman Publishers. Almeida, L. (987). A learning rule for asynchronous perceptrons with feedback in a combinatorial environment. In Caudill, M. and Butler, C., editors, IEEE International Conference on Neural Networks, volume 2, pages 609{68, San Diego 987. IEEE, New York. Bareiss, E. and Porter, B. (987). Protos: An exemplar-based learning apprentice. In Proceedings of the 4th International Workshop on Machine Learning, pages 2{23, Irvine, CA. Morgan Kaufmann. Bottou, L. and Gallinari, P. (99). A framework for the cooperation of learning algorithms. In Lippman, R. P., Moody, R., and Touretzky, D. S., editors, Advances in Neural Information Processing Systems 3, pages 78{788, Denver, CO. Ghahramani, Z. and Jordan, M. I. (994). Supervised learning from incomplete data via an EM approach. In Cowan, J., Tesauro, G., and Alspector, J., editors, Advances in Neural Information Processing Systems 6, page, San Mateo, CA. Morgan Kaufmann. Lang, K. J., Waibel, A. H., and Hinton, G. E. (990). A time-delay neural network architecture for isolated word recognition. Neural Networks, 3:23{43. Pineda, F. (989). Recurrent back-propagation and the dynamical approach to adaptive neural computation. Neural Computation, :6{72. Simard, P. and LeCun, Y. (992). Reverse TDNN: An architecture for trajectory generation. In Moody, J., Hanson, S., and Lipmann, R., editors, Advances in Neural Information Processing Systems 4, pages 579{588, Denver, CO. Morgan Kaufmann, San Mateo. Tresp, V., Ahmad, S., and Neuneier, R. (994). Training neural networks with decient data. In Cowan, J., Tesauro, G., and Alspector, J., editors, Advances in Neural Information Processing Systems 6, pages 28{35. Morgan Kaufman Publishers, San Mateo, CA.