ADAPTIVE INVERSE CONTROL BASED ON NONLINEAR ADAPTIVE FILTERING. Information Systems Lab., EE Dep., Stanford University

Similar documents
inear Adaptive Inverse Control

Adaptive Inverse Control

Adaptive Inverse Control based on Linear and Nonlinear Adaptive Filtering

EFFICIENT LINEAR MIMO ADAPTIVE ŜP INVERSE CONTROL. Gregory L. Plett E

Variable Learning Rate LMS Based Linear Adaptive Inverse Control *

Automatic Stabilization of an Unmodeled Dynamical System Final Report

Temporal Backpropagation for FIR Neural Networks

EFFICIENT LINEAR MIMO ADAPTIVE INVERSE CONTROL

SOME RESULTS CONCERNING FAST LINEAR MIMO ADAPTIVE INVERSE CONTROL

THE Back-Prop (Backpropagation) algorithm of Paul

Gradient-Adaptive Algorithms for Minimum Phase - All Pass Decomposition of an FIR System

New Recursive-Least-Squares Algorithms for Nonlinear Active Control of Sound and Vibration Using Neural Networks

Recurrent neural networks with trainable amplitude of activation functions

( t) Identification and Control of a Nonlinear Bioreactor Plant Using Classical and Dynamical Neural Networks

Neural Network Identification of Non Linear Systems Using State Space Techniques.

4.0 Update Algorithms For Linear Closed-Loop Systems

On the Use of A Priori Knowledge in Adaptive Inverse Control

CONTROL SYSTEMS, ROBOTICS, AND AUTOMATION - Vol. V - Prediction Error Methods - Torsten Söderström

Title without the persistently exciting c. works must be obtained from the IEE

Relating Real-Time Backpropagation and. Backpropagation-Through-Time: An Application of Flow Graph. Interreciprocity.

Application of Neuro Fuzzy Reduced Order Observer in Magnetic Bearing Systems

Bidirectional Representation and Backpropagation Learning

Modelling Time Series with Neural Networks. Volker Tresp Summer 2017

Adaptive Inverse Control Using Kernel Identification

Address for Correspondence

Identification of Nonlinear Systems Using Neural Networks and Polynomial Models

Correspondence. Nonlinear Modeling of Complex Large-Scale Plants Using Neural Networks and Stochastic Approximation

NONLINEAR PLANT IDENTIFICATION BY WAVELETS

A METHOD OF ADAPTATION BETWEEN STEEPEST- DESCENT AND NEWTON S ALGORITHM FOR MULTI- CHANNEL ACTIVE CONTROL OF TONAL NOISE AND VIBRATION

Gaussian Process for Internal Model Control

LECTURE # - NEURAL COMPUTATION, Feb 04, Linear Regression. x 1 θ 1 output... θ M x M. Assumes a functional form

TECHNICAL RESEARCH REPORT

NON-LINEAR CONTROL OF OUTPUT PROBABILITY DENSITY FUNCTION FOR LINEAR ARMAX SYSTEMS

Hand Written Digit Recognition using Kalman Filter

THE concept of active sound control has been known

SNR lidar signal improovement by adaptive tecniques

Autonomous learning algorithm for fully connected recurrent networks

ADAPTIVE FILTER THEORY

Bifurcation Analysis of Chaotic Systems using a Model Built on Artificial Neural Networks

AN ON-LINE ADAPTATION ALGORITHM FOR ADAPTIVE SYSTEM TRAINING WITH MINIMUM ERROR ENTROPY: STOCHASTIC INFORMATION GRADIENT

Direct Method for Training Feed-forward Neural Networks using Batch Extended Kalman Filter for Multi- Step-Ahead Predictions

System Identification Using a Retrospective Correction Filter for Adaptive Feedback Model Updating

On an internal multimodel control for nonlinear multivariable systems - A comparative study

Classification and Clustering of Printed Mathematical Symbols with Improved Backpropagation and Self-Organizing Map

CLOSE-TO-CLEAN REGULARIZATION RELATES

IS NEGATIVE STEP SIZE LMS ALGORITHM STABLE OPERATION POSSIBLE?

AN ALTERNATIVE FEEDBACK STRUCTURE FOR THE ADAPTIVE ACTIVE CONTROL OF PERIODIC AND TIME-VARYING PERIODIC DISTURBANCES

NEURAL NETWORKS FOR TRANSMISSION OVER NONLINEAR MIMO CHANNELS

IN neural-network training, the most well-known online

Neural Dynamic Optimization for Control Systems Part II: Theory

An Adaptive LQG Combined With the MRAS Based LFFC for Motion Control Systems

An artificial neural networks (ANNs) model is a functional abstraction of the

Submitted to Electronics Letters. Indexing terms: Signal Processing, Adaptive Filters. The Combined LMS/F Algorithm Shao-Jen Lim and John G. Harris Co

Equivalence of Backpropagation and Contrastive Hebbian Learning in a Layered Network

Modular Deep Recurrent Neural Network: Application to Quadrotors

Control of Nonlinear Stochastic Systems: Model-Free Controllers versus Linear Quadratic Regulators

A Novel 2-D Model Approach for the Prediction of Hourly Solar Radiation

A Minimum Error Entropy criterion with Self Adjusting. Step-size (MEE-SAS)

A Comparison of Nonlinear Kalman Filtering Applied to Feed forward Neural Networks as Learning Algorithms

Evaluation Performance of PID, LQR, Pole Placement Controllers for Heat Exchanger

Image Contrast Enhancement and Quantitative Measuring of Information Flow

Computational Intelligence Winter Term 2017/18

Recursive Least Squares for an Entropy Regularized MSE Cost Function

A STATE-SPACE NEURAL NETWORK FOR MODELING DYNAMICAL NONLINEAR SYSTEMS

ACTIVE noise control (ANC) ([1], [2]) is an established

Least Mean Squares Regression. Machine Learning Fall 2018

STRUCTURED NEURAL NETWORK FOR NONLINEAR DYNAMIC SYSTEMS MODELING

T Machine Learning and Neural Networks

Model-based Correlation Measure for Gain and Offset Nonuniformity in Infrared Focal-Plane-Array Sensors

CBE507 LECTURE III Controller Design Using State-space Methods. Professor Dae Ryook Yang

KNOWN approaches for improving the performance of

Computational Intelligence

Gaussian-Shaped Circularly-Symmetric 2D Filter Banks

Load Forecasting Using Artificial Neural Networks and Support Vector Regression

Nonlinear System Identification Based on a Novel Adaptive Fuzzy Wavelet Neural Network

Dictionary Learning for L1-Exact Sparse Coding

Pattern Classification

International Journal of Advanced Research in Computer Science and Software Engineering

Convergence of Hybrid Algorithm with Adaptive Learning Parameter for Multilayer Neural Network

ADAPTIVE INVERSE CONTROL OF PLANTS WITH DISTURBANCES

Application of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption

Lecture 16: Introduction to Neural Networks

Part 8: Neural Networks

Dynamic Operability for the Calculation of Transient Output Constraints for Non-Square Linear Model Predictive Controllers

On Identification of Cascade Systems 1

Artificial Neural Networks. Historical description

12.4 Known Channel (Water-Filling Solution)

Least Mean Squares Regression

2262 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 47, NO. 8, AUGUST A General Class of Nonlinear Normalized Adaptive Filtering Algorithms

ADAPTIVE FILTER ALGORITHMS. Prepared by Deepa.T, Asst.Prof. /TCE

ANN Control of Non-Linear and Unstable System and its Implementation on Inverted Pendulum

Neural Networks. Volker Tresp Summer 2015

Design of Linear Time Varying Flatness-Based Control for Single-Input Single-Output Systems

Iterative Learning Control for Tailor Rolled Blanks Zhengyang Lu DCT

Keywords- Source coding, Huffman encoding, Artificial neural network, Multilayer perceptron, Backpropagation algorithm

ON THE REALIZATION OF 2D LATTICE-LADDER DISCRETE FILTERS

λ-universe: Introduction and Preliminary Study

On Descent Spectral CG algorithms for Training Recurrent Neural Networks

Supervised (BPL) verses Hybrid (RBF) Learning. By: Shahed Shahir

Transfer Function Identification from Phase Response Data

Transcription:

ADAPTIVE INVERSE CONTROL BASED ON NONLINEAR ADAPTIVE FILTERING Bernard Widrow 1, Gregory Plett, Edson Ferreira 3 and Marcelo Lamego 4 Information Systems Lab., EE Dep., Stanford University Abstract: Many problems in adaptive control can be divided into two parts; the first part is the control of plant dynamics, and the second is the control of plant disturbance. Very often, a single system is utilized to achieve both of these control objectives. The approach of this paper treats each problem separately. Control of plant dynamics can be achieved by preceding the plant with an adaptive controller whose transfer function is the inverse of that of the plant. Control plant disturbance can be achieved by an adaptive feedbac process that minimizes plant output disturbance without altering plant dynamics. The adaptive controller is implemented using adaptive filters. right 1998 IFAC. Keywords: Control, Inverse Control, s, Neural Networs, Nonlinear Systems. 1. INTRODUCTION At present, the control of a dynamic system (the plant ) is generally done by means of feedbac. This paper proposes an alternative approach that uses adaptive filtering to achieve feedforward control. Precision is attained because of the feedbac incorporated in the adaptive filtering. The control of plant dynamic response is treated separately, without compromise, from the optimal control of the disturbance. All of the required operations are based on adaptive filtering techniques (Widrow and Walach, 1996). Following the proposed methodology, nowledge of adaptive signal processing allows one to go deeply into the field of adaptive control. In order for adaptive inverse control to wor, the plant must be stable. If the plant is not stable, then conventional feedbac methods should be used to stabilize it. Generally, the form of this feedbac is not critical and would not need to be optimized. If the plant is stable to begin with, no feedbac would be required. If the plant is linear, a linear control system would generally be used. The transfer function of the 1 Professor. Ph.D. student. 3 Visiting Professor sponsored by CAPES/UFES, Brazil. 4 Ph.D. student sponsored by CNPq/UFES, Brazil. controller converges to the reciprocal of that of the plant. If the plant is minimum phase, an inverse is easily obtained. If the plant is nonminimum phase, a delayed inverse can be obtained. The delay in the inverse results in a delay in overall system response, but this is inevitable with a nonminimumphase plant. The basic idea can be used to implement modelreference control by adapting the cascaded filter to cause the overall system response to match a preselected model response. Disturbance in a linear plant, whether minimum phase or nonminimum phase, can be optimally controlled by a special circuit that obtains the disturbance at the plant output, filters it, and feeds it bac into the plant input. The circuit wors in such a way that the feedbac does not alter the plant dynamic response. So disturbance control and control of dynamic response can be accomplished separately. The same ideas wor for MIMO systems as well as SISO systems. Control of nonlinear plants is an important subject that raises significant issues. Since a nonlinear plant does not have a transfer function, how could it have an inverse? By using a cascade of the nonlinear adaptive filter with the nonlinear plant, the filter can learn to drive the plant as if it were the plant s inverse. This wors surprisingly well for a range of training and operation signals. Control of dynamic response and

plant disturbance can be done. This paper introduces adaptive inverse control by first discussing adaptive filters. Then, inverse plant modeling for linear plants is described. The ideas are extended to nonlinear control, examples are presented and conclusions made.. ADAPTIVE FILTERS An adaptive digital filter, shown in fig. 1 has an input, an output, and another special input called the desired response. The desired response input is sometimes called the training signal. Input x Error e Desired Response d Fig. 1. Symbolic representation of an adaptive transversal filter adapted by the LMS algorithm The adaptive filter contains adjustable parameters that control its impulse response. These parameters could, for example, be variable weights connected to the taps of a tapped delay line. The filter would thus be FIR, finite impulse response. The adaptive filter also incorporates an adaptive algorithm whose purpose is to automatically adjust the parameters to minimize some function of the error (usually mean square error). The error is defined as the difference between the desired response and the actual filter response. Many such algorithms exist, a number of which are described in the textboos by Widrow and Stearns (1985) and by Hayin (1996). 3. INVERSE PLANT MODELING The plant s controller will be an inverse of the plant. Inverse plant modeling of a linear SISO plant is illustrated in Fig. The plant input is its control signal. The plant output, shown in the figure, is the input of an adaptive filter. The desired response for the adaptive filter is the plant input (sometimes delayed by a modeling delay, ). Minimizing mean square error causes the adaptive filter ˆ P to be the best least squares inverse to the plant P for the given input spectrum. The adaptive algorithm attempts to mae the cascade of plant and adaptive inverse behave lie a unit gain. This process is often called deconvolution. With the delay incorporated as shown, the inverse will be a delayed inverse. For sae of argument, the plant can be assumed to have y poles and zeros. An inverse, if it also had poles and zeros, would need to have zeros where the plant had poles and poles where the plant had zeros. Maing an inverse would be no problem except for the case of a nonminimum phase plant. It would seem that such an inverse would need to have unstable poles, and this would be true if the inverse were causal. If the inverse could be noncausal as well as causal, however, then a twosided stable inverse would exist for all linear timeinvariant plants in accord with the theory of twosided ztransforms. For useful realization, the twosided inverse response would need to be delayed by. A causal FIR filter can approximate the delayed version of the twosided plant inverse. The time span of the adaptive filter (the number of weights multiplied by the sampling period) should be made adequately long, and the delay needs to be chosen appropriately. The choice is generally not critical. Control Input Unnown P Modeling Delay Error Desired Input (Desired Response) Fig.. Delayed inverse modeling of an unnown plant The inverse filter is used as a controller in the present scheme, so that becomes the response delay of the controlled plant. Maing small is generally desirable, but the quality of control depends on the accuracy of the inversion process, which sometimes requires to be of the order of half the length of the adaptive filter. Command Input r u u Unnown P Model M Fig. 3. inverse model control system A modelreference inversion process is incorporated in the feedforward control system of Fig. 3. A reference model is used in place of the delay of Fig.. Minimizing mean square error with the system of Fig. 3 causes the cascade of the plant and its model Weights Error

reference inverse to closely approximate the response of the referencemodel M. Much is nown about the design of modelreference systems (Landau, 1979). The model is chosen to give a desirable response for the overall system. Thus far, the plant has been treated as disturbance free. But, if there is disturbance, the scheme of Fig. 4 can be used. A direct plant modeling process, not shown, yields Pˆ, a close fitting FIR model of the plant. The difference between the plant output and the output of Pˆ is essentially the plant disturbance. Control Signal P (z) Pˆ (z) Disturbance n nˆ close when the canceler is turned on, at 3 samplings. Impulse Response of Nonminimum Phase.4.3..1.1..3 1 3 4 5 6 7 8 9 1 15 1 5 5 (a) Dynamic Tracing of Desired Noise Q (z) 5 1 3 4 5 6 7 8 9 1 (b) Disturbance Canceling Pˆ (z) Q (z) 1 3 Off Line Process Fig. 4. Optimal adaptive plant disturbance canceler Now, using a digital copy of Pˆ in place of P, an off line process, shown in Fig. 4, calculates the best leastsquares plant inverse Q. The off line process can run much faster than real time, so that as Pˆ is calculated, the inverse Q is immediately obtained. The disturbance is filtered by digital copy of Q and subtracted from the plant input. For linear systems, the scheme of Fig. 4 has been shown to be optimal in the leastsquares sense (Widrow and Walach, 1996). To illustrate the effectiveness of adaptive inverse control, a nonminimum phase plant has been simulated, and its impulse response is shown in Fig. 5(a). the output of this plant and the output of its reference model are plotted in Fig. 5(b), showing dynamics tracing when the command input signal is a random firstorder Marov process. The gray line is the desired output and the blac line is the actual plant output. Tracing is quite good. With disturbance added to the plant output, Fig. 5(c) shows the effect of disturbance cancelation. Both the desired and actual plant outputs are plotted in the figure, and they become 4 5 1 3 4 5 6 7 8 9 1 (c) Fig. 5. (a) Impulse response of the nonminimum phase plant used in simulation; (b) Dynamics tracing of desired output by actual plant output when the plant was not disturbed. (c) Cancelation of plant disturbance. 4. NONLINEAR ADAPTIVE INVERSE CONTROL WITH NEURAL NETWORKS Nonlinear inverse controllers can be used to control nonlinear plants. Although the theory is in its infancy, experiments can be done to demonstrate this. A nonlinear adaptive filter is shown in Fig. 6. It is composed of a neural networ whose input is a tapped delay line connected to the exogenous input signal. In addition, the input to the networ might include a tapped delay line connected to its own output signal. This type of nonlinear filter is called a Nonlinear Auto Regressive with exogeneous Input (NARX) filter, and has recently been shown to be a universal dynamical system (Siegelmann, et al., 1997). Algorithms such as

realtimerecurrentlearning (RTRL) (Williams and Zipser, 1989) and the bacpropagationthroughtime (BPTT) (Werbos, 199) may be used to adapt the weights of the neural networ to minimize the mean squared error. If the feedbac connections are omitted, the familiar bacpropagation may be used (Werbos, 1974), (Rumelhart, et al., 1986). In the nonlinear adaptive inverse control scheme of Fig.7, such filters are used as the plant emulator and controller. Nonlinear systems do not commute. Therefore, the simple and intuitive blocdiagram method of Figs. and 3, for adapting a controller to be the inverse of the plant, will not wor if the plant is nonlinear. Instead, a lowerlevel mathematical approach is taen. We use an extension of the RTRL learning algorithm to train the controller. This method can be briefly summarized using the notation of ordered derivatives, proposed by Werbos networ. We also use the fact that the plant model computes a function of the form y = f ( y, y,..., y n, u, u,..., u 1 p The weights of the controller are adapted using steepest descent. The change in the weights at each time step is in the negative direction to the gradient of the system error with respect to the weights of the controller. To find the gradient, we use the chainrule expansion for ordered derivatives u u = e = e m u u j= 1 j y u j ). (1) 1 x 1 x x n w w 1 w w n Neuron y = p y u j = j n y y j = 1 u W y W j j () Each of the terms in Eqs. (1) and () is either a Jacobian matrix, which may be calculated using the dualsubroutine (Werbos, 199) of the bacpropagation algorithm, or is previously calculated value of or. u y z z x y Command Input Controller Emulator z Neural Networ Model Fig. 6. An adaptive nonlinear filter composed of a tapped delay line and a threelayer neural networ. (1974). The goal is to adapt the weights of the controller to minimize the mean squared error of the output of the system. We use the fact that the controller computes a function of the form u = g ( u 1 1, u,..., u m, r, r,..., r q, W ), where W are the weights of the controller s neural Fig. 7. A method for adapting a nonlinear controller To be more specific, the first term in Eq. (1) is the partial derivative of the controller s output with respect to its weights. This term is one of the Jacobian matrices of the controller and may be calculated with the dual subroutine of the bacpropagation algotithm. The second part of Eq. (1) is a summation. The first term of the summation is the partial derivative of the controller s current output with respect to a previous output. However, since the controller is externally recurrent, this previous output is also a current input.

Therefore the first term of the summation is really just a partial derivative of the output of the controller with respect to one of its inputs. By definition, this is a submatrix of the Jacobian matrix for the networ, and may be computed using the dualsubroutine of the bacpropagation algorithm. The second term of the summation in Eq. (1) is the ordered partial derivative of a previous output with respect to the weights of the controller. This term has already been computed in a previous evaluation of Eq. (1), and need not be recomputed. y y. 1 3 = u 1 y The method just described for adapting a controller and disturbance canceler were simulated for this plant, and the results are presented here. 8 6 4 Tracing Uniform White Input A similar analysis may be performed to determine all of the terms required to evaluate Eq. (). After calculating these terms, the weights of the controller may be adapted using the weightupdate equation 4 W = µ e y Continual adaptation will minimize the mean squared error at the system output. C u r Pˆ P Pˆ Q M sys e Dist. w sys e = sys e Fig. 8. A fully integrated nonlinear adaptive inverse control scheme. Disturbance canceling for a nonlinear system is performed by filtering an estimate of the disturbance with the nonlinear filter Q and adding the filter s output to the control signal. An additional input to Q is the control signal to the plant u, to allow the disturbance canceler nowledge of the plant state. The same algorithm which was used to adapt the controller can be used to adapt the disturbance canceling filter. The entire control system is shown in Fig. 8. An interesting discretetime nonlinear plant has been studied by Narendra and Parthasarathy (199) y 6 8 1 3 4 5 6 7 8 9 1 8 6 4 4 6 (a) Tracing Sinusoidal Input 8 1 3 4 5 6 7 8 9 1 8 7 6 5 4 3 1 (b) Tracing SquareWave Input 1 3 4 5 6 7 8 9 1 (c) Fig. 9. Feedforward control of a nonlinear system. The controller was trained with uniform distributed (white) random input (a). Plots (b) and (c) show the plant tracing sinusoidal and square waves, having been previously trained with the random imput. With the reference model being a simple delay, and the command input being an i.i.d (independent and identically distributed) uniform process, the system adapted and learned to trac the model output. The result is shown in Fig. 9(a). The desired plant output (gray line) and the true plant output (solid line) are shown, at the end of training, when the training signal was used to drive the controller. The gray line is

completely covered by the blac line, indicating nearperfect control. With the weights fixed at their trained values, the next two plots show the generalization ability of the controller. After training with the random input, the adaptive process was halted. With no further training, the system was tested with inputs of different character in order to demonstrate the generalization ability of the controller. The first test was a sinewave command input. Tracing was surprisingly good, as shown in Fig. 9(b). Again, without further training, the system was tested with a squarewave command input, and the results, shown in Fig. 9(c), are excellent. A disturbance canceler was also trained for this plant, were the disturbance was a firstorder Marov signal added to the plant output. Fig.1 shows the results of disturbance cancelation. The power of the system error is plotted versus time. The disturbance canceler was turned on at iteration 5. Dramatic improvement may be seen. Square 3 5 15 1 5 Power of Disturbance 1 3 4 5 6 7 8 9 1 Fig. 1. Cancelation of plant disturbance for a nonlinear plant. The disturbance canceler was turned on at iteration 5. CONCLUSIONS control is seem as a two part problem, (a) a control of plant dynamics, and (b) control of plant disturbance. Conventionally, one uses feedbac control to treat both problems simultaneously. Tradeoffs and compromises are necessary to achieve good solutions, however. The method proposed here, based on inverse control, treats the two problems separately without compromise. The method applies to SISO and MIMO linear plants, and to nonlinear plants. An unnown linear plant will trac an input command signal if the plant is driven by a controller whose transfer function approximates the inverse of the plant transfer function. An adaptive inverse identification process can be used to obtain a stable controller, even if the plant is nonminimum phase. A modelreference version of this idea allows system dynamics to closely approximate desired referencemodel dynamics. No direct feedbac is used, except that the plant output is monitored and utilized by an adaptive algorithm to adjust the parameters of the controller. Although nonlinear plants do not have transfer functions, the same idea wors well for nonlinear plants. Control of internal plant disturbance is accomplished with an adaptive disturbance canceler. The canceler does not affect plant dynamics, but feeds bac plant disturbance in a way that minimizes plant output disturbance power. This approach is optimal for linear plants and wors surprisingly well with nonlinear plants. A great deal of wor will be needed to gain greater understanding of this ind of behavior, but the prospects for useful and unusual performance and for development of this new approach seem very promising. REFERENCES Hayin, S. (1996). Theory. Prentice Hall, third edition, Upper Saddle River, NJ. Landau, I. D. (1979). Control. The Model Approach, volume VIII of Control and Systems Theory Series. Marcel Deer, New Yor. Narendra, K. S. and K. Parthasarathy (199). Identification and control of dynamical systems using neural networs. IEEE Transactions on Neural Networs, Vol. 1(1), March. Rumelhart, D. E., G. E. Hinton and R. J. Williams (1986). Learning internal representations by error propagation. Parallel Distributed Processing. (D. E. Rumelhart and J. L McClelland editors), volume 1, chapter 8. MIT Press, Cambridge, MA. Siegelmann, H. T., B. B. Horne and C. L. Giles (1997). Computational capabilities of recurrent NARX neural networs. IEEE Transactions on Systems, Man and Cybernetics Part B: Cybernetics, Vol. 7(), April, pp. 8 15. Werbos, P. (1974). Beyond Regression: New Tools for Prediction and Analysis in the Behavioral Sciences. PhD thesis, Harvard University, Cambridge, MA, August. Werbos, P. (199). Bacpropagation through time: What it does and how to do it. Proceedings of the IEEE, Vol. 78(1), October, pp. 1545 168. Werbos, P. (199). Neurocontrol and supervised learning: An overview and evaluation. Handboo of Intelligent Control: Neural, Fuzzy and Approaches. (D. White and D. Sofge editors), chapter 3. Van Nostrand Reinhold, New Yor. Widrow, B. and S. D. Stearns (1985). Signal Processing. Prentice Hall, Englewood Cliffs, NJ. Widrow, B. and E. Walach (1996). Inverse Control. Prentice Hall PTR, Upper Saddle River, NJ. Williams, R. J. and D. Zipser (1989). Experimental analysis of the realtime recurrent learning algorithm. Connection science, Vol. 1(1), pp.87 111.