Embedding Recurrent Neural Networks into Predator-Prey Models Yves Moreau, St phane Louies 2, Joos Vandewalle, L on renig 2 Katholieke Universiteit Le

Size: px
Start display at page:

Download "Embedding Recurrent Neural Networks into Predator-Prey Models Yves Moreau, St phane Louies 2, Joos Vandewalle, L on renig 2 Katholieke Universiteit Le"

Transcription

1 Katholieke Universiteit Leuven Departement Elektrotechniek ESAT-SISTA/TR Embedding Recurrent Neural Networks into Predator-Prey Models Yves Moreau 2, Stephane Louies 3, Joos Vandewalle 2, Leon renig 3 January 998 This report is available by anonymous ftp from ftp.esat.kuleuven.ac.be in the directory pub/sista/moreau/reports/. 2 ESAT - Katholieke Universiteit Leuven, Kardinaal Mercierlaan 94, 3 Leuven (Heverlee), elgium, tel +32/6/32 8 6, fax +32/6/ , yves.moreauesat.kuleuven.ac.be, www: 3 SO - Universite Libre de ruxelles, ampus Plaine, P 234, lvd du Triomphe, -5 ruxelles, elgium, tel +32/2/ , fax +32/2/65 5 3, lbrenigulb.ac.be, www:

2 Embedding Recurrent Neural Networks into Predator-Prey Models Yves Moreau, St phane Louies 2, Joos Vandewalle, L on renig 2 Katholieke Universiteit Leuven Department of Electrical Engineering Kardinaal Mercierlaan 94-3 Leuven, elgium 2 Universit Libre de ruxelles oulevard du Triomphe -5 ruxelles, elgium Running Title: Embedding Recurrent Neural Networks Acknowledgments Yves Moreau is a Research Assistant with the F.W.O.-Vlaanderen. Joos Vandewalle is Full Professor at the K.U.Leuven. Their work is supported by several institutions: the Flemish Government, GOA-MIPS (Model-based Information Processing Systems); the FWO Research ommunity, IoS (Identication and ontrol of omplex Systems); the elgian State, DWT - Interuniversity Poles of Attraction Programme (IUAP P4-2 (997-2): Modeling, Identication, Simulation and ontrol of omplex Systems. L on renig and St phane Louies are grateful to the ATHODE ESPRIT Working Group (WG #2449) for scientic and nancial support. They would like also to thank A. Figueiredo and T. M. da Rocha from the Physics Departement of the University of rasilia for fruitful scientic discussions. January 998 Requests for reprints should be addressed to Yves Moreau, Department of Electrical Engineering, Katholieke Universiteit Leuven, -3 Leuven, elgium. address: yves.moreauesat.kuleuven.ac.be.

3 Abstract We study changes of coordinates that allow the embedding of the ordinary dierential equations describing continuous-time recurrent neural networks into dierential equations describing predator-prey models also called Lotka-Volterra systems. We do this by transforming the equations for the neural network rst into quasi-monomial form (renig, 988), where we express the vector eld of the dynamical system as a linear combination of products of powers of the variables. In practice, this transformation is possible only if the activation function is the hyperbolic tangent or the logistic sigmo d. From this quasi-monomial form, we can directly transform the system further into Lotka-Volterra equations. The resulting Lotka-Volterra system is of higher dimension than the original system, but the behavior of its rst variables is equivalent to the behavior of the original neural network. We expect that this transformation will permit the application of existing techniques for the analysis of Lotka-Volterra systems to recurrent-neural networks. Furthermore, our result shows that Lotka-Volterra systems are universal approximators of dynamical systems, just as continuous-time neural networks. Keywordsontinuous-time neural networks, Equivalence of dynamical systems, Lotka-Volterra systems, Predator-prey models, Quasi-monomial forms. 2

4 Introduction Although major advances have been made, the eld of application of dynamical system theory to recurrent neural networks still has much uncharted territory. ohen and Grossberg studied a large class of competitive systems (which include many types of neural networks) (ohen & Grossberg, 983) with a symmetric interaction matrix for which they were able to construct a Lyapunov function guaranteeing global convergence. Hirsch considered cascades of networks and proved a cascade decomposition theorem (Hirsch, 989) giving conditions under which convergence of the individual networks in isolation guarantees the convergence of the whole cascade. For systems with inputs, Sontag, Sussman, and Albertini have studied their property from the point of view of nonlinear control system theory (Albertini & Sontag, 993, Sontag & Sussman, 997). ut for the analysis of many properties of recurrent networks, simple computational tools are still missing. Recently, we have started to investigate how changes of coordinates allow us to nd equivalences between dierent neural networks. We have shown in (Moreau & Vandewalle, 997) how to use linear changes of coordinates to compute a transformation that maps a continuous-time recurrent neural network with a hidden-layer onto a neural network without hidden-layer and with an extra output map. In this paper, we will show how to embed dierent types of neural networks (having the hyperbolic tangent or the logistic sigmo d as activation function) into predator-prey models (also called Lotka-Volterra systems) of the form: _z i = i z i + z i m X M ij z j ; with j = ; : : : ; m: The Lotka-Volterra system will be of higher dimension than the original system but its n rst variables will have a behavior equivalent or identical to that of the original system (if n was the number of variables of the original system). Such a Lotka-Volterra representation of neural networks is of interest for several reasons. Lotka-Volterra systems are a central tool in mathematical biology for the modeling of competition between species and a classical subject of dynamical system theory (MacArthur, 969, May & Leonard, 975). They have therefore been the object of intense scrutiny. These systems have simple quadratic nonlinearities (not all quadratic terms are possible in each equation), which might be easier to analyze than the sigmo dal saturations of the neural networks. We thus expect that this representation will allow us to apply methods and results from the study of Lotka-Volterra systems to neural networks. Indeed, early work by Grossberg (Grossberg, 978) exploited one of the changes of variables that we present here to study convergence and oscillations in generalized competitive networks and generalized Lotka-Volterra systems. Furthermore, since we know that dynamical neural networks can approximate arbitrary dynamical systems (Sontag, 992, Funahashi & Nakamura, 993) for any nite time because of their equivalence with dynamical single-hidden-layer perceptrons (Moreau & Vandewalle, 997), our result also serves as a simple 3

5 proof that Lotka-Volterra systems enjoy the same approximation property; and we might want to investigate them further from that point of view. We have structured this work as follows. We will rst consider the simple case of a dynamical perceptron. We will transform it into what we call a quasimonomial form in a rst step. This form will allow us to transform the system further into a Lotka-Volterra system. Second, we will consider the case of the dynamical perceptron with self-connections that is, where each neuron has a linear connection to itself. We will perform the same transformations rst into a quasi-monomial form and then into a Lotka-Volterra system. Finally, we will show that the same transformations apply to much more general types of continuous-time recurrent neural networks with hyperbolic tangent or logistic activation functions. In Appendix A and, we describe in details the structure of the quasi-monomial forms for the dynamical perceptron and the dynamical perceptron with self-connections. In Appendix, we clarify what the exact constraints on the activation function are and we show that both the hyperbolic tangent and the logistic sigmo d satisfy these. 2 The dynamical perceptron We consider rst an n-dimensional neural network of the following form, which we call a dynamical perceptron: where x 2 R n ; A 2 R nn ; a 2 R n ; and ~ is dened as _x = ~(Ax + a ); () ~(x) = ((x ); (x 2 ); : : : ; (x n )) T with a continuous sigmo dal function from R to R such that lim x!? (x) = a and lim x! (x) = b (a; b 2 R and a < b). We consider this type of system because its simplicity makes our calculations easier to follow, yet it permits us to illustrate the essential points of our method. In this paper, we do all developments for the special case of (x) = tanh(x). Another possibility is the use of the logistic sigmo d (x) = =( + e?x ). We will later discuss what constrains the choice of the activation function. For the dynamical perceptron (), the choice of the logistic sigmo d would be meaningless since the vector eld would then have positive components everywhere and the topology of the ow would be trivial (the hyperbolic tangent guarantees on the other hand the existence of an equilibrium). For the types of networks we describe in the following sections, this restriction to the hyperbolic tangent does not apply. ut since the developments depend on the choice of the activation function, we limit ourselves to the hyperbolic tangent; all developments can be done similarly for the logistic sigmo d. We rst apply the change of coordinates y = Ax + a (often called Sigma-Pi transformation (Grossberg, 988)) and get the following equation: _y = A(y): 4

6 As long as A is invertible (A is n n), we can directly revert the change of coordinates. Moreover, if A is singular, we in fact have an r-dimensional dynamics where r is the rank of the matrix A. Indeed, if we look at two dierent trajectories x(t) and ~x(t) of the system _x = (Ax+a ) having the property that their initial conditions x and ~x are such that Ax = A~x, the initial conditions y = Ax + a and ~y = A~x + a for the system _y = A(y) are equal. This means that y(t) = ~y(t) for all times. ut the vector eld of the initial system is _x = (Ax + a ) = (y), thus _x(t) = _~x(t) for all times. This means that the trajectories x(t) and ~x(t) through x and ~x always remain parallel. Since the kernel of A is of dimension n? r, this means that the dynamics is in fact r-dimensional. If we dene (x ) = x? A y Ax where A y is the Moore-Penrose pseudo-inverse of A, then x(t) = A y (y(t)? a ) + (x ) with the initial condition y = Ax + a. From here on, it will be necessary to look at the equations coordinate by coordinate, so that we have _y i = A ij (y j ); with i = ; : : : ; n: (2) Now, we decide to apply the change of coordinates z i = (y i ); we get _z i = (y i ) _y i : (3) It is here that we make use of our special choice of activation function, (y) = tanh(y): The essential property of the activation function we need is that we can express its derivative as a polynomial of the activation function: (y) = tanh (y) = = cosh 2 (y) =? tanh 2 (y) =? 2 (y): (4) Substituting (4) and (2) in Equation 3, we have _z i = (? z 2 i ) n X A ij z j : (5) We see that this form begins to resemble the Lotka-Volterra model except for the factor (? z 2 i ) instead of z i. This means that the system is polynomial of degree three. However, the vector eld of the system has now become a linear combination of products of powers of the variables, which is what we call a quasi-monomial form. Example We will follow our developments at the hand of an example to clarify the procedure. We look at the case n = 3. The equations of the recurrent neural network are 8 < : _x = tanh(a x + A 2 x 2 + A 3 x 3 + A ) _x 2 = tanh(a 2 x + A 22 x 2 + A 23 x 3 + A 2 ) _x 3 = tanh(a 3 x + A 32 x 2 + A 33 x 3 + A 3 ) 5

7 We can check that the equations become, after the ane and sigmo dal changes of coordinates, 8 < : _z = (? z 2)(A z + A 2 z 2 + A 3 z 3 ) _z 2 = (? z 2)(A 2 2z + A 22 z 2 + A 23 z 3 ) _z 3 = (? z3 2 )(A 3z + A 32 z 2 + A 33 z 3 ) 3 Transformation into a quasi-monomial form y quasi-monomial form, we mean a dynamical system that we can write as _z i = z i ( rx (6) A ij n Y k= z jk k ); with i = ; : : : ; n: (7) The number of dierent products of powers of the variables is r and we dene 2 R rn. It is important to note that the exponents jk are not restricted to integer values, hence the name quasimonomial form. Now, to reach this form, we write Equation 5 as A = A ij 2 R nr and = jk _z i = z i [ A ij (z? i z j? z i z j )]: (8) We then need to collect all possible monomials and use their exponents to construct the matrix and collect all the coecients of the monomials correspondingly to form the matrix A. We do this in a systematic fashion in Appendix A; and we illustrate it on the following example. renig (renig, 988) has previously shown that we can transform such a quasi-monomial form into a Lotka-Volterra system. Example (cont'd) To illustrate this somewhat obtuse notation (7), we take again the case n = 3. Expanding the expression (6), we get 8 < : _z = z (A? A z 2 + A 2z? z 2? A 2 z z 2 + A 3 z? z 3 + A 3 z z 3 ) z 2 3? A 23 z 2 z 3 ) 3? A 32 z 2 z 3 + A 33? A 33 z3 2) _z 2 = z 2 (A 2 z z? 2? A 2 z z 2 + A 22? A 22 z A 23z? _z 3 = z 3 (A 3 z z? 3? A 3 z z 3 + A 32 z 2 z? We see that there are 2n 2 = 8 terms, but only r = (3n 2? n + 2)=2 = 3 dierent monomials. The matrix is then = while the matrix A is A = 2?? 2?? 2?? T A ;?A?A 2?A 3 A 2 A 3 A?A 2?A 22?A 23 A 2 A 23 A 22?A 3?A 32?A 33 A 3 A 32 A 33 6 A :

8 How should we interpret these matrices? If we look at the rst row of, we have a row vector (2; ; ) (note that the notation above uses a transposition); and if we look at the rst column of A, we have a column vector (?A ; ; ) T. This means that the expression for the system contains a monomial of the form zz 2 2z 3 = z2 ; and that it appears in the rst equation with a coecient of?a and is absent from the second and third equations. hecking this for the whole P3 Q matrices, we indeed have _z i = z 3 i A ij k= z jk with i = ; 2; 3: 4 Transformation into a Lotka-Volterra system The whole transformation relies now on a simple trick: We look at each product of powers of the variables that appear in the quasi-monomial form as a new variable of our dynamical system. With r the number of dierent monomials in the quasi-monomial form, we dene the supplementary variables z n+ ; : : : ; z n+r as z n+k = ry l= z kl l ; with k = ; : : : r: If we now look at the system of n + r variables _z i, taking into account these new denitions, we have for the rst n variables _z i = z i rx A ij z n+j ; with i = ; : : : n: ecause of our special denition of the new variables, many simplications occur in the expression of their time derivative, as seen in Equation 9. We rst use the formula for the derivative of a product (). Then we try to reconstruct the original variables since they are almost preserved by the derivation (); the necessary correction leads to the appearance of a logarithmic derivative (2), which in turn transforms the exponent of the monomials into multiplicative coecients (3). Lastly, the remaining correction produces the new expression of the vector elds of the rst n variables, except for the disappearance of the multiplicative variable at the front (4). Writing these computations down, we have, for the r supplementary variables, _z n+j = ( = = ny k= l= l= ( k z jk k ) with j = ; : : : ; r (9) ny k = k 6= l z jk k )(z jl l ) () z n+j =z jl l (z jl l ) () 7

9 X n = z n+j (ln(z jl l )) (2) l= = z n+j n X l= jl _z l =z l (3) = z n+j n X l= = z n+j rx m= jl (( l= rx m= A lm z n+m (4) jl A lm )z n+m ) (5) So, we can regroup all these terms by dening the (n + r) (n + r) matrix M as O nn A M = O rn A since P n l= jl A lm = ( A) jm ; and the total system becomes then of the form X n+r _z i = z i M ij z j ; with i = ; : : : ; n + r; which is precisely an n + r-dimensional Lotka-Volterra system! Supposing we were to have a solution z(t) = (z (t); z 2 (t); : : : ; z n+r (t)) T, we could then retrieve the behavior of the original system by transforming back the n rst variables (if the matrix A is invertible): x(t) = A? (~? ((z (t); : : : ; z n (t)) T )? a ): If the matrix is not invertible, we can apply the argument from Section 2 to retrieve the solution. 5 The dynamical perceptron with self-connections One criticism that we can formulate about the dynamical perceptron is that its global behavior is often similar to the behavior around its xed point. Although we can obtain complex dynamics by a careful choice of parameters, this is a delicate operation. The dynamics of such systems appear to be essentially limited. One way to obtain complex oscillatory behaviors more easily is to add a proportional self-connection to each variable: _x i =? i x i + ( A ij x j + a i ): (6) The reason why we can obtain oscillatory behaviors more easily is that we can choose A to make one of the xed points unstable and then choose the i 8

10 suciently large so that the behavior of the system always remain bounded. A further remark is that this kind of system can have multiple isolated equilibria, while the previous system either had to have a single equilibrium or a connected innity of equilibria. The dynamics of the systems with inhibitory connections is thus richer. Also, the activation function for this system is not limited to the hyperbolic tangent anymore; in Appendix, we discuss other possible choices and conclude that we could use the logistic sigmo d as well. We will embed this system into a Lotka-Volterra system. The rst step is to dene the following auxiliary variables: w i = ( A ij x j + a i ): Looking at the derivative of w, we then have the following dynamics for the system of x and w variables: _x i = w i? i x i _w i = (? w 2 i ) n X A ij (w j? j x j ) This dynamics is essentially in quasi-monomial form and we can rewrite it as: _x i = x i (? i + x? i w i ) X n _w i = w i (? j A ij x j w? i? j A ij x j w i + A ij w? i w j + A ij w i w j ) In Appendix, we show how to write the matrices A and for the quasimonomial form of this system. Again, we simply have to carefully collect all the exponents of the monomials in the matrix and all their coecients in the matrix A. After this transformation into quasi-monomial form, the transformation of Section 4 directly applies in the same fashion except that the variable x is directly part of the Lotka-Volterra system and does not need to be retrieved by a change of coordinates (this advantage is paid by having a larger Lotka-Volterra system than for the dynamical perceptron). 6 Embedding a dynamical multi-layer perceptron into a Lotka-Volterra system The transformations we have showed for the dynamical perceptron and the dynamical perceptron with self-connections apply also to the dynamical Multi- Layer Perceptron (MLP). To avoid an unnecessary clutter of notation, we will take the specic example of a two-hidden layer continuous-time recurrent MLP: _x = (((Ax + a ) + b ) + c ) 9

11 where we have n states, p hidden units in the rst layer, q hidden units in the second layer, and the activation function is the hyperbolic tangent. This means x(t) 2 R n ; A 2 R pn ; 2 R qp ; 2 R nq ; a 2 R p ; b 2 R q ; c 2 R n. The argument applies to any kind of MLP by doing the same type of computations again. If we write this coordinate by coordinate, we get _x i = ( qx l= il ( px k= lk ( A kj x j + a k ) + b l ) + c i ); with i = ; : : : ; n: We dene the following vector of p new variables: y = (Ax + a ). Again, we will want to discuss the rank of the matrix A. If the matrix A is of rank n, then we can uniquely retrieve x from y = (Ax +a ). If the rank r of the matrix A is inferior to n, we have the same argument as in Section 2: Two initial conditions x and ~x such that Ax = A~x will have the same dynamics in the new variables y(t) = ~y(t) for all times, thus x(t) = (( y(t) + b ) + c ) = ~x(t) for all times. This means that the trajectories x(t) and ~x(t) through x and ~x always remain parallel. Since the kernel of A is of dimension n?r, this means that P the dynamics p is in fact r-dimensional. The dynamics of the variables y i = ( A ijx j +a i ) is thus _y i = ( A ij x j + a i )( = (? y 2 i )( n X A ij ( qx l= A ij _x j ) = (? 2 ( jl ( px k= lk y k + b l ) + c j )) A ij x j + a i ))( A ij _x j ) We can remark that the form of the dynamics of y closely resembles that of the dynamics of x; but that the rst hidden layer of the MLP has disappeared. This encourages us to dene, in the same fashion as before, a P new vector of q variables: p z = (y + b ). The dynamics of the variables z i = ( k= iky k + b i ) is thus _z i = ( px k= = (? z 2 i )( px ik y k + b i )( px k= k= ik (? y 2 k )( n X px ik _y k ) = (? zi 2 )( ik _y k ) A kj ( qx l= k= jl z l + c j ))) Enthusiastically, we pursue further by dening a new vector of n variables: w = (z + c ). The dynamics of the variables w i = ( P q l= ilz l + c i ) is thus _w i = ( qx l= il z l + c i )( qx l= qx px _w i = (? wi 2 )( il (? zl 2 )( l= qx il _z l ) = (? wi 2 )( il _z l ) l= k= lk (? y 2 k )( n X A kj w j )))

12 Now, substituting the denitions for y; z; w in the previous equations and noting that, because of the same denitions, _x = w, we get the following system: _x i = w i _y i = (? y 2 i )( n X A ij w j ) _z i = (? z 2 i )( px X n ik (? yk 2 )( k= qx px _w i = (? wi 2 )( il (? zl 2 )( l= A kj w j )) k= lk (? y 2 k )( n X A kj w j ))) The dynamics of the rst n variables of this (n + p + q + n)-dimensional system (with appropriate initial conditions: x = x ; y = (Ax + a ); z = (y + b ); w = (z + c )!) will be identical to the dynamics of the original system. Further, we can write this system in quasi-monomial form: _x i = x i (x i? w i ) _y i = y i ( _z i = z i ( _w i = w i ( px k= qx A ij (y? i w j? y i w j )) px l= k= ik A kj (z? i w j? y 2 kz? i w j? z i w j + y 2 kz i w j )) il lk A kj (? y 2 k )(? z2 l )(w? i? w i )w j ) We see that we have written the ordinary dierential equation for each variable as a product of the variable itself by a linear combination of products of the variables, which is precisely the denition of the quasi-monomial form. A last remark is that the convergence of x implies the convergence of the enlarged system (x; y; z; w) since _x = implies _y = ; _z =, and _w =. The main dierence is that, to check the convergence of the original system, we only need to check that all the trajectories of the enlarged system starting on the manifold of initial conditions m(x ) = (x ; (Ax + a ); ((Ax + a ) + b ); (((Ax + a ) + b ) + c )) converge. The transformation into a Lotka-Volterra system proceeds exactly in the same way as in Appendix A and. However, writing down the matrices A and is a lengthy and tedious procedure and does not bring any further insight. We therefore do not perform it here. This computation will be helpful only when we will use a computer algebra system to analyze the resulting Lotka-Volterra system.

13 7 Discussion and conclusion An important corollary of this embedding of continuous-time recurrent multilayer perceptrons into higher-dimensional Lotka-Volterra systems is that we can use Lotka-Volterra systems to approximate arbitrarily closely the dynamics of any nite-dimensional dynamical system for any nite time. The argument for this universal approximation property is exactly the same as in (Funahashi & Nakamura, 993) and says that, for any given nite time, the trajectories of two suciently close dynamical systems remain close (this is true only for a nite time); this statement is based on the classical use of the Gronwall inequality for dynamical systems. It could therefore be interesting to use Lotka-Volterra systems as approximators for dynamical systems. The restriction presented in Appendix on the choice of the activation function (which is in practice limited to the hyperbolic tangent or the logistic sigmo d) is theoretically strong: There is no guarantee that systems with dierent choices of activation function have qualitatively the same behavior. Indeed, Sontag and Sussman (Sontag & Sussman, 997) have shown that controllability properties of systems with inputs could dier, even for dierent types of monotone increasing sigmo ds. However, the limitation is fairly weak in the sense that the possible activation functions are the ones most commonly used and that we have universal approximation properties for dynamical systems. We have presented techniques to embed dierent types of neural networks (dynamical perceptrons, dynamical perceptrons with self-connections, and multilayer perceptrons) into predator-prey (Lotka-Volterra) systems. These techniques relied on looking at the dynamics of the output of every neuron in the network and using the special properties of the derivative of the hyperbolic tangent or logistic sigmo d activation function to write the system as what we called a quasi-monomial form that is a dynamical system where the vector eld is a linear combination of products of powers of the state variables. Once we had this quasi-monomial form, we looked at the dynamics of these products of powers of the state variables to obtain directly a Lotka-Volterra system. Our expectation is that these techniques will be useful to migrate techniques for the analysis of Lotka-Volterra systems into the eld of dynamical neural networks for example, techniques to study convergence or stability properties of dynamical systems, or discover rst integrals of motion. This is our next objective. References Albertini, F., & Sontag, E.D. (993). For Neural Networks Function Determines Form, Neural Networks, 6, renig, L. (988). omplete Factorization and Analytic Solutions of Generalized Lotka-Volterra Equations. Physics Letters A, 33 (78),

14 ohen, M. A., & Grossberg, S. (983). Absolute Stability of Global Pattern Formation and Parallel Memory Storage by ompetitive Neural Networks. IEEE Transactions on Systems, Man, and ybernetics, SM-3 (5), Funahashi, K.-I., & Nakamura, Y. (993). Approximation of Dynamical Systems by ontinuous-time Recurrent Neural Networks. Neural Networks, 6, 886. Grossberg, S. (978). Decisions, Patterns, and Oscillations in Nonlinear ompetitive Systems with Applications to Volterra-Lotka Systems. Journal of Theoretical iology, 73, 3. Grossberg, S. (988). Nonlinear Neural Networks: Principles, Mechanisms, and Architectures. Neural Networks, (), 76. Hirsch, M. W. (989). onvergent Activation Dynamics in ontinuous Time Networks. Neural Networks, 2, MacArthur, R. H. (969). Species Packing, or what ompetition Minimizes. Proceedings of the National Academy of Sciences, USA, 54, May, R. M., & Leonard, W. J. (975). Nonlinear Aspects of ompetition between Three Species. SIAM Journal of Applied Mathematics, 29, Moreau, Y., & Vandewalle, J. (997). When Do Dynamical Neural Networks with and without Hidden Layer Have Identical ehavior? (Tech. Rep. ESAT- SISTA TR97-5). Katholieke Universiteit Leuven: Department of Electrical Engineering, elgium. Sontag, E. D. (992). Neural Nets as System Models and ontrollers. Proceedings Seventh Yale Workshop on Adaptive and Learning Systems, Sontag, E. D., & Sussman, H. J. (997). omplete ontrollability of ontinuous-time Recurrent Neural Networks, Systems and ontrol Letters, 3, Appendix A: Transformation of the dynamical perceptron into a quasi-monomial form We derive here the structure of the matrices A and of the quasi-monomial form for the dynamical perceptron in a systematic fashion. Again, we write the equations as _z i = z i [ A ij (z? i z j? z i z j )]: (7) We see that the expression between brackets contains 2n monomials; for the n dierential equations this gives a total of 2n 2 terms. There are two types 3

15 of monomials: z i z j and z i? z j. The rst type of monomial can take two different forms: zi 2 (when i = j) or z iz j with j > i (when i 6= j). There are n monomials of the rst form and n(n? )=2 monomials of the second form. The second type of monomials can also take two dierent forms: (when i = j) or z i? z j (when i 6= j). There is one monomial of the rst form and n(n? ) monomials of the second form. This gives a total of (3n 2? n + 2)=2 dierent monomials to share among 2n 2 terms. arefully collecting all these terms into row vectors of exponents to form the matrix and all their corresponding coecients into column vectors to form the matrix A, we can write these matrices as = h ()T (2)T : : : (n)t D ()T D (2)T : : : D (n)t O nti T (8) A = h i S () S (2) : : : S (n) R () R (2) : : : R (n) V (9) where we dene the sub-matrices (i) ; D (i) ; O n ; and P (i) ; R (i) ; V as follows. In general, the matrix O nm is a matrix of zeros of dimension n m and the matrix I nm is a diagonal matrix of ones of dimension n m. The matrix (i) represents all the monomials of the form zi 2 and z i z j (with j > i) and thus the dimension of the matrix depends on the index i with the matrix n? i + rows and n columns. We can write it as (i) = : : : 2 : : : O (n?i)(i?). I (n?i)(n?i) A : Each of its rows is associated with the corresponding column of the matrix P (i), which collects all the coecients of the monomials zi 2 and z i z j (with j > i). Such a matrix has again variable dimension with n rows and n? i + columns. We can write it as P (i) = O (i?)(n?i+)?a ii?a i(i+)?a i(i+2) : : :?A in?a (i+)i : : :.?A (i+2)i : : :?A ni Turning back to the matrix of exponents, we collect all the exponents of the monomials of the form z? i z j (with j 6= i) in the matrix D (i) which has always A : 4

16 n? rows and n columns: D (i) = I (i?)(i?)?. O (n?i)(i?)..? O (i?)(n?i) I (n?i)(n?i) A : We collect the coecients of these monomials in the corresponding columns of R (i), which also has n? rows and n columns: R (i) = O (i?)(n?) A i : : : A i(i?) A i(i+) : : : A in O (n?i)(n?) We are now only left with the constant monomial (which we represent by the row vector O n ) and its coecients, which we collect in the column vector V = A ; : : : ; A nn T : Taking all this notation into consideration, we can represent the matrices A and according to Equations 8 and 9. Example We now present the dierent matrices for the case of a four dimensional system. This claries how we have collected the dierent terms in the matrices A and. The sub-matrices are () ; (2) ; (3) ; (4) and D () ; D (2) ; D (3) ; D (4) for : () = (3) = D () = D (3) = 2 2?????? A ; (2) = 2 A : A ; ; (4) =? 2 ; A ; D (2) = A ; D (4) =?????? A ; A : 5

17 We have decomposed the matrix A into the sub-matrices P () ; P (2) ; P (3) ; P (4) ; R () ; R (2) ; R (3) ; R (4) ; and V : P () = P (2) =?A?A 2?A 3?A 4?A 2?A 3?A 4 R () = R (3) =?A 22?A 23?A 24?A 32?A 42 A 2 A 3 A 4 A 3 A 32 A 34 A ; P (3) = A ; P (4) = A ; R(2) = A ; R(4) = V = (A ; A 22 ; A 33 ; A 44 ) T :?A 33?A 34?A 43?A 44 A 2 A 23 A 24 A 4 A 42 A 43 A ; A ; A ; A ; Appendix : Transformation of the dynamical perceptron with self-connections into a quasi-monomial form We recall that for the system (6), the quasi-monomial form is _x i = x i (? i + x? i w i ) _w i = w i ( P n (? ja ij x j w? i? j A ij x j w i + A ij w? i w j + A ij w i w j )) (2) To use the quasi-monomial formalism in the same way as in the previous section, we need to unify our formalism to use a single set of variables. We therefore dene the following variable u = (x; w) T : u i = x i ; with i = ; : : : ; n u n+i = w i ; with i = ; : : : ; n If the matrices A () and () contain respectively all the coecients and all the exponents of the dierent monomials, we can write the dynamics of u in 6

18 quasi-monomial form: _u i = u i? m X A () ij 2nY k= u () jk k : To build the matrices A () and (), we have to look at all the monomials in (2). If we rst focus on the terms w i? w j and w i w j, which already appeared in Section 3, we see that we have exactly the same contribution in the matrix () as in the case of the dynamical perceptron. We have the same cancellation of the factors of the terms in w i? w j or w i w j as in Section 3. They will contribute to the matrix () with a rst sub-matrix equal to. Notice that the left part of the matrix () is for powers of x while the right part is for powers of w. The next blocks take respectively into account the monomials of the form x i? w i ; x i w j? ; and x i w j. None of these terms ever adds up or cancels out, which explains the plain diagonal and column structure of these sub-matrices. We have () =?I nn I nn I nn?k () O rn. I nn.?k (n) I nn K (). I nn. K (n) A ; where we dene K (i) as K (i) = O n(i?). O n(n?i) A : Similarly, we collect all the coecients of the monomials in A ().We place the contributions relating to the equations for the x variables in the top part of the matrix while we place the contributions for w in the bottom part of the matrix. The contributions for the monomials w i? w j and w i w j are exactly the same for the variables w as before; the rst bottom sub-matrix is therefore equal to A. These monomials have no inuence on the variables x and the rst top sub-matrix is therefore equal to zero; except in the case where w i? w j = corresponding to the? i terms in the equations for x, which generate a column of? i in the otherwise empty matrix Q. The other sub-matrices take correspondingly into account the coecients of the monomials of the form x? i w i ; x i w j? ; and x i w j. We thus have A () = Q nr I nn O nn : : : O nn O nn : : : O nn A O nn ;?L () : : :?L (n) L () : : : L (n) 7

19 where we dene Q and L (i) as Q = O n(r?)?.? n A ; L (i) = O (i?)(n?) A i : : : A in n O (n?i)(n?) A : Appendix : hoice of the activation function We chose the activation function as (x) = tanh(x) for simplicity. ut we are in not directly limited to this activation function. The essential property that we used was the fact that (x) =? 2 (x). For the logistic sigmo d (x) = =(+e?x ), we have (x) = (x)(?(x)) and we can do the calculations again without diculty. The condition that the activation function has to satisfy is that its derivative be a sum of quasi-monomials of itself. If we do not require saturation at plus and minus innity, we can use (x) = e x, but the system is not globally Lipschitz and the solution might blow-up in nite time, which is an annoyance. A square function (x) = x 2 suers from the same problem. The hyperbola (x) = =x gives (x) =?= 2 (x), but then the vector eld is not even bounded, which is even worse. In general, we could consider the P condition (x) = L((x)), where L(:) is a nite Laurent series L(y) = (=y m n ) a k= ky k, or even more generally, a Puiseux series. For the Laurent series, the solution of the dierential equation is given by separation of variables m (x) d(x) = L((x)) dx. We can then expand the term m (x)=l((x)) as a polynomial plus a sum of simple fractions, and integrate them. This integration produces a rational function plus a sum of logarithmic terms and arctangent terms and (x) is found by inverting this relation. This inversion will not lead to an explicit solution in general. If we further require that the vector eld be globally Lipschitz to guarantee existence of the trajectories for all times, the possible solutions become scarce. From the most frequently used activation functions, we have been able to identify only the hyperbolic tangent and the logistic sigmo d as satisfying these conditions. 8

Departement Elektrotechniek ESAT-SISTA/TR Dynamical System Prediction: a Lie algebraic approach for a novel. neural architecture 1

Departement Elektrotechniek ESAT-SISTA/TR Dynamical System Prediction: a Lie algebraic approach for a novel. neural architecture 1 Katholieke Universiteit Leuven Departement Elektrotechniek ESAT-SISTA/TR 1995-47 Dynamical System Prediction: a Lie algebraic approach for a novel neural architecture 1 Yves Moreau and Joos Vandewalle

More information

Optimal trac light control for a single intersection Bart De Schutter and Bart De Moor y ESAT-SISTA, KU Leuven, Kardinaal Mercierlaan 94, B-3 Leuven (

Optimal trac light control for a single intersection Bart De Schutter and Bart De Moor y ESAT-SISTA, KU Leuven, Kardinaal Mercierlaan 94, B-3 Leuven ( Katholieke Universiteit Leuven Departement Elektrotechniek ESAT-SISTA/TR 97- Optimal trac light control for a single intersection Bart De Schutter and Bart De Moor Proceedings of the 997 International

More information

Departement Elektrotechniek ESAT-SISTA/TR About the choice of State Space Basis in Combined. Deterministic-Stochastic Subspace Identication 1

Departement Elektrotechniek ESAT-SISTA/TR About the choice of State Space Basis in Combined. Deterministic-Stochastic Subspace Identication 1 Katholieke Universiteit Leuven Departement Elektrotechniek ESAT-SISTA/TR 994-24 About the choice of State Space asis in Combined Deterministic-Stochastic Subspace Identication Peter Van Overschee and art

More information

The Linear Dynamic Complementarity Problem is a special case of the Extended Linear Complementarity Problem B. De Schutter 1 B. De Moor ESAT-SISTA, K.

The Linear Dynamic Complementarity Problem is a special case of the Extended Linear Complementarity Problem B. De Schutter 1 B. De Moor ESAT-SISTA, K. Katholieke Universiteit Leuven Departement Elektrotechniek ESAT-SISTA/TR 9-1 The Linear Dynamic Complementarity Problem is a special case of the Extended Linear Complementarity Problem 1 Bart De Schutter

More information

x 104

x 104 Departement Elektrotechniek ESAT-SISTA/TR 98-3 Identication of the circulant modulated Poisson process: a time domain approach Katrien De Cock, Tony Van Gestel and Bart De Moor 2 April 998 Submitted for

More information

only nite eigenvalues. This is an extension of earlier results from [2]. Then we concentrate on the Riccati equation appearing in H 2 and linear quadr

only nite eigenvalues. This is an extension of earlier results from [2]. Then we concentrate on the Riccati equation appearing in H 2 and linear quadr The discrete algebraic Riccati equation and linear matrix inequality nton. Stoorvogel y Department of Mathematics and Computing Science Eindhoven Univ. of Technology P.O. ox 53, 56 M Eindhoven The Netherlands

More information

Contents. 2.1 Vectors in R n. Linear Algebra (part 2) : Vector Spaces (by Evan Dummit, 2017, v. 2.50) 2 Vector Spaces

Contents. 2.1 Vectors in R n. Linear Algebra (part 2) : Vector Spaces (by Evan Dummit, 2017, v. 2.50) 2 Vector Spaces Linear Algebra (part 2) : Vector Spaces (by Evan Dummit, 2017, v 250) Contents 2 Vector Spaces 1 21 Vectors in R n 1 22 The Formal Denition of a Vector Space 4 23 Subspaces 6 24 Linear Combinations and

More information

Vector Space Basics. 1 Abstract Vector Spaces. 1. (commutativity of vector addition) u + v = v + u. 2. (associativity of vector addition)

Vector Space Basics. 1 Abstract Vector Spaces. 1. (commutativity of vector addition) u + v = v + u. 2. (associativity of vector addition) Vector Space Basics (Remark: these notes are highly formal and may be a useful reference to some students however I am also posting Ray Heitmann's notes to Canvas for students interested in a direct computational

More information

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) Contents 1 Vector Spaces 1 1.1 The Formal Denition of a Vector Space.................................. 1 1.2 Subspaces...................................................

More information

Ecient Higher-order Neural Networks. for Classication and Function Approximation. Joydeep Ghosh and Yoan Shin. The University of Texas at Austin

Ecient Higher-order Neural Networks. for Classication and Function Approximation. Joydeep Ghosh and Yoan Shin. The University of Texas at Austin Ecient Higher-order Neural Networks for Classication and Function Approximation Joydeep Ghosh and Yoan Shin Department of Electrical and Computer Engineering The University of Texas at Austin Austin, TX

More information

Master-Slave Synchronization using. Dynamic Output Feedback. Kardinaal Mercierlaan 94, B-3001 Leuven (Heverlee), Belgium

Master-Slave Synchronization using. Dynamic Output Feedback. Kardinaal Mercierlaan 94, B-3001 Leuven (Heverlee), Belgium Master-Slave Synchronization using Dynamic Output Feedback J.A.K. Suykens 1, P.F. Curran and L.O. Chua 1 Katholieke Universiteit Leuven, Dept. of Electr. Eng., ESAT-SISTA Kardinaal Mercierlaan 94, B-1

More information

Discrete time Generalized Cellular. Johan A.K. Suykens and Joos Vandewalle. Katholieke Universiteit Leuven

Discrete time Generalized Cellular. Johan A.K. Suykens and Joos Vandewalle. Katholieke Universiteit Leuven Discrete time Generalized Cellular Neural Networks within NL q Theory Johan A.K. Suykens and Joos Vandewalle Katholieke Universiteit Leuven Department of Electrical Engineering, ESAT-SISTA Kardinaal Mercierlaan

More information

Linear Regression and Its Applications

Linear Regression and Its Applications Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start

More information

Linear Algebra (part 1) : Matrices and Systems of Linear Equations (by Evan Dummit, 2016, v. 2.02)

Linear Algebra (part 1) : Matrices and Systems of Linear Equations (by Evan Dummit, 2016, v. 2.02) Linear Algebra (part ) : Matrices and Systems of Linear Equations (by Evan Dummit, 206, v 202) Contents 2 Matrices and Systems of Linear Equations 2 Systems of Linear Equations 2 Elimination, Matrix Formulation

More information

SVD-based optimal ltering with applications to noise reduction in speech signals Simon Doclo ESAT - SISTA, Katholieke Universiteit Leuven Kardinaal Me

SVD-based optimal ltering with applications to noise reduction in speech signals Simon Doclo ESAT - SISTA, Katholieke Universiteit Leuven Kardinaal Me Departement Elektrotechniek ESAT-SISTA/TR 999- SVD-based Optimal Filtering with Applications to Noise Reduction in Speech Signals Simon Doclo, Marc Moonen April, 999 Internal report This report is available

More information

Departement Elektrotechniek ESAT-SISTA/TR 98- Stochastic System Identication for ATM Network Trac Models: a Time Domain Approach Katrien De Cock and Bart De Moor April 998 Accepted for publication in roceedings

More information

R. Schaback. numerical method is proposed which rst minimizes each f j separately. and then applies a penalty strategy to gradually force the

R. Schaback. numerical method is proposed which rst minimizes each f j separately. and then applies a penalty strategy to gradually force the A Multi{Parameter Method for Nonlinear Least{Squares Approximation R Schaback Abstract P For discrete nonlinear least-squares approximation problems f 2 (x)! min for m smooth functions f : IR n! IR a m

More information

Matrix factorization and minimal state space realization in the max-plus algebra

Matrix factorization and minimal state space realization in the max-plus algebra KULeuven Department of Electrical Engineering (ESAT) SISTA Technical report 96-69 Matrix factorization and minimal state space realization in the max-plus algebra B De Schutter and B De Moor If you want

More information

8.1 Bifurcations of Equilibria

8.1 Bifurcations of Equilibria 1 81 Bifurcations of Equilibria Bifurcation theory studies qualitative changes in solutions as a parameter varies In general one could study the bifurcation theory of ODEs PDEs integro-differential equations

More information

Introduction to Neural Networks: Structure and Training

Introduction to Neural Networks: Structure and Training Introduction to Neural Networks: Structure and Training Professor Q.J. Zhang Department of Electronics Carleton University, Ottawa, Canada www.doe.carleton.ca/~qjz, qjz@doe.carleton.ca A Quick Illustration

More information

Method of Frobenius. General Considerations. L. Nielsen, Ph.D. Dierential Equations, Fall Department of Mathematics, Creighton University

Method of Frobenius. General Considerations. L. Nielsen, Ph.D. Dierential Equations, Fall Department of Mathematics, Creighton University Method of Frobenius General Considerations L. Nielsen, Ph.D. Department of Mathematics, Creighton University Dierential Equations, Fall 2008 Outline 1 The Dierential Equation and Assumptions 2 3 Main Theorem

More information

1 Matrices and Systems of Linear Equations

1 Matrices and Systems of Linear Equations Linear Algebra (part ) : Matrices and Systems of Linear Equations (by Evan Dummit, 207, v 260) Contents Matrices and Systems of Linear Equations Systems of Linear Equations Elimination, Matrix Formulation

More information

Chap. 3. Controlled Systems, Controllability

Chap. 3. Controlled Systems, Controllability Chap. 3. Controlled Systems, Controllability 1. Controllability of Linear Systems 1.1. Kalman s Criterion Consider the linear system ẋ = Ax + Bu where x R n : state vector and u R m : input vector. A :

More information

Departement Elektrotechniek ESAT-SISTA/TR Minimal state space realization of SISO systems in the max algebra 1.

Departement Elektrotechniek ESAT-SISTA/TR Minimal state space realization of SISO systems in the max algebra 1. Katholieke Universiteit Leuven Departement Elektrotechniek ESAT-SISTA/TR 9- inimal state space realization of SISO systems in the max algebra 1 Bart De Schutter and Bart De oor October 199 A short version

More information

Linearly-solvable Markov decision problems

Linearly-solvable Markov decision problems Advances in Neural Information Processing Systems 2 Linearly-solvable Markov decision problems Emanuel Todorov Department of Cognitive Science University of California San Diego todorov@cogsci.ucsd.edu

More information

The following can also be obtained from this WWW address: the papers [8, 9], more examples, comments on the implementation and a short description of

The following can also be obtained from this WWW address: the papers [8, 9], more examples, comments on the implementation and a short description of An algorithm for computing the Weierstrass normal form Mark van Hoeij Department of mathematics University of Nijmegen 6525 ED Nijmegen The Netherlands e-mail: hoeij@sci.kun.nl April 9, 1995 Abstract This

More information

In: Proc. BENELEARN-98, 8th Belgian-Dutch Conference on Machine Learning, pp 9-46, 998 Linear Quadratic Regulation using Reinforcement Learning Stephan ten Hagen? and Ben Krose Department of Mathematics,

More information

Spurious Chaotic Solutions of Dierential. Equations. Sigitas Keras. September Department of Applied Mathematics and Theoretical Physics

Spurious Chaotic Solutions of Dierential. Equations. Sigitas Keras. September Department of Applied Mathematics and Theoretical Physics UNIVERSITY OF CAMBRIDGE Numerical Analysis Reports Spurious Chaotic Solutions of Dierential Equations Sigitas Keras DAMTP 994/NA6 September 994 Department of Applied Mathematics and Theoretical Physics

More information

Math 1270 Honors ODE I Fall, 2008 Class notes # 14. x 0 = F (x; y) y 0 = G (x; y) u 0 = au + bv = cu + dv

Math 1270 Honors ODE I Fall, 2008 Class notes # 14. x 0 = F (x; y) y 0 = G (x; y) u 0 = au + bv = cu + dv Math 1270 Honors ODE I Fall, 2008 Class notes # 1 We have learned how to study nonlinear systems x 0 = F (x; y) y 0 = G (x; y) (1) by linearizing around equilibrium points. If (x 0 ; y 0 ) is an equilibrium

More information

Computational Aspects of Aggregation in Biological Systems

Computational Aspects of Aggregation in Biological Systems Computational Aspects of Aggregation in Biological Systems Vladik Kreinovich and Max Shpak University of Texas at El Paso, El Paso, TX 79968, USA vladik@utep.edu, mshpak@utep.edu Summary. Many biologically

More information

Handout 2: Invariant Sets and Stability

Handout 2: Invariant Sets and Stability Engineering Tripos Part IIB Nonlinear Systems and Control Module 4F2 1 Invariant Sets Handout 2: Invariant Sets and Stability Consider again the autonomous dynamical system ẋ = f(x), x() = x (1) with state

More information

7 Planar systems of linear ODE

7 Planar systems of linear ODE 7 Planar systems of linear ODE Here I restrict my attention to a very special class of autonomous ODE: linear ODE with constant coefficients This is arguably the only class of ODE for which explicit solution

More information

A general theory of discrete ltering. for LES in complex geometry. By Oleg V. Vasilyev AND Thomas S. Lund

A general theory of discrete ltering. for LES in complex geometry. By Oleg V. Vasilyev AND Thomas S. Lund Center for Turbulence Research Annual Research Briefs 997 67 A general theory of discrete ltering for ES in complex geometry By Oleg V. Vasilyev AND Thomas S. und. Motivation and objectives In large eddy

More information

1 The pendulum equation

1 The pendulum equation Math 270 Honors ODE I Fall, 2008 Class notes # 5 A longer than usual homework assignment is at the end. The pendulum equation We now come to a particularly important example, the equation for an oscillating

More information

On max-algebraic models for transportation networks

On max-algebraic models for transportation networks K.U.Leuven Department of Electrical Engineering (ESAT) SISTA Technical report 98-00 On max-algebraic models for transportation networks R. de Vries, B. De Schutter, and B. De Moor If you want to cite this

More information

IN THIS PAPER, we consider a class of continuous-time recurrent

IN THIS PAPER, we consider a class of continuous-time recurrent IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II: EXPRESS BRIEFS, VOL. 51, NO. 4, APRIL 2004 161 Global Output Convergence of a Class of Continuous-Time Recurrent Neural Networks With Time-Varying Thresholds

More information

Linear Algebra, 4th day, Thursday 7/1/04 REU Info:

Linear Algebra, 4th day, Thursday 7/1/04 REU Info: Linear Algebra, 4th day, Thursday 7/1/04 REU 004. Info http//people.cs.uchicago.edu/laci/reu04. Instructor Laszlo Babai Scribe Nick Gurski 1 Linear maps We shall study the notion of maps between vector

More information

. Consider the linear system dx= =! = " a b # x y! : (a) For what values of a and b do solutions oscillate (i.e., do both x(t) and y(t) pass through z

. Consider the linear system dx= =! =  a b # x y! : (a) For what values of a and b do solutions oscillate (i.e., do both x(t) and y(t) pass through z Preliminary Exam { 1999 Morning Part Instructions: No calculators or crib sheets are allowed. Do as many problems as you can. Justify your answers as much as you can but very briey. 1. For positive real

More information

= w 2. w 1. B j. A j. C + j1j2

= w 2. w 1. B j. A j. C + j1j2 Local Minima and Plateaus in Multilayer Neural Networks Kenji Fukumizu and Shun-ichi Amari Brain Science Institute, RIKEN Hirosawa 2-, Wako, Saitama 35-098, Japan E-mail: ffuku, amarig@brain.riken.go.jp

More information

Optimal traffic light control for a single intersection

Optimal traffic light control for a single intersection KULeuven Department of Electrical Engineering (ESAT) SISTA Technical report 97- Optimal traffic light control for a single intersection B De Schutter and B De Moor If you want to cite this report, please

More information

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel LECTURE NOTES on ELEMENTARY NUMERICAL METHODS Eusebius Doedel TABLE OF CONTENTS Vector and Matrix Norms 1 Banach Lemma 20 The Numerical Solution of Linear Systems 25 Gauss Elimination 25 Operation Count

More information

Economics 472. Lecture 10. where we will refer to y t as a m-vector of endogenous variables, x t as a q-vector of exogenous variables,

Economics 472. Lecture 10. where we will refer to y t as a m-vector of endogenous variables, x t as a q-vector of exogenous variables, University of Illinois Fall 998 Department of Economics Roger Koenker Economics 472 Lecture Introduction to Dynamic Simultaneous Equation Models In this lecture we will introduce some simple dynamic simultaneous

More information

INRIA Rocquencourt, Le Chesnay Cedex (France) y Dept. of Mathematics, North Carolina State University, Raleigh NC USA

INRIA Rocquencourt, Le Chesnay Cedex (France) y Dept. of Mathematics, North Carolina State University, Raleigh NC USA Nonlinear Observer Design using Implicit System Descriptions D. von Wissel, R. Nikoukhah, S. L. Campbell y and F. Delebecque INRIA Rocquencourt, 78 Le Chesnay Cedex (France) y Dept. of Mathematics, North

More information

Equations with regular-singular points (Sect. 5.5).

Equations with regular-singular points (Sect. 5.5). Equations with regular-singular points (Sect. 5.5). Equations with regular-singular points. s: Equations with regular-singular points. Method to find solutions. : Method to find solutions. Recall: The

More information

Objective. Single population growth models

Objective. Single population growth models Objective Single population growth models We are given a table with the population census at different time intervals between a date a and a date b, and want to get an expression for the population. This

More information

Identication and Control of Nonlinear Systems Using. Neural Network Models: Design and Stability Analysis. Marios M. Polycarpou and Petros A.

Identication and Control of Nonlinear Systems Using. Neural Network Models: Design and Stability Analysis. Marios M. Polycarpou and Petros A. Identication and Control of Nonlinear Systems Using Neural Network Models: Design and Stability Analysis by Marios M. Polycarpou and Petros A. Ioannou Report 91-09-01 September 1991 Identication and Control

More information

7.5 Partial Fractions and Integration

7.5 Partial Fractions and Integration 650 CHPTER 7. DVNCED INTEGRTION TECHNIQUES 7.5 Partial Fractions and Integration In this section we are interested in techniques for computing integrals of the form P(x) dx, (7.49) Q(x) where P(x) and

More information

Novel Approach to Analysis of Nonlinear Recursions. 1 Department of Physics, Bar-Ilan University, Ramat-Gan, ISRAEL

Novel Approach to Analysis of Nonlinear Recursions. 1 Department of Physics, Bar-Ilan University, Ramat-Gan, ISRAEL Novel Approach to Analysis of Nonlinear Recursions G.Berkolaiko 1 2, S. Rabinovich 1,S.Havlin 1 1 Department of Physics, Bar-Ilan University, 529 Ramat-Gan, ISRAEL 2 Department of Mathematics, Voronezh

More information

4.1 Eigenvalues, Eigenvectors, and The Characteristic Polynomial

4.1 Eigenvalues, Eigenvectors, and The Characteristic Polynomial Linear Algebra (part 4): Eigenvalues, Diagonalization, and the Jordan Form (by Evan Dummit, 27, v ) Contents 4 Eigenvalues, Diagonalization, and the Jordan Canonical Form 4 Eigenvalues, Eigenvectors, and

More information

Linearized methods for ordinary di erential equations

Linearized methods for ordinary di erential equations Applied Mathematics and Computation 104 (1999) 109±19 www.elsevier.nl/locate/amc Linearized methods for ordinary di erential equations J.I. Ramos 1 Departamento de Lenguajes y Ciencias de la Computacion,

More information

Minimum and maximum values *

Minimum and maximum values * OpenStax-CNX module: m17417 1 Minimum and maximum values * Sunil Kumar Singh This work is produced by OpenStax-CNX and licensed under the Creative Commons Attribution License 2.0 In general context, a

More information

Essentials of Intermediate Algebra

Essentials of Intermediate Algebra Essentials of Intermediate Algebra BY Tom K. Kim, Ph.D. Peninsula College, WA Randy Anderson, M.S. Peninsula College, WA 9/24/2012 Contents 1 Review 1 2 Rules of Exponents 2 2.1 Multiplying Two Exponentials

More information

Functional equations for Feynman integrals

Functional equations for Feynman integrals Functional equations for Feynman integrals O.V. Tarasov JINR, Dubna, Russia October 9, 016, Hayama, Japan O.V. Tarasov (JINR) Functional equations for Feynman integrals 1 / 34 Contents 1 Functional equations

More information

1. The Polar Decomposition

1. The Polar Decomposition A PERSONAL INTERVIEW WITH THE SINGULAR VALUE DECOMPOSITION MATAN GAVISH Part. Theory. The Polar Decomposition In what follows, F denotes either R or C. The vector space F n is an inner product space with

More information

2 EBERHARD BECKER ET AL. has a real root. Thus our problem can be reduced to the problem of deciding whether or not a polynomial in one more variable

2 EBERHARD BECKER ET AL. has a real root. Thus our problem can be reduced to the problem of deciding whether or not a polynomial in one more variable Deciding positivity of real polynomials Eberhard Becker, Victoria Powers, and Thorsten Wormann Abstract. We describe an algorithm for deciding whether or not a real polynomial is positive semidenite. The

More information

Error Empirical error. Generalization error. Time (number of iteration)

Error Empirical error. Generalization error. Time (number of iteration) Submitted to Neural Networks. Dynamics of Batch Learning in Multilayer Networks { Overrealizability and Overtraining { Kenji Fukumizu The Institute of Physical and Chemical Research (RIKEN) E-mail: fuku@brain.riken.go.jp

More information

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2. APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product

More information

ECS130 Scientific Computing. Lecture 1: Introduction. Monday, January 7, 10:00 10:50 am

ECS130 Scientific Computing. Lecture 1: Introduction. Monday, January 7, 10:00 10:50 am ECS130 Scientific Computing Lecture 1: Introduction Monday, January 7, 10:00 10:50 am About Course: ECS130 Scientific Computing Professor: Zhaojun Bai Webpage: http://web.cs.ucdavis.edu/~bai/ecs130/ Today

More information

October 7, :8 WSPC/WS-IJWMIP paper. Polynomial functions are renable

October 7, :8 WSPC/WS-IJWMIP paper. Polynomial functions are renable International Journal of Wavelets, Multiresolution and Information Processing c World Scientic Publishing Company Polynomial functions are renable Henning Thielemann Institut für Informatik Martin-Luther-Universität

More information

Control Systems I. Lecture 7: Feedback and the Root Locus method. Readings: Jacopo Tani. Institute for Dynamic Systems and Control D-MAVT ETH Zürich

Control Systems I. Lecture 7: Feedback and the Root Locus method. Readings: Jacopo Tani. Institute for Dynamic Systems and Control D-MAVT ETH Zürich Control Systems I Lecture 7: Feedback and the Root Locus method Readings: Jacopo Tani Institute for Dynamic Systems and Control D-MAVT ETH Zürich November 2, 2018 J. Tani, E. Frazzoli (ETH) Lecture 7:

More information

Neural Controller. Plant. Plant. Critic. evaluation signal. Reinforcement Learning Controller

Neural Controller. Plant. Plant. Critic. evaluation signal. Reinforcement Learning Controller Neural Control Theory: an Overview J.A.K. Suykens, H. Bersini Katholieke Universiteit Leuven Department of Electrical Engineering, ESAT-SISTA Kardinaal Mercierlaan 94, B-3001 Leuven (Heverlee), Belgium

More information

ELEMENTARY LINEAR ALGEBRA WITH APPLICATIONS. 1. Linear Equations and Matrices

ELEMENTARY LINEAR ALGEBRA WITH APPLICATIONS. 1. Linear Equations and Matrices ELEMENTARY LINEAR ALGEBRA WITH APPLICATIONS KOLMAN & HILL NOTES BY OTTO MUTZBAUER 11 Systems of Linear Equations 1 Linear Equations and Matrices Numbers in our context are either real numbers or complex

More information

Institute for Advanced Computer Studies. Department of Computer Science. On the Convergence of. Multipoint Iterations. G. W. Stewart y.

Institute for Advanced Computer Studies. Department of Computer Science. On the Convergence of. Multipoint Iterations. G. W. Stewart y. University of Maryland Institute for Advanced Computer Studies Department of Computer Science College Park TR{93{10 TR{3030 On the Convergence of Multipoint Iterations G. W. Stewart y February, 1993 Reviseed,

More information

Roots of Unity, Cyclotomic Polynomials and Applications

Roots of Unity, Cyclotomic Polynomials and Applications Swiss Mathematical Olympiad smo osm Roots of Unity, Cyclotomic Polynomials and Applications The task to be done here is to give an introduction to the topics in the title. This paper is neither complete

More information

The iterative convex minorant algorithm for nonparametric estimation

The iterative convex minorant algorithm for nonparametric estimation The iterative convex minorant algorithm for nonparametric estimation Report 95-05 Geurt Jongbloed Technische Universiteit Delft Delft University of Technology Faculteit der Technische Wiskunde en Informatica

More information

Problem Description The problem we consider is stabilization of a single-input multiple-state system with simultaneous magnitude and rate saturations,

Problem Description The problem we consider is stabilization of a single-input multiple-state system with simultaneous magnitude and rate saturations, SEMI-GLOBAL RESULTS ON STABILIZATION OF LINEAR SYSTEMS WITH INPUT RATE AND MAGNITUDE SATURATIONS Trygve Lauvdal and Thor I. Fossen y Norwegian University of Science and Technology, N-7 Trondheim, NORWAY.

More information

8 Periodic Linear Di erential Equations - Floquet Theory

8 Periodic Linear Di erential Equations - Floquet Theory 8 Periodic Linear Di erential Equations - Floquet Theory The general theory of time varying linear di erential equations _x(t) = A(t)x(t) is still amazingly incomplete. Only for certain classes of functions

More information

LECTURE NOTES IN CRYPTOGRAPHY

LECTURE NOTES IN CRYPTOGRAPHY 1 LECTURE NOTES IN CRYPTOGRAPHY Thomas Johansson 2005/2006 c Thomas Johansson 2006 2 Chapter 1 Abstract algebra and Number theory Before we start the treatment of cryptography we need to review some basic

More information

A characterization of consistency of model weights given partial information in normal linear models

A characterization of consistency of model weights given partial information in normal linear models Statistics & Probability Letters ( ) A characterization of consistency of model weights given partial information in normal linear models Hubert Wong a;, Bertrand Clare b;1 a Department of Health Care

More information

Circuit depth relative to a random oracle. Peter Bro Miltersen. Aarhus University, Computer Science Department

Circuit depth relative to a random oracle. Peter Bro Miltersen. Aarhus University, Computer Science Department Circuit depth relative to a random oracle Peter Bro Miltersen Aarhus University, Computer Science Department Ny Munkegade, DK 8000 Aarhus C, Denmark. pbmiltersen@daimi.aau.dk Keywords: Computational complexity,

More information

The characteristic equation and minimal state space realization of SISO systems in the max algebra

The characteristic equation and minimal state space realization of SISO systems in the max algebra KULeuven Department of Electrical Engineering (ESAT) SISTA Technical report 93-57a The characteristic equation and minimal state space realization of SISO systems in the max algebra B De Schutter and B

More information

output dimension input dimension Gaussian evidence Gaussian Gaussian evidence evidence from t +1 inputs and outputs at time t x t+2 x t-1 x t+1

output dimension input dimension Gaussian evidence Gaussian Gaussian evidence evidence from t +1 inputs and outputs at time t x t+2 x t-1 x t+1 To appear in M. S. Kearns, S. A. Solla, D. A. Cohn, (eds.) Advances in Neural Information Processing Systems. Cambridge, MA: MIT Press, 999. Learning Nonlinear Dynamical Systems using an EM Algorithm Zoubin

More information

REMARKS ON THE TIME-OPTIMAL CONTROL OF A CLASS OF HAMILTONIAN SYSTEMS. Eduardo D. Sontag. SYCON - Rutgers Center for Systems and Control

REMARKS ON THE TIME-OPTIMAL CONTROL OF A CLASS OF HAMILTONIAN SYSTEMS. Eduardo D. Sontag. SYCON - Rutgers Center for Systems and Control REMARKS ON THE TIME-OPTIMAL CONTROL OF A CLASS OF HAMILTONIAN SYSTEMS Eduardo D. Sontag SYCON - Rutgers Center for Systems and Control Department of Mathematics, Rutgers University, New Brunswick, NJ 08903

More information

1 Introduction It will be convenient to use the inx operators a b and a b to stand for maximum (least upper bound) and minimum (greatest lower bound)

1 Introduction It will be convenient to use the inx operators a b and a b to stand for maximum (least upper bound) and minimum (greatest lower bound) Cycle times and xed points of min-max functions Jeremy Gunawardena, Department of Computer Science, Stanford University, Stanford, CA 94305, USA. jeremy@cs.stanford.edu October 11, 1993 to appear in the

More information

Coins with arbitrary weights. Abstract. Given a set of m coins out of a collection of coins of k unknown distinct weights, we wish to

Coins with arbitrary weights. Abstract. Given a set of m coins out of a collection of coins of k unknown distinct weights, we wish to Coins with arbitrary weights Noga Alon Dmitry N. Kozlov y Abstract Given a set of m coins out of a collection of coins of k unknown distinct weights, we wish to decide if all the m given coins have the

More information

1. Affine Grassmannian for G a. Gr Ga = lim A n. Intuition. First some intuition. We always have to rst approximation

1. Affine Grassmannian for G a. Gr Ga = lim A n. Intuition. First some intuition. We always have to rst approximation PROBLEM SESSION I: THE AFFINE GRASSMANNIAN TONY FENG In this problem we are proving: 1 Affine Grassmannian for G a Gr Ga = lim A n n with A n A n+1 being the inclusion of a hyperplane in the obvious way

More information

Analog Neural Nets with Gaussian or other Common. Noise Distributions cannot Recognize Arbitrary. Regular Languages.

Analog Neural Nets with Gaussian or other Common. Noise Distributions cannot Recognize Arbitrary. Regular Languages. Analog Neural Nets with Gaussian or other Common Noise Distributions cannot Recognize Arbitrary Regular Languages Wolfgang Maass Inst. for Theoretical Computer Science, Technische Universitat Graz Klosterwiesgasse

More information

University of California. Berkeley, CA fzhangjun johans lygeros Abstract

University of California. Berkeley, CA fzhangjun johans lygeros Abstract Dynamical Systems Revisited: Hybrid Systems with Zeno Executions Jun Zhang, Karl Henrik Johansson y, John Lygeros, and Shankar Sastry Department of Electrical Engineering and Computer Sciences University

More information

1 Vectors. Notes for Bindel, Spring 2017 Numerical Analysis (CS 4220)

1 Vectors. Notes for Bindel, Spring 2017 Numerical Analysis (CS 4220) Notes for 2017-01-30 Most of mathematics is best learned by doing. Linear algebra is no exception. You have had a previous class in which you learned the basics of linear algebra, and you will have plenty

More information

Reduction of two-loop Feynman integrals. Rob Verheyen

Reduction of two-loop Feynman integrals. Rob Verheyen Reduction of two-loop Feynman integrals Rob Verheyen July 3, 2012 Contents 1 The Fundamentals at One Loop 2 1.1 Introduction.............................. 2 1.2 Reducing the One-loop Case.....................

More information

NL q theory: checking and imposing stability of recurrent neural networks for nonlinear modelling J.A.K. Suykens, J. Vandewalle, B. De Moor Katholieke

NL q theory: checking and imposing stability of recurrent neural networks for nonlinear modelling J.A.K. Suykens, J. Vandewalle, B. De Moor Katholieke NL q theory: checing and imposing stability of recurrent neural networs for nonlinear modelling J.A.K. Suyens, J. Vandewalle, B. De Moor Katholiee Universiteit Leuven, Dept. of Electr. Eng., ESAT-SISTA

More information

Pade approximants and noise: rational functions

Pade approximants and noise: rational functions Journal of Computational and Applied Mathematics 105 (1999) 285 297 Pade approximants and noise: rational functions Jacek Gilewicz a; a; b;1, Maciej Pindor a Centre de Physique Theorique, Unite Propre

More information

growth rates of perturbed time-varying linear systems, [14]. For this setup it is also necessary to study discrete-time systems with a transition map

growth rates of perturbed time-varying linear systems, [14]. For this setup it is also necessary to study discrete-time systems with a transition map Remarks on universal nonsingular controls for discrete-time systems Eduardo D. Sontag a and Fabian R. Wirth b;1 a Department of Mathematics, Rutgers University, New Brunswick, NJ 08903, b sontag@hilbert.rutgers.edu

More information

ON IDENTIFICATION OF DYNAMICAL SYSTEM PARAMETERS FROM EXPERIMENTAL DATA

ON IDENTIFICATION OF DYNAMICAL SYSTEM PARAMETERS FROM EXPERIMENTAL DATA ON IDENTIFICATION OF DYNAMICAL SYSTEM PARAMETERS FROM EXPERIMENTAL DATA M Shatalov*, I Fedotov** * CSIR Manufacturing and Materials PO Box 305, Pretoria 0001, SCIR, South Africa and Department of Mathematics

More information

Contents. 4 Arithmetic and Unique Factorization in Integral Domains. 4.1 Euclidean Domains and Principal Ideal Domains

Contents. 4 Arithmetic and Unique Factorization in Integral Domains. 4.1 Euclidean Domains and Principal Ideal Domains Ring Theory (part 4): Arithmetic and Unique Factorization in Integral Domains (by Evan Dummit, 018, v. 1.00) Contents 4 Arithmetic and Unique Factorization in Integral Domains 1 4.1 Euclidean Domains and

More information

A matrix over a field F is a rectangular array of elements from F. The symbol

A matrix over a field F is a rectangular array of elements from F. The symbol Chapter MATRICES Matrix arithmetic A matrix over a field F is a rectangular array of elements from F The symbol M m n (F ) denotes the collection of all m n matrices over F Matrices will usually be denoted

More information

Block-row Hankel Weighted Low Rank Approximation 1

Block-row Hankel Weighted Low Rank Approximation 1 Katholieke Universiteit Leuven Departement Elektrotechniek ESAT-SISTA/TR 03-105 Block-row Hankel Weighted Low Rank Approximation 1 Mieke Schuermans, Philippe Lemmerling and Sabine Van Huffel 2 July 2003

More information

Partial Fraction Decomposition Honors Precalculus Mr. Velazquez Rm. 254

Partial Fraction Decomposition Honors Precalculus Mr. Velazquez Rm. 254 Partial Fraction Decomposition Honors Precalculus Mr. Velazquez Rm. 254 Adding and Subtracting Rational Expressions Recall that we can use multiplication and common denominators to write a sum or difference

More information

B553 Lecture 5: Matrix Algebra Review

B553 Lecture 5: Matrix Algebra Review B553 Lecture 5: Matrix Algebra Review Kris Hauser January 19, 2012 We have seen in prior lectures how vectors represent points in R n and gradients of functions. Matrices represent linear transformations

More information

Experimental evidence showing that stochastic subspace identication methods may fail 1

Experimental evidence showing that stochastic subspace identication methods may fail 1 Systems & Control Letters 34 (1998) 303 312 Experimental evidence showing that stochastic subspace identication methods may fail 1 Anders Dahlen, Anders Lindquist, Jorge Mari Division of Optimization and

More information

State space transformations and state space realization in the max algebra

State space transformations and state space realization in the max algebra K.U.Leuven Department of Electrical Engineering (ESAT) SISTA Technical report 95-10 State space transformations and state space realization in the max algebra B. De Schutter and B. De Moor If you want

More information

ELEMENTARY LINEAR ALGEBRA

ELEMENTARY LINEAR ALGEBRA ELEMENTARY LINEAR ALGEBRA K R MATTHEWS DEPARTMENT OF MATHEMATICS UNIVERSITY OF QUEENSLAND First Printing, 99 Chapter LINEAR EQUATIONS Introduction to linear equations A linear equation in n unknowns x,

More information

March 16, Abstract. We study the problem of portfolio optimization under the \drawdown constraint" that the

March 16, Abstract. We study the problem of portfolio optimization under the \drawdown constraint that the ON PORTFOLIO OPTIMIZATION UNDER \DRAWDOWN" CONSTRAINTS JAKSA CVITANIC IOANNIS KARATZAS y March 6, 994 Abstract We study the problem of portfolio optimization under the \drawdown constraint" that the wealth

More information

Chapter 13 Internal (Lyapunov) Stability 13.1 Introduction We have already seen some examples of both stable and unstable systems. The objective of th

Chapter 13 Internal (Lyapunov) Stability 13.1 Introduction We have already seen some examples of both stable and unstable systems. The objective of th Lectures on Dynamic Systems and Control Mohammed Dahleh Munther A. Dahleh George Verghese Department of Electrical Engineering and Computer Science Massachuasetts Institute of Technology 1 1 c Chapter

More information

Linear Algebra: Characteristic Value Problem

Linear Algebra: Characteristic Value Problem Linear Algebra: Characteristic Value Problem . The Characteristic Value Problem Let < be the set of real numbers and { be the set of complex numbers. Given an n n real matrix A; does there exist a number

More information

is equal to = 3 2 x, if x < 0 f (0) = lim h = 0. Therefore f exists and is continuous at 0.

is equal to = 3 2 x, if x < 0 f (0) = lim h = 0. Therefore f exists and is continuous at 0. Madhava Mathematics Competition January 6, 2013 Solutions and scheme of marking Part I N.B. Each question in Part I carries 2 marks. p(k + 1) 1. If p(x) is a non-constant polynomial, then lim k p(k) (a)

More information

Groups. 3.1 Definition of a Group. Introduction. Definition 3.1 Group

Groups. 3.1 Definition of a Group. Introduction. Definition 3.1 Group C H A P T E R t h r e E Groups Introduction Some of the standard topics in elementary group theory are treated in this chapter: subgroups, cyclic groups, isomorphisms, and homomorphisms. In the development

More information

VICTORIA POWERS AND THORSTEN W ORMANN Suppose there exists a real, symmetric, psd matrix B such that f = x B x T and rank t. Since B is real symmetric

VICTORIA POWERS AND THORSTEN W ORMANN Suppose there exists a real, symmetric, psd matrix B such that f = x B x T and rank t. Since B is real symmetric AN ALGORITHM FOR SUMS OF SQUARES OF REAL POLYNOMIALS Victoria Powers and Thorsten Wormann Introduction We present an algorithm to determine if a real polynomial is a sum of squares (of polynomials), and

More information

[4] T. I. Seidman, \\First Come First Serve" is Unstable!," tech. rep., University of Maryland Baltimore County, 1993.

[4] T. I. Seidman, \\First Come First Serve is Unstable!, tech. rep., University of Maryland Baltimore County, 1993. [2] C. J. Chase and P. J. Ramadge, \On real-time scheduling policies for exible manufacturing systems," IEEE Trans. Automat. Control, vol. AC-37, pp. 491{496, April 1992. [3] S. H. Lu and P. R. Kumar,

More information

J. Symbolic Computation (1995) 11, 1{000 An algorithm for computing an integral basis in an algebraic function eld Mark van Hoeij Department of mathem

J. Symbolic Computation (1995) 11, 1{000 An algorithm for computing an integral basis in an algebraic function eld Mark van Hoeij Department of mathem J. Symbolic Computation (1995) 11, 1{000 An algorithm for computing an integral basis in an algebraic function eld Mark van Hoeij Department of mathematics University of Nijmegen 6525 ED Nijmegen The Netherlands

More information