Reachable Set Analysis for Dynamic Neural Networks with Polytopic Uncertainties

Similar documents
Exponential Tracking Control of Nonlinear Systems with Actuator Nonlinearity

Lie symmetry and Mei conservation law of continuum system

Adaptive Gain-Scheduled H Control of Linear Parameter-Varying Systems with Time-Delayed Elements

Lecture 2 Lagrangian formulation of classical mechanics Mechanics

Research Article Convex Polyhedron Method to Stability of Continuous Systems with Two Additive Time-Varying Delay Components

Generalized-Type Synchronization of Hyperchaotic Oscillators Using a Vector Signal

6 General properties of an autonomous system of two first order ODE

Switching Time Optimization in Discretized Hybrid Dynamical Systems

Introduction to the Vlasov-Poisson system

Results on stability of linear systems with time varying delay

Connections Between Duality in Control Theory and

NOTES ON EULER-BOOLE SUMMATION (1) f (l 1) (n) f (l 1) (m) + ( 1)k 1 k! B k (y) f (k) (y) dy,

Criteria for Global Stability of Coupled Systems with Application to Robust Output Feedback Design for Active Surge Control

Calculus of Variations

GLOBAL SOLUTIONS FOR 2D COUPLED BURGERS-COMPLEX-GINZBURG-LANDAU EQUATIONS

Chaos, Solitons and Fractals

Euler equations for multiple integrals

A Note on Exact Solutions to Linear Differential Equations by the Matrix Exponential

Math 342 Partial Differential Equations «Viktor Grigoryan

Slide10 Haykin Chapter 14: Neurodynamics (3rd Ed. Chapter 13)

A global Implicit Function Theorem without initial point and its applications to control of non-affine systems of high dimensions

Multi-agent Systems Reaching Optimal Consensus with Time-varying Communication Graphs

Stability region estimation for systems with unmodeled dynamics

Time-of-Arrival Estimation in Non-Line-Of-Sight Environments

An LMI Approach to Guaranteed Cost Control for Uncertain Delay Systems

Necessary and sufficient conditions for analysis and synthesis of markov jump linear systems with incomplete transition descriptions

INVERSE PROBLEM OF A HYPERBOLIC EQUATION WITH AN INTEGRAL OVERDETERMINATION CONDITION

SYSTEMS OF DIFFERENTIAL EQUATIONS, EULER S FORMULA. where L is some constant, usually called the Lipschitz constant. An example is

ELEC3114 Control Systems 1

. ISSN (print), (online) International Journal of Nonlinear Science Vol.6(2008) No.3,pp

Stabilization of Interconnected Systems with Decentralized State and/or Output Feedback

Nonlinear Adaptive Ship Course Tracking Control Based on Backstepping and Nussbaum Gain

II. First variation of functionals

All s Well That Ends Well: Supplementary Proofs

A DELAY-DEPENDENT APPROACH TO DESIGN STATE ESTIMATOR FOR DISCRETE STOCHASTIC RECURRENT NEURAL NETWORK WITH INTERVAL TIME-VARYING DELAYS

LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION

A nonlinear inverse problem of the Korteweg-de Vries equation

Minimum-time constrained velocity planning

θ x = f ( x,t) could be written as

Least-Squares Regression on Sparse Spaces

ALGEBRAIC AND ANALYTIC PROPERTIES OF ARITHMETIC FUNCTIONS

Approximate reduction of dynamic systems

TRAJECTORY TRACKING FOR FULLY ACTUATED MECHANICAL SYSTEMS

Remote Preparation of Multipartite Equatorial Entangled States in High Dimensions with Three Parties

Dissipative numerical methods for the Hunter-Saxton equation

Lecture 10: October 30, 2017

Robust Forward Algorithms via PAC-Bayes and Laplace Distributions. ω Q. Pr (y(ω x) < 0) = Pr A k

Tractability results for weighted Banach spaces of smooth functions

Characterizing Real-Valued Multivariate Complex Polynomials and Their Symmetric Tensor Representations

BEYOND THE CONSTRUCTION OF OPTIMAL SWITCHING SURFACES FOR AUTONOMOUS HYBRID SYSTEMS. Mauro Boccadoro Magnus Egerstedt Paolo Valigi Yorai Wardi

19 Eigenvalues, Eigenvectors, Ordinary Differential Equations, and Control

TOEPLITZ AND POSITIVE SEMIDEFINITE COMPLETION PROBLEM FOR CYCLE GRAPH

Delay-Dependent Stability Criteria for Linear Systems with Multiple Time Delays

Agmon Kolmogorov Inequalities on l 2 (Z d )

Optimized Schwarz Methods with the Yin-Yang Grid for Shallow Water Equations

Network Synchronization With Nonlinear Dynamics and Switching Interactions

APPROXIMATE SOLUTION FOR TRANSIENT HEAT TRANSFER IN STATIC TURBULENT HE II. B. Baudouy. CEA/Saclay, DSM/DAPNIA/STCM Gif-sur-Yvette Cedex, France

On the number of isolated eigenvalues of a pair of particles in a quantum wire

Dissipativity for dual linear differential inclusions through conjugate storage functions

Lectures - Week 10 Introduction to Ordinary Differential Equations (ODES) First Order Linear ODEs

Delay-Dependent Exponential Stability of Linear Systems with Fast Time-Varying Delay

Improved delay-dependent globally asymptotic stability of delayed uncertain recurrent neural networks with Markovian jumping parameters

J. Electrical Systems 11-4 (2015): Regular paper

Optimal Estimation for Continuous-Time Systems With Delayed Measurements

Qubit channels that achieve capacity with two states

Separation Principle for a Class of Nonlinear Feedback Systems Augmented with Observers

SINGULAR PERTURBATION AND STATIONARY SOLUTIONS OF PARABOLIC EQUATIONS IN GAUSS-SOBOLEV SPACES

Lower bounds on Locality Sensitive Hashing

A new four-dimensional chaotic system

Indirect Adaptive Fuzzy and Impulsive Control of Nonlinear Systems

Existence of equilibria in articulated bearings in presence of cavity

arxiv: v1 [math.mg] 10 Apr 2018

Damage identification based on incomplete modal data and constrained nonlinear multivariable function

WUCHEN LI AND STANLEY OSHER

Hyperbolic Moment Equations Using Quadrature-Based Projection Methods

JUST THE MATHS UNIT NUMBER DIFFERENTIATION 2 (Rates of change) A.J.Hobson

Polynomial Inclusion Functions

Continuous observer design for nonlinear systems with sampled and delayed output measurements

Stable and compact finite difference schemes

arxiv: v2 [math.dg] 16 Dec 2014

EXACT TRAVELING WAVE SOLUTIONS FOR A NEW NON-LINEAR HEAT TRANSFER EQUATION

The canonical controllers and regular interconnection

'HVLJQ &RQVLGHUDWLRQ LQ 0DWHULDO 6HOHFWLRQ 'HVLJQ 6HQVLWLYLW\,1752'8&7,21

arxiv:physics/ v2 [physics.ed-ph] 23 Sep 2003

Diagonalization of Matrices Dr. E. Jacobs

Linear First-Order Equations

Interconnected Systems of Fliess Operators

Lecture 1b. Differential operators and orthogonal coordinates. Partial derivatives. Divergence and divergence theorem. Gradient. A y. + A y y dy. 1b.

ECE 422 Power System Operations & Planning 7 Transient Stability

Lecture Introduction. 2 Examples of Measure Concentration. 3 The Johnson-Lindenstrauss Lemma. CS-621 Theory Gems November 28, 2012

Approximate Reduction of Dynamical Systems

Delay-dependent Stability Analysis for Markovian Jump Systems with Interval Time-varying-delays

Equations of lines in

Table of Common Derivatives By David Abraham

Hyperbolic Systems of Equations Posed on Erroneous Curved Domains

STABILITY ANALYSIS FOR SYSTEMS WITH LARGE DELAY PERIOD: A SWITCHING METHOD. Received March 2011; revised July 2011

On Characterizing the Delay-Performance of Wireless Scheduling Algorithms

PDE Notes, Lecture #11

2 GUANGYU LI AND FABIO A. MILNER The coefficient a will be assume to be positive, boune, boune away from zero, an inepenent of t; c will be assume con

Research Article Delay-Dependent Exponential Stability for Discrete-Time BAM Neural Networks with Time-Varying Delays

Transcription:

Commun. Theor. Phys. 57 (2012) 904 908 Vol. 57, No. 5, May 15, 2012 Reachable Set Analysis for Dynamic Neural Networks with Polytopic Uncertainties ZUO Zhi-Qiang ( ãö), CHEN Yin-Ping (í ), an WANG Yi-Jing ( è ) Tianjin Key Laboratory of Process Measurement an Control, School of Electrical Engineering an Automation, Tianjin University, Tianjin 300072, China (Receive November 7, 2011; revise manuscript receive December 22, 2011) Abstract n this paper, the reachable set estimation problem is stuie for a class of ynamic neural networks subject to polytopic uncertainties. The problem aresse here is to fin a set as small as possible to boun the states starting from the origin by inputs with peak values. The maximal Lyapunov functional is propose to erive a sufficient conition for the existence of a non-ellipsoial boun to estimate the states of neural networks. t is theoretically shown that this metho is superior to the traitional one base on the common Lyapunov function. Finally, two examples illustrate the avantages of our propose result. PACS numbers: 84.35.+i Key wors: neural networks, reachable set, polytopic uncertainties, maximal Lyapunov functional 1 ntrouction Dynamic neural networks analysis has attracte the increasing attention of researchers in the past ecaes ue to the successful applications in solving pattern recognition, associative memory, optimization problem, ientification an control, static image processing. Different kins of neural networks moels have been propose an stuie. To mention a few, we refer to [1] [3] references therein. Up to now, most works on ynamic neural networks have focuse on the stability analysis problem an a lot of results have been reporte for ifferent kins of neural networks, see, for example Refs. [4] [10]. Since neural networks can be viewe as a special class of nonlinear ynamic systems, many concepts originally appeare in control theory have been introuce to analyze the neural networks, such as, controllability, [11 12] state estimation, [13 15] input-tostate stability (SS), [16 17] passivity, [18 19] an so on. On the other han, uncertainties are inevitably encountere in practice. Generally speaking, there are mainly two kins of uncertainties, namely, the normboune uncertainty an the polytopic ones. t is well known the existence of uncertainty may lea to instability of neural networks. Therefore, much attention is evote to robust stability analysis of neural networks, for instance Refs. [20] [22]. As we known, the common Lyapunov function can lea to easily tractable matrix inequalities or linear matrix inequalities conitions. However, it probably yiels conservative evaluation of stability an performance. For example, systems with polytopic uncertainties may be stable without having a common Lyapunov function. As an important approach for state estimation an parameter estimation in control theory, a lot of works have been evote to the reachable set bouning, for example Refs. [23] [25]. All the above results on reachable set estimation are expresse in the form of ellipsoi. f there are polytopic uncertainties for systems, the common Lyapunov function has to be use for all the vertices of the polytope. To the best of our knowlege, there is no ocument available concerning reachable set estimation for uncertain neural networks, which motivates our stuy. n this paper, we evelop the reachable set bouning for ynamic neural networks subject to both polytopic uncertainties an boune peak inputs. We propose a maximal Lyapunov functional, which is obtaine by taking pointwise maximum over a family of Lyapunov functionals, to erive a tighter non-ellipsoial boun of the reachable sets. This metho shows great avantages over the traitional one base on common Lyapunov function. Finally, two numerical examples are provie to illustrate the usefulness of our theoretical results. Notations The notations use throughout the paper are stanar. R n enotes the n-imensional Eucliean space. The superscript T enotes matrix transposition, an an 0 enote the ientity an zero matrix with appropriate imension. The notation P > 0 (P 0) means that P is symmetric an positive efinite (positive semi-efinite). Co{ } enotes a convex hull. iag{l 1, l 2,...,l n } stans for a iagonal matrix with iagonal elements being the scalars l 1, l 2,...,l n. The symmetric terms in a symmetric matrix are enote by. Matri- Supporte by the National Natural Science Founation of China uner Grant Nos. 60774039, 60974024, 61074089, 61174129, Program for New Century Excellent Talents in University uner Grant No. NCET-11-0379, an the nepenent nnovation Founation of Tianjin University Corresponing author, E-mail: zqzuo@tju.eu.cn c 2011 Chinese Physical Society an OP Publishing Lt http://www.iop.org/ej/journal/ctp http://ctp.itp.ac.cn

No. 5 Communications in Theoretical Physics 905 ces, if not explicitly state, are assume to have compatible imensions. For a matrix P > 0, enote E(P, 1) : = {x R n : x T Px 1}. For a general Lyapunov function V, enote its 1-level set as L V : = {x R n : V (x) 1}. 2 Problem Formulation an Preliminaries Consier the following ynamic neural networks ẋ(t) = Cx(t) + Af(x(t)) + J(t), (1) where x(t) = [x 1 (t) x 2 (t) x n (t)] T R n is the state vector of the neural networks. C = iag{c 1, c 2,...,c n } is a constant iagonal matrix with c i > 0; f(x(t)) = [f 1 (x 1 (t)), f 2 (x 2 (t)),..., f n (x n (t))] T R n is the activation of neurons. A = (a ij ) n n is the connection weight matrix; J(t) represents external inputs with the form J T (t)j(t). (2) Furthermore, the activation function f i ( ) (i = 1, 2,...,n) is assume to satisfy the following conition 0 f i(s 1 ) f i (s 2 ) s 1 s 2 l i, (3) for any s 1, s 2 R, s 1 s 2, where l i > 0 (i = 1, 2,...,n) are positive scalars. Matrices C, A are uncertain but belong to a given polytope Φ = [C A], (4) Φ = θ i Φ i, θ i 0, θ i = 1, (5) where the N vertices of the polytope are escribe by Φ i = [C i A i ], (i = 1, 2,..., N). By (3), it is easy to see f j (x j (t))[f j (x j (t)) l j x j (t)] 0 (j = 1, 2,..., n). (6) From the above inequality (6), for any non-negative iagonal matrix H = iag{h 1, h 2,..., h n }, the following conition hols n 2 h j f j (x j (t))[l j x j (t) f j (x j (t))] 0, j=1 which can be rewritten as 2[x T (t)lhf(x(t)) f T (x(t))hf(x(t))] 0. (7) n this paper, we will construct a function which is the maximum of a family of Lyapunov functions. Each Lyapunov function correspons to a vertex of the polytope (i.e., Φ i ). Given a family of matrices P j > 0, (j = 1, 2,..., N), the pointwise maximum quaratic function is efine as V max = max{x T P j x}. The 1-level set of V max is enote as L Vmax : = {x R n : V max 1}. Since V max is strict convex an not ifferentiable everywhere, we introuce the efinitions of subgraient an subifferential to characterize the behavior at nonifferentiable points. Let g : R n R be a real-value convex function in R n, a vector v is calle a subgraient at a point x 0 if g(x) g(x 0 ) v T (x x 0 ), x R n. The set of all subgraients at x 0 is calle the subifferential at x 0 an is enote as f(x 0 ). The function f(x) is ifferentiable at x 0 if an only if the subifferential is mae up of only one point, which is the erivative at x 0. Consier x 0 R n. Suppose that there exists an integer s (1 < s N) such that V max (x 0 ) = x T 0 P jx 0 for j = 1,...,s an V max (x 0 ) > x T 0 P j x 0 for j > s. Base on the results in [26], we have the following properties: (i) for a vector ζ R n, the irectional erivative of V max at x 0 along ζ is ζ V max (x 0 ) = lim t 0 + V max (x 0 + tζ) V max (x 0 ) t = max ξ V max(x 0) {ξt ζ} ; (ii) V max (x 0 ) = Co{2P j x 0, j = 1,...,s}. The following lemma is an extension of the reachable set estimation for ynamic systems to that for neural networks (1). Lemma [27] Let V be a Lyapunov function for neural networks (1). f V (x) + αv (x) α J T (t)j(t) 0, t then we have V (x) 1. 3 Main Results Theorem 1 f there exist matrices P j > 0 with compatible imensions, a non-negative iagonal matrix H, an scalars α > 0, β ij 0 for all i, j = 1, 2,..., N such that Γ = Γ 11 P j A i + LH P j 2H 0 α Γ 11 = P j C i Ci T P j + αp j + β ij (P j P i ), < 0, (8) then for all the amissible polytopic uncertainties, the intersection of a family of ellipsois, i.e., E(P j, 1) (9) j=1, 2,..., N is a set which bouns the reachable set of neural networks (1), where E(P j, 1) = {x R n : x T P j x 1}. Proof Given a family of matrices P j > 0, the pointwise maximum Lyapunov functional is chosen as where V max (x) = max{v j (x)}, j = 1, 2,...,N, (10) For x 0, we efine V j (x) = x T (t)p j x(t). N max (x) : = {j {1, 2,..., N} : V j (x) = V max (x)}.

906 Communications in Theoretical Physics Vol. 57 t follows that V j (x) < V max if j / N max (x). Without loss of generality, we assume that N max (x) = {1, 2,..., n 1 } for some integer n 1 N. Here n 1 is the number of ellipsois E(P j, 1) s intersecte at x. Therefore, V j (x) = V max (x) for j n 1 an V j (x) < V max (x) for j > n 1. Now we know x T (P j P i )x 0, j n 1, i {1, 2,..., N}. (11) To better illustrate the thoughts of our proof, we efine a set Υ j = {x R n : V j (x) V i (x), i j}, an consier two ifferent cases for x. (i) f x Υ j \ i =j Υ i (i.e., x is ifferentiable), then where V max = x T (t)p j x(t), t V max + αv max α J T (t)j(t) ζ T (t) ζ T (t) = [ x T (t) f T (x(t)) J T (t) ]. f (8) is satisfie, then using convexity property, we have: t V max + αv max α J T (t)j(t) β ij x T (P i P j )x 0. The last inequality hols because x T (P i P j )x 0 if x Υ j \ i =j Υ i. This implies that V max 1 by Lemma 1. (ii) f x n1 j=1 Υ j\ N j=n Υ 1+1 j, it is clear that x belongs to the intersection of Υ j s (j = 1, 2,...,n 1). This means that x is nonifferentiable. As a result, V max = x T P j x, j N max (x) an V max = Co{2P j x : j N max (x)}. Therefore, provie that (8) is met, we obtain ẋv max (x) + αv max (x) α J T (t)j(t) max β ij x T (P i P j )x. j N max(x) Note that (11) is satisfie in such a case. Therefore we can raw the conclusion that V max 1. t is easy to see that L Vmax is the intersection of the ellipsois E(P j, 1) s. This point can be verifie as follows. On the one han, if x belongs to the intersection of E(P j, 1) s, then x belongs to any ellipsoi E(P j, 1), j = 1, 2,..., N. Therefore, x belongs to L Vmax. On the other han, we suppose that x L Vmax. Since V max = max{x T P j x}, it is obvious that x T P j x 1 for all j = 1, 2,..., N. This implies that x belongs to the intersection of E(P j, 1) s. Remark 1 We have constructe a function which is the maximum of a family of Lyapunov functions in Theorem 1. t V max = 2 x T (t)p j ẋ(t) Then we have = 2 x(t) T P j [ Cx(t) + Af(x(t)) + J(t)] 2x T (t)p j [ Cx(t) + Af(x(t)) + J(t)] + 2 [ x T (t)lhf(x(t)) f T (x(t))hf(x(t))]. (12) P j C C T P j + αp j P j A + LH P j 2H 0 α ζ(t), (13) t is well known that the common Lyapunov function will lea to easily tractable matrix inequality conitions an simplify computational issues even though it may bring conservatism to some extent. Remark 2 The neural networks with polytopic uncertainties of the form (4) an (5) is consiere in this paper. Both the maximum Lyapunov functional propose in (11) an the common Lyapunov function can be use to boun all the states starting from the origin. As for the robust boune uncertainties (i.e. norm-boune uncertainties), the neural networks can be presente by ẋ(t) = [C + C(t)]x(t) + [A + A(t)]f(x(t)) + J(t) with the parameter uncertainties C(t), A(t) which are normboune an satisfy [ C(t) A(t)] = DF(t)[E 1 E 2 ], where D, E 1, an E 2 are known constant real matrices an F(t) is an unknown matrix function with Lebesgue measurable elements an such that F T (t)f(t). For this case, we can only apply the common Lyapunov function to erive the reachable set of the networks. By setting P j = P, β ij = 0, (i, j = 1, 2,...,N), we can easily obtain the ellipsoial boun of a reachable set for neural network (1) by using the common Lyapunov function. Theorem 1 reuces to a simple form of set reachable criterion. Corollary 1 f there exists a matrix P > 0 with compatible imension, a non-negative iagonal matrix H, an a scalar α > 0 such that PC i Ci TP + αp PA i + LH P 2H 0 α < 0, (14) then for all the amissible polytopic uncertainties, the ellipsoi E(P, 1) = {x R n : x T Px 1}

No. 5 Communications in Theoretical Physics 907 is a set which bouns the reachable set of neural networks (1). Remark 3 As can be seen from Corollary 1, the conition obtaine by common Lyapunov function is a special case for that by maximal Lyapunov one. Therefore, we can expect that a tighter bouning set can be obtaine by using the maximal functional approach. However, the better result is achieve at the cost of a bigger computational complexity. Therefore, there is a traeoff between better performance an less computational buren for solving such a problem. n orer to fin the smallest boun for the reachable set of neural network (1), one may propose a simple optimization problem. That is, maximize δ subject to δ P j which can be transforme to the following optimization problem: minimize δ ( δ = 1/δ), [ ] δ (a) 0, s.t. P j (15) (b) (8) or (14). 4 llustrative Examples n this section, two numerical examples are provie to show the effectiveness of the metho evelope in this paper. Example 1 Consier the neural networks (1) with parameters [ ] 3.9 0 C =, 0 2 + ρ [ ] 0 1.72 A =, L = iag{3, 4}, 1.62 1 + ρ where ρ is a scalar parameter satisfying ρ 0.3 an = 2. Note that the uncertainty polytope in this case possesses N = 2 vertices with [ ] [ ] 3.9 0 3.9 0 C 1 =, C 2 =, 0 1.7 [ 0 1.72 A 1 = 1.62 0.7 ], A 2 = 0 2.3 [ ] 0 1.72. 1.62 1.3 By solving the optimization problems (15), the resulting δ s for ifferent values of α are liste in Table 1. t is verifie that the maximal Lyapunov functional metho yiels much tighter bouns than the common Lyapunov approach. We cannot fin any feasible solution when α is larger than 0.357 by using the common Lyapunov metho. The reachable set correspons to the system is epicte in Fig. 1. As we can see from the figure, the shae area compute by Theorem 1 is much tighter than the ellipsoi obtaine by the common Lyapunov metho. Table 1 The resulting δ for ifferent α of Example 1. α 0.20 0.25 0.27 0.30 0.35 0.357 Corollary 1 67.8008 79.5137 90.7370 124.5976 878.0109 Theorem 1 63.5997 72.5133 80.9227 105.1369 365.0691 569.4831 function. However, we can get the result of δ = 361.7478 by Theorem 1. Therefore, for this example, the maximal Lyapunov functional leas to a wier application range comparing with the common Lyapunov one. Fig. 1 The bouning reachable sets compute by ifferent methos for α = 0.3. Example 2 Consier the polytopic uncertain neural networks (1) with parameters: [ ] 0.0005 1.6560 A =, 1.5529 1 + ρ [ ] 3.9 0 C =, L = iag{3, 4}, 0 2 + ρ where ρ 0.3, = 2. t is verifie that there is no feasible solution using the metho of common Lyapunov 5 Conclusion The reachable set bouning for neural networks in the presence of polytopic uncertainties is consiere in this paper. By utilizing the maximal Lyapunov functional approach, we have erive a sufficient conition which results in a non-ellipsoial set to boun all the states of the neural networks starting from the origin by inputs with peak values. Since there exist some nonifferentiable points, the sub-graient an sub-ifferential concepts have been introuce to characterize the behavior at non-ifferentiable points. Our results have emonstrate much flexibility an less conservatism compare with the common Lyapunov function. The simulation illustrates the valiness an merit of the propose metho. The future work in this eneavor will focus on the following issues. First, reachable set estimation for uncertain neural networks with time elay. Time elay is not

908 Communications in Theoretical Physics Vol. 57 involve in this paper, however, it inevitably occurs in the practical application. Secon, some more complcate neural network moels can be consiere, for instance, networks of hogkin-huxley neurons, [28] scale-free Hopfiel neural network. [29] References [1] J. Cao an J. Wang, EEE Trans. Circuits Syst. 52 (2005) 417. [2] L.O. Chua an L. Yang, EEE Trans. Circuits an Syst. 35 (1988) 1257. [3] Z. Wang, Y. Liu, M. Li, an X. Liu, EEE Trans. Neural Netw. 17 (2006) 814. [4] J. Cao an J. Wang, EEE Trans. Circuits Syst. 52 (2005) 920. [5] T. Ensari an S. Arik, EEE Trans. Autom. Control 50 (2005) 1781. [6] Y. He, M. Wu, an J. She, EEE Trans. Circuits Syst. 53 (2006) 553. [7] H. Shao, EEE Trans. Neural Netw. 19 (2008) 1647. [8] S. Xu, J. Lam, D.W.C. Ho, an Y. Zou, EEE Trans. Syst. Man, Cybern. B Cybern. 35 (2005) 1317. [9] Z. Zeng, J. Wang, an X. Liao, EEE Trans. Circuits Syst. 50 (2003) 1353. [10] Z. Zuo, C. Yang, an Y. Wang, EEE Trans. Neural Netw. 21 (2010) 339. [11] E. Sontag an H. Sussmann, Syst. Contr. Lett. 30 (1997) 177. [12] E. Sontag an Y. Qian, Syst. Contr. Lett. 36 (1999) 121. [13] Y. He, Q.G. Wang, M. Wu, an C. Lin, EEE Trans. Neural Netw. 17 (2006) 1077. [14] Z. Wang, D.W.C. Ho, an X. Liu, EEE Trans. Neural Netw. 16 (2005) 279. [15] Z. Wang, Y. Liu, an X. Liu, Neural Networks 22 (2009) 41. [16] E.N. Sanchez an J.P. Perez, EEE Trans. Circuits Syst. 46 (1999) 1395. [17] E.N. Sanchez, J.P. Perez, EEE Trans. Syst., Man, Cybern. A, Syst., Humans 33 (2003) 532. [18] H. Li, H. Gao, an P. Shi, EEE Trans. Neural Netw. 21 (2010) 1842. [19] S. Xu, W. Zheng, an Y. Zou, EEE Trans. Circuits Syst. 56 (2009) 325. [20] C. Li an X. Liao, EEE Trans. Circuits Syst. 53 (2006) 2265. [21] N. Ozcan an S. Arik, EEE Trans. Circuits Syst. 53 (2006) 166. [22] Z. Zuo an Y. Wang, ET Control Theory Appl. 1 (2007) 387. [23] J. Abegor, K. Nagpal, an K. Poolla, nt. J. Robust an Nonlin. 6 (1996) 899. [24] C. Durieu, E. Walter, an B. Polyak, J. Optimiz. Theory Appl. 111 (2001) 273. [25] A. Kurzhanskiy an P. Varaiya, EEE Trans. Autom. Control 52 (2007) 26. [26] S. Boy, Lecture Notes for Convex Optimization, http://www.stanfor.eu/class/ee364b/lectures.html, 2007. [27] S. Boy, L. El Ghaoui, E. Feron, an V. Balakrishnan, Linear Matrix nequalities in System an Control Theory, SAM Stuies in Appl. Mathematics, Philaelphia (1994). [28] J. Ma, L. Yang, Y. Wu, an C. Zhang, Commun. Theor. Phys. 54 (2010) 583. [29] W. Chang, Commun. Theor. Phys. 53 (2010) 195.