Distributed Game Strategy Design with Application to Multi-Agent Formation Control

Similar documents
Multi-Pursuer Single-Evader Differential Games with Limited Observations*

Distributed Receding Horizon Control of Cost Coupled Systems

Distributed Coordinated Tracking With Reduced Interaction via a Variable Structure Approach Yongcan Cao, Member, IEEE, and Wei Ren, Member, IEEE

Complex Laplacians and Applications in Multi-Agent Systems

Distance-based Formation Control Using Euclidean Distance Dynamics Matrix: Three-agent Case

Adaptive State Feedback Nash Strategies for Linear Quadratic Discrete-Time Games

Zeno-free, distributed event-triggered communication and control for multi-agent average consensus

Automatica. Distributed discrete-time coordinated tracking with a time-varying reference state and limited communication

Decentralized Stabilization of Heterogeneous Linear Multi-Agent Systems

Formation Control and Network Localization via Distributed Global Orientation Estimation in 3-D

Consensus Algorithms are Input-to-State Stable

Consensus Tracking for Multi-Agent Systems with Nonlinear Dynamics under Fixed Communication Topologies

ANALYSIS OF CONSENSUS AND COLLISION AVOIDANCE USING THE COLLISION CONE APPROACH IN THE PRESENCE OF TIME DELAYS. A Thesis by. Dipendra Khatiwada

OUTPUT CONSENSUS OF HETEROGENEOUS LINEAR MULTI-AGENT SYSTEMS BY EVENT-TRIGGERED CONTROL

Consensus of Information Under Dynamically Changing Interaction Topologies

MULTI-AGENT TRACKING OF A HIGH-DIMENSIONAL ACTIVE LEADER WITH SWITCHING TOPOLOGY

Consensus Seeking in Multi-agent Systems Under Dynamically Changing Interaction Topologies

Consensus Protocols for Networks of Dynamic Agents

Consensus Seeking in Multi-agent Systems Under Dynamically Changing Interaction Topologies

Distributed Estimation for Motion Coordination in an Unknown Spatially Varying Flowfield

arxiv: v1 [cs.sy] 6 Jun 2016

Active Passive Networked Multiagent Systems

Discrete Double Integrator Consensus

arxiv: v1 [math.oc] 15 Jan 2016

Formation Control of Nonholonomic Mobile Robots

Tracking control for multi-agent consensus with an active leader and variable topology

Consensus Based Formation Control Strategies for Multi-vehicle Systems

Scaling the Size of a Multiagent Formation via Distributed Feedback

Target Tracking and Obstacle Avoidance for Multi-agent Networks with Input Constraints

Fast Linear Iterations for Distributed Averaging 1

2778 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 56, NO. 12, DECEMBER 2011

Combining distance-based formation shape control with formation translation

Distributed Computation of Minimum Time Consensus for Multi-Agent Systems

FORMATIONS OF FORMATIONS: HIERARCHY AND STABILITY

AFRL-RW-EG-TP

Non-Collision Conditions in Multi-agent Robots Formation using Local Potential Functions

Consensus Problem in Multi-Agent Systems with Communication Channel Constraint on Signal Amplitude

Networked Control Systems, Event-Triggering, Small-Gain Theorem, Nonlinear

EXPERIMENTAL ANALYSIS OF COLLECTIVE CIRCULAR MOTION FOR MULTI-VEHICLE SYSTEMS. N. Ceccarelli, M. Di Marco, A. Garulli, A.

1 Introduction. 2 Successive Convexification Algorithm

Safe Autonomous Agent Formation Operations Via Obstacle Collision Avoidance

Max-Consensus in a Max-Plus Algebraic Setting: The Case of Fixed Communication Topologies

Game Theoretic Approach to Power Control in Cellular CDMA

Distributed Structural Stabilization and Tracking for Formations of Dynamic Multi-Agents

Consensus Analysis of Networked Multi-agent Systems

Three-Dimensional Motion Coordination in a Spatiotemporal Flowfield

On Equilibria of Distributed Message-Passing Games

Multi-Robotic Systems

The Multi-Agent Rendezvous Problem - The Asynchronous Case

A Price-Based Approach for Controlling Networked Distributed Energy Resources

Notes on averaging over acyclic digraphs and discrete coverage control

Consensus of Hybrid Multi-agent Systems

Graph and Controller Design for Disturbance Attenuation in Consensus Networks

Min-max Time Consensus Tracking Over Directed Trees

Stability Analysis of Stochastically Varying Formations of Dynamic Agents

On the Scalability in Cooperative Control. Zhongkui Li. Peking University

Constrained Optimization and Distributed Computation Based Car-Following Control of A Connected and Autonomous Vehicle Platoon

IN the multiagent systems literature, the consensus problem,

Finite-Time Distributed Consensus in Graphs with Time-Invariant Topologies

Towards Decentralization of Multi-robot Navigation Functions

Target Localization and Circumnavigation Using Bearing Measurements in 2D

ON SEPARATION PRINCIPLE FOR THE DISTRIBUTED ESTIMATION AND CONTROL OF FORMATION FLYING SPACECRAFT

Stability and Disturbance Propagation in Autonomous Vehicle Formations : A Graph Laplacian Approach

Decentralized Control of Nonlinear Multi-Agent Systems Using Single Network Adaptive Critics

Coordinated Path Following for Mobile Robots

Stability and Performance of Non-Homogeneous Multi-Agent Systems on a Graph

Formation Stabilization of Multiple Agents Using Decentralized Navigation Functions

Passivity-based Formation Control for UAVs with a Suspended Load

Information Structures, the Witsenhausen Counterexample, and Communicating Using Actions

Periodic Behaviors in Multi-agent Systems with Input Saturation Constraints

Consensus Control of Multi-agent Systems with Optimal Performance

A Single-letter Upper Bound for the Sum Rate of Multiple Access Channels with Correlated Sources

Distributed LQR Design for Dynamically Decoupled Systems

Distributed Tracking ControlforLinearMultiagent Systems With a Leader of Bounded Unknown Input

Linear Algebra and its Applications

Noncooperative continuous-time Markov games

Research on Consistency Problem of Network Multi-agent Car System with State Predictor

Flocking of Discrete-time Multi-Agent Systems with Predictive Mechanisms

Average-Consensus of Multi-Agent Systems with Direct Topology Based on Event-Triggered Control

Learning Approaches to the Witsenhausen Counterexample From a View of Potential Games

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games

Bounded Budget Betweenness Centrality Game for Strategic Network Formations

An Optimal Tracking Approach to Formation Control of Nonlinear Multi-Agent Systems

Decentralized Formation Control and Connectivity Maintenance of Multi-Agent Systems using Navigation Functions

Multi-agent Second Order Average Consensus with Prescribed Transient Behavior

Distributed motion constraints for algebraic connectivity of robotic networks

Localization with Imprecise Distance Information in Sensor Networks

arxiv: v2 [cs.ro] 26 Sep 2016

Distributed model predictive control of large-scale systems

A Novel Integral-Based Event Triggering Control for Linear Time-Invariant Systems

A Robust Event-Triggered Consensus Strategy for Linear Multi-Agent Systems with Uncertain Network Topology

Risk-Sensitive and Robust Mean Field Games

Consensus seeking on moving neighborhood model of random sector graphs

Computation of Efficient Nash Equilibria for experimental economic games

OSNR Optimization in Optical Networks: Extension for Capacity Constraints

Using Orientation Agreement to Achieve Planar Rigid Formation

Almost Global Asymptotic Formation Stabilization Using Navigation Functions

Notes on Blackwell s Comparison of Experiments Tilman Börgers, June 29, 2009

Distributed Coordination Algorithms for Multiple Fractional-Order Systems

Dynamic region following formation control for a swarm of robots

Transcription:

5rd IEEE Conference on Decision and Control December 5-7, 4. Los Angeles, California, USA Distributed Game Strategy Design with Application to Multi-Agent Formation Control Wei Lin, Member, IEEE, Zhihua Qu, Fellow, IEEE,, and Marwan A. Simaan, Life Fellow, IEEE Abstract In this paper, we consider a multi-agent formation control problem from a game theory point of view. It is well known that a major difficulty in a communication network based formation control problem is that each agent is only able to exchange information with other agents according to the communication topology. This information constraint prevents many game strategy design approaches that require individual agents to have global information from being implemented in many cases. We formulate the formation control problem in such a way that individual agents try to minimize their locally measured formation errors and to solve it as a differential game problem. We consider two cases of non-cooperative and cooperative games and propose a novel distributed design approach that utilizes the relationship between the initial and terminal state variables. This approach is applied to an illustrative formation control example among three agents and the formation errors under various scenarios are compared and analyzed. I. INTRODUCTION An important application of cooperative control theory, is the multi-agent formation control problems, 4, 5. This problem revolves around designing local information constrained control inputs for the agents (mobile robots, unmanned vehicles, etc) and making their positions follow a prescribed formation. There is considerable work done in this field and the comprehensive review paper 6 on multi-agent distributed control systems is a good reference on recent progress in cooperative control theory including the formation control problem. Recently, multiple attempts 7, 8, 9 have been made to incorporate differential game theory, into the formation control design problem. A fundamental connection between these two problems can be explained as follows: in a multi-agent formation control problem, all the agents are basically trying to minimize the formation errors between each other. If every agent is able to observe all the other agents, then the problem can be essentially viewed as an optimal control problem where a common performance index of global formation errors can be assigned for all the agents. However, in most of realistic applications, since the agents are usually connected through a communication network with a certain topology, they are This work is supported in part by US National Science Foundation under grants ECCS-898 and CCF-9565, by US Department of Energys award DE-EE64, by US Department of Transportations award DTRT-G-UTC5, by L- Communications contract I4, and by Leidos contract P65. Wei Lin received his PhD degree from University of Central Florida in Dec. and is currently working in Western Digital Corporation, Irvine, CA, 96, USA. E-mail: weilin99@gmail.com. Zhihua Qu and Marwan A. Simaan are with the Department of EECS, University of Central Florida, Orlando, Florida, 86, USA, e-mails: qu@eecs.ucf.edu, simaan@eecs.ucf.edu. only able to exchange information with each other in a distributed manner (i.e., to share information locally). In this case, instead of minimizing a globally defined performance index, it is more realistic for each agent to only minimize a performance index of its locally measured formation errors. This scenario is exactly equivalent to that of a differential game where each player is trying to minimize its own performance index. Therefore, most multi-agent formation control design problems can be viewed and solved as differential game problems. The game solutions including Nash equilibria for non-cooperative agents and Pareto optimality for cooperative agents can be well established. However, solving for the Nash and Pareto solutions under distributed information is not a trivial task in the multi-agent environment because of the following two issues: () the strategy for each agent has to conform to its local information topology constraint and () local strategy computing within each agent is preferred rather than centralized strategy computing where strategies are designed by a centralized designer and assigned to individual agents. In 8, 9, the first issue has been addressed and the corresponding Nash and Pareto strategies are considered for the formation control problem as a differential game. In this present paper, we will focus more on the second aspect. By investigating the structure of the game solution in more details, a strategy design approach as well as an implementation algorithm are proposed for each agent to carry out its optimal control strategy design in a fully distributed manner. The remainder of this paper is organized as follows. A brief setup of the game problem is presented in Section II. The main results of the distributed Nash and Pareto strategy designs are presented in Section III and IV respectively. An illustrative example with three agents under noncooperative and cooperative cases is presented and analyzed in Section V. II. PROBLEM FORMULATION We consider the finite-time formation control problem between N agents in the n-dimensional Euclidean space where the motion dynamics of each agent is described as the following double integrator model: ṗi In pi = + u i () v i for i =,,N. Vector p i R n and v i R n are the position and velocity of agent i respectively. Vector u i R n is the acceleration control of agent i. We assume that the agents are able to exchange information with each other through a communication network whose topology can be v i I n 978--467-688-/4/$. 4 European Union 4

described by a set E in graph theory. An edge e ij E represents information being transmitted from agent j to agent i. The objective of the agents is to form a prescribed formation over a finite time interval, which is equivalent to individual agents minimizing the following performance indices: J i = p i (t f ) p j (t f ) µ ij + v i (t f ) v j (t f ) + r i tf u i dt () for i =,,N, where is the Euclidean norm, µ ij is the desired displacement vector pointing from agent j to agent i, and r i is a positive scalar. Minimizing the performance index J i in () means that every agent will try to minimize the total terminal formation error and terminal velocity error according to the graph while at the same time minimizing its control effort during the time interval. Since every agent is only aware of its local communication pattern and its own performance index, this formation problem can be regarded as a differential game problem where each player has its own objectives to pursue over a time interval. In this paper, we consider two types of open-loop (where the control inputs are functions of initial states and time only) formation control approaches through differential game strategy design. Specifically, if the control inputsu,,u N can be looked as strategies, then for the formulated problem, strategies u N,,uN N are the Nash strategies and form a Nash equilibrium if the inequality J i (u N,,u N i,,u N N) J i (u N,,u i,,u N N) () holds for all u i U i and for all i =,,N, where U i is the open-loop control set for agent i. The Nash equilibrium is an equilibrium where it is not possible for any agent to lower its performance index value by unilaterally deviating from its Nash strategy. The Nash equilibrium is a widely used solution in the situation where the players do not cooperate with each other. On the other hand, The strategies u P,u P,,u P N are the Pareto strategies and form a Pareto optimality if the inequality J i (u P,,u P N) < J i (u,,u N ) (4) holds for at least one i {,,N}. The Pareto optimality can be interpreted as a solution in which any changes made do not help lower every agent s performance index value. The Pareto optimality is a widely used solution in the situation where the players can cooperate with each other. In the following sections, two formation control approaches will be proposed based on the open-loop Nash strategy (assuming that the agents do not cooperate) and the open-loop Pareto strategy (assuming that the agents cooperate). III. DISTRIBUTED NASH STRATEGY DESIGN In this section, assuming that the agents do not corporate, we first consider the formation control through distributed open-loop Nash strategy design. We introduce the following new state variable: s i (t) = p i (t)+(t f t)v i (t) (5) for agent i. Then the performance index () becomes J i = s i(t f ) s j (t f ) µ ij + v i (t f ) v j (t f ) + r i tf u i dt. (6) for i =,,N. The dynamics of the new state variables is given by ṡ i = (t f t)u i and v i = u i. (7) Based on the above performance indices and state dynamics, we can solve inequality () for i =,,N as N simultaneous open-loop optimal control problems (by using the approach of calculus of variation ) and obtain the following open-loop Nash strategies: u N i = t f t r i s i (t f ) s j (t f ) µ ij v i (t f ) v j (t f ) i =,,N (8) r i which is analogous to the open-loop Nask solutions provided in. Theoretically, the terminal variabless i (t f ) andv i (t f ) can be obtained by substituting the open-loop control (8) back into the system dynamics (7) and solving the equations for all i =,,N simultaneously. However, solving these equations is beyond the capability of each agent because the agent is only aware of its own performance index and its local communication topology. To overcome this difficulty, one approach is to let all the agents exchange certain information across the communication network so that the terminal variable values s i (t f ) and v i (t f ) can be obtained by agent i in a distributed manner for all i =,,N. Toward this end, the following theorem is proposed. Theorem : If agent i updates vectors h si and continuously according to ḣsi ḣ vi + (t f /)ρ i r i (t f /)ρ i (W I n ) ( hsi r i =k i { s i () v i () hsi hsj h vj ) } (9) where k i is a positive scalar, t f / t f W = / t f / t, () f ρ i = µ ij i =,,N, () then lim τ hsi (τ) si (t f ) = (τ) v i (t f ) () 44

for i =,,N, where s i (t f ) and v i (t f ) are the terminal variables in Nash strategies (8). Proof: Firstly, substituting the open-loop strategies (8) back into system dynamics (7) and integrating the equations yields s(tf ) s() (t f /)(R I n ) M = + v(t f ) v() (t µ () f /)(R I n ) where s = s T s T N T, v = v T v T N T, R = diag{r,,r N }, µ = µ T µt N T, and M = I N +W (R L) I n, (4) where matrix L = l ij R N N is the Laplacian matrix associated with the information graph among the agents and if e ij E for j i if e ij / E for j i l ij = N, (5) l iq if j = i q=,q i Since all the eigenvalues of a Laplacian matrix have nonnegative real parts, all the eigenvalues of matrix M in (4) have positive real parts and matrix M is hence invertible. Secondly, combining equations (9) for i =,,N in a more compact vector form yields { ḣs s() hs (t f /)(R } I n ) =K M + ḣ v v() h v (t µ, f /)(R I n ) (6) where h s = h T s ht sn T, h v = h T v ht vn T, K = blkdiag{k (I n ),,k N (I n )}. Since all the eigenvalues of matrix M has positive real parts, system (6) is stable and the final convergent values of h s and h v are identical to the solution of s(t f ) and v(t f ) in equation (). Therefore, hs (τ) s(tf ) lim τ h v (τ) = v(t f ) and hence equation () holds for all i =,,N. Based on the above theorem, all the agents can exchange information with each other through the communication network and estimate the unknown terminal values in (8). Therefore, the following implementation algorithm is proposed. Algorithm :. All the agents initialize the state s i () = p i ()+t f v i () for all i =,,N.. For a certain period of time ( t), all the agents communicate with each other through the network and update the states h si and according to (9) and obtain the estimate of s i (t f ) and v i (t f ) for all i =,,N.. At t = t, all the agents implement the open-loop Nash strategies in (8) with s i (t f ) and v i (t f ) replaced by the estimated terminal states h si ( t) and ( t). The above algorithm is fully distributed (except for the knowledge of time instants and t) because agent i only. need to send out its own state information h si and to the agents that have direct information injection links from agent i and receive the state variables from the agents that have direct information links to agent i. Note that the terminal state estimate will approximate to the ideal Nash equilibrium strategies since the convergence of the linear system (9) is asymptotic in nature. Therefore, the strategies obtained in the above algorithm are also known to form an ǫ-nash equilibrium. A quantitative analysis is provided in 4. IV. DISTRIBUTED PARETO STRATEGY DESIGN In this section, assuming that the agents cooperate, we consider the formation control through distributed open-loop Pareto strategy design. The Pareto strategies can be derived through solving the following convex combination of all the agents performance indices J = N α j J j (7) j= where < α j < and N j= α j =. In this paper, the value of α j has to lie strictly within the interval between and for all j =,,N, which basically means that all agents have to be taken into account in (7). This minimization is essentially an optimal control problem with any fixed parameters α,,α N. Adopting the new state representation in (5), the convex combination in (7) with J i defined in (6) can be expressed as J = N j= α j { s j (t f ) s k (t f ) µ jk e jk E + v j (t f ) v k (t f ) +r j tf } u j dt (8) for i =,,N. Now, since all the agents will minimize the common performance index above, it is necessary to make several assumption in order to have a distributed design approach. Assumption : The communication link between any two agents is assumed to be bidirectional, that is, the information graph G is undirected. Assumption : Agent i has the knowledge of the convex parameter α j of agent j for all e ij E. Based on the above assumptions, the following Pareto strategies can be obtained by solving inequality (4) as an open-loop optimal control problem using the calculus of variation approach again: u P i = t f t (α i +α j )s i (t f ) s j (t f ) µ ij (α i +α j )v i (t f ) v j (t f ). (9) which is analogous to the non-inferior solutions provided in 5. Since the value of α i lies strictly within the interval between and for all i =,,N, α i can be safely 45

placed in the denominator in (9). The terminal states can be obtained by substituting the strategies (9) into the system dynamics (7) and solving the N equations. Due to the local information constraint, we propose the following theorem similar to Theorem to overcome agent i s inability to solve these equations. Theorem : If agent i updates vectors h si and continuously according to (t f /)ρ i hsi ḣsi ḣ vi = k i { s i () v i () (W I ) + (α i +α j ) (t f /)ρ i ( hsi hsj h vj ) } () where k i is a positive scalar, matrix W is defined in (), and ρ i = (α i +α j )µ ij, then the limit () holds for i =,,N, where s i (t f ) and v i (t f ) are the terminal variables in Pareto strategies (9). Proof: Similar to the proof of Theorem, substituting the open-loop strategies (9) into system dynamics (7) yields equation () where matrix M is defined in (4) with (α i +α j ) if e ij E for j i if e ij / E for j i l ij = N, () l iq if j = i q=,q i for the Laplacian matrix L. Stacking equation () from i = to i = N yields an equation similar to (6). Since matrix ( M) is Hurwitz. Therefore, the estimator is asymptotically convergent as shown in (). Based on the above theorem, the same algorithm as Algorithm can be carried out to obtain Pareto strategies in practice. A special case to consider is when the agent does not receive any information from the other agents. In this situation, according to the expressions in (8) and (9), either the Nash or Pareto strategy becomes zero and hence the agent will keep a constant velocity, which is still a reasonable choice in practice. V. ILLUSTRATIVE EXAMPLE In this section, we consider an illustrative formation control example in two-dimensional space and apply the proposed distributed Nash and Pareto strategy design approaches. The dynamics of the agents are given by () with n =. For the ease of illustration, the horizontal axis is called the x-axis and the vertical axis is called the y-axis. Various scenarios are considered as follows. Scenario : We start with a simple case, where there are three agents with directed communication topology as shown in Figure, where agent as the leader does not have information injection from the other agents but only sends Fig. : Three agents with undirected graph its information out to the other agents. Hence, the Laplacian matrix defined in (5) is given by L =. The initial states of the three agents are given by p = T,p = T,p = T, v = T,v = T,v = T, which means their initial positions are equally spaced and agent s initial velocity is upward and agents and s initial velocities are equal to zero. We assume that the desired formation among the agents is a triangle with the displacement vectors given by µ = T, µ = T. In performance indices (), the parameterr i is given by r = r = r = andt f = for the time interval of the formation control. Since the information graph is directed, we only consider the formation control using the distributed Nash strategy design in this scenario. For comparison purpose, we solve for the actual Nash strategies by directly solving the equations (7) with strategies in (8) substituted in. The Nash strategy components on x-axis and y-axis are shown as blue trajectories in Figure and the agents terminal states under the Nash strategies are s (t f ) =, v (t f ) =, 6.. s (t f ) =, v (t f ) =,.65.9.. s (t f ) =, v (t f ) =..65.9 If we implement the distributed Nash strategy design algorithm described in Algorithm with t =. and k = k = k = in (9), then the final terminal state estimate values are h s ( t) =, h v ( t) =, 5.5.84.9. h s ( t) =, h v ( t) =,.6.7.9. h s ( t) =, h v ( t) =.6.7 and the corresponding control strategies (8) with s i (t f ) and v i (t f ) replaced by the estimated terminal states h si ( t) and 46

( t) fori =,, are shown as red circled trajectories in figure. We can see that the largest terminal state estimation error is.4% and the resulting control inputs slightly deviate from the original Nash strategies. This estimation error can be dramatically reduced with larger values of k, k, and k. Therefore, if we implement the same algorithm with t =. and k = k = k = 4, then the final estimate values becomes h s ( t) =, h v ( t) =, 5.89.96.. h s ( t) =, h v ( t) =,.54.86.. h s ( t) =, h v ( t) =..54.86 and the largest terminal state estimation error is %. In this case, the corresponding control strategies (8) with s i (t f ) and v i (t f ) replaced by the estimated terminal states h si ( t) and ( t) for i =,, are shown as green crossed trajectories in figure, which are almost identical to original Nash strategies. The motion trajectories under various control.5.5 u x.4...4.4.. u x u x.4.5.5 u y.5.5 u y.5.5 u y Original Nash Control Proposed Design k i = Proposed Design k i =4 Fig. : Nash and proposed designed strategies inputs are shown in Figure. Graphically, there does not exist much difference between the trajectories under the original Nash strategies and the ones under designed strategies. Quantitatively, we are able to calculate the total terminal formation error with respect to the desired displaced among the agents, which is η = s (t f ) s i (t f ) µ i, () i= 6 5 4 Agent Agent Original Nash Control Proposed Design k i = Proposed Design k i =4 Agent Fig. : Motion trajectories under different control strategies where s i (t f ) is agent i s terminal position vector under certain control strategies. Therefore, the total terminal formation errors based on different values of r = r = r = r and k = k = k = k can be summarized into Table I. From Nash k = k = 4 η (r =.).6.77.6 η (r = ).864.85.879 η (r = 5).654.69.656 TABLE I: Total terminal formation errors in Scenario the table, it is clear to see that the total terminal formation error decreases ask i increase because as mentioned, largerk i will result in more accurate terminal state estimates and more accurate control inputs (with respect to the Nash strategies). It is also clear to see that the total terminal formation error increases as r increases. This is because that if agents put higher penalties on the control effort or energy (i.e., the quadratic term of the control input in ()), they will try to reduce more on control energy instead of the terminal formation errors in order to minimize their performance indices in () during the time period of the formation control. Based on more simulation data, the relationship among η, k, and r can be depicted in Figure 4. As we can see in the η 6 5 4 k=5 k=6 k=8 k= k= k=4 Nash.5.5.5.5 4 4.5 5 r Fig. 4: Relationship among η, r, and k 47

figure, when r is small, the variance of the total terminal formation error η among different k is small, and when r is large, the variance of the total terminal formation error η among different k becomes large. Another application of Figure 4 is to determine the lowest update gain k needs to achieve a desired terminal formation error. For instance, if we want to determine a update gain that can achieve a maximum terminal formation error of given r =, then we can find that a value of k less than 6 is good choice from the figure. Therefore, this figure provides a intuitive way of choosing a proper update gain, especially to avoid choosing a high gain which is undesired in many real life applications. Scenario : We now consider the scenario where there are three agents with undirected communication topology as shown in Figure 5. With this communication topology, both Fig. 5: Three agents with undirected graph Nash and Pareto strategies can be derived. The initial states of the agents and the performance index parameters are the same as the ones in Scenario. For both Nash strategy design and Pareto strategy design, the gain k i in the terminal state estimation equation (9) and () are set equal to 4 for all i =,,. For the Pareto strategies, the convex parameters in (7) are set to be α = α = α = /. The Laplacian matrix for the noncooperative in (5) and the one for the cooperative case in () are L = and L = 4, respectively. The resulting motion trajectories under both Nash and Pareto strategies are shown in Figure 6. The y.5.5.5.5 Nash Pareto x Fig. 6: Motion trajectories under Nash and Pareto strategies total terminal formation errors under different choices of the convex parameters are summarized in Table II. As we can clearly see in the table, the agents can actually achieve lower formation error using Pareto strategies than using Nash strategies. This is because the agents are assumed to η Nash.4856 Pareto (α = α = α = /). Pareto (α = /6,α = /,α = /6).8 Pareto (α = /,α = /6,α = /6).79 Pareto (α =.9,α =.5,α =.5).6 TABLE II: Total terminal formation errors in Scenario cooperate with each other in the Pareto optimality. Also note that as we place more weight on agent s performance index, the total formation error can be reduced. This is because it is much easier to move agent alone to form the desired formation instead of moving all the three agents together. VI. CONCLUSION This paper considers the design of open-loop Nash and Pareto strategies for multi-agent formation control problems. The proposed approach utilizes the distributed estimation of the terminal state variables among the agents through network-based information exchange. The convergence rate of the terminal state estimation algorithm is determined by a scalar gain and can be made sufficiently fast to get a better approximation of the real Nash and Pareto strategies. An illustrative example is solved with two scenarios and various comparisons have been made. REFERENCES W. Ren, R. W. Beard, and E. M. Atkins, Information consensus in multivehicle cooperative control: Collective group behavior through local interaction, IEEE Control Syst. Mag., vol. 7, pp. 7 8, 7. Z. Qu, Cooperative Control of Dynamical Systems: Applications to Autonomous Vehicles. London: Springer Verlag, 9. D. M. Stipanovic, G. Inalhan, R. Teo, and C. J. Tomlin, Decentralized overlapping control of a formation of unmanned aerial vehicles, Automatica, vol. 4, no. 8, pp. 85 96, 4. 4 T. Keviczky, F. Borrelli, K. Fregene, D. Godbole, and G. Balas, Decentralized receding horizon control and coordination of autonomous vehicle formations, IEEE Trans. Control System Technology, vol. 6, no., pp. 9, 8. 5 J. Wang and M. Xin, Integrated optimal formation control of multiple unmanned aerial vehicles, IEEE Trans. Control System Technology, vol. PP, no. 99, pp.,. 6 Y. Cao, W. Yu, W. Ren, and G. Chen, An overview of recent progress in the study of distributed multi-agent coordination, IEEE Trans. Industrial Informatics, vol. 9, no., pp. 47 48,. 7 M. R. Anderson and A. C. Robbins, Formation flight as a cooperative game, in AIAA Guidance, Navigation and Control Conference, vol., 998, pp. 44 5. 8 D. Gu, A differential game approach to formation control, IEEE Trans. Control Systems Technology, vol. 6, no., pp. 85 9, 8. 9 E. Semsar-Kazerooni and K. Khorasani, Multi-agent team cooperation: A game theory approach, Automatica, vol. 45, no., pp. 5, 9. R. Isaacs, Differential Games. John Wiley and Sons, 965. T. Basar and G. J. Olsder, Dynamic Noncooperative Game Theory, nd ed. Philadelphia, PA: SIAM, 998. E. Bryson and Y. C. Ho, Applied Optimal Control. Bristol, PA, USA: Hemisphere Publishing Corp., 975. Y. Ho, A. B. Jr., and S. Baron, Differential games and optimal pursuitevasion strategies, IEEE Trans. Automatic Control, vol. AC-, no. 4, pp. 85 89, 965. 4 W. Lin, Differential games for multi-agent systems under distributed information, Ph.D. dissertation, University of Central Florida,. 5 A. Starr and Y. Ho, Nonzero-sum differential games, Journal of Optimization Theory and Applications, vol., pp. 84 6, 969. 48