A Separation Principle for Decentralized State-Feedback Optimal Control

Similar documents
Decentralized LQG Control of Systems with a Broadcast Architecture

Structured State Space Realizations for SLS Distributed Controllers

McGill University Department of Electrical and Computer Engineering

arxiv: v1 [cs.sy] 24 May 2013

Time-Varying Systems and Computations Lecture 3

Optimal Decentralized Control with. Asymmetric One-Step Delayed Information Sharing

Adaptive State Feedback Nash Strategies for Linear Quadratic Discrete-Time Games

RECURSIVE ESTIMATION AND KALMAN FILTERING

Multi-Model Adaptive Regulation for a Family of Systems Containing Different Zero Structures

STATE ESTIMATION IN COORDINATED CONTROL WITH A NON-STANDARD INFORMATION ARCHITECTURE. Jun Yan, Keunmo Kang, and Robert Bitmead

Network Reconstruction from Intrinsic Noise: Non-Minimum-Phase Systems

Semidefinite Programming Duality and Linear Time-invariant Systems

arxiv: v2 [math.oc] 14 Dec 2015

OPTIMAL CONTROL AND ESTIMATION

RECURSIVE SUBSPACE IDENTIFICATION IN THE LEAST SQUARES FRAMEWORK

Zeros and zero dynamics

Linear Quadratic Gausssian Control Design with Loop Transfer Recovery

Stabilization with Disturbance Attenuation over a Gaussian Channel

arxiv: v1 [math.oc] 18 Mar 2014

Structured Stochastic Uncertainty

Model reduction for linear systems by balancing

Optimal Decentralized Control of Coupled Subsystems With Control Sharing

EE226a - Summary of Lecture 13 and 14 Kalman Filter: Convergence

6 OUTPUT FEEDBACK DESIGN

15 Singular Value Decomposition

Mathematical Relationships Between Representations of Structure in Linear Interconnected Dynamical Systems

Event-Triggered Decentralized Dynamic Output Feedback Control for LTI Systems

Chap. 3. Controlled Systems, Controllability

Decentralized Stochastic Control with Partial Sharing Information Structures: A Common Information Approach

EL2520 Control Theory and Practice

The norms can also be characterized in terms of Riccati inequalities.

Closed-Loop Structure of Discrete Time H Controller

Optimal Decentralized Control with Asymmetric One-Step Delayed Information Sharing

On the simultaneous diagonal stability of a pair of positive linear systems

Stability Analysis of Linear Systems with Time-varying State and Measurement Delays

NP-hardness of the stable matrix in unit interval family problem in discrete time

Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013

ON SEPARATION PRINCIPLE FOR THE DISTRIBUTED ESTIMATION AND CONTROL OF FORMATION FLYING SPACECRAFT

Sufficient Statistics in Decentralized Decision-Making Problems

Prediction-based adaptive control of a class of discrete-time nonlinear systems with nonlinear growth rate

Multi-Robotic Systems

Here represents the impulse (or delta) function. is an diagonal matrix of intensities, and is an diagonal matrix of intensities.

Balanced Truncation 1

Linear-Quadratic Optimal Control: Full-State Feedback

Identification and estimation of state variables on reduced model using balanced truncation method

A Partial Order Approach to Decentralized Control

OPTIMAL FUSION OF SENSOR DATA FOR DISCRETE KALMAN FILTERING Z. G. FENG, K. L. TEO, N. U. AHMED, Y. ZHAO, AND W. Y. YAN

Information Structures, the Witsenhausen Counterexample, and Communicating Using Actions

ASIGNIFICANT research effort has been devoted to the. Optimal State Estimation for Stochastic Systems: An Information Theoretic Approach

An LQ R weight selection approach to the discrete generalized H 2 control problem

A Characterization of the Hurwitz Stability of Metzler Matrices

Output Regulation of Uncertain Nonlinear Systems with Nonlinear Exosystems

Common Knowledge and Sequential Team Problems

Homework 2 Foundations of Computational Math 2 Spring 2019

K 1 K 2. System Level Synthesis: A Tutorial. John C. Doyle, Nikolai Matni, Yuh-Shyang Wang, James Anderson, and Steven Low

Static Output Feedback Stabilisation with H Performance for a Class of Plants

Model reduction of interconnected systems

LMI based output-feedback controllers: γ-optimal versus linear quadratic.

Minimal Interconnection Topology in Distributed Control Design

An alternative proof of the Barker, Berman, Plemmons (BBP) result on diagonal stability and extensions - Corrected Version

Linear-Quadratic-Gaussian (LQG) Controllers and Kalman Filters

w T 1 w T 2. w T n 0 if i j 1 if i = j

Denis ARZELIER arzelier

EL 625 Lecture 10. Pole Placement and Observer Design. ẋ = Ax (1)

SELECTION OF VARIABLES FOR REGULATORY CONTROL USING POLE VECTORS. Kjetil Havre 1 Sigurd Skogestad 2

DESIGNING A KALMAN FILTER WHEN NO NOISE COVARIANCE INFORMATION IS AVAILABLE. Robert Bos,1 Xavier Bombois Paul M. J. Van den Hof

Contents. 1 State-Space Linear Systems 5. 2 Linearization Causality, Time Invariance, and Linearity 31

Robust Control 5 Nominal Controller Design Continued

Derivation of the Maximum a Posteriori Estimate for Discrete Time Descriptor Systems

Backward Error Estimation

Linear Algebra and Robot Modeling

Optimal H Control Design under Model Information Limitations and State Measurement Constraints

LQR, Kalman Filter, and LQG. Postgraduate Course, M.Sc. Electrical Engineering Department College of Engineering University of Salahaddin

Lecture 9. Introduction to Kalman Filtering. Linear Quadratic Gaussian Control (LQG) G. Hovland 2004

6.241 Dynamic Systems and Control

Optimal control and estimation

Design of Observer-Based 2-d Control Systems with Delays Satisfying Asymptotic Stablitiy Condition

Chapter 2 Event-Triggered Sampling

Decentralized Multirate Control of Interconnected Systems

CALIFORNIA INSTITUTE OF TECHNOLOGY Control and Dynamical Systems. CDS 110b

Performance assessment of MIMO systems under partial information

State Regulator. Advanced Control. design of controllers using pole placement and LQ design rules

2. Matrix Algebra and Random Vectors

Linear Quadratic Zero-Sum Two-Person Differential Games

THE STABLE EMBEDDING PROBLEM

Unbiased minimum variance estimation for systems with unknown exogenous inputs

IEOR 265 Lecture 14 (Robust) Linear Tube MPC

FINITE HORIZON ROBUST MODEL PREDICTIVE CONTROL USING LINEAR MATRIX INEQUALITIES. Danlei Chu, Tongwen Chen, Horacio J. Marquez

Optimization-Based Control

An Optimization-based Approach to Decentralized Assignability

Discrete-Time H Gaussian Filter

Appendix A Solving Linear Matrix Inequality (LMI) Problems

Optimal Polynomial Control for Discrete-Time Systems

Towards control over fading channels

Robust Stabilization of Non-Minimum Phase Nonlinear Systems Using Extended High Gain Observers

Stability, Pole Placement, Observers and Stabilization

Iterative Learning Control Analysis and Design I

Graph and Controller Design for Disturbance Attenuation in Consensus Networks

Contents lecture 5. Automatic Control III. Summary of lecture 4 (II/II) Summary of lecture 4 (I/II) u y F r. Lecture 5 H 2 and H loop shaping

SYSTEMTEORI - KALMAN FILTER VS LQ CONTROL

Transcription:

A Separation Principle for Decentralized State-Feedbac Optimal Control Laurent Lessard Allerton Conference on Communication, Control, and Computing, pp. 528 534, 203 Abstract A cooperative control problem is considered in which dynamically decoupled subsystems must control their own states through state feedbac in order to optimize a global quadratic cost. The states of the subsystems are coupled only through the cost function and correlated external disturbances. The architecture is truly decentralized; no communication between subsystems or their controllers is permitted. The main result of this paper is that the optimal decentralized controller satisfies a new separation principle that is striingly similar to the celebrated result from centralized optimal control theory, but does not appear to follow from it. Roughly speaing, the optimal decentralized control strategy for each subsystem is the product of a static control gain and a global state estimate, and each can be separately computed. Introduction The separation principle is a classical and celebrated result in optimal control theory. When it holds, the optimal controller may be found by composing the solutions to two separate problems of control and filtering. For a recent historical perspective, see [ and references therein. Separation holds for linear systems with Gaussian noise and a classical information pattern [6. Additionally, if the cost is quadratic (LQG assumptions), the optimal controller is a linear function of the conditional mean of the state. So the optimal controller taes the form u opt K E(x Y) () where K is a static control gain, x is the state of the plant, and Y is the history of available noisy measurements. A ey requirement is that of a classical information pattern; all decision-maing stations must receive the same information and have perfect recall. Even under LQG assumptions, a nonclassical information pattern may result in a cost function with multiple local minima, and a nonlinear controllers may strictly outperform any linear controller [4. However, some structural results are nown for special classes of nonclassical information patterns. If the L. Lessard is with the Department of Mechanical Engineering at the University of California, Bereley, CA 94720, USA. lessard@bereley.edu information pattern is partially nested, then there exist linear optimal strategies [2. In particular, onetimestep delay-sharing architectures also satisfy a separation structure [5. More recent structural results for partially nested information patterns include: poset-causal systems [2, delayed systems characterized by an underlying directed graph [3, 4 and output-feedbac with sparsity [5, 7. These examples provide sufficient statistics expressible in terms of conditional means. For example, the twoplayer output-feedbac problem with triangular information structure [7 has an optimal controller of the form u opt K E(x Y ) + K 2 E(x Y, Y 2 ) (2) where Y and Y 2 are the histories of two different noisy measurements. This structure is similar to that of () in that it provides a separation between control and estimation. However, the separation is only superficial, as the associated estimation and control problems are coupled and must be jointly solved. Structural results have also been obtained for information patterns that are not partially nested, such as partial history sharing [0. In this paper, we study a problem that differs from those mentioned above in two important ways. First, the subsystems and controllers are dynamically decoupled and there is no communication or sharing of information between controllers. Second, the exogenous noise affecting the subsystems may be correlated. A statespace construction of the optimal linear controller for this problem was presented in [6 and it was found to have a state dimension that depends on the extent of the correlation in the exogenous noise. This is a curious finding, as it seems to defy the possibility of a structural result of the form () or (2). Indeed, a state estimate such as E(x Y) has a state dimension equal to the dimension of x, regardless of the noise statistics. Our main result is that the fully decentralized problem studied in [6 does indeed have a separation structure in the spirit of (), but it uses modified control and filtering problems. This separation structure was previously unnown, and gives an explanation for why the optimal controller s state dimension depends on the noise statistics as observed in [6. The paper is organized as follows. We begin with a formal description of the problem in Section 2, followed by main results in Section 3 and discussion in Section 4.

In Section 5 we relate the new separation result to the wor in [6, and we summarize our findings in Section 7. 2 Problem statement We consider the simplest case of maximal decentralization; multiple subsystems are independently controlled, with no possibility of communication or influence between subsystems. The goal is to optimize a global objective that depends on both subsystems. A bloc diagram is shown in Figure. The problem considered in [6 had vector disturbances, but we will assume for now that each system is subject to a scalar disturbance. We will examine the vector case later in Section 5. x, u x P u Quadratic K cost function. x N, u N P N η η N Gaussian process noise where A diag(a,..., A N ) and similarly for B and H. The objective is to find a fully decentralized stabilizing control policy that minimizes the following infinitehorizon LQR cost: T T [ x Q S x J lim E T 0 u S R T dt (6) u Note that Q, R, and S may be full matrices in general, so the cost couples the states and inputs of the various subsystems. We assume R 0 and Q SR S T 0 as is standard, to ensure strict convexity of the cost. We see fully decentralized policies u i γ i (X i ) that depend on the state histories of the respective plants. Naturally, a necessary condition for the existence of a stabilizing controller is that (A i, B i ) be stabilizable for each i. We assume without loss of generality that each A i is Hurwitz, for if it were not, we could perform a loopshifting change of coordinates by writing u i F i x i + ũ i, where F i is any matrix such that A i + B i F i is Hurwitz. No communication is permitted between the subsystems or their controllers. Although the plants (3) are dynamically decoupled, the performance of their controllers is coupled through the noise (4) and cost function (6). x N K N u N 3 Main result Figure : Diagram of the decentralized control architecture considered in this paper. The N subsystems are coupled only through the process noise and the cost function; no other communication is permitted. Each P i is a continuous-time linear dynamical system with state x i R ni, input u i R mi, and scalar disturbance η i R. The state-space equations are as follows P i : ẋ i A i x i + B i u i + h i η i for i,..., N (3) For each plant, A i R ni ni is the state transition matrix, B i R ni mi is the controlled input matrix, and h i R ni is the exogenous input vector. The noise terms η,..., η N are assumed to be zero-mean, jointly Gaussian, and IID in time. We characterize this property via the N N covariance matrix η σ 2... σ N η 2 Σ : cov. σ 2... σ 2N...... 0 (4) η N σ N σ N2... The assumption that E(ηi 2 ) is not restrictive, as we may simply scale each equation so that each η i has unit variance. By concatenating x x T... x T T N and similarly for u and η, we can write one compact equation that describes the global dynamics. ẋ Ax + Bu + Hη (5) The classical separation principle for LQG problems states that the optimal policy has the form u K ˆx.. K is a solution to the state-feedbac version of the problem. In other words, if we had perfect state information, the optimal policy would be u Kx. This is also nown as the LQR controller. 2. ˆx is the estimate of the state that minimizes the mean-squared error conditioned on the available measurements. This is equal to the conditional expectation ˆx E(x Y), and is also nown as the Kalman-Bucy filter. The main reason for calling this a separation principle is that finding the optimal controller is reduced to finding the solutions to two simpler problems that can be solved separately. Another reason is that it effectively gives a separation between estimation and control. Indeed, the control gain K does not depend on noise covariances, nor does the optimal estimate ˆx depend on the cost function. The main result of this paper is a separation principle for the decentralized problem described in Section 2. As in the classical case, it leads to an optimal controller of the form u K ˆx, where the control gain K and optimal estimate ˆx can be computed separately by solving separate simpler problems. There is one important difference however. In this new separation principle, K does depend on the noise covariance, and ˆx does depend on the cost function. In other words, even though K and 2

ˆx can be computed separately, they each depend on all parameters of the problem. There are two ingredients to the new separation principle, and each is a well-understood classical problem. We now describe these simpler problems and explain some notation used for each one. Centralized LQR control. If we relax the constraint that u i measures only x i, and instead allow u i to measure the global state x, the problem described by (3) (6) is a standard centralized LQR problem. The solution is given by u Kx where K R (B T P + S T ), and A BR P Ric S T B T R B (Q SR S T ) (A BR S T ) T where Ric(H) denotes the unique stabilizing solution to the algebraic Riccati equation with Hamiltonian matrix H. See [7 for more details. As a convention, we will summarize the above solution using the compact notation K LQR(A, B, Q, R, S). Decentralized estimation. Suppose two systems are described by the following state-space equations. ẋ A x + B u + w ẋ 2 A 2 x 2 + B 2 u 2 + w 2 where the A i are stable and both u and u 2 are nown to the estimator. The noises w and w 2 are zero-mean jointly Gaussian random variables with covariance w W W W : cov 2 (7) w 2 W 2 W 22 Then E(x 2 X ) is given by the output of the following dynamical estimator q A 2 q + (A 2 L LA )x + B 2 u 2 LB u E(x 2 X ) q + Lx where L is any solution to LW W 2. Note that such an L exists even when W is not invertible. The estimator error covariance is given by cov(x 2 E(x 2 X )) Y where Y is a solution to the continuous-time Lyapunov equation A 2 Y + Y A T 2 + W 22 LW L T 0. The matrix A 2 is stable by assumption, so the Lyapunov equation has a unique solution and it satisfies Y 0. See [8 for a reference on how to construct such observers. We now show how these two ingredients are assembled to give a solution to the decentralized control problem described in Section 2. What follows is our main result, a new separation principle for decentralized control. Theorem. Consider the decentralized control problem described by (3) (6), and perform these two steps. (i) Partition Q, R, S into N N bloc matrices that conform to the dimensions n i and m i. Therefore, Q ij R ni nj, R ij R mi mj and S ij R ni mj. Now define the blocwise Hadamard product Q σ 2 Q 2... σ N Q N σ 2 Q 2 Q 22... σ 2N Q 2N Q Σ :...... σ N Q N σ N2 Q N2... Q NN and similarly for R Σ and S Σ. Finally, define K σ : LQR(A, B, (Q Σ), (R Σ), (S Σ)) Hence, K σ is the solution to a centralized LQR problem where the cost matrices have been weighted. (ii) For each i,..., N, consider the following system defined by the state z i ż i Az i + Bu i + hη i u i K σ z i (8) where h is the vertical concatenation of h,..., h N. This is similar to (5), except η i is driving every state of the system, not just state i. Now define ˆx i as the decentralized estimate ˆx i : E(z i X i ). The optimal decentralized controller is given by u i K σ i ˆx i where K σ i is the i th bloc-row of K σ. Proof. See Section 6. 4 Discussion The result of Theorem bears a striing resemblance to the centralized separation principle in terms of structure, but there are some important differences. Firstly, the control gain K σ depends on the noise covariance (4). This is different from the centralized case, where the control gain is independent of the noise covariance. Indeed, the centralized LQR problem is certainty equivalent, so the optimal controller remains the same even if noise is removed entirely and we consider a deterministic formulation of the problem. This feature is clearly not present in the decentralized LQR problem considered herein. Secondly, the estimators actually depend on the cost function as well. Indeed, in estimating the global states, each assumes the control policy of the other agents is a function of K σ, and this depends on the cost matrices 3

Q, R, and S. So, even though Theorem expressed the optimal decentralized controller as a product of the solutions to a control and an estimation problem that can be solved separately, these problems each depend on all parameters of the original problem. It is worth noting that the estimates ˆx i E(z i X i ) are always exact. In other words, the covariance of the estimation error is zero; Y cov(z i E(z i X i )) 0. This is intuitive because the same η i drives the entire state. So measuring x i is equivalent to measuring any other component of x. One could perhaps call the ˆx i inferences rather than estimates. In the case where Σ T, the matrix of all-ones, all noise signals are perfectly correlated. Therefore, all η i are equal and the global estimates ˆξ i are each identically equal to the global state x. Also, Q Σ Q and similarly for R and S. Therefore, K σ K LQR(A, B, Q, R, S), and we recover the optimal centralized policy u Kx. In the case where Σ I, the noise signals are independent. Also, Q Σ diag(q,..., Q NN ) and similarly for R and S. We therefore obtain the optimal greedy gains defined by K i : LQR(A i, B i, Q ii, R ii, S ii ). In other words, K σ diag(k,..., K N ) and we recover the optimal greedy policy u i K i x i. The cases discussed above show that in the extremes where the noise is perfectly correlated or independent, we recover the optimal centralized policy or the optimal greedy one, respectively. While these edge cases correspond to classical problems, the intermediate values of Σ do not. The ey fact revealed in Theorem is that a unifying separation principle holds for all Σ. 5 Extension to the vector noise case The case of two subsystems. We begin with the case N 2, where the scalar noises η, η 2 are replaced by vector noises. This problem was solved explicitly in [6, and the optimal controller was found to have a state dimension that depends on the noise covariance W. We will now show that under an appropriate change of coordinates, this vector-noise problem can be reduced to decoupled instances of the scalar noise problem described by (3) (6). The vector-noise problem for N 2 is described by the following equations. [ẋ A 0 x B 0 u w + + (9) ẋ 2 0 A 2 x 2 0 B 2 u 2 w 2 with a cost function of the form (6). We assume each state equation has been scaled such that w I X W : cov w 2 X T (0) I We further assume that each input has been shifted using u i ũ i + K i x i where K i LQR(A i, B i, Q ii, R ii, S ii ). This ensures that the A i in (9) are stable. Note that in these new coordinates, LQR(A i, B i, Q ii, R ii, S ii ) 0. The vector noise w may be decomposed into a sum involving scalar noise terms. To see why, taing the full singular value decomposition (SVD) of X. Namely, write X [ˆΣ 0 [U2 T U Ū Ū 2 () 0 0 where ˆΣ diag(σ,..., σ r ), and σ σ r > 0 are the singular values of X, and r : ran X. Note that the various matrices have sizes U i R ni r, Ūi R ni (ni r). The following lemma gives the required decomposition. Lemma 2. Suppose that (w, w 2 ) are jointly Gaussian random vectors with a covariance matrix given by (0), and the SVD of X is given by (). Define the scalar noises η i,..., ηr i, v i for i, 2 using the formulas ηi : e T Ui T w i for,..., r v i : Ū for i, 2 i T w i Then we have the properties { } η ) η r,...,, v, v 2 are independent η 2 η 2) cov η2 η r 2 [ σ σ 3) cov v I and cov v 2 I. for,..., r Furthermore, (w, w 2 ) may be decomposed as follows. w w 2 U e 0 η [Ū 0 v 0 U 2 e η2 + 0 Ū 2 v 2 Proof. Expand (w, w 2 ) using (0) () and the orthogonality of the U i Ū i matrices to obtain w i U i e (e T Ui T w i ) + Ūi(Ū i T w i ) which is precisely the desired decomposition. The three properties follow from direct evaluation of the appropriate covariances. We now apply Lemma 2 to separate the state equations (9) into independent parts. Since (x, u ) is a linear function of w, we can consider the components that depend on each of the terms in the decomposition. Namely, we can write [ x ζ ζ r u ϕ + + ϕ r + ζ ϕ where (ζ, ϕ ) is a linear function of (η, η 2 ), and ( ζ, ϕ) is a linear function of (v, v 2 ). Furthermore, the terms in 4

the above decomposition are independent. We therefore have the dynamics ζ Aζ + Bϕ U e + 0 η 0 U 2 e η2 for,..., r [Ū 0 v ζ A ζ + B ϕ + 0 Ū 2 v 2 Each such equation is a decentralized LQR problem of the type described in (3) (6). The cost functions decouple as well, so we may apply Theorem for each of the r + equations above. Thus, the optimal controller for the N 2 case with vector noise can be written as the linear combination u i K σ i (ˆx ) i where each K σ i is generated by the corresponding σ according to the definition in the statement of Theorem. There is no term corresponding to the ζ i component of the decomposition because the associated gain would be LQR(A i, B i, Q ii, R ii, S ii ), which is zero due to the normalization discussed in the beginning of this section. Recall that (ˆx ) i is the estimate of ζ given ζi, where both noise inputs are set to ηi. As discussed in Section 4, this estimate is an exact inference. Furthermore, measuring ζi is equivalent to measuring ηi, which is in turn equivalent to measuring w i, and therefore x i. Thus, (ˆx ) i E(ξ X i ), where ẋ i A i x i + B i u i + w i ξ Aξ + BK σ ξ U e + η U 2 e i Applying the decentralized estimation result of Section 3 to evaluate each of the expectations, we obtain q (A + BK σ )q L i B i u i + (AL i L i A i )x i for,..., r u i K σ i (q + L i x i ) where L i : U T U2 T T e e T U i T. This is precisely the result derived in [6. In general, each state q has dimension n + n 2, so the overall state dimension of the controller K i : x i u i is r(n + n 2 ). Therefore, the case N 2 with vector noise is nothing more than r separate instances of the case N 2 with scalar noise, and each such instance satisfies the separation principle of Theorem. As discussed in [6, the vector noise problem can be formulated in such a way that it has a unique solution, and it can be shown that the optimal controller with state dimension r(n + n 2 ) is generically minimal. That is, there is a choice of problem parameters such that the optimal controller has a minimal realization with this state dimension. This means that the interpretation of the states in terms of separate filters is necessary; one may need each of these r estimates to accurately represent the optimal controller. The case of N subsystems. The case N 3 with vector noises is substantially more complicated than the case N 2 described above. The reason is that the decomposition of Lemma 2 does not hold for more than two subsystems. For example, if N 3, the noise covariance has the general form W cov w I X 2 X 3 w 2 X 2 T I X 23 w 3 X3 T X23 T I If the X ij happen to be diagonal, then the problem splits into separate instances of the N 3 scalar-noise case. However, the technique used in Lemma 2 can only diagonalize one of the X ij matrices at a time, so a new type of decomposition will be required to address the case N 3. 6 Proof of Theorem We prove the result by first explicitly constructing a state-space realization of the optimal controller using an approach similar to the one used in [6, and then showing that the solution satisfies the required separation structure. We begin by converting the problem (3) (6) into model-matching form. Define F and G as A H A B F and G (2) C 0 C D where C and D are defined such that T Q S C D C D S T R (3) Now consider the following decentralized model-matching problem with one-sided dynamics. min FΣ/2 + G Q 0... HΣ /2 0 Q N (4) H2 s.t. Q,..., Q N RH 2 We then have the following result. Lemma 3. Suppose a solution to (4) is given by Âi ˆBi Q i for i,..., N (5) Ĉ i 0 A solution to the decentralized LQR problem described by (3) (6) is given by u i K i x i for i,..., N, where [ Âi ˆB i B i Ĉ i (Âi ˆB i B i Ĉ i ) ˆB i ˆB i A i K i Ĉ i Ĉ i ˆBi (6) Proof. The LQR problem described by (3) (6) is quadratically invariant [, and the associated plant is 5

stable. Therefore, the set of stabilizing controllers is parameterized by K Q(I + (si A) B Q), where Q RH and Q is bloc-diagonal [7. The closedloop map e z where e is IID Gaussian with identity covariance and z Cx + Du is the cost functional is FΣ /2 + GK(I (si A) BK) (si A) HΣ /2 FΣ /2 + G Q(sI A) HΣ /2 Let Q : Q(sI A). Whenever Q RH, we have Q RH 2. Similarly, if Ā B Q RH 2 then C 0 [ Ā Ā Q B BA RH C C B Furthermore, Q is bloc-diagonal if and only if Q is blocdiagonal. Therefore, the set of closed-loop maps may be equivalently parameterized by FΣ /2 +GQHΣ /2, where Q RH 2 and is bloc-diagonal. Inverting the maps Q Q K yields the desired result. We will therefore concentrate our efforts on solving the decentralized model-matching problem (4). Viewed as an optimization problem in the Hilbert space H 2, the optimality condition for (4) is given by [9 G FΣH T + G GQHΣH T H 2 L 2... L 2 H2 Where Q : diag(q,..., Q N ). Extract the N blocdiagonal entries, and let E i be the (n + + n N ) n i matrix such that I E... E N. Similarly, let ei be the N n matrix such that I e... e N. Using the fact that H is bloc-diagonal, we obtain E T i G FΣe i + E T i G GQHΣe i H 2 Expanding, and maing use of the definition (7) Ei T G FΣe i + Ei T G G Q h σ i. H2 Q N h N σ Ni Commuting the scalars σ ji, Ei T G FΣe i + Ei T G G σ ii... h. H2 σnii Q N h N Where the identity matrices are sized according to the corresponding Q i blocs. Stacing these equations for i,..., N, we obtain ( ) ( ) (G F) Σ + (G G) Σ Q h. H2 (7) Q N h N where is the blocwise Hadamard product defined in the theorem statement, is the vector of all-ones, and we made use of the fact that Σ Σ T. This is a centralized LQR problem in model-matching form! It remains to find a joint realization for the blocs. Specifically, we would lie to find some F and Ḡ such that [( ) Ḡ F Ḡ (G F) Σ (G G) Σ (8) This will ensure that (7) is the optimality condition for the LQR problem determined by F and Ḡ. Substituting the definitions (2) (3), we obtain A T Q 0 G F 0 A H and B T S T 0 A T Q S G G 0 A B B T S T R Maing use of the fact that A, B, and H are blocdiagonal, we find after some algebra that A T Q Σ 0 (G F) Σ 0 A H and B T S T Σ 0 A T Q Σ S Σ (G G) Σ 0 A B B T S T Σ R Now, define C and D such that [ T [ Q Σ S Σ C D C D S T Σ R Σ (9) This new cost matrix satisfies the required definiteness constraints, see for example [3. It is straightforward to chec that A h A B F and Ḡ C 0 C D satisfy (8), where h : H is the vector obtained by stacing h,..., h N. The solution to this centralized model-matching problem is therefore Q h A + BK σ h. K σ (20) 0 Q N h N with K σ defined in the theorem statement. Extracting the i th bloc-row, one possible solution is given by [ A + BK σ h h i Q i Ei TKσ 0 Where h i (ht i h i) h T i. Note that this solution is not unique, and there are many possible choices of Q i that 6

satisfy (20). Now apply Lemma 3 to obtain a realization for K i. The state-space equations are q i (A + BK σ hh i B ie T i K σ )q i + ( (A + BK σ hh i B ie T i K σ )hh i hh i A i) xi u i E T i K σ q i + E T i K σ hh i x i Define L : hh i and ξi : q i + Lx i. The equations then simplify to q i Aq i + (AL LA i )x i + BK σ ξ i LB i u i ξ i q i + Lx i u i E T i K σ ξ i Upon comparing these equations with the decentralized estimation result from Section 3, we conclude that ξ i E(z i, X i ), where z i satisfies ż i Az i + Bu i + v i u i K σ z i (2) and v i has noise statistics that satisfy L E(h i η 2 i h T i ) E(v i η i h T i ) Substituting the definition L : hh i and using the fact that E(ηi 2), we deduce that vi hη i. Consequently, (2) becomes (8), as required. This completes the proof of Theorem. 7 Summary In this paper, we showed that the N-subsystem fully decentralized LQR problem with coupled cost and correlated scalar noise satisfies a new separation principle reminiscent of the classical separation principle for LQG control. We also showed that the N 2 case with correlated vector noise solved in [6 can be reduced to separate instances of the scalar-noise case. The estimation and separation structure of the general N-subsystem case with correlated vector noise remains an open problem, as our decomposition approach used for N 2 does not appear to generalize to N 3. References [ T. Georgiou and A. Lindquist. The separation principle in stochastic control, redux. IEEE Transactions on Automatic Control, 58(0):248 2494, 203. [2 Y.-C. Ho and K.-C. Chu. Team decision theory and information structures in optimal control problems Part I. IEEE Transactions on Automatic Control, 7():5 22, 972. [3 A. Lampersi and J. C. Doyle. Dynamic programming solutions for decentralized state-feedbac LQG problems with output feedbac. In American Control Conference, pages 6322 6327, 202. [4 A. Lampersi and L. Lessard. Optimal statefeedbac control under sparsity and delay constraints. In 3rd IFAC Worshop on Distributed Estimation and Control in Networed Systems, pages 204 209, 202. [5 L. Lessard. Decentralized LQG control of systems with a broadcast architecture. In IEEE Conference on Decision and Control, pages 624 6246, 202. [6 L. Lessard. Optimal control of a fully decentralized quadratic regulator. In Allerton Conference on Communication, Control, and Computing, pages 48 54, 202. [7 L. Lessard and S. Lall. A state-space solution to the two-player decentralized optimal control problem. In Allerton Conference on Communication, Control, and Computing, pages 559 564. IEEE, 20. [8 D. Luenberger. An introduction to observers. IEEE Transactions on Automatic Control, 6(6):596 602, 97. [9 D. Luenberger. Optimization by vector space methods. Wiley-Interscience, 997. [0 A. Nayyar, A. Mahajan, and D. Teneetzis. Decentralized stochastic control with partial history sharing: A common information approach. IEEE Transactions on Automatic Control, 58(7):644 658, 203. [ M. Rotowitz and S. Lall. A characterization of convex problems in decentralized control. IEEE Transactions on Automatic Control, 5(2):274 286, 2006. [2 P. Shah and P. A. Parrilo. An optimal architecture for poset-causal systems. In IEEE Conference on Decision and Control, pages 5522 5528, 20. [3 B. Wang and F. Zhang. Schur complements and matrix inequalities of hadamard products. Linear and Multilinear Algebra, 43(-3):35 326, 997. [4 H. S. Witsenhausen. A counterexample in stochastic optimum control. SIAM Journal on Control, 6():3 47, 968. [5 H. S. Witsenhausen. Separation of estimation and control for discrete time systems. Proceedings of the IEEE, 59():557 566, 97. [6 W. Wonham. On the separation theorem of stochastic control. SIAM Journal on Control, 6(2):32 326, 968. [7 K. Zhou, J. Doyle, and K. Glover. Robust and optimal control. Prentice-Hall, 995. 7