Time Varying Optimal Control with Packet Losses.

Similar documents
LQG CONTROL WITH MISSING OBSERVATION AND CONTROL PACKETS. Bruno Sinopoli, Luca Schenato, Massimo Franceschetti, Kameshwar Poolla, Shankar Sastry

Kalman Filtering with Intermittent Observations*

Foundations of Control and Estimation over Lossy Networks

IEEE Copyright Statement:

Kalman filtering with intermittent heavy tailed observations

Networked Control Systems: Estimation and Control over Lossy Networks

Via Gradenigo 6/b Department of Information Engineering University of Padova Padova, Italy

Kalman Filtering with Intermittent Observations

Lecture 5: Control Over Lossy Networks

LQG Control For MIMO Systems Over multiple TCP-like Erasure Channels

Estimation over Communication Networks: Performance Bounds and Achievability Results

Feedback Control over Packet Dropping Network Links

Optimal control and estimation

Time and Event-based Sensor Scheduling for Networks with Limited Communication Resources

Here represents the impulse (or delta) function. is an diagonal matrix of intensities, and is an diagonal matrix of intensities.

Networked Sensing, Estimation and Control Systems

Estimation for Nonlinear Dynamical Systems over Packet-Dropping Networks

Optimal LQG Control Across a Packet-Dropping Link

On Separation Principle for a Class of Networked Control Systems

Lecture 10 Linear Quadratic Stochastic Control with Partial State Observation

On Design of Reduced-Order H Filters for Discrete-Time Systems from Incomplete Measurements

A Stochastic Online Sensor Scheduler for Remote State Estimation with Time-out Condition

On the Optimality of Threshold Policies in Event Triggered Estimation with Packet Drops

State estimation of linear dynamic system with unknown input and uncertain observation using dynamic programming

4 Derivations of the Discrete-Time Kalman Filter

Lecture 3: Functions of Symmetric Matrices

False Data Injection Attacks in Control Systems

On Kalman filtering for detectable systems with intermittent observations

Riccati difference equations to non linear extended Kalman filter constraints

Power Spectral Analysis of Networked Control Systems with Data Dropouts

The Uncertainty Threshold Principle: Some Fundamental Limitations of Optimal Decision Making under Dynamic Uncertainty

Event-based State Estimation of Linear Dynamical Systems: Communication Rate Analysis

Discrete-Time H Gaussian Filter

Approximate Estimation of Distributed Networked Control Systems

STOCHASTIC STABILITY OF EXTENDED FILTERING FOR NONLINEAR SYSTEMS WITH MEASUREMENT PACKET LOSSES

State Estimation Utilizing Multiple Description Coding over Lossy Networks

Control Capacity. Gireeja Ranade and Anant Sahai UC Berkeley EECS

Optimal estimation in networked control systems subject to random delay and packet drop

Decentralized LQG Control of Systems with a Broadcast Architecture

Linear-quadratic-Gaussian control of linear system with Rayleigh flat fading channel

Estimation and Control across Analog Erasure Channels

Linear-Quadratic Optimal Control: Full-State Feedback

RECURSIVE ESTIMATION AND KALMAN FILTERING

Data Rate Theorem for Stabilization over Time-Varying Feedback Channels

FIR Filters for Stationary State Space Signal Models

Towards control over fading channels

CDS 270-2: Lecture 6-1 Towards a Packet-based Control Theory

Structured State Space Realizations for SLS Distributed Controllers

Control over finite capacity channels: the role of data loss, delay and signal-to-noise limitations

Energy Efficient Spectrum Sensing for State Estimation over A Wireless Channel

Kalman Filtering for networked control systems with random delay and packet loss

EE C128 / ME C134 Feedback Control Systems

Adaptive linear quadratic control using policy. iteration. Steven J. Bradtke. University of Massachusetts.

Kalman Filtering with Intermittent Observations: Tail Distribution and Critical Value

Optimal sensor fusion for distributed sensors subject to random delay and packet loss

arxiv: v1 [cs.sy] 22 Jan 2015

Consensus Control of Multi-agent Systems with Optimal Performance

Optimal estimation in networked control systems subject to random delay and packet drop

The Kalman filter is arguably one of the most notable algorithms

A Study of Covariances within Basic and Extended Kalman Filters

MUCH attention has been recently paid to networked control

Optimal estimation in networked control systems subject to random delay and packet loss

Markov decision processes with threshold-based piecewise-linear optimal policies

Networked Control Systems with Packet Delays and Losses

arxiv: v1 [cs.sy] 15 Oct 2017

Secure Control Against Replay Attacks

DESIGNING A KALMAN FILTER WHEN NO NOISE COVARIANCE INFORMATION IS AVAILABLE. Robert Bos,1 Xavier Bombois Paul M. J. Van den Hof

Alternative Characterization of Ergodicity for Doubly Stochastic Chains

On the Resource Allocation Problem in Wireless Networked Control Systems

Optimal Polynomial Control for Discrete-Time Systems


Lyapunov Stability of Linear Predictor Feedback for Distributed Input Delays

An Adaptive LQG Combined With the MRAS Based LFFC for Motion Control Systems

9 Multi-Model State Estimation

CALIFORNIA INSTITUTE OF TECHNOLOGY Control and Dynamical Systems. CDS 110b

ESTIMATOR STABILITY ANALYSIS IN SLAM. Teresa Vidal-Calleja, Juan Andrade-Cetto, Alberto Sanfeliu

Lecture 5 Linear Quadratic Stochastic Control

Output-based Event-triggered Control with Performance Guarantees

Lecture Outline. Target Tracking: Lecture 3 Maneuvering Target Tracking Issues. Maneuver Illustration. Maneuver Illustration. Maneuver Detection

State estimation and the Kalman filter

REAL-TIME STATE ESTIMATION OF LINEAR PROCESSES OVER A SHARED AWGN CHANNEL WITHOUT FEEDBACK

Outline. 1 Linear Quadratic Problem. 2 Constraints. 3 Dynamic Programming Solution. 4 The Infinite Horizon LQ Problem.

Steady State Kalman Filter

TWO SCHEMES OF DATA DROPOUT COMPENSATION FOR LQG CONTROL OF NETWORKED CONTROL SYSTEMS

EL2520 Control Theory and Practice

Adaptive Dual Control

RECENT advances in technology have led to increased activity

Level-triggered Control of a Scalar Linear System

Iterative Encoder-Controller Design for Feedback Control Over Noisy Channels

Chapter 3. LQ, LQG and Control System Design. Dutch Institute of Systems and Control

Kalman Filters with Uncompensated Biases

MOST control systems are designed under the assumption

IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 54, NO. 2, FEBRUARY

Optimal State Estimators for Linear Systems with Unknown Inputs

Optimal Decentralized Control of Coupled Subsystems With Control Sharing

Rank-one LMIs and Lyapunov's Inequality. Gjerrit Meinsma 4. Abstract. We describe a new proof of the well-known Lyapunov's matrix inequality about

ASIGNIFICANT research effort has been devoted to the. Optimal State Estimation for Stochastic Systems: An Information Theoretic Approach

Optimal mean-square performance for networked control systems with unreliable communication

Kalman Filtering with Intermittent Observations: Tail Distribution and Critical Value

Stability Conditions and Phase Transition for Kalman Filtering over Markovian Channels

Transcription:

Time Varying Optimal Control with Packet Losses. Bruno Sinopoli, Luca Schenato, Massimo Franceschetti, Kameshwar Poolla, Shankar S. Sastry Department of Electrical Engineering and Computer Sciences University of California at Berkeley {sinopoli,lusche,massimof,sastry}@eecs.berkeley.edu poolla@me.berkeley.edu Abstract The problem of using wireless sensor networks technology for estimation and control of dynamical systems has recently received widespread attention within the scientific community. Classical control theory is in general insufficient to model distributed control problems where issues of communication delay, jitter, and time synchronization between components cannot be ignored. The purpose of this paper is to extend our work on discrete time Kalman filtering with intermittent observations [] that was motivated by data losses in a communication channel. Accordingly, we consider the Linear Gaussian Quadratic (LQG optimal control problem in the discrete time setting, showing that the separation principle holds in the presence of data losses. Then, using our previous results, we show the existence of a critical arrival probability below which the resulting optimal controller fails to stabilize the system. This is done by providing analytic upper and lower bounds on the cost functional, and stochastically characterizing their convergence properties as k. I. INTRODUCTION Advances in VLSI and MEMS technology have boosted the development of micro sensor integrated systems. Such systems combine computing, storage, radio technology, and energy source on a single chip [2] [3]. When distributed over a wide area, networks of sensors can perform a variety of tasks that range from environmental monitoring and military surveillance, to navigation and control of a moving vehicle [4] [5] [6]. A common feature of these systems is the presence of significant communication delays and data loss across the network. From the point of view of control theory, significant delay is equivalent to loss, as data needs to arrive to its destination in time to be used for control. In short, communication and control become tightly coupled such that the two issues cannot be addressed independently. Consider, for example, the problem of navigating a vehicle based on the estimate from a sensor web of its current position and velocity. The measurements underlying this estimate can be lost or delayed due to the unreliability of the wireless links. What is the amount of data loss that the control loop can tolerate in order to reliably perform the navigation task? Can communication protocols be designed to satisfy this constraint? Practical advances in the design of these systems are described in [7]. The goal of this paper is to examine some control-theoretic implications of using sensor networks for control. These require a generalization of classical control techniques that explicitly take into account the stochastic nature of the communication channel. In our setting, the sensor network provides observed data that are used to estimate the state of a controlled system, and this estimate is then used for control. However, due to unreliability of the network links, data communicated from the sensors to the estimator can be lost and we want to characterize the impact of such event on the control performance. Accordingly, we generalize the Linear Quadratic Gaussian (LQG optimal control problem modeling the arrival of an observation as a random process whose parameters are related to the characteristics of the communication channel, as shown in Figure. The separation principle states that observer and plant of a linear system can be designed independently. We show that this principle continues to hold in the case of data loss between the sensor and the estimator. This allows to extend our results in [], [8] showing the existence of a critical loss probability below which a transition occurs and the resulting optimal controller fails to stabilize the system. We show the existence of such transition by finding deterministic upper and lower bounds for the expected optimal cost and their convergence conditions. In some related work [9] Nilsson presents the LQG optimal regulator with bounded delays between sensors and controller, and between the controller and the actuator, but he does not address the packet-loss case. This is considered by Hadijcostis and Touri [0]. Their analysis is restricted to the static scalar case. Other approaches include using the last received sample for control, or designing a dropout compensator [], [2]. We consider the alternative approach where the external compensator feeding the controller is the optimal time varying Kalman gain. Moreover, we analyze the proposed solution in state space domain rather than in frequency domain as it was presented in [2], and we consider the more general Multiple Input Multiple Output (MIMO case. The LQG optimal control problem with missing observations can also be modelled using the well known Jump Linear System (JLS theory [3], where the observer switches between open loop and closed loop configuration, depending on whether the packet containing the observation is lost, or arrives at the estimator in time. However, convergence results in this case can be obtained only when

Fig.. Overview of the system. We study the statistical convergence of the expected state covariance of the discrete time LQG, where the observation, travelling over an unreliable communication channel, can be lost at each time step with probability γ. each jump sub-system is stabilizable and detectable. The detectability assumption fails in our case, producing a nonstationary state random process. Finally, we mention that philosophically our result can be seen as another manifestation of the well known uncertainty threshold principle [4], [5]. This principle states that optimum long-range control of a dynamical system with uncertainty parameters is possible if and only if the uncertainty does not exceed a given threshold. The uncertainty is modelled as white noise scalar sequences acting on the system and control matrices. In our case the uncertainty is due to the random arrival of the observation, with the randomness arising from losses in the network. The paper is organized as follows. In section II we formulate the LQG optimal control problem with intermittent observations. In section III we prove some lemmas we need to prove the main theorems. In Section IV we compute the optimal controller, showing that it is linear. We also provide upper and lower bounds on the cost functional of the LQG problem, and find the conditions on the observation arrival probability γ for which the upper bound converges to a fixed point, and for which the lower bound diverges. Finally, in section V, we state our conclusions and give directions for future work. II. PROBLEM FORMULATION Consider the following linear stochastic system with intermittent observations: x k+ = Ax k + B + w k ( y k = Cx k + v k, (2 where x k R n is the state vector, y k R m is the output vector, R q is the input vector, x 0 R n, w k R n and v k R m are Gaussian, uncorrelated, white, with zero mean and covariance (P 0, Q, R k respectively, R k = γ k R + ( γ k σ 2 I, and γ k are i.i.d. Bernoulli random variable with P (γ k = = γ. Let us define the following information set: I k = {y k, γ k }, (3 where y k = (y k, y k,..., y, γ k = (γ k, γ k,..., γ. Consider also the following cost function: [ ] N J N (u N = E x N W N x N + (x kw k x k + u ku k IN (4 We now look for control input sequence u N that minimizes the above functional given that the information I k is available at time k, i.e. J N u N J N (u N = J N (u N (5 where u k = u k (I k and I k is defined in Equation 3. III. MATHEMATICAL BACKGROUND Before proceeding, let us define the following variables: ˆx k k = E[xk I k ], e k k = xk ˆx k k, P k k = E[ek k e k k I k]. In the following derivation we will make use of the following facts Lemma. The following facts are true: (a E [(x k ˆx k ˆx k I k] = E [ e k k ˆx k I k] = 0 (b E [x k Sx k I k ] = ˆx k Sˆx k + trace ( SP k k, (c E [E[ g(x k+ I k+ ] I k ] = E [g(x k+ I k ], (6 S g( Proof: (a It follows directly from the definition. In fact: E [(x k ˆx k ˆx k I k] = E [x k ˆx k ˆx k ˆx k I k] = E [x k I k ] ˆx k ˆx k ˆx k (b Using standard algebraic = 0 operations and the previous fact we have: E [ x ] ksx k I k = E [ (x k ˆx k + ˆx k ] S(x k ˆx k + ˆx k I k = ˆx ksˆx k + E [ (x k ˆx k S(x k ˆx k ] + + 2E [ˆx ks(x k ˆx k I k ] = ˆx ksˆx k + 2trace{SE[(x k ˆx k ˆx k I k ]} + + trace{se[(x k ˆx k (x k ˆx k I k ]} = ˆx ksˆx k + trace{sp k k } (c Let (X, Y, Z be any random vectors, g( any function, and p the probability distribution, then E Y,Z [g(x, Y, Z X] = = g(x, Y, Zp(Y, Z XdY dz Z Y = g(x, Y, Zp(Y Z, Xp(Z XdY dz Z Y [ ] = g(x, Y, Zp(Y Z, XdY p(z XdZ Z Y = E Z [ E Y [g(x, Y, Z Z, X] X].

where we used the Bayes Rule. Since by hypothesis I k I k+, then fact (c follows from the above equality by substituting I k = X and I k+ = (X, Z. IV. FINITE AND INFINITE HORIZON LQG We first start finding the optimal estimator, which will be needed to solve the LQG controller design, as it will be shown later. A. Estimator Design, σ + We derive the equations for optimal estimator using similar arguments used for the standard Kalman filtering equations. The innovation step is given by: ˆx k+ k = E[x k+ I k ] = E[Ax k + B + w k I k ] = AE[x k I k ] + B = Aˆx k k + B (7 e k+ k = x k+ ˆx k+ k (8 P k+ k = E[e k+ k e k+ k I k ] = [ (A(xk ( ] = E ˆx k k + w k A(xk ˆx k k + w k I k = AE[e k k e k k I k ]A + E[w k w k] = AP k k A + Q (9 where we used the independence of w k and I k. Since y k+, γ k+, w k and I k are all independent of each other and following the same approach described in [], then correction step is given by: ˆx k+ k+ = ˆx k+ k + γ k+ K k+ (y k+ C ˆx k+ k (0 P k+ k+ = P k+ k γ k+ K k+ CP k+ k, ( K k+ = P k+ k C (CP k+ k C + R, (2 where we took the limit σ +. The initial conditions for the estimator iterative equations are: B. Estimator properties ˆx 0 = 0 (3 P 0 = P 0 (4 Here we compute some quantities that will prove useful when deriving the equation for the optimal LQG controller. Consider first the following variable: k+ k = E[ˆxk+ k ˆx k+ k I k] = = E[(Aˆx k k + B (Aˆx k k + B I k ] = E[Aˆx k k ˆx k k A + 2Aˆx k k u kb + B u kb I k ] = Aˆx k k ˆx k k A + 2Aˆx k k u kb + B u kb (5 where we used the fact that ˆx k k is independent of I k. Equation (0 can be rewritten as follows: ˆx k+ k+ = = ˆx k+ k + γ k+ K k+ ( C(xk+ ˆx k+ k + v k+ = ˆx k+ k + γ k+ K k+ ( Cek+ k + v k+ (6 The first and the second terms in the previous equation and v k+ are independent of each other by Lemma (a and have zero mean. Moreover, the random variable γ k+ is independent of the previous terms. Therefore, we have: k+ k+ = E[ˆx k+ k+ˆx k+ k+ I k ] = = E[ˆx k+ k ˆx k+ k I k ] + + E[γ 2 k+k k+ (Ce k+ k e k+ kc + + v k+ v k+ K k+ I k ] = = k+ k + E[γ 2 k+k k+ (Ce k+ k e k+ kc + + v k+ v k+ K k+ I k ] = = k+ k + + E[γ 2 k+k k+ (CP k+ k C + RK k+ I k ] = = k+ k + + E[γ 2 k+p k+ k C (CP k+ k C + R CP k+ k I k ] = = k+ k + E[M k+ I k ] (7 M k+ = γ 2 k+p k+ k C (CP k+ k C + R CP k+ k (8 where M k+ is a stochastic matrix which depends on the sequence {γ k }, but is independent of. C. Controller design To derive the optimal feedback control law and the corresponding value for the objective function we will follow the dynamic programming approach based on the cost-to-go iterative procedure. Define the optimal value function V k (x k as follows: V N (x N = E[x NW N x N I N ] V k (x k E[x kw k x k + u ku k + V k+ (x k+ I k ] Using dynamic programming theory [6] [7], one can show that J N = V 0(x 0. We claim that the value function V k (x k can be written as: V k (x k = ˆx k k S k ˆx k k + c k, k = 0,..., N (9 where the matrix S k and the scalar c k are to be determined and are independent of the information set I. We will prove it by induction. The claim is certainly true for k = N, in fact: V N (x N = E[x N W N x N I N ] = ˆx N N W Nx N N + trace(w N P N N (20 where we used Lemma (b, with the following choice of parameters: S N = W N (2 c N = trace(w N P N N (22 Suppose now that the claim is true for k +, i.e. V k+ (x k+ = ˆx k+ k+ S k+ˆx k+ k+ + c k+, and we use it to compute the value function at time step k as follows:

V k (x k = E[x kw k x k + u ku k + V k+ (x k+ I k ] E[x kw k x k + u ku k + + E[x k+ k+s k+ x k+ k+ + c k+ I k+ ] I k ] E[x kw k x k + u ku k + + x k+ k+s k+ x k+ k+ + c k+ I k ] (ˆx k kw k ˆx k k + trace(w k P k k + u ku k + + trace(s k+ E[x k+ k+ x k+ k+ I k ] + E[c k+ I k ] (u ku k + trace(s k+ k+ k + + trace(s k+ M k+ + ˆx k kw k ˆx k k + + trace(w k P k k + E[c k+ I k ] (u ku k + u kb S k+ B + 2u kb S k+ Aˆx k k + + ˆx k ka S k+ Aˆx k k + trace(s k+ M k+ + + ˆx k kw k ˆx k k + trace(w k P k k + E[c k+ I k ] (23 where we used Lemma (c in the third line and Lemma (b in the fourth line. We substituted Equations (7 in fifth line and we separated terms which did not depend on. Finally, in sixth line we substituted Equations (5 and we used the property trace(ab = trace(ba. The value function is a convex quadratic function of the input, therefore the minimizer can be simply obtained by solving V k = 0, which gives: = (B S k+ B + U k B S k+ A ˆx k k = L k ˆx k k. (24 The optimal feedback is therefore a simple linear feedback of the estimated state. If we substitute the minimizer back into Equation (23 and we use Equation (9 we get: and V k (x k = = ˆx k ka S k+ B(B S k+ B + U k B S k+ A ˆx k k + + ˆx k ka S k+ Aˆx k k + ˆx k kw k ˆx k k + + trace(s k+ M k + trace(w k P k k + E[c k+ I k ], ˆx k ks k ˆx k k + c k = ˆx k k(a S k+ A + W k + A S k+ B(B S k+ B + U k B S k+ Aˆx k k + + trace(s k+ M k+ + trace(w k P k k + E[c k+ I k ] For the previous equation to hold for ˆx k k, we need to have: S k = A S k+ A + W k + A S k+ B(B S k+ B + U k B S k+ A (25 c k = E[c k+ I k ] + trace(s k+ M k+ + trace(w k P k k Therefore, the cost function for the optimal LQG is given by: JN = V 0 (x 0 = ˆx 0 0S 0ˆx 0 0 + N ( + trace(sk+ E γ[m k+ ] + trace(w k E γ[p k k ](26 where the first term is zero since by hypothesis ˆx 0 = 0. The matrices M k+ and P k k in the second and third term are stochastic since they are function of the sequence {γ k }, where the expectation with respect to this process is made explicit in E γ. The exact expected value of these matrices cannot be computed analytically, as shown in [], but it can be bounded by computable quantities. In fact let us consider the following three equations: P k+ k = A P k k A + Q + γa P k k C (C P k k C + R C P k k A P k+ k+ = P k k + γ P k k C (C P k k C + R C P k k P k+ k = ( γa P k k A + Q initialized to P 0 = P 0 = P 0. It is possible to show that: P k k E γ [P k k ] P k k (27 0 E γ [M k+ ] Therefore we have: γ P k k C (C P k k C + R C P k k = P k k P k+ k+ (28 JN min JN JN max (29 N JN max = (trace(s k+ ( P k k P k+ k+ + + trace(w k Pk k (30 N JN min = trace(w k Pk k (3 D. Finite and Infinite Horizon LQG control The previous equations were derived for the finite horizon LQG. The infinite horizon LQG can be obtained by taking the limit for N + of the previous equations. However, the matrices {M k+ } and {P k k } depend on the specific realization of the observation sequence {γ k }, therefore the minimal cost J N is a stochastic function and does not have a limit. Differently from standard LQG controller design where the controller always stabilizes the original system, in the case of control with packet losses, the stability can be lost if the arrival probability γ is below a certain threshold. In particular the Equation for the cost matrix S k is the solution of a modified Riccati Algebraic Equation (MARE which was already introduced and studied in our previous work []. In particular, Equation (25 is the dual of the estimator equation presented in []. Therefore, the same conclusions can be drawn and we are now ready summarize the previous result in the following theorem: Theorem (Finite Horizon LQG. Consider the system (-(2 and consider the problem of minimizing the cost function (4 with policy = f(i k, where I k is the information available under the communication model, as defined in Equation (3. Then, the optimal control is a linear

function of the estimated system state given by Equation (24, where the matrix S k can be computed iteratively using Equation (25. The separation principle still holds under our assumptions, since the optimal estimator is independent of the control input. The optimal state estimator is given by Equations (7-(0 and (9-(2, and the minimal achievable cost is given given by Equation (26. Theorem 2 (Infinite Horizon LQG. Consider the same systems as defined in the previous theorem with the following additional hypothesis: W N = W k = W and U k = U. Moreover, let (A, B and (A, Q 2 be controllable, and let (A, C and (A, W 2 be observable. Let us consider the limiting case N +, then, there exists a critical arrival probability γ min which satisfies the following property: min (, λ max (A 2 γ min (32 where λ max (A is the eigenvalue of matrix A with largest absolute value, such that for all γ > γ min we have: L k = L = (B S B + U B S A (33 N J N min N J N N J N max (34 where the matrix S, and the mean cost bounds JN min, J N max are given by: lim N + N J N max = J max = = γtrace((s W P C (CP C + R CP + + trace(w P (35 N J N min = J min = trace(w P, (36 lim N + and the matrices S,P,P are given by: S = A S A + W + A S B(B S B + U B S A (37 P = AP A + Q + γ AP C (CP C + R CP A (38 P = ( γap A + Q, (39 where S is the solution of a standard ARE which always exists [8], P is the solution of a MARE dual to the one proposed in [], and P is the solution of a Lyapunov Equation. Moreover, the assumptions above are necessary and sufficient conditions for the closed loop system under LQG feedback to be stable in mean sense. The critical probability γ min can be computed via the solution of the following LMIs optimization problem: γ min = argmin γ Ψ γ (Y, Z > 0, 0 Y I. Ψ γ (Y, Z = Y γ(y A + ZC γy A = γ(a Y + C Z Y 0 γa Y 0 Y V. CONCLUSION Motivated by applications where control is performed over a communication network, in this paper we extend our previous results on estimation with intermittent observations to the optimal control problem. First, we show that the separation principle holds also in the case when the observed state can be lost at each time step with some probability λ. Then, we show how the optimal control problem formally reduces to the solution of a standard Riccati equation for the controller and the same modified Riccati equation that was studied in [], [8] for the estimator. Accordingly, we provide upper and lower bounds on the expected optimal cost functional and characterize its convergence conditions, showing a transition to an unbounded cost beyond a critical arrival probability. We also provide upper and lower bounds for the cost in the finite horizon case. Future work will include stability analysis and LQG controller design for a more general class of systems, specifically the ones where also loss of control packets occurs between the controller and one or more actuators. REFERENCES [] B. Sinopoli, L. Schenato, M. Franceschetti, K. Poolla, M. Jordan, and S. Sastry, Kalman filtering with intermittent observations, IEEE Transactions on Automatic Control, September 2004, to appear. [2] Smart dust project home page. http://robotics.eecs.berkeley.edu/ pister/smartdust/. [3] NEST project at Berkeley home page. http://webs.cs.berkeley.edu/nest-index.html. [4] Seismic sensor research at berkeley, home page. http://www.berkeley.edu/news/media/releases /200/2/3 snsor.html. [5] P. Varaiya, Smart cars on smart roads: Problems of control, IEEE Transactions on Automatic Control, vol. 38(2, February 993. [6] J. Lygeros, D. N. Godbole, and S. S. Sastry, Verified hybrid controllers for automated vehicles, IEEE Transactions on Automatic Control, vol. 43(4, 998. [7] B. Sinopoli, C. Sharp, S. Schaffert, L. Schenato, and S. Sastry, Distributed control applications within sensor networks, IEEE Proceedings Special Issue on Distributed Sensor Networks, November 2003. [8] B. Sinopoli, L. Schenato, M. Franceschetti, K. Poolla, M. Jordan, and S. Sastry, Kalman filtering with intermittent observations, in IEEE Conference on Decision and Control, December 2003. [9] J. Nilsson, Real-time control systems with delays, Ph.D. dissertation, Department of Automatic Control, Lund Institute of Technology, 998. [0] C. N. Hadjicostis and R. Touri, Feedback control utilizing packet dropping network links, in Proceedings of the 4st IEEE Conference on Decision and Control, Las Vegas, NV, Dec 2002, invited. [] Q. Ling and M. Lemmon, Soft real-time scheduling of networked control systems with dropouts governed by a markov chain, in American Control Conference, June 2003, denver, CO. [2], Optimal dropout compensation in networked control systems, in IEEE conference on decision and control, December 2003, maui, HI. [3] M. Mariton, Jump Linear Systems in Automatic Control. Marcel Dekker, 990. [4] M. Athans, R. Ku, and S. B. Gershwin, The uncertainty threshold principle, some fundamental limitations of optimal decision making under dynamic uncertainty, IEEE Transactions on Automatic Control, vol. 22(3, pp. 49 495, June 977. [5] R. Ku and M. Athans, Further results on the uncertainty threshold principle, IEEE Transactions on Automatic Control, vol. 22(5, pp. 49 495, October 977.

[6] P. Kumar and P. Varaiya, Stochastic Systems: Estimation, Identification and Adaptive Control, ser. Information and System Science Series, T. Kailath, Ed. Englewood Cliffs, NJ 07632: Prentice Hall, 986. [7] D. Bertsekas and J. Tsitsiklis, Neuro-Dynamic Programming. Athena Scientific, September 996. [8] P. S. Maybeck, Stochastic models, estimation, and control, ser. Mathematics in Science and Engineering, 979, vol. 4.