Towards control over fading channels

Similar documents
Data rate theorem for stabilization over fading channels

Data Rate Theorem for Stabilization over Time-Varying Feedback Channels

IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 54, NO. 2, FEBRUARY

Stabilization Over Markov Feedback Channels: The general case.

Chapter 1 Elements of Information Theory for Networked Control Systems

Capacity-achieving Feedback Scheme for Flat Fading Channels with Channel State Information

Estimating a linear process using phone calls

Feedback Stabilization of Uncertain Systems Using a Stochastic Digital Link

Encoder Decoder Design for Feedback Control over the Binary Symmetric Channel

Encoder Decoder Design for Event-Triggered Feedback Control over Bandlimited Channels

Communication constraints and latency in Networked Control Systems

Universal Anytime Codes: An approach to uncertain channels in control

RECENT advances in technology have led to increased activity

On the Effect of Quantization on Performance at High Rates

ELEC546 Review of Information Theory

Control Over Noisy Channels

Approximately achieving the feedback interference channel capacity with point-to-point codes

Encoder Decoder Design for Event-Triggered Feedback Control over Bandlimited Channels

Lecture 4 Noisy Channel Coding

A digital interface for Gaussian relay and interference networks: Lifting codes from the discrete superposition model

Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information

Iterative Encoder-Controller Design for Feedback Control Over Noisy Channels

Random Access: An Information-Theoretic Perspective

Characterization of Information Channels for Asymptotic Mean Stationarity and Stochastic Stability of Nonstationary/Unstable Linear Systems

Time-triggering versus event-triggering control over communication channels

Upper Bounds on the Capacity of Binary Intermittent Communication

MARKOV CHAINS A finite state Markov chain is a sequence of discrete cv s from a finite alphabet where is a pmf on and for

The Capacity Region of the Gaussian Cognitive Radio Channels at High SNR

Appendix B Information theory from first principles

Shannon s A Mathematical Theory of Communication

Lecture 5 Channel Coding over Continuous Channels

Secure Degrees of Freedom of the MIMO Multiple Access Wiretap Channel

Chapter 8: Differential entropy. University of Illinois at Chicago ECE 534, Natasha Devroye

Lecture 8: Channel and source-channel coding theorems; BEC & linear codes. 1 Intuitive justification for upper bound on channel capacity

Stabilization over discrete memoryless and wideband channels using nearly memoryless observations

Feedback Capacity of a Class of Symmetric Finite-State Markov Channels

Stochastic Stability and Ergodicity of Linear Gaussian Systems Controlled over Discrete Channels

2. PROBLEM SETUP. <. The system gains {A t } t 0

Lecture 7 MIMO Communica2ons

Lecture 11: Quantum Information III - Source Coding

Lecture 6 I. CHANNEL CODING. X n (m) P Y X

EE/Stats 376A: Homework 7 Solutions Due on Friday March 17, 5 pm

Exercises with solutions (Set D)

Detectability and Output Feedback Stabilizability of Nonlinear Networked Control Systems

ON SCALABLE CODING OF HIDDEN MARKOV SOURCES. Mehdi Salehifar, Tejaswi Nanjundaswamy, and Kenneth Rose

Bounds on Mutual Information for Simple Codes Using Information Combining

On Scalable Coding in the Presence of Decoder Side Information

18.2 Continuous Alphabet (discrete-time, memoryless) Channel

Stabilization with Disturbance Attenuation over a Gaussian Channel

Amobile satellite communication system, like Motorola s

Event-triggered second-moment stabilization of linear systems under packet drops

The Poisson Channel with Side Information

Quantized average consensus via dynamic coding/decoding schemes

Feedback Stabilization over a First Order Moving Average Gaussian Noise Channel

ECE Information theory Final (Fall 2008)

Cooperative Communication with Feedback via Stochastic Approximation

Generalized Writing on Dirty Paper

The necessity and sufficiency of anytime capacity for control over a noisy communication link: Parts I and II

On Design of Reduced-Order H Filters for Discrete-Time Systems from Incomplete Measurements

A Mathematical Framework for Robust Control Over Uncertain Communication Channels

Optimality of Walrand-Varaiya Type Policies and. Approximation Results for Zero-Delay Coding of. Markov Sources. Richard G. Wood

A General Formula for Compound Channel Capacity

Variable Length Codes for Degraded Broadcast Channels

Nonstochastic Information Theory for Feedback Control

Feedback Capacity of the Gaussian Interference Channel to Within Bits: the Symmetric Case

Coding into a source: an inverse rate-distortion theorem

PREDICTIVE quantization is one of the most widely-used

Necessary and sufficient bit rate conditions to stabilize quantized Markov jump linear systems

Towards the Control of Linear Systems with Minimum Bit-Rate

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

Approximate Capacity of Fast Fading Interference Channels with no CSIT

On the Duality between Multiple-Access Codes and Computation Codes

Channel Estimation with Low-Precision Analog-to-Digital Conversion

Dispersion of the Gilbert-Elliott Channel

Control over finite capacity channels: the role of data loss, delay and signal-to-noise limitations

Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels

Distributed Stochastic Optimization in Networks with Low Informational Exchange

Cut-Set Bound and Dependence Balance Bound

A Framework for Control System Design Subject to Average Data-Rate Constraints Silva, Eduardo; Derpich, Milan; Østergaard, Jan

EVALUATING SYMMETRIC INFORMATION GAP BETWEEN DYNAMICAL SYSTEMS USING PARTICLE FILTER

Capacity of Memoryless Channels and Block-Fading Channels With Designable Cardinality-Constrained Channel State Feedback

Basic Principles of Video Coding

Energy State Amplification in an Energy Harvesting Communication System

Minimum Feedback Rates for Multi-Carrier Transmission With Correlated Frequency Selective Fading

Linear Codes, Target Function Classes, and Network Computing Capacity

Anytime Capacity of the AWGN+Erasure Channel with Feedback. Qing Xu. B.S. (Beijing University) 1997 M.S. (University of California at Berkeley) 2000

MOST control systems are designed under the assumption

Shannon s noisy-channel theorem

On Common Information and the Encoding of Sources that are Not Successively Refinable

5958 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 56, NO. 12, DECEMBER 2010

EECS 229A Spring 2007 * * (a) By stationarity and the chain rule for entropy, we have

Tracking and Control of Gauss Markov Processes over Packet-Drop Channels with Acknowledgments

Reliable Computation over Multiple-Access Channels

STOCHASTIC STABILITY OF EXTENDED FILTERING FOR NONLINEAR SYSTEMS WITH MEASUREMENT PACKET LOSSES

Secret Key Agreement Using Asymmetry in Channel State Knowledge

Control Capacity. Gireeja Ranade and Anant Sahai UC Berkeley EECS

Noise-Shaped Predictive Coding for Multiple Descriptions of a Colored Gaussian Source

ASIGNIFICANT research effort has been devoted to the. Optimal State Estimation for Stochastic Systems: An Information Theoretic Approach

AN INFORMATION THEORY APPROACH TO WIRELESS SENSOR NETWORK DESIGN

Stochastic Stabilization of a Noisy Linear System with a Fixed-Rate Adaptive Quantizer

Transcription:

Towards control over fading channels Paolo Minero, Massimo Franceschetti Advanced Network Science University of California San Diego, CA, USA mail: {minero,massimo}@ucsd.edu Invited Paper) Subhrakanti Dey, Girish Nair Dept. of lectrical ngineering University of Melbourne Parkville Victoria 31 Australia mail: {sdey,gnair}@ee.unimelb.edu.au Abstract Motivated by control problems over wireless fading channels, the mean square stabilization of a linear system over a time-varying finite rate feedback link is studied. Process and observation disturbances are allowed to occur over an unbounded support. Necessary and sufficient conditions to stabilize the system are derived, and their implications are discussed. I. INTRODUCTION In wireless communication, the quality of the communication link between transmitter and receiver varies over time because of random fading in the received signal. In the case of digital communication, this can reflect in a time variation of the rate supported by the wireless channel. The coherence time indicates the time interval over which the channel can be considered constant. If the coherence time is long enough, then transmitter and receiver can estimate the quality of the link by sending a known sequence called pilot, and can adapt the communication scheme to the channel condition. Let us assume the aim is to stabilize an unstable linear plant through a controller which receives measurements of a sensor over a fading wireless channel. Is it possible to design a communication scheme that changes dynamically according to the channel condition and, at the same time, is guaranteed to stabilize the system? This is the problem under investigation. As a first approach, it is assumed that the communication channel is noiseless and, at any given time k, allows transmission of R k bits without error. The case when R k changes deterministically has been well studied. It has been shown, under different notions of stability and system models see 1 3, and references therein), that a discrete linear plant with parameter λ > 1 can be stabilized if and only if the data rate R k over the digital link is no less than log 2 λ. This paper focuses on the case when R k fluctuates randomly. We model the coherence time by assuming that R k remains constant in blocks of n consecutive channel uses and varies according to an ergodic process across blocks. Channel state estimation at the transmitter and receiver is reinterpreted here as causal knowledge, at both encoder and decoder, of the rate supported by the communication link. A work directly related to ours is 2. These authors also considered the case of a random time-varying digital link and provided necessary and sufficient conditions for m-th moment stability. The main difference with our work is that we allow the system disturbance to have unbounded support, and the encoder has access to output feedback instead than state feedback. We point out that this is a more general model, which considerably complicates the design of the adaptive quantizer, as this must be capable of tracking the state when atypically large disturbances affect the system. Some preliminary results are presented next. They consist of necessary and sufficient conditions on the channel rate to achieve second moment stability. A necessary condition to stabilize a linear system with parameter λ > 1 in the second moment sense is, λ < 1. 1) 2 R Furthermore, it is shown that there exists a coder-decoder pair that stabilizes the system if, λ < c, 2) 2 R where c < 1 16. Condition 1) can be intuitively interpreted as follows. If no information is sent over the link during a transmission block, the estimation error at the controller grows by λ n. The information sent by the coder can reduce the this error by at most 2 nr, where nr is the total rate supported by the channel in a given block. However, if averaging over the fluctuation of the rate λ n exceeds 2 nr, then the information rate of the channel cannot compensate on average) the dynamics of the system and it is not possible to stabilize the plant. Notice that if the variation of the rate is deterministic, then our necessary condition reduces to the well known R > log 2 λ. As for condition 2), we give an explicit construction of a coder-decoder pair similar to the one by Nair and vans 1. However, while in 1 the time variation of the transmission rate is part of the coder design, in the model under consideration the variation of {R k } k=1 is uniquely determined by a stochastic process. As the rate process is assumed constant over blocks of n channels uses, coder and decoder exploit the causal knowledge of the rate, transmitting one packet per block, and by encoding each packet at the rate supported by the channel. We notice that there is a constant gap between our necessary and sufficient conditions 1-2). We are currently working towards the possibility of closing this gap and we conecture that 1) is indeed achievable using a more complex codingdecoding scheme than the one presented here. Although a sub-

optimal achievability result can be derived also in the case of multi-state discrete linear systems, in what follows, for ease of exposition we limit the treatment only to the scalar case. The rest of the paper is organized as follows. After introducing formally the stabilization problem, the necessity result is proved in Section III, while the sufficiency is developed in Section IV. II. PROBLM FORMULATION Consider the scalar partially-observed unstable stochastic linear system x k+1 = λx k + u k + v k, y k = x k + w k, k W 3) with state x k, process disturbance v k, control u k, measurement y k and measurement noise w k. It is assumed that λ > 1, so that the system is unstable. No Gaussian assumptions are made on the disturbances, but the following assumptions are supposed to hold A1. x,v k and w are mutually independent for all k, W. A2. ɛ > such that x, v k and w k have uniformly bounded 2 + ɛ)th absolute moments over k W. A3. inf k W hv k ) >. Thus, β > such that e 2hv k) > β for all k W suppose that observer and controller are connected by a timevarying digital link. Suppose that rate remains constant over blocks of n Z + channel uses but changes independently from block to block according to a given probability distribution. In each of the n uses of the digital link during the -th block symbols coming from an alphabet of size 2 R are transmitted without error from encoder to controller. Let S k {1,..., 2 R } denote the symbol sent at time k {n,..., + 1)n 1}. An underlying stochastic process determines the alphabet size 2 R in each block. For simplicity, suppose that for all W R are i.i.d random variables distributed as R, where R can only take finite number of integer values, such that PR = r i ) = p i, i I Z +, 4) for r i W, r min r i r max <, i I p i = 1, and where I is a finite set. At the beginning of each block, coder and decoder are assumed to know R and hence all {R i } i= ), while the realization of the rate process in future blocks, {R i } i=+1, is unknown to them. The aim is to construct a coder-decoder which stabilizes the plant in the mean square sense sup x k 2 < 5) k W using the finite data rate provided by the time-varying digital link. ach transmitted symbol can depend on all past and present measurements, the present channel state and the past symbols, s k =g k y,..., y k, s,..., s k 1, R ), 6) k {n,..., + 1)n 1}, W, where g k ) is the coder mapping at time k. The control sequence, on the other hand, can depend on all past and present channel symbols u k = f k s,..., s k ), k W, 7) where f k ) is the controller mapping at time k. From 7),6) and 3) observe that R affects only x k for all k n + 1. Thus, x n does not depend on R. In the sequel, expectation with respect to the random variable X will be denoted as X. III. NCSSITY Theorem 3.1: Under the conditions above, necessary condition for stabilizability in mean square sense is that ) λ 2 n < 1 8) 2 2R Proof: In order to prove the statement, we find a lower bound for the second moment of the state, and show that q. 8 is a necessary condition for the lower bound to be finite. We focus on the times k = n with W, i.e. the beginning of each channel block. Let S = {S,..., S +1)n 1 }, denote the symbols sent over the noiseless channel until the end of the -th channel block. From the close loop equation, we have +1)n 1 x +1)n = λ n x n + λ +1)n 1 k f k S,..., S k )+v k Let n = 1 2πe S 1 entropy power of x n k=n e 2hx n S 1 =s 1 ) s 1 }, averaged over all possible s 1 of x n is lower bounded by n : n = 1 2πe S 1 1 2πe S 1 = S 1 = x 2 n be the conditional conditioned on the event {S 1 e 2hxn S 1 =s 1 ) =. The second moment e ln2πex2 n S 1 =s 1 x 2 n S 1 = s 1 where the inequality follows from maximum entropy theorem 4, Theorem 9.4.1. Thus, necessary condition for 5) to hold is that sup W n <. The proof utilizes the following result. Lemma 3.2: S S 1,R e 2hx n S =s ) 1 1 2 e2hx n S R =s 1 ) 9) Proof: First, observe that the following chain of inequalities holds: S S 1,R hx n S 1 = s 1, S = s, R ) = hx n, S S 1 = s 1, R ) HS S 1 = s 1, R ) hx n S 1 = s 1, R ) HS S 1 = s 1, R ) hx n S 1 = s 1, R ) ln 2 nr = hx n S 1 = s 1 ) ln 2 nr 1)

where hx, A B) with A discrete denotes lnp A B f x A,B ). The last inequality follows from the fact that, given R, the cardinality of {S n,..., S +1)n 1 } is 2 nr, and where the last equality follows from the fact that, given the control input up to time n 1, 3) implies that x n is independent of R i.e. x n S 1 R is Markov). Then, S S 1,R e 2hx n S =s ) S S 1,R e 2hxn S 1 =s 1,S =s,r) e 2 S hx S 1,R n S 1 =s 1,S =s,r) e 2hX S 1 =s 1 ) ln 2 nr = 1 =s 1 ) 1 2 R e2hx n S where the first inequality follows from the fact that conditioning reduces the entropy; the second inequality follows from Jensen s inequality; finally, 1) implies the third inequality. Next, it is shown that n evolves according to a recursive equation. Using standard properties of entropy 4 translation invariance, conditional version of entropy power inequality), and assumptions A1. and A3., it follows that S e 2hx +1)n S =s ) = = S e 2hλn x n + +1)n 1 k=n λ +1)n 1 k v k S ) =s S = λ S λ S = λ S = λ S 1 e 2hλn x n S =s ) + +1)n 1 k=n e ) +1)n 1 2hxn S =s + k=n e ) +1)n 1 2hxn S =s + e 2hx n S =s ) + γ,r S S 1,R 1,R k=n λ S 1 2 nr 2 e2hx n S = λ 1 R 2 nr 2 S 1 e 2hλ+1)n 1 k v k S =s ) λ 2+1)n 1 k e 2hv k) λ 2+1)n 1 k β e ) 2hxn S =s + γ 1 =s 1 ) + γ e 2hx n S 1 =s 1 ) + γ, the second inequality uses assumption A3. above, i.e. e 2 f hv k) > β. The constant γ is defined as and γ := +1)n 1 k=n λ 2+1)n 1 k β. Finally, the last inequality follows from Lemma 3.2 and the fact that R is independent of x n and S 1. Thus, using the fact that the rate process is i.i.d., λ n +1 2 R n + γ 2πe. Therefore, λ 1 implies that sup 2 R W n =. IV. SUFFICINCY Proposition 4.1: Suppose that nr min 2. Under the conditions above, sufficient condition for stabilizability in mean square sense is that ) λ 2 n 2 2R < 1 φζ 11) where φ and ζ are constants such that φ = 2 1+ɛ 2 and ζ 8. The remaining part of the paper is dedicated to the proof of the above Proposition. The coder/decoder pair is based on the construction given in 1. The coder utilizes the quantizer that is described below. A. Quantizer The quantizer partitions the real line into non-uniform regions, in such a way that most of the regions are concentrated in a symmetrical way around the origin. A parameter ρ > 1 determines the speed at which the quantizer range increases. The quantizer generates 2 ν, ν 2, quantization intervals by partitioning the set 1, 1 into 2 ν 1 intervals of equal length, and the sets ρ i 2, ρ i 1, ρ i 2, ρ i 1 ) into 2 ν i intervals of equal length, i 2,..., ν. The two open sets ρ ν 1, ) and, ρ ν 1 are respectively the leftmost and rightmost intervals of the quantizer. Note that as i increases the size of the intervals increases. We assume that I),..., I2 ν 1) are labels for the intervals from left to right. Let κ ν ω) be half-length of interval Iω) for ω 1,..., 2 n 2, be equal to ρ ν ρ ν 1 when ω = 2 ν 1 and equal to ρ ν ρ ν 1 ) when ω =. q ν x) := ω ν ω) be midpoint of interval x Iω) for ω 1,..., 2 ν 2, be equal to ρ ν when ω = 2 ν 1 and equal to ρ ν when ω =. If a real number x is in Iω) for some ω {,..., 2 ν 1 }, then the quantizer approximates x with ω ν ω). The quantization error is not uniform over x R, but is bounded by κ ν ω) for all ω {1,..., 2 ν 2 }. Suppose now that X is a real-valued random variable. A fundamental property of the quantizer is that the average quantization error diminish like the inverse square of the number of levels, 2 ν. More precisely, if the 2+ ɛ)-th moment of X is bounded for some ɛ >, then an upper bound of the second moment of the estimation error decays as 2 ν. The existence of a higher moment of X is useful to bound the estimation error using Chebyshev s inequality) when X lies in one of the two open intervals ρ ν 1, ) and, ρ ν 1. Let L be a strictly positive random variable, define the functional M ɛ X, L L 2 + X 2+ɛ L ɛ. 12) The functional M ɛ X, L is an upper bound to the second moment of X: X 2 = X 2 1 X L + 1 X >L ) M ɛ {X L} 13)

Given a third random variable R, define the conditional version of M ɛ X, L given R as M ɛ X, L R L 2 + X 2+ɛ L ɛ R 14) The fundamental property of the quantizer described above is given by the following result: Lemma 4.2: Let X and L > and R 2/n be random variables with X 2+ɛ < for some ɛ >, and n W. If ρ > 2 2/ɛ, then the quantization error X Lq nr X/L) satisfies M ɛ X Lq ν X/L), Lκ nr ω) R ζ 2 R M ɛx, L 15) for some ζ 8 determined only by ɛ and ρ. Proof: 1, Lemma 5.2 Next, the coder and observer are described. B. Coder The first stage of the encoding process consists of computing the linear minimum variance estimator of the plant state based on the previous measurements and control sequences. The filter process satisfies a recursive equation of the same form as 3), namely x k+1 = λ x k + u k + z k, z W. 16) where the innovation z k is uniformly bounded. From the orthogonality principle the stability of x is equivalent to that of x. The output x k of the filter or a function of it) must be transmitted using the finite number of bits supported on the digital channel. Coder and decoder share a state estimator ˆx k based uniquely on the symbols sent on the digital link. Since ˆx k is available both at the coder and decoder, while the minimum variance estimator is available at the coder only, the encoder utilizes a predictive quantizer to encode the error between x k and ˆx k. The error is scaled by an appropriate coefficient and then quantized using the quantizer in Section IV-A. Just before the start of the -th block at time k = n the coder sets the quantization rate ν equal to nr and computes ) xn ˆx n ω nr ω ) = q nr 17) l where l is a scaling factor updated after each packet transmission. The index ω of the quantization level is converted into a string of nr bits and transmitted using the n channel uses of the -th block. Before the + 1)-th block, the coder updates the state estimator as follows where ˆx +1)n = λ n ˆx n + l ω nr ω )+ +1)n 1 + k=n λ +1)n 1 k Lˆx k 18) ˆx k+1 = λ + L)ˆx k, k n,..., n + n 2 19) and ˆx =. L is the certainty-equivalent control coefficient. Finally the scaling coefficient l is updated as follows l +1 = max{σ, l λ n κ nr ω )} 2) with l = σ, where σ 2+ɛ is a uniform bound in the 2 + ɛ)- moment of n 1 g := λ n i z n+i, W 21) i= It is important to note from 2) and the independence of the rate process that the l does not depend on R. C. Controller At time k = n coder and controller are synchronized and have common knowledge of the state estimator ˆx n. While receiving the digitalized index of 17) during times n,..., n+n 2, the controller sends to the plant a certaintyequivalent control signal u k = Lˆx k k n,..., n + n 2 22) where ˆx k is updated as in 19). Once received ω nr ω ), the controller updates the estimator ˆx +1)n using 18). Synchronism between coder and observer is ensured by the fact that the initial value ˆx is set equal to zero at both coder and decoder, and by the fact that the digital link is noiseless. D. Analysis In this section it is shown that the code/observer pair described above ensures that the second moment of x is uniformly bounded if 11) is satisfied. Define the coder error at time k W as f k = ˆx k x k. The following analysis is developed in three steps. First it is shown that f k is bounded for all times k = n W, i.e. the beginning of each channel block. Next, the analysis is extended to all k W. Finally the stability of f k for all k W is shown to imply that x is uniformly bounded. First we show that the coder error f k = ˆx k x k is bounded in mean square for all times k = n W. Substituting 22) into 16), and iterating over a block duration x +1)n = λ n x +1)n 1 n + λ +1)n 1 k Lˆx k + z k ) k=n = λ n x +1)n 1 n + g + λ +1)n 1 k Lˆx k ) 23) k=n where g is defined in 21). So subtracting 23) from 18) f +1)n = ˆx +1)n x +1)n = λ n f n l ω nr ω ) + g 24) From the inequality x + y ) α 2 α 1 x α + y α ) α >, f +1)n 2+ɛ φ λ n 2+ɛ f n l q nr f n /l ) 2+ɛ + z 2+ɛ)

with φ = 2 1+ɛ. Dividing by l+1 ɛ and taking expectations, f +1)n 2+ɛ l ɛ +1 ) φ λ n 2+ɛ fn l q nr f n /l ) 2+ɛ g 2+ɛ + φ λ n 2+ɛ = φ λ n 2 l ɛ +1 l κ nr ω ) ɛ l ɛ +1 fn l q nr f n /l ) 2+ɛ g 2+ɛ ) l λ n κ nr ω ) ɛ + σ ɛ fn l q nr f n /l ) 2+ɛ g 2+ɛ ) + σ ɛ 25) using the fact that, from 2), l +1 = max{σ, l λ n κ nr ω )} Now let Observe that θ := M ɛ {f n, l } l 2 + f n 2+ɛ l ɛ l 2 +1 max{σ 2, l 2 λ n 2 κ nr ω ) 2 } where the second inequality follows from Lemma 4.2), and the last equality uses the fact that the rate process is i.i.d. and that f n and l are independent of R. Therefore, λ θ +1 = φ2σ 2 n 2 + φζ θ and it follows that λ n 2 2 R 2 R < 1 φζ, where φ = 2 1+ɛ and ζ 8, is a sufficient condition to have sup W θ <. From 13) it follows that f n 2 is bounded for every W. Next, for any r {, n 1} triangle inequality implies r 1 f n+r λ r f n + λ r 1 k L z n+k, k= so the error f k is uniformely bounded for all k W. Finally, by rewriting 16) as x k+1 = λ + L) x k Lf k + z k the fact that f k and z k are uniformly bounded and the strict stability of λ + L ensures that x 2 k < for all k W. V. CONCLUSION In this paper we considered mean square stabilizability of a discrete-time, linear system with a noiseless time-varying digital communication link. Process and observation disturbances are allowed to occur over an unbounded support. Necessary and a sub-optimal sufficient conditions to stabilize the system are derived. VI. ACKNOWLDGMNT This work is partially supported by the National Science Foundation CARR award CNS-546235 RFRNCS σ 2 + λ n 2 l κ nr ω ) 2 26) 1 G. N. Nair and R. J. vans, Stabilizability of stochastic linear systems with finite feedback data rates, SIAM Journal on Control and Optimization, Society for Industrial and Applied Mathematics, USA, vol. 43, no. Adding this to 25), using g 2+ɛ < σ 2+ɛ and the definition 2, pp. 413-436, July 2, 24. of θ, 2 Martins N.C., Dahleh M.A. and lia, N., Feedback stabilization of uncertain systems in the presence of a direct link Martins, Automatic Control, +1 θ φσ 2 + λ n 2 { l κ nr ω ) 2} I Transactions on, Volume 51, Issue 3, March 26 Pages):438 - + 447. { } ) + φ λ n 2 fn l q nr f n/l ) 2+ɛ 3 Tatikonda, S.; Mitter, S.; Control under communication constraints, + σ 2 Automatic Control, I Transactions on, Volume 49, Issue 7, July 24 l κ nr ω ) ɛ Pages):156-168 { })4 T. Cover and J. Thomas, lements of Information Theory. New York: = φ 2σ 2 + λ n 2 fn l q nr f n/l ) 2+ɛ + l κ l κ nr ω k ) ɛ nr ω ) 2 Wiley, 1987. { ) }) = φ 2σ 2 + λ n 2 fn R M ɛ f n l q nr, l κ nr ω ) l R ) φ 2σ 2 + λ n 2 ζ R 2 M R ɛ { f n l } λ = φ2σ 2 n 2 + φζ θ 2 R