Quantization for Distributed Estimation

Similar documents
Design of Optimal Quantizers for Distributed Source Coding

Distributed Arithmetic Coding

A Practical and Optimal Symmetric Slepian-Wolf Compression Strategy Using Syndrome Formers and Inverse Syndrome Formers

ECE Information theory Final

ECE Information theory Final (Fall 2008)

On Side-Informed Coding of Noisy Sensor Observations

Network Distributed Quantization

On Scalable Source Coding for Multiple Decoders with Side Information

A Systematic Description of Source Significance Information

Multiuser Successive Refinement and Multiple Description Coding

Efficient Use of Joint Source-Destination Cooperation in the Gaussian Multiple Access Channel

Digital Communications III (ECE 154C) Introduction to Coding and Information Theory

Variable-Rate Universal Slepian-Wolf Coding with Feedback

Proceedings of the DATA COMPRESSION CONFERENCE (DCC 02) /02 $ IEEE

6196 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 9, SEPTEMBER 2011

A Hyper-Trellis based Turbo Decoder for Wyner-Ziv Video Coding

Distributed Functional Compression through Graph Coloring

Optimal matching in wireless sensor networks

Binary Rate Distortion With Side Information: The Asymmetric Correlation Channel Case

Performance of Polar Codes for Channel and Source Coding

QUANTIZATION FOR DISTRIBUTED ESTIMATION IN LARGE SCALE SENSOR NETWORKS

Data and error rate bounds for binar gathering wireless sensor networks. He, Xin; Zhou, Xiaobo; Juntti, Markk Author(s) Tad

Multi-Hypothesis based Distributed Video Coding using LDPC Codes

Multiterminal Source Coding with an Entropy-Based Distortion Measure

On Scalable Coding in the Presence of Decoder Side Information

On Common Information and the Encoding of Sources that are Not Successively Refinable

18.2 Continuous Alphabet (discrete-time, memoryless) Channel

Multimedia Communications. Scalar Quantization

Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information

Slepian-Wolf Code Design via Source-Channel Correspondence

Distributed Compression for Condition Monitoring of Wind Farms

Turbo Compression. Andrej Rikovsky, Advisor: Pavol Hanus

A New Achievable Region for Gaussian Multiple Descriptions Based on Subset Typicality

Soft-Output Trellis Waveform Coding

Approximately achieving the feedback interference channel capacity with point-to-point codes

Practical Coding Scheme for Universal Source Coding with Side Information at the Decoder

Distributed Lossy Interactive Function Computation

The Duality Between Information Embedding and Source Coding With Side Information and Some Applications

On Perceptual Audio Compression with Side Information at the Decoder

Communication Engineering Prof. Surendra Prasad Department of Electrical Engineering Indian Institute of Technology, Delhi

The Distributed Karhunen-Loève Transform

Cooperative Spectrum Sensing for Cognitive Radios under Bandwidth Constraints

Representation of Correlated Sources into Graphs for Transmission over Broadcast Channels

IN this paper, we study the problem of universal lossless compression

MMSE DECODING FOR ANALOG JOINT SOURCE CHANNEL CODING USING MONTE CARLO IMPORTANCE SAMPLING

arxiv: v2 [math.st] 11 Oct 2017

ON DISTRIBUTED ARITHMETIC CODES AND SYNDROME BASED TURBO CODES FOR SLEPIAN-WOLF CODING OF NON UNIFORM SOURCES

Interactive Decoding of a Broadcast Message

On Optimal Coding of Hidden Markov Sources

Rate-adaptive turbo-syndrome scheme for Slepian-Wolf Coding

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 12, DECEMBER

On The Binary Lossless Many-Help-One Problem with Independently Degraded Helpers

Constellation Shaping for Communication Channels with Quantized Outputs

Communication Cost of Distributed Computing

AN INFORMATION THEORY APPROACH TO WIRELESS SENSOR NETWORK DESIGN

RCA Analysis of the Polar Codes and the use of Feedback to aid Polarization at Short Blocklengths

Lattices for Distributed Source Coding: Jointly Gaussian Sources and Reconstruction of a Linear Function

MARKOV CHAINS A finite state Markov chain is a sequence of discrete cv s from a finite alphabet where is a pmf on and for

Distributed Lossless Compression. Distributed lossless compression system

ProblemsWeCanSolveWithaHelper

The Poisson Channel with Side Information

OPTIMUM fixed-rate scalar quantizers, introduced by Max

Lossy Distributed Source Coding

An Achievable Rate Region for the 3-User-Pair Deterministic Interference Channel

An Efficient SF-ISF Approach for the Slepian-Wolf Source Coding Problem

Shannon s noisy-channel theorem

SCALABLE AUDIO CODING USING WATERMARKING

EE368B Image and Video Compression

Green Traffic Compression in Wireless Sensor Networks

Lecture 7. Union bound for reducing M-ary to binary hypothesis testing

Gaussian source Assumptions d = (x-y) 2, given D, find lower bound of I(X;Y)

Graph Coloring and Conditional Graph Entropy

Encoder Decoder Design for Event-Triggered Feedback Control over Bandlimited Channels

C.M. Liu Perceptual Signal Processing Lab College of Computer Science National Chiao-Tung University

Cognitive Multiple Access Networks

The Gallager Converse

Side Information Aware Coding Strategies for Estimation under Communication Constraints

Extended Subspace Error Localization for Rate-Adaptive Distributed Source Coding

Sum Capacity of General Deterministic Interference Channel with Channel Output Feedback

Joint Source-Channel Coding for the Multiple-Access Relay Channel

PERFORMANCE COMPARISON OF DATA-SHARING AND COMPRESSION STRATEGIES FOR CLOUD RADIO ACCESS NETWORKS. Pratik Patil, Binbin Dai, and Wei Yu

An Uplink-Downlink Duality for Cloud Radio Access Network

Chapter 9 Fundamental Limits in Information Theory

arxiv: v1 [cs.it] 20 Jan 2018

On Optimum Conventional Quantization for Source Coding with Side Information at the Decoder

Minimum Necessary Data Rates for Accurate Track Fusion

A Half-Duplex Cooperative Scheme with Partial Decode-Forward Relaying

Side-information Scalable Source Coding

Approximating the Gaussian multiple description rate region under symmetric distortion constraints

Cooperative Communication with Feedback via Stochastic Approximation

Compressed Sensing Using Bernoulli Measurement Matrices

Decentralized Detection In Wireless Sensor Networks

Digital communication system. Shannon s separation principle

The Gaussian Many-Help-One Distributed Source Coding Problem Saurabha Tavildar, Pramod Viswanath, Member, IEEE, and Aaron B. Wagner, Member, IEEE

Wyner-Ziv Coding of Video with Unsupervised Motion Vector Learning

The Capacity Region of the Gaussian Cognitive Radio Channels at High SNR

Secret Key and Private Key Constructions for Simple Multiterminal Source Models

IN MOST problems pertaining to hypothesis testing, one

Reliable Computation over Multiple-Access Channels

Fixed-Length-Parsing Universal Compression with Side Information

Transcription:

0 IEEE International Conference on Internet of Things ithings 0), Green Computing and Communications GreenCom 0), and Cyber-Physical-Social Computing CPSCom 0) Quantization for Distributed Estimation uan-yu Chen, and wang-cheng Chen, Fellow,IEEE Graduate Institute of Communication Engineering, National Taiwan University, Taipei, Taiwan Email: r09@ntu.edu.tw and chenkc@cc.ee.ntu.edu.tw Abstract Distributed Estimation scheme has been widely used in sensor networks. Through various approaches, sum of bits, Sigma-Delta modulation, and source coding, this paper implements the above techniques for distributed estimation and provides a new one called Quantization Region Allocation. In the end, this paper analyzes the performance minimum mean squared error) and demonstrates performance approximating the soft decision scenario. Index Terms distributed estimation, source coding, Sigma- Delta modulation a) I. INTRODUCTION The problem we consider in this paper is associated with Distributed Source Coding DSC) for multiple sources. DSC problems regard the compression of multiple correlated information sources that do not communicate with each other, as Fig. a shows. By designing the encoders for different source and the joint decoder, DSC is able to shift the computational complexity from encoder side to decoder side, therefore provide appropriate frameworks for applications with complexity-constrained sender. The lossless version of this problem with discrete sources was solved by Slepian and Wolf [], using a random-binning approach. It is shown that for lossless scenarios, cooperation among encoders does not improve the sum-rate boundary. For =, the achievable rate region R,R for source Y,Y is given by R HY Y ) R HY Y ) R + R HY,Y ) Practical distributed lossless coding schemes have been proposed [], [3] that are close to the Slepian-Wolf bound. However, when lossless coding is not possible, rate distortion theory must be taken into account. This problem has been considered by Wyner and Ziv [] for the case where the decoder has access to side information about the source. When Y,Y,Y are influenced by a source X, sensors observe independently corrupted versions of X. The sensors encode their observations without cooperating with one another. This is the so-called CEO problem and was first studied by Berger, Zhang, and Viswanathan [5] in the context of discrete memoryless sources and observations. The fusion center and sensors are referred to as CEO and agents respectively, which is illustrated in Fig. b. Prabhakaran, Tse, Ramchandran [] have shown that if X is zero mean and if This work was also supported by National Science Council, National Taiwan University and Intel Corporation under Grants NSC0-9-I-00-00 and NTU0R750. ) b) c) Fig. : System structure for a) distributed source coding b) CEO problem c) quantized distributed estimation. the transmission from agents to CEO is error-free, the -tuple rate region R,R,R ) is given by RD) = R D r,r,r ) ) r,r,r ) FD) where R D r,r, r ) and FD) is given by 3) and ) respectively. The system structure for our problem is depicted in Fig. c. In the first approach, we simply use the number of s +A) and 0 s A) sent from sensors to estimate the value of X, namely, X = μ + 0 )B 5) where, 0 denote the number of s and 0 s respectively and B is a constant needed to be determined. This can give a satisfactory result. To improve the performance, Sigma-Delta modulation Σ-Δ modulation) [7] is introduced. In addition, 97--799-597-9/ $3.00 0 IEEE DOI 0.09/iThings.0.0 3

R D r,r,r )= R,R,R ): R k r k + log D k A k A log α + ) } exp r k ) σ, non empty A,,L} k A c } FD) = r,r,r ) R + : α + exp r k ) σ = D 3) ) when it comes to quantization at each sensor, jointly optimize the quantization region would be considered. However, to simplify the computation, we combine Lloyd Algorithm with weighted bits, providing a new approach called Quantization Region Allocation in section IV. II. PROBLEM STATEMENT As Fig. c shows, sensors is used to detect the source parameter X in the following way. Assume that the source X is Gaussian distributed with mean μ and variance α. The k-th sensor has noisy observation Y k = X + N k, where N k is i.i.d. Gaussian with mean 0 and variance σ. In addition, the data sent from each sensor Y k also suffers from noise N k, which is assumed i.i.d. and has same distribution as N k. The fusion center finally determines X based on Z k = Y k + N k,k =,,. When each sensor noiselessly sends its observation to Fusion Center, there are two possible ways: to quantize i.e. to determine between A, A}, also known as hard decision) or not i.e. just to send observed value, also known as soft decision). III. BASIC APPROACH A. Sum of Bits Assume that each sensor sends A to the receiver if Y k >μ, and sends A) otherwise. The fusion center first detect whether A or A) is sent from each path. Define the Bernoulli random variables I i }, i = 0,,, where I i =0if A is decided and I i =if A) is decided. Our estimator is given by X = μ + ) Ii B = μ + NB ) Furthermore, if we define the random variable M = I i, thenm X B, p), where p = PrI k = X). Since N = M, the mean squared error can be written as E[X μ NB) ] =E [ E[X μ M)B) X] ] 7) Differentiate 7) with respect to B yeild the optimal value E[X μ)p] B = )E[p] E[p ) ]+ The result is shown in Fig.. Here we set μ = and A = B. For the simulation curves, different values of B has Mean Squared Error... 0. 0. 0. 0. Theoretical Curve α=.0,σ=.0) Simulation Curve α=.0,σ=.0) Theoretical Curve α=.0,σ=.5) Simulation Curve α=.0,σ=.5) Theoretical Curve α=.5,σ=.0) Simulation Curve α=.5,σ=.0) 0 0 5 0 5 0 Fig. : by sum-of-bit approach. been tested, where B varies from 0.05 to 0.5 with step size 0.05, and the optimal value of B is selected. From Fig. it can be found that the rises a little when σ increases, while the rises more when α increases. This is intuitive since the variance of noise can be recovered with large. But it is harder to estimate X when the source itself has large variance. B. Σ-Δ Modulation Approach Sigma-Delta Modulation can also be used to deal with the problem. This is done by adding a filter and oversampling the observations in front of the fusion center from each path, as Fig. 3 shows. After collecting the -bit information from each path, we make the estimation by b i X = μ + k α)+k M 9) where k,k are parameters needed to be defined, M is the number of decimation, and b i is the number of ones density of input Z i. The value of k,k,m has influence on the performance. To minimize MSE, we first fix k. Then optimize k by simulation. Fig. a shows the performance for different k. In fact, the optimal value of k is about 0.. This may be associated with the optimal -bit quantization level for Gaussian random variable, i.e. the quantization level for zeromean Gaussian with variance σ is ±σ /π. On the other hand, the for different value of M is depicted in Fig. 5

0 Fig. 3: The modified fusion center. k =0.50 k =0.75 k =.00 k =.5 k =.50 0 0 0 a) M =, k =0.50, 0.75,.00,.5,.50. M= M= M=3 M= 0 0 b) k =0., M =,, 3,. Fig. : for ) different k s, ) different M s by Σ Δ modulation approach. b. It can be seen that the performance for M is about the same. Fig. 5: A distributed estimation system with general observation distribution. C. Source Coding Approach A more general system model for distributed source coding has been considered by Gubner [], as Fig. 5 shows. Each sensor k processes its measurement Y k to obtain an output Z k. Each Z k is then transmitted to the fusion center. Assume that the use of error-correcting codes permits us to view the channel as noiseless. For each Y k, A k, A kn is a partition of the real line, R. In other words, it quantizes Y k in N levels. The sensor output Z k is then defined by Z k i )I Aki Y k ) 0) where I A y) is the indicator function given by,y A I A y) = ) 0,y / A At the fusion center, Z, Z is collected and for simplicity, the estimator X is defined by N X = c ki I i } Z k ) ) The algorithm for optimizing the partitions A i } N, A i} N, A i} N and the parameters c ki, i =,,,N, k =,,, is given in [], which is straight-forward. However, the computation makes it not realistic for implementation. Thus, we introduce another method to define the partitions A ki } N. Consider a quantization problem: A continuous random variable X are needed to be quantized in bits. How to minimize the mean squared error by specifying the quantization levels and step sizes? If the step sizes are identical for each quantization region, it is called uniform quantization. Nevertheless, to minimize MSE, nonuniform quantization must be used, in which the step sizes may not be of equal length. This can be done by the well-known Lloyd algorithm [9]. The algorithm operates as follows. Given an initial quantization region R, R,, R M, where M =, we calculate the conditional mean for each region. xf X x) dx R a i = i,,,, 3) P x R i ) where a i is the quantization level for region i. Then the region R i = x b i x<b i } is updated by a i s. b i = a i + a i+,,,, ) 5

0 35 Mean Squared Error 30 5 0 5 0 0 Fig. 7: for μ =5, α =0, σ =by source coding approach. Fig. : Detection rule for sensor k, k =,, 3. with b 0, b + =. After the regions R i,i =,,, are decided, repeat the calculation of a i s. This iteration stops until the difference of MSE between two iterations is small enough. Next, we could connect this problem to our original problem. This is illustrated by Fig.. For instance, if =3,we first use Lloyd Algorithm to generate the quantization levels and regions for the random variable X. The detection rule at each sensor is described as follows. A bit 0), if Y R Z = R R3 R A bit ), otherwise A bit 0), if Y R Z = R R5 R A bit ), otherwise A bit 0), if Y3 R Z 3 = R3 R5 R7 A bit ), otherwise In other worlds, we use the quantization rule obtained by Lloyd Algorithm for X to quantize Y k in sensor k. If the k-th bit is 0, sensor k sends A), otherwise sends A. Finally, the fusion center combines these bits and maps back the quantization levels a i, which is used to estimate X. The simulation result is illustrated in Fig. 7, where μ =5, α =0, σ =. IV. QUANTIZATION REGION ALLOCATION As the Fig. 7 shows, the doesn t decrease monotonically while the number of sensors increase. This may be due to the fact that bits sent from various sensors are not equally weighted. Namely, the most significant bit, which decides whether the source X locates on the left of mean μ or on the right, depends on one sensor only. Though every sensor represents different bit, these bits sent by sensors are not trustworthy since the noise exists in two stages of transmission. Therefore, we must adjust our strategy. Only few bits are used in estimation. Besides, the number of sensors which a) α =0. b) α =. Fig. : for μ =5, σ =after bit allocation. are used to decide specific bit is optimized. The simulation goes as follows. We use three bits to quantize the source. Let a,a,a 3 be the number of bit,,3 repectively, where a stands for the most significant bit MSB) and a 3 represents the last significant bit LSB). Since MSB is more important, assume that a a a 3 and running simulation for all possible allocation of bits. The result is shown in Fig.. In Fig., we also attach the table for the bit allocation. It can be seen that two bits are enough for estimation. Namely, Instead of increasing the resolution, it is more important to increase the accuracy of every bits, especially the MSB. In addition, notice that the third bit is used in Fig. b. This is 7

due to the larger variance used in Fig. b than Fig. a. While dealing the source with larger variance, more bits are worthy to use. V. COMPARISON TO SOFT DECISION Since the data sent from each sensor Z k suffers from two noises, N k and N k, Z k given X is Gaussian distributed with variance σ. Therefore, the a posteriori probability of X given Z =[Z Z Z ] is f X Z x z) = f Z z) =gz)exp πσ e zk x) σ σp x σ p μ α + σ e x μ) α πα ) z k ) 5) where σp = ). σ + α Since the exponent in 5) is quadratic function of x, f X Z x z) is Gaussian. Note that the estimator under mean squared error criterion is the conditional mean of the a posteriori random variable. Thus the estimator is given by μ XZ) =σp α + ) σ Z k ) And the corresponding is E[X XZ)) ]= σ + ) α 7) This should be the optimal performance for estimating X since in soft decision, we have assumed that all of the sensors can directly transmit the unquantized data they receive, which is not true in realistic. Next, we compare the performance among optimal soft decision and the hard decision methods mensioned above. For soft decision, -FS is used to transmit quantized data at each sensor. Thus, in these hard decision methods, the received data is oversampled at each sensor and the -bit data is transmitted using BPS channels simultaneously for fair comparison. VI. CONCLUSION From Fig. 9, we can see that the performance of Σ-Δ method is almost unchanged since it has oversampled the data at each sensor and it oversamples again at the fusion center. Thus, adding more sensor has little effect on the performance. Besides, the for Σ-Δ method is lower than direct method when the source variance is 00. This is intuitively due to the fact that oversampling for Σ-Δ method has more effect on the while the variance is large. For bit allocation method, the decreases until more sensors are used. Since we let different sensors be responsible for different bits, which increases the resolution, and optimize the quantization region. The performance of bit allocation method is always better than other two hard decision methods when the number of sensors increases. Finally, proper windowing techniques may further improve the performance. 0 9 7 5 3 ΣΔ Modulation Direct Approach Bit Allocation Soft Decision 0 0 50 5 0 35 30 5 0 5 0 a) α =5. ΣΔ Modulation Direct Approach Bit Allocation Soft Decision 5 0 0 b) α =0. Fig. 9: comparison for μ =5, σ =. REFERENCES [] D. Slepian and J. Wolf, Noiseless coding of correlated information sources, IEEE Trans. Inf. Theory, vol. 9, no., pp. 7 0, Jul. 973. [] J. Garcia-Frias, Compression of correlated binary sources using turbo codes, Communications Letters, IEEE, vol. 5, no. 0, pp. 7 9, Oct 00. [3] A. Aaron and B. Girod, Compression with side information using turbo codes, in Data Compression Conference, 00. Proceedings. DCC 00, 00, pp. 5. [] A. Wyner and J. Ziv, The rate-distortion function for source coding with side information at the decoder, IEEE Trans. Inf. Theory, vol., no., pp. 0, Jan 97. [5] T. Berger, Z. Zhang, and H. Viswanathan, The ceo problem, IEEE Trans. Inf. Theory, vol., no. 3, pp. 7 90, May 99. [] V. Prabhakaran, D. Tse, and. Ramachandran, Rate region of the quadratic gaussian ceo problem, in Information Theory, 00. ISIT 00. Proceedings. International Symposium on, June 00, pp. 9. [7] W. ester, The Data Conversion Handbook. Newnes, 005. [] J. Gubner, Distributed estimation and quantization, IEEE Trans. Inf. Theory, vol. 39, no., pp. 5 59, Jul 993. [9] S. Lloyd, Least squares quantization in pcm, IEEE Trans. Inf. Theory, vol., no., pp. 9 37, Mar 9. 7