On MMSE Estimation from Quantized Observations in the Nonasymptotic Regime

Size: px
Start display at page:

Download "On MMSE Estimation from Quantized Observations in the Nonasymptotic Regime"

Transcription

1 On MMSE Estimation from Quantized Observations in the Nonasymptotic Regime Jaeho Lee, Maxim Raginsky, and Pierre Moulin Abstract This paper studies MMSE estimation on the basis of quantized noisy observations It presents nonasymptotic bounds on MMSE regret due to quantization for two settings: () estimation of a scalar random variable given a quantized vector of n conditionally independent observations, and () estimation of a p-dimensional random vector given a quantized vector of n observations (not necessarily independent) when the full MMSE estimator has a subgaussian concentration property Index Terms MMSE estimation, vector quantization, indirect rate distortion problems, nonasymptotic bounds I INTRODUCTION Minimum mean-square error (MMSE) estimation is a fundamental primitive in communications, signal processing, and data analytics In today s applications, the estimation task is often performed on high-dimensional data collected, and possibly preprocessed, at multiple remote locations Consequently, attention must be paid to communication constraints and their impact on the MMSE One strategy for reducing the communication burden is to compress the observations using vector quantization (VQ) The idea of quantization for estimation and control can be traced back to the work of Curry [, who considered in detail the jointly Gaussian case, derived a modification of the Kalman filter for use with quantized inputs, and developed approximation-based schemes for nonlinear systems Because VQ inevitably introduces loss, it is of interest to characterize the resulting MMSE regret, ie, the difference between the optimal performance achievable with quantized observations and the optimal performance that can be attained without quantization The problem of designing an optimal vector quantizer to minimize the MMSE regret is equivalent to a noisy source coding problem with the quadratic fidelity criterion, where the compressor acts on the vector of observations and the decompressor generates an estimate of the target random vector This equivalence was systematically studied by Wolf and Ziv [, who showed that there is no loss of optimality if we first compute the full MMSE estimate and then compress it using an optimal quantizer Ephraim and Gray [3 later extended this result to a more general class of weighted quadratic distortion functions and gave conditions for convergence of the Lloyd-type iterative algorithm for quantizer design The authors are with the Department of Electrical and Computer Engineering and the Coordinated Science Laboratory, University of Illinois, Urbana, IL 680, USA s: jlee60,maxim,pmoulin}@illinoisedu Research supported in part by DARPA under grant no N and in part by NSF under CAREER award no CCF-5404 The main message of the above works is that the problem of quantizer design for minimum MMSE regret is a functional compression problem: the compressed representation of X should retain as much ormation as possible about the conditional mean η(x) E[ X of the target given X, also known as the regression function However, there is an interesting tension between the ormation-theoretic and the statistical aspects of the problem: if the MMSE without quantization is sufficiently small, then the dominant contribution to the quantized MMSE should come from the quantization error; in the opposite regime of high-rate quantization, the quantized MMSE should not differ much from its unquantized counterpart In this paper, we study both the statistical and the functionalcompression aspects of MMSE estimation with quantized observation In particular, we obtain sharp upper bounds on the MMSE regret in two scenarios: MMSE estimation of a scalar random variable, where the vector X = (X,, X n ) of conditionally iid observations is passed through a k-ary vector quantizer In this setting, under mild regularity conditions on the conditional distribution P X, we obtain nonasymptotic bounds on the MMSE regret These bounds exhibit two distinct behaviors depending on whether the number of observations n is larger than the square of the codebook size k In the asymptotic regime of n, k, we recover an existing result of Samarov and Has minskii [6, who were the first to address this problem MMSE estimation of a p-dimensional random vector when the n-dimensional vector of observations X (not necessarily independent) is passed through a k-ary VQ In this setting, we derive an upper bound on the MMSE regret under the assumption that the l norm of the regression function exhibits subgaussian concentration around its expected value E η(x) Unlike some of the existing literature on high-resolution quantization for functional compression (eg, [4), we do not require smoothness of η(x) Notation We will always denote by the l (Euclidean) norm A k-ary quantizer on the Euclidean space R d is a Borelmeasurable mapping q : R d [k, where [k is shorthand for the set,, k} Any such q is characterized by its cells or bins, ie, the Borel sets C j = q (j}) = v R d : q(v) = j}, j [k The set of all k-ary quantizers on R d will be denoted by Q d k A k-ary reconstruction function on Rd is a mapping f : [k R d Any such f is characterized by its

2 reconstruction points c j = f(j) R d, j [k The set of all k-ary reconstruction functions on R d will be denoted by R d k Any finite set of points C = c,, c k } R d defines a quantizer q C Q d k and a reconstruction function f C R d k by q C (v) = arg min v c j, j [k u R d and f C (j) = c j for all j [k The composite mapping q C f C q C : R d R d is called a k-point nearest-neighbor quantizer with codebook C A well-known result is that, for any random vector V with E V <, f R d k q Q d k E V f(q(v )) = E V q C (V ), C R d : C =k and the imum is actually a minimum We refer the reader to the survey article by Gray and Neuhoff [5 for more details We use the following asymptotic order notation: for two sequences a m } and b m }, we write a m b m if a m = O(b m ), and a m b m if a m b m and b m a m II PERFORMANCE CRITERIA AND SOME BASIC RESULTS Suppose two random vectors X in R n and in R p are jointly distributed according to a given probability law P X The MMSE in estimating as a function of X is mmse(p X ) f E f(x), where the imum over all Borel-measurable functions f : R n R p is achieved by the regression function η(x) = E[ X = x We consider the problem of MMSE estimation of under the circumstances where only q(x), a quantized version of X, is accessible Thus, for each k Z, we are interested in the MMSE functional mmse k (P X ) f R p q Q n k k E f(q(x)) This problem can be cast as one-shot fixed-rate lossy coding of with squared-error distortion when the encoder only has access to X (see, eg, [, [3) We recall some known results that explicitly involve the regression function η(x) The first one, due to Wolf and Ziv [, is a useful decomposition of mmse k (P X ): Proposition For every k Z, where mmse k (P X ) = mmse(p X ) reg k (P X ), () reg k (P X ) = E E[ X E[ q(x) q Q n k is the MMSE regret due to quantization Remark A special case of this result for jointly Gaussian X and was obtained by Curry [, Sec 4 The second result, which was proved by Ephraim and Gray [3 for a more general class of weighted quadratic distortion functions, shows that there is no loss of optimality if we restrict our attention to schemes of the following type: given X, we first compute the regression function η(x), quantize it using a k-ary quantizer, and then estimate by its conditional mean given the cell index of X Proposition reg k (P X ) = f R p k q Q p k E η(x) f(q(η(x))) = C R p : C =k E η(x) q C(η(X)) III THE CASE OF n CONDITIONALL INDEPENDENT OBSERVATIONS (a) (b) We now consider the situation when the coordinates X,, X n of X are independent and identically distributed (iid) conditionally on To keep things simple, we consider the case when is a scalar random variable (ie, p = ), and its marginal distribution P has a probability density function f supported on a compact interval = [ A, A In the asymptotic regime as k and n, this setting was investigated by Samarov and Has minskii [6, who showed the following under mild regularity conditions: If n/k, then the dominant contribution to mmse k (P X ) comes from the minimum expected distortion C ; C =k E q C( ) k ( ) 3 f (y) /3 dy incurred on by any k-point quantizer [7, [8 If n/k 0, then the dominant contribution to mmse k (P X ) comes from mmse(p X ) n f (y) I(y) dy, where I(y) is the Fisher ormation in the conditional distribution P X =y [9 (see below for definitions) The proofs in [6 rely on the deep result, commonly referred to as the Bernstein von Mises theorem, which says that the posterior distribution of the normalized error ni( )(η(x) ) is asymptotically standard normal (see, eg, [0, Sec 0) In this section, we establish a finite n, finite k version of the results of Samarov Has minskii using a recent nonasymptotic generalization of the Bernstein von Mises theorem due to Spokoiny [ Our result requires a number of regularity assumptions The first two are standard: (C) The conditional distributions P X =y, y, are dominated by a common σ-finite measure µ on the real line The corresponding log-density l(u, y) log dp X =y (u) dµ is twice differentiable in y for every u We will denote the first and second derivatives with respect to y by and (C) For each y, the Fisher ormation [ exists and is positive I(y) E [l(x, ) = y

3 The remaining assumptions (see [3, Sec 5) are slightly stronger than the classical ones For each r > 0, define the local neighborhoods N r (y) y : } I(y) y y r The regularity assumptions can be split into two groups: identifiability conditions and exponential moment conditions We start with the former: (I) There are positive constants r 0 > 0 and δ > 0, such that, for every r r 0, sup sup D(P X =y P X =y) I(y)(y y ) δ r y y N r(y) (I) There is a constant b > 0, such that, for every r > 0, sup sup y y N r(y) D(P X =y P X =y) I(y)(y y ) b The exponential moment assumptions pertain to the random variables ζ y (X, y ) l(x, y ) E[l(X, y ) = y, [note that ζ y (Z, y) l(x, y) Let ξ y (X, y ) w ζ y(x, w) w=y y, y (E) There exist constants g > 0 and v 0 > 0, such that [ ( ) λξ y (X, y) sup log E exp = y v 0λ y I(y) y y N r(y) v 0λ for all λ g (E3) For every r > 0 there exists g (r) > 0, such that [ ( ) λξ y (X, y) sup sup log E exp = y I(y) for all λ g (r) v 0λ For example, in the case of additive Gaussian noise, ie, when P X =y = N(y, σ ) for all y, it is easy to verify that all of these assumptions are met We are now ready to state and prove the main result of this section Since both and η(x) are supported on the bounded interval [ A, A, there is no loss of generality in restricting our attention only to nearest-neighbor quantizers q C with codebooks C = y,, y k } [ A, A Moreover, we can assume that C is ordered in such a way that A y 0 y < y < < y k y k A Given such an ordered C, we define C max 0 j k (y j y j ) Theorem Suppose that Assumptions (C) (C), (I) (I), and (E) (E3) hold Suppose also that log f is Lipschitz on [ A, A Then there exists a constant L > 0 that depends only on the constants in the above assumptions, such that, for any k-point nearest-neighbor quantizer q C with C, we have E η(x) q C (η(x)) E q C ( ) ( [ L C min, E )} mmse(p X ) C n I( ) (3) In the additive Gaussian noise case, the bound (3) becomes E η(x) q C (η(x)) E q C ( ) L C min σ, C n }, where σ is the noise variance Proof: For any collection C = y,, y k } of k reconstruction points, define the function e C : R by e C (y) min j [k (y y j) Then a simple calculation shows that e C (y) e C (y ) min C, C y y }, (4) for all y, y The expected reconstruction error of the for all λ g nearest-neighbor quantizer q (E) There exists a constant ω C can be written as > 0, such that, for every r r 0, E η(x) q C (η(x)) = E[e C (η(x)) [ ( ) λ[ξ y (X, y ) ξ y (X, y) Using the law of iterated expectation and the smoothness sup sup log E exp y y N r(y) ω r I(y) = y estimate (4), we obtain E[e C (η(x)) E[e C ( ) E E[e C (η(x)) e C ( ) CE min, E [ η(x) } (5) C We now invoke Spokoiny s nonasymptotic Bernstein von Mises theorem [ Let Z n η(x) ( G n (X, )), where G n (X, ) n l(x i, ), ni( ) i= and for any L > 0 consider the event ( ) } ni( /4 log n A L n( ) ) Zn L, n Then there exists a choice L = L 0 that depends only on the constants in the regularity conditions, such that P [ A n ( ) C n, (6)

4 where A n ( ) A L0 n ( ), and C > 0 is an absolute constant [, Sec 43 Therefore, E [ η(x) (a) E [ Z n [ E Gn (X, ) = E [ A n ( )} Z n [ E A c n ( )} Z n E [ G n (X, ) (b) L 0 C ni( ) n E[ Z n [ E Gn (X, ) (c) L 0 C ni( ) n E[ η(x) C n E[ Gn (X, ) [ E Gn (X, ), (7) where (a) follows from the triangle inequality, (b) from (6) and Cauchy Schwarz, and (c) again from the triangle inequality Now, since E[ l(x, ) = 0 and Var[ l(x, ) = I( ) [, we have and E [ G n (X, ) = Var[Gn (X, ) = ni( ) E [ G n (X, ) ni( ) Substituting these estimates into (7) gives E [ η(x) L 0 C ni( ) n I( ) C E [ η(x) n Plugging this bound into (5), using Jensen s inequality, and simplifying, we get (3) Corollary Under the same assumptions as in Theorem, reg k(p X ) E q C( ) C ; C k ( [ min k, k E )} mmse(p X ) n I( ) (8) Remark Using the ormation inequality [9 mmse(p X ) [ n E I( ) and Jensen s inequality, the bound (8) can be weakened to reg k(p X ) E q C( ) C ; C k } min k, mmse(px ) k Proof: Let C = y,, yk } be the reconstruction points of an optimal k-point quantizer for arranged in increasing order From the work of Panter and Dite [7, we know that these points should be chosen in such a way that y j f (y) /3 dy = c j f (y) /3 dy, k y j for all j = 0,,, k, where c j = / for j = 0 and otherwise Therefore, C /k Using this and the definition of the MMSE regret in (3), we get (8) Observe that the value of the right-hand side of (8) is determined by whether the number of observations n is larger or smaller than k This agrees with the asymptotic results of Samarov and Has minskii [6 IV A HIGH-RESOLUTION BOUND FOR FUNCTIONAL COMPRESSION We now consider a general setting of p-dimensional and n-dimensional X without any conditional independence assumptions Instead, we focus on the scaling of the MMSE regret reg k (P X ) with k, while the values of n and p stay fixed Proposition shows that the MMSE regret is precisely the minimum expected distortion attainable by k-ary quantization of X in the problem of functional compression of the regression function η(x) The results we present in this section require only some regularity assumptions on the l norm η(x) : Assumption The random variable η(x) has a finite fourth moment: E η(x) 4 < Assumption The random variable η(x) is subgaussian: there exists a positive constant v > 0, such that log E [e λ( η(x) E η(x) ) vλ, λ R One sufficient condition for Assumption to hold is for to have a finite fourth moment Indeed, in that case, using Jensen s inequality, we have E η(x) 4 = E E[ X 4 E 4 < As for Assumption, it will be met, for example, if the regression function η is Lipschitz, ie, if there exists some finite constant L > 0, such that η(x) η(x ) L x x, x, x R n and if X is a Gaussian random vector with a nonsingular covariance matrix [4 Theorem Suppose Assumption holds Then reg k (P X ) ( E η(x) E η(x) 4) /3 k /3p (9) If Assumption also holds, then reg k (P X ) r>e η(x) r k /p E η(x) 4 e (r E η(x) ) /4v } (0) Remark 3 Here, p can be replaced by a suitable intrinsic dimension of the support of η(x) (eg, the rate-distortion dimension [5) For example, if η(x) is linear, ie, E[ X =

5 AX for some deterministic matrix A R p n, then we can replace p by rank(a) Moreover, using a suboptimal value of r, we can weaken the bound in (0) to reg k (P X ) log k k /p, where the hidden constant depends on p, on the first and fourth moments of η(x), and on the subgaussian constant v Apart from the logarithmic factor, this scaling of the MMSE regret agrees with the high-resolution approximation for VQ [5 and with the Shannon lower bound [5, [6 Proof: From Assumption and from Jensen s inequality, it follows that the first and second moments of η(x) are also finite Fix a positive real constant r > 0, which will be optimized later A simple volumetric estimate shows that the l ball of radius r in R p can be covered by at most ( ) r p ɛ balls of radius ɛ So, for a given k Z, we can cover the radius-r ball by k balls of radius ɛ rk /p Let y i } k i= be the centers of these k balls We now construct a quantizer q (r) Q p k as follows: arg min η(x) y j, if η(x) r q (r) (x) = j [k k, otherwise For this quantizer, we have E η(x) E[η(X) q (r) (X) = T T, where T E [ η(x) r} η(x) E[η(X) q (r) (X) [ T E η(x) > r} η(x) E[η(X) q (r) (X) For the first term, we have T 4ɛ P [ η(x) r 4ɛ r k /p, where the first inequality is true since η(x) and E[η(X) q (r) (X) are inside the same ɛ-ball in the covering for all X such that q (r) (X) [k For the second term, T P [ η(x) > r E η(x) E[η(X) q (r) (X) 4 P[ η(x) > r E( η(x) E[η(X) q (r) (X) ) 4 P[ η(x) > r 8 E η(x) 4 8 E E[η(X) q (r) 4 P[ η(x) > r 6 E η(x) 4, where the first line is by Cauchy Schwarz inequality, and the remaining steps follow from monotonicity and convexity By Markov s inequality, P[ η(x) > r = P[ η(x) > r E η(x) r Therefore, reg k (P X ) E η(x) E[η(X) q (r) r k /p E η(x) E η(x) 4 r Optimizing over all r > 0, we get (9) Now suppose Assumption also holds Let r = E η(x) t for some t > 0 Since η(x) is subgaussian, the Chernoff bounding technique gives P [ η(x) > r = P [ η(x) E η(x) > t e t /v = e (r E η(x) ) /v (see, eg, [4, Chap 3) Thus, we have reg k (P X ) E η(x) E[η(X) q (r) r k /p E η(x) 4 e (r E η(x) ) /4v for all r > E η(x) Optimizing over r, we get (0) ACKNOWLEDGMENTS The authors would like to thank Tamás Linder and Vladimir Spokoiny for helpful discussions REFERENCES [ R E Curry, Estimation and Control with Quantized Measurements, Cambridge, MA: MIT Press, 970 [ J K Wolf and J Ziv, Transmission of noisy ormation to a noisy receiver with minimum distortion, IEEE Trans Inform Theory, vol 6, no 4, pp 406-4, July 970 [3 Ephraim and R M Gray, A unified approach for encoding clean and noisy sources by means of waveform and autoregressive model vector quantization, IEEE Trans Inform Theory, vol 34, no 4, pp , July 988 [4 V Misra, V K Goyal, and L R Varshney, Distributed scalar quantization for computing: High-resolution analysis and extensions, IEEE Trans Inform Theory, vol 57, no 8, pp , August 0 [5 R M Gray and D L Neuhoff, Quantization, IEEE Trans Inform Theory, vol 44, no 6, October 998 [6 A M Samarov and R Z Has minskii, Estimation by means of statistics that take a finite number of values, Problems of Inform Transmission, vol 3, no 4, pp 60-65, 977 [7 P F Panter and W Dite, Quantisation distortion in pulse-count modulation with nonuniform spacing of levels, Proc IRE, vol 39, pp 44-48, January 95 [8 P Elias, Bounds on performance of optimum quantizers, IEEE Trans Inform Theory, vol 6, no, pp 7-84, March 970 [9 I A Ibragimov and R Z Hasminskii, On ormation inequalities and superefficient estimates, Problems of Inform Transmission, vol 9, no 3, pp 6-7, 973 [0 A W van der Vaart, Asymptotic Statistics, Cambridge University Press, 998 [ V Spokoiny, Bernstein von Mises theorem for growing parameter dimension, arxiv preprint , 03 [ I A Ibragimov and R Z Has minskii, Statistical Estimation: Asymptotic Theory, Springer, New ork, 98 [3 V Spokoiny, Parametric estimation Finite sample theory, Ann Statist, vol 40, no 6, pp , 0 [4 M Raginsky and I Sason, Concentration of Measure Inequalities in Information Theory, Communications, and Coding, nd ed, Now Publishers, 04 [5 T Kawabata, and A Dembo, The rate-distortion dimension of sets and measures, IEEE Trans Inform Theory, vol 40, no 5, pp , September 994 [6 T Linder and R Zamir, On the asymptotic tightness of the Shannon lower bound, IEEE Trans Inform Theory, vol 40, no 6, pp 06-03, November 994

Afundamental component in the design and analysis of

Afundamental component in the design and analysis of IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 45, NO. 2, MARCH 1999 533 High-Resolution Source Coding for Non-Difference Distortion Measures: The Rate-Distortion Function Tamás Linder, Member, IEEE, Ram

More information

Lattices for Distributed Source Coding: Jointly Gaussian Sources and Reconstruction of a Linear Function

Lattices for Distributed Source Coding: Jointly Gaussian Sources and Reconstruction of a Linear Function Lattices for Distributed Source Coding: Jointly Gaussian Sources and Reconstruction of a Linear Function Dinesh Krithivasan and S. Sandeep Pradhan Department of Electrical Engineering and Computer Science,

More information

QUANTIZATION FOR DISTRIBUTED ESTIMATION IN LARGE SCALE SENSOR NETWORKS

QUANTIZATION FOR DISTRIBUTED ESTIMATION IN LARGE SCALE SENSOR NETWORKS QUANTIZATION FOR DISTRIBUTED ESTIMATION IN LARGE SCALE SENSOR NETWORKS Parvathinathan Venkitasubramaniam, Gökhan Mergen, Lang Tong and Ananthram Swami ABSTRACT We study the problem of quantization for

More information

Design of Optimal Quantizers for Distributed Source Coding

Design of Optimal Quantizers for Distributed Source Coding Design of Optimal Quantizers for Distributed Source Coding David Rebollo-Monedero, Rui Zhang and Bernd Girod Information Systems Laboratory, Electrical Eng. Dept. Stanford University, Stanford, CA 94305

More information

Sequential prediction with coded side information under logarithmic loss

Sequential prediction with coded side information under logarithmic loss under logarithmic loss Yanina Shkel Department of Electrical Engineering Princeton University Princeton, NJ 08544, USA Maxim Raginsky Department of Electrical and Computer Engineering Coordinated Science

More information

Lossy Compression Coding Theorems for Arbitrary Sources

Lossy Compression Coding Theorems for Arbitrary Sources Lossy Compression Coding Theorems for Arbitrary Sources Ioannis Kontoyiannis U of Cambridge joint work with M. Madiman, M. Harrison, J. Zhang Beyond IID Workshop, Cambridge, UK July 23, 2018 Outline Motivation

More information

On Scalable Coding in the Presence of Decoder Side Information

On Scalable Coding in the Presence of Decoder Side Information On Scalable Coding in the Presence of Decoder Side Information Emrah Akyol, Urbashi Mitra Dep. of Electrical Eng. USC, CA, US Email: {eakyol, ubli}@usc.edu Ertem Tuncel Dep. of Electrical Eng. UC Riverside,

More information

arxiv: v1 [cs.it] 21 Feb 2013

arxiv: v1 [cs.it] 21 Feb 2013 q-ary Compressive Sensing arxiv:30.568v [cs.it] Feb 03 Youssef Mroueh,, Lorenzo Rosasco, CBCL, CSAIL, Massachusetts Institute of Technology LCSL, Istituto Italiano di Tecnologia and IIT@MIT lab, Istituto

More information

Estimation of Intrinsic Dimensionality Using High-Rate Vector Quantization

Estimation of Intrinsic Dimensionality Using High-Rate Vector Quantization Estimation of Intrinsic Dimensionality Using High-Rate Vector Quantization Maxim Raginsky and Svetlana Lazebnik Beckman Institute, University of Illinois 4 N Mathews Ave, Urbana, IL {maxim,slazebni}@uiuc.edu

More information

Randomized Quantization and Optimal Design with a Marginal Constraint

Randomized Quantization and Optimal Design with a Marginal Constraint Randomized Quantization and Optimal Design with a Marginal Constraint Naci Saldi, Tamás Linder, Serdar Yüksel Department of Mathematics and Statistics, Queen s University, Kingston, ON, Canada Email: {nsaldi,linder,yuksel}@mast.queensu.ca

More information

Lecture 5 Channel Coding over Continuous Channels

Lecture 5 Channel Coding over Continuous Channels Lecture 5 Channel Coding over Continuous Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 14, 2014 1 / 34 I-Hsiang Wang NIT Lecture 5 From

More information

Example: for source

Example: for source Nonuniform scalar quantizer References: Sayood Chap. 9, Gersho and Gray, Chap.'s 5 and 6. The basic idea: For a nonuniform source density, put smaller cells and levels where the density is larger, thereby

More information

MMSE DECODING FOR ANALOG JOINT SOURCE CHANNEL CODING USING MONTE CARLO IMPORTANCE SAMPLING

MMSE DECODING FOR ANALOG JOINT SOURCE CHANNEL CODING USING MONTE CARLO IMPORTANCE SAMPLING MMSE DECODING FOR ANALOG JOINT SOURCE CHANNEL CODING USING MONTE CARLO IMPORTANCE SAMPLING Yichuan Hu (), Javier Garcia-Frias () () Dept. of Elec. and Comp. Engineering University of Delaware Newark, DE

More information

Quantization. Introduction. Roadmap. Optimal Quantizer Uniform Quantizer Non Uniform Quantizer Rate Distorsion Theory. Source coding.

Quantization. Introduction. Roadmap. Optimal Quantizer Uniform Quantizer Non Uniform Quantizer Rate Distorsion Theory. Source coding. Roadmap Quantization Optimal Quantizer Uniform Quantizer Non Uniform Quantizer Rate Distorsion Theory Source coding 2 Introduction 4 1 Lossy coding Original source is discrete Lossless coding: bit rate

More information

Lecture 7 Introduction to Statistical Decision Theory

Lecture 7 Introduction to Statistical Decision Theory Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7

More information

Joint Universal Lossy Coding and Identification of Stationary Mixing Sources With General Alphabets Maxim Raginsky, Member, IEEE

Joint Universal Lossy Coding and Identification of Stationary Mixing Sources With General Alphabets Maxim Raginsky, Member, IEEE IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 5, MAY 2009 1945 Joint Universal Lossy Coding and Identification of Stationary Mixing Sources With General Alphabets Maxim Raginsky, Member, IEEE Abstract

More information

A new converse in rate-distortion theory

A new converse in rate-distortion theory A new converse in rate-distortion theory Victoria Kostina, Sergio Verdú Dept. of Electrical Engineering, Princeton University, NJ, 08544, USA Abstract This paper shows new finite-blocklength converse bounds

More information

The Duality Between Information Embedding and Source Coding With Side Information and Some Applications

The Duality Between Information Embedding and Source Coding With Side Information and Some Applications IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 49, NO. 5, MAY 2003 1159 The Duality Between Information Embedding and Source Coding With Side Information and Some Applications Richard J. Barron, Member,

More information

Mathematical Institute, University of Utrecht. The problem of estimating the mean of an observed Gaussian innite-dimensional vector

Mathematical Institute, University of Utrecht. The problem of estimating the mean of an observed Gaussian innite-dimensional vector On Minimax Filtering over Ellipsoids Eduard N. Belitser and Boris Y. Levit Mathematical Institute, University of Utrecht Budapestlaan 6, 3584 CD Utrecht, The Netherlands The problem of estimating the mean

More information

Information theoretic perspectives on learning algorithms

Information theoretic perspectives on learning algorithms Information theoretic perspectives on learning algorithms Varun Jog University of Wisconsin - Madison Departments of ECE and Mathematics Shannon Channel Hangout! May 8, 2018 Jointly with Adrian Tovar-Lopez

More information

March 1, Florida State University. Concentration Inequalities: Martingale. Approach and Entropy Method. Lizhe Sun and Boning Yang.

March 1, Florida State University. Concentration Inequalities: Martingale. Approach and Entropy Method. Lizhe Sun and Boning Yang. Florida State University March 1, 2018 Framework 1. (Lizhe) Basic inequalities Chernoff bounding Review for STA 6448 2. (Lizhe) Discrete-time martingales inequalities via martingale approach 3. (Boning)

More information

Random Bernstein-Markov factors

Random Bernstein-Markov factors Random Bernstein-Markov factors Igor Pritsker and Koushik Ramachandran October 20, 208 Abstract For a polynomial P n of degree n, Bernstein s inequality states that P n n P n for all L p norms on the unit

More information

Local Asymptotics and the Minimum Description Length

Local Asymptotics and the Minimum Description Length Local Asymptotics and the Minimum Description Length Dean P. Foster and Robert A. Stine Department of Statistics The Wharton School of the University of Pennsylvania Philadelphia, PA 19104-6302 March 27,

More information

1 Introduction One of the central problems in lossy data compression concerns designing vector quantizers from training data. The following question i

1 Introduction One of the central problems in lossy data compression concerns designing vector quantizers from training data. The following question i On the Performance of Vector Quantizers Empirically Designed From Dependent Sources Assaf J. Zeevi Information Systems Lab Stanford University Stanford, CA. 94305-9510 Abstract Suppose we are given n real

More information

Lecture 22: Variance and Covariance

Lecture 22: Variance and Covariance EE5110 : Probability Foundations for Electrical Engineers July-November 2015 Lecture 22: Variance and Covariance Lecturer: Dr. Krishna Jagannathan Scribes: R.Ravi Kiran In this lecture we will introduce

More information

Review of Quantization. Quantization. Bring in Probability Distribution. L-level Quantization. Uniform partition

Review of Quantization. Quantization. Bring in Probability Distribution. L-level Quantization. Uniform partition Review of Quantization UMCP ENEE631 Slides (created by M.Wu 004) Quantization UMCP ENEE631 Slides (created by M.Wu 001/004) L-level Quantization Minimize errors for this lossy process What L values to

More information

3 Integration and Expectation

3 Integration and Expectation 3 Integration and Expectation 3.1 Construction of the Lebesgue Integral Let (, F, µ) be a measure space (not necessarily a probability space). Our objective will be to define the Lebesgue integral R fdµ

More information

Compressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery

Compressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery Compressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery Jorge F. Silva and Eduardo Pavez Department of Electrical Engineering Information and Decision Systems Group Universidad

More information

THIS paper addresses the quadratic Gaussian two-encoder

THIS paper addresses the quadratic Gaussian two-encoder 1938 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL 54, NO 5, MAY 2008 Rate Region of the Quadratic Gaussian Two-Encoder Source-Coding Problem Aaron B Wagner, Member, IEEE, Saurabha Tavildar, Pramod Viswanath,

More information

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis. Vector spaces DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Vector space Consists of: A set V A scalar

More information

6196 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 9, SEPTEMBER 2011

6196 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 9, SEPTEMBER 2011 6196 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 9, SEPTEMBER 2011 On the Structure of Real-Time Encoding and Decoding Functions in a Multiterminal Communication System Ashutosh Nayyar, Student

More information

On Acceleration with Noise-Corrupted Gradients. + m k 1 (x). By the definition of Bregman divergence:

On Acceleration with Noise-Corrupted Gradients. + m k 1 (x). By the definition of Bregman divergence: A Omitted Proofs from Section 3 Proof of Lemma 3 Let m x) = a i On Acceleration with Noise-Corrupted Gradients fxi ), u x i D ψ u, x 0 ) denote the function under the minimum in the lower bound By Proposition

More information

On Competitive Prediction and Its Relation to Rate-Distortion Theory

On Competitive Prediction and Its Relation to Rate-Distortion Theory IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 49, NO. 12, DECEMBER 2003 3185 On Competitive Prediction and Its Relation to Rate-Distortion Theory Tsachy Weissman, Member, IEEE, and Neri Merhav, Fellow,

More information

MMSE Dimension. snr. 1 We use the following asymptotic notation: f(x) = O (g(x)) if and only

MMSE Dimension. snr. 1 We use the following asymptotic notation: f(x) = O (g(x)) if and only MMSE Dimension Yihong Wu Department of Electrical Engineering Princeton University Princeton, NJ 08544, USA Email: yihongwu@princeton.edu Sergio Verdú Department of Electrical Engineering Princeton University

More information

5682 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 12, DECEMBER /$ IEEE

5682 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 12, DECEMBER /$ IEEE 5682 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 12, DECEMBER 2009 Hyperplane-Based Vector Quantization for Distributed Estimation in Wireless Sensor Networks Jun Fang, Member, IEEE, and Hongbin

More information

Uniform Convergence of a Multilevel Energy-based Quantization Scheme

Uniform Convergence of a Multilevel Energy-based Quantization Scheme Uniform Convergence of a Multilevel Energy-based Quantization Scheme Maria Emelianenko 1 and Qiang Du 1 Pennsylvania State University, University Park, PA 16803 emeliane@math.psu.edu and qdu@math.psu.edu

More information

Empirical Risk Minimization as Parameter Choice Rule for General Linear Regularization Methods

Empirical Risk Minimization as Parameter Choice Rule for General Linear Regularization Methods Empirical Risk Minimization as Parameter Choice Rule for General Linear Regularization Methods Frank Werner 1 Statistical Inverse Problems in Biophysics Group Max Planck Institute for Biophysical Chemistry,

More information

EE368B Image and Video Compression

EE368B Image and Video Compression EE368B Image and Video Compression Homework Set #2 due Friday, October 20, 2000, 9 a.m. Introduction The Lloyd-Max quantizer is a scalar quantizer which can be seen as a special case of a vector quantizer

More information

Channel Polarization and Blackwell Measures

Channel Polarization and Blackwell Measures Channel Polarization Blackwell Measures Maxim Raginsky Abstract The Blackwell measure of a binary-input channel (BIC is the distribution of the posterior probability of 0 under the uniform input distribution

More information

1590 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 48, NO. 6, JUNE Source Coding, Large Deviations, and Approximate Pattern Matching

1590 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 48, NO. 6, JUNE Source Coding, Large Deviations, and Approximate Pattern Matching 1590 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 48, NO. 6, JUNE 2002 Source Coding, Large Deviations, and Approximate Pattern Matching Amir Dembo and Ioannis Kontoyiannis, Member, IEEE Invited Paper

More information

2 Statistical Estimation: Basic Concepts

2 Statistical Estimation: Basic Concepts Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:

More information

D I S C U S S I O N P A P E R

D I S C U S S I O N P A P E R I N S T I T U T D E S T A T I S T I Q U E B I O S T A T I S T I Q U E E T S C I E N C E S A C T U A R I E L L E S ( I S B A ) UNIVERSITÉ CATHOLIQUE DE LOUVAIN D I S C U S S I O N P A P E R 2014/06 Adaptive

More information

Optimal Linear Estimation under Unknown Nonlinear Transform

Optimal Linear Estimation under Unknown Nonlinear Transform Optimal Linear Estimation under Unknown Nonlinear Transform Xinyang Yi The University of Texas at Austin yixy@utexas.edu Constantine Caramanis The University of Texas at Austin constantine@utexas.edu Zhaoran

More information

Pointwise Redundancy in Lossy Data Compression and Universal Lossy Data Compression

Pointwise Redundancy in Lossy Data Compression and Universal Lossy Data Compression Pointwise Redundancy in Lossy Data Compression and Universal Lossy Data Compression I. Kontoyiannis To appear, IEEE Transactions on Information Theory, Jan. 2000 Last revision, November 21, 1999 Abstract

More information

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions International Journal of Control Vol. 00, No. 00, January 2007, 1 10 Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions I-JENG WANG and JAMES C.

More information

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,

More information

On Side-Informed Coding of Noisy Sensor Observations

On Side-Informed Coding of Noisy Sensor Observations On Side-Informed Coding of Noisy Sensor Observations Chao Yu and Gaurav Sharma C Dept, University of Rochester, Rochester NY 14627 ABSTRACT In this paper, we consider the problem of side-informed coding

More information

arxiv: v1 [cs.it] 20 Jan 2018

arxiv: v1 [cs.it] 20 Jan 2018 1 Analog-to-Digital Compression: A New Paradigm for Converting Signals to Bits Alon Kipnis, Yonina C. Eldar and Andrea J. Goldsmith fs arxiv:181.6718v1 [cs.it] Jan 18 X(t) sampler smp sec encoder R[ bits

More information

Sparse Regression Codes for Multi-terminal Source and Channel Coding

Sparse Regression Codes for Multi-terminal Source and Channel Coding Sparse Regression Codes for Multi-terminal Source and Channel Coding Ramji Venkataramanan Yale University Sekhar Tatikonda Allerton 2012 1 / 20 Compression with Side-Information X Encoder Rate R Decoder

More information

certain class of distributions, any SFQ can be expressed as a set of thresholds on the sufficient statistic. For distributions

certain class of distributions, any SFQ can be expressed as a set of thresholds on the sufficient statistic. For distributions Score-Function Quantization for Distributed Estimation Parvathinathan Venkitasubramaniam and Lang Tong School of Electrical and Computer Engineering Cornell University Ithaca, NY 4853 Email: {pv45, lt35}@cornell.edu

More information

ON SCALABLE CODING OF HIDDEN MARKOV SOURCES. Mehdi Salehifar, Tejaswi Nanjundaswamy, and Kenneth Rose

ON SCALABLE CODING OF HIDDEN MARKOV SOURCES. Mehdi Salehifar, Tejaswi Nanjundaswamy, and Kenneth Rose ON SCALABLE CODING OF HIDDEN MARKOV SOURCES Mehdi Salehifar, Tejaswi Nanjundaswamy, and Kenneth Rose Department of Electrical and Computer Engineering University of California, Santa Barbara, CA, 93106

More information

SHARED INFORMATION. Prakash Narayan with. Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe

SHARED INFORMATION. Prakash Narayan with. Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe SHARED INFORMATION Prakash Narayan with Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe 2/41 Outline Two-terminal model: Mutual information Operational meaning in: Channel coding: channel

More information

Statistics 300B Winter 2018 Final Exam Due 24 Hours after receiving it

Statistics 300B Winter 2018 Final Exam Due 24 Hours after receiving it Statistics 300B Winter 08 Final Exam Due 4 Hours after receiving it Directions: This test is open book and open internet, but must be done without consulting other students. Any consultation of other students

More information

The Moment Method; Convex Duality; and Large/Medium/Small Deviations

The Moment Method; Convex Duality; and Large/Medium/Small Deviations Stat 928: Statistical Learning Theory Lecture: 5 The Moment Method; Convex Duality; and Large/Medium/Small Deviations Instructor: Sham Kakade The Exponential Inequality and Convex Duality The exponential

More information

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions Pattern Recognition and Machine Learning Chapter 2: Probability Distributions Cécile Amblard Alex Kläser Jakob Verbeek October 11, 27 Probability Distributions: General Density Estimation: given a finite

More information

Towards stability and optimality in stochastic gradient descent

Towards stability and optimality in stochastic gradient descent Towards stability and optimality in stochastic gradient descent Panos Toulis, Dustin Tran and Edoardo M. Airoldi August 26, 2016 Discussion by Ikenna Odinaka Duke University Outline Introduction 1 Introduction

More information

L p Functions. Given a measure space (X, µ) and a real number p [1, ), recall that the L p -norm of a measurable function f : X R is defined by

L p Functions. Given a measure space (X, µ) and a real number p [1, ), recall that the L p -norm of a measurable function f : X R is defined by L p Functions Given a measure space (, µ) and a real number p [, ), recall that the L p -norm of a measurable function f : R is defined by f p = ( ) /p f p dµ Note that the L p -norm of a function f may

More information

LARGE DEVIATIONS OF TYPICAL LINEAR FUNCTIONALS ON A CONVEX BODY WITH UNCONDITIONAL BASIS. S. G. Bobkov and F. L. Nazarov. September 25, 2011

LARGE DEVIATIONS OF TYPICAL LINEAR FUNCTIONALS ON A CONVEX BODY WITH UNCONDITIONAL BASIS. S. G. Bobkov and F. L. Nazarov. September 25, 2011 LARGE DEVIATIONS OF TYPICAL LINEAR FUNCTIONALS ON A CONVEX BODY WITH UNCONDITIONAL BASIS S. G. Bobkov and F. L. Nazarov September 25, 20 Abstract We study large deviations of linear functionals on an isotropic

More information

( 1 k "information" I(X;Y) given by Y about X)

( 1 k information I(X;Y) given by Y about X) SUMMARY OF SHANNON DISTORTION-RATE THEORY Consider a stationary source X with f (x) as its th-order pdf. Recall the following OPTA function definitions: δ(,r) = least dist'n of -dim'l fixed-rate VQ's w.

More information

IEOR 265 Lecture 3 Sparse Linear Regression

IEOR 265 Lecture 3 Sparse Linear Regression IOR 65 Lecture 3 Sparse Linear Regression 1 M Bound Recall from last lecture that the reason we are interested in complexity measures of sets is because of the following result, which is known as the M

More information

A New Achievable Region for Gaussian Multiple Descriptions Based on Subset Typicality

A New Achievable Region for Gaussian Multiple Descriptions Based on Subset Typicality 0 IEEE Information Theory Workshop A New Achievable Region for Gaussian Multiple Descriptions Based on Subset Typicality Kumar Viswanatha, Emrah Akyol and Kenneth Rose ECE Department, University of California

More information

A New Metaconverse and Outer Region for Finite-Blocklength MACs

A New Metaconverse and Outer Region for Finite-Blocklength MACs A New Metaconverse Outer Region for Finite-Blocklength MACs Pierre Moulin Dept of Electrical Computer Engineering University of Illinois at Urbana-Champaign Urbana, IL 680 Abstract An outer rate region

More information

Multimedia Communications. Scalar Quantization

Multimedia Communications. Scalar Quantization Multimedia Communications Scalar Quantization Scalar Quantization In many lossy compression applications we want to represent source outputs using a small number of code words. Process of representing

More information

On the interior of the simplex, we have the Hessian of d(x), Hd(x) is diagonal with ith. µd(w) + w T c. minimize. subject to w T 1 = 1,

On the interior of the simplex, we have the Hessian of d(x), Hd(x) is diagonal with ith. µd(w) + w T c. minimize. subject to w T 1 = 1, Math 30 Winter 05 Solution to Homework 3. Recognizing the convexity of g(x) := x log x, from Jensen s inequality we get d(x) n x + + x n n log x + + x n n where the equality is attained only at x = (/n,...,

More information

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Lesson 7 Delta Modulation and DPCM Instructional Objectives At the end of this lesson, the students should be able to: 1. Describe a lossy predictive coding scheme.

More information

Concentration behavior of the penalized least squares estimator

Concentration behavior of the penalized least squares estimator Concentration behavior of the penalized least squares estimator Penalized least squares behavior arxiv:1511.08698v2 [math.st] 19 Oct 2016 Alan Muro and Sara van de Geer {muro,geer}@stat.math.ethz.ch Seminar

More information

Optimal Decentralized Control of Coupled Subsystems With Control Sharing

Optimal Decentralized Control of Coupled Subsystems With Control Sharing IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 58, NO. 9, SEPTEMBER 2013 2377 Optimal Decentralized Control of Coupled Subsystems With Control Sharing Aditya Mahajan, Member, IEEE Abstract Subsystems that

More information

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 54, NO. 7, JULY

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 54, NO. 7, JULY IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 54, NO. 7, JULY 2008 3059 Joint Fixed-Rate Universal Lossy Coding Identification of Continuous-Alphabet Memoryless Sources Maxim Raginsky, Member, IEEE Abstract

More information

THE INVERSE FUNCTION THEOREM

THE INVERSE FUNCTION THEOREM THE INVERSE FUNCTION THEOREM W. PATRICK HOOPER The implicit function theorem is the following result: Theorem 1. Let f be a C 1 function from a neighborhood of a point a R n into R n. Suppose A = Df(a)

More information

Nonparametric regression with martingale increment errors

Nonparametric regression with martingale increment errors S. Gaïffas (LSTA - Paris 6) joint work with S. Delattre (LPMA - Paris 7) work in progress Motivations Some facts: Theoretical study of statistical algorithms requires stationary and ergodicity. Concentration

More information

Statistical Data Analysis

Statistical Data Analysis DS-GA 0 Lecture notes 8 Fall 016 1 Descriptive statistics Statistical Data Analysis In this section we consider the problem of analyzing a set of data. We describe several techniques for visualizing the

More information

Information Embedding meets Distributed Control

Information Embedding meets Distributed Control Information Embedding meets Distributed Control Pulkit Grover, Aaron B Wagner and Anant Sahai Abstract We consider the problem of information embedding where the encoder modifies a white Gaussian host

More information

Summary and discussion of: Dropout Training as Adaptive Regularization

Summary and discussion of: Dropout Training as Adaptive Regularization Summary and discussion of: Dropout Training as Adaptive Regularization Statistics Journal Club, 36-825 Kirstin Early and Calvin Murdock November 21, 2014 1 Introduction Multi-layered (i.e. deep) artificial

More information

Vector Quantization. Institut Mines-Telecom. Marco Cagnazzo, MN910 Advanced Compression

Vector Quantization. Institut Mines-Telecom. Marco Cagnazzo, MN910 Advanced Compression Institut Mines-Telecom Vector Quantization Marco Cagnazzo, cagnazzo@telecom-paristech.fr MN910 Advanced Compression 2/66 19.01.18 Institut Mines-Telecom Vector Quantization Outline Gain-shape VQ 3/66 19.01.18

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

Stability and Sensitivity of the Capacity in Continuous Channels. Malcolm Egan

Stability and Sensitivity of the Capacity in Continuous Channels. Malcolm Egan Stability and Sensitivity of the Capacity in Continuous Channels Malcolm Egan Univ. Lyon, INSA Lyon, INRIA 2019 European School of Information Theory April 18, 2019 1 / 40 Capacity of Additive Noise Models

More information

Rényi Information Dimension: Fundamental Limits of Almost Lossless Analog Compression

Rényi Information Dimension: Fundamental Limits of Almost Lossless Analog Compression IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 56, NO. 8, AUGUST 2010 3721 Rényi Information Dimension: Fundamental Limits of Almost Lossless Analog Compression Yihong Wu, Student Member, IEEE, Sergio Verdú,

More information

VECTOR QUANTIZATION TECHNIQUES FOR MULTIPLE-ANTENNA CHANNEL INFORMATION FEEDBACK

VECTOR QUANTIZATION TECHNIQUES FOR MULTIPLE-ANTENNA CHANNEL INFORMATION FEEDBACK VECTOR QUANTIZATION TECHNIQUES FOR MULTIPLE-ANTENNA CHANNEL INFORMATION FEEDBACK June Chul Roh and Bhaskar D. Rao Department of Electrical and Computer Engineering University of California, San Diego La

More information

Decentralized Stochastic Control with Partial Sharing Information Structures: A Common Information Approach

Decentralized Stochastic Control with Partial Sharing Information Structures: A Common Information Approach Decentralized Stochastic Control with Partial Sharing Information Structures: A Common Information Approach 1 Ashutosh Nayyar, Aditya Mahajan and Demosthenis Teneketzis Abstract A general model of decentralized

More information

Optimization methods

Optimization methods Lecture notes 3 February 8, 016 1 Introduction Optimization methods In these notes we provide an overview of a selection of optimization methods. We focus on methods which rely on first-order information,

More information

j=1 r 1 x 1 x n. r m r j (x) r j r j (x) r j (x). r j x k

j=1 r 1 x 1 x n. r m r j (x) r j r j (x) r j (x). r j x k Maria Cameron Nonlinear Least Squares Problem The nonlinear least squares problem arises when one needs to find optimal set of parameters for a nonlinear model given a large set of data The variables x,,

More information

Math 5210, Definitions and Theorems on Metric Spaces

Math 5210, Definitions and Theorems on Metric Spaces Math 5210, Definitions and Theorems on Metric Spaces Let (X, d) be a metric space. We will use the following definitions (see Rudin, chap 2, particularly 2.18) 1. Let p X and r R, r > 0, The ball of radius

More information

Clustering by Mixture Models. General background on clustering Example method: k-means Mixture model based clustering Model estimation

Clustering by Mixture Models. General background on clustering Example method: k-means Mixture model based clustering Model estimation Clustering by Mixture Models General bacground on clustering Example method: -means Mixture model based clustering Model estimation 1 Clustering A basic tool in data mining/pattern recognition: Divide

More information

2. PROBLEM SETUP. <. The system gains {A t } t 0

2. PROBLEM SETUP. <. The system gains {A t } t 0 Rate-limited control of systems with uncertain gain Victoria Kostina, Yuval Peres, Miklós Z. Rácz, Gireeja Ranade California Institute of Technology, Pasadena, C Microsoft Research, Redmond, W vkostina@caltech.edu,

More information

PILCO: A Model-Based and Data-Efficient Approach to Policy Search

PILCO: A Model-Based and Data-Efficient Approach to Policy Search PILCO: A Model-Based and Data-Efficient Approach to Policy Search (M.P. Deisenroth and C.E. Rasmussen) CSC2541 November 4, 2016 PILCO Graphical Model PILCO Probabilistic Inference for Learning COntrol

More information

Channels with cost constraints: strong converse and dispersion

Channels with cost constraints: strong converse and dispersion Channels with cost constraints: strong converse and dispersion Victoria Kostina, Sergio Verdú Dept. of Electrical Engineering, Princeton University, NJ 08544, USA Abstract This paper shows the strong converse

More information

Functional Analysis. Franck Sueur Metric spaces Definitions Completeness Compactness Separability...

Functional Analysis. Franck Sueur Metric spaces Definitions Completeness Compactness Separability... Functional Analysis Franck Sueur 2018-2019 Contents 1 Metric spaces 1 1.1 Definitions........................................ 1 1.2 Completeness...................................... 3 1.3 Compactness......................................

More information

Semiparametric posterior limits

Semiparametric posterior limits Statistics Department, Seoul National University, Korea, 2012 Semiparametric posterior limits for regular and some irregular problems Bas Kleijn, KdV Institute, University of Amsterdam Based on collaborations

More information

Linear Regression and Its Applications

Linear Regression and Its Applications Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start

More information

A Systematic Description of Source Significance Information

A Systematic Description of Source Significance Information A Systematic Description of Source Significance Information Norbert Goertz Institute for Digital Communications School of Engineering and Electronics The University of Edinburgh Mayfield Rd., Edinburgh

More information

arxiv: v4 [cs.it] 17 Oct 2015

arxiv: v4 [cs.it] 17 Oct 2015 Upper Bounds on the Relative Entropy and Rényi Divergence as a Function of Total Variation Distance for Finite Alphabets Igal Sason Department of Electrical Engineering Technion Israel Institute of Technology

More information

Bayesian Aggregation for Extraordinarily Large Dataset

Bayesian Aggregation for Extraordinarily Large Dataset Bayesian Aggregation for Extraordinarily Large Dataset Guang Cheng 1 Department of Statistics Purdue University www.science.purdue.edu/bigdata Department Seminar Statistics@LSE May 19, 2017 1 A Joint Work

More information

Econometrics I, Estimation

Econometrics I, Estimation Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the

More information

PCA with random noise. Van Ha Vu. Department of Mathematics Yale University

PCA with random noise. Van Ha Vu. Department of Mathematics Yale University PCA with random noise Van Ha Vu Department of Mathematics Yale University An important problem that appears in various areas of applied mathematics (in particular statistics, computer science and numerical

More information

PROOF OF ZADOR-GERSHO THEOREM

PROOF OF ZADOR-GERSHO THEOREM ZADOR-GERSHO THEOREM FOR VARIABLE-RATE VQ For a stationary source and large R, the least distortion of k-dim'l VQ with nth-order entropy coding and rate R or less is δ(k,n,r) m k * σ 2 η kn 2-2R = Z(k,n,R)

More information

ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016

ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 ECE598: Information-theoretic methods in high-dimensional statistics Spring 06 Lecture : Mutual Information Method Lecturer: Yihong Wu Scribe: Jaeho Lee, Mar, 06 Ed. Mar 9 Quick review: Assouad s lemma

More information

SOURCE CODING WITH SIDE INFORMATION AT THE DECODER (WYNER-ZIV CODING) FEB 13, 2003

SOURCE CODING WITH SIDE INFORMATION AT THE DECODER (WYNER-ZIV CODING) FEB 13, 2003 SOURCE CODING WITH SIDE INFORMATION AT THE DECODER (WYNER-ZIV CODING) FEB 13, 2003 SLEPIAN-WOLF RESULT { X i} RATE R x ENCODER 1 DECODER X i V i {, } { V i} ENCODER 0 RATE R v Problem: Determine R, the

More information

Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses

Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses Ann Inst Stat Math (2009) 61:773 787 DOI 10.1007/s10463-008-0172-6 Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses Taisuke Otsu Received: 1 June 2007 / Revised:

More information

THIS paper is aimed at designing efficient decoding algorithms

THIS paper is aimed at designing efficient decoding algorithms IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 45, NO. 7, NOVEMBER 1999 2333 Sort-and-Match Algorithm for Soft-Decision Decoding Ilya Dumer, Member, IEEE Abstract Let a q-ary linear (n; k)-code C be used

More information

Gaussian Estimation under Attack Uncertainty

Gaussian Estimation under Attack Uncertainty Gaussian Estimation under Attack Uncertainty Tara Javidi Yonatan Kaspi Himanshu Tyagi Abstract We consider the estimation of a standard Gaussian random variable under an observation attack where an adversary

More information