Convexity/Concavity of Renyi Entropy and α-mutual Information

Size: px
Start display at page:

Download "Convexity/Concavity of Renyi Entropy and α-mutual Information"

Transcription

1 Convexity/Concavity of Renyi Entropy and -Mutual Information Siu-Wai Ho Institute for Telecommunications Research University of South Australia Adelaide, SA 5095, Australia Sergio Verdú Dept. of Electrical Engineering Princeton University Princeton, NJ 08544, U.S.A. Abstract Entropy is well known to be Schur concave on finite alphabets. Recently, the authors have strengthened the result by showing that for any pair of probability distributions P and with majorized by P, the entropy of is larger than the entropy of P by the amount of relative entropy DP. This result applies to P and defined on countable alphabets. This paper shows the counterpart of this result for the Rényi entropy and the Tsallis entropy. Lower bounds on the difference in the Rényi or Tsallis entropy are given in terms of a new divergence which is related to the Rényi or Tsallis divergence. This paper also considers a notion of generalized mutual information, namely -mutual information, which is defined through the Rényi divergence. The convexity/concavity for different ranges of is shown. A sufficient condition for the Schur concavity is discussed and upper bounds on -mutual information are given in terms of the Rényi entropy. I. INTRODUCTION The Rényi entropy H P of a probability distribution P of order was introduced in []. The concavity of H P for P defined on a finite support was shown in [2]. If 0, ], H P is strictly concave in P and it is also Schur concave [3]. When >, Rényi entropy is neither convex nor concave in general but it is pseudoconcave. Since all concave functions and pseudoconcave functions are quasiconcave, Rényi entropy is Schur concave for > 0 [4]. It is well known that the Shannon entropy of finitelyvalued random variables is a Schur-concave function [4]. This result was strengthened and generalized for countably infinite random variables in [5], which shows that if a probability distribution is majorized by a probability distribution P, their difference in entropy was shown to be lower bounded by the relative entropy DP. This motivates the search for strengthened results for the Schur-concavity of the Rényi entropy. In order to strengthen the Schur-concavity of the Rényi entropy, this paper investigates the Schur-concavity of the Tsallis entropy. Tsallis entropy was introduced by Tsallis [6] to study multifractals. The same paper showed that the Tsallis entropy of order β is concave for β > 0 and convex for β < 0. The Schur-concavity of the Tsallis entropy was discussed in [7]. Tsallis divergence of order β was introduced in [8] and it is also known as the Hellinger divergence of order β [9]. Indeed, the Tsallis divergence is an f-divergence with fx = x xβ β and the Tsallis entropy is the corresponding entropy with the form i fp i for any probability distribution P. This paper also considers the convexity of a notion of generalized mutual information, namely -mutual information. This notion of mutual information was introduced by Sibson [0] in the discrete case, as the information radius of order []. Alternative notions of what could be called Rényi mutual information were proposed by Csiszár [2] and Arimoto [3]. -mutual information is closely related to Gallager s exponent function and plays an important role in a number of applications in channel capacity problems e.g., [4]. In Section II, the Schur concavity of the Rényi entropy and the Tsallis entropy are revisited. A new divergence is defined and new bounds on entropy in terms of the new divergence are derived. The -mutual information is defined and discussed in Section III. Its concavity, convexity, Schur concavity and upper bounds under different settings are illustrated. Due to space limits, some of the proofs are omitted. For simplicity, only countable alphabets are considered in this paper. All log and exp have the same arbitrary base. We use P to denote that the support of P is a subset of the support of. Furthermore, if P, it means that P does not hold. II. SCHUR-CONCAVITY OF RÉNYI ENTROPY AND TSALLIS ENTROPY Definition : For any probability distribution P, the Rényi entropy of order 0,, is defined as H P = log P i, where A is the support of P. By convention, H P is the Shannon entropy of P. Definition 2: For any probability distributions P and, let A be the union of the supports of P and. The Rényi divergence [] of order 0,, is defined as D P = log P i i, 2 which becomes infinite if > and P.

2 Definition 3: Given P X, P Y X, Y X, the conditional Rényi divergence of order 0,, is D P Y X Y X P X 3 = D P Y X P X Y X P X 4 = log PY X y x Y X y xp Xx, x,y X Y where X and Y are the supports of P X and P Y, respectively. The Tsallis entropy is related to the Rényi entropy through the following one-to-one function. For 0,, and z > 0, let and 5 ϕ z = expz z 6 ϕ v = log v. 7 Definition 4: For any probability distribution P, the Tsallis entropy of order β 0,, is defined as S β P = ϕ βh β P = P i β, 8 β β where A is the support of P. Definition 5: For any probability distributions P and, let A be the support of P. The Tsallis divergence of order β 0,, is defined as T β P = ϕ β D β P 9 β = P i β i β, 0 β which becomes infinite if β > and P. Definition 6: For any probability distributions P and and strictly concave function φ : [0, ] IR, let φ P i, i =P i iφ i + φi φp i, where φ x is the derivative of φx and i A which is the union of the supports of P and. Define the divergence W φ P = φ P i, i. 2 Lemma : For any probability distributions P and and strictly concave function φ : [0, ] IR, and with equality if and only if P =. φ P i, i 0 3 W φ P 0, 4 In the following, we fix β 0,, and consider the strictly concave function φ β x = xβ x β. 5 Its derivative is denoted by φ β x = βxβ β. For 0 x, φ β x is a decreasing function. Note that the Tsallis divergence can be seen as an f-divergence with fx = φx. The Tsallis entropy is equal to fp i. Without loss of generality, we assume that P and are defined on the positive integers and labeled in decreasing probabilities, i.e., P i P i + 6 i i +. 7 We say that is majorized by P if for all k =, 2,... k i i= k P i. 8 i= If is majorized by P, then P. If a real-valued functional f is such that fp f whenever is majorized by P, we say that f is Schur-concave. The following result generalizes [5, Theorem 3]. Theorem : For β 0,,, if is majorized by P, then S β S β P + W φβ P. 9 Proof: We first consider the case where and therefore P takes values on a finite integer set {, 2,..., M}. S β S β P W φβ P 20 = i φ β i i φ β P i W φβ P 2 = φ βii P i 22 i = M i P i φ βk φ βk + + i k=i M i P iφ βm 23 i= M = k= k φ βk φ βk + i i= k P i i= 24 0, 25 where 25 follows from the fact that φ β x is a decreasing function and is majorized by P. Consider the case that is non-zero for all integers. We assume that S β is finite, as otherwise, there is nothing to prove. We need to take M in the foregoing expressions.

3 Although the last term in the right side of 23 can now be negative, it vanishes due to the finiteness of S β : M i P iφ βm i= i=m+ i=m+ = β β iφ βm iφ βi i=m i β, 28 where 27 follows from that φ β x is a decreasing function and 7. Therefore, the Tsallis entropy is a Schur-concave function. Due to the one-to-one correspondence between the Rényi entropy and the Tsallis entropy through 8, Theorem leads to the following bounds. If 0 < <, then ϕ H ϕ H P W φ P. 29 If >, then ϕ H P ϕ H W φ P. 30 These bounds are sufficient to show that the Rényi entropy is Schur-concave. They can further be used to obtain bounds on H H P in terms of W φ P as follows. Theorem 2: For P defined on a countable alphabet, the Rényi entropy H P is a Schur-concave function for 0,,. Furthermore, if is majorized by P, then a If 0 < <, then in nats H H P = ϕ A A W φ P 3 W φ P A 32 0, 33 where A is the support of. b If >, then in nats H H P ϕ W φ P 34 W φ P It would be interesting to obtain an inequality similar to [5, Theorem 3] for the Rényi entropy. However, the following example shows that it is not possible. Example : Consider P = {0.4, 0.35, 0.5, 0.} and = {0.3, 0.3, 0.3, 0.} so that is majorized by P. If = 0.6, then H H P > 0 but H H P < D P. To end this section, we illustrate the meaning of φ P i, i and how W φ P is related to relative a Fig.. An illustration of φx and φ in case a P i < i and case b i < P i. entropy and Tsallis divergence. Consider the tangent at φx when x = i as shown in Fig.. Lemma shows that the tangent is always above φp i and the amount is φ P i, i. Theorem 3: For any probability distributions P and with P and φx = x log x, b DP = W φ P 37 with the convention 0 log 0 = 0. Furthermore, if φ β x = xβ β, i β φβ P i, i = T β P, 38 where A is the support of, and hence for 0 < β <, T β P W φβ P. 39 III. -MUTUAL INFORMATION Definition 7: Let P X P Y X P Y, with X X and Y Y. The -mutual information with 0,, is I X; Y = min D P Y X P X 40 = log P X xpy X=x y. x X 4 By convention, I X; Y is the mutual information. Some properties of I X; Y related to the parameter are summarized in the following theorem. Theorem 4: For any fixed P X on X and P Y X : X Y, a I X; Y is continuous with 0, and lim X; Y = sup log 0 P X x. 42 lim I X; Y = I X; Y. 43 lim I X; Y = log sup x:p X x>0 P Y X y x. 44

4 b I X; Y is increasing with. Proof: a. The limit in 42 can be obtained by using L norm. For any ɛ > 0, there exists a sufficiently small 0 < such that P X x ɛ < P X xpy X=x y 45 x X < P X x. 46 Since ɛ > 0 is arbitrary, using L -norm gives lim P X xp 0 Y X=x y 47 y x X = sup P X x. 48 So the limit in 42 follows. The limits in 43 and 44 can be easily verified by using L Hôpital s rule. Due to 43, I X; Y is continuous with 0,. b. For < β, denote β = argmin D β P Y X P X so that I X; Y D P Y X β P X 49 D β P Y X β P X 50 = I β X; Y, 5 where 49 and 50 follow from 40 and [5, Theorem 3], respectively. Provided that C 0f > 0, Theorem 4 intriguingly reveals that the zero-error capacity of a discrete memoryless channel with feedback [6] is equal to C 0f = sup I 0 X; Y, 52 X where I 0 is defined as the continuous extension of I. Some upper bounds on -mutual information in terms of Rényi entropy are illustrated. These bounds provide some insights about generalizing conditional entropy. Theorem 5: For any given P X, P Y X and 0 < <, H P X = I X; X I P X, P Y X 53 Proof: It is easy to verify the equality in 53. The inequality can be seen by the following. Let U = X so that X U Y forms a Markov chain. The inequality in 53 follows from [4, Theorem 5.2] for and [7] for =. The difference between the sides in 53 is a good candidate for the conditional Rényi entropy for which several candidates have been suggested in the past e.g., [2], [3]. Since H P X is decreasing in [2], the following corollary holds. Corollary 6: For any given P X, P Y X and 0 <, H P X I P X, P Y X. 54 The following result shows the convexity or concavity of the conditional Rényi divergence, which proves to be useful to investigate the properties of -mutual information. Theorem 7: Given P X, P Y X, Y X, the conditional Rényi divergence D P Y X Y X P X is a concave in P X for and quasiconcave in P X for 0 < <. b convex in P Y X, Y X for 0 < and quasiconvex in P Y X, Y X for 0 < <. Proof: a. For 0 < <, log a is a decreasing function in a. The quasiconcavity of I X; Y can be seen from 5. For >, log a is concave in a so that the concavity of I X; Y can be seen from 5. Since I X; Y is concave in P X, part a is verified. Part b is a consequence of 4 together with [5, Theorem 3] and [5, Theorem ]. Theorem 8: Let P X P Y X P Y. For fixed P Y X, a I P X, P Y X = I X; Y is concave in P X for. b I P X, P Y X = I X; Y is quasiconcave in P X for 0 < <. Proof: Consider 0 < λ < and λ = λ. We first prove part a. I X; Y is known to be concave in P X for fixed P Y X [7]. Consider any probability distributions P X0 and P X and let = argmin D P Y X λp X0 + λp X. 55 Theorem 7 can be applied to show part a as follows: I λp X0 + λp X, P Y X = D P Y X λp X0 + λp X 56 λd P Y X P X0 + λd P Y X P X 57 λ min D P Y X P X0 + λ min D P Y X P X. 58 Similarly, part b can also be shown by using Theorem 7. Example 2: Although I P X, P Y X is concave in P X for, it can be shown that in general it is not Schur concave by considering λ = 0.3, = 2, P X0 0 = P X0 = 0.4, P X 0 = P X = 0.5, and [ ] 0 P Y X = However, the Schur concavity can be preserved if P Y X describes a symmetric channel. In other words, all the rows of the probability transition matrix P Y X are permutations of other rows and so are the columns. Since all symmetric quasiconcave functions are Schur concave [4], Theorem 8 implies the following corollary. Corollary 9: For any fixed symmetric P Y X, I P X, P Y X = I X; Y is Schur concave in P X for 0 < <. We now fix P X and consider the behavior of I P X,.

5 Theorem 0: Let P X P Y X P Y. For fixed P X, a I P X, P Y X = I X; Y is convex in P Y X for 0 <. b I P X, P Y X = I X; Y is quasiconvex in P Y X for 0 < <. Proof: Consider 0 < λ < and λ = λ. We first prove part a. I X; Y is known to be convex in P Y X [7]. Consider any conditional probability distributions P Y0 X and P Y X and let i = argmin D P Yi X P X 60 for i = 0 or. Due to 0 < < and Theorem 7, I P X, λp Y X + λp Y0 X 6 = min λp Y X + λp Y0 X P X 62 D λp Y X + λp Y0 X λ + λ 0 P X 63 λd P Y X P X + λd P Y0 X 0 P X 64 =λi P X, P Y X + λi P X, P Y0 X. 65 Similarly, Part b can be verifed by using Theorem 7. In Theorems 8 and 0, we have made different assumptions on the ranges of. The following examples justify that those assumptions are not superfluous. Example 3: Let λ = 0.3, P X0 0 = P X0 =, P X 0 = P X = 0.5, and [ ] 0 P Y X = I 0. P X, P Y X is not concave in P X and I 0.3 P X, P Y X is not convex in P X. Example 4: Let λ = 0.3, P X 0 = P X = 0.5, [ ] [ ] P Y0 X 0 = and P Y X = I 2 P X, P Y X is not concave in P Y X and I 3 P X, P Y X is not convex in P Y X. Although Examples 3 and 4 show some unpleasant behaviour of I P X, P Y X, the Arimoto-Blahut algorithm [8] can still be applied to maximize I P X, P Y X for > 0. The following theorem, motivated by [4], illustrates that some optimization problems about I P X, P Y X can be converted into convex optimization problems. Theorem : Consider 0,,. Let f P X, P Y X = ϕ I P X, P Y X 68 = P X xpy X=x y, x X z is a mono- where presents in 68 such that tonic increasing function in z. ϕ 69 a For a fixed P Y X, sup P X I P X, P Y X = ϕ sup f P X, P Y X, P X 70 where f P X, P Y X is concave in P X. b For a fixed P X, inf I P X, P Y X = ϕ P Y X inf f P X, P Y X P Y X, 7 where f P X, P Y X is convex in P Y X. Proof: Since ϕ z is a monotonic increasing function in z, 70 and 7 follow from the relation in 68. The function f P X, P Y X in 69 is concave in P X because z is concave in z. The function f P X, P Y X in 69 is convex in P Y X due to the Minkowski inequality for > and the reverse Minkowski inequality for <. Remark: Note that convexity/concavity were mistakenly reversed in [4, Theorem 5]. REFERENCES [] A. Rényi, On measures of entropy and information, Fourth Berkeley Symp. on Mathematical Statistics and Probability, pp , 96. [2] M. Ben-Bassat and J. Raviv, Rényi s entropy and the probability of error, IEEE Transactions on Information Theory, vol. 24, no. 3, pp , May 978. [3] Y. He, A. B. Hamza, and H. Krim, A generalized divergence measure for robust image registration, IEEE Transactions on Signal Processing, vol. 5, no. 5, pp , [4] A. W. Marshall, I. Olkin, and B. C. Arnold, Inequalities: Theory of Majorization and Its Applications. Springer, 200. [5] S.-W. Ho and S. Verdú, On the interplay between conditional entropy and error probability, IEEE Transactions on Information Theory, vol. 56, no. 2, pp , 200. [6] C. Tsallis, Possible generalization of Boltzmann-Gibbs statistics, Journal of statistical physics, vol. 52, no. -2, pp , 988. [7] S. Furuichi, K. Yanagi, and K. Kuriyama, Fundamental properties of Tsallis relative entropy, Journal of Mathematical Physics, vol. 45, no. 2, pp , [8] C. Tsallis, Generalized entropy-based criterion for consistent testing, Physical Review E, vol. 58, no. 2, p. 442, 998. [9] F. Liese and I. Vajda, On divergences and informations in statistics and information theory, IEEE Transactions on Information Theory, vol. 52, no. 0, pp , Oct [0] R. Sibson, Information radius, Zeitschrift für Wahrscheinlichkeitstheorie und verwandte Gebiete, vol. 4, no. 2, pp , 969. [] S. Verdú, -mutual information, 205 Information Theory and Applications Workshop, 205. [2] I. Csiszár, Generalized cutoff rates and Rényi s information measures, IEEE Trans. on Information Theory, vol. 4, no., pp , 995. [3] S. Arimoto, Information measures and capacity of order for discrete memoryless channels, Proc. Coll. Math. Soc. János Bolyai., 975. [4] Y. Polyanskiy and S. Verdú, Arimoto channel coding converse and Rényi divergence, th Annual Allerton Conference on Communication, Control, and Computing Allerton, pp , 200. [5] T. van Erven and P. Harremoës, Rényi divergence and Kullback-Leibler divergence, IEEE Transactions on Information Theory, vol. 60, no. 7, pp , July 204. [6] C. E. Shannon, The zero error capacity of a noisy channel, IRE Transactions on Information Theory, vol. 2, no. 3, pp. 8 9, 956. [7] T. M. Cover and J. A. Thomas, Elements of Information Theory, 2nd ed. John Wiley & Sons, 202. [8] S. Arimoto, Computation of random coding exponent functions, IEEE Transactions on Information Theory, vol. 22, no. 6, pp , 976.

Arimoto Channel Coding Converse and Rényi Divergence

Arimoto Channel Coding Converse and Rényi Divergence Arimoto Channel Coding Converse and Rényi Divergence Yury Polyanskiy and Sergio Verdú Abstract Arimoto proved a non-asymptotic upper bound on the probability of successful decoding achievable by any code

More information

arxiv: v4 [cs.it] 17 Oct 2015

arxiv: v4 [cs.it] 17 Oct 2015 Upper Bounds on the Relative Entropy and Rényi Divergence as a Function of Total Variation Distance for Finite Alphabets Igal Sason Department of Electrical Engineering Technion Israel Institute of Technology

More information

Arimoto-Rényi Conditional Entropy. and Bayesian M-ary Hypothesis Testing. Abstract

Arimoto-Rényi Conditional Entropy. and Bayesian M-ary Hypothesis Testing. Abstract Arimoto-Rényi Conditional Entropy and Bayesian M-ary Hypothesis Testing Igal Sason Sergio Verdú Abstract This paper gives upper and lower bounds on the minimum error probability of Bayesian M-ary hypothesis

More information

Entropy measures of physics via complexity

Entropy measures of physics via complexity Entropy measures of physics via complexity Giorgio Kaniadakis and Flemming Topsøe Politecnico of Torino, Department of Physics and University of Copenhagen, Department of Mathematics 1 Introduction, Background

More information

The Information Bottleneck Revisited or How to Choose a Good Distortion Measure

The Information Bottleneck Revisited or How to Choose a Good Distortion Measure The Information Bottleneck Revisited or How to Choose a Good Distortion Measure Peter Harremoës Centrum voor Wiskunde en Informatica PO 94079, 1090 GB Amsterdam The Nederlands PHarremoes@cwinl Naftali

More information

Strong Converse Theorems for Classes of Multimessage Multicast Networks: A Rényi Divergence Approach

Strong Converse Theorems for Classes of Multimessage Multicast Networks: A Rényi Divergence Approach Strong Converse Theorems for Classes of Multimessage Multicast Networks: A Rényi Divergence Approach Silas Fong (Joint work with Vincent Tan) Department of Electrical & Computer Engineering National University

More information

MMSE Dimension. snr. 1 We use the following asymptotic notation: f(x) = O (g(x)) if and only

MMSE Dimension. snr. 1 We use the following asymptotic notation: f(x) = O (g(x)) if and only MMSE Dimension Yihong Wu Department of Electrical Engineering Princeton University Princeton, NJ 08544, USA Email: yihongwu@princeton.edu Sergio Verdú Department of Electrical Engineering Princeton University

More information

Discrete Memoryless Channels with Memoryless Output Sequences

Discrete Memoryless Channels with Memoryless Output Sequences Discrete Memoryless Channels with Memoryless utput Sequences Marcelo S Pinho Department of Electronic Engineering Instituto Tecnologico de Aeronautica Sao Jose dos Campos, SP 12228-900, Brazil Email: mpinho@ieeeorg

More information

Functional Properties of MMSE

Functional Properties of MMSE Functional Properties of MMSE Yihong Wu epartment of Electrical Engineering Princeton University Princeton, NJ 08544, USA Email: yihongwu@princeton.edu Sergio Verdú epartment of Electrical Engineering

More information

Convergence of generalized entropy minimizers in sequences of convex problems

Convergence of generalized entropy minimizers in sequences of convex problems Proceedings IEEE ISIT 206, Barcelona, Spain, 2609 263 Convergence of generalized entropy minimizers in sequences of convex problems Imre Csiszár A Rényi Institute of Mathematics Hungarian Academy of Sciences

More information

Quantum Sphere-Packing Bounds and Moderate Deviation Analysis for Classical-Quantum Channels

Quantum Sphere-Packing Bounds and Moderate Deviation Analysis for Classical-Quantum Channels Quantum Sphere-Packing Bounds and Moderate Deviation Analysis for Classical-Quantum Channels (, ) Joint work with Min-Hsiu Hsieh and Marco Tomamichel Hao-Chung Cheng University of Technology Sydney National

More information

ECE 4400:693 - Information Theory

ECE 4400:693 - Information Theory ECE 4400:693 - Information Theory Dr. Nghi Tran Lecture 8: Differential Entropy Dr. Nghi Tran (ECE-University of Akron) ECE 4400:693 Lecture 1 / 43 Outline 1 Review: Entropy of discrete RVs 2 Differential

More information

Literature on Bregman divergences

Literature on Bregman divergences Literature on Bregman divergences Lecture series at Univ. Hawai i at Mānoa Peter Harremoës February 26, 2016 Information divergence was introduced by Kullback and Leibler [25] and later Kullback started

More information

Limits from l Hôpital rule: Shannon entropy as limit cases of Rényi and Tsallis entropies

Limits from l Hôpital rule: Shannon entropy as limit cases of Rényi and Tsallis entropies Limits from l Hôpital rule: Shannon entropy as it cases of Rényi and Tsallis entropies Frank Nielsen École Polytechnique/Sony Computer Science Laboratorie, Inc http://www.informationgeometry.org August

More information

Variable Length Codes for Degraded Broadcast Channels

Variable Length Codes for Degraded Broadcast Channels Variable Length Codes for Degraded Broadcast Channels Stéphane Musy School of Computer and Communication Sciences, EPFL CH-1015 Lausanne, Switzerland Email: stephane.musy@ep.ch Abstract This paper investigates

More information

Dispersion of the Gilbert-Elliott Channel

Dispersion of the Gilbert-Elliott Channel Dispersion of the Gilbert-Elliott Channel Yury Polyanskiy Email: ypolyans@princeton.edu H. Vincent Poor Email: poor@princeton.edu Sergio Verdú Email: verdu@princeton.edu Abstract Channel dispersion plays

More information

f-divergence Inequalities

f-divergence Inequalities SASON AND VERDÚ: f-divergence INEQUALITIES f-divergence Inequalities Igal Sason, and Sergio Verdú Abstract arxiv:508.00335v7 [cs.it] 4 Dec 06 This paper develops systematic approaches to obtain f-divergence

More information

ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016

ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 ECE598: Information-theoretic methods in high-dimensional statistics Spring 06 Lecture : Mutual Information Method Lecturer: Yihong Wu Scribe: Jaeho Lee, Mar, 06 Ed. Mar 9 Quick review: Assouad s lemma

More information

The Method of Types and Its Application to Information Hiding

The Method of Types and Its Application to Information Hiding The Method of Types and Its Application to Information Hiding Pierre Moulin University of Illinois at Urbana-Champaign www.ifp.uiuc.edu/ moulin/talks/eusipco05-slides.pdf EUSIPCO Antalya, September 7,

More information

Feedback Capacity of a Class of Symmetric Finite-State Markov Channels

Feedback Capacity of a Class of Symmetric Finite-State Markov Channels Feedback Capacity of a Class of Symmetric Finite-State Markov Channels Nevroz Şen, Fady Alajaji and Serdar Yüksel Department of Mathematics and Statistics Queen s University Kingston, ON K7L 3N6, Canada

More information

Tight Bounds for Symmetric Divergence Measures and a New Inequality Relating f-divergences

Tight Bounds for Symmetric Divergence Measures and a New Inequality Relating f-divergences Tight Bounds for Symmetric Divergence Measures and a New Inequality Relating f-divergences Igal Sason Department of Electrical Engineering Technion, Haifa 3000, Israel E-mail: sason@ee.technion.ac.il Abstract

More information

f-divergence Inequalities

f-divergence Inequalities f-divergence Inequalities Igal Sason Sergio Verdú Abstract This paper develops systematic approaches to obtain f-divergence inequalities, dealing with pairs of probability measures defined on arbitrary

More information

Entropy and Ergodic Theory Lecture 4: Conditional entropy and mutual information

Entropy and Ergodic Theory Lecture 4: Conditional entropy and mutual information Entropy and Ergodic Theory Lecture 4: Conditional entropy and mutual information 1 Conditional entropy Let (Ω, F, P) be a probability space, let X be a RV taking values in some finite set A. In this lecture

More information

Correlation Detection and an Operational Interpretation of the Rényi Mutual Information

Correlation Detection and an Operational Interpretation of the Rényi Mutual Information Correlation Detection and an Operational Interpretation of the Rényi Mutual Information Masahito Hayashi 1, Marco Tomamichel 2 1 Graduate School of Mathematics, Nagoya University, and Centre for Quantum

More information

Channels with cost constraints: strong converse and dispersion

Channels with cost constraints: strong converse and dispersion Channels with cost constraints: strong converse and dispersion Victoria Kostina, Sergio Verdú Dept. of Electrical Engineering, Princeton University, NJ 08544, USA Abstract This paper shows the strong converse

More information

Upper Bounds on the Capacity of Binary Intermittent Communication

Upper Bounds on the Capacity of Binary Intermittent Communication Upper Bounds on the Capacity of Binary Intermittent Communication Mostafa Khoshnevisan and J. Nicholas Laneman Department of Electrical Engineering University of Notre Dame Notre Dame, Indiana 46556 Email:{mhoshne,

More information

Lecture 2: August 31

Lecture 2: August 31 0-704: Information Processing and Learning Fall 206 Lecturer: Aarti Singh Lecture 2: August 3 Note: These notes are based on scribed notes from Spring5 offering of this course. LaTeX template courtesy

More information

On the Entropy of Sums of Bernoulli Random Variables via the Chen-Stein Method

On the Entropy of Sums of Bernoulli Random Variables via the Chen-Stein Method On the Entropy of Sums of Bernoulli Random Variables via the Chen-Stein Method Igal Sason Department of Electrical Engineering Technion - Israel Institute of Technology Haifa 32000, Israel ETH, Zurich,

More information

Simple Channel Coding Bounds

Simple Channel Coding Bounds ISIT 2009, Seoul, Korea, June 28 - July 3,2009 Simple Channel Coding Bounds Ligong Wang Signal and Information Processing Laboratory wang@isi.ee.ethz.ch Roger Colbeck Institute for Theoretical Physics,

More information

A Single-letter Upper Bound for the Sum Rate of Multiple Access Channels with Correlated Sources

A Single-letter Upper Bound for the Sum Rate of Multiple Access Channels with Correlated Sources A Single-letter Upper Bound for the Sum Rate of Multiple Access Channels with Correlated Sources Wei Kang Sennur Ulukus Department of Electrical and Computer Engineering University of Maryland, College

More information

Jensen-Shannon Divergence and Hilbert space embedding

Jensen-Shannon Divergence and Hilbert space embedding Jensen-Shannon Divergence and Hilbert space embedding Bent Fuglede and Flemming Topsøe University of Copenhagen, Department of Mathematics Consider the set M+ 1 (A) of probability distributions where A

More information

Chapter 2: Entropy and Mutual Information. University of Illinois at Chicago ECE 534, Natasha Devroye

Chapter 2: Entropy and Mutual Information. University of Illinois at Chicago ECE 534, Natasha Devroye Chapter 2: Entropy and Mutual Information Chapter 2 outline Definitions Entropy Joint entropy, conditional entropy Relative entropy, mutual information Chain rules Jensen s inequality Log-sum inequality

More information

Multiplicativity of Maximal p Norms in Werner Holevo Channels for 1 < p 2

Multiplicativity of Maximal p Norms in Werner Holevo Channels for 1 < p 2 Multiplicativity of Maximal p Norms in Werner Holevo Channels for 1 < p 2 arxiv:quant-ph/0410063v1 8 Oct 2004 Nilanjana Datta Statistical Laboratory Centre for Mathematical Sciences University of Cambridge

More information

5 Mutual Information and Channel Capacity

5 Mutual Information and Channel Capacity 5 Mutual Information and Channel Capacity In Section 2, we have seen the use of a quantity called entropy to measure the amount of randomness in a random variable. In this section, we introduce several

More information

Multiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets

Multiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets Multiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets Shivaprasad Kotagiri and J. Nicholas Laneman Department of Electrical Engineering University of Notre Dame Notre Dame,

More information

Reliable Computation over Multiple-Access Channels

Reliable Computation over Multiple-Access Channels Reliable Computation over Multiple-Access Channels Bobak Nazer and Michael Gastpar Dept. of Electrical Engineering and Computer Sciences University of California, Berkeley Berkeley, CA, 94720-1770 {bobak,

More information

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University Chapter 4 Data Transmission and Channel Capacity Po-Ning Chen, Professor Department of Communications Engineering National Chiao Tung University Hsin Chu, Taiwan 30050, R.O.C. Principle of Data Transmission

More information

Optimal Sequences and Sum Capacity of Synchronous CDMA Systems

Optimal Sequences and Sum Capacity of Synchronous CDMA Systems Optimal Sequences and Sum Capacity of Synchronous CDMA Systems Pramod Viswanath and Venkat Anantharam {pvi, ananth}@eecs.berkeley.edu EECS Department, U C Berkeley CA 9470 Abstract The sum capacity of

More information

(each row defines a probability distribution). Given n-strings x X n, y Y n we can use the absence of memory in the channel to compute

(each row defines a probability distribution). Given n-strings x X n, y Y n we can use the absence of memory in the channel to compute ENEE 739C: Advanced Topics in Signal Processing: Coding Theory Instructor: Alexander Barg Lecture 6 (draft; 9/6/03. Error exponents for Discrete Memoryless Channels http://www.enee.umd.edu/ abarg/enee739c/course.html

More information

Large Deviations Performance of Knuth-Yao algorithm for Random Number Generation

Large Deviations Performance of Knuth-Yao algorithm for Random Number Generation Large Deviations Performance of Knuth-Yao algorithm for Random Number Generation Akisato KIMURA akisato@ss.titech.ac.jp Tomohiko UYEMATSU uematsu@ss.titech.ac.jp April 2, 999 No. AK-TR-999-02 Abstract

More information

Some New Information Inequalities Involving f-divergences

Some New Information Inequalities Involving f-divergences BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 12, No 2 Sofia 2012 Some New Information Inequalities Involving f-divergences Amit Srivastava Department of Mathematics, Jaypee

More information

On Improved Bounds for Probability Metrics and f- Divergences

On Improved Bounds for Probability Metrics and f- Divergences IRWIN AND JOAN JACOBS CENTER FOR COMMUNICATION AND INFORMATION TECHNOLOGIES On Improved Bounds for Probability Metrics and f- Divergences Igal Sason CCIT Report #855 March 014 Electronics Computers Communications

More information

arxiv: v4 [cs.it] 8 Apr 2014

arxiv: v4 [cs.it] 8 Apr 2014 1 On Improved Bounds for Probability Metrics and f-divergences Igal Sason Abstract Derivation of tight bounds for probability metrics and f-divergences is of interest in information theory and statistics.

More information

Lecture 5 - Information theory

Lecture 5 - Information theory Lecture 5 - Information theory Jan Bouda FI MU May 18, 2012 Jan Bouda (FI MU) Lecture 5 - Information theory May 18, 2012 1 / 42 Part I Uncertainty and entropy Jan Bouda (FI MU) Lecture 5 - Information

More information

On Semicontinuity of Convex-valued Multifunctions and Cesari s Property (Q)

On Semicontinuity of Convex-valued Multifunctions and Cesari s Property (Q) On Semicontinuity of Convex-valued Multifunctions and Cesari s Property (Q) Andreas Löhne May 2, 2005 (last update: November 22, 2005) Abstract We investigate two types of semicontinuity for set-valued

More information

Tight Bounds for Symmetric Divergence Measures and a Refined Bound for Lossless Source Coding

Tight Bounds for Symmetric Divergence Measures and a Refined Bound for Lossless Source Coding APPEARS IN THE IEEE TRANSACTIONS ON INFORMATION THEORY, FEBRUARY 015 1 Tight Bounds for Symmetric Divergence Measures and a Refined Bound for Lossless Source Coding Igal Sason Abstract Tight bounds for

More information

A new converse in rate-distortion theory

A new converse in rate-distortion theory A new converse in rate-distortion theory Victoria Kostina, Sergio Verdú Dept. of Electrical Engineering, Princeton University, NJ, 08544, USA Abstract This paper shows new finite-blocklength converse bounds

More information

Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels

Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels Nan Liu and Andrea Goldsmith Department of Electrical Engineering Stanford University, Stanford CA 94305 Email:

More information

Equivalence for Networks with Adversarial State

Equivalence for Networks with Adversarial State Equivalence for Networks with Adversarial State Oliver Kosut Department of Electrical, Computer and Energy Engineering Arizona State University Tempe, AZ 85287 Email: okosut@asu.edu Jörg Kliewer Department

More information

Computing and Communications 2. Information Theory -Entropy

Computing and Communications 2. Information Theory -Entropy 1896 1920 1987 2006 Computing and Communications 2. Information Theory -Entropy Ying Cui Department of Electronic Engineering Shanghai Jiao Tong University, China 2017, Autumn 1 Outline Entropy Joint entropy

More information

EE5139R: Problem Set 7 Assigned: 30/09/15, Due: 07/10/15

EE5139R: Problem Set 7 Assigned: 30/09/15, Due: 07/10/15 EE5139R: Problem Set 7 Assigned: 30/09/15, Due: 07/10/15 1. Cascade of Binary Symmetric Channels The conditional probability distribution py x for each of the BSCs may be expressed by the transition probability

More information

SHARED INFORMATION. Prakash Narayan with. Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe

SHARED INFORMATION. Prakash Narayan with. Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe SHARED INFORMATION Prakash Narayan with Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe 2/41 Outline Two-terminal model: Mutual information Operational meaning in: Channel coding: channel

More information

Optimal global rates of convergence for interpolation problems with random design

Optimal global rates of convergence for interpolation problems with random design Optimal global rates of convergence for interpolation problems with random design Michael Kohler 1 and Adam Krzyżak 2, 1 Fachbereich Mathematik, Technische Universität Darmstadt, Schlossgartenstr. 7, 64289

More information

An Alternative Proof of Channel Polarization for Channels with Arbitrary Input Alphabets

An Alternative Proof of Channel Polarization for Channels with Arbitrary Input Alphabets An Alternative Proof of Channel Polarization for Channels with Arbitrary Input Alphabets Jing Guo University of Cambridge jg582@cam.ac.uk Jossy Sayir University of Cambridge j.sayir@ieee.org Minghai Qin

More information

On the Duality between Multiple-Access Codes and Computation Codes

On the Duality between Multiple-Access Codes and Computation Codes On the Duality between Multiple-Access Codes and Computation Codes Jingge Zhu University of California, Berkeley jingge.zhu@berkeley.edu Sung Hoon Lim KIOST shlim@kiost.ac.kr Michael Gastpar EPFL michael.gastpar@epfl.ch

More information

Information Theory and Hypothesis Testing

Information Theory and Hypothesis Testing Summer School on Game Theory and Telecommunications Campione, 7-12 September, 2014 Information Theory and Hypothesis Testing Mauro Barni University of Siena September 8 Review of some basic results linking

More information

Optimization in Information Theory

Optimization in Information Theory Optimization in Information Theory Dawei Shen November 11, 2005 Abstract This tutorial introduces the application of optimization techniques in information theory. We revisit channel capacity problem from

More information

EECS 750. Hypothesis Testing with Communication Constraints

EECS 750. Hypothesis Testing with Communication Constraints EECS 750 Hypothesis Testing with Communication Constraints Name: Dinesh Krithivasan Abstract In this report, we study a modification of the classical statistical problem of bivariate hypothesis testing.

More information

Intermittent Communication

Intermittent Communication Intermittent Communication Mostafa Khoshnevisan, Student Member, IEEE, and J. Nicholas Laneman, Senior Member, IEEE arxiv:32.42v2 [cs.it] 7 Mar 207 Abstract We formulate a model for intermittent communication

More information

Lecture 6: Gaussian Channels. Copyright G. Caire (Sample Lectures) 157

Lecture 6: Gaussian Channels. Copyright G. Caire (Sample Lectures) 157 Lecture 6: Gaussian Channels Copyright G. Caire (Sample Lectures) 157 Differential entropy (1) Definition 18. The (joint) differential entropy of a continuous random vector X n p X n(x) over R is: Z h(x

More information

Iterative Markov Chain Monte Carlo Computation of Reference Priors and Minimax Risk

Iterative Markov Chain Monte Carlo Computation of Reference Priors and Minimax Risk Iterative Markov Chain Monte Carlo Computation of Reference Priors and Minimax Risk John Lafferty School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 lafferty@cs.cmu.edu Abstract

More information

Lecture 4 Noisy Channel Coding

Lecture 4 Noisy Channel Coding Lecture 4 Noisy Channel Coding I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw October 9, 2015 1 / 56 I-Hsiang Wang IT Lecture 4 The Channel Coding Problem

More information

Channel Polarization and Blackwell Measures

Channel Polarization and Blackwell Measures Channel Polarization Blackwell Measures Maxim Raginsky Abstract The Blackwell measure of a binary-input channel (BIC is the distribution of the posterior probability of 0 under the uniform input distribution

More information

Lecture 6 I. CHANNEL CODING. X n (m) P Y X

Lecture 6 I. CHANNEL CODING. X n (m) P Y X 6- Introduction to Information Theory Lecture 6 Lecturer: Haim Permuter Scribe: Yoav Eisenberg and Yakov Miron I. CHANNEL CODING We consider the following channel coding problem: m = {,2,..,2 nr} Encoder

More information

H(X) = plog 1 p +(1 p)log 1 1 p. With a slight abuse of notation, we denote this quantity by H(p) and refer to it as the binary entropy function.

H(X) = plog 1 p +(1 p)log 1 1 p. With a slight abuse of notation, we denote this quantity by H(p) and refer to it as the binary entropy function. LECTURE 2 Information Measures 2. ENTROPY LetXbeadiscreterandomvariableonanalphabetX drawnaccordingtotheprobability mass function (pmf) p() = P(X = ), X, denoted in short as X p(). The uncertainty about

More information

Coding on Countably Infinite Alphabets

Coding on Countably Infinite Alphabets Coding on Countably Infinite Alphabets Non-parametric Information Theory Licence de droits d usage Outline Lossless Coding on infinite alphabets Source Coding Universal Coding Infinite Alphabets Enveloppe

More information

An upper bound on l q norms of noisy functions

An upper bound on l q norms of noisy functions Electronic Collouium on Computational Complexity, Report No. 68 08 An upper bound on l norms of noisy functions Alex Samorodnitsky Abstract Let T ɛ, 0 ɛ /, be the noise operator acting on functions on

More information

LECTURE 2. Convexity and related notions. Last time: mutual information: definitions and properties. Lecture outline

LECTURE 2. Convexity and related notions. Last time: mutual information: definitions and properties. Lecture outline LECTURE 2 Convexity and related notions Last time: Goals and mechanics of the class notation entropy: definitions and properties mutual information: definitions and properties Lecture outline Convexity

More information

An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if. 2 l i. i=1

An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if. 2 l i. i=1 Kraft s inequality An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if N 2 l i 1 Proof: Suppose that we have a tree code. Let l max = max{l 1,...,

More information

LECTURE 3. Last time:

LECTURE 3. Last time: LECTURE 3 Last time: Mutual Information. Convexity and concavity Jensen s inequality Information Inequality Data processing theorem Fano s Inequality Lecture outline Stochastic processes, Entropy rate

More information

Sequential prediction with coded side information under logarithmic loss

Sequential prediction with coded side information under logarithmic loss under logarithmic loss Yanina Shkel Department of Electrical Engineering Princeton University Princeton, NJ 08544, USA Maxim Raginsky Department of Electrical and Computer Engineering Coordinated Science

More information

Soft Covering with High Probability

Soft Covering with High Probability Soft Covering with High Probability Paul Cuff Princeton University arxiv:605.06396v [cs.it] 20 May 206 Abstract Wyner s soft-covering lemma is the central analysis step for achievability proofs of information

More information

Course 212: Academic Year Section 1: Metric Spaces

Course 212: Academic Year Section 1: Metric Spaces Course 212: Academic Year 1991-2 Section 1: Metric Spaces D. R. Wilkins Contents 1 Metric Spaces 3 1.1 Distance Functions and Metric Spaces............. 3 1.2 Convergence and Continuity in Metric Spaces.........

More information

Numerical Sequences and Series

Numerical Sequences and Series Numerical Sequences and Series Written by Men-Gen Tsai email: b89902089@ntu.edu.tw. Prove that the convergence of {s n } implies convergence of { s n }. Is the converse true? Solution: Since {s n } is

More information

Refined Bounds on the Empirical Distribution of Good Channel Codes via Concentration Inequalities

Refined Bounds on the Empirical Distribution of Good Channel Codes via Concentration Inequalities Refined Bounds on the Empirical Distribution of Good Channel Codes via Concentration Inequalities Maxim Raginsky and Igal Sason ISIT 2013, Istanbul, Turkey Capacity-Achieving Channel Codes The set-up DMC

More information

Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information

Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information 204 IEEE International Symposium on Information Theory Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information Omur Ozel, Kaya Tutuncuoglu 2, Sennur Ulukus, and Aylin Yener

More information

Metric Spaces and Topology

Metric Spaces and Topology Chapter 2 Metric Spaces and Topology From an engineering perspective, the most important way to construct a topology on a set is to define the topology in terms of a metric on the set. This approach underlies

More information

Research Collection. The relation between block length and reliability for a cascade of AWGN links. Conference Paper. ETH Library

Research Collection. The relation between block length and reliability for a cascade of AWGN links. Conference Paper. ETH Library Research Collection Conference Paper The relation between block length and reliability for a cascade of AWG links Authors: Subramanian, Ramanan Publication Date: 0 Permanent Link: https://doiorg/0399/ethz-a-00705046

More information

Math 118B Solutions. Charles Martin. March 6, d i (x i, y i ) + d i (y i, z i ) = d(x, y) + d(y, z). i=1

Math 118B Solutions. Charles Martin. March 6, d i (x i, y i ) + d i (y i, z i ) = d(x, y) + d(y, z). i=1 Math 8B Solutions Charles Martin March 6, Homework Problems. Let (X i, d i ), i n, be finitely many metric spaces. Construct a metric on the product space X = X X n. Proof. Denote points in X as x = (x,

More information

Czechoslovak Mathematical Journal

Czechoslovak Mathematical Journal Czechoslovak Mathematical Journal Oktay Duman; Cihan Orhan µ-statistically convergent function sequences Czechoslovak Mathematical Journal, Vol. 54 (2004), No. 2, 413 422 Persistent URL: http://dml.cz/dmlcz/127899

More information

An Extended Fano s Inequality for the Finite Blocklength Coding

An Extended Fano s Inequality for the Finite Blocklength Coding An Extended Fano s Inequality for the Finite Bloclength Coding Yunquan Dong, Pingyi Fan {dongyq8@mails,fpy@mail}.tsinghua.edu.cn Department of Electronic Engineering, Tsinghua University, Beijing, P.R.

More information

SHARED INFORMATION. Prakash Narayan with. Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe

SHARED INFORMATION. Prakash Narayan with. Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe SHARED INFORMATION Prakash Narayan with Imre Csiszár, Sirin Nitinawarat, Himanshu Tyagi, Shun Watanabe 2/40 Acknowledgement Praneeth Boda Himanshu Tyagi Shun Watanabe 3/40 Outline Two-terminal model: Mutual

More information

Introduction to Real Analysis

Introduction to Real Analysis Introduction to Real Analysis Joshua Wilde, revised by Isabel Tecu, Takeshi Suzuki and María José Boccardi August 13, 2013 1 Sets Sets are the basic objects of mathematics. In fact, they are so basic that

More information

Part III. 10 Topological Space Basics. Topological Spaces

Part III. 10 Topological Space Basics. Topological Spaces Part III 10 Topological Space Basics Topological Spaces Using the metric space results above as motivation we will axiomatize the notion of being an open set to more general settings. Definition 10.1.

More information

Mismatched Multi-letter Successive Decoding for the Multiple-Access Channel

Mismatched Multi-letter Successive Decoding for the Multiple-Access Channel Mismatched Multi-letter Successive Decoding for the Multiple-Access Channel Jonathan Scarlett University of Cambridge jms265@cam.ac.uk Alfonso Martinez Universitat Pompeu Fabra alfonso.martinez@ieee.org

More information

An Achievable Error Exponent for the Mismatched Multiple-Access Channel

An Achievable Error Exponent for the Mismatched Multiple-Access Channel An Achievable Error Exponent for the Mismatched Multiple-Access Channel Jonathan Scarlett University of Cambridge jms265@camacuk Albert Guillén i Fàbregas ICREA & Universitat Pompeu Fabra University of

More information

Threshold Optimization for Capacity-Achieving Discrete Input One-Bit Output Quantization

Threshold Optimization for Capacity-Achieving Discrete Input One-Bit Output Quantization Threshold Optimization for Capacity-Achieving Discrete Input One-Bit Output Quantization Rudolf Mathar Inst. for Theoretical Information Technology RWTH Aachen University D-556 Aachen, Germany mathar@ti.rwth-aachen.de

More information

A Holevo-type bound for a Hilbert Schmidt distance measure

A Holevo-type bound for a Hilbert Schmidt distance measure Journal of Quantum Information Science, 205, *,** Published Online **** 204 in SciRes. http://www.scirp.org/journal/**** http://dx.doi.org/0.4236/****.204.***** A Holevo-type bound for a Hilbert Schmidt

More information

Symmetric Characterization of Finite State Markov Channels

Symmetric Characterization of Finite State Markov Channels Symmetric Characterization of Finite State Markov Channels Mohammad Rezaeian Department of Electrical and Electronic Eng. The University of Melbourne Victoria, 31, Australia Email: rezaeian@unimelb.edu.au

More information

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 53, NO. 12, DECEMBER

IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 53, NO. 12, DECEMBER IEEE TRANSACTIONS ON INFORMATION THEORY, VOL 53, NO 12, DECEMBER 2007 4457 Joint Source Channel Coding Error Exponent for Discrete Communication Systems With Markovian Memory Yangfan Zhong, Student Member,

More information

Universal Estimation of Divergence for Continuous Distributions via Data-Dependent Partitions

Universal Estimation of Divergence for Continuous Distributions via Data-Dependent Partitions Universal Estimation of for Continuous Distributions via Data-Dependent Partitions Qing Wang, Sanjeev R. Kulkarni, Sergio Verdú Department of Electrical Engineering Princeton University Princeton, NJ 8544

More information

Journal of Inequalities in Pure and Applied Mathematics

Journal of Inequalities in Pure and Applied Mathematics Journal of Inequalities in Pure and Applied Mathematics http://jipamvueduau/ Volume 7, Issue 2, Article 59, 2006 ENTROPY LOWER BOUNDS RELATED TO A PROBLEM OF UNIVERSAL CODING AND PREDICTION FLEMMING TOPSØE

More information

THEOREM AND METRIZABILITY

THEOREM AND METRIZABILITY f-divergences - REPRESENTATION THEOREM AND METRIZABILITY Ferdinand Österreicher Institute of Mathematics, University of Salzburg, Austria Abstract In this talk we are first going to state the so-called

More information

Two Applications of the Gaussian Poincaré Inequality in the Shannon Theory

Two Applications of the Gaussian Poincaré Inequality in the Shannon Theory Two Applications of the Gaussian Poincaré Inequality in the Shannon Theory Vincent Y. F. Tan (Joint work with Silas L. Fong) National University of Singapore (NUS) 2016 International Zurich Seminar on

More information

SHANNON made the following well-known remarks in

SHANNON made the following well-known remarks in IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 50, NO. 2, FEBRUARY 2004 245 Geometric Programming Duals of Channel Capacity and Rate Distortion Mung Chiang, Member, IEEE, and Stephen Boyd, Fellow, IEEE

More information

Strong Converse and Stein s Lemma in the Quantum Hypothesis Testing

Strong Converse and Stein s Lemma in the Quantum Hypothesis Testing Strong Converse and Stein s Lemma in the Quantum Hypothesis Testing arxiv:uant-ph/9906090 v 24 Jun 999 Tomohiro Ogawa and Hiroshi Nagaoka Abstract The hypothesis testing problem of two uantum states is

More information

Speech Recognition Lecture 7: Maximum Entropy Models. Mehryar Mohri Courant Institute and Google Research

Speech Recognition Lecture 7: Maximum Entropy Models. Mehryar Mohri Courant Institute and Google Research Speech Recognition Lecture 7: Maximum Entropy Models Mehryar Mohri Courant Institute and Google Research mohri@cims.nyu.com This Lecture Information theory basics Maximum entropy models Duality theorem

More information

Recent Results on Input-Constrained Erasure Channels

Recent Results on Input-Constrained Erasure Channels Recent Results on Input-Constrained Erasure Channels A Case Study for Markov Approximation July, 2017@NUS Memoryless Channels Memoryless Channels Channel transitions are characterized by time-invariant

More information

Channel Dispersion and Moderate Deviations Limits for Memoryless Channels

Channel Dispersion and Moderate Deviations Limits for Memoryless Channels Channel Dispersion and Moderate Deviations Limits for Memoryless Channels Yury Polyanskiy and Sergio Verdú Abstract Recently, Altug and Wagner ] posed a question regarding the optimal behavior of the probability

More information

Week 2: Sequences and Series

Week 2: Sequences and Series QF0: Quantitative Finance August 29, 207 Week 2: Sequences and Series Facilitator: Christopher Ting AY 207/208 Mathematicians have tried in vain to this day to discover some order in the sequence of prime

More information