Entropy Power Inequalities: Results and Speculation

Size: px
Start display at page:

Download "Entropy Power Inequalities: Results and Speculation"

Transcription

1 Entropy Power Inequalities: Results and Speculation Venkat Anantharam EECS Department University of California, Berkeley December 12, 2013 Workshop on Coding and Information Theory Institute of Mathematical Research The University of Hong Kong Contributions involve joint work with Varun Jog V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

2 Outline 1 Setting the stage 2 The Brunn Minkowski inequality 3 Proofs of the EPI 4 Mrs. Gerber s Lemma 5 Young s inequality 6 Versions of the EPI for discrete random variables 7 EPI for Groups or Order 2 n 8 The Fourier transform on a finite abelian group 9 Young s inequality on a finite abelian group 10 Brunn Minkowski inequality on a finite abelian group 11 Speculation V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

3 Outline 1 Setting the stage 2 The Brunn Minkowski inequality 3 Proofs of the EPI 4 Mrs. Gerber s Lemma 5 Young s inequality 6 Versions of the EPI for discrete random variables 7 EPI for Groups or Order 2 n 8 The Fourier transform on a finite abelian group 9 Young s inequality on a finite abelian group 10 Brunn Minkowski inequality on a finite abelian group 11 Speculation V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

4 Differential entropy and entropy power V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

5 Differential entropy and entropy power Let X be an R n -valued random variable with density f. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

6 Differential entropy and entropy power Let X be an R n -valued random variable with density f. The differential entropy of X is h(x) := E[ log f (X)]. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

7 Differential entropy and entropy power Let X be an R n -valued random variable with density f. The differential entropy of X is h(x) := E[ log f (X)]. The entropy power of X is N(X) := 1 2πe e 2 n h(x). V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

8 Shannon s entropy power inequality V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

9 Shannon s entropy power inequality Theorem (Entropy Power Inequality) If X and Y are independent R n -valued random variables, then e 2 n h(x+y) e 2 n h(x) + e 2 n h(y), with equality if and only if X and Y are Gaussian with proportional covariance matrices. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

10 Notable applications of the EPI-1 V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

11 Notable applications of the EPI-2 V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

12 Notable applications of the EPI-3 V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

13 Notable applications of the EPI-4 V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

14 Outline 1 Setting the stage 2 The Brunn Minkowski inequality 3 Proofs of the EPI 4 Mrs. Gerber s Lemma 5 Young s inequality 6 Versions of the EPI for discrete random variables 7 EPI for Groups or Order 2 n 8 The Fourier transform on a finite abelian group 9 Young s inequality on a finite abelian group 10 Brunn Minkowski inequality on a finite abelian group 11 Speculation V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

15 The Minkowski sum of two subsets of R n V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

16 The Minkowski sum of two subsets of R n Minkowski sum of a square and a circle V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

17 The Minkowski sum of two subsets of R n Minkowski sum of a square and a circle In general the Minkowski sum of two sets A, B R n is A + B := {a + b : a A and b B}. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

18 The Brunn Minkowski inequality V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

19 The Brunn Minkowski inequality Let A, B R n be convex bodies (compact convex sets with nonempty interior). V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

20 The Brunn Minkowski inequality Let A, B R n be convex bodies (compact convex sets with nonempty interior). Then Vol(A + B) n Vol(A) n + Vol(B) n. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

21 The Brunn Minkowski inequality Let A, B R n be convex bodies (compact convex sets with nonempty interior). Then Vol(A + B) n Vol(A) n + Vol(B) n. Equality holds iff A and B are homothetic, i.e. equal up to translation and dilation. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

22 The Brunn Minkowski inequality Let A, B R n be convex bodies (compact convex sets with nonempty interior). Then Vol(A + B) n Vol(A) n + Vol(B) n. Equality holds iff A and B are homothetic, i.e. equal up to translation and dilation. An equivalent form is that for convex bodies A, B R n and 0 < λ < 1 we have Vol((1 λ)a + λb) 1 n (1 λ)vol(a) 1 n + λvol(b) 1 n. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

23 The Brunn Minkowski inequality Let A, B R n be convex bodies (compact convex sets with nonempty interior). Then Vol(A + B) n Vol(A) n + Vol(B) n. Equality holds iff A and B are homothetic, i.e. equal up to translation and dilation. An equivalent form is that for convex bodies A, B R n and 0 < λ < 1 we have Vol((1 λ)a + λb) 1 n (1 λ)vol(a) 1 n + λvol(b) 1 n. Another equivalent form is that for convex bodies A, B R n and 0 < λ < 1 we have Vol((1 λ)a + λb) 1 n min(vol(a), Vol(B)). V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

24 Parallels between the EPI and the Brunn Minkowski inequality V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

25 Parallels between the EPI and the Brunn Minkowski inequality The distribution of a random variable corresponds to a (measurable) subset of R n. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

26 Parallels between the EPI and the Brunn Minkowski inequality The distribution of a random variable corresponds to a (measurable) subset of R n. Addition of independent random variables corresponds to the Minkowski sum. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

27 Parallels between the EPI and the Brunn Minkowski inequality The distribution of a random variable corresponds to a (measurable) subset of R n. Addition of independent random variables corresponds to the Minkowski sum. Balls play the role of spherically symmetric Gaussians. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

28 Parallels between the EPI and the Brunn Minkowski inequality The distribution of a random variable corresponds to a (measurable) subset of R n. Addition of independent random variables corresponds to the Minkowski sum. Balls play the role of spherically symmetric Gaussians. For instance, an equivalent form of the EPI is that if X and Y are independent R n -valued random variables, and 0 < λ < 1, then h( λx + 1 λy) λh(x) + (1 λ)h(y). V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

29 Outline 1 Setting the stage 2 The Brunn Minkowski inequality 3 Proofs of the EPI 4 Mrs. Gerber s Lemma 5 Young s inequality 6 Versions of the EPI for discrete random variables 7 EPI for Groups or Order 2 n 8 The Fourier transform on a finite abelian group 9 Young s inequality on a finite abelian group 10 Brunn Minkowski inequality on a finite abelian group 11 Speculation V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

30 Some history V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

31 Some history The EPI was conjectured by Shannon, who showed that X and Y being Gaussian with proportional covariance matrices is a stationary point of the difference between the entropy power of the sum and sum of the entropy powers. This did not exclude the possibility of it being a local maximum or a saddle point. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

32 Some history The EPI was conjectured by Shannon, who showed that X and Y being Gaussian with proportional covariance matrices is a stationary point of the difference between the entropy power of the sum and sum of the entropy powers. This did not exclude the possibility of it being a local maximum or a saddle point. Stam gave the first rigorous proof of the EPI in 1959, using de Bruijn s identity which relates Fisher information and differential entropy. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

33 Some history The EPI was conjectured by Shannon, who showed that X and Y being Gaussian with proportional covariance matrices is a stationary point of the difference between the entropy power of the sum and sum of the entropy powers. This did not exclude the possibility of it being a local maximum or a saddle point. Stam gave the first rigorous proof of the EPI in 1959, using de Bruijn s identity which relates Fisher information and differential entropy. Blachman s 1965 paper gives a simplified and succinct version of Stam s proof. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

34 Some history The EPI was conjectured by Shannon, who showed that X and Y being Gaussian with proportional covariance matrices is a stationary point of the difference between the entropy power of the sum and sum of the entropy powers. This did not exclude the possibility of it being a local maximum or a saddle point. Stam gave the first rigorous proof of the EPI in 1959, using de Bruijn s identity which relates Fisher information and differential entropy. Blachman s 1965 paper gives a simplified and succinct version of Stam s proof. Subsequently, several proofs have appeared, some of which will be mentioned here for our story. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

35 Blachman s (1965) proof strategy for the EPI following Stam (1959) V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

36 Blachman s (1965) proof strategy for the EPI following Stam (1959) Step 1: Prove de Bruijn s identity for scalar random variables V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

37 Blachman s (1965) proof strategy for the EPI following Stam (1959) Step 1: Prove de Bruijn s identity for scalar random variables Let X t = X + tn, where N N (0, 1). Let X p(x), and X + tn p t (x t ). V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

38 Blachman s (1965) proof strategy for the EPI following Stam (1959) Step 1: Prove de Bruijn s identity for scalar random variables Let X t = X + tn, where N N (0, 1). Let X p(x), and X + tn p t (x t ). De Bruijn s Identity d dt h(x t) = 1 2 J(X t) = 1 2 ( p t (x t )) 2 dx t x t p t (x t ) V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

39 Blachman s (1965) proof strategy for the EPI following Stam (1959) Step 1: Prove de Bruijn s identity for scalar random variables Let X t = X + tn, where N N (0, 1). Let X p(x), and X + tn p t (x t ). De Bruijn s Identity d dt h(x t) = 1 2 J(X t) = 1 2 The proof follows from differentiating h(x t ) = ( p t (x t )) 2 dx t x t p t (x t ) p t (x t ) log p t (x t )dx t, with respect to t, and using the heat equation t p t(x t ) = t p t(x 2 t ). V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

40 Blachman s proof strategy (contd.) V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

41 Blachman s proof strategy (contd.) Step 2: Prove Stam s inequality V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

42 Blachman s proof strategy (contd.) Step 2: Prove Stam s inequality If X and Y are independent, and Z = X + Y then Stam s Inequality J(Z) J(X ) J(Y ) V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

43 Blachman s proof strategy (contd.) Step 2: Prove Stam s inequality If X and Y are independent, and Z = X + Y then Stam s Inequality J(Z) J(X ) J(Y ) The proof goes by showing that ( E a p X (x) p X (x) + b p Y (y) ) p Y (y) Z = z = (a + b) p Z (z) p Z (z), for all a, b. Then squaring both sides, using Jensen s inequality and optimizing over the choice of a, b establishes the inequality. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

44 Blachman s proof strategy (contd.) V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

45 Blachman s proof strategy (contd.) Step 3: Set up the one parameter flow defined by the heat equation starting with the two initial distributions, with suitable time reparametrizations f (t), g(t) for each, respectively V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

46 Blachman s proof strategy (contd.) Step 3: Set up the one parameter flow defined by the heat equation starting with the two initial distributions, with suitable time reparametrizations f (t), g(t) for each, respectively Let f (t) and g(t) be increasing functions tending to infinity with t, and consider the function s(t) = e2h(x f ) + e 2h(Yg ) e 2h(Z f +g ). V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

47 Blachman s proof strategy (contd.) Step 3: Set up the one parameter flow defined by the heat equation starting with the two initial distributions, with suitable time reparametrizations f (t), g(t) for each, respectively Let f (t) and g(t) be increasing functions tending to infinity with t, and consider the function s(t) = e2h(x f ) + e 2h(Yg ) e 2h(Z f +g ). Intuitively, lim t s(t) = 1, since both initial distributions become increasingly Gaussian as time progresses along the flow. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

48 Blachman s proof strategy (contd.) Step 3: Set up the one parameter flow defined by the heat equation starting with the two initial distributions, with suitable time reparametrizations f (t), g(t) for each, respectively Let f (t) and g(t) be increasing functions tending to infinity with t, and consider the function s(t) = e2h(x f ) + e 2h(Yg ) e 2h(Z f +g ). Intuitively, lim t s(t) = 1, since both initial distributions become increasingly Gaussian as time progresses along the flow. Choosing f (t) = exp(2h(x f )) and g (t) = exp(2h(y g )), and using Stam s inequality gives s (t) 0. This gives s(0) 1, proving the EPI. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

49 Lieb s EPI proof as interpreted by Verdú and Guo V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

50 Lieb s EPI proof as interpreted by Verdú and Guo Reparametrization of the de Bruijn identity gives the Guo-Shamai-Verdú I-MMSE relation: and thus d dγ I (X ; γx + N) = 1 mmse(x, γ), 2 h(x ) = 1 2 log 2πe γ mmse(x, γ)dγ. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

51 Lieb s EPI proof as interpreted by Verdú and Guo Reparametrization of the de Bruijn identity gives the Guo-Shamai-Verdú I-MMSE relation: and thus d dγ I (X ; γx + N) = 1 mmse(x, γ), 2 h(x ) = 1 2 log 2πe γ mmse(x, γ)dγ. The EPI in its equivalent form, h(x 1 cos α + X 2 sin α) cos 2 αh(x 1 ) + sin 2 αh(x 2 ), can be proved using these ideas. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

52 Verdú and Guo s proof: Proof strategy V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

53 Verdú and Guo s proof: Proof strategy h(x 1 cos α + X 2 sin α) cos 2 αh(x 1 ) + sin 2 αh(x 2 ) = ( 1 mmse(x 1 cos α + X 2 sin α) cos 2 α mmse(x 1, γ) 2 ) + sin 2 α mmse(x 2, γ) dγ V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

54 Verdú and Guo s proof: Proof strategy h(x 1 cos α + X 2 sin α) cos 2 αh(x 1 ) + sin 2 αh(x 2 ) = ( 1 mmse(x 1 cos α + X 2 sin α) cos 2 α mmse(x 1, γ) 2 ) + sin 2 α mmse(x 2, γ) dγ Consider Z 1 = γx 1 + N 1, Z 2 = γx 2 + N 2 and Z = Z 1 cos α + Z 2 sin α. Then mmse(x 1 cos α + X 2 sin α Z) mmse(x Z 1, Z 2 ) immediately gives the term inside the integral is 0. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

55 Szarek and Voiculescu s proof of the EPI V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

56 Szarek and Voiculescu s proof of the EPI Szarek and Voiculescu (1996) give a proof of the EPI working directly with typical sets, using a restricted version of the Brunn-Minkowski inequality. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

57 Szarek and Voiculescu s proof of the EPI Szarek and Voiculescu (1996) give a proof of the EPI working directly with typical sets, using a restricted version of the Brunn-Minkowski inequality. The restricted Minkowski sum of sets A and B, based on Θ A B is defined as A + Θ B := {x + y : (x, y) Θ}. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

58 Szarek and Voiculescu s proof of the EPI Szarek and Voiculescu (1996) give a proof of the EPI working directly with typical sets, using a restricted version of the Brunn-Minkowski inequality. The restricted Minkowski sum of sets A and B, based on Θ A B is defined as A + Θ B := {x + y : (x, y) Θ}. Szarek and Voiculescu s restricted B-M inequality says: For every ɛ, there exists δ such that if Θ is large enough, viz V 2n (Θ) (1 δ) n V n (A)V n (B), then V n (A + Θ B) 2 n (1 ɛ)(vn (A) 2 n + Vn (B) 2 n ). V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

59 Szarek and Voiculescu s proof (contd.) V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

60 Szarek and Voiculescu s proof (contd.) To prove the EPI, Szarek and Voiculescu replace A and B by typical sets for n i.i.d. copies of X and Y respectively, and define Θ as all pairs (x n, y n ) (each marginal typical) such that x n + y n is typical for X + Y. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

61 Szarek and Voiculescu s proof (contd.) To prove the EPI, Szarek and Voiculescu replace A and B by typical sets for n i.i.d. copies of X and Y respectively, and define Θ as all pairs (x n, y n ) (each marginal typical) such that x n + y n is typical for X + Y. For this choice of restriction, they determine suitable sequence of typicality defining constants (ɛ n, δ n ) going to 0, thus proving the EPI. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

62 Outline 1 Setting the stage 2 The Brunn Minkowski inequality 3 Proofs of the EPI 4 Mrs. Gerber s Lemma 5 Young s inequality 6 Versions of the EPI for discrete random variables 7 EPI for Groups or Order 2 n 8 The Fourier transform on a finite abelian group 9 Young s inequality on a finite abelian group 10 Brunn Minkowski inequality on a finite abelian group 11 Speculation V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

63 Mrs. Gerber s Lemma for Gaussian random variables V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

64 Mrs. Gerber s Lemma for Gaussian random variables n log(e 2 n x + e 2 n y ) is the minimum achievable differential entropy 2 of X + Y where X and Y are independent R n -valued random variables with differential entropies x and y respectively. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

65 Mrs. Gerber s Lemma for Gaussian random variables n log(e 2 n x + e 2 n y ) is the minimum achievable differential entropy 2 of X + Y where X and Y are independent R n -valued random variables with differential entropies x and y respectively. Note that (x, y) n 2 log(e 2 n x + e 2 n y ) is convex in (x, y). V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

66 Mrs. Gerber s Lemma for Gaussian random variables n log(e 2 n x + e 2 n y ) is the minimum achievable differential entropy 2 of X + Y where X and Y are independent R n -valued random variables with differential entropies x and y respectively. Note that (x, y) n 2 log(e 2 n x + e 2 n y ) is convex in (x, y). In particular, for fixed x, is convex in y and for fixed y, is convex in x. y n 2 log(e 2 n x + e 2 n y ) x n 2 log(e 2 n x + e 2 n y ) V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

67 Outline 1 Setting the stage 2 The Brunn Minkowski inequality 3 Proofs of the EPI 4 Mrs. Gerber s Lemma 5 Young s inequality 6 Versions of the EPI for discrete random variables 7 EPI for Groups or Order 2 n 8 The Fourier transform on a finite abelian group 9 Young s inequality on a finite abelian group 10 Brunn Minkowski inequality on a finite abelian group 11 Speculation V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

68 Young s inequality V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

69 Young s inequality The Hausdorff-Young inequality gives an estimate on the norm of the Fourier transform: Ff p f p, for 1 p 2. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

70 Young s inequality The Hausdorff-Young inequality gives an estimate on the norm of the Fourier transform: Ff p f p, for 1 p 2. The sharp constant in the inequality was found by Beckner Ff p C p f p, where C p := p 1/p. p 1/p V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

71 Young s inequality The Hausdorff-Young inequality gives an estimate on the norm of the Fourier transform: Ff p f p, for 1 p 2. The sharp constant in the inequality was found by Beckner Ff p C p f p, where C p := p 1/p. p 1/p This leads to Young s inequality. If p, q, r > 0 satisfy 1 p + 1 q = r, then f g r C pc q C r f p g q, if p, q, r 1, f g r C pc q C r f p g q, if 1 p, q, r. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

72 Young s inequality The Hausdorff-Young inequality gives an estimate on the norm of the Fourier transform: Ff p f p, for 1 p 2. The sharp constant in the inequality was found by Beckner Ff p C p f p, where C p := p 1/p. p 1/p This leads to Young s inequality. If p, q, r > 0 satisfy 1 p + 1 q = r, then f g r C pc q C r f p g q, if p, q, r 1, f g r C pc q C r f p g q, if 1 p, q, r. The second half of this is called the reverse Young s inequality. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

73 EPI from the Young s inequality V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

74 EPI from the Young s inequality For p > 1, h p (X) := X f. p 1 p f p is called the Renyi entropy of V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

75 EPI from the Young s inequality For p > 1, h p (X) := p f 1 p p is called the Renyi entropy of X f. N p (X) := 1 /p 2π p p f 2p /n p is the Renyi entropy power of X f. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

76 EPI from the Young s inequality For p > 1, h p (X) := p f 1 p p is called the Renyi entropy of X f. N p (X) := 1 /p 2π p p f 2p /n p is the Renyi entropy power of X f. As p 1 these converge respectively to the differential entropy and the entropy power of X. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

77 EPI from the Young s inequality For p > 1, h p (X) := p f 1 p p is called the Renyi entropy of X f. N p (X) := 1 /p 2π p p f 2p /n p is the Renyi entropy power of X f. As p 1 these converge respectively to the differential entropy and the entropy power of X. r Given r > 1 and 0 < λ < 1 define p(r) := and (1 λ)+λr q(r) := r. λ+(1 λ)r V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

78 EPI from the Young s inequality For p > 1, h p (X) := p f 1 p p is called the Renyi entropy of X f. N p (X) := 1 /p 2π p p f 2p /n p is the Renyi entropy power of X f. As p 1 these converge respectively to the differential entropy and the entropy power of X. r Given r > 1 and 0 < λ < 1 define p(r) := and (1 λ)+λr r q(r) :=. λ+(1 λ)r Young s inequality gives N r (X + Y) ( Np (X) 1 λ ) 1 λ ( ) λ Nq (Y). λ V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

79 EPI from the Young s inequality For p > 1, h p (X) := p f 1 p p is called the Renyi entropy of X f. N p (X) := 1 /p 2π p p f 2p /n p is the Renyi entropy power of X f. As p 1 these converge respectively to the differential entropy and the entropy power of X. r Given r > 1 and 0 < λ < 1 define p(r) := and (1 λ)+λr r q(r) :=. λ+(1 λ)r Young s inequality gives N r (X + Y) ( Np (X) 1 λ ) 1 λ ( ) λ Nq (Y). λ Optimize over λ and let r 1 to get the EPI. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

80 Brunn Minkowski inequality from the reverse Young s inequality V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

81 Brunn Minkowski inequality from the reverse Young s inequality The limit as r 0 in Young s inequality leads to the Prékopa-Leindler inequality: If 0 < λ < 1 and f, g, h are nonnegative integrable functions on R n satisfying h((1 λ)x + λy) f (x) 1 λ g(y) λ for all x, y R n, then ( ) 1 λ ( ) λ h(x)dx R n f (x)dx R n g(x)dx R n. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

82 Brunn Minkowski inequality from the reverse Young s inequality The limit as r 0 in Young s inequality leads to the Prékopa-Leindler inequality: If 0 < λ < 1 and f, g, h are nonnegative integrable functions on R n satisfying h((1 λ)x + λy) f (x) 1 λ g(y) λ for all x, y R n, then ( ) 1 λ ( ) λ h(x)dx R n f (x)dx R n g(x)dx R n. Setting f := 1 A, g := 1 B, and h := 1 (1 λ)a+λb, this gives Vol((1 λ)a+λb) Vol(A) 1 λ Vol(B) λ min(vol(a), Vol(B)). V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

83 Outline 1 Setting the stage 2 The Brunn Minkowski inequality 3 Proofs of the EPI 4 Mrs. Gerber s Lemma 5 Young s inequality 6 Versions of the EPI for discrete random variables 7 EPI for Groups or Order 2 n 8 The Fourier transform on a finite abelian group 9 Young s inequality on a finite abelian group 10 Brunn Minkowski inequality on a finite abelian group 11 Speculation V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

84 Mrs. Gerber s lemma V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

85 Mrs. Gerber s lemma Let h be the binary entropy function, and h 1 be its inverse. Wyner and Ziv (1973) showed that for a binary X and arbitrary U, if Z Bern(p) is independent of (X, U) then h(x Z U) h(h 1 (H(X U) p). V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

86 Mrs. Gerber s lemma Let h be the binary entropy function, and h 1 be its inverse. Wyner and Ziv (1973) showed that for a binary X and arbitrary U, if Z Bern(p) is independent of (X, U) then h(x Z U) h(h 1 (H(X U) p). The result follows immediately from the following lemma, using Jensen s inequality - Lemma The function is convex on 0 x log 2. x h(h 1 (x) p) V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

87 Wyner and Ziv s proof of Mrs. Gerber s lemma V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

88 Wyner and Ziv s proof of Mrs. Gerber s lemma Introduce the parametrization α = 1 h 1 (x), let p = 1 a, and 2 2 differentiate wrt α, to get that f (x) 0 a(1 α 2 ) log 1 + α 1 α (1 aα)2 log 1 + aα 1 aα, which can be verified using the series expansion of log 1+x 1 x. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

89 Shamai and Wyner s EPI V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

90 Shamai and Wyner s EPI Shamai and Wyner show that - EPI for binary sequences For binary independent processes with entropies H(X ) and H(Y ), σ(z) σ(x ) σ(y ), where Z = X Y and σ(x ) = h 1 (H(X )), σ(y ) = h 1 (H(Y ) and σ(z) = h 1 (H(Z)). V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

91 Shamai and Wyner s proof strategy for the binary EPI V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

92 Shamai and Wyner s proof strategy for the binary EPI H(Z 0 Z 1 n ) H(Z 0 Z 1 n, X 1 n, Y 1 n ) = H(Z 0 X 1 n, Y 1 n ), = P(X 1 n = x)p(y 1 n = y)h(z 0 X 1 n = x, Y 1 n = y) = P(X 1 n = x)p(y 1 n = y)h(α(x) β(y)) Where α(x) = P(X 0 = 1 X 1 n = x), β(x) = P(Y 0 = 1 Y 1 n = y). V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

93 Shamai and Wyner s proof strategy for the binary EPI H(Z 0 Z 1 n ) H(Z 0 Z 1 n, X 1 n, Y 1 n ) = H(Z 0 X 1 n, Y 1 n ), = P(X 1 n = x)p(y 1 n = y)h(z 0 X 1 n = x, Y 1 n = y) = P(X 1 n = x)p(y 1 n = y)h(α(x) β(y)) Where α(x) = P(X 0 = 1 X 1 n = x), β(x) = P(Y 0 = 1 Y 1 n = y). Using convexity in MGL twice, summing over x and then y, we get H(Z 0 Z 1 n ) h(h 1 (H(X 0 X 1 n ) h 1 (H(Y 0 Y 1 n )). Taking n proves the result. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

94 Harremoës s EPI for the Binomial family V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

95 Harremoës s EPI for the Binomial family Let X n Binomial(n, 1 ), then for all m, n 1, 2 e 2H(Xn+Xm) e 2H(Xn) + e 2H(Xm). V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

96 Harremoës s EPI for the Binomial family Let X n Binomial(n, 1 ), then for all m, n 1, 2 e 2H(Xn+Xm) e 2H(Xn) + e 2H(Xm). The proof follows by showing that Y n = e2h(xn) is an increasing n sequence, and thus is super additive, i.e Y m+n Y m + Y n. Figure : Values of Y n V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

97 Our approach V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

98 Our approach Suppose the random variables take values on a finite abelian group G. We define the function f G : [0, log G ] [0, log G ] [0, log G ] by f G (x, y) = min H(X + Y ) H(X )=x,h(y )=y V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

99 Our approach Suppose the random variables take values on a finite abelian group G. We define the function f G : [0, log G ] [0, log G ] [0, log G ] by f G (x, y) = We now ask the questions- min H(X + Y ) H(X )=x,h(y )=y Does f G have a succinct description? V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

100 Our approach Suppose the random variables take values on a finite abelian group G. We define the function f G : [0, log G ] [0, log G ] [0, log G ] by f G (x, y) = We now ask the questions- min H(X + Y ) H(X )=x,h(y )=y Does f G have a succinct description? Does f G have any nice properties? V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

101 Our approach Suppose the random variables take values on a finite abelian group G. We define the function f G : [0, log G ] [0, log G ] [0, log G ] by f G (x, y) = We now ask the questions- min H(X + Y ) H(X )=x,h(y )=y Does f G have a succinct description? Does f G have any nice properties? We try to exploit the group structure to answer these questions. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

102 Mimicking Blachman s proof? V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

103 Mimicking Blachman s proof? Find a 1-parameter group to evolve the probability distributions towards the uniform distribution on the group? V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

104 Mimicking Blachman s proof? Find a 1-parameter group to evolve the probability distributions towards the uniform distribution on the group? Perhaps the analogs of the Gaussian distributions are the distributions of the form 1 G 1 p(0) = 1 ɛ, p(g) = ɛ for g 0, 0 ɛ? G 1 G V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

105 Mimicking Blachman s proof? Find a 1-parameter group to evolve the probability distributions towards the uniform distribution on the group? Perhaps the analogs of the Gaussian distributions are the distributions of the form 1 G 1 p(0) = 1 ɛ, p(g) = ɛ for g 0, 0 ɛ? G 1 G But these are not extremal for the EPI. For instance, for G = Z 2 Z 2, if ɛ is chosen so that h(ɛ) + ɛ log 3 = log 2 (note that ɛ < 3 ), then we have 4 h(1 2ɛ ɛ2 ) + 2ɛ(1 2 ɛ) log 3 > log 2. 3 V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

106 Mimicking Blachman s proof? Find a 1-parameter group to evolve the probability distributions towards the uniform distribution on the group? Perhaps the analogs of the Gaussian distributions are the distributions of the form 1 G 1 p(0) = 1 ɛ, p(g) = ɛ for g 0, 0 ɛ? G 1 G But these are not extremal for the EPI. For instance, for G = Z 2 Z 2, if ɛ is chosen so that h(ɛ) + ɛ log 3 = log 2 (note that ɛ < 3 ), then we have 4 h(1 2ɛ ɛ2 ) + 2ɛ(1 2 ɛ) log 3 > log 2. 3 However, for the uniform distribution on a subgroup, its convolution with itself has entropy log 2. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

107 Mimicking Lieb s proof? V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

108 Mimicking Lieb s proof? Given group valued random variables X 1 and X 2, what is the analog of X 1 cos α + X 2 sin α? V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

109 Mimicking Lieb s proof? Given group valued random variables X 1 and X 2, what is the analog of X 1 cos α + X 2 sin α? There is a version of Stam s inequality for finite abelian groups (Gibilisco and Isola, 2008). Given a set of generators Γ := {γ 1, γ 2,..., γ m } for the group G, define the Fisher information of a random variable X by J(X ) := ( ) px (g) p X (γ 1 2 g) p X (g). p X (g) γ Γ g G 1 Then one has 1 J(X 1 +X 2 ) random variables X 1 and X J(X 1 ) J(X 2 ) for independent G-valued V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

110 Mimicking Lieb s proof? Given group valued random variables X 1 and X 2, what is the analog of X 1 cos α + X 2 sin α? There is a version of Stam s inequality for finite abelian groups (Gibilisco and Isola, 2008). Given a set of generators Γ := {γ 1, γ 2,..., γ m } for the group G, define the Fisher information of a random variable X by J(X ) := ( ) px (g) p X (γ 1 2 g) p X (g). p X (g) γ Γ g G 1 Then one has J(X 1 +X 2 ) J(X 1 ) J(X 2 for independent G-valued ) random variables X 1 and X 2. However this notion of Fisher information is merely mimicking formal properties of the Fisher information from the continuous case. It seems to have little to do with estimation, which, for finite groups ought to refer to likelihood ratio type quantities. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

111 Mimicking Szarek and Voiculescu s proof? V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

112 Mimicking Szarek and Voiculescu s proof? This is a promising direction. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

113 Mimicking Szarek and Voiculescu s proof? This is a promising direction. The question remains, what is the correct analog of the restricted Minkowski sum? V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

114 Mimicking Szarek and Voiculescu s proof? This is a promising direction. The question remains, what is the correct analog of the restricted Minkowski sum? The key lemma driving the proof of Szarek and Voiculescu appears to be Euclidean in character. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

115 Outline 1 Setting the stage 2 The Brunn Minkowski inequality 3 Proofs of the EPI 4 Mrs. Gerber s Lemma 5 Young s inequality 6 Versions of the EPI for discrete random variables 7 EPI for Groups or Order 2 n 8 The Fourier transform on a finite abelian group 9 Young s inequality on a finite abelian group 10 Brunn Minkowski inequality on a finite abelian group 11 Speculation V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

116 Our results V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

117 Our results For a group G of order 2 n, we explicitly describe f G in terms of f Z2. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

118 Our results For a group G of order 2 n, we explicitly describe f G in terms of f Z2. We also describe the minimum entropy achieving distributions on G, these distributions are analogs of Gaussians in this sense. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

119 Our results For a group G of order 2 n, we explicitly describe f G in terms of f Z2. We also describe the minimum entropy achieving distributions on G, these distributions are analogs of Gaussians in this sense. The f G (x, y) function obtained has the property that it is convex in x for a fixed y. This is yet another version of Mrs. Gerber s Lemma. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

120 Statement of our results V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

121 Statement of our results Theorem f G depends only on the size of G, and is denoted by f 2 n, where f 2 (x k log 2, y k log 2) + k log 2, f 2 n(x, y) = if k log 2 x, y (k + 1) log 2, max(x, y) otherwise. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

122 Visualizing our results V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

123 Visualizing our results (contd.) V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

124 A conjecture V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

125 A conjecture Generalized Mrs. Gerber s Lemma If G is a finite group, then f G (x, y) is convex in x for a fixed y, and by symmetry convex in y for a fixed x. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

126 A conjecture Generalized Mrs. Gerber s Lemma If G is a finite group, then f G (x, y) is convex in x for a fixed y, and by symmetry convex in y for a fixed x. If one is less optimistic, one might make this conjecture only for finite abelian groups. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

127 A conjecture Generalized Mrs. Gerber s Lemma If G is a finite group, then f G (x, y) is convex in x for a fixed y, and by symmetry convex in y for a fixed x. If one is less optimistic, one might make this conjecture only for finite abelian groups. Simulations for Z 3 and Z 5 seem to support this conjecture. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

128 A conjecture Generalized Mrs. Gerber s Lemma If G is a finite group, then f G (x, y) is convex in x for a fixed y, and by symmetry convex in y for a fixed x. If one is less optimistic, one might make this conjecture only for finite abelian groups. Simulations for Z 3 and Z 5 seem to support this conjecture. We saw already that f 2 n satisfies the conjecture. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

129 EPI for abelian groups of order 2 n : proof strategy V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

130 EPI for abelian groups of order 2 n : proof strategy First observe that for Z 2 the function f Z2 (x, y) is explicitly given by f Z2 (x, y) = h(h 1 (x) h 1 (y)). V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

131 EPI for abelian groups of order 2 n : proof strategy First observe that for Z 2 the function f Z2 (x, y) is explicitly given by f Z2 (x, y) = h(h 1 (x) h 1 (y)). The proof proceeds by first proving the EPI for Z 4, using convexity properties of f Z2 and concavity of entropy. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

132 EPI for abelian groups of order 2 n : proof strategy First observe that for Z 2 the function f Z2 (x, y) is explicitly given by f Z2 (x, y) = h(h 1 (x) h 1 (y)). The proof proceeds by first proving the EPI for Z 4, using convexity properties of f Z2 and concavity of entropy. We then use induction, where we assume that the result holds for all groups of size 2 n and prove it groups of size 2 n+1. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

133 EPI for abelian groups of order 2 n : proof strategy First observe that for Z 2 the function f Z2 (x, y) is explicitly given by f Z2 (x, y) = h(h 1 (x) h 1 (y)). The proof proceeds by first proving the EPI for Z 4, using convexity properties of f Z2 and concavity of entropy. We then use induction, where we assume that the result holds for all groups of size 2 n and prove it groups of size 2 n+1. The induction part of the proof has almost exactly the same structure as the proof for Z 4. Thus proving the result for Z 4 is the key step in our proof. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

134 Proof for Z 4 V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

135 Proof for Z 4 For Z 4, splitting the support of distributions on Z 4 into two parts, {0, 2} (even part) and{1, 3} (odd part) and using precisely two ingredients: concavity of entropy, and convexity in MGL we arrive at the lower bound Lower bound f 4 (x, y) min u,v f 2(u, v) + f 2 (x u, y v) V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

136 Proof for Z 4 For Z 4, splitting the support of distributions on Z 4 into two parts, {0, 2} (even part) and{1, 3} (odd part) and using precisely two ingredients: concavity of entropy, and convexity in MGL we arrive at the lower bound Lower bound Lemma f Z2 f 4 (x, y) min u,v f 2(u, v) + f 2 (x u, y v) To evaluate the minimum, we prove some additional properties of f Z2, specifically regarding its behavior along lines through the origin. The key Lemma we use is (and by symmetry, f Z 2 ) strictly decreases along lines through the x y origin V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

137 Proving the lemma V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

138 Proving the lemma We use the parametrization h 1 (x) = p and h 1 (y) = q. Our strategy involves brute force differentiation, and a series of steps where we conclude that proving the lemma is equivalent to proving F (p) = p 2 (log(p)) 2 (1 p) 2 (log(1 p)) 2 + (1 2p) (log(p) log(1 p) + p log(p) + (1 p) log(1 p)) 0 V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

139 Proving the lemma We use the parametrization h 1 (x) = p and h 1 (y) = q. Our strategy involves brute force differentiation, and a series of steps where we conclude that proving the lemma is equivalent to proving F (p) = p 2 (log(p)) 2 (1 p) 2 (log(1 p)) 2 + (1 2p) (log(p) log(1 p) + p log(p) + (1 p) log(1 p)) 0 We show that proving F (p) 0 is the same as proving d 5 dp 5 F (p) 0, we differentiate F five times, use a polynomial approximation of log, and finally use Sturm s theorem to prove that the resulting polynomial is non-positive. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

140 Outline 1 Setting the stage 2 The Brunn Minkowski inequality 3 Proofs of the EPI 4 Mrs. Gerber s Lemma 5 Young s inequality 6 Versions of the EPI for discrete random variables 7 EPI for Groups or Order 2 n 8 The Fourier transform on a finite abelian group 9 Young s inequality on a finite abelian group 10 Brunn Minkowski inequality on a finite abelian group 11 Speculation V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

141 Norm of the Fourier transform on a finite abelian group V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

142 Norm of the Fourier transform on a finite abelian group Every finite abelian group G has a dual group Ĝ comprised of the characters of G, i.e. the homomorphisms from G to the unit circle in the complex plane. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

143 Norm of the Fourier transform on a finite abelian group Every finite abelian group G has a dual group Ĝ comprised of the characters of G, i.e. the homomorphisms from G to the unit circle in the complex plane. Ĝ has the same cardinality and structure as G and G can be thought of as the dual of Ĝ. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

144 Norm of the Fourier transform on a finite abelian group Every finite abelian group G has a dual group Ĝ comprised of the characters of G, i.e. the homomorphisms from G to the unit circle in the complex plane. Ĝ has the same cardinality and structure as G and G can be thought of as the dual of Ĝ. The Fourier transform on the group is the map from functions on 1 G to functions on Ĝ given by F(f )(γ) := G g γ(g)f (g). 1/2 V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

145 Norm of the Fourier transform on a finite abelian group Every finite abelian group G has a dual group Ĝ comprised of the characters of G, i.e. the homomorphisms from G to the unit circle in the complex plane. Ĝ has the same cardinality and structure as G and G can be thought of as the dual of Ĝ. The Fourier transform on the group is the map from functions on 1 G to functions on Ĝ given by F(f )(γ) := G g γ(g)f (g). 1/2 With counting measure on G and Ĝ, we have, for p, q > 0, F : L p q (G) L (Ĝ). What is the norm of this map? V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

146 Norm of the Fourier transform on a finite abelian group Every finite abelian group G has a dual group Ĝ comprised of the characters of G, i.e. the homomorphisms from G to the unit circle in the complex plane. Ĝ has the same cardinality and structure as G and G can be thought of as the dual of Ĝ. The Fourier transform on the group is the map from functions on 1 G to functions on Ĝ given by F(f )(γ) := G g γ(g)f (g). 1/2 With counting measure on G and Ĝ, we have, for p, q > 0, F : L p q (G) L (Ĝ). What is the norm of this map? Resolved by Gilbert and Rzeszotnik (2010). V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

147 Visualizing the norm of the FT on a finite abelian group Norm of the Fourier Transform on a finite abelian group V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

148 Outline 1 Setting the stage 2 The Brunn Minkowski inequality 3 Proofs of the EPI 4 Mrs. Gerber s Lemma 5 Young s inequality 6 Versions of the EPI for discrete random variables 7 EPI for Groups or Order 2 n 8 The Fourier transform on a finite abelian group 9 Young s inequality on a finite abelian group 10 Brunn Minkowski inequality on a finite abelian group 11 Speculation V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

149 Young s inequality on a finite abelian group V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

150 Young s inequality on a finite abelian group There is a forward Young s inequality, which reads: For every p, q, r 1 such that 1 p + 1 q = r we have f g r f p g q. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

151 Outline 1 Setting the stage 2 The Brunn Minkowski inequality 3 Proofs of the EPI 4 Mrs. Gerber s Lemma 5 Young s inequality 6 Versions of the EPI for discrete random variables 7 EPI for Groups or Order 2 n 8 The Fourier transform on a finite abelian group 9 Young s inequality on a finite abelian group 10 Brunn Minkowski inequality on a finite abelian group 11 Speculation V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

152 Brunn Minkowski inequality on a finite abelian group V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

153 Brunn Minkowski inequality on a finite abelian group For a finite abelian group G, consider the function µ G (r, s) := min{ A + B : A, B G, A = r, B = s}. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

154 Brunn Minkowski inequality on a finite abelian group For a finite abelian group G, consider the function µ G (r, s) := min{ A + B : A, B G, A = r, B = s}. Then we have µ G (r, s) = min d divides G ( r d + s d 1) d. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

155 Brunn Minkowski inequality on a finite abelian group For a finite abelian group G, consider the function µ G (r, s) := min{ A + B : A, B G, A = r, B = s}. Then we have µ G (r, s) = min d divides G ( r d + s d 1) d. In particular for G a finite abelian group of size n = 2 k, we have µ G (r, s) = r s where r s is the Hopf-Stiefel function defined as the smallest positive integer m for which the polynomial (x + y) m falls in the ideal generated by x r and y s in the polynomial ring F 2 [x, y]. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

156 Brunn Minkowski inequality on a finite abelian group For a finite abelian group G, consider the function µ G (r, s) := min{ A + B : A, B G, A = r, B = s}. Then we have µ G (r, s) = min d divides G ( r d + s d 1) d. In particular for G a finite abelian group of size n = 2 k, we have µ G (r, s) = r s where r s is the Hopf-Stiefel function defined as the smallest positive integer m for which the polynomial (x + y) m falls in the ideal generated by x r and y s in the polynomial ring F 2 [x, y]. Involved history of partial results. See Eliahu and Kervaire (2007) for the history. V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

157 Outline 1 Setting the stage 2 The Brunn Minkowski inequality 3 Proofs of the EPI 4 Mrs. Gerber s Lemma 5 Young s inequality 6 Versions of the EPI for discrete random variables 7 EPI for Groups or Order 2 n 8 The Fourier transform on a finite abelian group 9 Young s inequality on a finite abelian group 10 Brunn Minkowski inequality on a finite abelian group 11 Speculation V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

158 Questions V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

159 Questions What form does the reverse Young s inequality take on finite abelian groups? V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

160 Questions What form does the reverse Young s inequality take on finite abelian groups? Less ambitiously, what form does the reverse Young s inequality take on finite groups of size 2 k? V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

161 Questions What form does the reverse Young s inequality take on finite abelian groups? Less ambitiously, what form does the reverse Young s inequality take on finite groups of size 2 k? Is the EPI we proved for finite abelian groups of size 2 k derivable as a limit from the Young s inequality for such groups? V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

162 Questions What form does the reverse Young s inequality take on finite abelian groups? Less ambitiously, what form does the reverse Young s inequality take on finite groups of size 2 k? Is the EPI we proved for finite abelian groups of size 2 k derivable as a limit from the Young s inequality for such groups? Is the Brunn Minkowski inequality for finite abelian groups of size 2 k (expressed via the Hopf-Stiefel function) derivable as a limit from the appropriate reverse Young s inequality for such groups? V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

163 Questions What form does the reverse Young s inequality take on finite abelian groups? Less ambitiously, what form does the reverse Young s inequality take on finite groups of size 2 k? Is the EPI we proved for finite abelian groups of size 2 k derivable as a limit from the Young s inequality for such groups? Is the Brunn Minkowski inequality for finite abelian groups of size 2 k (expressed via the Hopf-Stiefel function) derivable as a limit from the appropriate reverse Young s inequality for such groups? Is the general Brunn Minkowski inequality for finite abelian groups derivable as a limit from the appropriate reverse Young s inequality for such groups? V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

164 Questions What form does the reverse Young s inequality take on finite abelian groups? Less ambitiously, what form does the reverse Young s inequality take on finite groups of size 2 k? Is the EPI we proved for finite abelian groups of size 2 k derivable as a limit from the Young s inequality for such groups? Is the Brunn Minkowski inequality for finite abelian groups of size 2 k (expressed via the Hopf-Stiefel function) derivable as a limit from the appropriate reverse Young s inequality for such groups? Is the general Brunn Minkowski inequality for finite abelian groups derivable as a limit from the appropriate reverse Young s inequality for such groups? Will a limit of the Young s inequality for a general finite Abelian group give rise to an EPI for each such group? V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

165 The End V. Jog and V. Anantharam (UC Berkeley) EPI December 12, / 57

Steiner s formula and large deviations theory

Steiner s formula and large deviations theory Steiner s formula and large deviations theory Venkat Anantharam EECS Department University of California, Berkeley May 19, 2015 Simons Conference on Networks and Stochastic Geometry Blanton Museum of Art

More information

A glimpse into convex geometry. A glimpse into convex geometry

A glimpse into convex geometry. A glimpse into convex geometry A glimpse into convex geometry 5 \ þ ÏŒÆ Two basis reference: 1. Keith Ball, An elementary introduction to modern convex geometry 2. Chuanming Zong, What is known about unit cubes Convex geometry lies

More information

The Entropy Power Inequality and Mrs. Gerber s Lemma for Groups of order 2 n

The Entropy Power Inequality and Mrs. Gerber s Lemma for Groups of order 2 n The Entrop Power Inequalit and Mrs. Gerber s Lemma for Groups of order 2 n Varun Jog EECS, UC Berkele Berkele, CA-94720 Email: varunjog@eecs.berkele.edu Venkat Anantharam EECS, UC Berkele Berkele, CA-94720

More information

Entropy power inequality for a family of discrete random variables

Entropy power inequality for a family of discrete random variables 20 IEEE International Symposium on Information Theory Proceedings Entropy power inequality for a family of discrete random variables Naresh Sharma, Smarajit Das and Siddharth Muthurishnan School of Technology

More information

Yet Another Proof of the Entropy Power Inequality

Yet Another Proof of the Entropy Power Inequality IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 63, NO. 6, JUNE 2017 3595 Yet Another Proof of the Entropy Power Inequality Olivier Rioul, Member, IEEE Abstract Yet another simple proof of the entropy power

More information

Monotonicity of entropy and Fisher information: a quick proof via maximal correlation

Monotonicity of entropy and Fisher information: a quick proof via maximal correlation Communications in Information and Systems Volume 16, Number 2, 111 115, 2016 Monotonicity of entropy and Fisher information: a quick proof via maximal correlation Thomas A. Courtade A simple proof is given

More information

Around the Brunn-Minkowski inequality

Around the Brunn-Minkowski inequality Around the Brunn-Minkowski inequality Andrea Colesanti Technische Universität Berlin - Institut für Mathematik January 28, 2015 Summary Summary The Brunn-Minkowski inequality Summary The Brunn-Minkowski

More information

A concavity property for the reciprocal of Fisher information and its consequences on Costa s EPI

A concavity property for the reciprocal of Fisher information and its consequences on Costa s EPI A concavity property for the reciprocal of Fisher information and its consequences on Costa s EPI Giuseppe Toscani October 0, 204 Abstract. We prove that the reciprocal of Fisher information of a logconcave

More information

A Sharpened Hausdorff-Young Inequality

A Sharpened Hausdorff-Young Inequality A Sharpened Hausdorff-Young Inequality Michael Christ University of California, Berkeley IPAM Workshop Kakeya Problem, Restriction Problem, Sum-Product Theory and perhaps more May 5, 2014 Hausdorff-Young

More information

Probabilistic Methods in Asymptotic Geometric Analysis.

Probabilistic Methods in Asymptotic Geometric Analysis. Probabilistic Methods in Asymptotic Geometric Analysis. C. Hugo Jiménez PUC-RIO, Brazil September 21st, 2016. Colmea. RJ 1 Origin 2 Normed Spaces 3 Distribution of volume of high dimensional convex bodies

More information

Chapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space.

Chapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space. Chapter 1 Preliminaries The purpose of this chapter is to provide some basic background information. Linear Space Hilbert Space Basic Principles 1 2 Preliminaries Linear Space The notion of linear space

More information

ECE 4400:693 - Information Theory

ECE 4400:693 - Information Theory ECE 4400:693 - Information Theory Dr. Nghi Tran Lecture 8: Differential Entropy Dr. Nghi Tran (ECE-University of Akron) ECE 4400:693 Lecture 1 / 43 Outline 1 Review: Entropy of discrete RVs 2 Differential

More information

Fisher information and Stam inequality on a finite group

Fisher information and Stam inequality on a finite group Fisher information and Stam inequality on a finite group Paolo Gibilisco and Tommaso Isola February, 2008 Abstract We prove a discrete version of Stam inequality for random variables taking values on a

More information

Entropy and the Additive Combinatorics of Probability Densities on LCA groups

Entropy and the Additive Combinatorics of Probability Densities on LCA groups Entropy and the Additive Combinatorics of Probability Densities on LCA groups Mokshay Madiman University of Delaware Based on joint work with Ioannis Kontoyiannis, Athens Univ. of Economics Jiange Li,

More information

BRUNN-MINKOWSKI INEQUALITIES IN PRODUCT METRIC MEASURE SPACES

BRUNN-MINKOWSKI INEQUALITIES IN PRODUCT METRIC MEASURE SPACES BRUNN-MINKOWSKI INEQUALITIES IN PRODUCT METRIC MEASURE SPACES MANUEL RITORÉ AND JESÚS YEPES NICOLÁS Abstract. Given one metric measure space satisfying a linear Brunn-Minkowski inequality, and a second

More information

LECTURE 2. Convexity and related notions. Last time: mutual information: definitions and properties. Lecture outline

LECTURE 2. Convexity and related notions. Last time: mutual information: definitions and properties. Lecture outline LECTURE 2 Convexity and related notions Last time: Goals and mechanics of the class notation entropy: definitions and properties mutual information: definitions and properties Lecture outline Convexity

More information

Nonparametric estimation of. s concave and log-concave densities: alternatives to maximum likelihood

Nonparametric estimation of. s concave and log-concave densities: alternatives to maximum likelihood Nonparametric estimation of s concave and log-concave densities: alternatives to maximum likelihood Jon A. Wellner University of Washington, Seattle Cowles Foundation Seminar, Yale University November

More information

Probability and Measure

Probability and Measure Part II Year 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2018 84 Paper 4, Section II 26J Let (X, A) be a measurable space. Let T : X X be a measurable map, and µ a probability

More information

much more on minimax (order bounds) cf. lecture by Iain Johnstone

much more on minimax (order bounds) cf. lecture by Iain Johnstone much more on minimax (order bounds) cf. lecture by Iain Johnstone http://www-stat.stanford.edu/~imj/wald/wald1web.pdf today s lecture parametric estimation, Fisher information, Cramer-Rao lower bound:

More information

Log-concave distributions: definitions, properties, and consequences

Log-concave distributions: definitions, properties, and consequences Log-concave distributions: definitions, properties, and consequences Jon A. Wellner University of Washington, Seattle; visiting Heidelberg Seminaire, Institut de Mathématiques de Toulouse 28 February 202

More information

7.1 Coupling from the Past

7.1 Coupling from the Past Georgia Tech Fall 2006 Markov Chain Monte Carlo Methods Lecture 7: September 12, 2006 Coupling from the Past Eric Vigoda 7.1 Coupling from the Past 7.1.1 Introduction We saw in the last lecture how Markov

More information

Theory of Probability Fall 2008

Theory of Probability Fall 2008 MIT OpenCourseWare http://ocw.mit.edu 8.75 Theory of Probability Fall 008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. Section Prekopa-Leindler inequality,

More information

CONVOLUTION ON L p. JENS GERLACH CHRISTENSEN

CONVOLUTION ON L p. JENS GERLACH CHRISTENSEN CONVOLUTION ON L p. JENS GERLACH CHRISTENSEN 1. Introduction. This paper will look in depth at the L p -conjecture and Young s inequality. It s based on an article from 1990 by Sadahiro Saeki. I chose

More information

On duality gap in linear conic problems

On duality gap in linear conic problems On duality gap in linear conic problems C. Zălinescu Abstract In their paper Duality of linear conic problems A. Shapiro and A. Nemirovski considered two possible properties (A) and (B) for dual linear

More information

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016 Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016 1 Entropy Since this course is about entropy maximization,

More information

Logarithmic Sobolev Inequalities

Logarithmic Sobolev Inequalities Logarithmic Sobolev Inequalities M. Ledoux Institut de Mathématiques de Toulouse, France logarithmic Sobolev inequalities what they are, some history analytic, geometric, optimal transportation proofs

More information

Math 341: Convex Geometry. Xi Chen

Math 341: Convex Geometry. Xi Chen Math 341: Convex Geometry Xi Chen 479 Central Academic Building, University of Alberta, Edmonton, Alberta T6G 2G1, CANADA E-mail address: xichen@math.ualberta.ca CHAPTER 1 Basics 1. Euclidean Geometry

More information

An upper bound on l q norms of noisy functions

An upper bound on l q norms of noisy functions Electronic Collouium on Computational Complexity, Report No. 68 08 An upper bound on l norms of noisy functions Alex Samorodnitsky Abstract Let T ɛ, 0 ɛ /, be the noise operator acting on functions on

More information

On the Complete Monotonicity of Heat Equation

On the Complete Monotonicity of Heat Equation [Pop-up Salon, Maths, SJTU, 2019/01/10] On the Complete Monotonicity of Heat Equation Fan Cheng John Hopcroft Center Computer Science and Engineering Shanghai Jiao Tong University chengfan@sjtu.edu.cn

More information

LEGENDRE POLYNOMIALS AND APPLICATIONS. We construct Legendre polynomials and apply them to solve Dirichlet problems in spherical coordinates.

LEGENDRE POLYNOMIALS AND APPLICATIONS. We construct Legendre polynomials and apply them to solve Dirichlet problems in spherical coordinates. LEGENDRE POLYNOMIALS AND APPLICATIONS We construct Legendre polynomials and apply them to solve Dirichlet problems in spherical coordinates.. Legendre equation: series solutions The Legendre equation is

More information

Classnotes - MA Series and Matrices

Classnotes - MA Series and Matrices Classnotes - MA-2 Series and Matrices Department of Mathematics Indian Institute of Technology Madras This classnote is only meant for academic use. It is not to be used for commercial purposes. For suggestions

More information

Exercises: Brunn, Minkowski and convex pie

Exercises: Brunn, Minkowski and convex pie Lecture 1 Exercises: Brunn, Minkowski and convex pie Consider the following problem: 1.1 Playing a convex pie Consider the following game with two players - you and me. I am cooking a pie, which should

More information

MGMT 69000: Topics in High-dimensional Data Analysis Falll 2016

MGMT 69000: Topics in High-dimensional Data Analysis Falll 2016 MGMT 69000: Topics in High-dimensional Data Analysis Falll 2016 Lecture 14: Information Theoretic Methods Lecturer: Jiaming Xu Scribe: Hilda Ibriga, Adarsh Barik, December 02, 2016 Outline f-divergence

More information

Kernel Density Estimation

Kernel Density Estimation EECS 598: Statistical Learning Theory, Winter 2014 Topic 19 Kernel Density Estimation Lecturer: Clayton Scott Scribe: Yun Wei, Yanzhen Deng Disclaimer: These notes have not been subjected to the usual

More information

ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016

ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 ECE598: Information-theoretic methods in high-dimensional statistics Spring 06 Lecture : Mutual Information Method Lecturer: Yihong Wu Scribe: Jaeho Lee, Mar, 06 Ed. Mar 9 Quick review: Assouad s lemma

More information

Locally convex spaces, the hyperplane separation theorem, and the Krein-Milman theorem

Locally convex spaces, the hyperplane separation theorem, and the Krein-Milman theorem 56 Chapter 7 Locally convex spaces, the hyperplane separation theorem, and the Krein-Milman theorem Recall that C(X) is not a normed linear space when X is not compact. On the other hand we could use semi

More information

Remarks on Extremization Problems Related To Young s Inequality

Remarks on Extremization Problems Related To Young s Inequality Remarks on Extremization Problems Related To Young s Inequality Michael Christ University of California, Berkeley University of Wisconsin May 18, 2016 Part 1: Introduction Young s convolution inequality

More information

Lecture 5 - Information theory

Lecture 5 - Information theory Lecture 5 - Information theory Jan Bouda FI MU May 18, 2012 Jan Bouda (FI MU) Lecture 5 - Information theory May 18, 2012 1 / 42 Part I Uncertainty and entropy Jan Bouda (FI MU) Lecture 5 - Information

More information

EE/Stats 376A: Homework 7 Solutions Due on Friday March 17, 5 pm

EE/Stats 376A: Homework 7 Solutions Due on Friday March 17, 5 pm EE/Stats 376A: Homework 7 Solutions Due on Friday March 17, 5 pm 1. Feedback does not increase the capacity. Consider a channel with feedback. We assume that all the recieved outputs are sent back immediately

More information

Lecture 2: August 31

Lecture 2: August 31 0-704: Information Processing and Learning Fall 206 Lecturer: Aarti Singh Lecture 2: August 3 Note: These notes are based on scribed notes from Spring5 offering of this course. LaTeX template courtesy

More information

A Simple Proof of the Entropy-Power Inequality

A Simple Proof of the Entropy-Power Inequality A Simple Proof of the Entropy-Power Inequality via Properties of Mutual Information Olivier Rioul Dept. ComElec, GET/Telecom Paris (ENST) ParisTech Institute & CNRS LTCI Paris, France Email: olivier.rioul@enst.fr

More information

Algorithmic Convex Geometry

Algorithmic Convex Geometry Algorithmic Convex Geometry August 2011 2 Contents 1 Overview 5 1.1 Learning by random sampling.................. 5 2 The Brunn-Minkowski Inequality 7 2.1 The inequality.......................... 8 2.1.1

More information

Chapter 8: Differential entropy. University of Illinois at Chicago ECE 534, Natasha Devroye

Chapter 8: Differential entropy. University of Illinois at Chicago ECE 534, Natasha Devroye Chapter 8: Differential entropy Chapter 8 outline Motivation Definitions Relation to discrete entropy Joint and conditional differential entropy Relative entropy and mutual information Properties AEP for

More information

1 An Elementary Introduction to Monotone Transportation

1 An Elementary Introduction to Monotone Transportation An Elementary Introduction to Monotone Transportation after K. Ball [3] A summary written by Paata Ivanisvili Abstract We outline existence of the Brenier map. As an application we present simple proofs

More information

Heat equation and the sharp Young s inequality

Heat equation and the sharp Young s inequality Noname manuscript No. will be inserted by the editor) Heat equation and the sharp Young s inequality Giuseppe Toscani the date of receipt and acceptance should be inserted later Abstract We show that the

More information

High-dimensional distributions with convexity properties

High-dimensional distributions with convexity properties High-dimensional distributions with convexity properties Bo az Klartag Tel-Aviv University A conference in honor of Charles Fefferman, Princeton, May 2009 High-Dimensional Distributions We are concerned

More information

Exercises with solutions (Set D)

Exercises with solutions (Set D) Exercises with solutions Set D. A fair die is rolled at the same time as a fair coin is tossed. Let A be the number on the upper surface of the die and let B describe the outcome of the coin toss, where

More information

Tangent spaces, normals and extrema

Tangent spaces, normals and extrema Chapter 3 Tangent spaces, normals and extrema If S is a surface in 3-space, with a point a S where S looks smooth, i.e., without any fold or cusp or self-crossing, we can intuitively define the tangent

More information

(x, y) = d(x, y) = x y.

(x, y) = d(x, y) = x y. 1 Euclidean geometry 1.1 Euclidean space Our story begins with a geometry which will be familiar to all readers, namely the geometry of Euclidean space. In this first chapter we study the Euclidean distance

More information

Chapter 2: Unconstrained Extrema

Chapter 2: Unconstrained Extrema Chapter 2: Unconstrained Extrema Math 368 c Copyright 2012, 2013 R Clark Robinson May 22, 2013 Chapter 2: Unconstrained Extrema 1 Types of Sets Definition For p R n and r > 0, the open ball about p of

More information

Introduction to Real Analysis Alternative Chapter 1

Introduction to Real Analysis Alternative Chapter 1 Christopher Heil Introduction to Real Analysis Alternative Chapter 1 A Primer on Norms and Banach Spaces Last Updated: March 10, 2018 c 2018 by Christopher Heil Chapter 1 A Primer on Norms and Banach Spaces

More information

Estimation in Gaussian Noise: Properties of the Minimum Mean-Square Error

Estimation in Gaussian Noise: Properties of the Minimum Mean-Square Error Estimation in Gaussian Noise: Properties of the Minimum Mean-Square Error Dongning Guo, Yihong Wu, Shlomo Shamai (Shitz), and Sergio Verdú Abstract arxiv:1004.333v1 [cs.it] 0 Apr 010 Consider the minimum

More information

EECS 229A Spring 2007 * * (a) By stationarity and the chain rule for entropy, we have

EECS 229A Spring 2007 * * (a) By stationarity and the chain rule for entropy, we have EECS 229A Spring 2007 * * Solutions to Homework 3 1. Problem 4.11 on pg. 93 of the text. Stationary processes (a) By stationarity and the chain rule for entropy, we have H(X 0 ) + H(X n X 0 ) = H(X 0,

More information

Chapter 1 Preliminaries

Chapter 1 Preliminaries Chapter 1 Preliminaries 1.1 Conventions and Notations Throughout the book we use the following notations for standard sets of numbers: N the set {1, 2,...} of natural numbers Z the set of integers Q the

More information

Measure and Integration: Solutions of CW2

Measure and Integration: Solutions of CW2 Measure and Integration: s of CW2 Fall 206 [G. Holzegel] December 9, 206 Problem of Sheet 5 a) Left (f n ) and (g n ) be sequences of integrable functions with f n (x) f (x) and g n (x) g (x) for almost

More information

A New Tight Upper Bound on the Entropy of Sums

A New Tight Upper Bound on the Entropy of Sums Article A New Tight Upper Bound on the Entropy of Sums Jihad Fahs * and Ibrahim Abou-Faycal eceived: 18 August 015; Accepted: 14 December 015; Published: 19 December 015 Academic Editor: aúl Alcaraz Martínez

More information

B. Appendix B. Topological vector spaces

B. Appendix B. Topological vector spaces B.1 B. Appendix B. Topological vector spaces B.1. Fréchet spaces. In this appendix we go through the definition of Fréchet spaces and their inductive limits, such as they are used for definitions of function

More information

Dimensionality in the Stability of the Brunn-Minkowski Inequality: A blessing or a curse?

Dimensionality in the Stability of the Brunn-Minkowski Inequality: A blessing or a curse? Dimensionality in the Stability of the Brunn-Minkowski Inequality: A blessing or a curse? Ronen Eldan, Tel Aviv University (Joint with Bo`az Klartag) Berkeley, September 23rd 2011 The Brunn-Minkowski Inequality

More information

Math 321 Final Examination April 1995 Notation used in this exam: N. (1) S N (f,x) = f(t)e int dt e inx.

Math 321 Final Examination April 1995 Notation used in this exam: N. (1) S N (f,x) = f(t)e int dt e inx. Math 321 Final Examination April 1995 Notation used in this exam: N 1 π (1) S N (f,x) = f(t)e int dt e inx. 2π n= N π (2) C(X, R) is the space of bounded real-valued functions on the metric space X, equipped

More information

LECTURE 3. Last time:

LECTURE 3. Last time: LECTURE 3 Last time: Mutual Information. Convexity and concavity Jensen s inequality Information Inequality Data processing theorem Fano s Inequality Lecture outline Stochastic processes, Entropy rate

More information

Nonparametric estimation: s concave and log-concave densities: alternatives to maximum likelihood

Nonparametric estimation: s concave and log-concave densities: alternatives to maximum likelihood Nonparametric estimation: s concave and log-concave densities: alternatives to maximum likelihood Jon A. Wellner University of Washington, Seattle Statistics Seminar, York October 15, 2015 Statistics Seminar,

More information

Maximum likelihood estimation

Maximum likelihood estimation Maximum likelihood estimation Guillaume Obozinski Ecole des Ponts - ParisTech Master MVA Maximum likelihood estimation 1/26 Outline 1 Statistical concepts 2 A short review of convex analysis and optimization

More information

λ(x + 1)f g (x) > θ 0

λ(x + 1)f g (x) > θ 0 Stat 8111 Final Exam December 16 Eleven students took the exam, the scores were 92, 78, 4 in the 5 s, 1 in the 4 s, 1 in the 3 s and 3 in the 2 s. 1. i) Let X 1, X 2,..., X n be iid each Bernoulli(θ) where

More information

Discrete Geometry. Problem 1. Austin Mohr. April 26, 2012

Discrete Geometry. Problem 1. Austin Mohr. April 26, 2012 Discrete Geometry Austin Mohr April 26, 2012 Problem 1 Theorem 1 (Linear Programming Duality). Suppose x, y, b, c R n and A R n n, Ax b, x 0, A T y c, and y 0. If x maximizes c T x and y minimizes b T

More information

Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 2012

Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 2012 Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 202 BOUNDS AND ASYMPTOTICS FOR FISHER INFORMATION IN THE CENTRAL LIMIT THEOREM

More information

On the analogue of the concavity of entropy power in the Brunn-Minkowski theory

On the analogue of the concavity of entropy power in the Brunn-Minkowski theory On the analogue of the concavity of entropy power in the Brunn-Minkowski theory Matthieu Fradelizi and Arnaud Marsiglietti Université Paris-Est, LAMA (UMR 8050), UPEMLV, UPEC, CNRS, F-77454, Marne-la-Vallée,

More information

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product Chapter 4 Hilbert Spaces 4.1 Inner Product Spaces Inner Product Space. A complex vector space E is called an inner product space (or a pre-hilbert space, or a unitary space) if there is a mapping (, )

More information

Machine Learning. Lecture 02.2: Basics of Information Theory. Nevin L. Zhang

Machine Learning. Lecture 02.2: Basics of Information Theory. Nevin L. Zhang Machine Learning Lecture 02.2: Basics of Information Theory Nevin L. Zhang lzhang@cse.ust.hk Department of Computer Science and Engineering The Hong Kong University of Science and Technology Nevin L. Zhang

More information

The proximal mapping

The proximal mapping The proximal mapping http://bicmr.pku.edu.cn/~wenzw/opt-2016-fall.html Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes Outline 2/37 1 closed function 2 Conjugate function

More information

ECE534, Spring 2018: Solutions for Problem Set #3

ECE534, Spring 2018: Solutions for Problem Set #3 ECE534, Spring 08: Solutions for Problem Set #3 Jointly Gaussian Random Variables and MMSE Estimation Suppose that X, Y are jointly Gaussian random variables with µ X = µ Y = 0 and σ X = σ Y = Let their

More information

1 Lesson 1: Brunn Minkowski Inequality

1 Lesson 1: Brunn Minkowski Inequality 1 Lesson 1: Brunn Minkowski Inequality A set A R n is called convex if (1 λ)x + λy A for any x, y A and any λ [0, 1]. The Minkowski sum of two sets A, B R n is defined by A + B := {a + b : a A, b B}. One

More information

Statistics 612: L p spaces, metrics on spaces of probabilites, and connections to estimation

Statistics 612: L p spaces, metrics on spaces of probabilites, and connections to estimation Statistics 62: L p spaces, metrics on spaces of probabilites, and connections to estimation Moulinath Banerjee December 6, 2006 L p spaces and Hilbert spaces We first formally define L p spaces. Consider

More information

Countable subgroups of Euclidean space

Countable subgroups of Euclidean space Countable subgroups of Euclidean space Arnold W. Miller April 2013 revised May 21, 2013 In his paper [1], Konstantinos Beros proved a number of results about compactly generated subgroups of Polish groups.

More information

1.3.1 Definition and Basic Properties of Convolution

1.3.1 Definition and Basic Properties of Convolution 1.3 Convolution 15 1.3 Convolution Since L 1 (R) is a Banach space, we know that it has many useful properties. In particular the operations of addition and scalar multiplication are continuous. However,

More information

Department of Mathematics, University of California, Berkeley. GRADUATE PRELIMINARY EXAMINATION, Part A Fall Semester 2014

Department of Mathematics, University of California, Berkeley. GRADUATE PRELIMINARY EXAMINATION, Part A Fall Semester 2014 Department of Mathematics, University of California, Berkeley YOUR 1 OR 2 DIGIT EXAM NUMBER GRADUATE PRELIMINARY EXAMINATION, Part A Fall Semester 2014 1. Please write your 1- or 2-digit exam number on

More information

STEP Support Programme. Pure STEP 1 Questions

STEP Support Programme. Pure STEP 1 Questions STEP Support Programme Pure STEP 1 Questions 2012 S1 Q4 1 Preparation Find the equation of the tangent to the curve y = x at the point where x = 4. Recall that x means the positive square root. Solve the

More information

Chernoff s distribution is log-concave. But why? (And why does it matter?)

Chernoff s distribution is log-concave. But why? (And why does it matter?) Chernoff s distribution is log-concave But why? (And why does it matter?) Jon A. Wellner University of Washington, Seattle University of Michigan April 1, 2011 Woodroofe Seminar Based on joint work with:

More information

Chapter 4: Asymptotic Properties of the MLE

Chapter 4: Asymptotic Properties of the MLE Chapter 4: Asymptotic Properties of the MLE Daniel O. Scharfstein 09/19/13 1 / 1 Maximum Likelihood Maximum likelihood is the most powerful tool for estimation. In this part of the course, we will consider

More information

ST5215: Advanced Statistical Theory

ST5215: Advanced Statistical Theory Department of Statistics & Applied Probability Monday, September 26, 2011 Lecture 10: Exponential families and Sufficient statistics Exponential Families Exponential families are important parametric families

More information

10-704: Information Processing and Learning Fall Lecture 10: Oct 3

10-704: Information Processing and Learning Fall Lecture 10: Oct 3 0-704: Information Processing and Learning Fall 206 Lecturer: Aarti Singh Lecture 0: Oct 3 Note: These notes are based on scribed notes from Spring5 offering of this course. LaTeX template courtesy of

More information

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University Chapter 4 Data Transmission and Channel Capacity Po-Ning Chen, Professor Department of Communications Engineering National Chiao Tung University Hsin Chu, Taiwan 30050, R.O.C. Principle of Data Transmission

More information

A relative entropy characterization of the growth rate of reward in risk-sensitive control

A relative entropy characterization of the growth rate of reward in risk-sensitive control 1 / 47 A relative entropy characterization of the growth rate of reward in risk-sensitive control Venkat Anantharam EECS Department, University of California, Berkeley (joint work with Vivek Borkar, IIT

More information

A NICE PROOF OF FARKAS LEMMA

A NICE PROOF OF FARKAS LEMMA A NICE PROOF OF FARKAS LEMMA DANIEL VICTOR TAUSK Abstract. The goal of this short note is to present a nice proof of Farkas Lemma which states that if C is the convex cone spanned by a finite set and if

More information

Principle of Mathematical Induction

Principle of Mathematical Induction Advanced Calculus I. Math 451, Fall 2016, Prof. Vershynin Principle of Mathematical Induction 1. Prove that 1 + 2 + + n = 1 n(n + 1) for all n N. 2 2. Prove that 1 2 + 2 2 + + n 2 = 1 n(n + 1)(2n + 1)

More information

The Canonical Gaussian Measure on R

The Canonical Gaussian Measure on R The Canonical Gaussian Measure on R 1. Introduction The main goal of this course is to study Gaussian measures. The simplest example of a Gaussian measure is the canonical Gaussian measure P on R where

More information

Convex Sets with Applications to Economics

Convex Sets with Applications to Economics Convex Sets with Applications to Economics Debasis Mishra March 10, 2010 1 Convex Sets A set C R n is called convex if for all x, y C, we have λx+(1 λ)y C for all λ [0, 1]. The definition says that for

More information

Mathematical Preliminaries for Microeconomics: Exercises

Mathematical Preliminaries for Microeconomics: Exercises Mathematical Preliminaries for Microeconomics: Exercises Igor Letina 1 Universität Zürich Fall 2013 1 Based on exercises by Dennis Gärtner, Andreas Hefti and Nick Netzer. How to prove A B Direct proof

More information

Geometry of log-concave Ensembles of random matrices

Geometry of log-concave Ensembles of random matrices Geometry of log-concave Ensembles of random matrices Nicole Tomczak-Jaegermann Joint work with Radosław Adamczak, Rafał Latała, Alexander Litvak, Alain Pajor Cortona, June 2011 Nicole Tomczak-Jaegermann

More information

arxiv:math/ v1 [math.pr] 27 Oct 1995

arxiv:math/ v1 [math.pr] 27 Oct 1995 Volumes of Restricted Minkowski Sums and the Free Analogue of the Entropy Power Inequality arxiv:math/9510203v1 [math.pr] 27 Oct 1995 Stanislaw J. Szarek Department of Mathematics Case Western Reserve

More information

From the Brunn-Minkowski inequality to a class of Poincaré type inequalities

From the Brunn-Minkowski inequality to a class of Poincaré type inequalities arxiv:math/0703584v1 [math.fa] 20 Mar 2007 From the Brunn-Minkowski inequality to a class of Poincaré type inequalities Andrea Colesanti Abstract We present an argument which leads from the Brunn-Minkowski

More information

Lebesgue Integration on R n

Lebesgue Integration on R n Lebesgue Integration on R n The treatment here is based loosely on that of Jones, Lebesgue Integration on Euclidean Space We give an overview from the perspective of a user of the theory Riemann integration

More information

EE/Stat 376B Handout #5 Network Information Theory October, 14, Homework Set #2 Solutions

EE/Stat 376B Handout #5 Network Information Theory October, 14, Homework Set #2 Solutions EE/Stat 376B Handout #5 Network Information Theory October, 14, 014 1. Problem.4 parts (b) and (c). Homework Set # Solutions (b) Consider h(x + Y ) h(x + Y Y ) = h(x Y ) = h(x). (c) Let ay = Y 1 + Y, where

More information

The Entropy Per Coordinate of a Random Vector is Highly Constrained Under Convexity Conditions Sergey Bobkov and Mokshay Madiman, Member, IEEE

The Entropy Per Coordinate of a Random Vector is Highly Constrained Under Convexity Conditions Sergey Bobkov and Mokshay Madiman, Member, IEEE 4940 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 8, AUGUST 2011 The Entropy Per Coordinate of a Random Vector is Highly Constrained Under Convexity Conditions Sergey Bobkov and Mokshay Madiman,

More information

1 Lyapunov theory of stability

1 Lyapunov theory of stability M.Kawski, APM 581 Diff Equns Intro to Lyapunov theory. November 15, 29 1 1 Lyapunov theory of stability Introduction. Lyapunov s second (or direct) method provides tools for studying (asymptotic) stability

More information

TOOLS FROM HARMONIC ANALYSIS

TOOLS FROM HARMONIC ANALYSIS TOOLS FROM HARMONIC ANALYSIS BRADLY STADIE Abstract. The Fourier transform can be thought of as a map that decomposes a function into oscillatory functions. In this paper, we will apply this decomposition

More information

10-704: Information Processing and Learning Fall Lecture 9: Sept 28

10-704: Information Processing and Learning Fall Lecture 9: Sept 28 10-704: Information Processing and Learning Fall 2016 Lecturer: Siheng Chen Lecture 9: Sept 28 Note: These notes are based on scribed notes from Spring15 offering of this course. LaTeX template courtesy

More information

1 Directional Derivatives and Differentiability

1 Directional Derivatives and Differentiability Wednesday, January 18, 2012 1 Directional Derivatives and Differentiability Let E R N, let f : E R and let x 0 E. Given a direction v R N, let L be the line through x 0 in the direction v, that is, L :=

More information

Course 212: Academic Year Section 1: Metric Spaces

Course 212: Academic Year Section 1: Metric Spaces Course 212: Academic Year 1991-2 Section 1: Metric Spaces D. R. Wilkins Contents 1 Metric Spaces 3 1.1 Distance Functions and Metric Spaces............. 3 1.2 Convergence and Continuity in Metric Spaces.........

More information

1 Continuity Classes C m (Ω)

1 Continuity Classes C m (Ω) 0.1 Norms 0.1 Norms A norm on a linear space X is a function : X R with the properties: Positive Definite: x 0 x X (nonnegative) x = 0 x = 0 (strictly positive) λx = λ x x X, λ C(homogeneous) x + y x +

More information