On the Complete Monotonicity of Heat Equation

Size: px
Start display at page:

Download "On the Complete Monotonicity of Heat Equation"

Transcription

1 [Pop-up Salon, Maths, SJTU, 2019/01/10] On the Complete Monotonicity of Heat Equation Fan Cheng John Hopcroft Center Computer Science and Engineering Shanghai Jiao Tong University

2 Overview 2 f x, t = 2 x2 t (Heat equation) f(x, t) h X + tz (Gaussian Channel) Y = X + tz Z N(0,1) h X = xlogx dx CMI = 2 (H. P. McKean, 1966) CMI 4. Conjecture: CMI = + (Cheng, 2015)

3 Outline Super-H Theorem Boltzmann equation and heat equation Shannon Entropy Power Inequality Complete Monotonicity Conjecture

4 Fire and Civilization Drill Myth: west and east Steam engine James Watts The Wealth of Nations Independence of US 1776

5 Study of Heat t f x, t = 1 f(x, t) 2 x2 2 Heat transfer The history begins with the work of Joseph Fourier around 1807 In a remarkable memoir, Fourier invented both Heat equation and the method of Fourier analysis for its solution

6 Information Age A mathematical theory of communication, Bell System Technical Journal. 27 (3): Gaussian Channel: Z t N(0, t) X and Z are mutually independent. The p.d.f of X is g(x) Y is the convolution of X and Z t : Y = X+ The p.d.f. of Y 1 f(y; t) = g(x) e(y x)2 2t 2πt t f(y; t) = y2 f(y; t) Fundamentally, Gaussian channel and heat equation are identical in mathematics (Gaussian mixture model) Z t

7 Entropy Formula Second law of thermodynamics: one way only Entropy

8 Ludwig Boltzmann Boltzmann formula: Gibbs formula: S = k b i S = k B lnw p i lnp i Ludwig Eduard Boltzmann Vienna, Austrian Empire Boltzmann equation: H-theorem: df dt = ( f t ) force + ( f t ) diff + ( f t ) coll H(f(t))is non decreasing

9 Super H-theorem for Boltzmann Equation Notation A function is completely monotone (CM) iff all the signs of its derivatives are: +, -, +, -, (e.g., 1, t e t ) McKean s Conjecture on Boltzmann equation (1966): H(f(t)) is CM in t, when f t satisfies Boltzmann equation False, disproved by E. Lieb in 1970s the particular Bobylev-Krook-Wu explicit solutions, this theorem holds true for n 101 and breaks downs afterwards H. P. McKean, NYU. National Academy of Sciences

10 Super H-theorem for Heat Equation Heat equation: Is H(f(t)) CM in t, if f(t) satisfies heat equation Equivalently, is H(X + tz) CM in t? The signs of the first two order derivatives were obtained Failed to obtain the 3 rd and 4 th. (It is easy to compute the derivatives, it is hard to obtain their signs) This suggests that, etc., but I could not prove it C. Villani, 2010 Fields Medalist

11 Claude E. Shannon and EPI Central limit theorem Capacity region of Gaussian broadcast channel Capacity region of Gaussian Multiple-Input Multiple-Output broadcast channel Uncertainty principle All of them can be proved by Entropy power inequality (EPI) Entropy power inequality (Shannon 1948): For any two independent continuous random variables X and Y, e 2h(X+Y) e 2h(X) + e 2h(Y) Equalities holds iff X and Y are Gaussian Motivation: Gaussian noise is the worst noise. Impact: A new characterization of Gaussian distribution in information theory Comments: most profound! (Kolmogorov)

12 Entropy Power Inequality Shannon himself didn t give a proof but an explanation, which turned out to be wrong The first proof is given by A. J. Stam (1959), N. M. Blachman (1966) Research on EPI Generalization, new proof, new connection. E.g., Gaussian interference channel is open, some stronger EPI should exist. Stanford Information Theory School: Thomas Cover and his students: A. El Gamel, M. H. Costa, A. Dembo, A. Barron ( ) Princeton Information Theory School: Sergio Verdu, etc. (2000s) Battle field of Shannon theory

13 Ramification of EPI Gaussian perturbation: h(x + tz) Shannon EPI Fisher Information: I X + tz = t h(x + tz)/2 Fisher Information is decreasing in t Fisher information inequality (FII): I(X+Y) I(X) I(Y) e 2h(X+ tz) is concave in t Tight Young s inequality X + Y r c X p Y q Status Quo: FII can imply EPI and all its generalizations. However, life is always hard. FII is far from enough

14 On X + tz X is arbitrary and h(x) may not exist When t 0, X + tz X. When t, X + tz Gaussian When t > 0, X + tz and h(x + tz) are infinitely differential X + tz is called mixed gaussian distribution (Gaussian Mixed Model (GMM) in machine learning) X + tz is Gaussian channel/source in information theory Gaussian noise is the worst additive noise Gaussian distribution maximizes h(x) Entropy power inequality, central limit theorem, etc.

15 Where we take off Shannon Entropy power inequality Fisher information inequality h(x + tz) h f t is CM When f(t) satisfied Boltzmann equation, disproved When f(t) satisfied heat equation, unknown We even don t know what CM is! Motivation: to study some inequalities; e.g., the convexity of h(x + the concavity of Any progress? None I X+ tz t Information theorists got lost in the past 70 years Mathematicians ignored it e t Z), It is widely believed that there should be no new EPI except Shannon EPI and FII.

16 Discovery I X + tz = h X + tz 0 (de Bruijn, 1958) 2 t I (1) = I X + tz 0 (McKean1966, Costa 1985) t Observation: I(X + tz) is convex in t h X + tz = 1 ln 2πet, I X + tz = 1. I is CM: +, -, +, - 2 t If the observation is true, the first three derivatives are: +, -, + Q: Is the 4 th order derivative -? Because Z is Gaussian! The signs of derivatives of h(x + tz) are independent of X. Invariant! Exactly the same problem in McKean 1966 To convince people, we must prove its convexity

17 Challenge Let X g(x) h Y t = f(y, t) ln f(y, t) dy, no closed form except some special g(x). f(y, t) satisfies heat equation. I Y t = f 1 2 dy f I 1 Y t = f 2 f 2 1 f f 2 2 dy So what is I (2)? (Heat equation, integration by parts)

18 Challenge (cont d) It is trivial to calculate derivatives. It is hard to prove their signs

19 Breakthrough Integration by parts: udv = uv vdu First breakthrough since McKean 1966

20

21 GCMC Gaussian complete monotonicity conjecture: I(X + tz) is CM in t Conjecture: logi(x + tz) is convex in t Pointed out by C. Villani and G. Toscani the connection with McKean s paper A general form: number partition. Hard to determine the coefficients. Hard to find β k,j!

22 Complete monotone function How to construct g(x)? A new expression for entropy involved special functions in mathematical physics Herbert R. Stahl, 2013

23 Complete monotone function A function f(t) is CM, then logf(t) is convex in t I Y t is CM in t, then log I(Y t ) is convex in t A function f(t) is CM, a Schur-convex function can be obtained by f(t) Schur-convex Majority theory Remarks: The current tools in information theory don t work. More sophisticated tools should be built to attack this problem. A new mathematical theory of information theory

24 Potential application: Interference channel A challenge question: what is the application of GCMC? Mathematical speaking, a beautiful result on a fundamental problem will be very useful Potential Application Central limit theorem Capacity region of Gaussian broadcast channel Capacity region of Gaussian Multiple-Input Multiple-Output broadcast channel Uncertainty principle Where EPI works Gaussian interference channel: open since 1970s Where EPI fails CM is considered to be much more powerful than EPI

25 Remarks If GCMC is true A fundamental breakthrough in mathematical physics, information theory and any disciplines related to Gaussian distribution A new expression for Fisher information Derivatives are an invariant Though h(x + tz) looks very messy, certain regularity exists Application: Gaussian interference channel? If GCMC is false No Failure, as heat equation is a physical phenomenon A lucky number (e.g. 2019) where Gaussian distribution fails. Painful!

A concavity property for the reciprocal of Fisher information and its consequences on Costa s EPI

A concavity property for the reciprocal of Fisher information and its consequences on Costa s EPI A concavity property for the reciprocal of Fisher information and its consequences on Costa s EPI Giuseppe Toscani October 0, 204 Abstract. We prove that the reciprocal of Fisher information of a logconcave

More information

Computing and Communications 2. Information Theory -Entropy

Computing and Communications 2. Information Theory -Entropy 1896 1920 1987 2006 Computing and Communications 2. Information Theory -Entropy Ying Cui Department of Electronic Engineering Shanghai Jiao Tong University, China 2017, Autumn 1 Outline Entropy Joint entropy

More information

Principles of Communications

Principles of Communications Principles of Communications Weiyao Lin Shanghai Jiao Tong University Chapter 10: Information Theory Textbook: Chapter 12 Communication Systems Engineering: Ch 6.1, Ch 9.1~ 9. 92 2009/2010 Meixia Tao @

More information

Machine Learning Srihari. Information Theory. Sargur N. Srihari

Machine Learning Srihari. Information Theory. Sargur N. Srihari Information Theory Sargur N. Srihari 1 Topics 1. Entropy as an Information Measure 1. Discrete variable definition Relationship to Code Length 2. Continuous Variable Differential Entropy 2. Maximum Entropy

More information

Heat equation and the sharp Young s inequality

Heat equation and the sharp Young s inequality Noname manuscript No. will be inserted by the editor) Heat equation and the sharp Young s inequality Giuseppe Toscani the date of receipt and acceptance should be inserted later Abstract We show that the

More information

ECE 4400:693 - Information Theory

ECE 4400:693 - Information Theory ECE 4400:693 - Information Theory Dr. Nghi Tran Lecture 8: Differential Entropy Dr. Nghi Tran (ECE-University of Akron) ECE 4400:693 Lecture 1 / 43 Outline 1 Review: Entropy of discrete RVs 2 Differential

More information

Information Theoretic Limits of Randomness Generation

Information Theoretic Limits of Randomness Generation Information Theoretic Limits of Randomness Generation Abbas El Gamal Stanford University Shannon Centennial, University of Michigan, September 2016 Information theory The fundamental problem of communication

More information

3. If a choice is broken down into two successive choices, the original H should be the weighted sum of the individual values of H.

3. If a choice is broken down into two successive choices, the original H should be the weighted sum of the individual values of H. Appendix A Information Theory A.1 Entropy Shannon (Shanon, 1948) developed the concept of entropy to measure the uncertainty of a discrete random variable. Suppose X is a discrete random variable that

More information

Monotonicity of entropy and Fisher information: a quick proof via maximal correlation

Monotonicity of entropy and Fisher information: a quick proof via maximal correlation Communications in Information and Systems Volume 16, Number 2, 111 115, 2016 Monotonicity of entropy and Fisher information: a quick proof via maximal correlation Thomas A. Courtade A simple proof is given

More information

Lecture 8: Channel Capacity, Continuous Random Variables

Lecture 8: Channel Capacity, Continuous Random Variables EE376A/STATS376A Information Theory Lecture 8-02/0/208 Lecture 8: Channel Capacity, Continuous Random Variables Lecturer: Tsachy Weissman Scribe: Augustine Chemparathy, Adithya Ganesh, Philip Hwang Channel

More information

A New Tight Upper Bound on the Entropy of Sums

A New Tight Upper Bound on the Entropy of Sums Article A New Tight Upper Bound on the Entropy of Sums Jihad Fahs * and Ibrahim Abou-Faycal eceived: 18 August 015; Accepted: 14 December 015; Published: 19 December 015 Academic Editor: aúl Alcaraz Martínez

More information

ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016

ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016 ECE598: Information-theoretic methods in high-dimensional statistics Spring 06 Lecture : Mutual Information Method Lecturer: Yihong Wu Scribe: Jaeho Lee, Mar, 06 Ed. Mar 9 Quick review: Assouad s lemma

More information

MMSE Dimension. snr. 1 We use the following asymptotic notation: f(x) = O (g(x)) if and only

MMSE Dimension. snr. 1 We use the following asymptotic notation: f(x) = O (g(x)) if and only MMSE Dimension Yihong Wu Department of Electrical Engineering Princeton University Princeton, NJ 08544, USA Email: yihongwu@princeton.edu Sergio Verdú Department of Electrical Engineering Princeton University

More information

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy Coding and Information Theory Chris Williams, School of Informatics, University of Edinburgh Overview What is information theory? Entropy Coding Information Theory Shannon (1948): Information theory is

More information

Chapter 8: Differential entropy. University of Illinois at Chicago ECE 534, Natasha Devroye

Chapter 8: Differential entropy. University of Illinois at Chicago ECE 534, Natasha Devroye Chapter 8: Differential entropy Chapter 8 outline Motivation Definitions Relation to discrete entropy Joint and conditional differential entropy Relative entropy and mutual information Properties AEP for

More information

An Alternative Proof for the Capacity Region of the Degraded Gaussian MIMO Broadcast Channel

An Alternative Proof for the Capacity Region of the Degraded Gaussian MIMO Broadcast Channel IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 58, NO. 4, APRIL 2012 2427 An Alternative Proof for the Capacity Region of the Degraded Gaussian MIMO Broadcast Channel Ersen Ekrem, Student Member, IEEE,

More information

d(x n, x) d(x n, x nk ) + d(x nk, x) where we chose any fixed k > N

d(x n, x) d(x n, x nk ) + d(x nk, x) where we chose any fixed k > N Problem 1. Let f : A R R have the property that for every x A, there exists ɛ > 0 such that f(t) > ɛ if t (x ɛ, x + ɛ) A. If the set A is compact, prove there exists c > 0 such that f(x) > c for all x

More information

A Simple Proof of the Entropy-Power Inequality

A Simple Proof of the Entropy-Power Inequality A Simple Proof of the Entropy-Power Inequality via Properties of Mutual Information Olivier Rioul Dept. ComElec, GET/Telecom Paris (ENST) ParisTech Institute & CNRS LTCI Paris, France Email: olivier.rioul@enst.fr

More information

Variable selection and feature construction using methods related to information theory

Variable selection and feature construction using methods related to information theory Outline Variable selection and feature construction using methods related to information theory Kari 1 1 Intelligent Systems Lab, Motorola, Tempe, AZ IJCNN 2007 Outline Outline 1 Information Theory and

More information

G 8243 Entropy and Information in Probability

G 8243 Entropy and Information in Probability G 8243 Entropy and Information in Probability I. Kontoyiannis Columbia University Spring 2009 What is Information Theory? Born in 1948, it is the mathematical foundation of communication theory; it quantifies

More information

Limits on classical communication from quantum entropy power inequalities

Limits on classical communication from quantum entropy power inequalities Limits on classical communication from quantum entropy power inequalities Graeme Smith, IBM Research (joint work with Robert Koenig) QIP 2013 Beijing Channel Capacity X N Y p(y x) Capacity: bits per channel

More information

Ch. 8 Math Preliminaries for Lossy Coding. 8.4 Info Theory Revisited

Ch. 8 Math Preliminaries for Lossy Coding. 8.4 Info Theory Revisited Ch. 8 Math Preliminaries for Lossy Coding 8.4 Info Theory Revisited 1 Info Theory Goals for Lossy Coding Again just as for the lossless case Info Theory provides: Basis for Algorithms & Bounds on Performance

More information

Discrete Memoryless Channels with Memoryless Output Sequences

Discrete Memoryless Channels with Memoryless Output Sequences Discrete Memoryless Channels with Memoryless utput Sequences Marcelo S Pinho Department of Electronic Engineering Instituto Tecnologico de Aeronautica Sao Jose dos Campos, SP 12228-900, Brazil Email: mpinho@ieeeorg

More information

Chapter I: Fundamental Information Theory

Chapter I: Fundamental Information Theory ECE-S622/T62 Notes Chapter I: Fundamental Information Theory Ruifeng Zhang Dept. of Electrical & Computer Eng. Drexel University. Information Source Information is the outcome of some physical processes.

More information

Information maximization in a network of linear neurons

Information maximization in a network of linear neurons Information maximization in a network of linear neurons Holger Arnold May 30, 005 1 Introduction It is known since the work of Hubel and Wiesel [3], that many cells in the early visual areas of mammals

More information

6.1 Main properties of Shannon entropy. Let X be a random variable taking values x in some alphabet with probabilities.

6.1 Main properties of Shannon entropy. Let X be a random variable taking values x in some alphabet with probabilities. Chapter 6 Quantum entropy There is a notion of entropy which quantifies the amount of uncertainty contained in an ensemble of Qbits. This is the von Neumann entropy that we introduce in this chapter. In

More information

Fisher information and Stam inequality on a finite group

Fisher information and Stam inequality on a finite group Fisher information and Stam inequality on a finite group Paolo Gibilisco and Tommaso Isola February, 2008 Abstract We prove a discrete version of Stam inequality for random variables taking values on a

More information

Medical Imaging. Norbert Schuff, Ph.D. Center for Imaging of Neurodegenerative Diseases

Medical Imaging. Norbert Schuff, Ph.D. Center for Imaging of Neurodegenerative Diseases Uses of Information Theory in Medical Imaging Norbert Schuff, Ph.D. Center for Imaging of Neurodegenerative Diseases Norbert.schuff@ucsf.edu With contributions from Dr. Wang Zhang Medical Imaging Informatics,

More information

Yet Another Proof of the Entropy Power Inequality

Yet Another Proof of the Entropy Power Inequality IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 63, NO. 6, JUNE 2017 3595 Yet Another Proof of the Entropy Power Inequality Olivier Rioul, Member, IEEE Abstract Yet another simple proof of the entropy power

More information

Introduction to Information Theory

Introduction to Information Theory Introduction to Information Theory Gurinder Singh Mickey Atwal atwal@cshl.edu Center for Quantitative Biology Kullback-Leibler Divergence Summary Shannon s coding theorems Entropy Mutual Information Multi-information

More information

Lecture 2: August 31

Lecture 2: August 31 0-704: Information Processing and Learning Fall 206 Lecturer: Aarti Singh Lecture 2: August 3 Note: These notes are based on scribed notes from Spring5 offering of this course. LaTeX template courtesy

More information

Entropy Power Inequalities: Results and Speculation

Entropy Power Inequalities: Results and Speculation Entropy Power Inequalities: Results and Speculation Venkat Anantharam EECS Department University of California, Berkeley December 12, 2013 Workshop on Coding and Information Theory Institute of Mathematical

More information

A CLASSROOM NOTE: ENTROPY, INFORMATION, AND MARKOV PROPERTY. Zoran R. Pop-Stojanović. 1. Introduction

A CLASSROOM NOTE: ENTROPY, INFORMATION, AND MARKOV PROPERTY. Zoran R. Pop-Stojanović. 1. Introduction THE TEACHING OF MATHEMATICS 2006, Vol IX,, pp 2 A CLASSROOM NOTE: ENTROPY, INFORMATION, AND MARKOV PROPERTY Zoran R Pop-Stojanović Abstract How to introduce the concept of the Markov Property in an elementary

More information

Chapter 2: Entropy and Mutual Information. University of Illinois at Chicago ECE 534, Natasha Devroye

Chapter 2: Entropy and Mutual Information. University of Illinois at Chicago ECE 534, Natasha Devroye Chapter 2: Entropy and Mutual Information Chapter 2 outline Definitions Entropy Joint entropy, conditional entropy Relative entropy, mutual information Chain rules Jensen s inequality Log-sum inequality

More information

Ch. 8 Math Preliminaries for Lossy Coding. 8.5 Rate-Distortion Theory

Ch. 8 Math Preliminaries for Lossy Coding. 8.5 Rate-Distortion Theory Ch. 8 Math Preliminaries for Lossy Coding 8.5 Rate-Distortion Theory 1 Introduction Theory provide insight into the trade between Rate & Distortion This theory is needed to answer: What do typical R-D

More information

1.6. Information Theory

1.6. Information Theory 48. INTRODUCTION Section 5.6 Exercise.7 (b) First solve the inference problem of determining the conditional density p(t x), and then subsequently marginalize to find the conditional mean given by (.89).

More information

Convexity/Concavity of Renyi Entropy and α-mutual Information

Convexity/Concavity of Renyi Entropy and α-mutual Information Convexity/Concavity of Renyi Entropy and -Mutual Information Siu-Wai Ho Institute for Telecommunications Research University of South Australia Adelaide, SA 5095, Australia Email: siuwai.ho@unisa.edu.au

More information

3F1 Information Theory, Lecture 1

3F1 Information Theory, Lecture 1 3F1 Information Theory, Lecture 1 Jossy Sayir Department of Engineering Michaelmas 2013, 22 November 2013 Organisation History Entropy Mutual Information 2 / 18 Course Organisation 4 lectures Course material:

More information

Information Theory Primer:

Information Theory Primer: Information Theory Primer: Entropy, KL Divergence, Mutual Information, Jensen s inequality Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro,

More information

Information in Biology

Information in Biology Information in Biology CRI - Centre de Recherches Interdisciplinaires, Paris May 2012 Information processing is an essential part of Life. Thinking about it in quantitative terms may is useful. 1 Living

More information

Kinetic models of Maxwell type. A brief history Part I

Kinetic models of Maxwell type. A brief history Part I . A brief history Part I Department of Mathematics University of Pavia, Italy Porto Ercole, June 8-10 2008 Summer School METHODS AND MODELS OF KINETIC THEORY Outline 1 Introduction Wild result The central

More information

Information. = more information was provided by the outcome in #2

Information. = more information was provided by the outcome in #2 Outline First part based very loosely on [Abramson 63]. Information theory usually formulated in terms of information channels and coding will not discuss those here.. Information 2. Entropy 3. Mutual

More information

A Gentle Tutorial on Information Theory and Learning. Roni Rosenfeld. Carnegie Mellon University

A Gentle Tutorial on Information Theory and Learning. Roni Rosenfeld. Carnegie Mellon University A Gentle Tutorial on Information Theory and Learning Roni Rosenfeld Mellon University Mellon Outline First part based very loosely on [Abramson 63]. Information theory usually formulated in terms of information

More information

Entropy power inequality for a family of discrete random variables

Entropy power inequality for a family of discrete random variables 20 IEEE International Symposium on Information Theory Proceedings Entropy power inequality for a family of discrete random variables Naresh Sharma, Smarajit Das and Siddharth Muthurishnan School of Technology

More information

EE/Stat 376B Handout #5 Network Information Theory October, 14, Homework Set #2 Solutions

EE/Stat 376B Handout #5 Network Information Theory October, 14, Homework Set #2 Solutions EE/Stat 376B Handout #5 Network Information Theory October, 14, 014 1. Problem.4 parts (b) and (c). Homework Set # Solutions (b) Consider h(x + Y ) h(x + Y Y ) = h(x Y ) = h(x). (c) Let ay = Y 1 + Y, where

More information

Functional Properties of MMSE

Functional Properties of MMSE Functional Properties of MMSE Yihong Wu epartment of Electrical Engineering Princeton University Princeton, NJ 08544, USA Email: yihongwu@princeton.edu Sergio Verdú epartment of Electrical Engineering

More information

Quantitative Biology Lecture 3

Quantitative Biology Lecture 3 23 nd Sep 2015 Quantitative Biology Lecture 3 Gurinder Singh Mickey Atwal Center for Quantitative Biology Summary Covariance, Correlation Confounding variables (Batch Effects) Information Theory Covariance

More information

Estimation of information-theoretic quantities

Estimation of information-theoretic quantities Estimation of information-theoretic quantities Liam Paninski Gatsby Computational Neuroscience Unit University College London http://www.gatsby.ucl.ac.uk/ liam liam@gatsby.ucl.ac.uk November 16, 2004 Some

More information

Draft. On Limiting Expressions for the Capacity Region of Gaussian Interference Channels. Mojtaba Vaezi and H. Vincent Poor

Draft. On Limiting Expressions for the Capacity Region of Gaussian Interference Channels. Mojtaba Vaezi and H. Vincent Poor Preliminaries Counterexample Better Use On Limiting Expressions for the Capacity Region of Gaussian Interference Channels Mojtaba Vaezi and H. Vincent Poor Department of Electrical Engineering Princeton

More information

Heat equation and the sharp Young s inequality

Heat equation and the sharp Young s inequality Heat equation and the sharp Young s inequality arxiv:1204.2086v2 [math-ph] 11 Apr 2012 Giuseppe Toscani April 12, 2012 Abstract. We show that the sharp Young s inequality for convolutions first obtained

More information

Information in Biology

Information in Biology Lecture 3: Information in Biology Tsvi Tlusty, tsvi@unist.ac.kr Living information is carried by molecular channels Living systems I. Self-replicating information processors Environment II. III. Evolve

More information

Lecture 14 February 28

Lecture 14 February 28 EE/Stats 376A: Information Theory Winter 07 Lecture 4 February 8 Lecturer: David Tse Scribe: Sagnik M, Vivek B 4 Outline Gaussian channel and capacity Information measures for continuous random variables

More information

Outline of the Lecture. Background and Motivation. Basics of Information Theory: 1. Introduction. Markku Juntti. Course Overview

Outline of the Lecture. Background and Motivation. Basics of Information Theory: 1. Introduction. Markku Juntti. Course Overview : Markku Juntti Overview The basic ideas and concepts of information theory are introduced. Some historical notes are made and the overview of the course is given. Source The material is mainly based on

More information

Gaussian channel. Information theory 2013, lecture 6. Jens Sjölund. 8 May Jens Sjölund (IMT, LiU) Gaussian channel 1 / 26

Gaussian channel. Information theory 2013, lecture 6. Jens Sjölund. 8 May Jens Sjölund (IMT, LiU) Gaussian channel 1 / 26 Gaussian channel Information theory 2013, lecture 6 Jens Sjölund 8 May 2013 Jens Sjölund (IMT, LiU) Gaussian channel 1 / 26 Outline 1 Definitions 2 The coding theorem for Gaussian channel 3 Bandlimited

More information

How to Quantitate a Markov Chain? Stochostic project 1

How to Quantitate a Markov Chain? Stochostic project 1 How to Quantitate a Markov Chain? Stochostic project 1 Chi-Ning,Chou Wei-chang,Lee PROFESSOR RAOUL NORMAND April 18, 2015 Abstract In this project, we want to quantitatively evaluate a Markov chain. In

More information

Entropy and Limit theorems in Probability Theory

Entropy and Limit theorems in Probability Theory Entropy and Limit theorems in Probability Theory Introduction Shigeki Aida Important Notice : Solve at least one problem from the following Problems -8 and submit the report to me until June 9. What is

More information

Thermodynamics/Optics

Thermodynamics/Optics Thermodynamics/Optics Survey! Survey! I hate calculus! Survey! I hate calculus! Overview of Last Lecture Foundation of physics, classical mechanics Copernican Revolution Galileo/Newton experimentation

More information

Strong log-concavity is preserved by convolution

Strong log-concavity is preserved by convolution Strong log-concavity is preserved by convolution Jon A. Wellner Abstract. We review and formulate results concerning strong-log-concavity in both discrete and continuous settings. Although four different

More information

Some New Results on Information Properties of Mixture Distributions

Some New Results on Information Properties of Mixture Distributions Filomat 31:13 (2017), 4225 4230 https://doi.org/10.2298/fil1713225t Published by Faculty of Sciences and Mathematics, University of Niš, Serbia Available at: http://www.pmf.ni.ac.rs/filomat Some New Results

More information

Concavity of entropy under thinning

Concavity of entropy under thinning Concavity of entropy under thinning Yaming Yu Department of Statistics University of California Irvine, CA 92697, USA Email: yamingy@uci.edu Oliver Johnson Department of Mathematics University of Bristol

More information

Submodularity in Machine Learning

Submodularity in Machine Learning Saifuddin Syed MLRG Summer 2016 1 / 39 What are submodular functions Outline 1 What are submodular functions Motivation Submodularity and Concavity Examples 2 Properties of submodular functions Submodularity

More information

ELEC546 Review of Information Theory

ELEC546 Review of Information Theory ELEC546 Review of Information Theory Vincent Lau 1/1/004 1 Review of Information Theory Entropy: Measure of uncertainty of a random variable X. The entropy of X, H(X), is given by: If X is a discrete random

More information

Lecture 2. Capacity of the Gaussian channel

Lecture 2. Capacity of the Gaussian channel Spring, 207 5237S, Wireless Communications II 2. Lecture 2 Capacity of the Gaussian channel Review on basic concepts in inf. theory ( Cover&Thomas: Elements of Inf. Theory, Tse&Viswanath: Appendix B) AWGN

More information

Steiner s formula and large deviations theory

Steiner s formula and large deviations theory Steiner s formula and large deviations theory Venkat Anantharam EECS Department University of California, Berkeley May 19, 2015 Simons Conference on Networks and Stochastic Geometry Blanton Museum of Art

More information

Lecture 11: Continuous-valued signals and differential entropy

Lecture 11: Continuous-valued signals and differential entropy Lecture 11: Continuous-valued signals and differential entropy Biology 429 Carl Bergstrom September 20, 2008 Sources: Parts of today s lecture follow Chapter 8 from Cover and Thomas (2007). Some components

More information

The Effect upon Channel Capacity in Wireless Communications of Perfect and Imperfect Knowledge of the Channel

The Effect upon Channel Capacity in Wireless Communications of Perfect and Imperfect Knowledge of the Channel The Effect upon Channel Capacity in Wireless Communications of Perfect and Imperfect Knowledge of the Channel Muriel Medard, Trans. on IT 000 Reading Group Discussion March 0, 008 Intro./ Overview Time

More information

VARIANTS OF ENTROPY POWER INEQUALITY

VARIANTS OF ENTROPY POWER INEQUALITY VARIANTS OF ENTROPY POWER INEQUALITY Sergey G. Bobkov and Arnaud Marsiglietti Abstract An extension of the entroy ower inequality to the form N α r (X + Y ) N α r (X) + N α r (Y ) with arbitrary indeendent

More information

Lecture 22: Final Review

Lecture 22: Final Review Lecture 22: Final Review Nuts and bolts Fundamental questions and limits Tools Practical algorithms Future topics Dr Yao Xie, ECE587, Information Theory, Duke University Basics Dr Yao Xie, ECE587, Information

More information

Be sure this exam has 8 pages including the cover The University of British Columbia MATH 103 Midterm Exam II Mar 14, 2012

Be sure this exam has 8 pages including the cover The University of British Columbia MATH 103 Midterm Exam II Mar 14, 2012 Be sure this exam has 8 pages including the cover The University of British Columbia MATH Midterm Exam II Mar 4, 22 Family Name Student Number Given Name Signature Section Number This exam consists of

More information

Chaos, Complexity, and Inference (36-462)

Chaos, Complexity, and Inference (36-462) Chaos, Complexity, and Inference (36-462) Lecture 7: Information Theory Cosma Shalizi 3 February 2009 Entropy and Information Measuring randomness and dependence in bits The connection to statistics Long-run

More information

CS6304 / Analog and Digital Communication UNIT IV - SOURCE AND ERROR CONTROL CODING PART A 1. What is the use of error control coding? The main use of error control coding is to reduce the overall probability

More information

Plan Martingales cont d. 0. Questions for Exam 2. More Examples 3. Overview of Results. Reading: study Next Time: first exam

Plan Martingales cont d. 0. Questions for Exam 2. More Examples 3. Overview of Results. Reading: study Next Time: first exam Plan Martingales cont d 0. Questions for Exam 2. More Examples 3. Overview of Results Reading: study Next Time: first exam Midterm Exam: Tuesday 28 March in class Sample exam problems ( Homework 5 and

More information

Logarithmic Sobolev Inequalities

Logarithmic Sobolev Inequalities Logarithmic Sobolev Inequalities M. Ledoux Institut de Mathématiques de Toulouse, France logarithmic Sobolev inequalities what they are, some history analytic, geometric, optimal transportation proofs

More information

Entropy and Ergodic Theory Lecture 4: Conditional entropy and mutual information

Entropy and Ergodic Theory Lecture 4: Conditional entropy and mutual information Entropy and Ergodic Theory Lecture 4: Conditional entropy and mutual information 1 Conditional entropy Let (Ω, F, P) be a probability space, let X be a RV taking values in some finite set A. In this lecture

More information

Functional Analysis HW 2

Functional Analysis HW 2 Brandon Behring Functional Analysis HW 2 Exercise 2.6 The space C[a, b] equipped with the L norm defined by f = b a f(x) dx is incomplete. If f n f with respect to the sup-norm then f n f with respect

More information

Learning Objectives for Math 165

Learning Objectives for Math 165 Learning Objectives for Math 165 Chapter 2 Limits Section 2.1: Average Rate of Change. State the definition of average rate of change Describe what the rate of change does and does not tell us in a given

More information

EE 4TM4: Digital Communications II Scalar Gaussian Channel

EE 4TM4: Digital Communications II Scalar Gaussian Channel EE 4TM4: Digital Communications II Scalar Gaussian Channel I. DIFFERENTIAL ENTROPY Let X be a continuous random variable with probability density function (pdf) f(x) (in short X f(x)). The differential

More information

The Central Limit Theorem: More of the Story

The Central Limit Theorem: More of the Story The Central Limit Theorem: More of the Story Steven Janke November 2015 Steven Janke (Seminar) The Central Limit Theorem:More of the Story November 2015 1 / 33 Central Limit Theorem Theorem (Central Limit

More information

QB LECTURE #4: Motif Finding

QB LECTURE #4: Motif Finding QB LECTURE #4: Motif Finding Adam Siepel Nov. 20, 2015 2 Plan for Today Probability models for binding sites Scoring and detecting binding sites De novo motif finding 3 Transcription Initiation Chromatin

More information

FOR a continuous random variable X with density g(x), the differential entropy is defined as

FOR a continuous random variable X with density g(x), the differential entropy is defined as Higher Order Derivatives in Costa s Entrop Power Inequalit Fan Cheng, Member, IEEE and Yanlin Geng, Member, IEEE arxiv:09.v [cs.it] Aug 0 Abstract Let X be an arbitrar continuous random variable and Z

More information

Lecture 6: Gaussian Channels. Copyright G. Caire (Sample Lectures) 157

Lecture 6: Gaussian Channels. Copyright G. Caire (Sample Lectures) 157 Lecture 6: Gaussian Channels Copyright G. Caire (Sample Lectures) 157 Differential entropy (1) Definition 18. The (joint) differential entropy of a continuous random vector X n p X n(x) over R is: Z h(x

More information

High-dimensional distributions with convexity properties

High-dimensional distributions with convexity properties High-dimensional distributions with convexity properties Bo az Klartag Tel-Aviv University A conference in honor of Charles Fefferman, Princeton, May 2009 High-Dimensional Distributions We are concerned

More information

Example: Letter Frequencies

Example: Letter Frequencies Example: Letter Frequencies i a i p i 1 a 0.0575 2 b 0.0128 3 c 0.0263 4 d 0.0285 5 e 0.0913 6 f 0.0173 7 g 0.0133 8 h 0.0313 9 i 0.0599 10 j 0.0006 11 k 0.0084 12 l 0.0335 13 m 0.0235 14 n 0.0596 15 o

More information

Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 2012

Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 2012 Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 202 BOUNDS AND ASYMPTOTICS FOR FISHER INFORMATION IN THE CENTRAL LIMIT THEOREM

More information

3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions

3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions 3. Convex functions Convex Optimization Boyd & Vandenberghe basic properties and examples operations that preserve convexity the conjugate function quasiconvex functions log-concave and log-convex functions

More information

Example: Letter Frequencies

Example: Letter Frequencies Example: Letter Frequencies i a i p i 1 a 0.0575 2 b 0.0128 3 c 0.0263 4 d 0.0285 5 e 0.0913 6 f 0.0173 7 g 0.0133 8 h 0.0313 9 i 0.0599 10 j 0.0006 11 k 0.0084 12 l 0.0335 13 m 0.0235 14 n 0.0596 15 o

More information

Channel capacity. Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5.

Channel capacity. Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5. Channel capacity Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5. Exercices Exercise session 11 : Channel capacity 1 1. Source entropy Given X a memoryless

More information

Nonstochastic Information Theory for Feedback Control

Nonstochastic Information Theory for Feedback Control Nonstochastic Information Theory for Feedback Control Girish Nair Department of Electrical and Electronic Engineering University of Melbourne Australia Dutch Institute of Systems and Control Summer School

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Introduction to Probabilistic Methods Varun Chandola Computer Science & Engineering State University of New York at Buffalo Buffalo, NY, USA chandola@buffalo.edu Chandola@UB

More information

Capacity Bounds for Diamond Networks

Capacity Bounds for Diamond Networks Technische Universität München Capacity Bounds for Diamond Networks Gerhard Kramer (TUM) joint work with Shirin Saeedi Bidokhti (TUM & Stanford) DIMACS Workshop on Network Coding Rutgers University, NJ

More information

Part I. Entropy. Information Theory and Networks. Section 1. Entropy: definitions. Lecture 5: Entropy

Part I. Entropy. Information Theory and Networks. Section 1. Entropy: definitions. Lecture 5: Entropy and Networks Lecture 5: Matthew Roughan http://www.maths.adelaide.edu.au/matthew.roughan/ Lecture_notes/InformationTheory/ Part I School of Mathematical Sciences, University

More information

An Extended Fano s Inequality for the Finite Blocklength Coding

An Extended Fano s Inequality for the Finite Blocklength Coding An Extended Fano s Inequality for the Finite Bloclength Coding Yunquan Dong, Pingyi Fan {dongyq8@mails,fpy@mail}.tsinghua.edu.cn Department of Electronic Engineering, Tsinghua University, Beijing, P.R.

More information

Introduction to Information Entropy Adapted from Papoulis (1991)

Introduction to Information Entropy Adapted from Papoulis (1991) Introduction to Information Entropy Adapted from Papoulis (1991) Federico Lombardo Papoulis, A., Probability, Random Variables and Stochastic Processes, 3rd edition, McGraw ill, 1991. 1 1. INTRODUCTION

More information

On some special cases of the Entropy Photon-Number Inequality

On some special cases of the Entropy Photon-Number Inequality On some special cases of the Entropy Photon-Number Inequality Smarajit Das, Naresh Sharma and Siddharth Muthukrishnan Tata Institute of Fundamental Research December 5, 2011 One of the aims of information

More information

Chernoff s distribution is log-concave. But why? (And why does it matter?)

Chernoff s distribution is log-concave. But why? (And why does it matter?) Chernoff s distribution is log-concave But why? (And why does it matter?) Jon A. Wellner University of Washington, Seattle University of Michigan April 1, 2011 Woodroofe Seminar Based on joint work with:

More information

The Capacity Region of the Gaussian MIMO Broadcast Channel

The Capacity Region of the Gaussian MIMO Broadcast Channel 0-0 The Capacity Region of the Gaussian MIMO Broadcast Channel Hanan Weingarten, Yossef Steinberg and Shlomo Shamai (Shitz) Outline Problem statement Background and preliminaries Capacity region of the

More information

x log x, which is strictly convex, and use Jensen s Inequality:

x log x, which is strictly convex, and use Jensen s Inequality: 2. Information measures: mutual information 2.1 Divergence: main inequality Theorem 2.1 (Information Inequality). D(P Q) 0 ; D(P Q) = 0 iff P = Q Proof. Let ϕ(x) x log x, which is strictly convex, and

More information

Functions & Graphs. Section 1.2

Functions & Graphs. Section 1.2 Functions & Graphs Section 1.2 What you will remember Functions Domains and Ranges Viewing and Interpreting Graphs Even Functions and Odd Functions Symmetry Functions Defined in Pieces Absolute Value Functions

More information

Communication Theory II

Communication Theory II Communication Theory II Lecture 15: Information Theory (cont d) Ahmed Elnakib, PhD Assistant Professor, Mansoura University, Egypt March 29 th, 2015 1 Example: Channel Capacity of BSC o Let then: o For

More information

Hands-On Learning Theory Fall 2016, Lecture 3

Hands-On Learning Theory Fall 2016, Lecture 3 Hands-On Learning Theory Fall 016, Lecture 3 Jean Honorio jhonorio@purdue.edu 1 Information Theory First, we provide some information theory background. Definition 3.1 (Entropy). The entropy of a discrete

More information