Information in Biology

Similar documents
Information in Biology

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Dept. of Linguistics, Indiana University Fall 2015

Lecture 11: Information theory THURSDAY, FEBRUARY 21, 2019

Introduction to Information Theory

Quantitative Biology Lecture 3

Lecture 2: August 31

Lecture 1: Introduction, Entropy and ML estimation

6.02 Fall 2011 Lecture #9

Entropies & Information Theory

3F1 Information Theory, Lecture 1

Part I. Entropy. Information Theory and Networks. Section 1. Entropy: definitions. Lecture 5: Entropy

Classification & Information Theory Lecture #8

Murray Gell-Mann, The Quark and the Jaguar, 1995

Thermodynamics Second Law Entropy

Outline of the Lecture. Background and Motivation. Basics of Information Theory: 1. Introduction. Markku Juntti. Course Overview

Introduction to Information Theory. B. Škorić, Physical Aspects of Digital Security, Chapter 2

Information Theory Primer:

Machine Learning. Lecture 02.2: Basics of Information Theory. Nevin L. Zhang

Lecture 5 - Information theory

AQI: Advanced Quantum Information Lecture 6 (Module 2): Distinguishing Quantum States January 28, 2013

Machine Learning Srihari. Information Theory. Sargur N. Srihari

Information. = more information was provided by the outcome in #2

A Gentle Tutorial on Information Theory and Learning. Roni Rosenfeld. Carnegie Mellon University

Chapter 2: Entropy and Mutual Information. University of Illinois at Chicago ECE 534, Natasha Devroye

Neural coding Ecological approach to sensory coding: efficient adaptation to the natural environment

COMPSCI 650 Applied Information Theory Jan 21, Lecture 2

CS 630 Basic Probability and Information Theory. Tim Campbell

Principles of Communications

3. If a choice is broken down into two successive choices, the original H should be the weighted sum of the individual values of H.

Lecture 27: Entropy and Information Prof. WAN, Xin

How to Quantitate a Markov Chain? Stochostic project 1

4F5: Advanced Communications and Coding Handout 2: The Typical Set, Compression, Mutual Information

What is Entropy? Jeff Gill, 1 Entropy in Information Theory

Biology as Information Dynamics

Information Theory in Intelligent Decision Making

CHAPTER 4 ENTROPY AND INFORMATION

5 Mutual Information and Channel Capacity

Mutual Information & Genotype-Phenotype Association. Norman MacDonald January 31, 2011 CSCI 4181/6802

Chapter I: Fundamental Information Theory

Thermodynamics: More Entropy

(Classical) Information Theory III: Noisy channel coding

Quantitative Biology II Lecture 4: Variational Methods

Channel capacity. Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5.

UNIT I INFORMATION THEORY. I k log 2

Biology as Information Dynamics

Lecture 1: Shannon s Theorem

INTRODUCTION TO INFORMATION THEORY

Noisy-Channel Coding

Information Theory (Information Theory by J. V. Stone, 2015)

Information Theory, Statistics, and Decision Trees

6.1 Main properties of Shannon entropy. Let X be a random variable taking values x in some alphabet with probabilities.

PROBABILITY AND INFORMATION THEORY. Dr. Gjergji Kasneci Introduction to Information Retrieval WS

Thermodynamics: More Entropy

Chapter 2 Review of Classical Information Theory

Lecture 7: DecisionTrees

Medical Imaging. Norbert Schuff, Ph.D. Center for Imaging of Neurodegenerative Diseases

Historical Perspectives John von Neumann ( )

6.02 Fall 2012 Lecture #1

Shannon's Theory of Communication

Topics. Probability Theory. Perfect Secrecy. Information Theory

Massachusetts Institute of Technology

MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK UNIT V PART-A. 1. What is binary symmetric channel (AUC DEC 2006)

Transmitting and Hiding Quantum Information

Computing and Communications 2. Information Theory -Entropy

MAHALAKSHMI ENGINEERING COLLEGE QUESTION BANK. SUBJECT CODE / Name: EC2252 COMMUNICATION THEORY UNIT-V INFORMATION THEORY PART-A

Beyond the Second Law of Thermodynamics

An Introduction to Information Theory: Notes

Natural Image Statistics and Neural Representations

Information Theory - Entropy. Figure 3

Welcome to Comp 411! 2) Course Objectives. 1) Course Mechanics. 3) Information. I thought this course was called Computer Organization

1 Introduction to information theory

The physics of information: from Maxwell s demon to Landauer. Eric Lutz University of Erlangen-Nürnberg

Thermodynamics of feedback controlled systems. Francisco J. Cao

Communication Theory and Engineering

An Elementary Notion and Measurement of Entropy

Information & Correlation

Chapter 9 Fundamental Limits in Information Theory

Lecture 8: Shannon s Noise Models

Introduction to Information Entropy Adapted from Papoulis (1991)

Some Concepts in Probability and Information Theory

A Mathematical Theory of Communication

Quantum Thermodynamics

The Measure of Information Uniqueness of the Logarithmic Uncertainty Measure

Intro to Information Theory

3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions

CSE468 Information Conflict

Lecture 22: Final Review

9. Distance measures. 9.1 Classical information measures. Head Tail. How similar/close are two probability distributions? Trace distance.

Lecture 18: Quantum Information Theory and Holevo s Bound

Massachusetts Institute of Technology

1 Ex. 1 Verify that the function H(p 1,..., p n ) = k p k log 2 p k satisfies all 8 axioms on H.

Information Theory and Coding Techniques: Chapter 1.1. What is Information Theory? Why you should take this course?

Information Theory and ID3 Algo.

Introduction to Information Theory

Information Theory. David Rosenberg. June 15, New York University. David Rosenberg (New York University) DS-GA 1003 June 15, / 18

Fundamentals of Information Theory Lecture 1. Introduction. Prof. CHEN Jie Lab. 201, School of EIE BeiHang University

Some Basic Concepts of Probability and Information Theory: Pt. 2

Gambling and Information Theory

Noisy channel communication

Transcription:

Information in Biology CRI - Centre de Recherches Interdisciplinaires, Paris May 2012 Information processing is an essential part of Life. Thinking about it in quantitative terms may is useful. 1

Living information is carried by molecular channels Living systems I. Self-replicating information processors Environment II. III. Evolve collectively. Made of molecules. Generic properties of molecular channels subject to evolution? Information theory approach? Other biological information channels. 2

Outline Information in Biology Information in Biology Concept of information is found in many living systems: DNA, signaling, neuron, ribosomes, evolution. Goals: (1) Formalize and quantify biological information. (2) Application to various biological systems. (3) Looking for common principles. I. Information and Statistical Mechanics: Shannon s information theory and its relation to statistical mechanics. II. III. Overview: Living Information: molecules, neurons, population and evolution. Living systems as information sources, channels and processors. Molecular information and noise. IV. Neural networks and coding theory. V. Population dynamics, social interaction and sensing. 3

I. Basics of Information Theory (Shannon) 4

Shannon s Information theory Information theory: a branch of applied math and electrical engineering. Developed by Claude E. Shannon. Main results: fundamental limits on signal processing such as, How well data can be compressed? What is the reliability of communicating signals? Numerous applications (besides communication eng.): Physics (stat mech), Math (statistical inference), linguistics, Computer science (cryptography, complexity), Economics (portfolio theory). The key quantity which measures information is entropy: Quantifies the uncertainty involved in predicting the value of a random variable (e.g., a coin flip or a die). What are the biological implications? 5

Claude Elwood Shannon (1916-2001) 1937 master's thesis: A Symbolic Analysis of Relay and Switching Circuits. 1940 Ph.D. thesis: An Algebra for Mandelian Genetics. WWII (Bell labs) works on cryptography and fire-control systems: Data Smoothing and Prediction in Fire-Control Systems. Communication theory of secrecy systems. 1948: Mathematical Theory of Communication. 1949: Sampling theory: Analog to digital. 1951: Prediction and Entropy of Printed English. 1950: Shannon s mouse: 1 st artificial learning machine. 1950: Programming a Computer for Playing Chess. 1960: 1 st wearable computer, Las Vegas. 6

A Mathematical Theory of Communication Shannon s seminal paper: "A Mathematical Theory of Communication". Bell System Technical Journal 27 (3): 379 423 (1948). Basic scheme of communication: Information source produces messages. Transmitter converts message to signal. Channel conveys the signal with Noise. Receiver transforms the signal back into the message Destination: machine, person, organism receiving the message. Introduces information entropy measured in bits. 7

What is information? "The fundamental problem of communication is that of reproducing at one point, either exactly or approximately, a message selected at another point." Engineering perspective How to make good transmission channels Problem with telegraph lines, Define information as measure for the surprise If a binary channel transmits only 1 s there is no information (no surprise). If the channel transmits 0 s and 1 s with equal probability max. information. 8

Intuition: 20 questions game Try to guess the object from: {barrel, cat, dog, ball, fish, box, building}. First strategy: wild guess. 9

Intuition: 20 questions game Optimal strategy: equalized tree Information = # of yes/no questions in an optimal tree. 10

Introducing the bit If I have a (equal) choice between two alternatives the information is: I=1 bit = log 2 (#Alternatives) Harry Nyquist (1924): 1 bit = Certain Factors Affecting Telegraph Speed Example: How many bits are in a genome of length N? 11

Shannon s axioms of information For random variable that distributes as p 1 p n,search for a function H(p 1,..,p n ) that measures information that obeys: 1) Continuous in p i 2) For equal probabilities it is Nyquist expression: H(1/n,,1/n)=log 2 (n). 3) Tree property : invariant to redistribution into stages. 12

Information from Shannon s axioms Shannon showed that the only function that obeys these axioms is H p log p log p i i 2 i 2 i (up to proportion constant). Example : if X is uniformly distributed over 128 outcomes 1 1 7 H p log p log log 2 7 bits i i 2 i 7 2 7 2 i 2 2 Example: very uneven coin shows head only 1 in 1024 times 10 10 10 10 10 ilog2 i 2 log22 (1 2 )log 2(1 2 ) (10 log 2 )2 bits 13 i H p p e

Entropy of a Binary Channel H plog p (1 p)log (1 p) 2 2 14

Why call it entropy? Shannon discussed this problem with John von Neumann: My greatest concern was what to call it. I thought of calling it information, but the word was overly used, so I decided to call it uncertainty. When I discussed it with John von Neumann, he had a better idea. Von Neumann told me, You should call it entropy, for two reasons. In the first place your uncertainty function has been used in statistical mechanics under that name, so it already has a name. In the second place, and more important, nobody knows what entropy really is, so in a debate you will always have the advantage. M. Tribus, E.C. McIrvine, Energy and information, Scientific American, 224 (1971). 15

Conditional probability Consider two random variables X,Y with a joint probability distribution P(X,Y) The joint entropy is H(X,Y) The mutual entropy is H(X,Y)-H(X)-H(Y) The conditional entropies are H(X Y) and H(Y X) 16

Joint entropy H(X,Y) measures total entropy of the joint distribution P(X,Y) Mutual entropy I(X;Y) measures correlations ( P(x,y)=P(x)P(y) => I=0 ) Conditional entropy H(X Y) measures remaining uncertainty of X given Y 17

More information measures 18

Whose entropy? E. T. Jaynes 19

Kullback Leibler entropy Problem: Suppose a random variable X is distributed according to P(X) but we expect it to be distributed according to Q(X). What is the level of our surprise? Answer: The Kullback Leibler divergence pi DKL( p q) pi log2 q i i Mutual information I( X, Y ) D ( p( x, y) p( x) p( y)) KL Appears in many circumstances Example Markov chains 20

Shannon Entropy and Statistical Mechanics 21

I. Maxwell s Demon Entropy in statistical mechanics: measure of uncertainty of a system after specifying its macroscopic observables such as temperature and pressure. Given macroscopic variables, entropy measures the degree of spreading probability over different possible states. Boltzmann famous formula: S ln 22

The second law of thermodynamics The second law of thermodynamics: In general, the total entropy of a system isolated from its environment, the entropy of that system will tend not to decrease. Consequences: (i) (ii) heat will not flow from a colder body to a hotter body without work. No perpetuum mobile: One cannot produce net work from a single temperature reservoir (production of net work requires flow of heat from a hotter reservoir to a colder reservoir). 23

Maxwell s thought experiment How to violate the Second Law? - Container divided by an insulated wall. - Door can be opened and closed by a demon. - Demon opens the door to allow only "hot" molecules of gas to flow to the favored side. - One side heats up while other side cools down: decreasing entropy. Breaking 2 nd law! 24

Solution: entropy = -information Demon reduces the thermodynamic entropy of a system using information about individual molecules (their direction) Landauer (1961) showed that he demon must increase thermodynamic entropy by at least the amount of Shannon information he acquires and stores; total thermodynamic entropy does not decrease! Landauer s principle : Any logically irreversible manipulation of information, such as the erasure of a bit, must be accompanied by a corresponding entropy increase in the information processing apparatus or its environment. min E k T ln 2 B 25

II. Second law in Markov chains Random walk on a graph:. W is transition matrix : p(t+1)=wp(t) p* be the steady state solution: Wp*=p* Theorem: distribution approaches steady-state t D(p p*) <=0 Also t D(p p*)=0 <=> p=p* In other words: Markov dynamics dissipates any initial information. 26

Maximum entropy inference (E. T. Jaynes) Problem: Given partial knowledge e.g. the average value <X> of X how should we assign probabilities to outcomes P(X)? Answer: choose the probability distribution that maximizes the entropy (surprise) and is consistent with what we already know. Example: given energies E i and measurement <E> what is p i? Exercise: which dice X={1,2,3,4,5,6} gives <X>=3? 27

Maximum Entropy principle relates thermodynamics and information At equilibrium, TD entropy = Shannon information needed to define the microscopic state of the system, given its macroscopic description. Gain in entropy always means loss of information about this state. Equivalently, TD entropy = minimum number of yes/no questions needed to be answered in order to fully specify the microstate. 28