Towards a Theory of Information Flow in the Finitary Process Soup

Similar documents
Channel capacity. Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5.

(Classical) Information Theory III: Noisy channel coding

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Dept. of Linguistics, Indiana University Fall 2015

Shannon s noisy-channel theorem

Noisy channel communication

Solutions to Homework Set #3 Channel and Source coding

ITCT Lecture IV.3: Markov Processes and Sources with Memory

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy

PROBABILITY AND INFORMATION THEORY. Dr. Gjergji Kasneci Introduction to Information Retrieval WS

Lecture 2: August 31

Computing and Communications 2. Information Theory -Entropy

Cut-Set Bound and Dependence Balance Bound

Chapter 2: Entropy and Mutual Information. University of Illinois at Chicago ECE 534, Natasha Devroye

1 Introduction to information theory

The Method of Types and Its Application to Information Hiding

Hierarchical Self-Organization in the Finitary Process Soup

arxiv: v1 [q-bio.pe] 28 Apr 2007

Shannon s Noisy-Channel Coding Theorem

Upper Bounds on the Capacity of Binary Intermittent Communication

Chapter 9 Fundamental Limits in Information Theory

Variable Length Codes for Degraded Broadcast Channels

INTRODUCTION TO INFORMATION THEORY

EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018

Universal Anytime Codes: An approach to uncertain channels in control

Revision of Lecture 5

EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY

Chapter 3 Source Coding. 3.1 An Introduction to Source Coding 3.2 Optimal Source Codes 3.3 Shannon-Fano Code 3.4 Huffman Code

Lecture 1: Introduction, Entropy and ML estimation

Information Theory and Statistics Lecture 2: Source coding

(Preprint of paper to appear in Proc Intl. Symp. on Info. Th. and its Applications, Waikiki, Hawaii, Nov , 1990.)

EE 4TM4: Digital Communications II. Channel Capacity

Entropies & Information Theory

Information Theory in Intelligent Decision Making

Exercise 1. = P(y a 1)P(a 1 )

An Introduction to Information Theory: Notes

Lecture 6 I. CHANNEL CODING. X n (m) P Y X

Equivalence of Regular Expressions and FSMs

Superposition Encoding and Partial Decoding Is Optimal for a Class of Z-interference Channels

CS256/Spring 2008 Lecture #11 Zohar Manna. Beyond Temporal Logics

Also, in recent years, Tsallis proposed another entropy measure which in the case of a discrete random variable is given by

MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK UNIT V PART-A. 1. What is binary symmetric channel (AUC DEC 2006)

Communication Theory II

ELEC546 Review of Information Theory

The Poisson Channel with Side Information

Capacity of a channel Shannon s second theorem. Information Theory 1/33

(each row defines a probability distribution). Given n-strings x X n, y Y n we can use the absence of memory in the channel to compute

5 Mutual Information and Channel Capacity

MAHALAKSHMI ENGINEERING COLLEGE QUESTION BANK. SUBJECT CODE / Name: EC2252 COMMUNICATION THEORY UNIT-V INFORMATION THEORY PART-A

Lecture 14 February 28

Coding into a source: an inverse rate-distortion theorem

Lecture 3: Nondeterministic Finite Automata

Information Theory - Entropy. Figure 3

CS 154 Formal Languages and Computability Assignment #2 Solutions

Capacity of the Discrete Memoryless Energy Harvesting Channel with Side Information

EE229B - Final Project. Capacity-Approaching Low-Density Parity-Check Codes

Principles of Communications

Chapter I: Fundamental Information Theory

UNIT I INFORMATION THEORY. I k log 2

Block 2: Introduction to Information Theory

LECTURE 3. Last time:

ECE 4400:693 - Information Theory

CS 630 Basic Probability and Information Theory. Tim Campbell

Lecture 4 Noisy Channel Coding

3F1 Information Theory, Lecture 3

Lecture 8: Channel Capacity, Continuous Random Variables

Lecture 2. Capacity of the Gaussian channel

Noisy-Channel Coding

Relay Networks With Delays

Feedback Capacity of a Class of Symmetric Finite-State Markov Channels

Module 1. Introduction to Digital Communications and Information Theory. Version 2 ECE IIT, Kharagpur

Equivalence for Networks with Adversarial State

1 Ex. 1 Verify that the function H(p 1,..., p n ) = k p k log 2 p k satisfies all 8 axioms on H.

Entropy Rate of Stochastic Processes

Bioinformatics: Biology X

UCSD ECE 255C Handout #12 Prof. Young-Han Kim Tuesday, February 28, Solutions to Take-Home Midterm (Prepared by Pinar Sen)

Revision of Lecture 4

Information in Biology

Directed and Undirected Graphical Models

Computation Theory Finite Automata

Deterministic Finite Automata (DFAs)

An Outer Bound for the Gaussian. Interference channel with a relay.

Can Feedback Increase the Capacity of the Energy Harvesting Channel?

Entropy and Ergodic Theory Lecture 4: Conditional entropy and mutual information

Towards control over fading channels

Deterministic Finite Automata (DFAs)

Discrete Memoryless Channels with Memoryless Output Sequences

Complexity 6: AIT. Outline. Dusko Pavlovic. Kolmogorov. Solomonoff. Chaitin: The number of wisdom RHUL Spring Complexity 6: AIT.

3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions

Lecture 18: Shanon s Channel Coding Theorem. Lecture 18: Shanon s Channel Coding Theorem

(Classical) Information Theory II: Source coding

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

The Modeling and Complexity of Dynamical Systems by Means of Computation and Information Theories

Information in Biology

Computational Systems Biology: Biology X

We define the multi-step transition function T : S Σ S as follows. 1. For any s S, T (s,λ) = s. 2. For any s S, x Σ and a Σ,

An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if. 2 l i. i=1

X 1 : X Table 1: Y = X X 2

A Mathematical Theory of Communication

Energy State Amplification in an Energy Harvesting Communication System

Transcription:

Towards a Theory of in the Finitary Process Department of Computer Science and Complexity Sciences Center University of California at Davis June 1, 2010

Goals Analyze model of evolutionary self-organization in terms of information flow. Measure channel capacity of elementary. Develop functional partitioning based on this and language properties.

Outline 1 2 3 4 5 6

ǫ-machines ǫ-machines Defined T = {S,T } S is a set of causal states T is the set of transitions between them: T (s) ij, s A ǫ-machine Properties All of their recurrent states form a single strongly connected component. Transitions are deterministic. S is minimal: an ǫ-machine is the smallest causal representation of the transformation it implements.

Transducers We interpret the symbols labeling the transitions in the alphabet A as consisting of two parts: an input symbol that determines which transition to take from a state and an output symbol which is emitted on taking that transition. Transducers implement functions: Character to character Input string to output string Map sets to sets (languages)

Set of Machines 15 0 0 0 1 1 0 1 1 14 0 1 1 0 1 1 13 0 0 1 0 1 1 12 1 0 1 1 11 0 0 0 1 1 1 10 0 1 1 1 9 0 0 1 1 8 1 1 7 0 0 0 1 1 0 6 0 1 1 0 5 0 0 1 0 4 1 0 3 0 0 0 1 2 0 1 1 0 0

Transducer Composition Create new mapping, input language of one machine becomes input language of another Not commutative Possible exponential growth in number of states.

Interaction Network Interaction Matrix G (k) G (k) ij = { 1 iftk = T j T i 0 otherwise

Transition Matrix T 1 T 2 T 3 T 0 T 1 T 2 T 3 T 0 T 1 T 2 T 3 T 0 T 1 T 2 T 3 T 0 T 0 T 0 T 1 T 1 T 1 T 1 T 2 T 2 T 2 T 2 T 3 T 3 T 3 T 3 T 1 T 2 T 3 T 1 T 1 T 3 T 3 T 2 T 3 T 2 T 3 T 3 T 3 T 3 T 3 T 4 T 8 T 12 T 0 T 4 T 8 T 12 T 0 T 4 T 8 T 12 T 0 T 4 T 8 T 12 T 5 T 10 T 15 T 0 T 5 T 10 T 15 T 0 T 5 T 10 T 15 T 0 T 5 T 10 T 15 T 4 T 8 T 12 T 1 T 5 T 9 T 13 T 2 T 6 T 10 T 14 T 3 T 7 T 11 T 15 T 5 T 10 T 15 T 1 T 5 T 11 T 15 T 2 T 7 T 10 T 15 T 3 T 7 T 11 T 15 T 0 T 0 T 0 T 4 T 4 T 4 T 4 T 8 T 8 T 8 T 8 T 12 T 12 T 12 T 12 T 1 T 2 T 3 T 4 T 5 T 6 T 7 T 8 T 9 T 10 T 11 T 12 T 13 T 14 T 15 T 0 T 0 T 0 T 5 T 5 T 5 T 5 T 10 T 10 T 10 T 10 T 15 T 15 T 15 T 15 T 1 T 2 T 3 T 5 T 5 T 7 T 7 T 10 T 11 T 10 T 11 T 15 T 15 T 15 T 15 T 4 T 8 T 12 T 4 T 4 T 12 T 12 T 8 T 12 T 8 T 12 T 12 T 12 T 12 T 12 T 5 T 10 T 15 T 4 T 5 T 14 T 15 T 8 T 13 T 10 T 15 T 12 T 13 T 14 T 15 T 4 T 8 T 12 T 5 T 5 T 13 T 13 T 10 T 14 T 10 T 14 T 15 T 15 T 15 T 15 T 5 T 10 T 15 T 5 T 5 T 15 T 15 T 10 T 15 T 10 T 15 T 15 T 15 T 15 T 15

Interaction Network Interaction Graph G Nodes correspond to ǫ-transducers If T C = T B T A, place directed edge connecting node T A, to T C, labeled with the transforming machine T B.

T_12 T_8 T_14 T_6 T_1 T_4 T_13 T_15 T_3 T_7 T_11 T_1 T_4 T_5 T_2 T_8T_9T_10T_11 T_15 T_3T_6T_7T_9T_11T_12T_13T_14T_15 T_2 T_8 T_10 T_3 T_1 T_1T_5T_9T_13 T_10 T_4 T_5 T_6 T_7 T_2 T_2 T_4 T_3 T_8 T_5 T_12 T_5 T_2 T_7 T_9T_13 T_6 T_14 T_3 T_1 T_7 T_11 T_15 T_1 T_3 T_5 T_6 T_7 T_11 T_14 T_9T_11 T_7 T_11 T_12 T_13 T_14 T_15 T_2 T_9 T_10 T_10 T_3 T_7 T_11 T_12 T_13 T_14 T_15 T_1 T_4 T_5 T_3 T_7 T_11 T_15 T_3T_6T_7T_9T_11T_12T_13T_14T_15 T_4 T_15 T_8 T_10 T_15 T_9 T_6 T_6 T_15 T_13 T_6 T_14 T_5 T_10 T_9 T_7 T_2 T_13 T_6 T_9T_13 T_11 T_14 T_6 T_7 T_14 T_9T_11 T_10 T_8 T_10 T_4 T_3 T_4 T_13 T_14 T_10 T_8T_9T_10T_11 T_15 T_12 T_1 T_1T_5 T_12 T_3T_6T_7T_9T_11T_12T_13T_14T_15 T_8 T_12 T_2 T_8 T_10 T_13 T_14 T_15 T_14 T_5 T_1T_5T_9T_13 T_6 T_7 T_4 T_2 T_6 T_10 T_4 T_5 T_2 T_12 T_12 T_4 T_6 T_10 T_14 T_5 T_3 T_7 T_11 T_15 T_1T_5T_9T_13 T_3 T_4 T_5 T_4 T_5 T_6 T_7 T_1 T_2 T_8 T_8T_9T_10T_11 T_8 T_2 T_12 T_13 T_14 T_15 T_8 T_10 T_1 T_2 Interaction Network G

Population Population P N individuals

Population A single replication is determined through compositions and replacements in a two-step sequence: 1 Construct ǫ-machine T C by forming the composition T C = T B T A from T A and T B randomly selected from the population and minimizing. 2 Replace a randomly selected ǫ-machine, T D, with T C. Note that there is no imposed notion of fitness nor spatial component.

Relaxation to Steady State ( of Size 100,000) 0.25 T 15 0.2 p 0.15 T 3, T 5, T 10, T 12 0.1 T 1, T 2, T 4, T 8 0.05 0 T 7, T 11, T 13, T 14 T 6, T 9 0 2 4 6 8 10 t/n

Communication Discrete [Cover and Thomas, 2006] (X,p(y x), Y): (X and Y) are finite sets p(y x) are probability mass functions

Cover and Thomas [Cover and Thomas, 2006] define the channel capacity of a discrete memoryless channel as: C = max I(X;Y ) (1) p(x) This capacity specifies the highest rate, in bits, at which information may be reliably transmitted through the channel, and Shannon s second theorem states that his rate is achievable in practice.

Mutual Remember that mutual information is the reduction in uncertainty in one random variable due to knowledge of another. I(X;Y ) = H(Y ) H(Y X) (2)

Discrete Noiseless Discrete Noiseless One to one correspondence between input and output symbols

Languages based on language L L in = Σ L out = Σ union of these possibility for positive channel capacity

Capacities 1 bits 0.5 0 1 0.5 0 1 0.5 0 0 0.5 1 0 0.5 1 0 0.5 1 0 0.5 1 0 0.5 1 P(X)

Machines of Maximal 1 bits 0.5 0 0 0.5 1 P(X)

Partially Noisy s 1 bits 0.5 0 0 0.5 1 P(X)

Partially Noisy s 1 bits 0.5 0 0 0.5 1 P(X)

Capacities Most are zero, except for... Machine Number P(X=1) 6 1.0 0.5 7 0.32 0.6 9 1.0 0.5 11 0.32 0.6 13 0.32 0.4 14 0.32 0.4

Meta-machines Meta-machine A set of machines that is closed and self-maintained under composition Ω P is a meta-machine if and only if (i) T i T j Ω, for all T i,t j Ω and (ii) For all T k Ω, there exists T i,t j Ω, such that T k = T i T j. Captures notion of invariant set: Ω = G Ω

T_12 T_8 T_14 T_6 T_1 T_4 T_13 T_15 T_3 T_7 T_11 T_1 T_4 T_5 T_2 T_8T_9T_10T_11 T_15 T_3T_6T_7T_9T_11T_12T_13T_14T_15 T_2 T_8 T_10 T_3 T_1 T_1T_5T_9T_13 T_10 T_4 T_5 T_6 T_7 T_2 T_2 T_4 T_3 T_8 T_5 T_12 T_5 T_2 T_7 T_9T_13 T_6 T_14 T_3 T_1 T_7 T_11 T_15 T_1 T_3 T_6 T_7 T_11 T_14 T_9T_11 T_7 T_11 T_12 T_13 T_14 T_15 T_2 T_9 T_10 T_10 T_3 T_7 T_11 T_12 T_13 T_14 T_15 T_1 T_4 T_5 T_3 T_7 T_11 T_15 T_3T_6T_7T_9T_11T_12T_13T_14T_15 T_5 T_4 T_15 T_8 T_10 T_15 T_9 T_6 T_6 T_15 T_13 T_6 T_14 T_5 T_10 T_9 T_7 T_2 T_13 T_6 T_9T_13 T_11 T_14T_6 T_7 T_14 T_9T_11 T_10 T_8 T_10 T_4 T_3 T_4 T_13 T_14 T_10 T_8T_9T_10T_11 T_15 T_12 T_1 T_1T_5 T_12 T_3T_6T_7T_9T_11T_12T_13T_14T_15 T_8 T_12 T_2 T_8 T_10 T_13 T_14 T_15 T_14 T_5 T_6 T_7 T_1T_5T_9T_13 T_4 T_2 T_6 T_10 T_4 T_5 T_2 T_12 T_12 T_4 T_6 T_10 T_14 T_5 T_1T_5T_9T_13 T_3 T_3 T_7 T_11 T_15 T_4 T_5 T_4 T_5 T_6 T_7 T_1 T_2 T_8 T_8T_9T_10T_11 T_8 T_2 T_12 T_13 T_14 T_15 T_8 T_10 T_1 T_2

T A T D T E T C T O T O T L T C T A (0.125) T A T E T T L T O T J T H B T E (0.125) TCTO TA TD T E T D T E T B T J T O (0.225) T O T L T J TH T J (0.125) T B T J TH T A T D T E T C T O T O T L T C T C (0.125) T T L T O T J T H B T A T E T D (0.069) T D T E T B T J T H (0.069) T J T H T O T L T C T J T H T J (0.069) T D T E T J T B T B (0.069) T H T J [Crutchfield, 2006]

Summary capacity is a function of machine structure. High channel capacity does not necessarily lead to persistence. All of the transducers in the single state meta-machine have zero channel capacity. Composition never increases channel capacity.

For Further Reading T. M. Cover and J. A. Thomas Elements of Theory Wiley-Interscience, 2006 J. P. Crutchfield and O. Görnerup Objects That Make Objects: The Population of Structural Complexity Journal of the Royal Society Interface, 3 (2006) 345-349.