Introduction to Information Theory. Part 2

Similar documents
Introduction to Information Theory. Part 3

Introduction to Information Theory. Part 4

Chapter 2: Source coding

Multimedia Communications. Mathematical Preliminaries for Lossless Compression

Lecture 1: Shannon s Theorem

3F1 Information Theory, Lecture 3

6.02 Fall 2012 Lecture #1

Entropy as a measure of surprise

10-704: Information Processing and Learning Fall Lecture 10: Oct 3

COMM901 Source Coding and Compression. Quiz 1

Bandwidth: Communicate large complex & highly detailed 3D models through lowbandwidth connection (e.g. VRML over the Internet)

3F1 Information Theory, Lecture 3

Chapter 2 Date Compression: Source Coding. 2.1 An Introduction to Source Coding 2.2 Optimal Source Codes 2.3 Huffman Code

UNIT I INFORMATION THEORY. I k log 2

DCSP-3: Minimal Length Coding. Jianfeng Feng

Data Compression. Limit of Information Compression. October, Examples of codes 1

Information Theory (Information Theory by J. V. Stone, 2015)

Entropies & Information Theory

INTRODUCTION TO INFORMATION THEORY

An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if. 2 l i. i=1

Shannon s Noisy-Channel Coding Theorem

A Mathematical Theory of Communication

Chapter 3 Source Coding. 3.1 An Introduction to Source Coding 3.2 Optimal Source Codes 3.3 Shannon-Fano Code 3.4 Huffman Code

CSCI 2570 Introduction to Nanocomputing

An introduction to basic information theory. Hampus Wessman

A Simple Introduction to Information, Channel Capacity and Entropy

Information Theory and Coding Techniques

Chapter 9 Fundamental Limits in Information Theory

Shannon's Theory of Communication

Information Theory. David Rosenberg. June 15, New York University. David Rosenberg (New York University) DS-GA 1003 June 15, / 18

(Classical) Information Theory III: Noisy channel coding

ECE INFORMATION THEORY. Fall 2011 Prof. Thinh Nguyen

1 Ex. 1 Verify that the function H(p 1,..., p n ) = k p k log 2 p k satisfies all 8 axioms on H.

Information in Biology

Lecture 16. Error-free variable length schemes (contd.): Shannon-Fano-Elias code, Huffman code

EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018

EE376A: Homework #2 Solutions Due by 11:59pm Thursday, February 1st, 2018

PERFECT SECRECY AND ADVERSARIAL INDISTINGUISHABILITY

Information in Biology

10-704: Information Processing and Learning Fall Lecture 9: Sept 28

Digital Communications III (ECE 154C) Introduction to Coding and Information Theory

Run-length & Entropy Coding. Redundancy Removal. Sampling. Quantization. Perform inverse operations at the receiver EEE

Source Coding. Master Universitario en Ingeniería de Telecomunicación. I. Santamaría Universidad de Cantabria

Lecture 2: August 31

Chapter 2 Source Models and Entropy. Any information-generating process can be viewed as. computer program in executed form: binary 0

Topics. Probability Theory. Perfect Secrecy. Information Theory

Information Theory. Week 4 Compressing streams. Iain Murray,

3F1 Information Theory, Lecture 1

Lecture 4 : Adaptive source coding algorithms

U Logo Use Guidelines

Intro to Information Theory

1. Basics of Information

Basic Principles of Lossless Coding. Universal Lossless coding. Lempel-Ziv Coding. 2. Exploit dependences between successive symbols.

Outline of the Lecture. Background and Motivation. Basics of Information Theory: 1. Introduction. Markku Juntti. Course Overview

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy

17.1 Binary Codes Normal numbers we use are in base 10, which are called decimal numbers. Each digit can be 10 possible numbers: 0, 1, 2, 9.

CHAPTER 4 ENTROPY AND INFORMATION

Exercises with solutions (Set B)

Coding of memoryless sources 1/35

Source Coding Techniques

6.02 Fall 2011 Lecture #9

Welcome to Comp 411! 2) Course Objectives. 1) Course Mechanics. 3) Information. I thought this course was called Computer Organization

Entropy Coding. Connectivity coding. Entropy coding. Definitions. Lossles coder. Input: a set of symbols Output: bitstream. Idea

Lecture 4 Channel Coding

( c ) E p s t e i n, C a r t e r a n d B o l l i n g e r C h a p t e r 1 7 : I n f o r m a t i o n S c i e n c e P a g e 1

SIGNAL COMPRESSION Lecture Shannon-Fano-Elias Codes and Arithmetic Coding

Information Theory CHAPTER. 5.1 Introduction. 5.2 Entropy

Lecture 18: Shanon s Channel Coding Theorem. Lecture 18: Shanon s Channel Coding Theorem

2018/5/3. YU Xiangyu

D CUME N TAT ION C 0 R P 0 R A T E D 2,21 CONNECTICUT AVENUE, N. W,

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Dept. of Linguistics, Indiana University Fall 2015

3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions

Homework Set #2 Data Compression, Huffman code and AEP

Introduction to Information Theory

Information & Correlation

EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY

Murray Gell-Mann, The Quark and the Jaguar, 1995

MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK UNIT V PART-A. 1. What is binary symmetric channel (AUC DEC 2006)

Multimedia. Multimedia Data Compression (Lossless Compression Algorithms)

PROBABILITY AND INFORMATION THEORY. Dr. Gjergji Kasneci Introduction to Information Retrieval WS

Information. = more information was provided by the outcome in #2

A Gentle Tutorial on Information Theory and Learning. Roni Rosenfeld. Carnegie Mellon University

at Some sort of quantization is necessary to represent continuous signals in digital form

Data Compression Techniques

Quantitative Biology Lecture 3

Fibonacci Coding for Lossless Data Compression A Review

Do we live in a world of facts or information? What information anyway?

Entropy in Classical and Quantum Information Theory

Chaos, Complexity, and Inference (36-462)

! Where are we on course map? ! What we did in lab last week. " How it relates to this week. ! Compression. " What is it, examples, classifications

Huffman Coding. C.M. Liu Perceptual Lab, College of Computer Science National Chiao-Tung University

CSE468 Information Conflict

Information and Entropy. Professor Kevin Gold

LECTURE 15. Last time: Feedback channel: setting up the problem. Lecture outline. Joint source and channel coding theorem

Sequences and Information

1 Introduction to information theory

Lecture 6: Kraft-McMillan Inequality and Huffman Coding

AQI: Advanced Quantum Information Lecture 6 (Module 2): Distinguishing Quantum States January 28, 2013

CSE 421 Greedy: Huffman Codes

Transcription:

Introduction to Information Theory Part 2 1

A General Communication System CHANNEL Information Source Transmitter Channel Receiver Destination 2

Information: Definition Information is quantified using probabilities. Given a finite set of possible messages, associate a probability with each message. A message with low probability represents more information than one with high probability. Definition of Information: Where is the probability of the message Base 2 is used for the logarithm so is measured in bits Trits for base 3, nats for base, Hartleys for base 10 3

Some properties of Information is non negative. Information we get from observing two independent events occurring is the sum of two information(s). is monotonic and continuous in Slight changes in probability incur slight changes in information. We get zero information from an event whose probability is 1. 4

Example: Information in a coin flip 5

Independent Events: 2 Coin flips There are four possibilities: HH, HT, TH, TT i.e. Additive property: 6

Example: Text Analysis a 0.06428 b 0.01147 c 0.02413 d 0.03188 e 0.10210 f 0.01842 g 0.01543 h 0.04313 i 0.05767 j 0.00082 k 0.00514 l 0.03338 m 0.01959 n 0.05761 o 0.06179 p 0.01571 q 0.00084 r 0.04973 s 0.05199 t 0.07327 u 0.02201 v 0.00800 w 0.01439 x 0.00162 y 0.01387 z 0.00077 SPC 0.20096 0.25000 0.20000 0.15000 0.10000 0.05000 0.00000 SPC SPC e e t t a o a o i n s r i n s r h h l l d c u m f p g w y b v k x q j z d c u m f p g w y b v k x q j z 7

Example: Text Analysis Letter Freq. I a 0.06428 3.95951 b 0.01147 6.44597 c 0.02413 5.37297 d 0.03188 4.97116 e 0.10210 3.29188 f 0.01842 5.76293 g 0.01543 6.01840 h 0.04313 4.53514 i 0.05767 4.11611 j 0.00082 10.24909 k 0.00514 7.60474 l 0.03338 4.90474 m 0.01959 5.67385 n 0.05761 4.11743 o 0.06179 4.01654 p 0.01571 5.99226 q 0.00084 10.21486 r 0.04973 4.32981 s 0.05199 4.26552 t 0.07327 3.77056 u 0.02201 5.50592 v 0.00800 6.96640 w 0.01439 6.11899 x 0.00162 9.26697 y 0.01387 6.17152 z 0.00077 10.34877 SPC 0.20096 2.31502 8

Definition of Entropy Information ( ) is associated with known events/messages Entropy ( ) is the average information w.r.to all possible outcomes. Given, Characterizes an information source. 9

Example: A 3 event Source ) 10

Example: Text Analysis Letter Freq. I a 0.06428 3.95951 b 0.01147 6.44597 c 0.02413 5.37297 d 0.03188 4.97116 e 0.10210 3.29188 f 0.01842 5.76293 g 0.01543 6.01840 h 0.04313 4.53514 i 0.05767 4.11611 j 0.00082 10.24909 k 0.00514 7.60474 l 0.03338 4.90474 m 0.01959 5.67385 n 0.05761 4.11743 o 0.06179 4.01654 p 0.01571 5.99226 q 0.00084 10.21486 r 0.04973 4.32981 s 0.05199 4.26552 t 0.07327 3.77056 u 0.02201 5.50592 v 0.00800 6.96640 w 0.01439 6.11899 x 0.00162 9.26697 y 0.01387 6.17152 z 0.00077 10.34877 SPC 0.20096 2.31502 Aka, First Order Entropy. 11

Entropy (2 outcomes) 12

Entropy: Properties Entropy is maximized if is uniform. Additive property for independent events. If and are not independent. 13

Entropy of things Entropy of English text is approx 1.5 bits Entropy of the human genome <= 2 bits Entropy of a black hole is ¼ of the area of the outer event horizon. Value of information in economics is defined in terms of entropy. E.g. Scarcity 14

Entropy: What about it? Does have a maximum? Where? Is entropy a good name for this stuff? How is it related to entropy in thermodynamics? How does entropy help in communication? What else can we do with it? Why use the letter? 15

Entropy & Codes Entropy is closely related to the design of efficient codes for random sources. Provides foundations for techniques of compression, data search, encryption, correction of communication errors, etc. Essential to the study of life sciences, economics, etc. 16

Coding: Basics Events of an information source:,, A code is made up of codewords from a code alphabet (e.g. {0, 1}, {., }, etc.) A code is an assignment of codewords to source symbols. 17

Coding: Basics Block code: When all codes have the same length. For example, ASCII (8 bits) Average Word Length: More generally, 1 A code is efficient if it has the smallest average word length. (Turns out entropy is the benchmark ) 18

Coding: Basics Singular (not unique) codes Nonsingular (unique) codes Symbol Singular Code Nonsingular Code A 00 0 B 10 10 C 01 00 D 10 01 19

Coding: Basics Singular (not unique) codes Nonsingular (unique) codes instantaneous codes (every word can be decoded as soon as it is received) Symbol Singular Code Nonsingular Code A 00 0 B 10 10 C 01 00 D 10 01 Not an instantaneous Code! 20

Example: Avg. Code Length (L) Symbol p Codeword A 0.3 00 B 0.2 10 C 0.2 11 D 0.2 010 E 0.1 011 0.3 2 0.2 2 0.2 2 0.2 3 0.1 3 2.3 21

Example: Source Entropy (H) Symbol p Codeword A 0.3 00 B 0.2 10 C 0.2 11 D 0.2 010 E 0.1 011 0.3 2 0.2 2 0.2 2 0.2 3 0.1 3 2.3 0.3log 1 0.3 0.2log 1 0.2 30.1log 1 0.1 2.246 22

Example: & Symbol p Codeword A 0.3 00 B 0.2 10 C 0.2 11 D 0.2 010 E 0.1 011 0.3 2 0.2 2 0.2 2 0.2 3 0.1 3 2.3 0.3log Is there a relationship between and? 1 0.3 0.2log 1 0.2 30.1log 1 0.1 2.246 23

Average Code Length & Entropy Average length bounds: Grouping symbols together: 24

Average Code Length & Entropy Average length bounds: Grouping symbols together: 25

Average Code Length & Entropy Average length bounds: Grouping symbols together: 26

Average Code Length & Entropy Average length bounds: Grouping symbols together: 27

Average Code Length & Entropy Average length bounds: 1 Grouping symbols together: 1 1 1 lim H 28

Shannon s First Theorem By coding sequences of independent symbols (in ), it is possible to construct codes such that The price paid for such improvement is increased coding complexity (due to increased ) and increased delay in coding. 29

Entropy & Coding: Central Ideas Use short codes for highly likely events. This shortens the average length of coded messages. Code several events at a time. Provides greater flexibility in code design. 30

Data Compression: Huffman Coding A 0.3 B 0.2 C 0.2 D 0.2 E 0.1 31

Huffman Coding: Reduction Phase A 0.3 0.3 0.4 0.6 B 0.2 0.3 0.3 0.4 C 0.2 0.2 0.3 D 0.2 0.2 E 0.1 32

Huffman Coding: SplittingPhase A 0.3 00 0.3 00 0.4 1 0.6 0 B 0.2 10 0.3 01 0.3 00 0.4 1 C 0.2 11 0.2 10 0.3 01 D 0.2 010 0.2 11 E 0.1 011 33

Huffman Coding: SplittingPhase A 0.3 00 0.3 00 0.4 1 0.6 0 B 0.2 10 0.3 01 0.3 00 0.4 1 C 0.2 11 0.2 10 0.3 01 D 0.2 010 0.2 11 E 0.1 011 34

Huffman Codes Nonsingular Instantaneous Efficient Non unique Powers of a source lead closer to Requires knowledge of symbol probabilities 35

Design Huffman Codes 36

References Eugene Chiu, Jocelyn Lin, Brok Mcferron, Noshirwan Petigara, Satwiksai Seshasai: Mathematical Theory of Claude Shannon: A study of the style and context of his work up to the genesis of information theory. MIT 6.933J / STS.420J The Structure of Engineering Revolutions Luciano Floridi, 2010: Information: A Very Short Introduction, Oxford University Press, 2011. Luciano Floridi, 2011: The Philosophy of Information, Oxford University Press, 2011. James Gleick, 2011: The Information: A History, A Theory, A Flood, Pantheon Books, 2011. Zhandong Liu, Santosh S Venkatesh and Carlo C Maley, 2008: Sequence space coverage, entropy of genomes and the potential to detect non human DNA in human samples, BMC Genomics 2008, 9:509 David Luenberger, 2006: Information Science, Princeton University Press, 2006. David J.C. MacKay, 2003: Information Theory, Inference, and Learning Algorithms, Cambridge University Press, 2003. Claude Shannon & Warren Weaver, 1949: The Mathematical Theory of Communication, University of Illinois Press, 1949. W. N. Francis and H. Kucera: Brown University Standard Corpus of Present Day American English, Brown University, 1967. 37

38

Additive property.,, log log 39