Introduction to Information Theory. Part 4

Similar documents
Introduction to Information Theory. Part 3

Introduction to Information Theory. Part 2

Digital Communications III (ECE 154C) Introduction to Coding and Information Theory

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy


INTRODUCTION TO INFORMATION THEORY

EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY

Shannon's Theory of Communication

CSCI 2570 Introduction to Nanocomputing

(Classical) Information Theory III: Noisy channel coding

6.02 Fall 2012 Lecture #1

ELECTRONICS & COMMUNICATIONS DIGITAL COMMUNICATIONS

Noisy channel communication

Turbo Compression. Andrej Rikovsky, Advisor: Pavol Hanus

UNIT I INFORMATION THEORY. I k log 2

16.36 Communication Systems Engineering

Massachusetts Institute of Technology

Lecture 8: Shannon s Noise Models

4 An Introduction to Channel Coding and Decoding over BSC

Physical Layer and Coding

Chapter 9 Fundamental Limits in Information Theory

Massachusetts Institute of Technology

Revision of Lecture 5

Digital communication system. Shannon s separation principle

Entropies & Information Theory

One Lesson of Information Theory

1 Background on Information Theory

MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK UNIT V PART-A. 1. What is binary symmetric channel (AUC DEC 2006)

Ch 0 Introduction. 0.1 Overview of Information Theory and Coding

Information Theory and Coding Techniques: Chapter 1.1. What is Information Theory? Why you should take this course?

1 Ex. 1 Verify that the function H(p 1,..., p n ) = k p k log 2 p k satisfies all 8 axioms on H.

ECE Information theory Final (Fall 2008)

MAHALAKSHMI ENGINEERING COLLEGE QUESTION BANK. SUBJECT CODE / Name: EC2252 COMMUNICATION THEORY UNIT-V INFORMATION THEORY PART-A

Midterm Exam Information Theory Fall Midterm Exam. Time: 09:10 12:10 11/23, 2016

4F5: Advanced Communications and Coding Handout 2: The Typical Set, Compression, Mutual Information

ECE Information theory Final

Lecture 12. Block Diagram

Introduction to Information Theory. By Prof. S.J. Soni Asst. Professor, CE Department, SPCE, Visnagar

SIDDHARTH GROUP OF INSTITUTIONS :: PUTTUR Siddharth Nagar, Narayanavanam Road UNIT I

CSE468 Information Conflict

3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions

Introduction to Information Theory

Belief-Propagation Decoding of LDPC Codes

Quantitative Biology Lecture 3

Exercise 1. = P(y a 1)P(a 1 )

Information Theory (Information Theory by J. V. Stone, 2015)

An introduction to basic information theory. Hampus Wessman

6.02 Fall 2011 Lecture #9

Information Theory - Entropy. Figure 3

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Dept. of Linguistics, Indiana University Fall 2015

Roll No. :... Invigilator's Signature :.. CS/B.TECH(ECE)/SEM-7/EC-703/ CODING & INFORMATION THEORY. Time Allotted : 3 Hours Full Marks : 70

Shannon s Noisy-Channel Coding Theorem

ELEMENTS O F INFORMATION THEORY

Channel capacity. Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5.

9. Distance measures. 9.1 Classical information measures. Head Tail. How similar/close are two probability distributions? Trace distance.

LECTURE 15. Last time: Feedback channel: setting up the problem. Lecture outline. Joint source and channel coding theorem

Information Theory and Coding Techniques

ECS 332: Principles of Communications 2012/1. HW 4 Due: Sep 7

Capacity of a channel Shannon s second theorem. Information Theory 1/33

Lecture 7 September 24

Block 2: Introduction to Information Theory

Information in Biology

CS4800: Algorithms & Data Jonathan Ullman

Lecture 1. Introduction

Coding, Information Theory (and Advanced Modulation) Prof. Jay Weitzen Ball 411

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

Module 2 LOSSLESS IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Non-binary Distributed Arithmetic Coding

Information in Biology

Image and Multidimensional Signal Processing

Compression and Coding

transmission and coding of information problem list José Luis Ruiz july 2018

Principles of Communications

channel of communication noise Each codeword has length 2, and all digits are either 0 or 1. Such codes are called Binary Codes.

Chapter 2: Source coding

Communication Theory II

Welcome to Comp 411! 2) Course Objectives. 1) Course Mechanics. 3) Information. I thought this course was called Computer Organization

Digital Image Processing Lectures 25 & 26

Lecture 4 Noisy Channel Coding

9 THEORY OF CODES. 9.0 Introduction. 9.1 Noise

Chaos, Complexity, and Inference (36-462)

Notes 3: Stochastic channels and noisy coding theorem bound. 1 Model of information communication and noisy channel

LDPC Codes. Slides originally from I. Land p.1

Classical Information Theory Notes from the lectures by prof Suhov Trieste - june 2006

ECEN 655: Advanced Channel Coding

Lecture 11: Information theory THURSDAY, FEBRUARY 21, 2019

Entropy as a measure of surprise

Multimedia. Multimedia Data Compression (Lossless Compression Algorithms)

Introduction to Wireless & Mobile Systems. Chapter 4. Channel Coding and Error Control Cengage Learning Engineering. All Rights Reserved.

Lecture 2: August 31

3F1 Information Theory, Lecture 3

Lecture 9 Polar Coding

Coding for Discrete Source

Information and Entropy

Massachusetts Institute of Technology. Solution to Problem 1: The Registrar s Worst Nightmare

Infor mation Theor y. Outline

exercise in the previous class (1)

Module 1. Introduction to Digital Communications and Information Theory. Version 2 ECE IIT, Kharagpur

Part I. Entropy. Information Theory and Networks. Section 1. Entropy: definitions. Lecture 5: Entropy

Transcription:

Introduction to Information Theory Part 4

A General Communication System CHANNEL Information Source Transmitter Channel Receiver Destination 10/2/2012 2

Information Channel Input X Channel Output Y 10/2/2012 3

Information Channel Input X Cholesterol Levels Channel Output Y Condition of Arteries 10/2/2012 4

Information Channel Input X Symptoms or Test results Channel Output Y Diagnosis 10/2/2012 5

Information Channel Input X Geological Structure Channel Output Y Presence of oil deposits 10/2/2012 6

Information Channel Input X Opinion Poll Channel Output Y Next President 10/2/2012 7

Perfect Communication (Discrete Noiseless Channel) X Transmitted Symbol 0 1 0 1 Y Received Symbol 10/2/2012 8

NOISE 10/2/2012 9

Motivating Noise X Transmitted Symbol 0 1 0 1 Y Received Symbol 10/2/2012 10

Motivating Noise f = 0.1, n = ~10,000 0 f 1 f 0 1 f 1 f 1 10/2/2012 11

Motivating Noise Message: $5213.75 Received: $5293.75 1. Detect that an error has occurred. 2. Correct the error. 3. Watch out for the overhead. 10/2/2012 12

Error Detection by Repetition In the presence of 20% noise Message : $ 5 2 1 3. 7 5 Transmission 1: $ 5 2 9 3. 7 5 Transmission 2: $ 5 2 1 3. 7 5 Transmission 3: $ 5 2 1 3. 1 1 Transmission 4: $ 5 4 4 3. 7 5 Transmission 5: $ 7 2 1 8. 7 5 There is no way of knowing where the errors are. 10/2/2012 13

Error Detection by Repetition In the presence of 20% noise Message : $ 5 2 1 3. 7 5 Transmission 1: $ 5 2 9 3. 7 5 Transmission 2: $ 5 2 1 3. 7 5 Transmission 3: $ 5 2 1 3. 1 1 Transmission 4: $ 5 4 4 3. 7 5 Transmission 5: $ 7 2 1 8. 7 5 Most common: $ 5 2 1 3. 7 5 1. Guesswork is involved. 2. There is overhead. 10/2/2012 14

Error Detection by Repetition In the presence of 50% noise Message : $ 5 2 1 3. 7 5 Repeat 1000 times! 1. Guesswork is involved. But it will almost never be wrong! 2. There is overhead. A LOT of it! 10/2/2012 15

Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol 0 1 1 1 0 1 Received Symbol 10/2/2012 16

Binary Symmetric Channel (BSC) (Discrete Memoryless Channel) Transmitted Symbol 0 1 1 1 0 1 Received Symbol Defined by a set of conditional probabilities (aka transitional probabilities) The probability of occurring at the output when is the input to the channel. 10/2/2012 17

A General Discrete Channel p p p p input symbols transition probabilities output symbols 10/2/2012 18

Channel With Internal Structure 0 1 0 1 0 0 1 1 0 1 1 1 1 1 1 1 1 1 10/2/2012 19

The Weather Channel SUNNY 0.5 p(hot SUNNY)=0.5 HOT p(warm SUNNY)=0.5 CLOUDY 0.25 p(warm CLOUDY)=0.5 p(cold CLOUDY)=0.5 WARM RAINY 0.25 p(warm RAINY)=0.5 p(cold RAINY)=0.5 COLD 10/2/2012 20

Entropy, random variables with entropy and Conditional Entropy: Average entropy in, given knowledge of., where, Joint Entropy:, Entropy of the pair, 10/2/2012 21

The Weather Channel SUNNY 0.5 p(hot SUNNY)=0.5 HOT p(warm SUNNY)=0.5 CLOUDY 0.25 p(warm CLOUDY)=0.5 p(cold CLOUDY)=0.5 WARM RAINY 0.25 p(warm RAINY)=0.5 p(cold RAINY)=0.5 COLD Q. What is the Entropy,? 10/2/2012 22

The Weather Channel SUNNY 0.5 p(hot SUNNY)=0.5 HOT p(warm SUNNY)=0.5 CLOUDY 0.25 p(warm CLOUDY)=0.5 p(cold CLOUDY)=0.5 WARM RAINY 0.25 p(warm RAINY)=0.5 p(cold RAINY)=0.5 COLD Q. What is the Entropy,? 0.5log 2 0.25 log 4 0.25 log 4 0.50.50.5 1.5 bits 10/2/2012 23

Example Entropy of a toss of die, is log 6 2.59 If outcome is HIGH (either 5 or 6): 21 If outcome is LOW (either 1, 2, 3, or 4): 42 Conditional Entropy: 1 3 2 2 3 4 5 3 1.67 10/2/2012 24

Example Entropy of a toss of die, is log 6 2.59 If outcome is HIGH (either 5 or 6): 21 If outcome is LOW (either 1, 2, 3, or 4): 42 Entropy Reduction: Entropy of a variable is, on average, never increased by knowledge of another variable. Conditional Entropy: 1 3 2 2 3 4 5 3 1.67 10/2/2012 25

The Weather Channel SUNNY 0.5 p(hot SUNNY)=0.5 HOT p(warm SUNNY)=0.5 CLOUDY 0.25 p(warm CLOUDY)=0.5 p(cold CLOUDY)=0.5 WARM RAINY 0.25 p(warm RAINY)=0.5 p(cold RAINY)=0.5 COLD Q. What is the Entropy,? 1 log log 1 0.5 log 2 0.5log 2 1 1 10/2/2012 26

The Weather Channel SUNNY 0.5 p(hot SUNNY)=0.5 HOT p(warm SUNNY)=0.5 CLOUDY 0.25 p(warm CLOUDY)=0.5 p(cold CLOUDY)=0.5 WARM RAINY 0.25 p(warm RAINY)=0.5 p(cold RAINY)=0.5 COLD Q. What is the Entropy,? 0.5 0.5 0.25 0.5 0.5 0.25 0.5 025 0.5 0.5 0.25 0.50.25 0.50.25 1.5 10/2/2012 27

Mutual Information The mutual information of a random variable given the random variable is It is the information about transmitted by. 10/2/2012 28

Mutual Information: Properties is symmetric in and, 10/2/2012 29

The Weather Channel SUNNY 0.5 p(hot SUNNY)=0.5 HOT p(warm SUNNY)=0.5 CLOUDY 0.25 p(warm CLOUDY)=0.5 p(cold CLOUDY)=0.5 WARM RAINY 0.25 p(warm RAINY)=0.5 p(cold RAINY)=0.5 COLD Q. What is the mutual information ;? ; 1.51.00.5 Also ; ; 0.5 10/2/2012 30

Entropy Concepts ;, ; ; ; is symmetric in and ;,, ; 0 ; 10/2/2012 31

Channel Capacity The capacity of a channel is the maximum possible mutual information that can be achieved between input and output by varying the probabilities of the input symbols. If X is the input channel and Y is the output, the capacity C is 10/2/2012 32

Channel Capacity Mutual information about X given Y is the information transmitted by the channel and depends on the probability structure Input probabilities Transition probabilities Output probabilities 10/2/2012 33

Channel Capacity Mutual information about X given Y is the information transmitted by the channel and depends on the probability structure Input probabilities Transition probabilities: fixed by properties of channel Output probabilities: determined by input and transition probabilities 10/2/2012 34

Channel Capacity Mutual information about X given Y is the information transmitted by the channel and depends on the probability structure Input probabilities: can be adjusted by suitable coding Transition probabilities: fixed by properties of channel Output probabilities: determined by input and transition probabilities 10/2/2012 35

Channel Capacity ; Mutual information about X given Y is the information transmitted by the channel and depends on the probability structure Input probabilities: can be adjusted by suitable coding Transition probabilities: fixed by properties of channel Output probabilities: determined by input and transition probabilities That is, input probabilities determine mutual information and can be varied by coding. The maximum mutual information with respect to these input probabilities is the channel capacity. 10/2/2012 36

Shannon s Second Theorem Suppose a discrete channel has capacity C and the source has entropy H If H < C there is a coding scheme such that the source can be transmitted over the channel with an arbitrarily small frequency of error. If H > C, it is not possible to achieve arbitrarily small error frequency. 10/2/2012 37

Detailed Communication Model TRANSMITTER DISCRETE CHANNEL information source data reduction source coding encipherment channel coding modulation distortion (noise) RECEIVER destination data reconstruction source coding decipherment channel decoding demodulation 10/2/2012 38

Error Correcting Codes: Checksum ISBN: 0 691 12418 3 1*0+2*6+3*9+4*1+5*1+6*2+7*4+8*1+9*8 = 168 mod 11 = 3 This is a staircase checksum 10/2/2012 39

Error Correcting Codes Hamming Codes (1950) Linear Codes Low Density Parity Codes (1960) Convolutional Codes Turbo Codes (1993) 10/2/2012 40

MTC: Summary Information Entropy Source Coding Theorem Redundancy Compression Huffman Encoding Lempel Ziv Coding Channel Conditional Entropy Joint Entropy Mutual Information Channel Capacity Shannon s Second Theorem Error Correction Codes 10/2/2012 41

References Eugene Chiu, Jocelyn Lin, Brok Mcferron, Noshirwan Petigara, Satwiksai Seshasai: Mathematical Theory of Claude Shannon: A study of the style and context of his work up to the genesis of information theory. MIT 6.933J / STS.420J The Structure of Engineering Revolutions Luciano Floridi, 2010: Information: A Very Short Introduction, Oxford University Press, 2011. Luciano Floridi, 2011: The Philosophy of Information, Oxford University Press, 2011. James Gleick, 2011: The Information: A History, A Theory, A Flood, Pantheon Books, 2011. Zhandong Liu, Santosh S Venkatesh and Carlo C Maley, 2008: Sequence space coverage, entropy of genomes and the potential to detect non human DNA in human samples, BMC Genomics 2008, 9:509 David Luenberger, 2006: Information Science, Princeton University Press, 2006. David J.C. MacKay, 2003: Information Theory, Inference, and Learning Algorithms, Cambridge University Press, 2003. Claude Shannon & Warren Weaver, 1949: The Mathematical Theory of Communication, University of Illinois Press, 1949. W. N. Francis and H. Kucera: Brown University Standard Corpus of Present Day American English, Brown University, 1967. Edward L. Glaeser: A Tale of Many Cities, New York Times, April 10, 2010. Available at: http://economix.blogs.nytimes.com/2010/04/20/a tale of many cities/ Alan Rimm Kaufman, The Long Tail of Search. Search Engine Land Website, September 18, 2007. Available at: http://searchengineland.com/the long tail of search 12198 42