CSCI 2570 Introduction to Nanocomputing

Similar documents
UNIT I INFORMATION THEORY. I k log 2

Digital Communications III (ECE 154C) Introduction to Coding and Information Theory

Chapter 9 Fundamental Limits in Information Theory

1 Introduction to information theory

An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if. 2 l i. i=1

16.36 Communication Systems Engineering

Bandwidth: Communicate large complex & highly detailed 3D models through lowbandwidth connection (e.g. VRML over the Internet)

Lecture 8: Shannon s Noise Models

Lecture 5: Channel Capacity. Copyright G. Caire (Sample Lectures) 122

EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018

Entropies & Information Theory

COMM901 Source Coding and Compression. Quiz 1

Information and Entropy

Chapter 3 Source Coding. 3.1 An Introduction to Source Coding 3.2 Optimal Source Codes 3.3 Shannon-Fano Code 3.4 Huffman Code

Communications Theory and Engineering

Chapter 5: Data Compression

3F1 Information Theory, Lecture 3

Coding for Discrete Source

1 Ex. 1 Verify that the function H(p 1,..., p n ) = k p k log 2 p k satisfies all 8 axioms on H.

Exercise 1. = P(y a 1)P(a 1 )

Digital communication system. Shannon s separation principle

2018/5/3. YU Xiangyu

Multimedia Communications. Mathematical Preliminaries for Lossless Compression

3F1 Information Theory, Lecture 3

(Classical) Information Theory III: Noisy channel coding

Capacity of a channel Shannon s second theorem. Information Theory 1/33

Run-length & Entropy Coding. Redundancy Removal. Sampling. Quantization. Perform inverse operations at the receiver EEE

Delay, feedback, and the price of ignorance

Entropy and Ergodic Theory Lecture 3: The meaning of entropy in information theory

4. Quantization and Data Compression. ECE 302 Spring 2012 Purdue University, School of ECE Prof. Ilya Pollak

An introduction to basic information theory. Hampus Wessman

Information Theory and Statistics Lecture 2: Source coding

Lecture 3 : Algorithms for source coding. September 30, 2016

Physical Layer and Coding

Introduction to Information Theory. By Prof. S.J. Soni Asst. Professor, CE Department, SPCE, Visnagar

Coding theory: Applications

Chapter 2: Source coding

Information Theory with Applications, Math6397 Lecture Notes from September 30, 2014 taken by Ilknur Telkes

Lecture 6: Expander Codes

Information Theory CHAPTER. 5.1 Introduction. 5.2 Entropy

Source Coding: Part I of Fundamentals of Source and Video Coding

Electrical and Information Technology. Information Theory. Problems and Solutions. Contents. Problems... 1 Solutions...7

Autumn Coping with NP-completeness (Conclusion) Introduction to Data Compression

Information Theory and Coding Techniques

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Shannon s noisy-channel theorem

Dept. of Linguistics, Indiana University Fall 2015


Intro to Information Theory

Chapter 7. Error Control Coding. 7.1 Historical background. Mikael Olofsson 2005

Basic Principles of Lossless Coding. Universal Lossless coding. Lempel-Ziv Coding. 2. Exploit dependences between successive symbols.

A Mathematical Theory of Communication

3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions

MARKOV CHAINS A finite state Markov chain is a sequence of discrete cv s from a finite alphabet where is a pmf on and for

Lecture 1: Shannon s Theorem

Compression and Coding

List Decoding of Reed Solomon Codes

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

6.02 Fall 2012 Lecture #1

Entropy as a measure of surprise

ELEC 515 Information Theory. Distortionless Source Coding

MATH3302. Coding and Cryptography. Coding Theory

Source Coding. Master Universitario en Ingeniería de Telecomunicación. I. Santamaría Universidad de Cantabria

Lecture 18: Shanon s Channel Coding Theorem. Lecture 18: Shanon s Channel Coding Theorem

Coding of memoryless sources 1/35

Lecture 14 February 28

ELEMENT OF INFORMATION THEORY

Chapter 2 Date Compression: Source Coding. 2.1 An Introduction to Source Coding 2.2 Optimal Source Codes 2.3 Huffman Code

Optimum Soft Decision Decoding of Linear Block Codes

SIGNAL COMPRESSION Lecture Shannon-Fano-Elias Codes and Arithmetic Coding

Lecture 4 Noisy Channel Coding

10-704: Information Processing and Learning Fall Lecture 9: Sept 28

EE5139R: Problem Set 4 Assigned: 31/08/16, Due: 07/09/16

LECTURE 15. Last time: Feedback channel: setting up the problem. Lecture outline. Joint source and channel coding theorem

exercise in the previous class (1)

Binary Convolutional Codes

4 An Introduction to Channel Coding and Decoding over BSC

Lecture 18: Gaussian Channel

Lecture 4 Channel Coding

Uncertainity, Information, and Entropy

Ch 0 Introduction. 0.1 Overview of Information Theory and Coding

Notes 3: Stochastic channels and noisy coding theorem bound. 1 Model of information communication and noisy channel

Shannon's Theory of Communication

Introduction to Wireless & Mobile Systems. Chapter 4. Channel Coding and Error Control Cengage Learning Engineering. All Rights Reserved.

Roll No. :... Invigilator's Signature :.. CS/B.TECH(ECE)/SEM-7/EC-703/ CODING & INFORMATION THEORY. Time Allotted : 3 Hours Full Marks : 70

Module 1. Introduction to Digital Communications and Information Theory. Version 2 ECE IIT, Kharagpur

(each row defines a probability distribution). Given n-strings x X n, y Y n we can use the absence of memory in the channel to compute

Homework Set #2 Data Compression, Huffman code and AEP

Principles of Communications

National University of Singapore Department of Electrical & Computer Engineering. Examination for

A One-to-One Code and Its Anti-Redundancy

18.2 Continuous Alphabet (discrete-time, memoryless) Channel

MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK UNIT V PART-A. 1. What is binary symmetric channel (AUC DEC 2006)

Block 2: Introduction to Information Theory

Lecture 11: Quantum Information III - Source Coding

Noisy channel communication

EE5585 Data Compression January 29, Lecture 3. x X x X. 2 l(x) 1 (1)

Coding Theory and Applications. Linear Codes. Enes Pasalic University of Primorska Koper, 2013

EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY

channel of communication noise Each codeword has length 2, and all digits are either 0 or 1. Such codes are called Binary Codes.

Transcription:

CSCI 2570 Introduction to Nanocomputing Information Theory John E Savage

What is Information Theory Introduced by Claude Shannon. See Wikipedia Two foci: a) data compression and b) reliable communication through noisy channels. Data compression important today for storage and transmission, e.g., audio & video (JPEG). Reliable communication used in memories, CDs, Internet, and deep-space probes. Lect 09 Information Theory CSCI 2570 @John E Savage 2

Source Models Memoryless sources generate successive outcomes that are independent and identically distributed. Source (S, p) has outcomes S = {1,2,, n} that occur with probabilities p = {p 1, p 2,, p n } E.g. Binary Source: S = {H,T}, p H =1-p T Lect 09 Information Theory CSCI 2570 @John E Savage 3

Entropy A Measure of Information Entropy of source (S,p) in bits is Binary Source: H(S) = - p H log p H (1- p H )log (1- p H ) The larger the entropy, the less predictable is the source output and the more information is produced by seeing it. If base two logarithms used, entropy measured in bits (binary digits). Lect 09 Information Theory CSCI 2570 @John E Savage 4

What are Codes? A code is a set of words (codewords). Source codes compress data probabilistically E.g. Outputs, probabilities and codes: P(a) =.5, P(b) =.25, P(c) =.125, P(d) =.125 w(a) =1, w(b) = 01, w(c) = 001, w(d) = 000 Channel codes add redundancy for error correction and detection purposes. 0 000, 1 111; decide by majority rule Codes can be used for detection or correction. d 0 0 1 0 c 1 b 1 a Lect 09 Information Theory CSCI 2570 @John E Savage 5

Source Coding Prefix condition: no codeword is a prefix for another. Needed to decode a source code. A source with n i.i.d. outputs and entropy H(S) can be compressed to a string of length nh(s) for large n. Huffman coding algorithm gives the most efficient prefix source encoding. For binary source code, combine two least probable outcomes and give them both the same prefix. Repeat using the prefix as a new outcome with probability equal to the sum of the two least probable outcomes. Algorithm was used on previous page. Lect 09 Information Theory CSCI 2570 @John E Savage 6

Source Coding Theorem Theorem Let codeword a i over alphabet of b symbols encode ith output where s i = a i. Let p(a i ) be probability of a i. Let E(X) = i s i p(a i ) be average codeword length. Let be the source entropy. Then, Lect 09 Information Theory CSCI 2570 @John E Savage 7

Source Coding Theorem Let where C such that Using log x x-1, we have This implies Using we have Lect 09 Information Theory CSCI 2570 @John E Savage 8

Source Coding Theorem Let codewords {a i }, s i = a i, satisfy the prefix condition. Theorem Lengths {s i } satisfy Kraft s Inequality and and for any {s i } satisfying Kraft s Inequaliy, a prefix code can be constructed for them. Lect 09 Information Theory CSCI 2570 @John E Savage 9

Source Coding Theorem Proof Consider complete tree on b letters of depth s n =max i s i. If A i are leaves of the complete tree that are leaves of a i, Since the number of descendants in the complete tree is exactly and the A i are disjoint, Kraft s Inequality follows. Lect 09 Information Theory CSCI 2570 @John E Savage 10

Source Coding Theorem Proof(cont.) Let s 1 s 2 s n. To construct a prefix code, assign codeword to nth word, w(n), which is the set of labels to a vertex at depth s n. Assign a codeword to the (n-1)st word, w(n-1), by picking a vertex at depth s n-1 and deleting all of its leaves in the complete tree. Continue in this fashion. The fact that Kraft s Inequality is satisfied ensures that this process can go to completion. Lect 09 Information Theory CSCI 2570 @John E Savage 11

Discrete Memoryless Channels Inputs are discrete; noise on successive transmissions is i.i.d. Codeword + Received Word s Noise e r = s e Memoryless channels have a capacity, C, a maximum rate at which a source can transmit reliably through the channel, as we shall see. Lect 09 Information Theory CSCI 2570 @John E Savage 12

Codes A code is a set of words (codewords). Block codes and convolutional codes (n,k,d) q block codes k inputs over alphabet of size q are encoded into codewords of length n over the same alphabet. The minimum Hamming distance (no. differences) between two codewords is d. k = message length n = block length R = k/n = rate d = minimum distance q = alphabet size Lect 09 Information Theory CSCI 2570 @John E Savage 13

Binary Symmetric Channel q = 1-p 0 0 p 1 p q = 1-p 1 r = s e. Error vector e identifies errors. Average number of errors in n transmissions is np. Standard deviation is σ = (npq) 1/2. If codewords are more than 2(np + tσ)+1 bits apart, very likely can decode correctly. Lect 09 Information Theory CSCI 2570 @John E Savage 14

Sphere Packing Argument Likely error vectors form sphere around each codeword. If the spheres are disjoint, the probability of decoding the received word correctly will be high. Lect 09 Information Theory CSCI 2570 @John E Savage 15

Memoryless Channel Coding Theorem There exists an infinite family of codes of rate R < C such that nth code achieves a prob. Of error P(E) satisfying where E(R) > 0 for R < C All codes with rate R > C require P(E) > ε > 0. Capacity of BSC C = 1 H(p) = 1 + p log p + (1-p) log (1-p). Lect 09 Information Theory CSCI 2570 @John E Savage 16

The Hamming Code - Example Encode b = (b 0, b 1, b 2, b 3 ) as bg where G is the generator matrix. This is a (7,4,3) 2 code. Why is d = 3? Compare b 1 G and b 2 G where b 1 b 2. Note that b 1 G b 2 G (term-by-term XOR) is equivalent to b 3 G where b 3 = b 1 b 2. Lect 09 Information Theory CSCI 2570 @John E Savage 17

Other Methods of Reliable Communication Automatic Repeat Request (ARQ) The receiver checks to see if the received is a codeword. If not, it requests retransmission of the message. This method can detect d-1 errors when an (n,k,d) block code is used. Requires buffering of data, which may result in loss of data. Lect 09 Information Theory CSCI 2570 @John E Savage 18