Chapter 9 Fundamental Limits in Information Theory

Similar documents
Information Theory CHAPTER. 5.1 Introduction. 5.2 Entropy

Chapter 3 Source Coding. 3.1 An Introduction to Source Coding 3.2 Optimal Source Codes 3.3 Shannon-Fano Code 3.4 Huffman Code

CSCI 2570 Introduction to Nanocomputing

Chapter 2 Date Compression: Source Coding. 2.1 An Introduction to Source Coding 2.2 Optimal Source Codes 2.3 Huffman Code

UNIT I INFORMATION THEORY. I k log 2


Digital Communications III (ECE 154C) Introduction to Coding and Information Theory

Multimedia Communications. Mathematical Preliminaries for Lossless Compression

MAHALAKSHMI ENGINEERING COLLEGE-TRICHY QUESTION BANK UNIT V PART-A. 1. What is binary symmetric channel (AUC DEC 2006)

EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY

Coding for Discrete Source

MAHALAKSHMI ENGINEERING COLLEGE QUESTION BANK. SUBJECT CODE / Name: EC2252 COMMUNICATION THEORY UNIT-V INFORMATION THEORY PART-A

Chapter 2: Source coding

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

Channel capacity. Outline : 1. Source entropy 2. Discrete memoryless channel 3. Mutual information 4. Channel capacity 5.

An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if. 2 l i. i=1

Basic Principles of Video Coding

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy

Run-length & Entropy Coding. Redundancy Removal. Sampling. Quantization. Perform inverse operations at the receiver EEE

18.2 Continuous Alphabet (discrete-time, memoryless) Channel

Compression and Coding

MARKOV CHAINS A finite state Markov chain is a sequence of discrete cv s from a finite alphabet where is a pmf on and for

3F1 Information Theory, Lecture 3

Block 2: Introduction to Information Theory

Revision of Lecture 5

Lecture 4 Noisy Channel Coding

Source Coding. Master Universitario en Ingeniería de Telecomunicación. I. Santamaría Universidad de Cantabria

Digital communication system. Shannon s separation principle

Principles of Communications

1 Ex. 1 Verify that the function H(p 1,..., p n ) = k p k log 2 p k satisfies all 8 axioms on H.

Entropies & Information Theory

1 Introduction to information theory

EE5139R: Problem Set 4 Assigned: 31/08/16, Due: 07/09/16

An introduction to basic information theory. Hampus Wessman

Lecture 18: Gaussian Channel

4F5: Advanced Communications and Coding Handout 2: The Typical Set, Compression, Mutual Information

ELECTRONICS & COMMUNICATIONS DIGITAL COMMUNICATIONS

Lecture 5 Channel Coding over Continuous Channels

Lecture 2. Capacity of the Gaussian channel

(Classical) Information Theory III: Noisy channel coding

lossless, optimal compressor

Lecture 4 Channel Coding

Appendix B Information theory from first principles

Shannon s noisy-channel theorem

ECE Information theory Final

Information Theory - Entropy. Figure 3

COMM901 Source Coding and Compression. Quiz 1

ECE Information theory Final (Fall 2008)

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Dept. of Linguistics, Indiana University Fall 2015

Lecture 22: Final Review

3F1 Information Theory, Lecture 3

Bandwidth: Communicate large complex & highly detailed 3D models through lowbandwidth connection (e.g. VRML over the Internet)

Introduction to Information Theory. By Prof. S.J. Soni Asst. Professor, CE Department, SPCE, Visnagar

Ch 0 Introduction. 0.1 Overview of Information Theory and Coding

PART III. Outline. Codes and Cryptography. Sources. Optimal Codes (I) Jorge L. Villar. MAMME, Fall 2015

Lecture 5: Channel Capacity. Copyright G. Caire (Sample Lectures) 122

ELEC546 Review of Information Theory

10-704: Information Processing and Learning Fall Lecture 10: Oct 3

Chapter 5: Data Compression

Lower Bounds on the Graphical Complexity of Finite-Length LDPC Codes

Information and Entropy

X 1 : X Table 1: Y = X X 2

Lecture 8: Shannon s Noise Models

Lecture 3. Mathematical methods in communication I. REMINDER. A. Convex Set. A set R is a convex set iff, x 1,x 2 R, θ, 0 θ 1, θx 1 + θx 2 R, (1)

Entropy as a measure of surprise

Lecture 1: Shannon s Theorem

Lecture 14 February 28

Communication Theory II

Massachusetts Institute of Technology

Introduction to information theory and coding

Lecture 1. Introduction

ELEMENT OF INFORMATION THEORY

Noisy channel communication

Information Theory and Coding Techniques

Noisy-Channel Coding

BASICS OF COMPRESSION THEORY

3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions

Lecture 2: August 31

Module 1. Introduction to Digital Communications and Information Theory. Version 2 ECE IIT, Kharagpur

Lecture 6: Gaussian Channels. Copyright G. Caire (Sample Lectures) 157

Exercises with solutions (Set B)

Lecture 4 : Adaptive source coding algorithms

Source Coding: Part I of Fundamentals of Source and Video Coding

Gaussian channel. Information theory 2013, lecture 6. Jens Sjölund. 8 May Jens Sjölund (IMT, LiU) Gaussian channel 1 / 26

ITCT Lecture IV.3: Markov Processes and Sources with Memory

Lecture Notes on Digital Transmission Source and Channel Coding. José Manuel Bioucas Dias

Revision of Lecture 4

Basic Principles of Lossless Coding. Universal Lossless coding. Lempel-Ziv Coding. 2. Exploit dependences between successive symbols.

Physical Layer and Coding

Coding of memoryless sources 1/35

Notes 3: Stochastic channels and noisy coding theorem bound. 1 Model of information communication and noisy channel

10-704: Information Processing and Learning Fall Lecture 9: Sept 28

Information Theory, Statistics, and Decision Trees

Lecture 6 Channel Coding over Continuous Channels

Energy State Amplification in an Energy Harvesting Communication System

Information Theory and Statistics Lecture 2: Source coding

2018/5/3. YU Xiangyu

Homework Set #2 Data Compression, Huffman code and AEP

Electrical and Information Technology. Information Theory. Problems and Solutions. Contents. Problems... 1 Solutions...7

Transcription:

Chapter 9 Fundamental Limits in Information Theory Information Theory is the fundamental theory behind information manipulation, including data compression and data transmission. 9.1 Introduction o For years, researchers wish to seek answers to some fundamental questions on information manipulation: n What is the irreducible complexity below which an informational signal cannot be compressed? Entropy. n What is the ultimate transmission rate for reliable communication over a noisy channel? Capacity. n A more striking result o If the entropy of the source is less than the capacity of the channel, then (asymptotic) error-free (or arbitrarily small error) communication over the channel can be achieved. Po-Ning Chen@cm.nctu Chapter 9-2 1

9.2 Uncertainty, information, and entropy o Uncertainty = Information n When one gains information, he/she shall lose uncertainty. n Example. Suppose a discrete random variable S takes value from with probability How much information we gain if we observe the outcome of S (if we already know the statistics of S)? Po-Ning Chen@cm.nctu Chapter 9-3 9.2 Uncertainty, information, and entropy n Case 1: p 0 = 1. o Since we know that s 0 will be observed before we observe it, no uncertainty loses (namely, no information gains) for this observation. n Case 2: p k < 1 for every k. o Can we quantitatively measure the information amount we gain after observing S? Po-Ning Chen@cm.nctu Chapter 9-4 2

9.2 Uncertainty, information, and entropy o Axioms for information (uncertainty) measure n Monotonicity in event probability o If an event is less likely to happen, it is more uncertain that the event would happen. Therefore, its degree of uncertainty should be higher. n Additivity o If S and T are two independent random variables, then the uncertainty loss due to the observations of both S and T should be equal to the sum of the uncertainty loss due to the observation of S and the uncertainty loss due to the observation of T. Po-Ning Chen@cm.nctu Chapter 9-5 9.2 Uncertainty, information, and entropy n Continuity o A small adjustment in event probability should induce a small change in event uncertainty. Po-Ning Chen@cm.nctu Chapter 9-6 3

9.2 Uncertainty, information, and entropy o Mathematically, the three axioms are transformed to: o It can be proved that the only function that satisfies the above three conditions is: Po-Ning Chen@cm.nctu Chapter 9-7 9.2 Uncertainty, information, and entropy o Hence, for random variable S, each outcome, when it does happen, will respectively give the observer information o In expectation value, the random variable S will give the observer information amount Po-Ning Chen@cm.nctu Chapter 9-8 4

9.2 Uncertainty, information, and entropy o When p 0 = ½, s 0 is either observed or not observed with equal probable. Hence, intuitively, one should learn one bit of information after this observation. o As a result, take C = 1/log(2). o This is named entropy of S. Po-Ning Chen@cm.nctu Chapter 9-9 9.2 Uncertainty, information, and entropy o Relation of entropy and data compression n For a single random variable with K possible outcomes, a straightforward representation is to use For example, K = 8. Then, use 3 bits to represent each outcome. n For a sequence of observations on independent and identically distributed (i.i.d.) S 1, S 2, S 3,, we will use to represent the sequence. Can we reduce this number? Po-Ning Chen@cm.nctu Chapter 9-10 5

9.2 Uncertainty, information, and entropy o (Section 9.3 Source-Coding Theorem) In 1948, Shannon proved that n (Converse) the minimum average number of bits per observation to losslessly and uniquely-decodably represent an i.i.d. sequence S 1, S 2, S 3,, is lower-bounded by the entropy H(S). n (Achievability) the minimum average number of bits per observation to losslessly and uniquely-decodably represent an i.i.d. sequence S 1, S 2, S 3,, can be made arbitrarily close to H(S). o Thus, the information quantitative measure finds its theoretical footing! Po-Ning Chen@cm.nctu Chapter 9-11 o Unique decodability is an essential premise for Shannon s source coding theorem n Unique decodability = Concatenation of codewords (without punctuation mechanism) can be uniquely decodable. Po-Ning Chen@cm.nctu Chapter 9-12 6

o A uniquely decodable code must satisfy the Kraft-McMillan inequality (or Kraft s inequality). A violation to the Kraft-McMillan ineq. Po-Ning Chen@cm.nctu Chapter 9-13 o Arbitrarily close to entropy of average codeword length for a sequence of i.i.d. random variables Po-Ning Chen@cm.nctu Chapter 9-14 7

By increasing the number of letters encoded at a time (at the price of memory consumption), one can make the average codeword length approaching the source entropy. Codeword length Number of codewords Po-Ning Chen@cm.nctu Chapter 9-15 9.2 Uncertainty, information, and entropy o Some properties regarding entropy Po-Ning Chen@cm.nctu Chapter 9-16 8

Po-Ning Chen@cm.nctu Chapter 9-17 9.2 Uncertainty, information, and entropy o Definition of discrete memoryless source (DMS) S 1, S 2, S 3, n Discrete = The alphabet of each S i is discrete. n Memoryless = Independent among S i (from the text) o Memoryless = Identically distributed in addition to independent (otherwise we need to check the time instance i in order to identify the statistics) Po-Ning Chen@cm.nctu Chapter 9-18 9

9.4 Data compaction o Data compaction or lossless data compression n To remove the redundancy with no loss of information o A sample uniquely decodable code Prefix code n Prefix condition: No codeword is the prefix of other codeword. n A prefix code satisfies the prefix condition. Po-Ning Chen@cm.nctu Chapter 9-19 9.4 Data compaction o A prefix code is uniquely decodable, thereby satisfying the Kraft-McMillan inequality. o Converse is not necessary true. n A uniquely decodable code is not necessarily a prefix code. For example, a uniquely decodable non-prefix code Codeword of A = 0 Codeword of B = 01 Codeword of C = 011 Codeword of D = 0111 Po-Ning Chen@cm.nctu Chapter 9-20 10

9.4 Data compaction The codewords of this prefix code are 00, 01, 10, 110, 1110 and 1111. o The codewords of a prefix code, when the code is represented by a code tree, are always located at the leaves. o Conversely, a prefix code can be formed by selecting the binary sequences corresponding to leaves on a code tree. Po-Ning Chen@cm.nctu Chapter 9-21 9.4 Data compaction o Enforced by the prefix condition (or more specifically, enforced by the fact that the codewords are all residing at the leaves of a code tree), the prefix codeword can be instantaneously decoded upon the reception of the last bit. o For this reason, it is also named the instantaneous code. Po-Ning Chen@cm.nctu Chapter 9-22 11

9.4 Data compaction o Due to the tree-leave graphical representation of prefix codes, it can be shown that: o With this property, we can prove that there exists a prefix code whose average codeword length satisfies Po-Ning Chen@cm.nctu Chapter 9-23 Po-Ning Chen@cm.nctu Chapter 9-24 12

With this property, if we compress two symbols at a time, then With this property, if we compress three symbols at a time, then With this property, if we compress n symbols at a time, then As a result, we can make the average codeword length per observation arbitrarily close to the source entropy. Po-Ning Chen@cm.nctu Chapter 9-25 9.4 Data compaction o The optimal code for lossless compression Huffman code Po-Ning Chen@cm.nctu Chapter 9-26 13

The idea behind the proof is to show that any code that violates any of the three conditions cannot be the one with the smallest average codeword length. Po-Ning Chen@cm.nctu Chapter 9-27 Example. Consider a source with alphabet {1, 2, 3, 4, 5, 6} with probability 0.25, 0.25, 0.25, 0.1, 0.1, 0.05, respectively. Step 1: Po-Ning Chen@cm.nctu Chapter 9-28 14

Step 2: Po-Ning Chen@cm.nctu Chapter 9-29 9.4 Data compaction o Variance of average codeword length of Huffman code n When the probability of a newly combined symbol is found to be equal to another probability in the list, the selection of any one of them as the next symbol to be combined will yield the same average codeword length but different variance. n A trick is to avoid using the newly combined symbol in order to minimize the variance. (See the example in the text.) Po-Ning Chen@cm.nctu Chapter 9-30 15

9.4 Data compaction o Lempel-Ziv code An asymptotically optimal (i.e., achieving source entropy) universal code n Huffman coding requires the knowledge of probability of occurrence for each symbol; so, it is not universally good. n Can we design a coding scheme that is universally good (achieving source entropy) for a class of source, such as, for all i.i.d. sources? Po-Ning Chen@cm.nctu Chapter 9-31 Lempel-Ziv coding scheme Chapter 9-32 16

Po-Ning Chen@cm.nctu Chapter 9-33 9.5 Discrete memoryless channels o Discrete memoryless channels (DMCs) n Discrete input and output alphabets n Current output only depends on current input, where the dependence is time invariant. Po-Ning Chen@cm.nctu Chapter 9-34 17

9.5 Discrete memoryless channels o Example. (Memoryless) Binary symmetric channel n The simplest discrete memoryless channel Po-Ning Chen@cm.nctu Chapter 9-35 9.6 Mutual information o Assumptions in transceiving operation n The receiver o Does not know the channel input x o Does not know the channel noise as well as how the channel noise affects x o Do know the channel output y o Do know the distribution of input x o Do know the transition probability of output y given input x Po-Ning Chen@cm.nctu Chapter 9-36 18

9.6 Mutual information o Assumptions in transceiving operation n The receiver o Do know the information content of channel input to be transmitted, i.e., the input entropy H(X). o After observing y, i.e., after transmission, the remaining uncertainty for receiver about the input is H(X Y). Po-Ning Chen@cm.nctu Chapter 9-37 9.6 Mutual information o Conditional entropy H(X Y) Po-Ning Chen@cm.nctu Chapter 9-38 19

9.6 Mutual information o Mutual information Po-Ning Chen@cm.nctu Chapter 9-39 9.6 Mutual information o Properties of mutual information No information (H(X) H(X Y)) can be conveyed from input to output if Y is independent of X. Po-Ning Chen@cm.nctu Chapter 9-40 20

Po-Ning Chen@cm.nctu Chapter 9-41 9.7 Channel capacity o Channel capacity n A new terminology baptized by Shannon n n n n p(y x) is fixed and given by the channel. I(X;Y) is the information that can be conveyed through the channel. p(x) is the way we use the channel. How about we choose the right p(x) to maximize the information conveyed through the channel! Po-Ning Chen@cm.nctu Chapter 9-42 21

9.7 Channel capacity o So, Shannon calculated o Then, he asked himself What is the operational meaning of this quantity? Po-Ning Chen@cm.nctu Chapter 9-43 9.8 Channel coding theorem o Shannon found that Word error rate (WER)= Codeword error rate Po-Ning Chen@cm.nctu Chapter 9-44 22

9.8 Channel coding theorem o Code rate in data transmission n Binary symmetric channel a b a b n WER for one information bit per channel usage (i.e., for an uncoded system) is e. Po-Ning Chen@cm.nctu Chapter 9-45 a b a b n WER for 1/3 information bit per channel usage is given by: n WER is reduced at the price of code rate reduction (i.e., (transmission speed reduction). Po-Ning Chen@cm.nctu Chapter 9-46 23

n How about comparing the WERs of all codes with the same code rate (i.e., the same transmission speed)? For example, R = 1/3. n Shannon proved that under, if P e can be made arbitrarily close to zero, n In terminology, reliable in data transmission means that WER can be made arbitrarily small. n Thus R = 1/3 is a reliable code rate (i.e., a reliable data transmission rate). Po-Ning Chen@cm.nctu Chapter 9-47 9.8 Channel-coding theorem o Note that the left-hand-side (code rate) and the right-handside (channel capacity) should be measured (or calculated) in the same unit. Po-Ning Chen@cm.nctu Chapter 9-48 24

9.9 Differential entropy and mutual information for continuous ensembles o Entropy of a continuous source n Suppose a continuous source X has continuous density f. n Transform the continuous source X into a discrete one X D by quantization with step size D. (Mean-value theorem) Po-Ning Chen@cm.nctu Chapter 9-49 The entropy of X D is therefore given by: Po-Ning Chen@cm.nctu Chapter 9-50 25

o Two observations can be made on the entropy of a continuous source. n Its entropy is infinite (due to the second term ); so, a continuous source contains infinite number of information bits (amount). n The first term may be viewed as the quantization efficiency for the source, and is named the differential entropy h(x). o To uniformly quantize X up to n-bit accuracy requires D = 2 -n. o However, to losslessly express the quantization result (so that the accuracy remains) requires (approximately) Po-Ning Chen@cm.nctu Chapter 9-51 9.9 Differential entropy and mutual information for continuous ensembles o Richness in information content in Gauassian source n Example 9.8 o Let X be a Gaussian random variable with mean µ and variance s 2. o Let Y be a random variable with mean µ and variance s 2. o Then, Po-Ning Chen@cm.nctu Chapter 9-52 26

Po-Ning Chen@cm.nctu Chapter 9-53 We can similarly prove that among all distributions with support in [a, b), the uniform distribution has the largest differential entropy. Po-Ning Chen@cm.nctu Chapter 9-54 27

9.9 Differential entropy and mutual information for continuous ensembles o Mutual information for continuous ensembles Therefore, Po-Ning Chen@cm.nctu Chapter 9-55 9.10 Information capacity theorem o Band-limited, power-limited (also, time-limited) Gaussian channels 1 Sampling at rate 2B samples per second, starting from t = 0 to t = T AWGN -B B f Sampling at rate 2B samples per second, starting from t = 0 to t = T Po-Ning Chen@cm.nctu Chapter 9-56 28

9.10 Information capacity theorem o What is the channel capacity (information bit per channel usage) for band-limited, power-limited (time-limited) Gaussian channels? Answer: o Optimization of mutual information Po-Ning Chen@cm.nctu Chapter 9-57 Po-Ning Chen@cm.nctu Chapter 9-58 29

The upper bound can be achieved by making Y k Gaussian (i.e., by taking X k Gaussian with variance P) Note that Po-Ning Chen@cm.nctu Chapter 9-59 9.10 Information capacity theorem o Sphere packing argument With high probability Although the receiver does not know the transmitted X, it knows that with high probability, the transmitted one will be in the sphere Po-Ning Chen@cm.nctu Chapter 9-60 30

9.10 Information capacity theorem o Hence, if the spheres centered at each (possibly) transmitted X with radius (ns 2 ) 1/2 do not overlap with each other, the decoding error will be small. Po-Ning Chen@cm.nctu Chapter 9-61 9.10 Information capacity theorem o Question: What the maximum number of spheres we can place inside the Y-space is? Po-Ning Chen@cm.nctu Chapter 9-62 31

9.10 Information capacity theorem o Hence, the code rate is given by Po-Ning Chen@cm.nctu Chapter 9-63 9.11 Implications of the information capacity theorem o In order to have arbitrarily small error, it requires: o With P = E b R, the above requirement is equivalent to: o We can then plot the relation between R/B (bits per second per hertz) and E b /N 0. Po-Ning Chen@cm.nctu Chapter 9-64 32

Reliable transmission is not feasible! Reliable transmission is feasible! Po-Ning Chen@cm.nctu Chapter 9-65 Implication 1: implies E b /N 0 must exceed -1.6 db in order to make possible the arbitrarily small error transmission. -1.6 db is named the Shannon limit for an AWGN channel. Implication 2: The lower the E b /N 0 (even if it exceeds -1.6 db), the lower the bandwidth efficiency R/B for reliable transmission. Po-Ning Chen@cm.nctu Chapter 9-66 33

Implication 3: implies that when bandwidth B is very large, the capacity is proportional to P/N 0. Po-Ning Chen@cm.nctu Chapter 9-67 9.11 Implications of the information capacity theorem o Example 9.10 M-ary PCM n Examination of the relation between transmission rate, bandwidth and power. Po-Ning Chen@cm.nctu Chapter 9-68 34

Po-Ning Chen@cm.nctu Chapter 9-69 This implies exactly the same relation between transmission rate, power and bandwidth as Shannon s capacity formula! Po-Ning Chen@cm.nctu Chapter 9-70 35

9.11 Implications of the information capacity theorem o Comparison of bandwidth efficiency n Example 9.11 M-ary PSK and M-ary FSK o From Slide 6-66, the bandwidth efficiency of M-ary PSK is: o From Slide 6-156, the bandwidth efficiency of M-ary FSK is: Po-Ning Chen@cm.nctu Chapter 9-71 M-ary PSK M-ary FSK R/B versus E b /N 0 required for P e = 10-5 Po-Ning Chen@cm.nctu Chapter 9-72 36

9.11 Implications of the information capacity theorem o Remark n By increasing M, M-ary FSK approaches Shannon limit, while M-ary PSK deviates from Shannon limit. Po-Ning Chen@cm.nctu Chapter 9-73 9.11 Implications of the information capacity theorem o Example 9.12 Capacity of binary-input AWGN channel Po-Ning Chen@cm.nctu Chapter 9-74 37

We then plot the minimum E b /N 0 that is required to validate the above inequality for a given code rate R. Po-Ning Chen@cm.nctu Chapter 9-75 0.2 db -1.6 db R Po-Ning Chen@cm.nctu Chapter 9-76 38

9.12 Information capacity of colored noise channel o Band-limited, power-limited color Gaussian channels Sampling at rate 2B samples per second, starting from t = 0 to t = T Color Gaussian 1 f Sampling at rate 2B samples per second, starting from t = 0 to t = T Po-Ning Chen@cm.nctu Chapter 9-77 Treated white in this sub-h(f) Assume that there are N sub-intervals. Po-Ning Chen@cm.nctu Chapter 9-78 39

9.12 Information capacity of colored noise channel o Band-limited, power-limited Gaussian channels 1 Sampling at rate 2B samples per second, starting from t = 0 to t = T AWGN -B B f Sampling at rate 2B samples per second, starting from t = 0 to t = T Po-Ning Chen@cm.nctu Chapter 9-79 Po-Ning Chen@cm.nctu Chapter 9-80 40

Lagrange multipliers technique 0 Po-Ning Chen@cm.nctu Chapter 9-81 Equivalently, Po-Ning Chen@cm.nctu Chapter 9-82 41

9.12 Information capacity of colored noise channel o Water-filling interpretation of information-capacity theorem for a colored noise channel The area of the two shaded bowls is P. Po-Ning Chen@cm.nctu Chapter 9-83 Po-Ning Chen@cm.nctu Chapter 9-84 42

9.12 Information capacity of colored noise channel o This makes Band-limited, power-limited Gaussian channels a special case. 1 -B B f AWGN Po-Ning Chen@cm.nctu Chapter 9-85 Po-Ning Chen@cm.nctu Chapter 9-86 43

1 -B B f Color Gaussian channel In the previous derivation, we assume that H(f) is an ideal lowpass filter with bandwidth B. What if H(f) does not assume the shape of an ideal lowpass filter? Will channel capacity be affected? Po-Ning Chen@cm.nctu Chapter 9-87 1 -B B f Color Gaussian channel Answer: Since we assume that H(f) is known to the system, we can add an equalizer 1/H(f) at the receiver to compensate it. Accordingly, under such perfectly known H(f) assumption, the non-idealshape of H(f) will not affect the channel capacity, namely, channel capacity is completely determined by the power spectrum of the color Gaussian. Po-Ning Chen@cm.nctu Chapter 9-88 44

9.13 Rate distortion function Source with entropy H > C Channel with capacity C Output with unmanageable error Source with entropy H > C Lossy data compression that introduces distortion D Compressed output with reduced entropy H r < C Channel with capacity C Output with arbitrarily small error Po-Ning Chen@cm.nctu Chapter 9-89 9.13 Rate distortion function o Lossy data compression with a fidelity criterion Non-reversible function mapping f Po-Ning Chen@cm.nctu Chapter 9-90 45

Question: What is the smallest data compression rate (number of output bits per input symbol) such that the average distortion is less than D? Answer: Po-Ning Chen@cm.nctu Chapter 9-91 Po-Ning Chen@cm.nctu Chapter 9-92 46

Po-Ning Chen@cm.nctu Chapter 9-93 Po-Ning Chen@cm.nctu Chapter 9-94 47

o Remarks n R(D) reduces to H(X) when D = 0. n As D (the acceptable distortion level) increases, the data compression rate can be reduced to nearly zero. Po-Ning Chen@cm.nctu Chapter 9-95 48