J. E. Cunningham, U. F. Gronemann, T. S. Huang, J. W. Pan, O. J. Tretiak, W. F. Schreiber

Similar documents
In this report we shall attempt to answer this question.

(1) QPR No. 67 PROCESSING AND TRANSMISSION OF INFORMATION' XI.

PROCESSING AND TRANSMISSION OF INFORMATION* XV.

S. Arnstein D. Bridwell A. Bucher Chase D. Falconer L. Greenspan Haccoun M. Heggestad

Introduction to Convolutional Codes, Part 1

PROCESSING AND TRANSMISSION OF INFORMATION*

Roll No. :... Invigilator's Signature :.. CS/B.TECH(ECE)/SEM-7/EC-703/ CODING & INFORMATION THEORY. Time Allotted : 3 Hours Full Marks : 70

P/, t. co P9 ALGEBRAIC DECODING FOR A BINARY ERASURE CHANNEL TECHNICAL REPORT 340 RESEARCH LABORATORY OF ELECTRONICS M. A. EPSTEIN

Polar Coding. Part 1 - Background. Erdal Arıkan. Electrical-Electronics Engineering Department, Bilkent University, Ankara, Turkey

coe'p O hs THRESHOLD DECODING I. 1% JAMES L. MASSEY MASSACHUSETTS INSTITUTE OF TECHNOLOGY TECHNICAL REPORT 410

EE 4TM4: Digital Communications II. Channel Capacity

Chapter 2 Review of Classical Information Theory


Lecture 5: Channel Capacity. Copyright G. Caire (Sample Lectures) 122

LECTURE 10. Last time: Lecture outline

Capacity of a channel Shannon s second theorem. Information Theory 1/33

The sequential decoding metric for detection in sensor networks

A Heuristic Discussion of Probabilistic Decoding*

Channel combining and splitting for cutoff rate improvement

Lecture 4 Channel Coding

Exercise 1. = P(y a 1)P(a 1 )

1 Introduction to information theory

CSCI 2570 Introduction to Nanocomputing

UNIT I INFORMATION THEORY. I k log 2

Approaching Blokh-Zyablov Error Exponent with Linear-Time Encodable/Decodable Codes

One Lesson of Information Theory

Exponential Error Bounds for Block Concatenated Codes with Tail Biting Trellis Inner Codes

1 1 0, g Exercise 1. Generator polynomials of a convolutional code, given in binary form, are g

Multiaccess Channels with State Known to One Encoder: A Case of Degraded Message Sets

Information Theory and Statistics Lecture 2: Source coding

The PPM Poisson Channel: Finite-Length Bounds and Code Design

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Information and Entropy

Dept. of Linguistics, Indiana University Fall 2015

Network coding for multicast relation to compression and generalization of Slepian-Wolf

EE229B - Final Project. Capacity-Approaching Low-Density Parity-Check Codes

PROCESSING AND TRANSMISSION OF INFORMATION* A. A METHOD OF PLANAR PREDICTION FOR CODING PICTURES

SIDDHARTH GROUP OF INSTITUTIONS :: PUTTUR Siddharth Nagar, Narayanavanam Road UNIT I

Information Theory. Lecture 10. Network Information Theory (CT15); a focus on channel capacity results

T h e C S E T I P r o j e c t

Shannon s noisy-channel theorem

Error Exponent Region for Gaussian Broadcast Channels

Generalized Kraft Inequality and Arithmetic Coding

Turbo Codes for Deep-Space Communications

Shannon s Noisy-Channel Coding Theorem

Module 1. Introduction to Digital Communications and Information Theory. Version 2 ECE IIT, Kharagpur

for some error exponent E( R) as a function R,

COMMUNICATION SCIENCES

Lecture 4 Noisy Channel Coding

ECEN 655: Advanced Channel Coding

Evaluation of Error probability for Quantum. Gaussian Channels Using Gallager Functions

(each row defines a probability distribution). Given n-strings x X n, y Y n we can use the absence of memory in the channel to compute

ELEC546 Review of Information Theory

SOFT DECISION FANO DECODING OF BLOCK CODES OVER DISCRETE MEMORYLESS CHANNEL USING TREE DIAGRAM

THIS paper is aimed at designing efficient decoding algorithms

BASICS OF DETECTION AND ESTIMATION THEORY

Mismatched Multi-letter Successive Decoding for the Multiple-Access Channel

LDPC Codes. Intracom Telecom, Peania

Reliable Computation over Multiple-Access Channels

Serially Concatenated Polar Codes

Chapter 9 Fundamental Limits in Information Theory

Lecture 6 I. CHANNEL CODING. X n (m) P Y X

Universal Anytime Codes: An approach to uncertain channels in control

EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018

ELECTRONICS & COMMUNICATIONS DIGITAL COMMUNICATIONS

Variable Length Codes for Degraded Broadcast Channels

Tightened Upper Bounds on the ML Decoding Error Probability of Binary Linear Block Codes and Applications

SIPCom8-1: Information Theory and Coding Linear Binary Codes Ingmar Land

Joint Source-Channel Coding for the Multiple-Access Relay Channel

Maximum Likelihood Decoding of Codes on the Asymmetric Z-channel

Chapter 3 Linear Block Codes

Lecture 18: Shanon s Channel Coding Theorem. Lecture 18: Shanon s Channel Coding Theorem

ScienceDirect. ScienceDirec

Logarithmic and Exponential Equations and Change-of-Base

An Achievable Error Exponent for the Mismatched Multiple-Access Channel

Motivation for Arithmetic Coding

A L A BA M A L A W R E V IE W

INFORMATION PROCESSING ABILITY OF BINARY DETECTORS AND BLOCK DECODERS. Michael A. Lexa and Don H. Johnson

A Simple Converse of Burnashev s Reliability Function

An analysis of the computational complexity of sequential decoding of specific tree codes over Gaussian channels

Practical Polar Code Construction Using Generalised Generator Matrices

Performance Analysis and Code Optimization of Low Density Parity-Check Codes on Rayleigh Fading Channels

3F1 Information Theory, Lecture 3

Digital Communications III (ECE 154C) Introduction to Coding and Information Theory

3F1 Information Theory, Lecture 3

National University of Singapore Department of Electrical & Computer Engineering. Examination for

Research Collection. The relation between block length and reliability for a cascade of AWGN links. Conference Paper. ETH Library

Channel Coding 1. Sportturm (SpT), Room: C3165

Chapter 7: Channel coding:convolutional codes

Exact Probability of Erasure and a Decoding Algorithm for Convolutional Codes on the Binary Erasure Channel

Can Feedback Increase the Capacity of the Energy Harvesting Channel?

Lecture 1: The Multiple Access Channel. Copyright G. Caire 12

Channel Coding I. Exercises SS 2017

Coding for Discrete Source

Solutions to Homework Set #3 Channel and Source coding

LECTURE 13. Last time: Lecture outline

Channel Coding I. Exercises SS 2017

Lecture 7. Union bound for reducing M-ary to binary hypothesis testing

Bounds on Mutual Information for Simple Codes Using Information Combining

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

Transcription:

X. PROCESSNG AND TRANSMSSON OF NFORMATON D. N. Arden E. Arthurs J. B. Dennis M. Eden P. Elias R. M. Fano R. G. Gallager E. M. Hofstetter D. A. Huffman W. F. Schreiber C. E. Shannon J. M. Wozencraft S. Asano G. Cheadle J. E. Cunningham P. M. Ebert H. A. Ernst E. F. Ferretti T. J. Goblick, Jr. U. F. Gronemann F. C. Hennie S. Huang Jelinek Kailath Kleinrock H. Loomis, Jr. L. Massey W. Pan L. Reich G. Roberts J. Tretiak Ziv A. PCTURE PROCESSNG RESEARCH We have now made computer tapes from original pictures and played the tapes back for photography. Operations on the BM 709 computer will begin during the next reporting period. An example of a 1024 X 1024 element picture is shown in Fig. X-1. Fig. X-1. J. E. Cunningham, U. F. Gronemann, T. S. Huang, J. W. Pan, O. J. Tretiak, W. F. Schreiber *This research was supported in part by Purchase Order DDL B-00306 with Lincoln Laboratory, a center for research operated by Massachusetts nstitute of Technology, with the joint support of the U.S. Army, Navy, and Air Force under Air Force Contract AF19(604)-7400. 119

B. NEW SYSTEMATC DECODNG FOR MEMORYLESS CHANNELS n this report a sequential decoding scheme for random convolutional codes, which is to be used for memoryless channels, is described. The average number of computations does not grow exponentially with n; it is upper-bounded by a quantity proportional 2 to n, for all rates below some cutoff rate R (n is the block length). When this decoding scheme is averaged over a suitably defined ensemble of code words it has an average probability of error with an upper bound whose logarithm is -ne(r). E(R) is dependent on the data rate. (E(R) > 0 for rates below channel capacity.) The decoding scheme is different from other effective decoding schemes such as sequen- 1 2 tial decoding and low-density parity-check codes. 2 The lower bound on R comp of the systematic decoding scheme that is presented in this report is the same as the Rcomp of sequential decoding for asymmetric channels. However, in the case of sequential decoding, R is valid only for the incorrect comp subset of code words. The existence of Rcomp for the correct subset has not yet been established., the systematic decoding scheme yields a better bound on the average number of computations for asymmetric channels. (This is not the case when the channel is symmetric, since the modified sequential decoding scheme after Gallager 3 may be used.) S* * Fig. X-2. A convolutional tree code. A convolutional tree code may be viewed topologically as shown in Fig. X-2. The words are all the directed paths from the input node to the output nodes of the tree (there are no closed paths). From all nontrivial intermediate nodes there emerge f directed links, one for each of nodes. Let the number of input symbols per code word be n. Let the number of input symbols per link be d. Then the number of links per word is m = n/d (m is the number of information digits per word). n Fig. X-2, n = 6; m = 2; d= 3; f = 2. Reiffen 4 has shown that the convolutional codes may be generated sequentially. 1. The Decoding Procedure The decoding procedure consists of the following successive operations: Step 1: The a posteriori probability of each one of the links of length d that 120

emerge from the input node to the first k nodes in the tree is computed. The one link that yields the largest a posteriori probability is chosen to represent the corresponding part of the transmitted code word. This detected link connects the input node with one of the f nodes of the next set of nodes (set in Fig. X-3). The same procedure is then repeated with the detected node of set as a starting point., the a posteriori probability of each one of the f links emerging from the d * Fig. X-3. The decoding procedure of step 1. d N T node that was previously detected, is now computed, and a decision is made. This procedure is then repeated again and again until termination (i. e., until the detected path reaches one of the output nodes). A metric of the form D(u, v) = n P(v/u) is then com- L P(v) puted. Here, P(v/u) is the a posteriori probability of the complete detected word, u, and P(v) is the probability of the output symbol v. f D(u, v) is larger than some preset threshold D o, a final decision is made and the detection of the first information digit is completed. f D(u, v) < D o, the computation procedure is then to go to step 2. 2 Step 2: The a posteriori probability of each one of the 1 links of length 2d that emerge from the input node to set (that consist of f2 nodes) is computed (Fig. X-4). The one link that yields the largest a posteriori probability is chosen to represent the corresponding part of the transmitted code word. The same procedure is then repeated with the detected node of set as a starting point, and so on. path reaches one of the output nodes). This procedure is continued until termination (i. e., until the detected 121

D(u, v) is then computed for the detected path and compared with the threshold D. o f D(u, v) > Do, a final decision is made with respect to the first information digit. f Fig. X-4. The decoding procedure of step 2. t X D(u, v) < D O the computation procedure is then to go to step 3. Step 3: This step is similar to step 2, except that the computation length in this case is 3d, and the number of links involved at each stage is. f no termination occurs at step 3, the computation procedure then reaches step 4, and so on. Following the detection of the first information digit, the whole procedure is repeated for the next digit, and so forth. a. The Average Number of Computations per nformation Digit The number of computations that are involved in step 1 is equal to m.. The number of computations in step 2 is equal to (m. 2). The number of computations in step 3 is equal to (m.3 ). n general, the number of computations in step k is Nk = mfk Let Ck be the condition that no termination occur at step k. Step k will be used only if there were no terminations at all in the previous k - 1 steps. the probability of step k being used is P(k) = Pr(C 1,C 2' C2, 3 C4. Ck- ) 122

The average number of computations is given by N = N 1 P(1 ) + N 2 P(2) +... + NkP(k) +... NmP(m) m m = NkP(k) = mfkp(k). (2) k=l k= 1 b. An Upper Bound on the Average Number of Computations P(k) may be bounded by P(k) = Pr(C1, C2 C3 C4 ' Ck-l) < Pr(Ck- 1 ). (3) m N= N Pr(Ck-1) = (m-k+l) k r(ck-l1) (4) k= 1 k= 1 Now let uk be the code word detected at step k, and let u be the transmitted code word. Then Pr(Ckl) = Pr(D(uk, v).< Do) = Pr(D(uk' v) Do;uk=u) + Pr(D(uk, v) 2< Do;Uku) = Pr(D(u, v) < Do;uk=u) +Pr(D(uk, v) - < Do;Uk#u) Pr(Ckl) - Pr[D(u,v) <Do] + Pr[uku] (5) The threshold D 0 is set so as to make Pr[D(u, v)-<do] < e - (R)n, (6) where -E(R) is a function of the rate R. Now, The number of detected links needed to construct uk is, as we have shown, (m-k+1). u is equal to uk only if all of the (m-k+1) links of uk are equal to the corresponding links of u. k-1 Let e. be the condition that the a posteriori probability of one of the (f-l)k links 1 th th emerging from the i node of u and not including the (i+l) node of u is greater or equal to that of the corresponding link of u. Then 123

Pr(Uk=U) = Pr(not el; not e 2 ; not e 3 ;... ; not e i... ; not em-k+1) Pr(uk*u) = 1 - P r(uk=u) = Pr U L i=1 {ei}. The probability of the union of events is upper-bounded by the sum of the probabilities of the individual events. Pr(uk*u) -< m-k+ z 1 Pr(ei) The rate per symbol is defined as R n M 1 n m n /d 1 n n n n d where M is the number of code words of length n. 1 R = - n d (8) Fano has shown that for a given rate R and a given word length n the probability of error is bounded by P(e) < 2 e opt(r). n(9) E(R), the rate R. is the optimum exponent of the average probability of error and is a function of Now, in the case of Pr(ei), the number of the involved links is (-1) and the length of each link is kd; thus 1 k 1 Rk < -k nk =- nk = R jk-1 Therefore -E opt(r)kd Pr(ei) - 2e opt (10), by Eqs. 7 and 10, -E (R)kd P(uk u) < 2m e opt (11) Therefore, by Eqs. 5, 6, and 11, Pr(Ck) < e - ( R)n + 2m e opt (12) 124

Pr(Ck) < 2m ee(r)kd (13) because, as we shall show, by Eq. 21, E(R) -< E opt(r). opt Pr(Ck - 1 ) < 2 me - E (R)( k - 1 )d (14) The average number of computations, by Eqs. 4 and 14, is therefore bounded by m N < 2m 2 p k=l ke-e(r)(k-l)d (15) Now, R = 1 (k-1)d (k-l) d = ek-1 e(k-1)dr. Therefore m 2m 2 1 e(k-1)d[r-e(r)]. k= (16) Let R comp be the solution of R = E(R) (17) Then, for all rates below Rcomp, R - E(R) < 0, and N 2m2 1 (18) 1_ e[r-e(r)]d. The average amount of computations is therefore an algebraic function of m for all rates below R comp c. Evaluation of R comp Fano has shown that E op t (R ) = E op t ( 0 ) - R; opt opt for R < Rit crit (19) and opt opt for R.< R < C. crit (20) Let us set D o so as to make E (R) of Eq. 6 equal to E(R) = Eopt(0) - Rcomp. (21) 125

, by Eqs. 12 and 21, or R =E (O)-R comp opt comp R E (0). comp 2 opt (22) Also, E(R) =E (O)-R - E (0). (23) opt comp 2 opt N < 2m 2 1 S- exp d R- Eopt(0) E opt(0) is the zero-rate exponent of the upper bound on the average probability of error Pe of the same channel, when the optimal decoding scheme is used. 5 2. The Average Probability of Error Let u be the transmitted code word. Let v be the corresponding channel output vector. Let u' be one of the M(t-1) code words which starts with an information letter other than that of u. The probability of error is bounded by Pe < Pr(D(u, v) Do) + M(-)Pr(D(u', v)>do;d(u, v) >D 0 ) (24) P e Pr(D(u,v)<Do) + e n R Pr(D(u',v) >D ;D(u,v)>D ) (25) or Let Pe < P (D(u,v)Do) + e n R Pr(D(u ', v) >D;D(u, v) > D ) (26) D(u, v) = n D(u', v) = n P(v/u) P(v) P(v/u') (v) where n P(v/u) = - P(yi/xi) i=l 1 126

and n P(v) = -[ P(Yi ) i=l P(Yi/xi) is the probability that the output is yi, given that the input symbol x i was transmitted. X. P(x i ) P(Yi/Xi) = P(yi) D(u, v)= i=l di(xi' Yi),' where di(xi, yi) = P(yi/xi ) n P(yi) by the use of the Chernoff Bound, Pr(D(u, v) < Do) < e n ( G ( s )-s '(s)); s < 0 Pr(D(u', v) -< D )-< e n ( y(t)-ty'(t)).; t 0 where (s) = P(x) P(y/x) esd(x, y) xy y(t) = xxy P(x) P(x') P(y/x) e s d( x ' y) p(s) = P(x) P(y/x) l + s p(y)-s xy and y(t) = x'y P(x') P(y)l-t P(y/x')t = P(x) P(y)l-t P(y/x)t xy 127

Also, '(s) = y'(t) D no f we let t = 1 + s, it can be shown that y'(t) = 9'(t) (also y(t) = (s)). P (D(u, v) Do) - en((s) - s F'(s)) P (D(u', v) ad ) < e n ( ) s - + s ) [ ( s ' ) )' where D 1(s) - n f(s) = ) P(x) P(y/x) 1 + s p(y)-. xy P e < en(4(s)-s '(s))+ en(r+[(s)-(l+s)'(s)) (27) Now, by Eqs. 6 and 23, D o is set so as to get Pr(D(u, v) < D )< en(4(s )- s - '(s)) = -n(r), (28) where E(R) = Eopt(); for all R < R comp (28a) and, as shown by Fano, = -n xp(x) 2 E opt(0) = E opt(r) P(Y/x) R=0 y x We shall now prove that, once D is set so as to make -i(s) + as in Eq. 28, then -l(s) 1 + (i+s) ['(s) > 2 E opt(0). PROOF: The minimum of {(2s+l)p'(s)-2i(s)} occurs at that s for which (28b) s'(s) = E(R) 1 Eopt(0), SC(S) :~E(R) (29) a s [(2s+1) '(s)-2z(s)] = 0 (2s+l) '"(s ) + 2F'(s) - 2l'(s) = 0 1 s = - 1 2 128

Also, [(1+2s)~(s)-2[(s)] s=-1/2 2, (-- 0, since i" l(-) is the variance (see Fano 5 ) of a random variable. a minimum point. Now [(2s+1)i'(s)-2[i(s) > -24 -. s = -. 1. is indeed (30) and therefore - n = P (x) (y)/2 P(y/x)1/2 xy 4 1 = n P(x) P(y) 1 2 xy /2 p(y/x)l/2 2 24 n f(y) P(y)1/ y where f(y) = P(x) P(y/x) 1/ 2. -J By the Schwarz inequality, { f(y) P(y)1/2 2 y < f(y )2 1p(y ) Y Y y P(y) 2 7 P(y) y y Y 2 since P(y) = 1. Therefore, by Eq. 28b, 2 (- n P(x) y x P(y/x)1/2 =-E opt(0). opt -2z(-1) > Eopt(O), and therefore, by Eq. 30, 129

-[ '(s)i'(s)+(s) +s) (s)] > E(s)-s Eopt(0). (31) But, by Eq. 28, -E(R) = t(s) - s4'(s) 1 2 Eopt(0). Therefore, by Eq. 31, 4(s) - (l+s) 4'(s) > -- Eopt(0) Q.E.D. (32), by Eqs. 27, 28, and 32, P < e opt(0)n + en - Eopt(0) e n E (0) n(r-r p ) P 2 e n(r 2 Eopt = n(r-r comp e n(r-rcomp P e <2e comp L(s)- f D o is set so as to minimize the probability of error by making R+ [(s) - (1+s)['(s) = s '(s) = E(R), Shannon 7 has shown that P <2e2 < -P -ne(r) e where E(R) > 0 for R < C, and E(0) >Z E (0). However, R is then lower-bounded opt comp 1 1 by R comp > -2 -E(0) E() > >4 Ep E (0). t (O). This research was supported in part by a fellowship from the Government of srael. References 1. J. M. Wozencraft and B. Reiffen, Sequential Decoding (The M..T. Press, Cambridge, Mass., and John Wiley and Sons, nc., New York, 1961). 2. R. G. Gallager, Low Density Parity Check Codes, Sc.D. Thesis, Department of Electrical Engineering, M..T., September 1960. J. Ziv 3. B. Reiffen, Sequential Encoding and Decoding for the Discrete Memoryless Channel, Technical Report 374, Research Laboratory of Electronics, M..T.; Technical Report 231, Lincoln Laboratory, M..T., August 12, 1960, p. 29. 4. bid., see Chapter 2. 5. R. M. Fano, Transmission of nformation (The M..T. Press, Cambridge, Mass., and John Wiley and Sons, nc., New York, 1961), see Chapter 9. 6. bid., see Chapter 8; B. Reiffen, o F. cit., see Appendix B. 7. C. E. Shannon, Error Probability Bounds for Noisy Channels (unpublished report, 1959). 130