,. WOLFOWITZ. 0 < h < 1. A"code" of length is set l(x,a),i 1,...,t},where CORRECTING CODE THE MAXIMUM ACHIEVABLE LENGTH OF AN ERROR

Similar documents
CSCI 2570 Introduction to Nanocomputing

Lesson 2: Introduction to Variables

An instantaneous code (prefix code, tree code) with the codeword lengths l 1,..., l N exists if and only if. 2 l i. i=1

AUTOMORPHISMS OF ORDER TWO*

Exercise 1. = P(y a 1)P(a 1 )

STRONG CONVERSE OF THE CODING THEOREM FOR SEMICONTINUOUS CHANNELS

EE376A: Homework #3 Due by 11:59pm Saturday, February 10th, 2018

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

EE 4TM4: Digital Communications II. Channel Capacity

Module 1. Introduction to Digital Communications and Information Theory. Version 2 ECE IIT, Kharagpur

1 Introduction to information theory

12x + 18y = 50. 2x + v = 12. (x, v) = (6 + k, 2k), k Z.

(Classical) Information Theory III: Noisy channel coding

18.310A Final exam practice questions

School of Computer and Communication Sciences. Information Theory and Coding Notes on Random Coding December 12, 2003.

Fundamentals. Copyright Cengage Learning. All rights reserved.

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Dept. of Linguistics, Indiana University Fall 2015

Codes for Partially Stuck-at Memory Cells

Cryptography and Number Theory

Chapter 9 Fundamental Limits in Information Theory

Noisy channel communication

1 Basic Combinatorics

MATH 433 Applied Algebra Lecture 4: Modular arithmetic (continued). Linear congruences.

DEPARTMENT OF EECS MASSACHUSETTS INSTITUTE OF TECHNOLOGY. 6.02: Digital Communication Systems, Fall Quiz I. October 11, 2012

Shannon s noisy-channel theorem

A Generalized Uncertainty Principle and Sparse Representation in Pairs of Bases

Lecture 4 Noisy Channel Coding

LECTURE 13. Last time: Lecture outline

PERFECTLY secure key agreement has been studied recently

Variable Length Codes for Degraded Broadcast Channels

SOME FORMULAE FOR THE FIBONACCI NUMBERS

3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions

Bounds on Mutual Information for Simple Codes Using Information Combining

Lesson 2: Introduction to Variables

Notes 6 : First and second moment methods

Arimoto Channel Coding Converse and Rényi Divergence

Graph Theory. Dr. Saad El-Zanati, Faculty Mentor Ryan Bunge Graduate Assistant Illinois State University REU. Graph Theory

SERIES

New Trellis Codes Based on Lattices and Cosets *

Shannon's Theory of Communication

Anatomy of a Deterministic Finite Automaton. Deterministic Finite Automata. A machine so simple that you can understand it in less than one minute

The cocycle lattice of binary matroids

Midterm Exam Information Theory Fall Midterm Exam. Time: 09:10 12:10 11/23, 2016

IMC 2015, Blagoevgrad, Bulgaria

Introduction To Matrices MCV 4UI Assignment #1

A Piggybank Protocol for Quantum Cryptography

Lecture 4 Channel Coding

AN INVESTIGATION OF THE CHUNG-FELLER THEOREM AND SIMILAR COMBINATORIAL IDENTITIES

Chapter 1. Sets and Numbers

Noisy-Channel Coding

z-transforms 21.1 The z-transform Basics of z-transform Theory z-transforms and Difference Equations 36

6196 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 9, SEPTEMBER 2011

Math 430 Midterm II Review Packet Spring 2018 SOLUTIONS TO PRACTICE PROBLEMS

Network coding for multicast relation to compression and generalization of Slepian-Wolf

2x 1 7. A linear congruence in modular arithmetic is an equation of the form. Why is the solution a set of integers rather than a unique integer?

Decentralized Detection in Sensor Networks

Lecture 6 I. CHANNEL CODING. X n (m) P Y X

Jim Lambers MAT 169 Fall Semester Lecture 4 Notes

Finite Mathematics. Nik Ruškuc and Colva M. Roney-Dougal

Math 564 Homework 1. Solutions.

P/, t. co P9 ALGEBRAIC DECODING FOR A BINARY ERASURE CHANNEL TECHNICAL REPORT 340 RESEARCH LABORATORY OF ELECTRONICS M. A. EPSTEIN

MATH CSE20 Homework 5 Due Monday November 4

Frans M.J. Willems. Authentication Based on Secret-Key Generation. Frans M.J. Willems. (joint work w. Tanya Ignatenko)

Executive Assessment. Executive Assessment Math Review. Section 1.0, Arithmetic, includes the following topics:

Information Theory and Statistics Lecture 2: Source coding

Proof Techniques (Review of Math 271)

Discrete Math, Fourteenth Problem Set (July 18)

Geometric Sequences. Geometric Sequence a sequence whose consecutive terms have a common ratio.

Representation of Correlated Sources into Graphs for Transmission over Broadcast Channels

Chapter 6. Infinite series

11 Division Mod n, Linear Integer Equations, Random Numbers, The Fundamental Theorem of Arithmetic

Information Theory CHAPTER. 5.1 Introduction. 5.2 Entropy

Advanced Mathematics Unit 2 Limits and Continuity

Advanced Mathematics Unit 2 Limits and Continuity

, p 1 < p 2 < < p l primes.

Chapter 7. Error Control Coding. 7.1 Historical background. Mikael Olofsson 2005

And for polynomials with coefficients in F 2 = Z/2 Euclidean algorithm for gcd s Concept of equality mod M(x) Extended Euclid for inverses mod M(x)

Massachusetts Institute of Technology

Phase Transitions of Random Codes and GV-Bounds

Lecture 1: September 25, A quick reminder about random variables and convexity

Non Deterministic Automata. Linz: Nondeterministic Finite Accepters, page 51

Solving a linear equation in a set of integers II

p-adic Egyptian Fractions

NC Common Core Middle School Math Compacted Curriculum 7 th Grade Model 1 (3:2)

a 2 + b 2 = (p 2 q 2 ) 2 + 4p 2 q 2 = (p 2 + q 2 ) 2 = c 2,

ECE 564/645 - Digital Communications, Spring 2018 Midterm Exam #1 March 22nd, 7:00-9:00pm Marston 220

38 Riemann sums and existence of the definite integral.

Some remarks on the Shannon capacity of odd cycles Bruno Codenotti Ivan Gerace y Giovanni Resta z Abstract We tackle the problem of estimating the Sha

Shannon s Noisy-Channel Coding Theorem

Capacity of a Two-way Function Multicast Channel

A ZERO ENTROPY T SUCH THAT THE [T,ID] ENDOMORPHISM IS NONSTANDARD

An introduction to basic information theory. Hampus Wessman

MATH 573 FINAL EXAM. May 30, 2007

ALGEBRA. 1. Some elementary number theory 1.1. Primes and divisibility. We denote the collection of integers

Modular numbers and Error Correcting Codes. Introduction. Modular Arithmetic.

MATH3302. Coding and Cryptography. Coding Theory

Lecture 15: Conditional and Joint Typicaility

Math 308 Midterm Answers and Comments July 18, Part A. Short answer questions

THESIS. Presented in Partial Fulfillment of the Requirements for the Degree Master of Science in the Graduate School of The Ohio State University

Transcription:

THE MAXIMUM ACHIEVABLE LENGTH OF AN ERROR CORRECTING CODE BY,. WOLFOWITZ 1. Introduction The present pper continues the investigation initiated in [1]. It is written so that it cn be red without knowledge of [1] t least until the details of the proof begin. We shll assume that the lphbets involved contain exactly two symbols, 0 nd 1. This involves no loss of generality, s the extension to lphbets with ny finite number of symbols is immediate nd straightforward. Suppose that person hs vocabulary of S words (or messages), ny or ll of which he may wnt to transmit, in ny frequency nd in ny order, over "noisy channel" of memory m (m n integer => 0). We shll now explain wht is meant by noisy channel of memory m. A sequence of (m 1) elements, each zero or one, will be clled n asequence, Let the asequences be numbered in some fixed but arbitrary mnner from 1 to 2 +. The channel probability function p(ilj is nonnegtive function defined for i 0, 1 nd j 1, 2m+, such that for every j, p(oij) " p(llj 1. Any sequence of n elements, ech of which is zero or one, is clled n xsequence. To describe the operation of the channel it will be sufficient to describe how it transmits ny given xsequence, say x. Let a be the asequence of the first (m + 1) elements of x. The channel "performs" chance experiment with possible outcomes 0 and 1 nd respective probabilities p(0 a) and p(1 a), and transmits the outcome of this chance experiment. It then performs nother chance experiment, independently of the first, with possible 2, outcomes 0 nd 1 nd respective probabilities p(0 a) nd p(1 a) where a is the asequence of the 3r, (m 2) elements of the sequence x. This is repeted until (n m) independent experiments hve been i, performed. The probabilities of the outcomes 0 nd 1 in the i experiment re, respectively, p(0 a) nd p(1 a), where a is the asequence of the (i 1), (i k m) t elements of x. The xsequence x is clled the transmitted sequence. The chance sequence Y(x) of outcomes of the experiments in consecutive order is clled the received sequence. Any sequence of (n m) elements, each zero or one, is called ysequence. Let P/ denote the probability of the relation in braces, nd let 0 < h < 1. A"code" of length is set l(x,a),i 1,...,t},where Received Jnury 13, 1958. This research ws supported by the U. S. Air Force through the Office of Scientific Research. 454

LENGTH OF AN ERROR CORRECTING CODE 455 each x is an xsequence, each A is a set of ysequences,, for each i we huve P{ Y(x) e A} > 1 and the sets A, A are disjoint. The coding problem which is central concern of the theory of transmission of messages my be described as follows: For given S, to find an n and then a code of length S. The practical applications of this will be as follows" When one wishes to transmit the i word one transmits the xsequence x. Whenever the receiver receives ysequence which is in A, he always concludes that the jt word has been sent. Whenever the receiver receives ysequence not in A1 u A...u As, he may draw ny conclusion he wishes about the word that has been sent. The probability that any word transmitted will be correctly received is _>_ 1 The purpose of this pper is to prove the following" THEOREM 5. There exists a functional J of the channel probability function p with the following property" Let 0 be arbitrary, and let n be suciently large. There exists a code of length 2 (z). No code can have a length greater than 2 n(z+). The functional J will be defined below by specified algebraic and analytic operations on p. It seems reasonable to call J "the capacity" of the channel. 2. Relation to previous results When m 0, a functional Co of p is defined in [1] (Section 6) and there called the capacity (when m 0). It is proved in [1] that, when m 0, there is a functional K > 0 of p such that, for any n, there exists a code of length 2 nv:n11 (Theorem 1), and no code can have a length greater than 2 nc+11 (Theorem 2; see also Theorem 4 of [2]). Hence, when m 0, Co J, and Theorem 5 is weaker than Theorems 1 and 2 of [1]. When m => 0, functionals C1 and C2 of the channel probability function are defined respectively in footnotes 5 and 9 of [1]. (Unfortunately the letters C1 and C2 are also used in another sense in [1], but only in Section 4 and for particular argument. We shall not make use of these latter quantities again). From their definition we always have Co =< C1 < C2. It was proved in [3] and [1] that, for any > 0 and n sufficiently large, there exists a code of length greater than 2 c. Hence Co < C =< C =< J. We have seen that when m 0 these four quantities coincide. What their relation is for general m is at present unknown. The functional C is easier to compute than J, and the functional C is at least as difficult to compute as J. If further investigation should prove that C1 < J, then it may not be of practical wlue to determine whether C < J. If, however, it should turn out that C J, then this would be very useful for computational purposes. Moreover, then the positive part of Theorem 5 could be strengthened by using Theorem 1 (of [1]) which asserts that there is a functional K > 0 of the channel prob The first four theorems of the present investigation ppered in [1] and [2].

456. WOLFOWITZ ability function such that there exists a code of length 2cl nl. The author coniectures that, for m > 1, C1 < J. If one neglects the s which occur in the statement of Theorem 5, then the latter roughly states that 2 nj is the maximum achievable length of code. Any improvement on this result must consist in a more precise analysis of the n term in the exponent. For example, in some of the theorems cited above, the n term was replaced by a Kn 11 term. One could also consider a channel such that the probability that a particular symbol be received in a given place is a function not only of the symbol transmitted in that place and the m preceding symbols transmitted, as in the present case, but also of the m preceding symbols received. Such a channel can be treated by the same method as that used below, with a result analogous to Theorem 5. 3. Proof of the theorem Let be any integer > m. Only for the purposes of this proof let us call any sequence of elements, each zero or one, an/sequence, and any sequence of (l m) elements, each zero or one, an/ sequence. Let the/sequences, (/ sequences) be numbered in some fixed but arbitrary manner from 1 to 2 (fromlto2m). Let(ilj),i= 1,..,2m;j 1,...,2 be the probability that the it/ sequence is received when the j/sequence is sent. Suppose we restrict ourselves to n which are multiples of 1. Then any xsequence can be regarded as a sequence of nil disjoint/sequences. Suppose further that in any ysequence (which of course has (n m) elements) we "pay no attention" to the elements in the places whose serial numbers are congruent to (l m + 1) or (l m + 2) or or l, modulo l; this means that we do not distinguish between ysequences which differ only in elements in such places. Then what we have is a transmission system in which the,. input alphabet (alphabet of symbols transmitted) has the elements 1, 2 the output alphabet (alphabet of symbols received) has the elements 1, 2 m, the memory is zero, and the channel probability function is The place of n in the original system is taken by n/1 in the new system. Since ysequences different in the original system may be identified in the new system, but never vice versa, it follows that, for any fixed n, any length of code achievable in the new system is also achievable in the original system. We now proceed with the new system exactly as in Section 6 of [1], except for the unimportant difference that the alphabets may now contain more than two symbols. Let X1, X, be independent, identically distributed, chance variables such that P{X, i} p i= 1,,2 and p 1. Let Y,, Y., be chance variables which take only the values 1, 2 m, and such that the conditional probability that Y i, givenx1,x,..., Y,,..., Y l,is(ilx). Then the entropy of the

LENGTH OF AN ERROR CORRECTING CODE Y process is ([1], equation (6.8)) H(Yo) ( po(j i)) log. (, where the summation with respect to i is from 1 to 2l, and the summation with respect to j is from 1 to 2 zre. Also the conditional entropy of the Yprocess relative to the Xprocess is ([1], equation (6.9)) Hx(fo) j p(j i) logs(jli) with the same limits of summation as before. Let Co(l) be the maximum of H(Yo) Hx(Yo) with respect to pl, p, which are subject to the conditions all p 0, p 1. Then Co(l) is the capacity ([1], Section 6) of the new system described in the preceding paragraph. It may be mentioned in passing that C(ll) <= C(l) Although we shall not use this fact below, it may make easier the J sup > Co(l)/1. > 0 be given. Let be an integer > m such that Co(l)/1 > J /2. C(l l). computation of J. We now define Let c Then (n/l) (Co(l) s/2) n(co(l)/l /2l) By Theorem 1 of [1], for n sufficiently large, there exists in the new system code of length greater than 2 (n/l)(c(l)e/2) 2 n(j). Hence for the same n there exists in the original system a code of length greater than 2 n(j). This establishes the first part of the theorem. We now prove the second part of the theorem. Suppose that, from every ysequence, we delete the elements in all places whose serial numbers are congruent to (l m 1), (1 m 2),..., l, modulo l. Let us call the resulting sequences y sequences. Every xsequence may be considered as sequence of nil disjoint /sequences. If, for transmission, we used xsequences and y sequences, then there would be no memory between (or among) the /sequences which compose the xsequence, and Theorem 2 of [1] would apply. (While Theorem 2 was proved for the case when both sending and receiving alphabets contain two symbols, i obviously applies when there are any finite number of symbols in each; it is especially easy to see how the proof of Theorem 4 of [2] carries over, and the latter implies Theorem 2.) We obtain the following result" For > 0 and n sufficiently large, there cannot exist a code of length greater than 2(n)(c()+). However, this is not yet the desired result, because in the system under consider

458 j. WOLFOWITZ ation by us it is ysequences and not y sequences which are received sequences. If one considers the proof of Theorem 2 of [1], or, better still, of the more one observes that the proof rests on bounding the general Theorem 4 of [2], number of ysequences generated by an xsequence. Every y sequence is a subsequence of 2 (/) ysequences. Suppose we change the definition of a ysequence generated by an xsequence as follows" Consider a ysequence to be the composition of the y sequence which is a subsequence of it, hereafter to be called the first sequence, and the subsequence of all elements in the y hereafter to be called the second sequence. Say that the ysequence is generated by the xsequence if the first and second sequences ech fulfill the corresponding requirements of being generated in the original sense (by the sequences of asequences which produced them). Then all the arguments of [1] and [2] are valid with this new sequence but not in the y sequence, definition. It is easy to see that the number of ysequences generated (in the new sense) by all xsequences is less than 2 (/) times the number of y sequences generated (in the old sense) by all xsequences. Consequently the proof of Theorem 4 of [2], which applies to the system of xsequences and y sequences, applies also to our original system of xsequences and ysequences, if we multiply by 2 (n/)n the number of y sequences generated by all xsequences. This will multiply the upper bound on the length of the We conclude that in the original system of xsequences and code by 2 (n/l)m. ysequences no code can have a length greater than 2 n] l) m. 2 (n/l) (Co (l) "ts/2). We have not yet specified 1. Let be so large that m/1 /2. Then (n/l).m + (n/1)(co(1) + /2) < n(j + ). This completes the proof of the second part of the theorem and of the theorem itself. From the argument of the second part of the theorem and the conclusion of the first par, one easily obtains that, whenever is so large that m/1 < el, J 1 < Co (l)/1. This fact simplifies the task of computing J. REFERENCES 1. J. WOLFOWlTZ, The coding of messages subject to chance errors, Illinois J. Math., vol. 1 (1957), pp. 591606. 2., An upper bound on the rate of transmission of messages, Illinois J. Math., vol. 2 (1958), pp. 137141. 3. A. KHINTCHINE, On the fundamental theorems of the theory of information, Uspehi Matem. Nauk (N.S.), vol. 11 no. 1 (67), (1956), pp. 1775. COl{NELL UNIVERSITY ITHACA, NEW YORK ISRAEL INSTITUTE OF TECHNOLOGY HAIFA, ISRAEL