Aalborg Universitet. Bounds on information combining for parity-check equations Land, Ingmar Rüdiger; Hoeher, A.; Huber, Johannes

Similar documents
Bounds on Mutual Information for Simple Codes Using Information Combining

Decoding of LDPC codes with binary vector messages and scalable complexity

Turbo Compression. Andrej Rikovsky, Advisor: Pavol Hanus

Computation of Symbol-Wise Mutual Information in Transmission Systems with LogAPP Decoders and Application to EXIT Charts

Efficient Computation of EXIT Functions for Non-Binary Iterative Decoding

Successive Cancellation Decoding of Single Parity-Check Product Codes

Convergence analysis for a class of LDPC convolutional codes on the erasure channel

Höst, Stefan; Johannesson, Rolf; Zigangirov, Kamil; Zyablov, Viktor V.

On the Computation of EXIT Characteristics for Symbol-Based Iterative Decoding

A Relation between Conditional and Unconditional Soft Bit Densities of Binary Input Memoryless Symmetric Channels

Computation of Information Rates from Finite-State Source/Channel Models

Why We Can Not Surpass Capacity: The Matching Condition

Aalborg Universitet. Lecture 14 - Introduction to experimental work Kramer, Morten Mejlhede. Publication date: 2015

Maximum Likelihood Decoding of Codes on the Asymmetric Z-channel

Soft-Output Decision-Feedback Equalization with a Priori Information

An Alternative Proof of Channel Polarization for Channels with Arbitrary Input Alphabets

The Geometry of Turbo-Decoding Dynamics

LOW-density parity-check (LDPC) codes were invented

Citation for pulished version (APA): Henning, M. A., & Yeo, A. (2016). Transversals in 4-uniform hypergraphs. Journal of Combinatorics, 23(3).

QPP Interleaver Based Turbo-code For DVB-RCS Standard

Aalborg Universitet. Land-use modelling by cellular automata Hansen, Henning Sten. Published in: NORDGI : Nordic Geographic Information

Analysis of Sum-Product Decoding of Low-Density Parity-Check Codes Using a Gaussian Approximation

1e

Turbo Codes for Deep-Space Communications

Exact Probability of Erasure and a Decoding Algorithm for Convolutional Codes on the Binary Erasure Channel

Aalborg Universitet. All P3-equipackable graphs Randerath, Bert; Vestergaard, Preben Dahl. Publication date: 2008

Sound absorption properties of a perforated plate and membrane ceiling element Nilsson, Anders C.; Rasmussen, Birgit

THE seminal paper of Gallager [1, p. 48] suggested to evaluate

On The Binary Lossless Many-Help-One Problem with Independently Degraded Helpers

Trellis-based Detection Techniques

An Introduction to Low Density Parity Check (LDPC) Codes

On Generalized EXIT Charts of LDPC Code Ensembles over Binary-Input Output-Symmetric Memoryless Channels

A Note on Powers in Finite Fields

Construction of low complexity Array based Quasi Cyclic Low density parity check (QC-LDPC) codes with low error floor

Published in: Proceedings of the 21st Symposium on Mathematical Theory of Networks and Systems

Optimized Symbol Mappings for Bit-Interleaved Coded Modulation with Iterative Decoding

State-of-the-Art Channel Coding

SIPCom8-1: Information Theory and Coding Linear Binary Codes Ingmar Land

Aalborg Universitet. Published in: IEEE International Conference on Acoustics, Speech, and Signal Processing. Creative Commons License Unspecified

Unknown input observer based scheme for detecting faults in a wind turbine converter Odgaard, Peter Fogh; Stoustrup, Jakob

Time-invariant LDPC convolutional codes

Distributed Source Coding Using LDPC Codes

Estimation of Peak Wave Stresses in Slender Complex Concrete Armor Units

On the Block Error Probability of LP Decoding of LDPC Codes

The Effect of Memory Order on the Capacity of Finite-State Markov and Flat-Fading Channels

Development of a Dainish infrastructure for spatial information (DAISI) Brande-Lavridsen, Hanne; Jensen, Bent Hulegaard

Feedback Capacity of a Class of Symmetric Finite-State Markov Channels

Cut-Set Bound and Dependence Balance Bound

Polar codes for the m-user MAC and matroids

Interleaver Properties and Their Applications to the Trellis Complexity Analysis of Turbo Codes

Error Floor Analysis of Coded Slotted ALOHA over Packet Erasure Channels Ivanov, Mikhail; Graell i Amat, Alexandre ; Brannstrom, F.

An Introduction to Factor Graphs

Bifurcations and Chaos in Turbo Decoding Algorithms

Belief-Propagation Decoding of LDPC Codes

TURBO codes, which were introduced in 1993 by Berrou

Control Configuration Selection for Multivariable Descriptor Systems Shaker, Hamid Reza; Stoustrup, Jakob

Upper and Lower Bounds of Frequency Interval Gramians for a Class of Perturbed Linear Systems Shaker, Hamid Reza

Upper Bounds on the Capacity of Binary Intermittent Communication

Multi (building)physics modeling

Aalborg Universitet. CLIMA proceedings of the 12th REHVA World Congress Heiselberg, Per Kvols. Publication date: 2016

LDPC Codes. Slides originally from I. Land p.1

Geometry of power flows and convex-relaxed power flows in distribution networks with high penetration of renewables

EE229B - Final Project. Capacity-Approaching Low-Density Parity-Check Codes

Citation for published version (APA): Burcharth, H. F. (1988). Stress Analysis. Aalborg: Department of Civil Engineering, Aalborg University.

A Hyper-Trellis based Turbo Decoder for Wyner-Ziv Video Coding

Graph-based codes for flash memory

Factor Graphs and Message Passing Algorithms Part 1: Introduction

Multiple Scenario Inversion of Reflection Seismic Prestack Data

A New Performance Evaluation Metric for Sub-Optimal Iterative Decoders

Online algorithms for parallel job scheduling and strip packing Hurink, J.L.; Paulus, J.J.

Low Density Parity Check (LDPC) Codes and the Need for Stronger ECC. August 2011 Ravi Motwani, Zion Kwok, Scott Nelson

Modelling Salmonella transfer during grinding of meat

Tectonic Thinking Beim, Anne; Bech-Danielsen, Claus; Bundgaard, Charlotte; Madsen, Ulrik Stylsvig

Noisy channel communication

OPTIMUM fixed-rate scalar quantizers, introduced by Max

Capacity-approaching codes

On Achievable Rates and Complexity of LDPC Codes over Parallel Channels: Bounds and Applications

arxiv:cs/ v2 [cs.it] 1 Oct 2006

Construction of Protographs for QC LDPC Codes With Girth Larger Than 12 1

Distributed Functional Compression through Graph Coloring

Lecture 4 : Introduction to Low-density Parity-check Codes

Generating the optimal magnetic field for magnetic refrigeration

Lidar calibration What s the problem?

THE EXIT CHART INTRODUCTION TO EXTRINSIC INFORMATION TRANSFER IN ITERATIVE PROCESSING

IN MOST OF the theory and practice of error-control

Minimum Distance and Convergence Analysis of Hamming-Accumulate-Acccumulate Codes

Approaching Blokh-Zyablov Error Exponent with Linear-Time Encodable/Decodable Codes

Published in: Proceedings of the 17th International Modal Analysis Conference (IMAC), Kissimmee, Florida, USA, February 8-11, 1999

Tilburg University. Strongly Regular Graphs with Maximal Energy Haemers, W. H. Publication date: Link to publication

Performance of Polar Codes for Channel and Source Coding

Fountain Uncorrectable Sets and Finite-Length Analysis

Multi-Kernel Polar Codes: Proof of Polarization and Error Exponents

Characterization of microporous silica-based membranes by calorimetric analysis Boffa, Vittorio; Yue, Yuanzheng

Graph-based Codes and Iterative Decoding

Graph Coloring and Conditional Graph Entropy

Fast Joint Source-Channel Decoding of

GIS Readiness Survey 2014 Schrøder, Anne Lise; Hvingel, Line Træholt; Hansen, Henning Sten; Skovdal, Jesper Høi

TURBO codes [7] and low-density parity-check (LDPC)

A note on non-periodic tilings of the plane

Optimal Natural Encoding Scheme for Discrete Multiplicative Degraded Broadcast Channels

Transcription:

Aalborg Universitet Bounds on information combining for parity-check equations Land, Ingmar Rüdiger; Hoeher, A.; Huber, Johannes Published in: 2004 International Seminar on Communications DOI link to publication from Publisher): 10.1109/IZS.2004.1287390 Publication date: 2004 Document Version Publisher's PDF, also known as Version of record Link to publication from Aalborg University Citation for published version APA): Land, I., Hoeher, A., & Huber, J. 2004). Bounds on information combining for parity-check equations. In 2004 International Seminar on Communications pp. 68-71). IEEE Signal Processing Society. DOI: 10.1109/IZS.2004.1287390 General rights Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights.? Users may download and print one copy of any publication from the public portal for the purpose of private study or research.? You may not further distribute the material or use it for any profit-making activity or commercial gain? You may freely distribute the URL identifying the publication in the public portal? Take down policy If you believe that this document breaches copyright please contact us at vbn@aub.aau.dk providing details, and we will remove access to the work immediately and investigate your claim. Downloaded from vbn.aau.dk on: december 30, 2017

Bounds on Information Combining for Parity-Check Equations Ingmar Land and Peter A. Hoeher Information and Coding Theory Lab University of Kiel, Germany il,ph}@tf.uni-kiel.de www.tf.uni-kiel.de/ict Johannes Huber Chair of Information Transmission University Erlangen-Nürnberg, Germany huber@lnt.de www.lnt.de/lit Abstract When several independent channels are coupled by a parity check constraint on their inputs, the mutual information between the input of one channel and the outputs of all other channels can be expressed as a combination of the mutual information between the input and the output of each individual channel. This concept is denoted as information combining. For binary-input symmetric discrete memoryless channels, we present bounds on the combined information which are only based on the mutual information of the channels. Furthermore, we show that these bounds cannot be further improved. Exact expressions are provided for the case that all channels are binary symmetric channels and for the case that all channels are binary erasure channels. I. INTRODUCTION Consider coded transmission over one discrete memoryless channel or over multiple parallel discrete memoryless channels. Each noisy observation of a code bit contains information on this code bit. Furthermore, each observation also contains information on other code bits and information bits or for short, info bits) due to the code constraints. Accordingly, the overall information on a certain code bit or info bit is a combination of the information on code bits. In this paper, we will use the term info bit instead of information bit to avoid confusion with mutual information.) This combining of information under code constraints is used in every decoder, but the way code constraints are taken into account may differ. Whereas an APP decoder, e.g. [1], considers all constraints at once, iterative decoding algorithms consider only a subset of all code constraints in each decoding step. An iterative decoder for parallel and serially concatenated codes turbo codes), e.g. [2], [3], [4], takes into account only the constraints of the constituent codes. An iterative decoder for low-density parity-check codes, e.g. [5], [6], or in general, any iterative decoder operating on graphs, e.g. [7], [8], [9], takes into account only local code constraints. Such a basic local constraint between code bits is given by a parity check equation. If the information on the code bits is represented by probabilities, log-likelihood ratios, or similar measures, then the combined information on one code bit can easily be computed by an appropriate operation as given, e.g., in [9]. In this paper, information combining is used in a very strict sense, namely only for combining of mutual information. This notion of information combining was introduced in [10], [11] for design and analysis of parallel concatenated coding schemes. Since not all statistical properties of the channels are taken into account, but only their mutual information, only bounds on the combined information can be given. In [12], bounds on information combining were presented for the case that a binary symbol is transmitted over two independent channels. It was shown that these bounds cannot be further improved, and that the lower bound corresponds to the case that both channels are binary symmetric channels BSCs), and that the upper bound corresponds to the case that both channels are binary erasure channels BECs). Two independent channels having the same input can be interpreted as two independent channels with a parity-check constraint on their inputs. In this paper, this concept is generalized to the case of an arbitrary number of independent channels with their inputs fulfilling a parity-check equation; the inputs can be regarded as the code bits of a parity check code. This scenario can also be regarded as decoding on a local parity check constraint in a graph, as mentioned above. The channels under consideration are binary-input symmetric discrete memoryless channels BISDMC). Bounds are presented for the mutual information between a certain code bit and the observations of the other code bits; this information is denoted as extrinsic information on this code bit. Here is an outline of the paper: In Section II, some definitions and properties for BISDMCs are given. Section III addresses information combining across the parity check equation. Bounds for the general case and exact expressions for the only-bsc case and for the only-bec case are given. Finally, conclusions are drawn in Section IV. II. BINARY-INPUT DISCRETE MEMORYLESS CHANNELS: DEFINITIONS AND PROPERTIES Let X i Y i denote a binary-input symmetric discrete memoryless channel BISDMC) with inputs X i X := 1, +1} and outputs Y i Y i R, where X and Y i denote the input and the output alphabet of the channel, respectively. The transition probabilities are given by p Yi X i y x), denoting the probability density function for continuous output alphabets and denoting the probability mass function for discrete output alphabets. Since the channel is symmetric, we can assume p Yi X i y x) = p Yi X i y x)

for all x X and y Y i without significant loss of generality. The mutual information MI) of the channel is defined as I i := IX i ; Y i ). 1) Let define the random variable J i J i := y Y i : y 0} as the magnitude of Y i, J i := Y i. Using J i, the elements of the output alphabet Y i can be partitioned into the subsets +j, j} for j J i \0}, Y i j) := 0, 0} for j = 0. With these definitions, J i indicates which output set Y i j) the output symbol Y i belongs to. The random variable J i separates the symmetric channel X i Y i into strongly symmetric sub-channels X i Y i J i = j; it is therefore denoted as sub-channel indicator. The subchannels are binary symmetric channels. Their conditional crossover probabilities ɛ i j) are defined as p ɛ i j) := Yi X i,j i j + 1, j) for j J i \0}, 1 2 for j = 0. Let hx) := x ld x 1 x) ld 1 x), x [0, 1], denote the binary entropy function, and let h 1 y), y [0, 1], denote its inverse for x [0, 1 2 ]. Then, the MI of sub-channel j is given as I i j) := IX i ; Y J i = j) = 1 hɛ i j)). 2) Note that for J i = 0, the BEC with zero MI is transformed into an equivalent BSC with zero MI, so that all sub-channels are BSCs.) Using the above definitions, the MI of the channel can be written as the expected value of the MI of its sub-channels, I i = E j Ji I i j)}. 3) The separation of a BISDMC into sub-channels which are BSCs is exploited in the following. III. EXTRINSIC INFORMATION FOR PARITY CHECK EQUATIONS Let C 1, C 2,..., C N 0, 1} =: F 2 denote N binary code bits, which are uniformly distributed and independent except for the parity-check constraint C 1 C 2 C N = 0. 4) The code bits are mapped to the channel inputs X i according to the one-to-one mapping 0 +1, 1 1. For the sake of convenience, also X i will be denoted as code bit when possible without causing ambiguity.) The code bits X i are transmitted over independent BISDMCs X i Y i having MI I i := IX i ; Y i ), i = 1, 2,..., N. Note that this includes also the special case that all code bits are transmitted over the same channel, as we consider memoryless channels. The relations are depicted in Figure 1. C 1 C 2 C N X 1 X 2 X N Fig. 1. N independent channels concatenated by a parity-check constraint on the inputs. The notion of extrinsic probability plays an important role in the context of iterative decoding, see e.g. [2], [13]. The extrinsic probability of a bit is defined as the probability of this bit given all observations except for its direct observation. Similarly, an extrinsic information on a bit can be defined as also done in [14]). In the following, the short hand notation a j i := [a i, a i+1,..., a j 1, a j ] is used for subsequences. Definition 1 The extrinsic information I e1 on bit X 1 is defined as I e1 := IX 1 ; Y2 N ), i.e., all channel outputs except the direct observation Y 1 are taken into account. The extrinsic information for X i, i = 2, 3,..., N, is defined analogous. Y 1 Y 2 Y N The following class of functions will show to be useful. Definition 2 For x 1, x 2,..., x n [0, 1], n 2, the function f 2 x 1, x 2 ) is defined as f 2 x 1, x 2 ) := 1 h [1 h 1 1 x 1 )] h 1 1 x 2 ) ) + h 1 1 x 1 ) [1 h 1 1 x 2 )], and the function f n x 1, x 2,..., x n ), n > 2, is recursively defined as f n x 1, x 2,..., x n ) := f 2 x1, f n 1 x 2, x 3,..., x n ) ). An interpretation of these functions is as follows. Consider n BSCs S i R i, S i, R i 1, +1}, having MI I i := IS i ; R i ), i = 1, 2,..., n. These BSCs are serially concatenated such that R i = S i+1 for i = 1, 2,..., n 1. Then, the end-to-end MI is given as IS 1 ; R n ) = f n I 1, I 2,..., I n ). Using the above definitions, the main theorem of this paper can be stated as follows: Theorem 1 Bounds on extrinsic information) If the channels X i Y i are BISDMCs having MI I i := IX i ; Y i ), i = 2, 3,..., N, then the extrinsic information on code bit X 1, I e1 := IX 1 ; Y2 N ), is bounded as I 2 I 3 I N I e1 f N 1 I 2, I 3,..., I N ). The upper bound is achieved if all channels are BSCs, as will be shown in Theorem 2, and the lower bound is achieved if all channels are BECs, as will be shown in Theorem 3. Since we have examples achieving the bounds, these bounds cannot be further improved.

g replacements The bounds are illustrated for the case of I 2 = I 3 = = I N in Figure 2. Note that this corresponds to the case that the code bits are transmitted over channels which have the same MI but may differ in other statistical properties. Ie1 1 0.8 0.6 0.4 0.2 N = 2 N = 3 N = 5 N = 9 0 0 0.2 0.4 0.6 0.8 1 I i = IX i ; Y i ) Fig. 2. Bounds on extrinsic information I e1 = IX 1 ; Y2 N ) vs I 2 = I 3 = = I N. The lower bounds dashed lines) correspond to the case of BECs, the upper bounds solid lines) correspond to the case of BSCs. In the following, first the extrinsic information for two special cases is computed: a) all channels are BSCs; b) all channels are BECs. Then, these results are used to prove the bounds for the general case, i.e., to prove Theorem 1. A. Only Binary Symmetric Channels If all channels X i Y i are BSCs, then the extrinsic information can be expressed using function f n x 1, x 2,..., x n ) according to Definition 2. Theorem 2 BSC case) If the channels X i Y i are BSCs having MI I i := IX i ; Y i ), i = 2, 3,..., N, then the extrinsic information on code bit X 1, I e1 := IX 1 ; Y2 N ), is given as I e1 = f N 1 I 2, I 3,..., I N ). Proof: Using the chain rule of mutual information, the extrinsic information can be written as IX 1 ; Y2 N ) = IX 1 ; Y2 N 1 ) + IX 1 ; Y N Y2 N 1 ). Due to the independence of X 1, X 2,..., X N 1, we have IX 1 ; Y2 N 1 ) = 0. Let binary random variables Z i F 2, i = 1, 2,..., N, be defined as C 1 = Z 1, Z 1 C 2 = Z 2, Z 2 C 3 = Z 3,... Z N 2 C N 1 = Z N 1, Z N 1 = C N, Z N = Y N. It follows from the definitions that all Z i distributed and that are uniformly IX 1 ; Y N Y N 1 2 ) = IZ 1 ; Z N Y N 1 2 ). 5) For the time being, assume Y2 N 1 = y2 N 1, where y2 N 1 denotes an arbitrary but fixed realization of Y2 N 1. Then, the random variables Z i form a chain of BSCs, Z 1 Z 2 Z 3 Z N 2 Z N 1 Z N. Consider now the MI of each BSC: Z 1 Z 2 : Since C 2 represents the error bit, the crossover probability is given by ɛ 2 := min pc2 Y y 2 2 1 y 2) } = h 1 1 I 2 ). Thus, the MI for this channel is simply given by 1 hɛ 2 ) = I 2, which is independent from y 2. Z i Z i+1, i = 2, 3,..., N 2: Similarly to Z 1 Z 2, the MI is given by I i, respectively. Z N 1 Z N : This channel is equal to the channel X N Y N, and thus the MI is I N. Using the interpretation of Definition 2, we immediately get IZ 1 ; Z N Y N 1 2 = y N 1 2 ) = f N 1 I 2, I 3,..., I N ). Due to the independence from y N 1 2, we have IZ 1 ; Z N Y2 N 1 ) = E IZ 1 ; Z N Y2 N 1 = y2 N 1 ) } = f N 1 I 2, I 3,..., I N ). Together with 5), this concludes the proof. B. Only Binary Erasure Channels In this subsection, it is assumed that all channels X i Y i are BECs. In this case, the extrinsic information can be derived using a simple combinatorial approach. Theorem 3 BEC case) If the channels X i Y i are BECs having MI I i := IX i ; Y i ), i = 2, 3,..., N, then the extrinsic information on code bit X 1, I e1 := IX 1 ; Y2 N ), is given as I e1 = I 2 I 3 I N. Proof: Let δ i = PrY i = ) denote the erasure probability of channel X i Y i. X can be recovered from Y2 N only if no erasure occurred, due to the independence of X i and the parity check constraint. This happens with probability 1 δ 2 )1 δ 3 ) 1 δ N ), and the corresponding MI is then equal to 1. In all other cases, the MI is equal to zero. Thus, we have IX 1 ; Y N 2 ) = 1 δ 2 )1 δ 3 ) 1 δ N ). Regarding δ i = 1 I i concludes the proof.

C. General Symmetric Memoryless Channels If the channels X i Y i are only assumed to be BISDMCs without further restrictions, the extrinsic information cannot be computed exactly. Nevertheless, bounds can be given according to Theorem 1. In this section, the proof of this theorem is given. First, two properties of the functions according to Definition 2 are needed. Lemma 1 The function f n x 1, x 2,..., x n ), x 1, x 2,..., x n [0, 1], n 2, has the following two properties: a) f n x 1, x 2,..., x n ) is convex- in each x i, i = 1, 2,..., n; b) f n x 1, x 2,..., x n ) is not less than the product of its arguments: f n x 1, x 2,..., x n ) x 1 x 2 x n. The proof follows immediately from Lemma 2 in [12] or Lemma 2 in [15]. Using the above lemma, Theorem 1 can be proved as follows. Proof: The extrinsic information does not change if it is written conditional on the sub-channel indicators J2 N as IX 1 ; Y2 N ) = IX 1 ; Y2 N J2 N ) = E IX 1 ; Y N j2 N 2 J2 N = j2 N )} = E I2 j 2 ), I 3 j 3 ),..., I N j N ) )}. 6) The argument in the second line corresponds to the case, where all channels are BSCs, due to the conditions. Therefore, this expression can be replaced by the function f N 1...) according to Theorem 2. In the next step, the two properties of function f N 1...) given in Lemma 1 are exploited. First, using the lower bound for this function in 6), we get E I2 j 2 ), I 3 j 3 ),..., I N j N ) )} E I2 j 2 )I 3 j 3 ) I N j N ) } = j2 N = I 2 I 3 I N, where in the last line, 3) was used. Second, since the function f N 1...) is convex-, Jensen s inequality can be applied in 6), and we get E I2 j 2 ), I 3 j 3 ),..., I N j N ) )} f N 1 Ej2 I 2 j 2 )}, E j3 I 3 j 3 )},..., E jn I N j N )} ) = IV. CONCLUSIONS = f N 1 I 2, I 3,..., I N ). In this paper, binary-input symmetric discrete memoryless channels connected by a parity-check constraint on their inputs were considered. Bounds on the combined extrinsic information on one bit based on the mutual information of each channel were presented. For the cases that all BISDMCs are BSCs or BECs, respectively, the extrinsic information was computed exactly. Since these two cases correspond to the two bounds, the given bounds cannot be further improved. In this paper, we focused on the case that the channels are coupled by a parity-check constraint on their inputs. But the applied techniques may be used or extended to yield similar results for other constraints. Of special interest is certainly the case that all channels have the same inputs, which may be denoted as equality constraint [12]. In [11], the concept of information combining and the corresponding bounds were applied to analysis and design of concatenated codes. Using the bounds for the parity-check constraint, presented in this paper, and the bounds for the equality constraint, mentioned above, the iterative decoder for low-density parity-check codes can be analyzed without the commonly used Gaussian approximation for the a-priori distribution. REFERENCES [1] L. Bahl, J. Cocke, F. Jelinek, and J. Raviv, Optimal decoding of linear codes for minimizing symbol error rate, IEEE Trans. Inform. Theory, pp. 284 287, Mar. 1974. [2] C. Berrou and A. Glavieux, Near optimum error correcting coding and decoding: Turbo-codes, IEEE Trans. Commun., vol. 44, no. 10, pp. 1261 1271, Oct. 1996. [3] S. Benedetto and G. Montorsi, Unveiling turbo codes: Some results on parallel concatenated coding schemes, IEEE Trans. Inform. Theory, vol. 42, no. 2, pp. 409 428, Mar. 1996. [4] S. Benedetto, D. Divsalar, G. Montorsi, and F. Pollara, Serial concatenation of interleaved codes: Performance analysis, design, and iterative decoding, IEEE Trans. Inform. Theory, vol. 44, no. 3, pp. 909 926, May 1998. [5] R. Gallager, Low-density parity-check codes, IEEE Trans. Inform. Theory, vol. 8, no. 1, pp. 21 28, Jan. 1962. [6] D. J. MacKay, Good error-correcting codes based on very sparse matrices, IEEE Trans. Inform. Theory, vol. 45, no. 2, pp. 399 431, Mar. 1999. [7] R. M. Tanner, A recursive approach to low complexity codes, IEEE Trans. Inform. Theory, vol. 27, no. 5, pp. 533 547, Sept. 1981. [8] N. Wiberg, H.-A. Loeliger, and R. Koetter, Codes and iterative decoding on general graphs, in Proc. IEEE Int. Symp. Inform. Theory ISIT), Sept. 1995, p. 468. [9] F. Kschischang, B. Frey, and H.-A. Loeliger, Factor graphs and the sum-product algorithm, IEEE Trans. Inform. Theory, vol. 47, no. 2, pp. 498 519, Feb. 2001. [10] S. Huettinger, J. Huber, R. Johannesson, and R. Fischer, Information processing in soft-output decoding, in Proc. Allerton Conf. on Communications, Control, and Computing, Monticello, Illinois, USA, Oct. 2001. [11] S. Huettinger, J. Huber, R. Fischer, and R. Johannesson, Soft-outputdecoding: Some aspects from information theory, in Proc. Int. ITG Conf. on Source and Channel Coding, Berlin, Germany, Jan. 2002, pp. 81 90. [12] I. Land, S. Huettinger, P. A. Hoeher, and J. Huber, Bounds on information combining, in Proc. Int. Symp. on Turbo Codes & Rel. Topics, Brest, France, Sept. 2003, pp. 39 42. [13] J. Hagenauer, E. Offer, and L. Papke, Iterative decoding of binary block and convolutional codes, IEEE Trans. Inform. Theory, vol. 42, no. 2, pp. 429 445, Mar. 1996. [14] S. ten Brink, Convergence behavior of iteratively decoded parallel concatenated codes, IEEE Trans. Commun., vol. 49, no. 10, pp. 1727 1737, Oct. 2001. [15] A. D. Wyner and J. Ziv, A theorem on the entropy of certain binary sequences and applications: Part I, IEEE Trans. Inform. Theory, vol. 19, no. 6, pp. 769 772, Nov. 1973.