Lecture 9 Polar Coding
|
|
- April Hines
- 6 years ago
- Views:
Transcription
1 Lecture 9 Polar Coding I-Hsiang ang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 29, / 25 I-Hsiang ang IT Lecture 9
2 In Pursuit of Shannon s Limit Since 1948, Shannon s theory has drawn the sharp boundary between the possible and the impossible in data compression and data transmission. Once fundamental limits are characterized, the next natural question is: How to achieve these limits with acceptable complexity? For source coding, soon after Shannon s 1948 paper, information and coding theorists found optimal compression schemes with low complexity: Huffman Code (1952): optimal for memoryless source Lempel-Ziv (1977): optimal for stationary ergodic source On the other hand, for channel coding, it turns out be a much harder problem. It has been the holy grail for coding theorist to find a coding scheme that achieves Shannon s limit with low complexity. 2 / 25 I-Hsiang ang IT Lecture 9
3 In Pursuit of Capacity-Achieving Codes Two barriers in pursuing a low-complexity capacity-achieving codes: 1 Lack of explicit construction. In Shannon s proof, it is only proved that there exists coding schemes that achieve capacity. 2 Lack of structure to reduce complexity. In the proof of coding theorems, complexity issues are often neglected, while codes with structures are hard to prove to achieve capacity. Since 1990 s, there are several practical codes found to approach capacity, including turbo code, low-density parity-check (LDPC) code, etc. These codes perform very well empirically, but still in lack of theoretical investigation on the performances and even proof of optimality. The first provably capacity-achieving coding scheme with acceptable complexity is polar code, introduced by Erdal Arıkan in Later in 2012, spatially coupled LDPC codes were also shown to achieve capacity (Shrinivas Kudekar, Tom Richardson, and Rüediger Urbanke). 3 / 25 I-Hsiang ang IT Lecture 9
4 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 7, JULY : A Method for Constructing Capacity-Achieving Codes for Symmetric Binary-Input Memoryless Channels Erdal Arıkan, Senior Member, IEEE Abstract A method is proposed, called channel polarization, to construct code sequences that achieve the symmetric capacity of any given binary-input discrete memoryless channel (B-DMC). The symmetric capacity is the highest rate achievable subject to using the input letters of the channel with equal probability. Channel polarization refers to the fact that it is possible to synthesize, out of independent copies of a given B-DMC, a second set of binary-input channels such that, as becomes large, the fraction of indices for which is near approaches and the fraction for which is near approaches. The polarized channels are well-conditioned for channel coding: one need only send data at rate through those with capacity near and at rate through the remaining. Codes constructed on the basis of this idea are called polar codes. The paper proves that, given any B-DMC with and any target rate, there exists a sequence ofpolar codes suchthat hasblock-length, rate, and probability of block error under successive cancellation decoding bounded as independently of the code rate. This performance is achievable by encoders and decoders with complexity for each. A. Preliminaries e write to denote a generic B-DMC with input alphabet, output alphabet, and transition probabilities. The input alphabet will always be, the output alphabet and the transition probabilities may be arbitrary. e write to denote the channel corresponding to uses of ; thus, with. Given a B-DMC, there are two channel parameters of primary interest in this paper: the symmetric capacity and the Bhattacharyya parameter The paper wins the 2010 Information Theory Society Best Paper Award. These parameters are used as measures of rate and reliability, Index Terms Capacity-achieving codes, channel capacity, channel polarization, Plotkin construction, polar codes, Reed respectively. is the highest rate at which reliable communication is possible across using the inputs of with equal Muller (RM) codes, successive cancellation decoding. 4 / 25 I-Hsiang ang frequency. IT Lecture 9 is an upper bound on the probability of max-
5 Overview hen Arıkan introduced polar codes in 2007, he focus on achieving capacity for the general binary-input memoryless symmetric channels (BMS), including BSC, BEC, etc. Later, polar codes are shown to be optimal in many other settings, including lossy source coding, non-binary-input channels, multiple access channels, source coding with side information (yner-ziv problem), etc. Instead of giving a comprehensive introduction, we shall introduce channel polarization and polar coding for BMS, in the following order: 1 First we introduce the concept of channel polarization. 2 Then we explore polar coding. 5 / 25 I-Hsiang ang IT Lecture 9
6 Notations Recall in channel coding, we use the DMC N times with N being the blocklength of the coding scheme. Since the channel is the main focus, we shall use the following notations throughout this lecture: to denote the channel p Y X P to denote the input distribution p X I (P, ) to denote I (X ; Y ). Beside, since we focus on BMS channels, and it is not difficult to prove that X Ber ( 1 2) achieves the channel capacity of any BMS, we shall use I ( ) (abuse of notation) to denote I (P, ) when the input P is Ber ( 1 2). In other words, the channel capacity of the BMS channel is I ( ). 6 / 25 I-Hsiang ang IT Lecture 9
7 1 Polarization 7 / 25 I-Hsiang ang IT Lecture 9
8 Single Usage of Channel X Y N Usage of Channel X 1 Y 1 M ENC X 2. Y 2 DEC ˆM X N Y N 8 / 25 I-Hsiang ang IT Lecture 9
9 Arıkan s Idea U 1 X 1 Y 1 V 1 U 2 Pre- Processing X 2. Y 2 Post- Processing V 2 U N X N Y N V N Apply special transforms to both input and output 9 / 25 I-Hsiang ang IT Lecture 9
10 Arıkan s Idea U 1 V 1 1 U 2 2. V 2 U N N V N 10 / 25 I-Hsiang ang IT Lecture 9
11 Arıkan s Idea Roughly NI ( ) channels with capacity 1 U 1 V 1 1 U 2 2. V 2 U N N V N 11 / 25 I-Hsiang ang IT Lecture 9
12 Arıkan s Idea Roughly NI ( ) channels with capacity 1 U 1 V 1 1 U 2 2 V 2. Roughly N (1 I ( )) channels with capacity 0 U N N V N Equivalently some perfect channels and some useless channels Polarization Coding becomes extremely simple: simply use those perfect channels for uncoded transmission, and throw those useless channels away. 12 / 25 I-Hsiang ang IT Lecture 9
13 1 Polarization 13 / 25 I-Hsiang ang IT Lecture 9
14 Arıkan s Consider two channel uses of : X 1 Y 1 X 2 Y 2 14 / 25 I-Hsiang ang IT Lecture 9
15 Arıkan s Consider two channel uses of : Apply the pre-processor: U 1 Y 1 X 1 = U 1 U 2, X 2 = U 2, where U 1 U 2, U 1, U 2 Ber ( 1 2). U 2 Y 2 e now have two synthetic channels induced by the above procedure: : U 1 V 1 (Y 1, Y 2 ) + : U 2 V 2 (Y 1, Y 2, U 1 ) The above transform yields the following two crucial phenomenon: I ( ) I ( ) I ( + ) I ( ) + I ( + ) = 2I ( ) (Polarization) (Conservation of Information) 15 / 25 I-Hsiang ang IT Lecture 9
16 Example: Binary Erasure Channel Example 1 Let be a BEC with erasure probability ε (0, 1), and I ( ) = 1 ε. Find the values of I ( ) and I ( + ), and verify the above properties. sol: Intuitively is worse than and + is better than : For, input is U 1, output is (Y 1, Y 2 ). Only when both Y 1 and Y 2 are not erased, one can figure out U 1! = is BEC with erasure probability 1 (1 ε) 2 = 2ε ε 2. For +, input is U 2, output is (Y 1, Y 2, U 1 ). As long as one of Y 1 and Y 2 are not erased, one can figure out U 2! = + is BEC with erasure probability ε 2. Hence, I ( ) = 1 2ε + ε 2 and I ( + ) = 1 ε / 25 I-Hsiang ang IT Lecture 9
17 Example: Binary Symmetric Channel Example 2 Let be a BSC with crossover probability p (0, 1), and I ( ) = 1 H b (p). Find the values of I ( ) and I ( + ). 17 / 25 I-Hsiang ang IT Lecture 9
18 Basic Properties Theorem 1 For any BMS channel and the induced channels {, + } from Arıkan s basic transformation, we have I ( ) I ( ) I ( + ) with equality iff I ( ) = 0 or 1. I ( ) + I ( + ) = 2I ( ) pf: e prove the conservation of information first: I ( ) + I ( + ) = I (U 1 ; Y 1, Y 2 ) + I (U 2 ; Y 1, Y 2, U 1 ) = I (U 1 ; Y 1, Y 2 ) + I (U 2 ; Y 1, Y 2 U 1 ) = I (U 1, U 2 ; Y 1, Y 2 ) = I (X 1, X 2 ; Y 1, Y 2 ) = I (X 1 ; Y 1 ) + I (X 2 ; Y 2 ) = 2I ( ). I ( + ) = I (X 2 ; Y 1, Y 2, U 1 ) I (X 2 ; Y 2 ) = I ( ), and hence the first property holds. (Proof of the condition for equality is left as exercise.) 18 / 25 I-Hsiang ang IT Lecture 9
19 Extremal Channels I( + ) I( )[bits] BSC BEC If we plot the information stretch I ( + ) I ( ) versus the original information I ( ), it can be shown that among all BMS channels: BEC maximizes the stretch BSC minimizes the stretch Lower boundary: I() [bits] (Taken from Chap of Moser[4].) 2H b (2p(1 p)) 2H b (p), where p = H b 1 (1 I ( )). Upper boundary: 2I ( ) (1 I ( )). 19 / 25 I-Hsiang ang IT Lecture 9
20 1 Polarization 20 / 25 I-Hsiang ang IT Lecture 9
21 Recursive Application of Arıkan s Transformation Duplicate, apply the transformation, and get and / 25 I-Hsiang ang IT Lecture 9
22 Recursive Application of Arıkan s Transformation Duplicate, apply the transformation, and get and +. Duplicate (and + ). 22 / 25 I-Hsiang ang IT Lecture 9
23 Recursive Application of Arıkan s Transformation Duplicate, apply the transformation, and get and +. Duplicate (and + ). Apply the transformation on, and get and / 25 I-Hsiang ang IT Lecture 9
24 Recursive Application of Arıkan s Transformation Duplicate, apply the transformation, and get and +. Duplicate (and + ). Apply the transformation on, and get and +. Apply the transformation on +, and get + and / 25 I-Hsiang ang IT Lecture 9
25 Recursive Application of Arıkan s Transformation Duplicate, apply the transformation, and get and +. Duplicate (and + ). Apply the transformation on, and get and +. Apply the transformation on +, and get + and ++.. e can keep going and going, until the desired blocklength is reached. 25 / 25 I-Hsiang ang IT Lecture 9
Lecture 4 Noisy Channel Coding
Lecture 4 Noisy Channel Coding I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw October 9, 2015 1 / 56 I-Hsiang Wang IT Lecture 4 The Channel Coding Problem
More informationPractical Polar Code Construction Using Generalised Generator Matrices
Practical Polar Code Construction Using Generalised Generator Matrices Berksan Serbetci and Ali E. Pusane Department of Electrical and Electronics Engineering Bogazici University Istanbul, Turkey E-mail:
More informationLecture 4 Channel Coding
Capacity and the Weak Converse Lecture 4 Coding I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw October 15, 2014 1 / 16 I-Hsiang Wang NIT Lecture 4 Capacity
More informationDigital Communications III (ECE 154C) Introduction to Coding and Information Theory
Digital Communications III (ECE 154C) Introduction to Coding and Information Theory Tara Javidi These lecture notes were originally developed by late Prof. J. K. Wolf. UC San Diego Spring 2014 1 / 8 I
More informationChapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University
Chapter 4 Data Transmission and Channel Capacity Po-Ning Chen, Professor Department of Communications Engineering National Chiao Tung University Hsin Chu, Taiwan 30050, R.O.C. Principle of Data Transmission
More informationLecture 4: Proof of Shannon s theorem and an explicit code
CSE 533: Error-Correcting Codes (Autumn 006 Lecture 4: Proof of Shannon s theorem and an explicit code October 11, 006 Lecturer: Venkatesan Guruswami Scribe: Atri Rudra 1 Overview Last lecture we stated
More informationLecture 5 Channel Coding over Continuous Channels
Lecture 5 Channel Coding over Continuous Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 14, 2014 1 / 34 I-Hsiang Wang NIT Lecture 5 From
More informationLecture 8: Shannon s Noise Models
Error Correcting Codes: Combinatorics, Algorithms and Applications (Fall 2007) Lecture 8: Shannon s Noise Models September 14, 2007 Lecturer: Atri Rudra Scribe: Sandipan Kundu& Atri Rudra Till now we have
More informationThe Compound Capacity of Polar Codes
The Compound Capacity of Polar Codes S. Hamed Hassani, Satish Babu Korada and Rüdiger Urbanke arxiv:97.329v [cs.it] 9 Jul 29 Abstract We consider the compound capacity of polar codes under successive cancellation
More informationChapter 9 Fundamental Limits in Information Theory
Chapter 9 Fundamental Limits in Information Theory Information Theory is the fundamental theory behind information manipulation, including data compression and data transmission. 9.1 Introduction o For
More informationPolar Coding. Part 1 - Background. Erdal Arıkan. Electrical-Electronics Engineering Department, Bilkent University, Ankara, Turkey
Polar Coding Part 1 - Background Erdal Arıkan Electrical-Electronics Engineering Department, Bilkent University, Ankara, Turkey Algorithmic Coding Theory Workshop June 13-17, 2016 ICERM, Providence, RI
More informationBelief propagation decoding of quantum channels by passing quantum messages
Belief propagation decoding of quantum channels by passing quantum messages arxiv:67.4833 QIP 27 Joseph M. Renes lempelziv@flickr To do research in quantum information theory, pick a favorite text on classical
More informationChannel combining and splitting for cutoff rate improvement
Channel combining and splitting for cutoff rate improvement Erdal Arıkan Electrical-Electronics Engineering Department Bilkent University, Ankara, 68, Turkey Email: arikan@eebilkentedutr arxiv:cs/5834v
More informationAdvances in Error Control Strategies for 5G
Advances in Error Control Strategies for 5G Jörg Kliewer The Elisha Yegal Bar-Ness Center For Wireless Communications And Signal Processing Research 5G Requirements [Nokia Networks: Looking ahead to 5G.
More informationAn Alternative Proof of Channel Polarization for Channels with Arbitrary Input Alphabets
An Alternative Proof of Channel Polarization for Channels with Arbitrary Input Alphabets Jing Guo University of Cambridge jg582@cam.ac.uk Jossy Sayir University of Cambridge j.sayir@ieee.org Minghai Qin
More informationECEN 655: Advanced Channel Coding
ECEN 655: Advanced Channel Coding Course Introduction Henry D. Pfister Department of Electrical and Computer Engineering Texas A&M University ECEN 655: Advanced Channel Coding 1 / 19 Outline 1 History
More informationTurbo Compression. Andrej Rikovsky, Advisor: Pavol Hanus
Turbo Compression Andrej Rikovsky, Advisor: Pavol Hanus Abstract Turbo codes which performs very close to channel capacity in channel coding can be also used to obtain very efficient source coding schemes.
More informationNotes 3: Stochastic channels and noisy coding theorem bound. 1 Model of information communication and noisy channel
Introduction to Coding Theory CMU: Spring 2010 Notes 3: Stochastic channels and noisy coding theorem bound January 2010 Lecturer: Venkatesan Guruswami Scribe: Venkatesan Guruswami We now turn to the basic
More informationPerformance of Polar Codes for Channel and Source Coding
Performance of Polar Codes for Channel and Source Coding Nadine Hussami AUB, Lebanon, Email: njh03@aub.edu.lb Satish Babu Korada and üdiger Urbanke EPFL, Switzerland, Email: {satish.korada,ruediger.urbanke}@epfl.ch
More informationVariable Rate Channel Capacity. Jie Ren 2013/4/26
Variable Rate Channel Capacity Jie Ren 2013/4/26 Reference This is a introduc?on of Sergio Verdu and Shlomo Shamai s paper. Sergio Verdu and Shlomo Shamai, Variable- Rate Channel Capacity, IEEE Transac?ons
More informationConstructing Polar Codes Using Iterative Bit-Channel Upgrading. Arash Ghayoori. B.Sc., Isfahan University of Technology, 2011
Constructing Polar Codes Using Iterative Bit-Channel Upgrading by Arash Ghayoori B.Sc., Isfahan University of Technology, 011 A Thesis Submitted in Partial Fulfillment of the Requirements for the Degree
More informationLecture 6 I. CHANNEL CODING. X n (m) P Y X
6- Introduction to Information Theory Lecture 6 Lecturer: Haim Permuter Scribe: Yoav Eisenberg and Yakov Miron I. CHANNEL CODING We consider the following channel coding problem: m = {,2,..,2 nr} Encoder
More informationPolar Codes are Optimal for Lossy Source Coding
Polar Codes are Optimal for Lossy Source Coding Satish Babu Korada and Rüdiger Urbanke EPFL, Switzerland, Email: satish.korada,ruediger.urbanke}@epfl.ch Abstract We consider lossy source compression of
More informationLower Bounds on the Graphical Complexity of Finite-Length LDPC Codes
Lower Bounds on the Graphical Complexity of Finite-Length LDPC Codes Igal Sason Department of Electrical Engineering Technion - Israel Institute of Technology Haifa 32000, Israel 2009 IEEE International
More informationOn Bit Error Rate Performance of Polar Codes in Finite Regime
On Bit Error Rate Performance of Polar Codes in Finite Regime A. Eslami and H. Pishro-Nik Abstract Polar codes have been recently proposed as the first low complexity class of codes that can provably achieve
More informationMidterm Exam Information Theory Fall Midterm Exam. Time: 09:10 12:10 11/23, 2016
Midterm Exam Time: 09:10 12:10 11/23, 2016 Name: Student ID: Policy: (Read before You Start to Work) The exam is closed book. However, you are allowed to bring TWO A4-size cheat sheet (single-sheet, two-sided).
More informationChannel Polarization and Blackwell Measures
Channel Polarization Blackwell Measures Maxim Raginsky Abstract The Blackwell measure of a binary-input channel (BIC is the distribution of the posterior probability of 0 under the uniform input distribution
More informationLecture 8: Channel Capacity, Continuous Random Variables
EE376A/STATS376A Information Theory Lecture 8-02/0/208 Lecture 8: Channel Capacity, Continuous Random Variables Lecturer: Tsachy Weissman Scribe: Augustine Chemparathy, Adithya Ganesh, Philip Hwang Channel
More informationPolar Codes: Construction and Performance Analysis
Master Project Polar Codes: Construction and Performance Analysis Ramtin Pedarsani ramtin.pedarsani@epfl.ch Supervisor: Prof. Emre Telatar Assistant: Hamed Hassani Information Theory Laboratory (LTHI)
More informationLecture 8: Channel and source-channel coding theorems; BEC & linear codes. 1 Intuitive justification for upper bound on channel capacity
5-859: Information Theory and Applications in TCS CMU: Spring 23 Lecture 8: Channel and source-channel coding theorems; BEC & linear codes February 7, 23 Lecturer: Venkatesan Guruswami Scribe: Dan Stahlke
More informationRCA Analysis of the Polar Codes and the use of Feedback to aid Polarization at Short Blocklengths
RCA Analysis of the Polar Codes and the use of Feedback to aid Polarization at Short Blocklengths Kasra Vakilinia, Dariush Divsalar*, and Richard D. Wesel Department of Electrical Engineering, University
More informationOne Lesson of Information Theory
Institut für One Lesson of Information Theory Prof. Dr.-Ing. Volker Kühn Institute of Communications Engineering University of Rostock, Germany Email: volker.kuehn@uni-rostock.de http://www.int.uni-rostock.de/
More informationEE229B - Final Project. Capacity-Approaching Low-Density Parity-Check Codes
EE229B - Final Project Capacity-Approaching Low-Density Parity-Check Codes Pierre Garrigues EECS department, UC Berkeley garrigue@eecs.berkeley.edu May 13, 2005 Abstract The class of low-density parity-check
More informationPolar Codes are Optimal for Lossy Source Coding
Polar Codes are Optimal for Lossy Source Coding Satish Babu Korada and Rüdiger Urbanke Abstract We consider lossy source compression of a binary symmetric source using polar codes and the low-complexity
More informationOn the Polarization Levels of Automorphic-Symmetric Channels
On the Polarization Levels of Automorphic-Symmetric Channels Rajai Nasser Email: rajai.nasser@alumni.epfl.ch arxiv:8.05203v2 [cs.it] 5 Nov 208 Abstract It is known that if an Abelian group operation is
More informationLecture 3: Channel Capacity
Lecture 3: Channel Capacity 1 Definitions Channel capacity is a measure of maximum information per channel usage one can get through a channel. This one of the fundamental concepts in information theory.
More informationLecture 4 : Adaptive source coding algorithms
Lecture 4 : Adaptive source coding algorithms February 2, 28 Information Theory Outline 1. Motivation ; 2. adaptive Huffman encoding ; 3. Gallager and Knuth s method ; 4. Dictionary methods : Lempel-Ziv
More informationfor some error exponent E( R) as a function R,
. Capacity-achieving codes via Forney concatenation Shannon s Noisy Channel Theorem assures us the existence of capacity-achieving codes. However, exhaustive search for the code has double-exponential
More informationDiscrete Memoryless Channels with Memoryless Output Sequences
Discrete Memoryless Channels with Memoryless utput Sequences Marcelo S Pinho Department of Electronic Engineering Instituto Tecnologico de Aeronautica Sao Jose dos Campos, SP 12228-900, Brazil Email: mpinho@ieeeorg
More informationLecture 5: Channel Capacity. Copyright G. Caire (Sample Lectures) 122
Lecture 5: Channel Capacity Copyright G. Caire (Sample Lectures) 122 M Definitions and Problem Setup 2 X n Y n Encoder p(y x) Decoder ˆM Message Channel Estimate Definition 11. Discrete Memoryless Channel
More informationError Detection and Correction: Hamming Code; Reed-Muller Code
Error Detection and Correction: Hamming Code; Reed-Muller Code Greg Plaxton Theory in Programming Practice, Spring 2005 Department of Computer Science University of Texas at Austin Hamming Code: Motivation
More informationFountain Uncorrectable Sets and Finite-Length Analysis
Fountain Uncorrectable Sets and Finite-Length Analysis Wen Ji 1, Bo-Wei Chen 2, and Yiqiang Chen 1 1 Beijing Key Laboratory of Mobile Computing and Pervasive Device Institute of Computing Technology, Chinese
More informationNoisy channel communication
Information Theory http://www.inf.ed.ac.uk/teaching/courses/it/ Week 6 Communication channels and Information Some notes on the noisy channel setup: Iain Murray, 2012 School of Informatics, University
More informationLow-density parity-check codes
Low-density parity-check codes From principles to practice Dr. Steve Weller steven.weller@newcastle.edu.au School of Electrical Engineering and Computer Science The University of Newcastle, Callaghan,
More informationBounds on Achievable Rates of LDPC Codes Used Over the Binary Erasure Channel
Bounds on Achievable Rates of LDPC Codes Used Over the Binary Erasure Channel Ohad Barak, David Burshtein and Meir Feder School of Electrical Engineering Tel-Aviv University Tel-Aviv 69978, Israel Abstract
More informationOn the Typicality of the Linear Code Among the LDPC Coset Code Ensemble
5 Conference on Information Sciences and Systems The Johns Hopkins University March 16 18 5 On the Typicality of the Linear Code Among the LDPC Coset Code Ensemble C.-C. Wang S.R. Kulkarni and H.V. Poor
More informationHow to Achieve the Capacity of Asymmetric Channels
How to Achieve the Capacity of Asymmetric Channels Marco Mondelli, S. Hamed Hassani, and Rüdiger Urbanke Abstract arxiv:406.7373v5 [cs.it] 3 Jan 208 We survey coding techniques that enable reliable transmission
More informationSuccessive Cancellation Decoding of Single Parity-Check Product Codes
Successive Cancellation Decoding of Single Parity-Check Product Codes Mustafa Cemil Coşkun, Gianluigi Liva, Alexandre Graell i Amat and Michael Lentmaier Institute of Communications and Navigation, German
More informationShannon s noisy-channel theorem
Shannon s noisy-channel theorem Information theory Amon Elders Korteweg de Vries Institute for Mathematics University of Amsterdam. Tuesday, 26th of Januari Amon Elders (Korteweg de Vries Institute for
More informationLecture 12. Block Diagram
Lecture 12 Goals Be able to encode using a linear block code Be able to decode a linear block code received over a binary symmetric channel or an additive white Gaussian channel XII-1 Block Diagram Data
More information5. Density evolution. Density evolution 5-1
5. Density evolution Density evolution 5-1 Probabilistic analysis of message passing algorithms variable nodes factor nodes x1 a x i x2 a(x i ; x j ; x k ) x3 b x4 consider factor graph model G = (V ;
More informationEE376A - Information Theory Final, Monday March 14th 2016 Solutions. Please start answering each question on a new page of the answer booklet.
EE376A - Information Theory Final, Monday March 14th 216 Solutions Instructions: You have three hours, 3.3PM - 6.3PM The exam has 4 questions, totaling 12 points. Please start answering each question on
More informationUnified Scaling of Polar Codes: Error Exponent, Scaling Exponent, Moderate Deviations, and Error Floors
Unified Scaling of Polar Codes: Error Exponent, Scaling Exponent, Moderate Deviations, and Error Floors Marco Mondelli, S. Hamed Hassani, and Rüdiger Urbanke Abstract Consider transmission of a polar code
More informationPOLAR CODES FOR ERROR CORRECTION: ANALYSIS AND DECODING ALGORITHMS
ALMA MATER STUDIORUM UNIVERSITÀ DI BOLOGNA CAMPUS DI CESENA SCUOLA DI INGEGNERIA E ARCHITETTURA CORSO DI LAUREA MAGISTRALE IN INGEGNERIA ELETTRONICA E TELECOMUNICAZIONI PER L ENERGIA POLAR CODES FOR ERROR
More informationCodes on graphs and iterative decoding
Codes on graphs and iterative decoding Bane Vasić Error Correction Coding Laboratory University of Arizona Funded by: National Science Foundation (NSF) Seagate Technology Defense Advanced Research Projects
More informationInvestigation of the Elias Product Code Construction for the Binary Erasure Channel
Investigation of the Elias Product Code Construction for the Binary Erasure Channel by D. P. Varodayan A THESIS SUBMITTED IN PARTIAL FULFILLMENT OF THE REQUIREMENTS FOR THE DEGREE OF BACHELOR OF APPLIED
More informationOn Generalized EXIT Charts of LDPC Code Ensembles over Binary-Input Output-Symmetric Memoryless Channels
2012 IEEE International Symposium on Information Theory Proceedings On Generalied EXIT Charts of LDPC Code Ensembles over Binary-Input Output-Symmetric Memoryless Channels H Mamani 1, H Saeedi 1, A Eslami
More information1 Background on Information Theory
Review of the book Information Theory: Coding Theorems for Discrete Memoryless Systems by Imre Csiszár and János Körner Second Edition Cambridge University Press, 2011 ISBN:978-0-521-19681-9 Review by
More informationExact Probability of Erasure and a Decoding Algorithm for Convolutional Codes on the Binary Erasure Channel
Exact Probability of Erasure and a Decoding Algorithm for Convolutional Codes on the Binary Erasure Channel Brian M. Kurkoski, Paul H. Siegel, and Jack K. Wolf Department of Electrical and Computer Engineering
More informationLecture 18: Shanon s Channel Coding Theorem. Lecture 18: Shanon s Channel Coding Theorem
Channel Definition (Channel) A channel is defined by Λ = (X, Y, Π), where X is the set of input alphabets, Y is the set of output alphabets and Π is the transition probability of obtaining a symbol y Y
More informationOn Fast Channel Polarization of Double-layer Binary Discrete Memoryless Channels
Appl. Math. Inf. Sci. 9, o., 037-06 (05) 037 Applied Mathematics & Information Sciences An International Journal http://dx.doi.org/0.785/amis/0903 On Fast Channel Polarization of Double-layer Binary Discrete
More informationAPPLICATIONS. Quantum Communications
SOFT PROCESSING TECHNIQUES FOR QUANTUM KEY DISTRIBUTION APPLICATIONS Marina Mondin January 27, 2012 Quantum Communications In the past decades, the key to improving computer performance has been the reduction
More informationEE5139R: Problem Set 7 Assigned: 30/09/15, Due: 07/10/15
EE5139R: Problem Set 7 Assigned: 30/09/15, Due: 07/10/15 1. Cascade of Binary Symmetric Channels The conditional probability distribution py x for each of the BSCs may be expressed by the transition probability
More informationReliable Computation over Multiple-Access Channels
Reliable Computation over Multiple-Access Channels Bobak Nazer and Michael Gastpar Dept. of Electrical Engineering and Computer Sciences University of California, Berkeley Berkeley, CA, 94720-1770 {bobak,
More informationDistributed Arithmetic Coding
Distributed Arithmetic Coding Marco Grangetto, Member, IEEE, Enrico Magli, Member, IEEE, Gabriella Olmo, Senior Member, IEEE Abstract We propose a distributed binary arithmetic coder for Slepian-Wolf coding
More informationLecture 11: Polar codes construction
15-859: Information Theory and Applications in TCS CMU: Spring 2013 Lecturer: Venkatesan Guruswami Lecture 11: Polar codes construction February 26, 2013 Scribe: Dan Stahlke 1 Polar codes: recap of last
More informationTime-invariant LDPC convolutional codes
Time-invariant LDPC convolutional codes Dimitris Achlioptas, Hamed Hassani, Wei Liu, and Rüdiger Urbanke Department of Computer Science, UC Santa Cruz, USA Email: achlioptas@csucscedu Department of Computer
More informationCSCI 2570 Introduction to Nanocomputing
CSCI 2570 Introduction to Nanocomputing Information Theory John E Savage What is Information Theory Introduced by Claude Shannon. See Wikipedia Two foci: a) data compression and b) reliable communication
More informationDesign and Decoding of Polar Codes
TECHICAL UIVERSITY OF CRETE Department of Electronic and Computer Engineering Design and Decoding of Polar Codes By: ikolaos Tsagkarakis Submitted on July 6th, 0 in partial fulfillment of the requirements
More informationBelief-Propagation Decoding of LDPC Codes
LDPC Codes: Motivation Belief-Propagation Decoding of LDPC Codes Amir Bennatan, Princeton University Revolution in coding theory Reliable transmission, rates approaching capacity. BIAWGN, Rate =.5, Threshold.45
More informationAn Introduction to Low-Density Parity-Check Codes
An Introduction to Low-Density Parity-Check Codes Paul H. Siegel Electrical and Computer Engineering University of California, San Diego 5/ 3/ 7 Copyright 27 by Paul H. Siegel Outline Shannon s Channel
More informationLecture 6 Channel Coding over Continuous Channels
Lecture 6 Channel Coding over Continuous Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 9, 015 1 / 59 I-Hsiang Wang IT Lecture 6 We have
More information1 Ex. 1 Verify that the function H(p 1,..., p n ) = k p k log 2 p k satisfies all 8 axioms on H.
Problem sheet Ex. Verify that the function H(p,..., p n ) = k p k log p k satisfies all 8 axioms on H. Ex. (Not to be handed in). looking at the notes). List as many of the 8 axioms as you can, (without
More informationBasic Principles of Lossless Coding. Universal Lossless coding. Lempel-Ziv Coding. 2. Exploit dependences between successive symbols.
Universal Lossless coding Lempel-Ziv Coding Basic principles of lossless compression Historical review Variable-length-to-block coding Lempel-Ziv coding 1 Basic Principles of Lossless Coding 1. Exploit
More informationPolar codes for reliable transmission
Polar codes for reliable transmission Theoretical analysis and applications Jing Guo Department of Engineering University of Cambridge Supervisor: Prof. Albert Guillén i Fàbregas This dissertation is submitted
More informationEC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY
EC2252 COMMUNICATION THEORY UNIT 5 INFORMATION THEORY Discrete Messages and Information Content, Concept of Amount of Information, Average information, Entropy, Information rate, Source coding to increase
More informationEfficient Computation of EXIT Functions for Non-Binary Iterative Decoding
TO BE PUBLISHED IN IEEE TRANSACTIONS ON COMMUNCATIONS, DECEMBER 2006 Efficient Computation of EXIT Functions for Non-Binary Iterative Decoding Jörg Kliewer, Senior Member, IEEE, Soon Xin Ng, Member, IEEE,
More informationEfficiently decodable codes for the binary deletion channel
Efficiently decodable codes for the binary deletion channel Venkatesan Guruswami (venkatg@cs.cmu.edu) Ray Li * (rayyli@stanford.edu) Carnegie Mellon University August 18, 2017 V. Guruswami and R. Li (CMU)
More informationCodes on graphs and iterative decoding
Codes on graphs and iterative decoding Bane Vasić Error Correction Coding Laboratory University of Arizona Prelude Information transmission 0 0 0 0 0 0 Channel Information transmission signal 0 0 threshold
More informationSpatially Coupled LDPC Codes
Spatially Coupled LDPC Codes Kenta Kasai Tokyo Institute of Technology 30 Aug, 2013 We already have very good codes. Efficiently-decodable asymptotically capacity-approaching codes Irregular LDPC Codes
More informationPolar Codes: Graph Representation and Duality
Polar Codes: Graph Representation and Duality arxiv:1312.0372v1 [cs.it] 2 Dec 2013 M. Fossorier ETIS ENSEA/UCP/CNRS UMR-8051 6, avenue du Ponceau, 95014, Cergy Pontoise, France Email: mfossorier@ieee.org
More informationAppendix B Information theory from first principles
Appendix B Information theory from first principles This appendix discusses the information theory behind the capacity expressions used in the book. Section 8.3.4 is the only part of the book that supposes
More informationThreshold Saturation on Channels with Memory via Spatial Coupling
Threshold Saturation on Channels with Memory via Spatial Coupling Shrinivas Kudekar and Kenta Kasai New Mexico Consortium and Center for Non-linear Studies, Los Alamos National Laboratory, NM, USA Email:
More information(Classical) Information Theory III: Noisy channel coding
(Classical) Information Theory III: Noisy channel coding Sibasish Ghosh The Institute of Mathematical Sciences CIT Campus, Taramani, Chennai 600 113, India. p. 1 Abstract What is the best possible way
More informationCoding Techniques for Data Storage Systems
Coding Techniques for Data Storage Systems Thomas Mittelholzer IBM Zurich Research Laboratory /8 Göttingen Agenda. Channel Coding and Practical Coding Constraints. Linear Codes 3. Weight Enumerators and
More informationarxiv: v3 [cs.it] 1 Apr 2014
Universal Polar Codes for More Capable and Less Noisy Channels and Sources David Sutter and Joseph M. Renes Institute for Theoretical Physics, ETH Zurich, Switzerland arxiv:32.5990v3 [cs.it] Apr 204 We
More informationECE Advanced Communication Theory, Spring 2009 Homework #1 (INCOMPLETE)
ECE 74 - Advanced Communication Theory, Spring 2009 Homework #1 (INCOMPLETE) 1. A Huffman code finds the optimal codeword to assign to a given block of source symbols. (a) Show that cannot be a Huffman
More informationCh 0 Introduction. 0.1 Overview of Information Theory and Coding
Ch 0 Introduction 0.1 Overview of Information Theory and Coding Overview The information theory was founded by Shannon in 1948. This theory is for transmission (communication system) or recording (storage
More informationNonlinear Codes Outperform the Best Linear Codes on the Binary Erasure Channel
Nonear odes Outperform the Best Linear odes on the Binary Erasure hannel Po-Ning hen, Hsuan-Yin Lin Department of Electrical and omputer Engineering National hiao-tung University NTU Hsinchu, Taiwan poning@faculty.nctu.edu.tw,.hsuanyin@ieee.org
More informationConstruction of Polar Codes with Sublinear Complexity
1 Construction of Polar Codes with Sublinear Complexity Marco Mondelli, S. Hamed Hassani, and Rüdiger Urbanke arxiv:1612.05295v4 [cs.it] 13 Jul 2017 Abstract Consider the problem of constructing a polar
More informationCompound Polar Codes
Compound Polar Codes Hessam Mahdavifar, Mostafa El-Khamy, Jungwon Lee, Inyup Kang Mobile Solutions Lab, Samsung Information Systems America 4921 Directors Place, San Diego, CA 92121 {h.mahdavifar, mostafa.e,
More informationPerformance-based Security for Encoding of Information Signals. FA ( ) Paul Cuff (Princeton University)
Performance-based Security for Encoding of Information Signals FA9550-15-1-0180 (2015-2018) Paul Cuff (Princeton University) Contributors Two students finished PhD Tiance Wang (Goldman Sachs) Eva Song
More informationCS 229r Information Theory in Computer Science Feb 12, Lecture 5
CS 229r Information Theory in Computer Science Feb 12, 2019 Lecture 5 Instructor: Madhu Sudan Scribe: Pranay Tankala 1 Overview A universal compression algorithm is a single compression algorithm applicable
More information4216 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 51, NO. 12, DECEMBER Density Evolution for Asymmetric Memoryless Channels
4216 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 51, NO. 12, DECEMBER 2005 Density Evolution for Asymmetric Memoryless Channels Chih-Chun Wang, Sanjeev R. Kulkarni, Fellow, IEEE, and H. Vincent Poor,
More informationOptimal Rate and Maximum Erasure Probability LDPC Codes in Binary Erasure Channel
Optimal Rate and Maximum Erasure Probability LDPC Codes in Binary Erasure Channel H. Tavakoli Electrical Engineering Department K.N. Toosi University of Technology, Tehran, Iran tavakoli@ee.kntu.ac.ir
More informationIterative Encoder-Controller Design for Feedback Control Over Noisy Channels
IEEE TRANSACTIONS ON AUTOMATIC CONTROL 1 Iterative Encoder-Controller Design for Feedback Control Over Noisy Channels Lei Bao, Member, IEEE, Mikael Skoglund, Senior Member, IEEE, and Karl Henrik Johansson,
More informationPermuted Successive Cancellation Decoder for Polar Codes
Permuted Successive Cancellation Decoder for Polar Codes Harish Vangala, Emanuele Viterbo, and Yi Hong, Dept. of ECSE, Monash University, Melbourne, VIC 3800, Australia. Email: {harish.vangala, emanuele.viterbo,
More informationAsynchronous Decoding of LDPC Codes over BEC
Decoding of LDPC Codes over BEC Saeid Haghighatshoar, Amin Karbasi, Amir Hesam Salavati Department of Telecommunication Systems, Technische Universität Berlin, Germany, School of Engineering and Applied
More informationRevision of Lecture 5
Revision of Lecture 5 Information transferring across channels Channel characteristics and binary symmetric channel Average mutual information Average mutual information tells us what happens to information
More informationThe Least Degraded and the Least Upgraded Channel with respect to a Channel Family
The Least Degraded and the Least Upgraded Channel with respect to a Channel Family Wei Liu, S. Hamed Hassani, and Rüdiger Urbanke School of Computer and Communication Sciences EPFL, Switzerland Email:
More information