List Decoding: Geometrical Aspects and Performance Bounds
|
|
- August Hines
- 5 years ago
- Views:
Transcription
1 List Decoding: Geometrical Aspects and Performance Bounds Maja Lončar Department of Information Technology Lund University, Sweden Summer Academy: Progress in Mathematics for Communication Systems Bremen, Germany, July 007
2 Outline Maximum-likelihood decoding List decoding List configuration matrix List distance List error probability for a worst-case list New bound for the list decoding error probability Summary
3 Maximum-Likelihood Decoding Let S = {s 0, s 1,..., s M 1 } be a set of M distinct signal vectors used to communicate over the AWGN channel. Let s S denote the transmit signal. Then the received signal is r = s + n. Optimum decision strategy at the receiver: minimize the error probability { } { ŝ MAP =arg min Pr(ŝ s) =arg max Pr(ŝ = s) ŝ S ŝ S that is, for every r decide in favour the most probable signal s S } ŝ MAP = arg max s S {p(s r)} This is the maximum a posteriori probability decoder. { } =arg max Pr(ŝ = s r)p(r) ŝ S r From Bayes formula we have { p(r s)p(s) } ŝ MAP = arg max s S p(r) { } = arg max p(r s)p(s) s S
4 Maximum-Likelihood Decoding If the signals s S are a priori equiprobable: p(s) = 1/M. Then the MAP decoder reduces to ŝ ML = arg max {p(r s)} s S This is the maximum likelihood decoder. For the AWGN channel p(r s) e 1 N 0 r s ŝ ML = arg min s S { r s } ML decoder is the minimum Euclidean distance decoder. It chooses the point with the smallest d E (r, s) = r s. Decoding error occurs if d E (r, s i ) d E (r, s 0 ), for some i. Let ε i s 0 denote this event. Union bound on the ML error probability: M 1 P e s0 Pr(ε i s 0 ) i=1 d E (r, s 1 ) s 0 d E (r, s ) s 1 s d E (r, s 0 ) r d E (r, s 3 ) s 3
5 List Decoding Generalization of ML decoding for L 1 most likely codewords. List decoder finds a list of the L best codewords. For Gaussian channel, these are L codewords s i closest to the received vector r. List decoding error occurs if transmitted signal s 0 is not on the list, that is, if, d E (r, s i ) d E (r, s 0 ), i = 1,,..., L or, equivalently, if projections of the noise n along (s i s 0 ) are larger than d E (s i, s 0 )/ n, s i s 0 d }{{} E(s i, s 0 )/, i = 1,,..., L }{{} t i d E0i / that is, (t 1 t... t L ) (d E01 d E0... d E0L )/ t w/ d E (r, s 1 ) s 0 d E (r, s ) s 1 s d E (r, s 0 ) r d E (r, s 3 ) s 3
6 List Configuration Matrix Components of t, n, s i s 0, are Gaussian distributed with covariance matrix E[t T t] = σ W where W is the Gram matrix of the vectors (s i s 0 ), i = 1,,..., L with elements w ij = s i s 0, s j s 0 = (d E0i + d E0j d Eij )/ W = d E01 (d E01 + d E0 d E1 )/... (d E01 + d E0L d E1L )/ (d E01 + d E0 d E1 )/ d E0... (d E0 + d E0L d EL )/ (d E01 + d E0L d E1L )/ (d E0 + d E0L d EL )/... d E0L W is a list configuration matrix. It determines the list error probability for a given list: P el (W ) = Pr(t w/), Union-type bound for list decoding error probability: P el N(W )P el (W ) W
7 List Distance List radius R L for a given signal set S L = {s 0, s 1,..., s L } and a given reference (transmit) signal s 0, is the point of the list-error region that is closest to s 0. This is a radius of the smallest sphere S that contains (encompasses) all the signal points from S L (points on or inside the sphere). Euclidean list distance for a given signal set S L is d EL = R L. Theorem 1: The radius R L of the sphere S that passes through all the points of S L (all the points on the sphere) is given by R L (W ) = 1 ww 1 w T Theorem : The list radius R L of the smallest sphere S that passes through s 0 and encompasses the points s i, i = 1,,..., L, is given by { } 1 R L (W ) = max w I W 1 I I : w I adj(w I ) 0 wt I
8 List Distance Example: L = s 0 = (0 0) s 1 = (0 ) s = (1 3) ( ) 4 6 W = 6 10 R L = 1 ww 1 w T = s 1 R L s RL list error region w adj(w )=( 0 16) R L = 10/ s 0 1 d EL =R L = 10
9 List Distance Example: L = 3 s s 1 = (0 0) s 0 = (0 ) s = (1 3) ( ) 4 W = 1 s 0 R L = 1 ww 1 w T = 5 R L w adj(w )=(1 16) R L = R L = 5 s 1 1 d EL =R L = 0
10 Minimum List Distance Minimum list radius R L for a list size L is R L min = min {R L(W )} = min W W max I : w I adj(w I ) 0 { 1 } w I W 1 I wt I Minimum Euclidean list distance for a list size L is d ELmin = R L min. For binary linear codes and BPSK signaling, d Eij = 4E sd Hij. Minimum Hamming list distance of a code, for list size L is d HLmin = d ELmin /(4E s). Minimum list distance determines the performance of the list decoder at higher SNR, in the same way as the minimum distance determines the performance of the ML decoder. Theorem 3: For any binary linear code the minimum Hamming list distance is d HLmin L L + 1 For any binary linear code with odd, we have d HLmin L L L 1 L + 1
11 Minimum List Distance For even Worst-case list configuration yielding minimum list distance d HLmin = L L + 1 consists of L codewords of weight at pairwise distances : W H = The codewords form an L-dimensional simplex
12 Minimum List Distance Example: even L = R L min = 1 3 d Emin d ELmin = 3 d Emin s 1 s R L min d E min s 0
13 Minimum List Distance For odd Worst-case list configuration yielding minimum list distance consists of d HLmin = L L L 1 L + 1 (L + 1)/ codewords of weight (L 1)/ codewords of weight + 1 at pairwise distances and + 1. For example, for L = 5: W H = d Hmin 1 d 1 +1 Hmin 1 1 d +1 Hmin
14 List Error Probability for Worst-Case List Worst-case list configuration determines the performance of list decoder at high SNR Therefore, we want to estimate the list error probability Pr(t w/) for the worst-case list configuration Problem: Finding Pr(t w/) involves L-dimensional integration over the PDF of t Solution: Consider instead the orthogonalized vector q = tv The components of q are uncorrelated and its PDF breaks up into a product of L 1-dimensional PDFs By estimating the integration limits for q we obtain an upper bound on Pr(t w/).
15 List Error Probability for Worst-Case List For even Lemma 1: The list-error probability Pr(t w/), for a worst-case list configuration, for a code with even is upper-bounded by ( L ) ul (y) Pr(t w/) f(x)dx dy f(y) αl σ1 l= v l (y) with equality for L. f(x) and f(y) denote Gaussian N(0, 1) PDF. The integration intervals are determined by the eigenvalues of W. For odd Lemma : The list-error probability Pr(t w/), for a worst-case list configuration, for a code with odd is upper-bounded by Pr(t w/) φ(α,η) σl f(y 1 ) h(y 1 ) g(y 1 ) f(y )dy L 1 l=1 u l (y 1 ) v l (y 1 ) f(x)dx L l= L+1 z l (y 1 ) w l (y 1 ) f(x)dx dy 1
16 Generalized Tangential Bound on List Decoding Error Probability We want to improve the union bound on the list-error probability P el for binary codes Tangential-bound approach: Decompose noise vector n into one radial component x, along transmitted signal s 0 : x = n, s 0 L components y l orthogonal to the radial component: y l = n, s l (1 d H0l /N)x, l = 1,,..., L List decoding error probability P el Pr(ε) is upper-bounded by (few many errors) This yields P el min T P el = Pr (ε, x T) + Pr (ε, x > T) T T Pr (ε x) f(x)dx + Pr (x > T) min 1, N(K y )P el (K y, x) f(x)dx + Q(T) K y
17 Generalized Tangential Bound on List Decoding Error Probability New generalized tangential union bound for list decoding error probability { } P el min T T min 1, N(K y )P el (K y, x)+ W N(W )P el (d HL (W ), x) f(x)dx+q(t) The dominant term is estimated using the upper bound on worst-case-list error probability The remaining terms are upper-bounded only by using the list distance d HL (W ), which is equivalent to replacing the codeword sets with list configuration matrix W by an average codeword s at distance d HL (W ) from the transmitted codeword. s1 R L s D s D s 0
18 Generalized Tangential Bound on List Decoding Error Probability Comparison of bounds for (4, 1, 8) Golay code with list size L = P el Simulations Union bound Tangential bound New bound E b /N 0 [db]
19 Summary List configuration matrix describes the list geometry and list error probability List radius and list distance are defined via the list configuration matrix Minimum list distance is determined by the worst-case list configuration Upper bound on the list error probability for the worst-case lists is derived New generalized tangential union bound for the list decoding error probability is derived
An improved bound on the list error probability and list distance properties. Bocharova, Irina; Kudryashov, Boris; Johannesson, Rolf; Loncar, Maja
An improved bound on the list error probability and list distance properties Bocharova, Irina; Kudryashov, Boris; Johannesson, Rolf; Loncar, Maja Published in: IEEE Transactions on Information Theory DOI:
More informationA Simple Example Binary Hypothesis Testing Optimal Receiver Frontend M-ary Signal Sets Message Sequences. possible signals has been transmitted.
Introduction I We have focused on the problem of deciding which of two possible signals has been transmitted. I Binary Signal Sets I We will generalize the design of optimum (MPE) receivers to signal sets
More informationSIPCom8-1: Information Theory and Coding Linear Binary Codes Ingmar Land
SIPCom8-1: Information Theory and Coding Linear Binary Codes Ingmar Land Ingmar Land, SIPCom8-1: Information Theory and Coding (2005 Spring) p.1 Overview Basic Concepts of Channel Coding Block Codes I:
More informationThese outputs can be written in a more convenient form: with y(i) = Hc m (i) n(i) y(i) = (y(i); ; y K (i)) T ; c m (i) = (c m (i); ; c m K(i)) T and n
Binary Codes for synchronous DS-CDMA Stefan Bruck, Ulrich Sorger Institute for Network- and Signal Theory Darmstadt University of Technology Merckstr. 25, 6428 Darmstadt, Germany Tel.: 49 65 629, Fax:
More informationBASICS OF DETECTION AND ESTIMATION THEORY
BASICS OF DETECTION AND ESTIMATION THEORY 83050E/158 In this chapter we discuss how the transmitted symbols are detected optimally from a noisy received signal (observation). Based on these results, optimal
More informationOptimum Soft Decision Decoding of Linear Block Codes
Optimum Soft Decision Decoding of Linear Block Codes {m i } Channel encoder C=(C n-1,,c 0 ) BPSK S(t) (n,k,d) linear modulator block code Optimal receiver AWGN Assume that [n,k,d] linear block code C is
More informationECE 564/645 - Digital Communications, Spring 2018 Homework #2 Due: March 19 (In Lecture)
ECE 564/645 - Digital Communications, Spring 018 Homework # Due: March 19 (In Lecture) 1. Consider a binary communication system over a 1-dimensional vector channel where message m 1 is sent by signaling
More informationDigital Transmission Methods S
Digital ransmission ethods S-7.5 Second Exercise Session Hypothesis esting Decision aking Gram-Schmidt method Detection.K.K. Communication Laboratory 5//6 Konstantinos.koufos@tkk.fi Exercise We assume
More informationLecture 12. Block Diagram
Lecture 12 Goals Be able to encode using a linear block code Be able to decode a linear block code received over a binary symmetric channel or an additive white Gaussian channel XII-1 Block Diagram Data
More informationTightened Upper Bounds on the ML Decoding Error Probability of Binary Linear Block Codes and Applications
on the ML Decoding Error Probability of Binary Linear Block Codes and Department of Electrical Engineering Technion-Israel Institute of Technology An M.Sc. Thesis supervisor: Dr. Igal Sason March 30, 2006
More informationUTA EE5362 PhD Diagnosis Exam (Spring 2011)
EE5362 Spring 2 PhD Diagnosis Exam ID: UTA EE5362 PhD Diagnosis Exam (Spring 2) Instructions: Verify that your exam contains pages (including the cover shee. Some space is provided for you to show your
More informationChapter 7: Channel coding:convolutional codes
Chapter 7: : Convolutional codes University of Limoges meghdadi@ensil.unilim.fr Reference : Digital communications by John Proakis; Wireless communication by Andreas Goldsmith Encoder representation Communication
More information16.36 Communication Systems Engineering
MIT OpenCourseWare http://ocw.mit.edu 16.36 Communication Systems Engineering Spring 2009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 16.36: Communication
More informationLecture 6 Channel Coding over Continuous Channels
Lecture 6 Channel Coding over Continuous Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 9, 015 1 / 59 I-Hsiang Wang IT Lecture 6 We have
More informationPerformance of small signal sets
42 Chapter 5 Performance of small signal sets In this chapter, we show how to estimate the performance of small-to-moderate-sized signal constellations on the discrete-time AWGN channel. With equiprobable
More informationIntroduction to Low-Density Parity Check Codes. Brian Kurkoski
Introduction to Low-Density Parity Check Codes Brian Kurkoski kurkoski@ice.uec.ac.jp Outline: Low Density Parity Check Codes Review block codes History Low Density Parity Check Codes Gallager s LDPC code
More informationEE4304 C-term 2007: Lecture 17 Supplemental Slides
EE434 C-term 27: Lecture 17 Supplemental Slides D. Richard Brown III Worcester Polytechnic Institute, Department of Electrical and Computer Engineering February 5, 27 Geometric Representation: Optimal
More informationEstimation techniques
Estimation techniques March 2, 2006 Contents 1 Problem Statement 2 2 Bayesian Estimation Techniques 2 2.1 Minimum Mean Squared Error (MMSE) estimation........................ 2 2.1.1 General formulation......................................
More informationDigital Modulation 1
Digital Modulation 1 Lecture Notes Ingmar Land and Bernard H. Fleury Navigation and Communications () Department of Electronic Systems Aalborg University, DK Version: February 5, 27 i Contents I Basic
More informationIntroduction to Convolutional Codes, Part 1
Introduction to Convolutional Codes, Part 1 Frans M.J. Willems, Eindhoven University of Technology September 29, 2009 Elias, Father of Coding Theory Textbook Encoder Encoder Properties Systematic Codes
More informationDecoding the Tail-Biting Convolutional Codes with Pre-Decoding Circular Shift
Decoding the Tail-Biting Convolutional Codes with Pre-Decoding Circular Shift Ching-Yao Su Directed by: Prof. Po-Ning Chen Department of Communications Engineering, National Chiao-Tung University July
More informationApplications of Lattices in Telecommunications
Applications of Lattices in Telecommunications Dept of Electrical and Computer Systems Engineering Monash University amin.sakzad@monash.edu Oct. 2013 1 Sphere Decoder Algorithm Rotated Signal Constellations
More informationChapter 7. Error Control Coding. 7.1 Historical background. Mikael Olofsson 2005
Chapter 7 Error Control Coding Mikael Olofsson 2005 We have seen in Chapters 4 through 6 how digital modulation can be used to control error probabilities. This gives us a digital channel that in each
More informationProblem Set 7 Due March, 22
EE16: Probability and Random Processes SP 07 Problem Set 7 Due March, Lecturer: Jean C. Walrand GSI: Daniel Preda, Assane Gueye Problem 7.1. Let u and v be independent, standard normal random variables
More information392D: Coding for the AWGN Channel Wednesday, January 24, 2007 Stanford, Winter 2007 Handout #6. Problem Set 2 Solutions
392D: Coding for the AWGN Channel Wednesday, January 24, 2007 Stanford, Winter 2007 Handout #6 Problem Set 2 Solutions Problem 2.1 (Cartesian-product constellations) (a) Show that if A is a K-fold Cartesian
More informationPerformance Enhancements for Algebraic Soft Decision Decoding of Reed-Solomon Codes
Performance Enhancements for Algebraic Soft Decision Decoding of Reed-Solomon Codes ISIT 2004 Mostafa El-Khamy, Robert J. McEliece, and Jonathan Harel fmostafa, rjm, harelg@systems.caltech.edu California
More informationChapter 2 Signal Processing at Receivers: Detection Theory
Chapter Signal Processing at Receivers: Detection Theory As an application of the statistical hypothesis testing, signal detection plays a key role in signal processing at receivers of wireless communication
More informationCoding Techniques for Data Storage Systems
Coding Techniques for Data Storage Systems Thomas Mittelholzer IBM Zurich Research Laboratory /8 Göttingen Agenda. Channel Coding and Practical Coding Constraints. Linear Codes 3. Weight Enumerators and
More informationLecture 17: Perfect Codes and Gilbert-Varshamov Bound
Lecture 17: Perfect Codes and Gilbert-Varshamov Bound Maximality of Hamming code Lemma Let C be a code with distance 3, then: C 2n n + 1 Codes that meet this bound: Perfect codes Hamming code is a perfect
More informationTHIS paper is aimed at designing efficient decoding algorithms
IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 45, NO. 7, NOVEMBER 1999 2333 Sort-and-Match Algorithm for Soft-Decision Decoding Ilya Dumer, Member, IEEE Abstract Let a q-ary linear (n; k)-code C be used
More informationAdaptive Cut Generation for Improved Linear Programming Decoding of Binary Linear Codes
Adaptive Cut Generation for Improved Linear Programming Decoding of Binary Linear Codes Xiaojie Zhang and Paul H. Siegel University of California, San Diego, La Jolla, CA 9093, U Email:{ericzhang, psiegel}@ucsd.edu
More informationA First Course in Digital Communications
A First Course in Digital Communications Ha H. Nguyen and E. Shwedyk February 9 A First Course in Digital Communications 1/46 Introduction There are benefits to be gained when M-ary (M = 4 signaling methods
More informationOn the Joint Decoding of LDPC Codes and Finite-State Channels via Linear Programming
On the Joint Decoding of LDPC Codes and Finite-State Channels via Linear Programming Byung-Hak Kim (joint with Henry D. Pfister) Texas A&M University College Station International Symposium on Information
More informationEE456 Digital Communications
EE456 Digital Communications Professor Ha Nguyen September 5 EE456 Digital Communications Block Diagram of Binary Communication Systems m ( t { b k } b k = s( t b = s ( t k m ˆ ( t { bˆ } k r( t Bits in
More informationLECTURE 10. Last time: Lecture outline
LECTURE 10 Joint AEP Coding Theorem Last time: Error Exponents Lecture outline Strong Coding Theorem Reading: Gallager, Chapter 5. Review Joint AEP A ( ɛ n) (X) A ( ɛ n) (Y ) vs. A ( ɛ n) (X, Y ) 2 nh(x)
More informationLecture 4. Capacity of Fading Channels
1 Lecture 4. Capacity of Fading Channels Capacity of AWGN Channels Capacity of Fading Channels Ergodic Capacity Outage Capacity Shannon and Information Theory Claude Elwood Shannon (April 3, 1916 February
More informationA Hilbert Space for Random Processes
Gaussian Basics Random Processes Filtering of Random Processes Signal Space Concepts A Hilbert Space for Random Processes I A vector space for random processes X t that is analogous to L 2 (a, b) is of
More informationSummary: SER formulation. Binary antipodal constellation. Generic binary constellation. Constellation gain. 2D constellations
TUTORIAL ON DIGITAL MODULATIONS Part 8a: Error probability A [2011-01-07] 07] Roberto Garello, Politecnico di Torino Free download (for personal use only) at: www.tlc.polito.it/garello 1 Part 8a: Error
More informationLinear Block Codes. Saravanan Vijayakumaran Department of Electrical Engineering Indian Institute of Technology Bombay
1 / 26 Linear Block Codes Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay July 28, 2014 Binary Block Codes 3 / 26 Let F 2 be the set
More informationReceived Signal, Interference and Noise
Optimum Combining Maximum ratio combining (MRC) maximizes the output signal-to-noise ratio (SNR) and is the optimal combining method in a maximum likelihood sense for channels where the additive impairment
More informationLecture 15: Thu Feb 28, 2019
Lecture 15: Thu Feb 28, 2019 Announce: HW5 posted Lecture: The AWGN waveform channel Projecting temporally AWGN leads to spatially AWGN sufficiency of projection: irrelevancy theorem in waveform AWGN:
More informationLecture 9: Diversity-Multiplexing Tradeoff Theoretical Foundations of Wireless Communications 1
: Diversity-Multiplexing Tradeoff Theoretical Foundations of Wireless Communications 1 Rayleigh Friday, May 25, 2018 09:00-11:30, Kansliet 1 Textbook: D. Tse and P. Viswanath, Fundamentals of Wireless
More informationA General Procedure to Design Good Codes at a Target BER
A General Procedure to Design Good odes at a Target BER Speaker: Xiao Ma 1 maxiao@mail.sysu.edu.cn Joint work with: hulong Liang 1, Qiutao Zhuang 1, and Baoming Bai 2 1 Dept. Electronics and omm. Eng.,
More informationAnalytical Bounds on Maximum-Likelihood Decoded Linear Codes: An Overview
Analytical Bounds on Maximum-Likelihood Decoded Linear Codes: An Overview Igal Sason Department of Electrical Engineering, Technion Haifa 32000, Israel Sason@ee.technion.ac.il December 21, 2004 Background
More informationUniversity of Siena. Multimedia Security. Watermark extraction. Mauro Barni University of Siena. M. Barni, University of Siena
Multimedia Security Mauro Barni University of Siena : summary Optimum decoding/detection Additive SS watermarks Decoding/detection of QIM watermarks The dilemma of de-synchronization attacks Geometric
More informationOptimal Receiver for MPSK Signaling with Imperfect Channel Estimation
This full text paper was peer reviewed at the direction of IEEE Communications Society subject matter experts for publication in the WCNC 27 proceedings. Optimal Receiver for PSK Signaling with Imperfect
More informationExercise 1. = P(y a 1)P(a 1 )
Chapter 7 Channel Capacity Exercise 1 A source produces independent, equally probable symbols from an alphabet {a 1, a 2 } at a rate of one symbol every 3 seconds. These symbols are transmitted over a
More informationLecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable
Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed
More informationECEN 655: Advanced Channel Coding
ECEN 655: Advanced Channel Coding Course Introduction Henry D. Pfister Department of Electrical and Computer Engineering Texas A&M University ECEN 655: Advanced Channel Coding 1 / 19 Outline 1 History
More informationLecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable
Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed
More informationLecture 9: Diversity-Multiplexing Tradeoff Theoretical Foundations of Wireless Communications 1. Overview. Ragnar Thobaben CommTh/EES/KTH
: Diversity-Multiplexing Tradeoff Theoretical Foundations of Wireless Communications 1 Rayleigh Wednesday, June 1, 2016 09:15-12:00, SIP 1 Textbook: D. Tse and P. Viswanath, Fundamentals of Wireless Communication
More informationELEG 5633 Detection and Estimation Signal Detection: Deterministic Signals
ELEG 5633 Detection and Estimation Signal Detection: Deterministic Signals Jingxian Wu Department of Electrical Engineering University of Arkansas Outline Matched Filter Generalized Matched Filter Signal
More informationA Design of High-Rate Space-Frequency Codes for MIMO-OFDM Systems
A Design of High-Rate Space-Frequency Codes for MIMO-OFDM Systems Wei Zhang, Xiang-Gen Xia and P. C. Ching xxia@ee.udel.edu EE Dept., The Chinese University of Hong Kong ECE Dept., University of Delaware
More informationEE 229B ERROR CONTROL CODING Spring 2005
EE 229B ERROR CONTROL CODING Spring 2005 Solutions for Homework 1 1. Is there room? Prove or disprove : There is a (12,7) binary linear code with d min = 5. If there were a (12,7) binary linear code with
More informationRevision of Lecture 5
Revision of Lecture 5 Information transferring across channels Channel characteristics and binary symmetric channel Average mutual information Average mutual information tells us what happens to information
More informationLecture 4: Proof of Shannon s theorem and an explicit code
CSE 533: Error-Correcting Codes (Autumn 006 Lecture 4: Proof of Shannon s theorem and an explicit code October 11, 006 Lecturer: Venkatesan Guruswami Scribe: Atri Rudra 1 Overview Last lecture we stated
More informationModulation & Coding for the Gaussian Channel
Modulation & Coding for the Gaussian Channel Trivandrum School on Communication, Coding & Networking January 27 30, 2017 Lakshmi Prasad Natarajan Dept. of Electrical Engineering Indian Institute of Technology
More informationLecture 18: Gaussian Channel
Lecture 18: Gaussian Channel Gaussian channel Gaussian channel capacity Dr. Yao Xie, ECE587, Information Theory, Duke University Mona Lisa in AWGN Mona Lisa Noisy Mona Lisa 100 100 200 200 300 300 400
More informationPerformance of Sphere Decoding of Block Codes
1 Performance of Sphere Decoding of Block Codes Mostafa El-Khamy, Haris Vikalo, Babak Hassibi and Robert J. McEliece Electrical Engineering Department California Institute of Technology Pasadena, CA 91125,
More informationELEC546 Review of Information Theory
ELEC546 Review of Information Theory Vincent Lau 1/1/004 1 Review of Information Theory Entropy: Measure of uncertainty of a random variable X. The entropy of X, H(X), is given by: If X is a discrete random
More informationLower Bounds on the Graphical Complexity of Finite-Length LDPC Codes
Lower Bounds on the Graphical Complexity of Finite-Length LDPC Codes Igal Sason Department of Electrical Engineering Technion - Israel Institute of Technology Haifa 32000, Israel 2009 IEEE International
More informationComputing Probability of Symbol Error
Computing Probability of Symbol Error I When decision boundaries intersect at right angles, then it is possible to compute the error probability exactly in closed form. I The result will be in terms of
More informationA Relation between Conditional and Unconditional Soft Bit Densities of Binary Input Memoryless Symmetric Channels
A Relation between Conditional and Unconditional Soft Bit Densities of Binary Input Memoryless Symmetric Channels Wolfgang Rave Vodafone Chair Mobile Communications Systems, Technische Universität Dresden
More informationL11: Pattern recognition principles
L11: Pattern recognition principles Bayesian decision theory Statistical classifiers Dimensionality reduction Clustering This lecture is partly based on [Huang, Acero and Hon, 2001, ch. 4] Introduction
More informationLecture 5 Channel Coding over Continuous Channels
Lecture 5 Channel Coding over Continuous Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 14, 2014 1 / 34 I-Hsiang Wang NIT Lecture 5 From
More informationMATH3302 Coding Theory Problem Set The following ISBN was received with a smudge. What is the missing digit? x9139 9
Problem Set 1 These questions are based on the material in Section 1: Introduction to coding theory. You do not need to submit your answers to any of these questions. 1. The following ISBN was received
More informationAn Introduction to Low Density Parity Check (LDPC) Codes
An Introduction to Low Density Parity Check (LDPC) Codes Jian Sun jian@csee.wvu.edu Wireless Communication Research Laboratory Lane Dept. of Comp. Sci. and Elec. Engr. West Virginia University June 3,
More informationDiversity Performance of a Practical Non-Coherent Detect-and-Forward Receiver
Diversity Performance of a Practical Non-Coherent Detect-and-Forward Receiver Michael R. Souryal and Huiqing You National Institute of Standards and Technology Advanced Network Technologies Division Gaithersburg,
More informationOn Design Criteria and Construction of Non-coherent Space-Time Constellations
On Design Criteria and Construction of Non-coherent Space-Time Constellations Mohammad Jaber Borran, Ashutosh Sabharwal, and Behnaam Aazhang ECE Department, MS-366, Rice University, Houston, TX 77005-89
More informationCapacity of AWGN channels
Chapter 3 Capacity of AWGN channels In this chapter we prove that the capacity of an AWGN channel with bandwidth W and signal-tonoise ratio SNR is W log 2 (1+SNR) bits per second (b/s). The proof that
More informationShannon meets Wiener II: On MMSE estimation in successive decoding schemes
Shannon meets Wiener II: On MMSE estimation in successive decoding schemes G. David Forney, Jr. MIT Cambridge, MA 0239 USA forneyd@comcast.net Abstract We continue to discuss why MMSE estimation arises
More informationChannel Estimation under Asynchronous Packet Interference
SUBMITTED TO IEEE TRANSACTION ON SIGNAL PROCESSING, APRIL 2003, REVISED JULY 2003 Channel Estimation under Asynchronous Packet Interference Cristian Budianu and Lang Tong Abstract This paper investigates
More informationProblem Set 2. MAS 622J/1.126J: Pattern Recognition and Analysis. Due: 5:00 p.m. on September 30
Problem Set 2 MAS 622J/1.126J: Pattern Recognition and Analysis Due: 5:00 p.m. on September 30 [Note: All instructions to plot data or write a program should be carried out using Matlab. In order to maintain
More informationOn the Error Exponents of ARQ Channels with Deadlines
On the Error Exponents of ARQ Channels with Deadlines Praveen Kumar Gopala, Young-Han Nam and Hesham El Gamal arxiv:cs/06006v [cs.it] 8 Oct 2006 March 22, 208 Abstract We consider communication over Automatic
More informationDiversity-Multiplexing Tradeoff in MIMO Channels with Partial CSIT. ECE 559 Presentation Hoa Pham Dec 3, 2007
Diversity-Multiplexing Tradeoff in MIMO Channels with Partial CSIT ECE 559 Presentation Hoa Pham Dec 3, 2007 Introduction MIMO systems provide two types of gains Diversity Gain: each path from a transmitter
More informationError Correcting Codes: Combinatorics, Algorithms and Applications Spring Homework Due Monday March 23, 2009 in class
Error Correcting Codes: Combinatorics, Algorithms and Applications Spring 2009 Homework Due Monday March 23, 2009 in class You can collaborate in groups of up to 3. However, the write-ups must be done
More information5. Discriminant analysis
5. Discriminant analysis We continue from Bayes s rule presented in Section 3 on p. 85 (5.1) where c i is a class, x isap-dimensional vector (data case) and we use class conditional probability (density
More informationLecture 7. Union bound for reducing M-ary to binary hypothesis testing
Lecture 7 Agenda for the lecture M-ary hypothesis testing and the MAP rule Union bound for reducing M-ary to binary hypothesis testing Introduction of the channel coding problem 7.1 M-ary hypothesis testing
More informationOptimized Symbol Mappings for Bit-Interleaved Coded Modulation with Iterative Decoding
Optimized Symbol Mappings for Bit-nterleaved Coded Modulation with terative Decoding F. Schreckenbach, N. Görtz, J. Hagenauer nstitute for Communications Engineering (LNT) Munich University of Technology
More informationChannel Coding and Interleaving
Lecture 6 Channel Coding and Interleaving 1 LORA: Future by Lund www.futurebylund.se The network will be free for those who want to try their products, services and solutions in a precommercial stage.
More informationTurbo Codes. Manjunatha. P. Professor Dept. of ECE. June 29, J.N.N. College of Engineering, Shimoga.
Turbo Codes Manjunatha. P manjup.jnnce@gmail.com Professor Dept. of ECE J.N.N. College of Engineering, Shimoga June 29, 2013 [1, 2, 3, 4, 5, 6] Note: Slides are prepared to use in class room purpose, may
More informationAssignment 9. Due: July 14, 2017 Instructor: Dr. Mustafa El-Halabi. ! A i. P (A c i ) i=1
CCE 40: Communication Systems Summer 206-207 Assignment 9 Due: July 4, 207 Instructor: Dr. Mustafa El-Halabi Instructions: You are strongly encouraged to type out your solutions using mathematical mode
More informationEXTENDING THE DORSCH DECODER FOR EFFICIENT SOFT DECISION DECODING OF LINEAR BLOCK CODES SEAN MICHAEL COLLISON
EXTENDING THE DORSCH DECODER FOR EFFICIENT SOFT DECISION DECODING OF LINEAR BLOCK CODES By SEAN MICHAEL COLLISON A thesis submitted in partial fulfillment of the requirements for the degree of MASTER OF
More informationMATH/MTHE 406 Homework Assignment 2 due date: October 17, 2016
MATH/MTHE 406 Homework Assignment 2 due date: October 17, 2016 Notation: We will use the notations x 1 x 2 x n and also (x 1, x 2,, x n ) to denote a vector x F n where F is a finite field. 1. [20=6+5+9]
More informationProblem Set 3 Due Oct, 5
EE6: Random Processes in Systems Lecturer: Jean C. Walrand Problem Set 3 Due Oct, 5 Fall 6 GSI: Assane Gueye This problem set essentially reviews detection theory. Not all eercises are to be turned in.
More informationDefinition 2.1. Let w be a word. Then the coset C + w of w is the set {c + w : c C}.
2.4. Coset Decoding i 2.4 Coset Decoding To apply MLD decoding, what we must do, given a received word w, is search through all the codewords to find the codeword c closest to w. This can be a slow and
More informationThis examination consists of 10 pages. Please check that you have a complete copy. Time: 2.5 hrs INSTRUCTIONS
THE UNIVERSITY OF BRITISH COLUMBIA Department of Electrical and Computer Engineering EECE 564 Detection and Estimation of Signals in Noise Final Examination 08 December 2009 This examination consists of
More informationA Thesis for the Degree of Master. An Improved LLR Computation Algorithm for QRM-MLD in Coded MIMO Systems
A Thesis for the Degree of Master An Improved LLR Computation Algorithm for QRM-MLD in Coded MIMO Systems Wonjae Shin School of Engineering Information and Communications University 2007 An Improved LLR
More informationDetecting Parametric Signals in Noise Having Exactly Known Pdf/Pmf
Detecting Parametric Signals in Noise Having Exactly Known Pdf/Pmf Reading: Ch. 5 in Kay-II. (Part of) Ch. III.B in Poor. EE 527, Detection and Estimation Theory, # 5c Detecting Parametric Signals in Noise
More informationBounds on the Error Probability of ML Decoding for Block and Turbo-Block Codes
Bounds on the Error Probability of ML Decoding for Block and Turbo-Block Codes Igal Sason and Shlomo Shamai (Shitz) Department of Electrical Engineering Technion Israel Institute of Technology Haifa 3000,
More informationOn the Optimum Asymptotic Multiuser Efficiency of Randomly Spread CDMA
On the Optimum Asymptotic Multiuser Efficiency of Randomly Spread CDMA Ralf R. Müller Friedrich-Alexander-Universität Erlangen-Nürnberg (FAU) Lehrstuhl für Digitale Übertragung 13 December 2014 1. Introduction
More informationThe E8 Lattice and Error Correction in Multi-Level Flash Memory
The E8 Lattice and Error Correction in Multi-Level Flash Memory Brian M. Kurkoski kurkoski@ice.uec.ac.jp University of Electro-Communications Tokyo, Japan ICC 2011 IEEE International Conference on Communications
More informationChapter [4] "Operations on a Single Random Variable"
Chapter [4] "Operations on a Single Random Variable" 4.1 Introduction In our study of random variables we use the probability density function or the cumulative distribution function to provide a complete
More informationChapter 2. Error Correcting Codes. 2.1 Basic Notions
Chapter 2 Error Correcting Codes The identification number schemes we discussed in the previous chapter give us the ability to determine if an error has been made in recording or transmitting information.
More informationCodes on Graphs. Telecommunications Laboratory. Alex Balatsoukas-Stimming. Technical University of Crete. November 27th, 2008
Codes on Graphs Telecommunications Laboratory Alex Balatsoukas-Stimming Technical University of Crete November 27th, 2008 Telecommunications Laboratory (TUC) Codes on Graphs November 27th, 2008 1 / 31
More information4 An Introduction to Channel Coding and Decoding over BSC
4 An Introduction to Channel Coding and Decoding over BSC 4.1. Recall that channel coding introduces, in a controlled manner, some redundancy in the (binary information sequence that can be used at the
More informationEfficient Joint Maximum-Likelihood Channel. Estimation and Signal Detection
Efficient Joint Maximum-Likelihood Channel Estimation and Signal Detection H. Vikalo, B. Hassibi, and P. Stoica Abstract In wireless communication systems, channel state information is often assumed to
More informationUC Riverside UC Riverside Previously Published Works
UC Riverside UC Riverside Previously Published Works Title Soft-decision decoding of Reed-Muller codes: A simplied algorithm Permalink https://escholarship.org/uc/item/5v71z6zr Journal IEEE Transactions
More informationCHAPTER 14. Based on the info about the scattering function we know that the multipath spread is T m =1ms, and the Doppler spread is B d =0.2 Hz.
CHAPTER 4 Problem 4. : Based on the info about the scattering function we know that the multipath spread is T m =ms, and the Doppler spread is B d =. Hz. (a) (i) T m = 3 sec (ii) B d =. Hz (iii) ( t) c
More informationIntroduction to binary block codes
58 Chapter 6 Introduction to binary block codes In this chapter we begin to study binary signal constellations, which are the Euclidean-space images of binary block codes. Such constellations have nominal
More information