Modulation & Coding for the Gaussian Channel

Similar documents
Lattices and Lattice Codes

Capacity Bounds for Power- and Band-Limited Wireless Infrared Channels Corrupted by Gaussian Noise

Summary II: Modulation and Demodulation

A First Course in Digital Communications

that efficiently utilizes the total available channel bandwidth W.

Digital Band-pass Modulation PROF. MICHAEL TSAI 2011/11/10

Digital Modulation 1

Signal Design for Band-Limited Channels

a) Find the compact (i.e. smallest) basis set required to ensure sufficient statistics.

A Simple Example Binary Hypothesis Testing Optimal Receiver Frontend M-ary Signal Sets Message Sequences. possible signals has been transmitted.

Digital Transmission Methods S

Lecture 15: Thu Feb 28, 2019

ELEC546 Review of Information Theory

7 The Waveform Channel

Lecture 12. Block Diagram

392D: Coding for the AWGN Channel Wednesday, January 24, 2007 Stanford, Winter 2007 Handout #6. Problem Set 2 Solutions

EE 574 Detection and Estimation Theory Lecture Presentation 8

Performance of small signal sets

CHANNEL CAPACITY CALCULATIONS FOR M ARY N DIMENSIONAL SIGNAL SETS

Upper and Lower Bounds on the Capacity of Wireless Optical Intensity Channels Corrupted by Gaussian Noise

ECE 564/645 - Digital Communications, Spring 2018 Midterm Exam #1 March 22nd, 7:00-9:00pm Marston 220

ECE 564/645 - Digital Communications, Spring 2018 Homework #2 Due: March 19 (In Lecture)

ρ = sin(2π ft) 2π ft To find the minimum value of the correlation, we set the derivative of ρ with respect to f equal to zero.

GEORGIA INSTITUTE OF TECHNOLOGY SCHOOL OF ELECTRICAL AND COMPUTER ENGINEERING Final Examination - Fall 2015 EE 4601: Communication Systems

EE4304 C-term 2007: Lecture 17 Supplemental Slides

Square Root Raised Cosine Filter

Diversity Performance of a Practical Non-Coherent Detect-and-Forward Receiver

Chapter 2 Signal Processing at Receivers: Detection Theory

BASICS OF DETECTION AND ESTIMATION THEORY

PSK bit mappings with good minimax error probability

Revision of Lecture 5

CHAPTER 14. Based on the info about the scattering function we know that the multipath spread is T m =1ms, and the Doppler spread is B d =0.2 Hz.

Introduction to Convolutional Codes, Part 1

5 Analog carrier modulation with noise

This examination consists of 11 pages. Please check that you have a complete copy. Time: 2.5 hrs INSTRUCTIONS

Summary: SER formulation. Binary antipodal constellation. Generic binary constellation. Constellation gain. 2D constellations

Lecture 7. Union bound for reducing M-ary to binary hypothesis testing

Es e j4φ +4N n. 16 KE s /N 0. σ 2ˆφ4 1 γ s. p(φ e )= exp 1 ( 2πσ φ b cos N 2 φ e 0

SIPCom8-1: Information Theory and Coding Linear Binary Codes Ingmar Land

These outputs can be written in a more convenient form: with y(i) = Hc m (i) n(i) y(i) = (y(i); ; y K (i)) T ; c m (i) = (c m (i); ; c m K(i)) T and n

The E8 Lattice and Error Correction in Multi-Level Flash Memory

EE456 Digital Communications

Applications of Lattices in Telecommunications


Kevin Buckley a i. communication. information source. modulator. encoder. channel. encoder. information. demodulator decoder. C k.

Computing Probability of Symbol Error

Data Detection for Controlled ISI. h(nt) = 1 for n=0,1 and zero otherwise.

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Lecture Notes 1: Vector spaces

Problem 7.7 : We assume that P (x i )=1/3, i =1, 2, 3. Then P (y 1 )= 1 ((1 p)+p) = P (y j )=1/3, j=2, 3. Hence : and similarly.

Appendix B Information theory from first principles

On Design Criteria and Construction of Non-coherent Space-Time Constellations

On the Low-SNR Capacity of Phase-Shift Keying with Hard-Decision Detection

Rapport technique #INRS-EMT Exact Expression for the BER of Rectangular QAM with Arbitrary Constellation Mapping

Performance Analysis of Spread Spectrum CDMA systems

A Systematic Description of Source Significance Information

19. Channel coding: energy-per-bit, continuous-time channels

Error Exponent Region for Gaussian Broadcast Channels

ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process

Detecting Parametric Signals in Noise Having Exactly Known Pdf/Pmf

Optimum Soft Decision Decoding of Linear Block Codes

Outline Digital Communications. Lecture 02 Signal-Space Representation. Energy Signals. MMSE Approximation. Pierluigi SALVO ROSSI

LIKELIHOOD RECEIVER FOR FH-MFSK MOBILE RADIO*

Dr. Cathy Liu Dr. Michael Steinberger. A Brief Tour of FEC for Serial Link Systems

X b s t w t t dt b E ( ) t dt

EE6604 Personal & Mobile Communications. Week 15. OFDM on AWGN and ISI Channels

Consider a 2-D constellation, suppose that basis signals =cosine and sine. Each constellation symbol corresponds to a vector with two real components

Trellis Coded Modulation

Mapper & De-Mapper System Document

EE4512 Analog and Digital Communications Chapter 4. Chapter 4 Receiver Design

MITOCW ocw-6-451_ feb k_512kb-mp4

Information Theory - Entropy. Figure 3

8 PAM BER/SER Monte Carlo Simulation

Lecture 9: Diversity-Multiplexing Tradeoff Theoretical Foundations of Wireless Communications 1

Example: Bipolar NRZ (non-return-to-zero) signaling

Lecture # 3 Orthogonal Matrices and Matrix Norms. We repeat the definition an orthogonal set and orthornormal set.

Chapter 9 Fundamental Limits in Information Theory

EE Introduction to Digital Communications Homework 8 Solutions

EE 661: Modulation Theory Solutions to Homework 6

The E8 Lattice and Error Correction in Multi-Level Flash Memory

Lecture 9: Diversity-Multiplexing Tradeoff Theoretical Foundations of Wireless Communications 1. Overview. Ragnar Thobaben CommTh/EES/KTH

Upper Bounds for the Average Error Probability of a Time-Hopping Wideband System

Problem Set 2. MAS 622J/1.126J: Pattern Recognition and Analysis. Due: 5:00 p.m. on September 30

ECE Information theory Final (Fall 2008)

Digital Communications

4184 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 51, NO. 12, DECEMBER Pranav Dayal, Member, IEEE, and Mahesh K. Varanasi, Senior Member, IEEE

Advanced Digital Communications

Memo of J. G. Proakis and M Salehi, Digital Communications, 5th ed. New York: McGraw-Hill, Chenggao HAN

ECE6604 PERSONAL & MOBILE COMMUNICATIONS. Week 3. Flat Fading Channels Envelope Distribution Autocorrelation of a Random Process

Principles of Communications Lecture 8: Baseband Communication Systems. Chih-Wei Liu 劉志尉 National Chiao Tung University

2. SPECTRAL ANALYSIS APPLIED TO STOCHASTIC PROCESSES

Communication Theory Summary of Important Definitions and Results

A Design of High-Rate Space-Frequency Codes for MIMO-OFDM Systems

Signals and Spectra - Review

Communication Engineering Prof. Surendra Prasad Department of Electrical Engineering Indian Institute of Technology, Delhi

2016 Spring: The Final Exam of Digital Communications

IEEE C80216m-09/0079r1

4 An Introduction to Channel Coding and Decoding over BSC

Decision-Point Signal to Noise Ratio (SNR)

Shannon meets Wiener II: On MMSE estimation in successive decoding schemes

Transcription:

Modulation & Coding for the Gaussian Channel Trivandrum School on Communication, Coding & Networking January 27 30, 2017 Lakshmi Prasad Natarajan Dept. of Electrical Engineering Indian Institute of Technology Hyderabad lakshminatarajan@iith.ac.in 1 / 54

Digital Communication Convey a message from transmitter to receiver in a finite amount of time, where the message can assume only finitely many values. ˆ time can be replaced with any resource: space available in a compact disc, number of cells in flash memory Picture courtesy brigetteheffernan.wordpress.com 2 / 54

The Additive Noise Channel ˆ Message m takes finitely many, say M, distinct values Usually, not always, M = 2 k, for some integer k assume m is uniformly distributed over {1,..., M} ˆ Time duration T transmit signal s(t) is restricted to 0 t T ˆ Number of message bits k = log 2 M (not always an integer) 3 / 54

Modulation Scheme ˆ The transmitter & receiver agree upon a set of waveforms {s 1 (t),..., s M (t)} of duration T. ˆ The transmitter uses the waveform s i (t) for the message m = i. ˆ The receiver must guess the value of m given r(t). ˆ We say that a decoding error occurs if the guess ˆm m. Definition An M-ary modulation scheme is simply a set of M waveforms {s 1 (t),..., s M (t)} each of duration T. Terminology ˆ Binary: M = 2, modulation scheme {s 1 (t), s 2 (t)} ˆ Antipodal: M = 2 and s 2 (t) = s 1 (t) ˆ Ternary: M = 3, Quaternary: M = 4 4 / 54

Parameters of Interest ˆ Bit rate R = log 2 M T bits/sec Energy of the i th waveform E i = s i (t) 2 = ˆ Average Energy i=1 i=1 T t=0 s 2 i (t)dt M M 1 T E = P (m = i)e i = s i (t) 2 M t=0 ˆ Energy per message bit E b = E log 2 M ˆ Probability of error P e = P (m ˆm) Note P e depends on the modulation scheme, noise statistics and the demodulator. 5 / 54

Example: On-Off Keying, M = 2 6 / 54

Objectives 1 Characterize and analyze a modulation scheme in terms of energy, rate and error probability. What is the best/optimal performance that one can expect? 2 Design a good modulation scheme that performs close to the theoretical optimum. Key tool: Signal Space Representation ˆ Represent waveforms as vectors: geometry of the problem ˆ Simplifies performance analysis and modulation design ˆ Leads to efficient modulation/demodulation implementations 7 / 54

1 Signal Space Representation 2 Vector Gaussian Channel 3 Vector Gaussian Channel (contd.) 4 Optimum Detection 5 Probability of Error 8 / 54

References ˆ I. M. Jacobs and J. M. Wozencraft, Principles of Communication Engineering, Wiley, 1965. ˆ G. D. Forney and G. Ungerboeck, Modulation and coding for linear Gaussian channels, in IEEE Transactions on Information Theory, vol. 44, no. 6, pp. 2384-2415, Oct 1998. ˆ D. Slepian and H. O. Pollak, Prolate spheroidal wave functions, Fourier analysis and uncertainty I, in The Bell System Technical Journal, vol. 40, no. 1, pp. 43-63, Jan. 1961. ˆ H. J. Landau and H. O. Pollak, Prolate spheroidal wave functions, Fourier analysis and uncertainty III: The dimension of the space of essentially time- and band-limited signals, in The Bell System Technical Journal, vol. 41, no. 4, pp. 1295-1336, July 1962. 8 / 54

1 Signal Space Representation 2 Vector Gaussian Channel 3 Vector Gaussian Channel (contd.) 4 Optimum Detection 5 Probability of Error 9 / 54

Goal Map waveforms s 1 (t),..., s M (t) to M vectors in a Euclidean space R N, so that the map preserves the mathematical structure of the waveforms. 9 / 54

Quick Review of R N : N-Dimensional Euclidean Space R N = { (x 1, x 2,..., x N ) x 1,..., x N R } Notation: x = (x 1, x 2,..., x N ) and 0 = (0, 0,..., 0) Addition Properties: ˆ x + y = (x 1,..., x N ) + (y 1,..., y N ) = (x 1 + y 1,..., x N + y N ) ˆ x y = (x 1,..., x N ) (y 1,..., y N ) = (x 1 y 1,..., x N y N ) ˆ x + 0 = x for every x R N Multiplication Properties: ˆ a x = a (x 1,..., x N ) = (ax 1,..., ax N ), where a R ˆ a(x + y) = ax + ay ˆ (a + b)x = ax + bx ˆ ax = 0 if and only if a = 0 or x = 0 10 / 54

Quick Review of R N : Inner Product and Norm Inner Product ˆ x, y = y, x = x 1 y 1 + x 2 y 2 + + x N y N ˆ x, y + z = x, y + x, z (distributive law) ˆ ax, y = a x, y ˆ If x, y = 0 we say that x and y are orthogonal Norm ˆ x = x 2 1 + + x2 N = x, x denotes the length of x ˆ x 2 = x, x denotes the energy of the vector x ˆ x 2 = 0 if and only if x = 0 ˆ If x = 1 we say that x is of unit norm ˆ x y is the distance between two vectors. Cauchy-Schwarz Inequality ˆ x, y x y ˆ Or equivalently, 1 x, y x y 1 11 / 54

Waveforms as Vectors The set of all finite-energy waveforms of duration T and the Euclidean space R N share many structural properties. Addition Properties ˆ We can add and subtract two waveforms x(t) + y(t), x(t) y(t) ˆ The all-zero waveform 0(t) = 0 for 0 t T is the additive identity x(t) + 0(t) = x(t) for any waveform x(t) Multiplication Properties ˆ We can scale x(t) using a real number a and obtain a x(t) ˆ a(x(t) + y(t)) = ax(t) + ay(t) ˆ (a + b)x(t) = ax(t) + bx(t) ˆ ax(t) = 0(t) if and only if a = 0 or x(t) = 0(t) 12 / 54

Inner Product and Norm of Waveforms Inner Product ˆ x(t), y(t) = y(t), x(t) = T t=0 x(t)y(t)dt ˆ x(t), y(t) + z(t) = x(t), y(t) + x(t), z(t) (distributive law) ˆ ax(t), y(t) = a x(t), y(t) ˆ If x(t), y(t) = 0 we say that x(t) and y(t) are orthogonal Norm ˆ x(t) = x(t), x(t) = T t=0 x2 (t)dt is the norm of x(t) ˆ x(t) 2 = T t=0 x2 (t)dt denotes the energy of x(t) ˆ If x(t) = 1 we say that x(t) is of unit norm ˆ x(t) y(t) is the distance between two waveforms Cauchy-Schwarz Inequality ˆ x(t), y(t) x(t) y(t) for any two waveforms x(t), y(t) We want to map s 1 (t),..., s M (t) to vectors s 1,..., s M R N so that the addition, multiplication, inner product and norm properties are preserved. 13 / 54

Orthonormal Waveforms Definition A set of N waveforms {φ 1 (t),..., φ N (t)} is said to be orthonormal if 1 φ 1 (t) = φ 2 (t) = = φ N (t) = 1 (unit norm) 2 φ i (t), φ j (t) = 0 for all i j (orthogonality) The role of orthonormal waveforms is similar to that of the standard basis e 1 = (1, 0, 0,..., 0), e 2 = (0, 1, 0,..., 0),, e N = (0, 0,..., 0, 1) Remark Say x(t) = x 1 φ 1 (t) + x N φ N (t), y(t) = y 1 φ 1 (t) + + y N φ N (t) N N x(t), y(t) = x i φ i (t), y j φ j (t) = x i y j φ i (t), φ j (t) i=1 j=1 i j = x i y j = x i y i i j=i i = x, y 14 / 54

Example 15 / 54

Orthonormal Basis Definition An orthonormal basis for {s 1 (t),..., s M (t)} is an orthonormal set {φ 1 (t),..., φ N (t)} such that s i (t) = s i,1 φ i (t) + s i,2 φ 2 (t) + + s i,m φ N (t) for some choice of s i,1, s i,2,..., s i,n R ˆ We associate s i (t) s i = (s i,1, s i,2,..., s i,n ) ˆ A given modulation scheme can have many orthonormal bases. ˆ The map s 1 (t) s 1, s 2 (t) s 2,..., s M (t) s M depends on the choice of orthonormal basis. 16 / 54

Modulation Scheme Example: M-ary Phase Shift Keying ˆ s i (t) = A cos(2πf c t + 2πi M ), i = 1,..., M ˆ Expanding s i (t) using cos(c + D) = cos C cos D sin C sin D s i (t) = A cos Orthonormal Basis ( 2πi M ) cos(2πf c t) A sin ( 2πi M ) sin(2πf c t) ˆ Use φ 1 (t) = 2/T cos(2πf c t) and φ 2 (t) = 2/T sin(2πf c t) ( ) ( ) T 2πi T 2πi s i (t) = A 2 cos φ 1 (t) + A M 2 sin φ 2 (t) M ˆ Dimension N = 2 Waveform to Vector s i (t) ( A2 T 2 cos ( ) 2πi A2 T, M 2 sin ( ) ) 2πi M 17 / 54

8-ary Phase Shift Keying 18 / 54

How to find an orthonormal basis Gram-Schmidt Procedure Given a modulation scheme {s 1 (t),..., s M (t)}, constructs an orthonormal basis φ 1 (t),..., φ N (t) for the scheme. Similar to QR factorization of matrices r 1,1 r 1,2 r 1,M r 2,1 r 2,2 r 2,M A = [a 1 a 2 a M ] = [q 1 q 2 q N ]... = QR r N,1 r N,2 r N,M s 1,1 s 2,1 s M,1 s 1,2 s 2,2 s M,2 [s 1 (t) s M (t)] = [φ 1 (t) φ N (t)]... s 1,N s 2,N s M,N 19 / 54

Waveforms to Vectors, and Back Say {φ 1 (t),..., φ N (t)} is an orthonormal basis for {s 1 (t),..., s M (t)}. N Then, s i (t) = s i,j φ j (t) for some choice of {s i,j } Waveform to Vector j=1 s i (t), φ j (t) = k s i,k φ k (t), φ j (t) = k s i,k φ k (t), φ j (t) = s i,j s i (t) (s i,1, s i,2,..., s i,n ) = s i where s i,1 = s i (t), φ 1 (t), s i,2 = s i (t), φ 2 (t),..., s i,n = s i (t), φ N (t) Vector to Waveform s i = (s i,1,..., s i,n ) s i,1 φ 1 (t) + s i,2 φ 2 (t) + + s i,n φ N (t) ˆ Every point in R N corresponds to a unique waveform. ˆ Going back and forth between vectors and waveforms is easy. 20 / 54

Waveforms to Vectors, and Back Caveat v(t) Waveform to vector v v Vector to waveform ˆv(t) ˆv(t) = v(t) iff v(t) is some linear combination of φ 1 (t),..., φ N (t), or equivalently, v(t) is some linear combination of s 1 (t),..., s M (t) 21 / 54

Equivalence Between Waveform and Vector Representations Say v(t) = v 1 φ 1 (t) + + v N φ N (t) and u(t) = u 1 φ 1 (t) + + u N φ N (t) Addition v(t) + u(t) v + u Scalar Multiplication a v(t) a v Energy v(t) 2 v 2 Inner product v(t), u(t) v, u Distance v(t) u(t) v u Basis φ i (t) e i (Std. basis) 22 / 54

1 Signal Space Representation 2 Vector Gaussian Channel 3 Vector Gaussian Channel (contd.) 4 Optimum Detection 5 Probability of Error 23 / 54

Vector Gaussian Channel Definition An M-ary modulation scheme of dimension N is a set of M vectors {s 1,..., s M } in R N ˆ Average energy E = 1 ( s 1 2 + + s M 2) M 23 / 54

Vector Gaussian Channel Relation between received vector r and transmit vector s i The j th component of received vector r = (r 1,..., r N ) r j = r(t), φ i (t) = s i (t) + n(t), φ j (t) = s i (t), φ j (t) + n(t), φ j (t) = s i,j + n j Denoting n = (n 1,..., n N ) we obtain r = s i + n If n(t) is a Gaussian random process, noise vector n follows Gaussian distribution. Note Effective noise at the receiver ˆn(t) = n 1 φ 1 (t) + + n N φ N (t) In general, n(t) not a linear combination of basis, and ˆn(t) n(t), 24 / 54

Designing a Modulation Scheme 1 Choose an orthonormal basis φ 1 (t),..., φ N (t) Determines bandwidth of transmit signals, signalling duration T 2 Construct a (vector) modulation scheme s 1,..., s M R N Determines the signal energy, probability of error An N-dimensional modulation scheme exploits N uses of a scalar Gaussian channel r j = s i,j + n j where j = 1,..., N With limits on bandwidth and signal duration, how large can N be? 25 / 54

Dimension of Time/Band-limited Signals Say transmit signals s(t) must be time/band limited 1 s(t) = 0 if t < 0 or t T, and (time-limited) 2 S(f) = 0 if f < f c W 2 or f > f c + W 2 (band-limited) Uncertainty Principle: No non-zero signal is both time- and band-limited. No signal transmission is possible! We relax the constraint to approximately band-limited 1 s(t) = 0 if t < 0 or t > T, and (time-limited) 2 fc+w/2 S(f) 2 df (1 δ) f=f c W/2 0 + Here δ > 0 is the fraction of out-of-band signal energy. S(f) 2 df (approx. band-lim.) What is the largest dimension N of time-limited/approximately band-limited signals? 26 / 54

Dimension of Time/band-limited Signals Let T > 0 and W > 0 be given, and consider any δ, ɛ > 0. Theorem (Landau, Pollak & Slepian 1961-62) If T W is sufficiently large, there exists N = 2T W (1 ɛ) orthonormal waveforms φ 1 (t),..., φ N (t) such that 1 φ i (t) = 0 if t < 0 or t > T, and (time-limited) 2 fc+w/2 f=f c W/2 In summary Φ i (f) 2 df (1 δ) + 0 Φ i (f) 2 df (approx. band-lim.) ˆ We can pack N 2T W dimensions if the time-bandwidth product T W is large enough. ˆ Number of dimensions/channel uses normalized to 1 sec of transmit duration and 1 Hz of bandwidth N T W 2 dim/sec/hz 27 / 54

Relation between Waveform & Vector Channels Assume N = 2T W Signal energy E i s i (t) 2 s i 2 1 Avg. energy E M i s i(t) 2 1 M i s i 2 Transmit Power Rate S R E T log 2 M T E N 2W log 2 M N 2W Parameters for Vector Gaussian Channel ˆ Spectral Efficiency η = 2 log 2 M/N (unit: bits/sec/hz) Allows comparison between schemes with different bandwidths. Related to rate as η = R/W ˆ Power P = E/N (unit: Watt/Hz) Related to actual transmit power as S = 2W P 28 / 54

1 Signal Space Representation 2 Vector Gaussian Channel 3 Vector Gaussian Channel (contd.) 4 Optimum Detection 5 Probability of Error 29 / 54

Detection in the Gaussian Channel Definition Detection/Decoding/Demodulation is the process of estimating the message m given the received waveform r(t) and the modulation scheme {s 1 (t),..., s M (t)}. Objective: Design the decoder to minimize P e = P ( ˆm m). 29 / 54

The Gaussian Random Variable ˆ P (X < a) = P (X > a) = Q(a) ˆ Q( ) is a decreasing function ˆ Y = σx is Gaussian with mean 0 and var σ 2, i.e., N (0, σ 2 ) ˆ P (Y > b) = P (σx > b) = P (X > b σ ) = Q ( ) b σ 30 / 54

White Gaussian Noise Process n(t) Noise waveform n(t) modelled as a white Gaussian random process, i.e., as a a collection of random variables {n(τ) < τ < + } such that ˆ Stationary random process Statistics of the processes n(t) and n(t constant) are identical ˆ Gaussian random process Any linear combination of finitely many samples of n(t) is Gaussian a 1 n(t 1 ) + a 2 n(t 2 ) + + a l n(t l ) Gaussian distributed ˆ White random process The power spectrum N(f) of the noise process is flat N(f) = N o 2 W/Hz, for < f < + 31 / 54

32 / 54

Noise Process Through Waveform-to-Vector Converter Properties of the noise vector n = (n 1,..., n N ) ˆ n 1, n 2,..., n N are independent N (0, N o /2) random variables f(n i ) = 1 ( ) exp n2 i πno N o ˆ Noise vector n describes only a part of n(t) ˆn(t) = n 1 φ 1 (t) + + n N φ N (t) n(t) The noise component not captured by waveform-to-vector converter: n(t) = n(t) ˆn(t) 0 33 / 54

White Gaussian Noise Vector n n = (n 1,..., n N ) ˆ Probability density of n = (n 1,..., n N ) in R N N ) 1 f noise (n) = f(n 1,..., n N ) = f(n i ) = ( ( πn o ) exp n 2 N i=1 N o Probability density depends only on n 2 Spherically symmetric: Isotropic distribution Density highest near 0 and decreasing in n 2 noise vector of larger norm less likely than a vector with smaller norm ( ˆ For any a R N, n, a N 0, a 2 N ) o 2 ˆ a 1,..., a K are orthonormal n, a 1,..., n, a K are independent N (0, N o /2) 34 / 54

n(t) Carries Irrelevant Information ˆ r = s i + n does not carry all the information in r(t) ˆr(t) = r 1 φ 1 (t) + + r N φ N (t) r(t) ˆ The information about r(t) not contained in r r(t) j r j φ j (t) = s i (t) + n(t) j s i,j φ j (t) j n j φ j (t) = n(t) Theorem The vector r contains all the information in r(t) that is relevant to the transmitted message. ˆ n(t) is irrelevant for the optimum detection of transmit message. 35 / 54

The (Effective) Vector Gaussian Channel ˆ Modulation Scheme/Code is a set {s 1,..., s M } of M vectors in R N ˆ Power P = 1 N s 1 2 + + s M 2 M ˆ Noise variance σ 2 = N o (per dimension) 2 ˆ Signal to noise ratio SNR = P σ 2 = 2P N o ˆ Spectral Efficiency η = 2 log 2 M bits/s/hz (assuming N = 2T W ) N 36 / 54

1 Signal Space Representation 2 Vector Gaussian Channel 3 Vector Gaussian Channel (contd.) 4 Optimum Detection 5 Probability of Error 37 / 54

Optimum Detection Rule Objective Given {s 1,..., s M } & r, provide an estimate ˆm of the transmit message m, so that P e = P ( ˆm m) is as small as possible. Optimal Detection: Maximum a posteriori (MAP) detector Given received vector r, choose the vector s j that has the highest probability of being transmitted ˆm = arg max P (s k transmitted r received ) k {1,...,M} In other words, choose ˆm = k if P (s k transmitted r received ) > P (s j transmitted r received ) for every j k ˆ In case of a tie, can choose one of the indices arbitrarily. This does not increase P e. 37 / 54

Use Bayes rule P (A B) = Optimum Detection Rule P (A)P (B A) P (B) P (s k )f(r s k ) ˆm = arg max P (s j r) = arg max k k f(r) P (s j ) = Probability of transmitting s j = 1/M (equally likely messages) f(r s k ) = Probability density of r when s k is transmitted f(r) = Probability density of r averaged over all possible transmissions ˆm = arg max k 1/M f(r s k ) f(r) = arg max f(r s k ) k Likelihood function f(r s k ), Max. likelihood rule ˆm = arg max k f(r s k ) If all the M messages are equally likely Max. a posteriori detection = Max. likelihood (ML) detection 38 / 54

Maximum Likelihood Detection in Vector Gaussian Channel Use the model r = s i + n and the assumption n is independent of s i ˆm = arg max k f(r s k ) = arg max f noise (r s k s k ) k = arg max f noise (r s k ) k = arg max k 1 ( πn o ) N exp = arg min k r s k 2 ML Detection Rule for Vector Gaussian Channel Choose ˆm = k if r s k < r s j for every j k ( r s k 2 ) ˆ Also called minimum distance/nearest neighbor decoding ˆ In case of a tie, choose one of the contenders arbitrarily. N o 39 / 54

Example: M = 6 vectors in R 2 The k th Decision region D k D k = set of all points closer to s k than any other s j = { r R N r s k < r s j for all j k } The ML detector outputs ˆm = k if r D k. 40 / 54

Examples in R 2 41 / 54

1 Signal Space Representation 2 Vector Gaussian Channel 3 Vector Gaussian Channel (contd.) 4 Optimum Detection 5 Probability of Error 42 / 54

Error Probability when M = 2 Scenario Let {s 1, s 2 } R N be a binary modulation scheme with ˆ P (s 1 ) = P (s 2 ) = 1/2, and ˆ detected using the nearest neighbor decoder ˆ Error E occurs if (s 1 tx, ˆm = 2) or (s 2 tx, ˆm = 1) ˆ Conditional error probability P (E s 1 ) = P ( ˆm = 2 s 1 ) = P ( r s 2 < r s 1 s 1 ) ˆ Note that P (E) = P (s 1 )P (E s 1 ) + P (s 2 )P (E s 2 ) = P (E s 1 ) + P (E s 2 ) 2 ˆ P (E s i ) can be easy to analyse 42 / 54

Conditional Error Probability when M = 2 E s 1 : s 1 is transmitted r = s 1 + n, and r s 1 2 > r s 2 2 (E s 1 ) : s 1 + n s 1 2 > s 1 + n s 2 2 n 2 > s 1 s 2 + n, s 1 s 2 + n n 2 > s 1 s 2, s 1 s 2 + s 1 s 2, n + n, s 1 s 2 + n, n n 2 > s 1 s 2 2 + 2 n, s 1 s 2 + n 2 n, s 1 s 2 < s 1 s 2 2 2 s 1 s 2 2 n, s 1 s 2 < s 1 s 2 2 N o 2 2 n, s 1 s 2 s 1 s 2 N o < s 1 s 2 2No 1 2 s 1 s 2 N o 43 / 54

ˆ Z = n, N o 2 Error Probability when M = 2 s 1 s 2 s 1 s 2 ˆ P (E s 1 ) = P 2 N o s 1 s 2 2 s 1 s 2 N o ( ) s1 s 2 ˆ P (E s 2 ) = Q 2No is Gaussian with zero mean and variance 2 = N o 2 N o ( Z < s ) ( ) 1 s 2 s1 s 2 = Q 2No 2No P (E) = P (E s 1 ) + P (E s 2 ) 2 2 s 1 s 2 s 1 s 2 ( ) s1 s 2 = Q 2No 2 = 1 ˆ Error probability decreasing function of distance s 1 s 2 44 / 54

Bound on Error Probability when M > 2 Scenario Let C = {s 1,..., s M } R N be a modulation/coding scheme with ˆ P (s 1 ) = = P (s M ) = 1/M, and ˆ detected using the nearest neighbor decoder ˆ Minimum distance d min = smallest Euclidean distance between any pair of vectors in C d min = min i j s i s j ˆ Observe that s i s j d min for any i j ˆ Since Q( ) is a decreasing function ( ) ( ) si s j dmin Q Q 2No 2No for any i j ˆ Bound based only on d min Simple calculations, not tight, intuitive 45 / 54

46 / 54

Union Bound on Conditional Error Probability Assume that s 1 is transmitted, i.e., r = s 1 + n. We know that ( ) s1 s j P ( r s j < r s j s 1 ) = Q 2No Decoding error occurs if r is closer some s j than s 1, j = 2, 3,..., M M P (E s 1 ) = P (r / D 1 s 1 ) = P r s j < r s 1 s 1 j=2 From union bound P (A 2 A M ) P (A 2 ) + + P (A M ) M M ( ) s1 s j P (E s 1 ) P ( r s j < r s j s 1 ) = Q 2No j=2 j=2 47 / 54

Union Bound on Error Probability Since Q( ) is a decreasing function and s 1 s j d min M ( ) s1 s j M ( ) dmin P (E s 1 ) Q Q 2No 2No j=2 j=2 ( ) dmin P (E s 1 ) (M 1)Q 2No Upper bound on average error probability P (E) = M i=1 P (s i )P (E s i ) ( ) dmin P (E) (M 1)Q 2No Note ˆ Exact P e (or good approximations better than the union bound) can be derived for several constellations, for example PAM, QAM and PSK. ˆ Chernoff bound can be useful: Q(a) 1 2 exp( a2 /2) for a 0 ˆ Union bound, in general, is loose. 48 / 54

ˆ Abscissa is [SNR] db = 10 log 10 SNR ˆ The union bound is a reasonable approximation for large values of SNR 49 / 54

Performance of QAM and FSK η = 2 log 2 M N 50 / 54

Performance of QAM and FSK Probability of Error P e = 10 5 Modulation/Code Spectral Efficiency Signal-to-Noise Ratio η (bits/sec/hz) SNR (db) 16-QAM 4 20 4-QAM 2 13 2-FSK 1 12.6 8-FSK 3/4 7.5 16-FSK 1/2 4.6 How good are these modulation schemes? What is the best trade-off between SNR and η? 51 / 54

Capacity of the (Vector) Gaussian Channel Let the maximum allowable power be P and noise variance be N o /2. SNR = P N o /2 = 2P N o What is the highest η achievable while ensuring that P e is small? Theorem Given an ɛ > 0 and any constant η such that η < log 2 (1 + SNR), there exists a coding scheme with P e ɛ and spectral efficiency at least η. Conversely, for any coding scheme with η > log 2 (1 + SNR) and M sufficiently large, P e is close to 1. C(SNR) = log 2 (1 + SNR) is the capacity of the Gaussian channel. 52 / 54

How Good/Bad are QAM and FSK? Least SNR required to communicate reliably with spectral efficiency η is SNR (η) = 2 η 1 Probability of Error P e = 10 5 Modulation/Code η SNR (db) SNR (η) 16-QAM 4 20 11.7 4-QAM 2 13 4.7 2-FSK 1 12.6 0 8-FSK 3/4 7.5 1.7 16-FSK 1/2 4.6 3.8 53 / 54

How to Perform Close to Capacity? ˆ We need P e to be small at a fixed finite SNR d min must be large to ensure that P e is small ˆ It is necessary to use coding schemes in high dimensions N 1 Can ensure that d min constant N ˆ If N is large it is possible to pack vectors {s i } in R N such that Average power is at the most P d min is large η is close to log 2 (1 + SNR) P e is small ˆ A large N implies that M = 2 ηn/2 is also large. We must ensure that such a large code can be encoded/decoded with practical complexity Several known coding techniques η > 1: Trellis coded modulation, multilevel codes, lattice codes, bit-interleaved coded modulation, etc. η < 1: Low-density parity-check codes, turbo codes, polar codes, etc. Thank You! 54 / 54