Linear Gaussian Channels

Size: px
Start display at page:

Download "Linear Gaussian Channels"

Transcription

1 Chapter 15 Linear Gaussian Channels In this chapter we consider a more general linear Gaussian channel model in which the signal s(t) is passed through a linear filter h(t) with Fourier transform H (f ), and the noise power spectral density S n (f ) may be non-white. By a simple whitening filter argument, we show that without loss of generality or optimality we may consider only the white noise case. The channel capacity of a linear channel is derived via Shannon s well-known water-pouring argument, which determines the capacity-achieving transmit power spectral density subject to a constraint on total transmit power. This gives good practical guidelines as to which frequency band(s) should be used for transmission. Multicarrier modulation is one straightforward method for approaching this capacity. Using the principles of optimum detection theory, we show that for single-carrier PAM or QAM transmission through such a channel with a given symbol interval T, without loss of optimality we may reduce the channel model to an equivalent discrete-time channel model by use of a matched filter or a whitened matched filter. This development uses the principles of discretetime spectral factorization, which may be viewed as an instance of Cholesky factorization. Optimum sequence detection may then be performed by the Viterbi algorithm (VA). However, while this gives bounds on the best possible performance, in practice the VA is too complex. The simplest receiver is simply a linear receiver that removes intersymbol interference. We show that an optimal zero-forcing (ZF-LE) receiver exists when there are no nulls in the discrete-time channel spectrum. We determine its performance when it does exist, and show that it is near-canonical when there are no near-nulls in the channel spectrum. The simplest nonlinear receiver is a zero-forcing decision-feedback equalizer (ZF-DFE). We show that an optimal ZF-DFE exists when there is no null band in the transmit spectrum, and we compute its performance. One striking result ( Price s result ) is that at high SNRs the gap to capacity with uncoded PAM or QAM and ZF-DFE is approximately the same for all linear Gaussian channels; i.e., the ZF-DFE is effectively canonical. The above developments apply to uncoded PAM or QAM. There are no practical difficulties involved in combining coding with linear equalization. However, decision-feedback equalization does not combine naturally with coding, since immediate decisions are needed. We show that in point-to-point applications where equalization may be performed in the transmitter, DFE-equivalent equalization performance may be obtained in conjunction with coding by a transmitter equalization technique called precoding. 205

2 206 CHAPTER 15. LINEAR GAUSSIAN CHANNELS 15.1 Noise-whitening filters The general linear Gaussian channel model is r (t ) = s (t ) h (t )+ n (t ), (15.1) where the transmitted signal s (t) has power spectral density (p.s.d.) S x (f) and is subject to a power constraint S x (f) df P, the channel impulse response h (t) is L 2 and has Fourier transform H (f), and the additive noise n (t) is a Gaussian process with p.s.d. S n (f). The channel band is defined as B c = supp H (f) = {f : H (f) 0}. The channel SNR function is defined as SNR c (f ) = H (f ) 2 /S n (f ). We assume that SNR c (f ) < for all f ; i.e., S n (f ) > 0 whenever H (f ) 0. A noise-whitening filter is any filter with response g (t ) and transform G (f ) such that G (f ) 2 = S n /S n (f ) for some constant S n > 0 for all frequencies f B c. The filter is applied to the received signal r (t ) togive r (t ) = r (t ) g (t ) = s (t ) h (t ) g (t )+ n (t ) g (t ). Since G (f ) is nonzero in the channel band B c,it is invertible in B c and thus information-lossless. The filtered noise n (t ) = n (t ) g (t ) then has constant p.s.d. S n (f ) = S n for all f B c. Since the receiver may without loss of optimality filter out all frequencies for which H (f ) =0, it follows that the whitened noise n (t ) is equivalent to white Gaussian noise with p.s.d. S n with respect to the channel band B c. The channel response is now h (t ) = h (t ) g (t ), with spectrum H (f ) = H (f )G (f ). The channel SNR function is unchanged, since H (f) G (f) H (f) SNR = 2 = S =SNR c (f). (15.2) c(f ) = H (f) S n (f) G(f) n (f) S n We conclude that an equivalent channel model is r (t ) = s (t ) h (t )+ n (t ), (15.3) where h (t ) = h (t ) g (t ) and n (t ) is AWGN with p.s.d. S n. Thus without loss of generality or optimality, we may continue to assume that the noise is white Channel capacity and water-pouring Shannon showed that the total capacity of this channel is maximized subject to a power constraint S x (f) df P if the transmit p.s.d. S x (f) is chosen as { S x (f) = K (P ) 1/ SNR c (f), if K (P ) 1/ SNR c (f) 0; 0, if K (P ) 1/ SNR c (f) 0, (15.4) where K (P ) is a constant chosen so that the power constraint S x (f) df P is satisfied. Note that 1/ SNR 2 2 c (f) = S n / H (f) = S n (f)/ H (f). See Appendix 10-A for a proof. This result is illustrated by the water-pouring diagram of Figure 10.1.

3 15.2. CHANNEL CAPACITY AND WATER-POURING 207 K(P) 1/SNR c (f) S x (f) W frequency f Figure Determination of optimum transmit p.s.d. S x (f ) by water-pouring. On a typical channel, the function 1/ SNR c (f) forms a bowl into which water S x (f) is poured. The height of the water is a constant K, and the depth is S x (f) = K 1/ SNR c (f) for all frequencies for which K 1/ SNR c (f) 0. Water is poured until its total volume S x (f) df is equal to P, atwhich time itsheightis K (P ). The set of frequencies B = {f S x (f ) > 0} is called the capacity-achieving band. On a typical channel, B is one continuous band of width W, as shown in Figure However, if the function 1/ SNR c (f ) has multiple local minima, then depending on P the capacity-achieving band may consist of two or more frequency intervals. Note that if H (f ) is an ideal brick-wall channel over some band B of width W and S n (f ) is constant over B so that SNR c (f ) is constant, then the water-pouring result prescribes that S x (f ) should be constant over the band B and zero elsewhere, as expected. There is usually little loss in capacity if S x (f ) is chosen to be flat over the capacity-achieving band B. This observation justifies the use of conventional PAM or QAM modulation over B. If B is made up of more than one frequency interval, then PAM or QAM modulation may be used over each interval independently, in a form of frequency division multiplexing. In practice, a band somewhat narrower than B is often used. To achieve capacity, the transmitted signal s (t ) should actually be a zero-mean Gaussian process with p.s.d. S x (f ). Of course a digitally modulated signal s (t ) cannot actually be Gaussian. All that is required in practice is that the original data sequence {a k } resemble an i.i.d. Gaussian sequence statistically, to the accuracy of the continuous approximation. This is accomplished by shaping techniques such as those mentioned in Section 9.3. Multicarrier transmission is an alternative modulation technique that is directly inspired by water-pouring. The band B is effectively split into many parallel subbands of small width W. Within each subband, the function 1/ SNR c (f ) may be assumed to be nearly constant, so modulation and coding techniques suitable for ideal band-limited channels may be used. Intersymbol interference is avoided by allowing an appropriate guard band between symbols; the relative rate loss due to this guard band becomes negligible as the signaling interval T =1/ f becomes large. Capacity may be approached if the power P (f ) in a subband at frequency f is allocated in water-pouring fashion, P (f ) = S x (f ) f, and if the rate R (f ) is made to approach the subband capacity C (f ) = f log (1 + SNR(f )) b/s, where SNR(f ) = S x (f )SNR c (f ) isthe signal-to-noise ratio in the subband at frequency f.

4 208 CHAPTER 15. LINEAR GAUSSIAN CHANNELS 15.3 Normalized SNR for linear Gaussian channels The capacity in bits per two dimensions with the optimum S x (f ) is where we define W = B = B df C = log 2 (1 + SNR(f )), (15.5) W df and B SNR(f ) = S x (f )SNR c (f ). (15.6) We may translate this into an equivalent signal-to-noise ratio SNR eq satisfying df log (1 + SNR eq )= log (1 + SNR(f )), (15.7) W B where the logarithms may have any common base. In other words, 1 + SNR eq is the geometric mean of 1 + SNR(f ) over B. Equivalently, C = log (1 + SNR eq ) is the arithmetic mean of C (f ) = log (1 + SNR(f )). Note that if SNR(f ) is constant over B, thensnr eq is equal to this constant. In this way we may roughly characterize an arbitrary linear Gaussian channel by a bandwidth W and a signal-to-noise ratio SNR eq such that the number of available dimensions per second is 2W and the capacity in b/2d is C = log (1 + SNR eq ), as in the ideal channel case. We may proceed to define the normalized signal-to-noise ratio as before: SNR eq SNR norm = 2 ρ 1, (15.8) where ρ is the actual spectral efficiency of a given transmission scheme in b/2d. Since by the Shannon limit we must have ρ <C for reliable transmission, we may again express the Shannon limit by the lower bound SNR norm > 1, and measure the distance to the Shannon limit of a given transmission scheme by its required SNR norm,indb Reduction to equivalent discrete-time model The water-pouring argument allows us to determine the capacity-achieving band B andthusto choose one or more frequency intervals to be used by one or more PAM or QAM modulation schemes. Therefore, following Chapter 3, we henceforth assume that s (t ) is a real or complex equivalent baseband signal of the form s (t) = a k p (t kt ), k where {a k } is a real or complex data sequence and p (t ) is an equivalent baseband impulse response. The transmit power spectral density is then 2 S x (f ) = S x P (f ),

5 15.4. REDUCTION TO EQUIVALENT DISCRETE-TIME MODEL 209 where S x is the average energy of the signal alphabet A per two dimensions and P (f) isthe transform of p(t). To obtain the water-pouring transmit p.s.d. S x (f), assuming that the capacityachieving band B is asinglepassbandintervalofwidth W, the modulation interval should be chosen as T =1/W. This implies that the Nyquist band is B, and that S x (f) =0 outside the Nyquist band, so there is no aliasing. Furthermore, typically S x (f) = 0 at the Nyquist band edges. The channel output is then r(t) = a k f(t kt )+ n(t), k where f(t) = p(t) h(t) is the composite response of the transmitter and channel, and n(t) is additive white Gaussian noise with p.s.d. S n. At the channel output, the signal p.s.d. is S x F (f) 2 = S x P (f) 2 H(f) 2.The output signalto-noise ratio is therefore B S x F (f) 2 df df SNR o = = SNR(f). S n W W In other words, SNR o is the arithmetic mean of SNR(f) over B. We show in Appendix 10-B that this implies that SNR o SNR eq. We now use the principles of optimum detection theory to reduce the continuous-time model r(t) = k a kf(t kt )+ n(t) to an equivalent discrete-time model, without loss of optimality. The signal space is now the Hilbert space W (f) spanned by the set of shifted impulse responses f = {f(t kt ),k Z}. By the results of Chapter 4, the set r = {r k } of outputs of a bank of matched filters is a set of sufficient statistics for detection of the sequence a = {a k },where r k = r(t)f (t kt ) dt. B The set r = {r k } may be obtained by sampling the output z(t) of a single matched filter with response f ( t) attimes kt to give r k = z(kt ), since z(t) = r(τ)f (τ t) dτ. The composite response is then R ff (t) = f(t) f ( t), the autocorrelation function of f(t), whose transform is F (f) 2. Define R k = R ff (kt ); (15.9) thus by Chapter 2 where In D-transform notation, r k = a j R k j + n k, j n k = n(t)f (t kt ) dt. r(d) = a(d)r(d)+ n(d). By the aliasing theorem, the transform of the sequence {R k } is the aliased spectrum 1 S(u) = F ( u m ) 2. T T m

6 210 CHAPTER 15. LINEAR GAUSSIAN CHANNELS If there is no aliasing e.g., if S x (f ) is the water-pouring transmit spectrum then SNR o = S x S(u) = 1 F ( u ) 2. T T In this case the output signal-to-noise ratio SNR o is the same in discrete time, since 2 df = S x F (f) S(u) du, W the arithmetic mean of SNR(u) =(S x /S n )S(u). Furthermore, by Chapter 5, since n k = n(t)f (t kt ) dt, wehave S n B S n E[n j n k ] = S n f (t jt )f(t kt ) dt = S n R ff ((k j)t ) = S n R k j. Thus, up to the constant factor S n, the autocorrelation sequence of the noise sequence n(d) is the sampled autocorrelation sequence {R k = R ff (kt )}. To summarize: Equivalent discrete-time model (via MF): The output sequence {r k } of a T -sampled matched filter is a set of sufficient statistics and has D-transform r(d) = a(d)r(d)+ n (D), (15.10) where R(D) isthe D-transform of the autocorrelation sequence {R ff (kt )}, with spectrum S(u) equal to the aliased spectrum of F (u/t ) 2,and n (D) is a stationary Gaussian sequence with autocorrelation function R n n (D) = S nr(d). (15.11) 15.5 Spectral factorization in discrete time We now give a unique factorization theorem for a discrete-time autocorrelation function R(D), or equivalently for its spectrum S(u). In order for this theorem to hold, we require two technical assumptions, called the Paley-Wiener conditions: (a) The spectrum S(u) must be integrable: S(u) du <. (This always holds when S(u) is the aliased spectrum of an integrable spectrum F (f) 2.) (b) The spectrum S(u) must be log-integrable: log S(u) du >. (This holds if S(u) > 0 everywhere, or even if S(u) has only a countable set of algebraic zeroes; it fails if S(u) =0 over any band of frequencies of nonzero measure.) Then we have the following discrete-time spectral factorization theorem:

7 15.5. SPECTRAL FACTORIZATION IN DISCRETE TIME 211 Theorem 15.1 (Discrete-time spectral factorization) Let {R k } be any autocorrelation function with D-transform R(D) = R (D 1 ) and real spectrum S(u) 0, and assume that S(u) is integrable and log-integrable. Then there exists a unique canonical (causal, monic, minimumphase) impulse response {h k } with D-transform h(d) =1 + h 1 D + h 2 D and spectrum H(u) such that R(D) = A 2 h(d)h (D 1 ); (15.12) S(u) = A 2 H(u) 2, (15.13) where the constant A 2 > 0 is the geometric mean of the spectrum S(u); i.e., log A 2 = log S(u) du, (15.14) where the logarithms may have any common base. Appendix 10-C sketches a proof of this theorem. Appendix 10-D shows how this factorization may be understood as a Cholesky factorization of a Toeplitz Gram matrix. For uniqueness and stability, h(d) must be minimum-phase. If h(d) is rational, then minimum-phase means that all poles and zeroes are on or outside the complex unit circle. (In order for h(d) tobein L 2, all poles must be outside the unit circle.) A causal, monic, minimum-phase h(d) is called canonical. The spectral factorization theorem thus says that if a real spectrum S(u) satisfies the Paley- Wiener conditions, then there exists a square root AH(u) of S(u) such that S(u) = A 2 H(u) 2. By requiring that h(d) be canonical, the factorization is made unique. Spectral factorization is straightforward if R(D) is finite; i.e., R k =0 for k> L for some degree L. (Note that R(D) must be finite if f(t) is a finite impulse response, and that R(D) may be approximated arbitrarily closely by a sufficiently long finite sequence.) Then R(D) has precisely 2L complex roots. By the Hermitian symmetry of autocorrelation functions, R(D) = R (D 1 ), which implies that if α is arootthen(α ) 1 is a root; i.e., the roots are grouped in L conjugate inverse pairs. Since α 1 implies (α ) 1 1, at least one root of each pair satisfies α 1. We may thus write R(D) = K (1 α 1 D)(1 (α i ) 1 D 1 ) 1 i L where K is some constant and {α i, 1 i L} is a set of L roots with α i 1. Then R(D) = Kh(D)h (D 1 ), where h(d) = (1 αi 1 D) 1 i L is causal, monic and minimum-phase i.e., canonical. By uniqueness, h(d) must be the canonical factor of R(D), and K must equal A 2. Note that h(d) has degree L. This procedure generalizes straightforwardly to rational R(D); then h(d) is the rational canonical response that has as its poles and zeroes one of each pair of poles or zeroes of R(D) that lies on or outside the complex unit circle, and h (D 1 ) has as its poles and zeroes the conjugate inverses of the poles and zeroes of h(d). Note that if S(u) is integrable, then R(D) can have no poles on the complex unit circle. i

8 212 CHAPTER 15. LINEAR GAUSSIAN CHANNELS D 1 + Example 15.1 (First-order spectra) Let R(D) havedegree1; i.e., let R(D) = R 1 R 0 + R 1 D,where R 0 is real and positive. In order for the spectrum 2πiu S(u) = R 1 e + R 0 + R 1 e 2πiu = R 0 +2 R 1 cos(2πu + θ(r 1 )) to be nonnegative for all u, we must have 2 R 1 R 0. By the quadratic formula, the roots of R(D) are α = (α ) 1 = R 0 R R 1 2 ; 2R 1 R 0 + R R 1 2 2R 1 = 2R 1 R 0 R R 1. 2 Since 2 R 1 R 0, the discriminant R R 1 2 is nonnegative and its square root is real (and by convention positive). Therefore the root α has larger magnitude, α 1. The unique canonical spectral factorization is therefore where R(D) = A 2 (1 α 1 D)(1 (α ) 1 D 1 ), 1 ( ) A 2 = R 2 1 α = R 0 + R R 1. 2 Interestingly, this yields a closed-form formula for the definite integral log (R 0 +2 R 1 cos(2πu)) du = log S(u) du = log A 2. It can be shown that this formula (which may be found in standard integral tables) implies the validity of the formula (??) for A 2 for any rational R(D) The whitened matched filter and the canonical model Section 10.4 shows that without loss of optimality, using the T -sampled outputs of an MF, we obtain the equivalent discrete-time channel model r(d) = a(d)r(d)+ n (D), where n (D) is a stationary Gaussian sequence with autocorrelation function R n n (D) = S nr(d). By the spectral factorization theorem, if R(D) satisfies the Paley-Wiener conditions, then we may write R(D) = A 2 h(d)h (D 1 ) for some canonical h(d) and A 2 > 0. Moreover, we may represent n (D) as n (D) = n(d)ah (D 1 ),

9 15.6. THE WHITENED MATCHED FILTER AND THE CANONICAL MODEL 213 where n(d) is i.i.d. Gaussian with mean zero and symbol variance S n, since then R n n (D) = R nn(d)a 2 h (D 1 )h(d) = S n R(D). Consequently r(d) = a(d)a 2 h(d)h (D 1 )+ n(d)ah (D 1 ). Suppose h(d) has no zeroes on the unit circle; then it is invertible. We may thus filter r(d) by the anticausal filter 1/Ah (D 1 ) to obtain r (D) = a(d)ah(d)+ n(d), where h(d) is canonical and n(d) is i.i.d. Gaussian with symbol variance S n. We call this the canonical discrete-time channel model. It is the discrete-time equivalent of the continuous-time AWGN channel model r(t) = k a k f(t kt ) + n(t) withwhich we started. In this model, the power spectrum 2 S(u) = A 2 H(u) is a T -aliased version of the continuous-time power spectrum F (f) 2. As previously noted, if there is no aliasing, then the output signal-to-noise ratio is unchanged: SNR o = S x F (f) 2 df T = S x S(u) du; S n B S n [0,1) i.e., SNR o is the mean of the discrete-time SNR function SNR(u) =(S x /S n )S(u). The cascade of the continuous-time matched filter with anticausal response f ( t) with the discrete-time filter with anticausal response 1/Ah (D 1 ) is a continuous-time filter with anticausal response ε ( t), called a whitened matched filter (WMF). The samples of the whitened matched filter are the correlations r k = r(t)ε (t kt ) dt of the received sequence r(t) with the time-shifts {ε(t kt ), k Z}. The set ε = {ε(t kt ), k Z} may be seen to be an orthonormal basis for the signal space W (f ) generated by the set f = {f(t kt ),k Z}, as follows. Explicitly, since ε(t) is defined as the cascade of f(t) with the discrete-time filter 1/Ah(D), we have ε(t kt ) =(1/A)f(t kt ) h 1 ε(t (k +1)T ) h 2 ε(t (k +2)T ).... Inverting this equation, we have f(t kt ) = A h j ε(t (k + j)t ). From this it follows that the autocorrelation functions of f and ε are related by j R ff (D) = A 2 h(d)h (D 1 )R εε (D). But R ff (D) = R(D) = A 2 h(d)h (D 1 ). It follows that R εε (D) = 1; i.e., the generators {ε(t kt ), k Z} are orthonormal.

10 214 CHAPTER 15. LINEAR GAUSSIAN CHANNELS The fact that n(d) is an i.i.d. Gaussian sequence with symbol variance S n then follows directly from the orthonormality of the basis {ε(t kt ), k Z}. It may be shown that this continuous-time definition of the WMF and the canonical discretetime model are valid whenever R(D) satisfies the Paley-Wiener conditions, so that the canonical spectral factor h(d) is well-defined, regardless of whether h(d) has zeroes on the unit circle. To summarize: Canonical discrete-time model (via WMF): Provided that the discrete-time autocorrelation function R(D) satisfies the Paley-Wiener conditions (i.e., its spectrum S(u) is log-integrable), the output sequence {r k } of a T -sampled whitened matched filter is well defined and is a set of sufficient statistics with D-transform r (D) = a(d)ah(d)+ n(d), (15.15) where R(D) = A 2 h(d)h (D 1 ) is the canonical spectral factorization of R(D) (so h(d) is canonical and log A 2 = log S(u) du) and n(d) is i.i.d. Gaussian with symbol variance S n Maximum-likelihood sequence detection Given this canonical discrete-time channel model, we now show that we may use the Viterbi algorithm to perform maximum-likelihood sequence detection. We must assume that R(D) and hence h(d) is finite; however, any well-defined autocorrelation sequence R(D) may be approximated arbitrarily closely by a finite sequence. Suppose that h(d) has degree L; i.e., h(d) =1+h 1 D+ + h L D L. Define the signal sequence s(d) = a(d)ah(d). Then s(d) is the output of a finite-impulse-response (FIR) filter with memory-l response Ah(D) when the input is the data sequence a(d). Such a filter may be realized by a shift register with L memory elements, each of which stores an element of the input alphabet A; i.e., the memory at time k stores (a k 1,...,a k L ). Thus the filter is a finite-state machine with A L states. The set of all possible output sequences s(d) may thus be represented by a trellis with A L states. The problem of finding the a(d) that maximizes the likelihood p(r(d) a(d)) is thus a problem of maximum-likelihood sequence detection (MLSD) of the output of a finite-state machine in memoryless (i.i.d. Gaussian) noise. It may therefore be solved by applying the Viterbi algorithm (VA) to this A L -state trellis. In most practical settings, A L is so large so that the VA is impractical. However, since the VA is optimal, we may obtain upper bounds on performance by analyzing its error probability. Also, numerous suboptimal approximations to the VA that approach the optimal performance have been developed. The error probability with the VA is governed by the minimum squared Euclidean distance between possible signal sequences s(d) = a(d)ah(d). If the canonical channel model r(d) = a(d)ah(d)+ n(d) is real and the input alphabet A is an M-PAM alphabet A = b{±1, ±3,..., ±(M 1)}, then a pair of nearby signal sequences s(d) = a(d)ah(d) and s (D) = a (D)Ah(D) is obtained if a(d) and a (D) differ at only one

11 15.8. THE MATCHED-FILTER LOWER BOUND 215 time k by a single level; i.e., a(d) a (D) = ±2bD k. This is called a single-symbol error event. Then s(d) s (D) =(a(d) a (D))Ah(D) = ±2bD k Ah(D). It follows that the squared distance between these two signals is s s 2 =4b 2 A 2 h 2 =4b 2 A 2 H(u) 2 du = 4b 2 S(u) du. The probability of error for such a pair is ( ) (( ) ) s s 2 b 2 Pr(s s) = Q 4σ 2 = Q S(u) du σ 2, where σ 2 = N 0 /2 is the variance per dimension. Moreover, there are two such single-symbol error events at each time k if a k is an interior point in the M -PAM constellation, or one such event if a k is a boundary point. With a complex canonical channel model and a QAM signal alphabet A, the same development goes through, except that there may as many as 4 single-symbol error events if a k is an interior point in the QAM constellation. In many cases, single-symbol error events will be the minimum-distance error events. However, especially if there is severe intersymbol interference in the canonical response h(d), other types of error events may have smaller distance. Whatever the minimum squared distance d 2 min (s, s ) between error events, we also have the pairwise lower bound ( ) d 2 min (s, s ) Pr(E) Q 4σ 2. Thus, as is true in general, we have a pairwise lower bound, an estimate (the union bound estimate), and an upper bound (the union bound) that are all dominated by a term Q (d 2 min (s, s )/4σ 2 ), and that differ only in their multiplicative coefficients (plus additional terms, in the case of the union bound) The matched-filter lower bound Suppose that only a single non-zero signal is sent from an M -PAM or M M -QAM alphabet A with d 2 min =4b 2. Then the analysis above shows that an optimum detector will have error probability approximately equal to (( ) ) b 2 Pr(E) KQ S(u) du σ 2, (15.16) where K is the average number of nearest neighbors in A. Since no sequence detector can possibly outperform this optimal one-shot detector, this value of Pr(E) is called the matchedfilter lower bound (MFB). We define SNR MFB =SNR o =(S x /S n )( S(u) du). Since b 2 /σ 2 =3(S x /S n )/(2 ρ 1) for uncoded M -PAM or M M -QAM transmission, and SNR norm =SNR eq /(2 ρ 1), we may write the MFB as (( ) ) SNRMFB Pr(E) KQ 3SNR norm. SNR eq

12 216 CHAPTER 15. LINEAR GAUSSIAN CHANNELS which shows that the factor γ MFB =SNR MFB /SNR eq may be regarded as a potential ISI gain due to the intersymbol interference in the equivalent discrete-time channel response h(d). Note that γ MFB = 1 and h(d) = 1 if and only if the channel is ideal (S(u) is constant); otherwise γ MFB > 1. From the discussion of the previous section, to the accuracy of the union bound estimate, this potential ISI gain may indeed be realized with uncoded M -PAM or M M -QAM transmission and MLSD, provided that single-symbol error events of the type ±2bD k Ah(D) are theonly minimum-distance error events. We shall see subsequently that with powerful coded modulation this potential ISI gain cannot be realized; the minimum distance between error events is determined by the code and not by fortuitous intersymbol interference Zero-forcing linear equalization The canonical channel model is obtained by linearly filtering the received signal r(t) with a WMF and sampling. Since this may be done without loss of optimality, the optimum linear receiver under any conditions of optimality consists of the cascade of a WMF, a T -spaced sampler, and a further discrete-time filter f (D). The optimum linear receiver under the Nyquist zero-forcing criterion (i.e., no intersymbol interference) therefore must be the cascade of a WMF, a T -spaced sampler, and an ISIeliminating discrete-time filter f (D) = 1/Ah(D). This is called an (optimal) zero-forcing linear equalizer (ZF-LE). If h(d) has zeroes on the unit circle, or equivalently if the spectrum S(u) = A 2 H(u) 2 has nulls (frequencies u where S(u) =0), then 1/h(D) is not well defined and the ZF-LE does not exist. If the ZF-LE exists, then its output sequence is r(d) r (D) = Ah(D) = a(d)+ n(d). Ah(D) Thus r k = r k + n k,where n (D) = n(d)/ah(d) is a stationary Gaussian sequence with autocorrelation function R nn (D) S n R n n (D) = A 2 h(d)h (D 1 ) = R(D). Thus n k is a Gaussian variable with variance R n n,0 = S n n (u) du = S n S(u) 1 du. The signal-to-noise ratio at the decision point with a ZF-LE is therefore ( ) 1 SNR ZF LE = S x S(u) 1 du. (15.17) S n With uncoded transmission using a PAM or QAM constellation with minimum squared distance 4b 2 between signal points, independent symbol-by-symbol minimum-distance decisions

13 ZERO-FORCING DECISION-FEEDBACK EQUALIZATION 217 may be made on each received symbol (disregarding the memory of the noise). The error probability is approximately ( Pr(E) = KQ b 2 ) (( ) ) σ 2 = KQ SNR ZF LE 3SNR norm. SNR eq where σ 2 =(N 0 /2) S(u) 1 du is the variance per dimension, and K is the appropriate error coefficient. In Appendix 10-B we show that SNR ZF LE SNR eq, with equality if and only if S(u) is constant. The factor SNR ZF LE γ ZF LE = 1 SNR eq represents a noise enhancement loss due to noise enhancement in the linear receiver. The factor γ ZF LE =1 if S(u) is flat and 0 if S(u) has nulls (where the ZF-LE does not even exist); γ ZF LE will tend to be near 1 if S(u) is near-flat and near 0 if S(u) has near-nulls Zero-forcing decision-feedback equalization A decision-feedback equalizer is the simplest form of nonlinear equalizer. The basic idea is that if the receiver is operating with a low error rate, then almost all decisions are correct. By assuming correct decisions, we may eliminate all intersymbol interference due to past (decided) symbols. With the canonical discrete-time channel model r(d) = a(d)ah(d)+ n(d), decision-feedback equalization is implemented by the receiver shown in Figure r(d) = a(d)ah(d) + n(d) 1/A + - h(d) - 1 decision â(d) Figure Decision-feedback equalization with canonical channel model. The received sequence is scaled by 1/A to give r (D) = a(d)h(d) + n(d)/a. Given the decisions (..., aˆ k 2, aˆ k 1 ) prior to time k, the intersymbol interference due to these past symbols is subtracted from r k : r k = r k h j â k j. j 1 If the previous decisions are correct, â(d) = a(d), then r (D) = r (D) a(d)(h(d) 1) = a(d)+ n(d)/a, the postcursor interference is completely removed, and r k = a k + n k /A. The signal-to-noise ratio at the decision point of a ZF-DFE receiver is therefore SNR ZF DFE In Appendix 10-B, we show that = S x S n /A 2 = S x S n exp ( SNR ZF LE SNR ZF DFE SNR eq SNR o, with all inequalities strict if S(u) is not a constant. ) log S(u) du. (15.18)

14 218 CHAPTER 15. LINEAR GAUSSIAN CHANNELS We note however that if SNR(u) =(S x /S n )S(u) is large for all u, thensnr ZF DFE SNR eq. In this sense, at high SNRs on channels without nulls, the ZF-DFE is a canonical receiver. Assuming that all past decisions have been correct, with uncoded M -PAM or M M -QAM transmission and symbol-by-symbol decisions the probability of error is (( ) ) SNRZF DFE Pr(E) = K min Q 3SNR norm. SNR eq Therefore if SNR ZF DFE = SNR eq, which on most channels will be true for large SNR, then the curve of Pr(E) vs. SNR norm is independent of the channel characteristics with uncoded transmission and an optimum ZF-DFE receiver; i.e., the distance to the Shannon limit at a given Pr(E) is independent of the channel. (This is Price s result. ) It can be shown that over all factorizations R(D) = A 2 h(d)h (D 1 ), the canonical (minimumphase) factor h(d) maximizes the ratio h 0 2 / h 2, the fraction of the total response in the first coefficient h 0, which yields the largest SNR for a ZF-DFE. What happens when a decision error is made? In practice, errors may propagate for some time, but a DFE will always resynchronize eventually. Thus the probability of an error event given no past error is given by the Pr(E) above; however, the average number of symbol errors per error event will increase. In other words, the error probability exponent (argument of the Q ( ) function) is unaffected, but the error coefficient for bit or symbol error probability will increase. The reader should therefore not be too concerned about error propagation in decision-feedback equalizers Tomlinson-Harashima precoding Decision-feedback equalizers are inherently incompatible with coding. A DFE requires decisions with no delay; but if symbol-by-symbol decisions are made with a coded input sequence, then the resulting raw error rate is often quite high, so error propagation can no longer be regarded as insignificant. If the data sequence is trellis-coded and the VA is used for decoding, one method of resolving this incompatibility is to implement a separate DFE for each surviving path in the decoder ( persurvivor processing ). The computational complexity of course increases by a factor equal to the number of trellis states, which may be prohibitive. Many reduced-complexity approximations to this strategy have been investigated. A preferable method is to use a form of transmitter precoding, sometimes called DFE in the transmitter. Such precoding techniques effectively implement the feedback part of the DFE in the transmitter, where a(d) is known and no errors can occur. This method is applicable only on point-to-point (as opposed to multipoint or broadcast) channels, and requires that the channel characteristics (the equivalent discrete-time response Ah(D)) be known in the transmitter. In general, precoding aims to achieve the following two objectives: (a) a sequence of transmit signals x(d) is sent such that at the channel output, in the absence of noise, an apparently ISI-free sequence y(d) = x(d)h(d) is received, where y(d) may be uncoded or trellis-coded; (b) by allowing redundancy in the channel output sequence y(d), a power constraint on the transmitted sequence x(d) and possibly other desirable criteria are met.

15 TOMLINSON-HARASHIMA PRECODING 219 In this section we describe the simplest kind of transmitter precoding, which was invented independently by Tomlinson and by Harashima about TH precoding was originally introduced for uncoded transmission using M-PAM constellations, but may also be used with coded M-PAM or M M-QAM transmission. It cannot be used with more general shaped constellations, and thus cannot achieve shaping gain. An M-PAM signal set is a scaled version of the lattice constellation C(Z,R(MZ)) = (Z +1/2) [ M/2,M/2] consisting of the M points in the translated integer lattice Z + 1/2 that lie in the length-m interval [ M/2,M/2], which is the Voronoi region R(MZ) of the sublattice MZ. Every real number r maybewrittenuniquelyas x = r + Ma for r R(MZ), a Z. The unique r R(MZ) so defined is called x mod M. Given a signal point y k C(Z,R(MZ)) to convey to the receiver, and knowing the ISI p k = h j x k j due to previous signals, the transmitter generates the transmit signal point j 1 x k = y k p k mod M. Thus x k is in the Voronoi region R(MZ). In fact, the signal points will typically be uniformly distributed over R(MZ). The received point will then be r k = x k + p k + n k = y k + Ma k + n k /A. for some a k Z. Since MZ is a sublattice of Z, the noise-free output y k + Ma k is still in Z +1/2, and a decision d k may be made to the closest point in Z +1/2. The resulting decision d k is in Z + 1/2, but may lie outside R(MZ). The decision d k maythenbereduced to d k mod M R(MZ), which must equal y k if d k = y k + Ma k. Note that all operations are memoryless, so that no error propagation can occur. TH precoding works perfectly well when y(d) is a trellis-coded signal sequence, provided only that y (D) = y(d) + Ma(D) is a valid code sequence whenever y(d) is a valid code sequence and Ma(D) is any sequence of elements of MZ, which holds for practically all trellis codes. The decoder may then decode to the closest code sequence d(d) to the received sequence z(d), and reduce d(d) mod M to y(d) symbolbysymbol, as above. Since r(d) = y (D) + n(d)/a as with a ZF-DFE, the performance will be the same as on a ZF-DFE channel with effective signal-to-noise ratio SNR ZF DFE, which as we have seen is approximately equal to SNR eq at large SNRs. Thus TH precoding permits the combination of trellis coding and ISI-cancelling (DFEequivalent) equalization. Its only problem is that it requires the constellation shape in one dimension to be the Voronoi region R(MZ), which has no shaping gain. It also slightly increases transmit power from (M 2 1)/12, the average energy of an equiprobable discrete M-PAM constellation, to M 2 /12, the average energy of a uniform continuous distribution over R(MZ), but for large constellations this continuous approximation increase is negligible.

16 220 CHAPTER 15. LINEAR GAUSSIAN CHANNELS Summary It may be helpful to summarize these results. The equivalent signal-to-noise ratio SNR eq is the effective SNR of a linear Gaussian channel. It is given by ( ) SNR eq =exp log (1 + SNR(u)) du 1, where SNR(u) =(S x /S n )S(u), where S x and S n are the signal and noise variances, and S(u) = F (u/t ) 2 /T is the spectrum of the discrete-time channel response. The signal-to-noise ratios achieved with uncoded M -PAM or M M -QAM modulation and zero-forcing linear equalization (ZF-LE), zero-forcing decision-feedback equalization (ZF-DFE), or maximum-likelihood sequence detection (assuming that performance is governed by the matched filter lower bound) are given respectively by the harmonic, geometric and arithmetic means ( ) 1 SNR ZF LE = SNR(u) 1 du ; ( ) SNR ZF DFE = exp log SNR(u) du ; SNR MFB = SNR(u) du = SNR o. At high SNRs, SNR ZF DFE SNR eq, and we obtain approximately the same error probability as on an ideal Gaussian channel with SNR equal to SNR eq. The performance with ZF-LE is always worse and the MFB is always better, although as SNR(u) becomes flat all SNRs become the same. On the other hand, if SNR(u) has a near-null, then SNR ZF LE goes to zero, while if SNR(u) has a null, then the ZF-LE does not even exist. The ZF-DFE exists provided that SNR(u) is log-integrable, which will be the case provided that SNR(u) has only algebraic nulls. With TH precoding, we can use trellis codes to achieve the same coding gains as on an ideal channel. With more advanced precoding, we can achieve shaping gains (up to 1.53 db) as well. Therefore in principle precoding allows us to approach the Shannon limit as closely as on an arbitrary high-snr AWGN channel as we can on an ideal AWGN channel. In the following chapters, we will show that the signal-to-noise ratios achieved with MMSE linear equalization (MMSE-LE) and MMSE decision-feedback equalization (MMSE-DFE) are given respectively by ( ) 1 SNR MMSE LE = (1 + SNR(u)) 1 du 1; ( ) SNR MMSE DFE = exp log (1 + SNR(u)) du 1 =SNR eq. Since 1 + SNR(u) has no nulls, both equalizers unconditionally exist. Using the methods of Appendix 10-B, we can show that both MMSE SNRs are better than their ZF counterparts. Moreover, the MMSE-DFE is canonical at all SNRs, so MMSE-type precoding can be used to approach the Shannon limit on all linear Gaussian channels.

17 SUMMARY 221 Appendix 10-A. Proof of the water-pouring result After noise-whitening, we obtain an equivalent channel model r (t ) = s (t ) h (t )+ n (t ), where h (t) is a composite response with spectrum H (f), and n (t) is AWGN with p.s.d. S n.the channel SNR function is then 2 SNR c (f) = H (f). S n We wish to maximize the total information rate through this channel subject to a power constraint S x (f) df P on the transmit p.s.d. S x (f). Note that S x (f) must be nonnegative. We approach this problem by dividing the channel into independent parallel subchannels of width f. We assume that f is small enough and H (f ) smooth enough so that H (f ) 2 is approximately constant within each subchannel. Then each subchannel becomes an ideal AWGN channel. If the transmit power allocated to a subchannel of width f at frequency f is P (f ) = S x (f ) f, then the capacity of the subchannel in b/s is C (f ) = f log 2 (1 + S x (f )SNR c (f )). The total capacity in b/s and total transmit power are C = f log 2 (1 + S x (f)snr c (f)); P = S x (f) f. As f 0, these expressions become integrals: C = df log 2 (1 + S x (f)snr c (f)); P = S x (f) df. Using a Lagrange multiplier λ to account for the constraint S x (f) df P, and shifting for convenience to natural logarithms, our problem becomes to choose S x (f) to maximize C + λp = (ln (1 + S x (f)snr c (f)) + λs x (f)) df. Differentiating with respect to S x (f ) at each f and setting each derivative to zero, we obtain which yields SNR c (f) + λ = 0, 1+ S x (f)snr c (f) 1 S x (f ) = K (P ) SNRc (f ), where the constant K (P ) = λ 1 > 0 is chosen so that S x (f) df = P. This equation for S x (f) can be satisfied only when it yields a nonnegative value; if it yields a negative value, then S x (f) must be set to zero. (This implies that the derivative SNR c (f) K (P ) 1 is negative, so the so-called Kuhn-Tucker conditions for a minimum over a convex region are satisfied.) Thus we obtain the water-pouring solution for S x (f) given in (??).

18 222 CHAPTER 15. LINEAR GAUSSIAN CHANNELS Appendix 10-B. Spectral mean inequalities The parameters SNR eq,snr o,snr ZF LE and SNR ZF DFE all have the form of spectral means of the channel SNR function {SNR(f ),f B} (or equivalently of its discrete-time analogue {S(u),u [0, 1)}). In this appendix we give a general treatment of such spectral means, and a general convexity inequality between them. When we speak of a mean of SNR(f ) over B, we are regarding SNR(f ) as a random variable X defined on R + =[0, ) with probability distribution Pr{X x} = f : SNR(f ) x, B where B = W =1/T. This random variable X is deterministic with value K if and only if SNR(f ) is equal to a constant K over B. Given a continuous, real-valued, strictly monotonic function g : R + R, the g-mean M g of SNR(f ) over B is defined such that df g(m g )= g(snr(f )). W B In other words, g(m g ) is equal to the mean of the random variable g(x). More explicitly, since a continuous, strictly monotonic function g has a well-defined inverse g 1 : g(r + ) R +,we may write ( ) df M g = g 1 g(snr(f )). W B If X is deterministic with value K, then g(x) is deterministic with value g(k), so in this case M g = K, which justifies its being called a mean. A continuous function g is convex ( convex cap, concave ) if whenever X is a random variable equal to x 1 with probability (w.p.) θ and x 2 w.p. 1 θ, then E[g(x)] g(e[x]); see Figure It is strictly convex if strict inequality holds whenever x 1 x 2 and 0 < θ < 1. A twice-differentiable function is convex if its second derivative is nonpositive and strictly convex if its second derivative is strictly negative. Similarly, g is (strictly) convex ( convex cup, convex ) if and only if g is (strictly) convex. g(x 2 ) g(x) g(e[x]) E[g(X)] g(x 1 ) x 1 E[X] x 2 Figure Jensen s inequality E[g(X)] g(e[x]) for a convex function g, when X is a random variable equal to x 1 w.p. θ and x 2 w.p. 1 θ.

19 SUMMARY 223 The definition extends immediately to the following more general inequality: Theorem 15.2 (Convexity inequality) Let g be a convex function of x, and let E[X] and E[g(X)] be the means of X and g(x), respectively, under an arbitrary probability distribution. Then E[g(X)] g(e[x]). If g is strictly convex, then equality holds if and only if X is deterministic. In other words, the g-mean M g = g 1 (E[g(X)]) is overbounded by the mean E[X] when g is convex, with strict inequality if g is strictly convex and X is non-deterministic. A more general inequality is the following [Hardy, Littlewood and Pólya, Inequalities]: Theorem 15.3 (g-mean inequality) Let g and h be continuous, strictly monotonic functions such that g is increasing and g(h 1 (x)) is convex. Then M g M h. If g(h 1 (x)) is strictly convex and X is non-deterministic, then the inequality is strict: M g < M h. To apply these inequalities, we make the following observations: SNR o =SNR MFB is the mean of SNR(f ). SNR eq is the g-mean of SNR(f ) with g(x) =log (1 + x). SNR ZF LE is the g-mean of SNR(f ) with g(x) = x 1 (the harmonic mean ). SNR ZF DFE is the g-mean of SNR(f ) with g(x) =log x (the geometric mean ). The g-mean inequality then yields: Since log (1 + x) is strictly convex, SNR eq SNR o =SNR MFB. Since log (1 + x 1 ) is strictly convex, SNR ZF LE SNR eq. Since log (1 + e x ) is strictly convex, SNR ZF DFE SNR eq. Since log x is strictly convex, SNR ZF DFE SNR o. Since log (x 1 ) is strictly convex, SNR ZF LE SNR ZF DFE. Strict inequality holds in all cases unless SNR(f ) is constant.

20 224 CHAPTER 15. LINEAR GAUSSIAN CHANNELS Appendix 10-C. Proof of discrete-time spectral factorization The spectral factorization of Theorem?? may be obtained by expressing log S(u) as a discretetime Fourier transform log S(u) = α k e 2πiuk, whose coefficients {α k } are given by the inverse transform α k = log S(u) e 2πiuk du. Since log S(u) isreal, α k = α k. The Paley-Wiener conditions ensure that log S(u) isin L 1 so that this transform pair exists. The factorization results from grouping the zero, positive and negative terms of {α k } as follows: log A 2 = α 0 = log S(u) du; log H(u) = α k e 2πiuk ; which yields and the formula (??) for A 2. k>0 log H (u) = α k e 2πiuk, To obtain an explicit expression for the coefficients of h(d), consider the sequence ψ(d) = α k D k =log h(d), k<0 2 log S(u) =log A 2 +log H(u)+log H (u) =log A 2 H(u) k>0 whose transform is ψ(e 2πiu )=log H(u) =log h(e 2πiu ). Taking formal derivatives, we have ψ (1) (D) = h(d) h (1) (D), or equivalently h (1) (D) = h(d)ψ (1) (D). Repeated formal differentiation yields the recursive equation k 1 ( k 1 ) h (k) (D) = h (i) (D)ψ (k i) (D), k 1. i i=0 Evaluating this equation at D = 0 and noting that h (k) (0) = (k!)h k and ψ (k) (0) = (k!)α k,we obtain finally h 0 = 1 and the recursive expression k 1 k i h k = h i α k i, k k i=0 which explicitly determines h(d) in terms of the coefficients {α k } of the transform of log S(u). This proves that h(d) is causal and monic. For a proof that h(d) is minimum-phase, see [Papoulis, Signal Analysis, McGraw-Hill, 1984]. 1,

21 SUMMARY 225 Appendix 10-D. Spectral factorization as Cholesky factorization The spectral factorization of Theorem?? may be understood as a Cholesky factorization of a Toeplitz Gram matrix, as follows. The Gram matrix A(f ) of the infinite set of signal space generators f = {f(t kt), k Z} is the infinite Hermitian-symmetric non-negative definite matrix A(f ) = {R k j,j Z,k Z}. We note that each row or column of this matrix is just a shift of the autocorrelation sequence {R k } i.e., A(f ) is an infinite Toeplitz matrix. Infinite Toeplitz matrices have the following properties: (a) An infinite Toeplitz matrix T = {t k j } is completely characterized by the sequence {t k }, or by its D-transform t(d), or by its spectrum T (u); (b) The eigenvectors of any Toeplitz matrix are the vectors {e 2πiuk, k Z} for u [0, 1), and the associated eigenvalues are the corresponding values T (u) of its spectrum; (c) The product TU of two such matrices is characterized by the product t(d)u(d) ofthe associated D-transforms, or by the product T (u)u(u) of the associated spectra, and does not depend on order i.e., T and U commute: TU = UT. The Cholesky factorization of A(f ) has the form A(f ) = LD 2 L, where L is a lower triangular monic matrix and D is a nonnegative diagonal matrix. Because of the Toeplitz property of A(f ), both L and D are Toeplitz; i.e., L = {h k j,j Z,k Z}; D = {d k j,j Z,k Z}. Now the monic lower triangular property of L implies that h 0 = 1 and h k =0 for k < 0. Similarly the diagonal property of D implies that d k =0 for k 0. In terms of D-transforms, the factorization A(f ) = LD 2 L therefore becomes R(D) = A 2 h(d)h (D 1 ), where A = d 0 is the D-transform of {d k } and h(d) =1 + h 1 D + h 2 D is the D-transform of {h k }. Hence h(d) iscausal(h k =0 for k< 0) and monic (h 0 = 1).

22 226 CHAPTER 15. LINEAR GAUSSIAN CHANNELS Appendix 10-E. Spectral factorization and innovations sequences Spectral factorization permits a canonical representation of stationary Gaussian sequences in terms of innovations sequences, as follows. Let {n k } be an i.i.d. Gaussian sequence, where each n k has mean zero and variance 1. The autocorrelation sequence of {n k } is then { R nn,j = E[n k n 1, if j = 0; k j ]= 0, if j 0; i.e., R nn (D) = 1. Now let {n k } be a sequence obtained by passing {n k } through a filter with response {g k }; i.e., n (D) = n(d)g(d). Then it is easily shown that {n k } is a stationary (i.e., E[n k n k j ]= R n n,j, independent of k) Gaussian sequence with R n n (D) = R nn(d)g(d)g (D 1 )= g(d)g (D 1 ). k Conversely, suppose that {n } is a stationary Gaussian sequence with a given autocorrelation function R n n (D). Assuming that S n n (u) satisfies the Paley-Wiener conditions, we may write R n n (D) = A2 h(d)h (D 1 ) for some canonical h(d) and A 2 > 0. Since a zero-mean Gaussian sequence is entirely characterized by its second-order statistics, we may represent n (D) as the output of a filter with response Ah(D) (or Ah (D 1 )) when the input is an i.i.d. Gaussian innovations sequence with R nn (D) =1, n (D) = n(d)ah(d), since the autocorrelation sequence of n (D) will then be R n n (D) = A2 h(d)h (D 1 ).

Data Detection for Controlled ISI. h(nt) = 1 for n=0,1 and zero otherwise.

Data Detection for Controlled ISI. h(nt) = 1 for n=0,1 and zero otherwise. Data Detection for Controlled ISI *Symbol by symbol suboptimum detection For the duobinary signal pulse h(nt) = 1 for n=0,1 and zero otherwise. The samples at the output of the receiving filter(demodulator)

More information

Capacity Penalty due to Ideal Zero-Forcing Decision-Feedback Equalization

Capacity Penalty due to Ideal Zero-Forcing Decision-Feedback Equalization Capacity Penalty due to Ideal Zero-Forcing Decision-Feedback Equalization John R. Barry, Edward A. Lee, and David. Messerschmitt John R. Barry, School of Electrical Engineering, eorgia Institute of Technology,

More information

Shannon meets Wiener II: On MMSE estimation in successive decoding schemes

Shannon meets Wiener II: On MMSE estimation in successive decoding schemes Shannon meets Wiener II: On MMSE estimation in successive decoding schemes G. David Forney, Jr. MIT Cambridge, MA 0239 USA forneyd@comcast.net Abstract We continue to discuss why MMSE estimation arises

More information

a) Find the compact (i.e. smallest) basis set required to ensure sufficient statistics.

a) Find the compact (i.e. smallest) basis set required to ensure sufficient statistics. Digital Modulation and Coding Tutorial-1 1. Consider the signal set shown below in Fig.1 a) Find the compact (i.e. smallest) basis set required to ensure sufficient statistics. b) What is the minimum Euclidean

More information

This examination consists of 11 pages. Please check that you have a complete copy. Time: 2.5 hrs INSTRUCTIONS

This examination consists of 11 pages. Please check that you have a complete copy. Time: 2.5 hrs INSTRUCTIONS THE UNIVERSITY OF BRITISH COLUMBIA Department of Electrical and Computer Engineering EECE 564 Detection and Estimation of Signals in Noise Final Examination 6 December 2006 This examination consists of

More information

Lecture 12. Block Diagram

Lecture 12. Block Diagram Lecture 12 Goals Be able to encode using a linear block code Be able to decode a linear block code received over a binary symmetric channel or an additive white Gaussian channel XII-1 Block Diagram Data

More information

Signal Design for Band-Limited Channels

Signal Design for Band-Limited Channels Wireless Information Transmission System Lab. Signal Design for Band-Limited Channels Institute of Communications Engineering National Sun Yat-sen University Introduction We consider the problem of signal

More information

Statistical and Adaptive Signal Processing

Statistical and Adaptive Signal Processing r Statistical and Adaptive Signal Processing Spectral Estimation, Signal Modeling, Adaptive Filtering and Array Processing Dimitris G. Manolakis Massachusetts Institute of Technology Lincoln Laboratory

More information

BASICS OF DETECTION AND ESTIMATION THEORY

BASICS OF DETECTION AND ESTIMATION THEORY BASICS OF DETECTION AND ESTIMATION THEORY 83050E/158 In this chapter we discuss how the transmitted symbols are detected optimally from a noisy received signal (observation). Based on these results, optimal

More information

392D: Coding for the AWGN Channel Wednesday, January 24, 2007 Stanford, Winter 2007 Handout #6. Problem Set 2 Solutions

392D: Coding for the AWGN Channel Wednesday, January 24, 2007 Stanford, Winter 2007 Handout #6. Problem Set 2 Solutions 392D: Coding for the AWGN Channel Wednesday, January 24, 2007 Stanford, Winter 2007 Handout #6 Problem Set 2 Solutions Problem 2.1 (Cartesian-product constellations) (a) Show that if A is a K-fold Cartesian

More information

that efficiently utilizes the total available channel bandwidth W.

that efficiently utilizes the total available channel bandwidth W. Signal Design for Band-Limited Channels Wireless Information Transmission System Lab. Institute of Communications Engineering g National Sun Yat-sen University Introduction We consider the problem of signal

More information

Decision-Point Signal to Noise Ratio (SNR)

Decision-Point Signal to Noise Ratio (SNR) Decision-Point Signal to Noise Ratio (SNR) Receiver Decision ^ SNR E E e y z Matched Filter Bound error signal at input to decision device Performance upper-bound on ISI channels Achieved on memoryless

More information

Appendix B Information theory from first principles

Appendix B Information theory from first principles Appendix B Information theory from first principles This appendix discusses the information theory behind the capacity expressions used in the book. Section 8.3.4 is the only part of the book that supposes

More information

Performance of small signal sets

Performance of small signal sets 42 Chapter 5 Performance of small signal sets In this chapter, we show how to estimate the performance of small-to-moderate-sized signal constellations on the discrete-time AWGN channel. With equiprobable

More information

Physical Layer and Coding

Physical Layer and Coding Physical Layer and Coding Muriel Médard Professor EECS Overview A variety of physical media: copper, free space, optical fiber Unified way of addressing signals at the input and the output of these media:

More information

Power Spectral Density of Digital Modulation Schemes

Power Spectral Density of Digital Modulation Schemes Digital Communication, Continuation Course Power Spectral Density of Digital Modulation Schemes Mikael Olofsson Emil Björnson Department of Electrical Engineering ISY) Linköping University, SE-581 83 Linköping,

More information

Direct-Sequence Spread-Spectrum

Direct-Sequence Spread-Spectrum Chapter 3 Direct-Sequence Spread-Spectrum In this chapter we consider direct-sequence spread-spectrum systems. Unlike frequency-hopping, a direct-sequence signal occupies the entire bandwidth continuously.

More information

RADIO SYSTEMS ETIN15. Lecture no: Equalization. Ove Edfors, Department of Electrical and Information Technology

RADIO SYSTEMS ETIN15. Lecture no: Equalization. Ove Edfors, Department of Electrical and Information Technology RADIO SYSTEMS ETIN15 Lecture no: 8 Equalization Ove Edfors, Department of Electrical and Information Technology Ove.Edfors@eit.lth.se Contents Inter-symbol interference Linear equalizers Decision-feedback

More information

Es e j4φ +4N n. 16 KE s /N 0. σ 2ˆφ4 1 γ s. p(φ e )= exp 1 ( 2πσ φ b cos N 2 φ e 0

Es e j4φ +4N n. 16 KE s /N 0. σ 2ˆφ4 1 γ s. p(φ e )= exp 1 ( 2πσ φ b cos N 2 φ e 0 Problem 6.15 : he received signal-plus-noise vector at the output of the matched filter may be represented as (see (5-2-63) for example) : r n = E s e j(θn φ) + N n where θ n =0,π/2,π,3π/2 for QPSK, and

More information

Chapter [4] "Operations on a Single Random Variable"

Chapter [4] Operations on a Single Random Variable Chapter [4] "Operations on a Single Random Variable" 4.1 Introduction In our study of random variables we use the probability density function or the cumulative distribution function to provide a complete

More information

SOLUTIONS TO ECE 6603 ASSIGNMENT NO. 6

SOLUTIONS TO ECE 6603 ASSIGNMENT NO. 6 SOLUTIONS TO ECE 6603 ASSIGNMENT NO. 6 PROBLEM 6.. Consider a real-valued channel of the form r = a h + a h + n, which is a special case of the MIMO channel considered in class but with only two inputs.

More information

ECE 564/645 - Digital Communications, Spring 2018 Homework #2 Due: March 19 (In Lecture)

ECE 564/645 - Digital Communications, Spring 2018 Homework #2 Due: March 19 (In Lecture) ECE 564/645 - Digital Communications, Spring 018 Homework # Due: March 19 (In Lecture) 1. Consider a binary communication system over a 1-dimensional vector channel where message m 1 is sent by signaling

More information

ACOMMUNICATION situation where a single transmitter

ACOMMUNICATION situation where a single transmitter IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 50, NO. 9, SEPTEMBER 2004 1875 Sum Capacity of Gaussian Vector Broadcast Channels Wei Yu, Member, IEEE, and John M. Cioffi, Fellow, IEEE Abstract This paper

More information

MAXIMUM LIKELIHOOD SEQUENCE ESTIMATION FROM THE LATTICE VIEWPOINT. By Mow Wai Ho

MAXIMUM LIKELIHOOD SEQUENCE ESTIMATION FROM THE LATTICE VIEWPOINT. By Mow Wai Ho MAXIMUM LIKELIHOOD SEQUENCE ESTIMATION FROM THE LATTICE VIEWPOINT By Mow Wai Ho A thesis submitted in partial fulfillment of the requirements for the Degree of Master of Philosophy Department of Information

More information

Square Root Raised Cosine Filter

Square Root Raised Cosine Filter Wireless Information Transmission System Lab. Square Root Raised Cosine Filter Institute of Communications Engineering National Sun Yat-sen University Introduction We consider the problem of signal design

More information

Chapter 9 Fundamental Limits in Information Theory

Chapter 9 Fundamental Limits in Information Theory Chapter 9 Fundamental Limits in Information Theory Information Theory is the fundamental theory behind information manipulation, including data compression and data transmission. 9.1 Introduction o For

More information

3.9 Diversity Equalization Multiple Received Signals and the RAKE Infinite-length MMSE Equalization Structures

3.9 Diversity Equalization Multiple Received Signals and the RAKE Infinite-length MMSE Equalization Structures Contents 3 Equalization 57 3. Intersymbol Interference and Receivers for Successive Message ransmission........ 59 3.. ransmission of Successive Messages.......................... 59 3.. Bandlimited Channels..................................

More information

Lecture 15: Thu Feb 28, 2019

Lecture 15: Thu Feb 28, 2019 Lecture 15: Thu Feb 28, 2019 Announce: HW5 posted Lecture: The AWGN waveform channel Projecting temporally AWGN leads to spatially AWGN sufficiency of projection: irrelevancy theorem in waveform AWGN:

More information

LECTURE 16 AND 17. Digital signaling on frequency selective fading channels. Notes Prepared by: Abhishek Sood

LECTURE 16 AND 17. Digital signaling on frequency selective fading channels. Notes Prepared by: Abhishek Sood ECE559:WIRELESS COMMUNICATION TECHNOLOGIES LECTURE 16 AND 17 Digital signaling on frequency selective fading channels 1 OUTLINE Notes Prepared by: Abhishek Sood In section 2 we discuss the receiver design

More information

SIGNAL SPACE CONCEPTS

SIGNAL SPACE CONCEPTS SIGNAL SPACE CONCEPTS TLT-5406/0 In this section we familiarize ourselves with the representation of discrete-time and continuous-time communication signals using the concepts of vector spaces. These concepts

More information

Modulation & Coding for the Gaussian Channel

Modulation & Coding for the Gaussian Channel Modulation & Coding for the Gaussian Channel Trivandrum School on Communication, Coding & Networking January 27 30, 2017 Lakshmi Prasad Natarajan Dept. of Electrical Engineering Indian Institute of Technology

More information

Capacity of AWGN channels

Capacity of AWGN channels Chapter 3 Capacity of AWGN channels In this chapter we prove that the capacity of an AWGN channel with bandwidth W and signal-tonoise ratio SNR is W log 2 (1+SNR) bits per second (b/s). The proof that

More information

This examination consists of 10 pages. Please check that you have a complete copy. Time: 2.5 hrs INSTRUCTIONS

This examination consists of 10 pages. Please check that you have a complete copy. Time: 2.5 hrs INSTRUCTIONS THE UNIVERSITY OF BRITISH COLUMBIA Department of Electrical and Computer Engineering EECE 564 Detection and Estimation of Signals in Noise Final Examination 08 December 2009 This examination consists of

More information

Signal Processing for Digital Data Storage (11)

Signal Processing for Digital Data Storage (11) Outline Signal Processing for Digital Data Storage (11) Assist.Prof. Piya Kovintavewat, Ph.D. Data Storage Technology Research Unit Nahon Pathom Rajabhat University Partial-Response Maximum-Lielihood (PRML)

More information

which arises when we compute the orthogonal projection of a vector y in a subspace with an orthogonal basis. Hence assume that P y = A ij = x j, x i

which arises when we compute the orthogonal projection of a vector y in a subspace with an orthogonal basis. Hence assume that P y = A ij = x j, x i MODULE 6 Topics: Gram-Schmidt orthogonalization process We begin by observing that if the vectors {x j } N are mutually orthogonal in an inner product space V then they are necessarily linearly independent.

More information

Lecture 8: MIMO Architectures (II) Theoretical Foundations of Wireless Communications 1. Overview. Ragnar Thobaben CommTh/EES/KTH

Lecture 8: MIMO Architectures (II) Theoretical Foundations of Wireless Communications 1. Overview. Ragnar Thobaben CommTh/EES/KTH MIMO : MIMO Theoretical Foundations of Wireless Communications 1 Wednesday, May 25, 2016 09:15-12:00, SIP 1 Textbook: D. Tse and P. Viswanath, Fundamentals of Wireless Communication 1 / 20 Overview MIMO

More information

Diversity Multiplexing Tradeoff in ISI Channels Leonard H. Grokop, Member, IEEE, and David N. C. Tse, Senior Member, IEEE

Diversity Multiplexing Tradeoff in ISI Channels Leonard H. Grokop, Member, IEEE, and David N. C. Tse, Senior Member, IEEE IEEE TRANSACTIONS ON INFORMATION THEORY, VOL 55, NO 1, JANUARY 2009 109 Diversity Multiplexing Tradeoff in ISI Channels Leonard H Grokop, Member, IEEE, and David N C Tse, Senior Member, IEEE Abstract The

More information

4184 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 51, NO. 12, DECEMBER Pranav Dayal, Member, IEEE, and Mahesh K. Varanasi, Senior Member, IEEE

4184 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 51, NO. 12, DECEMBER Pranav Dayal, Member, IEEE, and Mahesh K. Varanasi, Senior Member, IEEE 4184 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 51, NO. 12, DECEMBER 2005 An Algebraic Family of Complex Lattices for Fading Channels With Application to Space Time Codes Pranav Dayal, Member, IEEE,

More information

L interférence dans les réseaux non filaires

L interférence dans les réseaux non filaires L interférence dans les réseaux non filaires Du contrôle de puissance au codage et alignement Jean-Claude Belfiore Télécom ParisTech 7 mars 2013 Séminaire Comelec Parts Part 1 Part 2 Part 3 Part 4 Part

More information

Weiyao Lin. Shanghai Jiao Tong University. Chapter 5: Digital Transmission through Baseband slchannels Textbook: Ch

Weiyao Lin. Shanghai Jiao Tong University. Chapter 5: Digital Transmission through Baseband slchannels Textbook: Ch Principles of Communications Weiyao Lin Shanghai Jiao Tong University Chapter 5: Digital Transmission through Baseband slchannels Textbook: Ch 10.1-10.5 2009/2010 Meixia Tao @ SJTU 1 Topics to be Covered

More information

Digital Baseband Systems. Reference: Digital Communications John G. Proakis

Digital Baseband Systems. Reference: Digital Communications John G. Proakis Digital Baseband Systems Reference: Digital Communications John G. Proais Baseband Pulse Transmission Baseband digital signals - signals whose spectrum extend down to or near zero frequency. Model of the

More information

Multicarrier transmission DMT/OFDM

Multicarrier transmission DMT/OFDM W. Henkel, International University Bremen 1 Multicarrier transmission DMT/OFDM DMT: Discrete Multitone (wireline, baseband) OFDM: Orthogonal Frequency Division Multiplex (wireless, with carrier, passband)

More information

Symbol Interleaved Parallel Concatenated Trellis Coded Modulation

Symbol Interleaved Parallel Concatenated Trellis Coded Modulation Symbol Interleaved Parallel Concatenated Trellis Coded Modulation Christine Fragouli and Richard D. Wesel Electrical Engineering Department University of California, Los Angeles christin@ee.ucla. edu,

More information

Chapter 7. Error Control Coding. 7.1 Historical background. Mikael Olofsson 2005

Chapter 7. Error Control Coding. 7.1 Historical background. Mikael Olofsson 2005 Chapter 7 Error Control Coding Mikael Olofsson 2005 We have seen in Chapters 4 through 6 how digital modulation can be used to control error probabilities. This gives us a digital channel that in each

More information

A Family of Nyquist Filters Based on Generalized Raised-Cosine Spectra

A Family of Nyquist Filters Based on Generalized Raised-Cosine Spectra Proc. Biennial Symp. Commun. (Kingston, Ont.), pp. 3-35, June 99 A Family of Nyquist Filters Based on Generalized Raised-Cosine Spectra Nader Sheikholeslami Peter Kabal Department of Electrical Engineering

More information

Applications of Lattices in Telecommunications

Applications of Lattices in Telecommunications Applications of Lattices in Telecommunications Dept of Electrical and Computer Systems Engineering Monash University amin.sakzad@monash.edu Oct. 2013 1 Sphere Decoder Algorithm Rotated Signal Constellations

More information

MMSE Decision Feedback Equalization of Pulse Position Modulated Signals

MMSE Decision Feedback Equalization of Pulse Position Modulated Signals SE Decision Feedback Equalization of Pulse Position odulated Signals AG Klein and CR Johnson, Jr School of Electrical and Computer Engineering Cornell University, Ithaca, NY 4853 email: agk5@cornelledu

More information

The E8 Lattice and Error Correction in Multi-Level Flash Memory

The E8 Lattice and Error Correction in Multi-Level Flash Memory The E8 Lattice and Error Correction in Multi-Level Flash Memory Brian M Kurkoski University of Electro-Communications Tokyo, Japan kurkoski@iceuecacjp Abstract A construction using the E8 lattice and Reed-Solomon

More information

MITOCW ocw-6-451_ feb k_512kb-mp4

MITOCW ocw-6-451_ feb k_512kb-mp4 MITOCW ocw-6-451_4-261-09feb2005-220k_512kb-mp4 And what we saw was the performance was quite different in the two regimes. In power-limited regime, typically our SNR is much smaller than one, whereas

More information

Digital Modulation 1

Digital Modulation 1 Digital Modulation 1 Lecture Notes Ingmar Land and Bernard H. Fleury Navigation and Communications () Department of Electronic Systems Aalborg University, DK Version: February 5, 27 i Contents I Basic

More information

Lecture 9: Diversity-Multiplexing Tradeoff Theoretical Foundations of Wireless Communications 1. Overview. Ragnar Thobaben CommTh/EES/KTH

Lecture 9: Diversity-Multiplexing Tradeoff Theoretical Foundations of Wireless Communications 1. Overview. Ragnar Thobaben CommTh/EES/KTH : Diversity-Multiplexing Tradeoff Theoretical Foundations of Wireless Communications 1 Rayleigh Wednesday, June 1, 2016 09:15-12:00, SIP 1 Textbook: D. Tse and P. Viswanath, Fundamentals of Wireless Communication

More information

Introduction to binary block codes

Introduction to binary block codes 58 Chapter 6 Introduction to binary block codes In this chapter we begin to study binary signal constellations, which are the Euclidean-space images of binary block codes. Such constellations have nominal

More information

Digital Band-pass Modulation PROF. MICHAEL TSAI 2011/11/10

Digital Band-pass Modulation PROF. MICHAEL TSAI 2011/11/10 Digital Band-pass Modulation PROF. MICHAEL TSAI 211/11/1 Band-pass Signal Representation a t g t General form: 2πf c t + φ t g t = a t cos 2πf c t + φ t Envelope Phase Envelope is always non-negative,

More information

EE4304 C-term 2007: Lecture 17 Supplemental Slides

EE4304 C-term 2007: Lecture 17 Supplemental Slides EE434 C-term 27: Lecture 17 Supplemental Slides D. Richard Brown III Worcester Polytechnic Institute, Department of Electrical and Computer Engineering February 5, 27 Geometric Representation: Optimal

More information

3. ESTIMATION OF SIGNALS USING A LEAST SQUARES TECHNIQUE

3. ESTIMATION OF SIGNALS USING A LEAST SQUARES TECHNIQUE 3. ESTIMATION OF SIGNALS USING A LEAST SQUARES TECHNIQUE 3.0 INTRODUCTION The purpose of this chapter is to introduce estimators shortly. More elaborated courses on System Identification, which are given

More information

Maximum Likelihood Sequence Detection

Maximum Likelihood Sequence Detection 1 The Channel... 1.1 Delay Spread... 1. Channel Model... 1.3 Matched Filter as Receiver Front End... 4 Detection... 5.1 Terms... 5. Maximum Lielihood Detection of a Single Symbol... 6.3 Maximum Lielihood

More information

EE 574 Detection and Estimation Theory Lecture Presentation 8

EE 574 Detection and Estimation Theory Lecture Presentation 8 Lecture Presentation 8 Aykut HOCANIN Dept. of Electrical and Electronic Engineering 1/14 Chapter 3: Representation of Random Processes 3.2 Deterministic Functions:Orthogonal Representations For a finite-energy

More information

ANALYSIS OF A PARTIAL DECORRELATOR IN A MULTI-CELL DS/CDMA SYSTEM

ANALYSIS OF A PARTIAL DECORRELATOR IN A MULTI-CELL DS/CDMA SYSTEM ANAYSIS OF A PARTIA DECORREATOR IN A MUTI-CE DS/CDMA SYSTEM Mohammad Saquib ECE Department, SU Baton Rouge, A 70803-590 e-mail: saquib@winlab.rutgers.edu Roy Yates WINAB, Rutgers University Piscataway

More information

On the Shamai-Laroia Approximation for the Information Rate of the ISI Channel

On the Shamai-Laroia Approximation for the Information Rate of the ISI Channel On the Shamai-Laroia Approximation for the Information Rate of the ISI Channel Yair Carmon and Shlomo Shamai (Shitz) Department of Electrical Engineering, Technion - Israel Institute of Technology 2014

More information

A First Course in Digital Communications

A First Course in Digital Communications A First Course in Digital Communications Ha H. Nguyen and E. Shwedyk February 9 A First Course in Digital Communications 1/46 Introduction There are benefits to be gained when M-ary (M = 4 signaling methods

More information

Equalization. Contents. John Barry. October 5, 2015

Equalization. Contents. John Barry. October 5, 2015 Equalization John Barry October 5, 205 School of Electrical and Computer Engineering Georgia Institute of Technology Atlanta, GA 30332-0250 barry@ece.gatech.edu Contents Motivation 2 2 Models and Metrics

More information

Filterbank Optimization with Convex Objectives and the Optimality of Principal Component Forms

Filterbank Optimization with Convex Objectives and the Optimality of Principal Component Forms 100 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 49, NO. 1, JANUARY 2001 Filterbank Optimization with Convex Objectives and the Optimality of Principal Component Forms Sony Akkarakaran, Student Member,

More information

Theory and Problems of Signals and Systems

Theory and Problems of Signals and Systems SCHAUM'S OUTLINES OF Theory and Problems of Signals and Systems HWEI P. HSU is Professor of Electrical Engineering at Fairleigh Dickinson University. He received his B.S. from National Taiwan University

More information

The Sorted-QR Chase Detector for Multiple-Input Multiple-Output Channels

The Sorted-QR Chase Detector for Multiple-Input Multiple-Output Channels The Sorted-QR Chase Detector for Multiple-Input Multiple-Output Channels Deric W. Waters and John R. Barry School of ECE Georgia Institute of Technology Atlanta, GA 30332-0250 USA {deric, barry}@ece.gatech.edu

More information

Timing Recovery at Low SNR Cramer-Rao bound, and outperforming the PLL

Timing Recovery at Low SNR Cramer-Rao bound, and outperforming the PLL T F T I G E O R G A I N S T I T U T E O H E O F E A L P R O G R ESS S A N D 1 8 8 5 S E R V L O G Y I C E E C H N O Timing Recovery at Low SNR Cramer-Rao bound, and outperforming the PLL Aravind R. Nayak

More information

Lecture 9: Diversity-Multiplexing Tradeoff Theoretical Foundations of Wireless Communications 1

Lecture 9: Diversity-Multiplexing Tradeoff Theoretical Foundations of Wireless Communications 1 : Diversity-Multiplexing Tradeoff Theoretical Foundations of Wireless Communications 1 Rayleigh Friday, May 25, 2018 09:00-11:30, Kansliet 1 Textbook: D. Tse and P. Viswanath, Fundamentals of Wireless

More information

A Thesis for the Degree of Master. An Improved LLR Computation Algorithm for QRM-MLD in Coded MIMO Systems

A Thesis for the Degree of Master. An Improved LLR Computation Algorithm for QRM-MLD in Coded MIMO Systems A Thesis for the Degree of Master An Improved LLR Computation Algorithm for QRM-MLD in Coded MIMO Systems Wonjae Shin School of Engineering Information and Communications University 2007 An Improved LLR

More information

Time Series Analysis: 4. Linear filters. P. F. Góra

Time Series Analysis: 4. Linear filters. P. F. Góra Time Series Analysis: 4. Linear filters P. F. Góra http://th-www.if.uj.edu.pl/zfs/gora/ 2012 Linear filters in the Fourier domain Filtering: Multiplying the transform by a transfer function. g n DFT G

More information

Digital Communications

Digital Communications Digital Communications Chapter 9 Digital Communications Through Band-Limited Channels Po-Ning Chen, Professor Institute of Communications Engineering National Chiao-Tung University, Taiwan Digital Communications:

More information

IN this paper, we show that the scalar Gaussian multiple-access

IN this paper, we show that the scalar Gaussian multiple-access 768 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 50, NO. 5, MAY 2004 On the Duality of Gaussian Multiple-Access and Broadcast Channels Nihar Jindal, Student Member, IEEE, Sriram Vishwanath, and Andrea

More information

Lattice Reduction Aided Precoding for Multiuser MIMO using Seysen s Algorithm

Lattice Reduction Aided Precoding for Multiuser MIMO using Seysen s Algorithm Lattice Reduction Aided Precoding for Multiuser MIMO using Seysen s Algorithm HongSun An Student Member IEEE he Graduate School of I & Incheon Korea ahs3179@gmail.com Manar Mohaisen Student Member IEEE

More information

Contents. A Information Measures 544

Contents. A Information Measures 544 Contents 5 Generalized Decision Feedback Equalization 424 5.1 Information Theoretic Approach to Decision Feedback.................... 425 5.1.1 MMSE Estimation and Conditional Entropy.....................

More information

Lattices for Communication Engineers

Lattices for Communication Engineers Lattices for Communication Engineers Jean-Claude Belfiore Télécom ParisTech CNRS, LTCI UMR 5141 February, 22 2011 Nanyang Technological University - SPMS Part I Introduction Signal Space The transmission

More information

Minimum BER Linear Transceivers for Block. Communication Systems. Lecturer: Tom Luo

Minimum BER Linear Transceivers for Block. Communication Systems. Lecturer: Tom Luo Minimum BER Linear Transceivers for Block Communication Systems Lecturer: Tom Luo Outline Block-by-block communication Abstract model Applications Current design techniques Minimum BER precoders for zero-forcing

More information

GEORGIA INSTITUTE OF TECHNOLOGY SCHOOL OF ELECTRICAL AND COMPUTER ENGINEERING Final Examination - Fall 2015 EE 4601: Communication Systems

GEORGIA INSTITUTE OF TECHNOLOGY SCHOOL OF ELECTRICAL AND COMPUTER ENGINEERING Final Examination - Fall 2015 EE 4601: Communication Systems GEORGIA INSTITUTE OF TECHNOLOGY SCHOOL OF ELECTRICAL AND COMPUTER ENGINEERING Final Examination - Fall 2015 EE 4601: Communication Systems Aids Allowed: 2 8 1/2 X11 crib sheets, calculator DATE: Tuesday

More information

Solutions to Selected Problems

Solutions to Selected Problems Solutions to Selected Problems from Madhow: Fundamentals of Digital Communication and from Johannesson & Zigangirov: Fundamentals of Convolutional Coding Saif K. Mohammed Department of Electrical Engineering

More information

Equalization. Chapter 3. x k. y k. z k. ( t ) mod#

Equalization. Chapter 3. x k. y k. z k. ( t ) mod# Contents 3 Equalization 6 3. Intersymbol Interference and Receivers for Successive Message ransmission........ 63 3.. ransmission of Successive Messages.......................... 63 3.. Bandlimited Channels..................................

More information

Sample Problems for the 9th Quiz

Sample Problems for the 9th Quiz Sample Problems for the 9th Quiz. Draw the line coded signal waveform of the below line code for 0000. (a Unipolar nonreturn-to-zero (NRZ signaling (b Polar nonreturn-to-zero (NRZ signaling (c Unipolar

More information

Truncation for Low Complexity MIMO Signal Detection

Truncation for Low Complexity MIMO Signal Detection 1 Truncation for Low Complexity MIMO Signal Detection Wen Jiang and Xingxing Yu School of Mathematics Georgia Institute of Technology, Atlanta, Georgia, 3033 Email: wjiang@math.gatech.edu, yu@math.gatech.edu

More information

Error Correction and Trellis Coding

Error Correction and Trellis Coding Advanced Signal Processing Winter Term 2001/2002 Digital Subscriber Lines (xdsl): Broadband Communication over Twisted Wire Pairs Error Correction and Trellis Coding Thomas Brandtner brandt@sbox.tugraz.at

More information

Linear Optimum Filtering: Statement

Linear Optimum Filtering: Statement Ch2: Wiener Filters Optimal filters for stationary stochastic models are reviewed and derived in this presentation. Contents: Linear optimal filtering Principle of orthogonality Minimum mean squared error

More information

Example: Bipolar NRZ (non-return-to-zero) signaling

Example: Bipolar NRZ (non-return-to-zero) signaling Baseand Data Transmission Data are sent without using a carrier signal Example: Bipolar NRZ (non-return-to-zero signaling is represented y is represented y T A -A T : it duration is represented y BT. Passand

More information

UNBIASED MAXIMUM SINR PREFILTERING FOR REDUCED STATE EQUALIZATION

UNBIASED MAXIMUM SINR PREFILTERING FOR REDUCED STATE EQUALIZATION UNBIASED MAXIMUM SINR PREFILTERING FOR REDUCED STATE EQUALIZATION Uyen Ly Dang 1, Wolfgang H. Gerstacker 1, and Dirk T.M. Slock 1 Chair of Mobile Communications, University of Erlangen-Nürnberg, Cauerstrasse

More information

EE4512 Analog and Digital Communications Chapter 4. Chapter 4 Receiver Design

EE4512 Analog and Digital Communications Chapter 4. Chapter 4 Receiver Design Chapter 4 Receiver Design Chapter 4 Receiver Design Probability of Bit Error Pages 124-149 149 Probability of Bit Error The low pass filtered and sampled PAM signal results in an expression for the probability

More information

Chapter 7: Channel coding:convolutional codes

Chapter 7: Channel coding:convolutional codes Chapter 7: : Convolutional codes University of Limoges meghdadi@ensil.unilim.fr Reference : Digital communications by John Proakis; Wireless communication by Andreas Goldsmith Encoder representation Communication

More information

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2. APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product

More information

Approximately achieving the feedback interference channel capacity with point-to-point codes

Approximately achieving the feedback interference channel capacity with point-to-point codes Approximately achieving the feedback interference channel capacity with point-to-point codes Joyson Sebastian*, Can Karakus*, Suhas Diggavi* Abstract Superposition codes with rate-splitting have been used

More information

Time Series Analysis: 4. Digital Linear Filters. P. F. Góra

Time Series Analysis: 4. Digital Linear Filters. P. F. Góra Time Series Analysis: 4. Digital Linear Filters P. F. Góra http://th-www.if.uj.edu.pl/zfs/gora/ 2018 Linear filters Filtering in Fourier domain is very easy: multiply the DFT of the input by a transfer

More information

Contents Equalization 148

Contents Equalization 148 Contents 3 Equalization 48 3. Intersymbol Interference and Receivers for Successive Message ransmission........ 50 3.. ransmission of Successive Messages.......................... 50 3.. Bandlimited Channels..................................

More information

SIPCom8-1: Information Theory and Coding Linear Binary Codes Ingmar Land

SIPCom8-1: Information Theory and Coding Linear Binary Codes Ingmar Land SIPCom8-1: Information Theory and Coding Linear Binary Codes Ingmar Land Ingmar Land, SIPCom8-1: Information Theory and Coding (2005 Spring) p.1 Overview Basic Concepts of Channel Coding Block Codes I:

More information

Digital Transmission Methods S

Digital Transmission Methods S Digital ransmission ethods S-7.5 Second Exercise Session Hypothesis esting Decision aking Gram-Schmidt method Detection.K.K. Communication Laboratory 5//6 Konstantinos.koufos@tkk.fi Exercise We assume

More information

INFORMATION PROCESSING ABILITY OF BINARY DETECTORS AND BLOCK DECODERS. Michael A. Lexa and Don H. Johnson

INFORMATION PROCESSING ABILITY OF BINARY DETECTORS AND BLOCK DECODERS. Michael A. Lexa and Don H. Johnson INFORMATION PROCESSING ABILITY OF BINARY DETECTORS AND BLOCK DECODERS Michael A. Lexa and Don H. Johnson Rice University Department of Electrical and Computer Engineering Houston, TX 775-892 amlexa@rice.edu,

More information

Optimal Sequences and Sum Capacity of Synchronous CDMA Systems

Optimal Sequences and Sum Capacity of Synchronous CDMA Systems Optimal Sequences and Sum Capacity of Synchronous CDMA Systems Pramod Viswanath and Venkat Anantharam {pvi, ananth}@eecs.berkeley.edu EECS Department, U C Berkeley CA 9470 Abstract The sum capacity of

More information

The Capacity of Finite Abelian Group Codes Over Symmetric Memoryless Channels Giacomo Como and Fabio Fagnani

The Capacity of Finite Abelian Group Codes Over Symmetric Memoryless Channels Giacomo Como and Fabio Fagnani IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 55, NO. 5, MAY 2009 2037 The Capacity of Finite Abelian Group Codes Over Symmetric Memoryless Channels Giacomo Como and Fabio Fagnani Abstract The capacity

More information

ADAPTIVE FILTER ALGORITHMS. Prepared by Deepa.T, Asst.Prof. /TCE

ADAPTIVE FILTER ALGORITHMS. Prepared by Deepa.T, Asst.Prof. /TCE ADAPTIVE FILTER ALGORITHMS Prepared by Deepa.T, Asst.Prof. /TCE Equalization Techniques Fig.3 Classification of equalizers Equalizer Techniques Linear transversal equalizer (LTE, made up of tapped delay

More information

Lecture 6: Gaussian Channels. Copyright G. Caire (Sample Lectures) 157

Lecture 6: Gaussian Channels. Copyright G. Caire (Sample Lectures) 157 Lecture 6: Gaussian Channels Copyright G. Caire (Sample Lectures) 157 Differential entropy (1) Definition 18. The (joint) differential entropy of a continuous random vector X n p X n(x) over R is: Z h(x

More information

EE6604 Personal & Mobile Communications. Week 15. OFDM on AWGN and ISI Channels

EE6604 Personal & Mobile Communications. Week 15. OFDM on AWGN and ISI Channels EE6604 Personal & Mobile Communications Week 15 OFDM on AWGN and ISI Channels 1 { x k } x 0 x 1 x x x N- 2 N- 1 IDFT X X X X 0 1 N- 2 N- 1 { X n } insert guard { g X n } g X I n { } D/A ~ si ( t) X g X

More information

Supplementary Figure 1: Scheme of the RFT. (a) At first, we separate two quadratures of the field (denoted by and ); (b) then, each quadrature

Supplementary Figure 1: Scheme of the RFT. (a) At first, we separate two quadratures of the field (denoted by and ); (b) then, each quadrature Supplementary Figure 1: Scheme of the RFT. (a At first, we separate two quadratures of the field (denoted by and ; (b then, each quadrature undergoes a nonlinear transformation, which results in the sine

More information

2A1H Time-Frequency Analysis II

2A1H Time-Frequency Analysis II 2AH Time-Frequency Analysis II Bugs/queries to david.murray@eng.ox.ac.uk HT 209 For any corrections see the course page DW Murray at www.robots.ox.ac.uk/ dwm/courses/2tf. (a) A signal g(t) with period

More information

Principles of Communications Lecture 8: Baseband Communication Systems. Chih-Wei Liu 劉志尉 National Chiao Tung University

Principles of Communications Lecture 8: Baseband Communication Systems. Chih-Wei Liu 劉志尉 National Chiao Tung University Principles of Communications Lecture 8: Baseband Communication Systems Chih-Wei Liu 劉志尉 National Chiao Tung University cwliu@twins.ee.nctu.edu.tw Outlines Introduction Line codes Effects of filtering Pulse

More information