The Viterbi Algorithm and Markov Noise Memory

Similar documents
Performance evaluation for ML sequence detection in ISI channels with Gauss Markov Noise

BASICS OF DETECTION AND ESTIMATION THEORY

Data Detection for Controlled ISI. h(nt) = 1 for n=0,1 and zero otherwise.

Signal Processing for Digital Data Storage (11)

Decision-Point Signal to Noise Ratio (SNR)

Maximum Likelihood Sequence Detection

Statistical and Adaptive Signal Processing

Introduction to Convolutional Codes, Part 1

12.4 Known Channel (Water-Filling Solution)

RADIO SYSTEMS ETIN15. Lecture no: Equalization. Ove Edfors, Department of Electrical and Information Technology

Shannon meets Wiener II: On MMSE estimation in successive decoding schemes

The Viterbi Algorithm EECS 869: Error Control Coding Fall 2009

Lessons in Estimation Theory for Signal Processing, Communications, and Control

Upper Bounds on the Capacity of Binary Intermittent Communication

A t super-channel. trellis code and the channel. inner X t. Y t. S t-1. S t. S t+1. stages into. group two. one stage P 12 / 0,-2 P 21 / 0,2

Soft-Output Trellis Waveform Coding

Turbo Codes. Manjunatha. P. Professor Dept. of ECE. June 29, J.N.N. College of Engineering, Shimoga.

Channel Probing in Communication Systems: Myopic Policies Are Not Always Optimal

6196 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 9, SEPTEMBER 2011

Expectation propagation for signal detection in flat-fading channels

RAO-BLACKWELLISED PARTICLE FILTERS: EXAMPLES OF APPLICATIONS

ADAPTIVE FILTER THEORY

Linear Prediction Theory

IN this paper, we show that the scalar Gaussian multiple-access

The Sorted-QR Chase Detector for Multiple-Input Multiple-Output Channels

Performance of Optimal Digital Page Detection in a Two-Dimensional ISI/AWGN Channel

A Piggybacking Design Framework for Read-and Download-efficient Distributed Storage Codes

IN this paper, we consider the capacity of sticky channels, a

Regression. Oscar García

Machine Learning Lecture Notes

ASIGNIFICANT research effort has been devoted to the. Optimal State Estimation for Stochastic Systems: An Information Theoretic Approach

Sequential Monte Carlo methods for filtering of unobservable components of multidimensional diffusion Markov processes

Digital Baseband Systems. Reference: Digital Communications John G. Proakis

LECTURE 16 AND 17. Digital signaling on frequency selective fading channels. Notes Prepared by: Abhishek Sood

INFORMATION PROCESSING ABILITY OF BINARY DETECTORS AND BLOCK DECODERS. Michael A. Lexa and Don H. Johnson

Recursive Least Squares for an Entropy Regularized MSE Cost Function

Signal Design for Band-Limited Channels

Asymptotic Filtering and Entropy Rate of a Hidden Markov Process in the Rare Transitions Regime

Lecture 8: MIMO Architectures (II) Theoretical Foundations of Wireless Communications 1. Overview. Ragnar Thobaben CommTh/EES/KTH

that efficiently utilizes the total available channel bandwidth W.

Bounds on Mutual Information for Simple Codes Using Information Combining

Direct-Sequence Spread-Spectrum

MODULE -4 BAYEIAN LEARNING

1 1 0, g Exercise 1. Generator polynomials of a convolutional code, given in binary form, are g

THIS paper is aimed at designing efficient decoding algorithms

arxiv:cs/ v1 [cs.it] 11 Sep 2006

ADAPTIVE FILTER ALGORITHMS. Prepared by Deepa.T, Asst.Prof. /TCE

NUMERICAL COMPUTATION OF THE CAPACITY OF CONTINUOUS MEMORYLESS CHANNELS

DETECTION theory deals primarily with techniques for

1 Introduction to information theory

Maximum Likelihood Diffusive Source Localization Based on Binary Observations

Chapter 7: Channel coding:convolutional codes

Inference and estimation in probabilistic time series models

Ralf Koetter, Andrew C. Singer, and Michael Tüchler

Module 1. Introduction to Digital Communications and Information Theory. Version 2 ECE IIT, Kharagpur

BLIND DECONVOLUTION ALGORITHMS FOR MIMO-FIR SYSTEMS DRIVEN BY FOURTH-ORDER COLORED SIGNALS

Model-based Correlation Measure for Gain and Offset Nonuniformity in Infrared Focal-Plane-Array Sensors

PERFECTLY secure key agreement has been studied recently

Capacity Penalty due to Ideal Zero-Forcing Decision-Feedback Equalization

Lecture 8: Bayesian Estimation of Parameters in State Space Models

component risk analysis

Lecture Notes 1: Vector spaces

4 Derivations of the Discrete-Time Kalman Filter

Chapter [4] "Operations on a Single Random Variable"

Introduction to Binary Convolutional Codes [1]

The sequential decoding metric for detection in sensor networks

UNIFORMLY MOST POWERFUL CYCLIC PERMUTATION INVARIANT DETECTION FOR DISCRETE-TIME SIGNALS

STA 4273H: Statistical Machine Learning

Encoder. Encoder 2. ,...,u N-1. 0,v (0) ,u 1. ] v (0) =[v (0) 0,v (1) v (1) =[v (1) 0,v (2) v (2) =[v (2) (a) u v (0) v (1) v (2) (b) N-1] 1,...

Statistical Signal Processing Detection, Estimation, and Time Series Analysis

An Efficient Low-Complexity Technique for MLSE Equalizers for Linear and Nonlinear Channels

MAXIMUM LIKELIHOOD SEQUENCE ESTIMATION FROM THE LATTICE VIEWPOINT. By Mow Wai Ho

EEG- Signal Processing

New Introduction to Multiple Time Series Analysis

Computation of Information Rates from Finite-State Source/Channel Models

LTI Systems, Additive Noise, and Order Estimation

These outputs can be written in a more convenient form: with y(i) = Hc m (i) n(i) y(i) = (y(i); ; y K (i)) T ; c m (i) = (c m (i); ; c m K(i)) T and n

Performance Analysis and Code Optimization of Low Density Parity-Check Codes on Rayleigh Fading Channels

Minimum Necessary Data Rates for Accurate Track Fusion

Optimal Decentralized Control of Coupled Subsystems With Control Sharing

Algebraic Soft-Decision Decoding of Reed Solomon Codes

Revision of Lecture 4

A Randomized Algorithm for the Approximation of Matrices

Temporal Backpropagation for FIR Neural Networks

9 Forward-backward algorithm, sum-product on factor graphs

MMSE Decision Feedback Equalization of Pulse Position Modulated Signals

ADVANCES IN MULTIUSER DETECTION

IN the mobile communication systems, the channel parameters

A New SLNR-based Linear Precoding for. Downlink Multi-User Multi-Stream MIMO Systems

Convolutional Codes. Telecommunications Laboratory. Alex Balatsoukas-Stimming. Technical University of Crete. November 6th, 2008

PREDICTIVE quantization is one of the most widely-used

Joint Optimum Bitwise Decomposition of any. Memoryless Source to be Sent over a BSC. Ecole Nationale Superieure des Telecommunications URA CNRS 820

MMSE DECISION FEEDBACK EQUALIZER FROM CHANNEL ESTIMATE

STA 4273H: Statistical Machine Learning

THE potential for large-scale sensor networks is attracting

392D: Coding for the AWGN Channel Wednesday, January 24, 2007 Stanford, Winter 2007 Handout #6. Problem Set 2 Solutions

WE consider the problem of estimating a time varying

On the Performance of. Golden Space-Time Trellis Coded Modulation over MIMO Block Fading Channels

F denotes cumulative density. denotes probability density function; (.)

3. ESTIMATION OF SIGNALS USING A LEAST SQUARES TECHNIQUE

Transcription:

IEEE RANSACIONS ON INFORMAION HEORY, VOL. 46, NO., JANUARY 2000 29 he Viterbi Algorithm and Marov Noise Memory Alesandar Kavčić, Member, IEEE, and José M. F. Moura, Fellow, IEEE Abstract his wor designs sequence detectors for channels with intersymbol interference (ISI) and correlated (and/or signal-dependent) noise. We describe three major contributions. i) First, by modeling the noise as a finite-order Marov process, we derive the optimal maximum-lielihood sequence detector (MLSD) and the optimal maximum a posteriori (MAP) sequence detector, extending to the correlated noise case the Viterbi algorithm. We show that, when the signal-dependent noise is conditionally Gauss Marov, the branch metrics in the MLSD are computed from the conditional second-order noise statistics. We evaluate the branch metrics using a ban of finite impulse response (FIR) filters. ii) Second, we characterize the error performance of the MLSD and MAP sequence detector. he error analysis of these detectors is complicated by the correlation asymmetry of the channel noise. We derive upper and lower bounds and computationally efficient approximations to these bounds based on the banded structure of the inverses of Gauss Marov covariance matrices. An experimental study shows the tightness of these bounds. iii) Finally, we derive several classes of suboptimal sequence detectors, and demonstrate how these and others available in the literature relate to the MLSD. We compare their error rate performance and their relative computational complexity, and show how the structure of the MLSD and the performance evaluation guide us in choosing a best compromise between several types of suboptimal sequence detectors. Index erms Correlated noise, Gauss Marov processes, intersymbol interference, Marov channel noise, Marov memory, maximum-lielihood sequence detection, Viterbi algorithm. I. INRODUCION In digital communications, the Viterbi algorithm [] is the maximum-lielihood sequence detector (MLSD), [2], [3], for channels with intersymbol interference (ISI) and memorylessn noise. Under these conditions, if fx g is the state-channel input sequence, and fz g is the observable sequence, we have that f (z jz 0; ;z0; x0; ;x)=f (z jx 0; x ) () where f denotes the conditional probability densify function (pdf) of the observable channel output, conditioned on the infinite past observable sequence and on the whole channel input sequence. Equation () states that under an appropriate definition of the channel state-sequence, the ISI white Gauss noise channel is memoryless [4]. In many applications, () is not an appropriate model, because, besides ISI, the channel noise is correlated, with correlation statistics that are possibly signal-dependent. Our research is motivated by detection in high-density magnetic recording [5] where, for example, the statistics of percolation effects between transitions depend strongly on Manuscript received December 22, 997; revised February 7, 999. his wor was supported in part by the National Science Foundation under Grant ECD-8907068. A. Kavčić is with the Division of Engineering and Applied Sciences, Harvard University, Cambridge, MA 0238 USA (e-mail: avcic@hrl.harvard.edu). J. M. F. Moura is with LIDS and the Department of Electrical Engineering and Computer Science, Massachusetts Institute of echnology, Cambridge, MA 0239 USA, on sabbatical leave from the Department of Electrical and Computer Engineering, Carnegie Mellon University, Pittsburgh, PA 523 USA (e-mail: moura@mit.edu). Communicated by E. Soljanin, Associate Editor for Coding echniques. Publisher Item Identifier S 008-9448(00)00033-X. the particular data sequence recorded: percolation is more liely in closely separated transitions (i.e., successive s in the nonreturn to zero inverted (NRZI) recorded data sequence) than in widely separated recorded transitions (i.e., successive data symbols are 0 s). Reference [6] illustrates with experimental evidence this type of signal-dependent noise in high-density magnetic recording. Due to percolation (but also due to nonlinear transition shifts), the noise is correlated across transitions, with the correlation statistics depending on the signal recorded. his signal-dependent correlated noise may overwhelm the white-noise component, becoming the dominant noise feature, and may severely degrade the performance of the sequence detector designed for a white Gauss noise ISI channel. Generalizing the Viterbi algorithm to these more complex channels raises the difficulty that, due to the noise correlation, it is no longer valid to truncate the conditioning in () to a finite sequence. In other words, with general correlated noise there is no appropriate definition for the state sequence under which the channel becomes memoryless, and so the model () becomes invalid. Early wor on this problem includes [7] [9]. More recently, the noise prediction method [0] [2], and the so-called K-step Viterbi detector [3] are attempts to combat the noise correlation in magnetic recording. In this wor, we extend the Viterbi algorithm to channels with ISI and signal-dependent correlated noise: we develop the optimal maximum-lielihood sequence detector (MLSD) and study its error performance. he ey to our approach is our modeling of the noise: we consider ISI channels with correlated noise, which goes well beyond white noise ISI channels, but we impose structure on the statistical properties of the correlated noise. We describe the noise, not as a general Gaussian correlated noise, but as a Gauss Marov correlated noise whose second-order statistics are signal-dependent. We use the Marov assumption to reduce the memory of the noise to finite length. We show in Section II that under the noise Marovianity property and with an appropriate definition for the input state sequence, the following holds: f (z jz 0; ;z0; x0; ;x) = f (z jz 0; ;z 0L; x 0; x ) (2) where L is the so-called Marov memory length of the noise. In Section II, we describe the basic Marovian-memory channel model, and then develop examples of channels with such characteristics. For these Marov channels, the derivation of the Viterbi algorithm becomes a simple exercise in application of Bayes law. We present the Viterbi algorithm in Section III as the solution to maximum a posteriori (MAP) and maximum-lielihood (ML) sequence detection in Marovian-memory channels. In this section, we consider the general case of Marov noise, not necessarily Gauss. he recent paper [4] also notices that the Viterbi algorithm solves the MLSD in channels with additive Marov noise. In Section IV, we derive explicit expressions for the branch metric when the noise in the channel is signal-dependent additive Gauss Marov. o compute efficiently the branch metrics, we use the properties of Gauss Marov processes, namely, the structure of the inverses of the covariance matrices of such processes [5]; we compute the branch metrics by using finite impulse response (FIR) filters. We also discuss in Section IV adaptive implementations of these FIR filters that trac the signal-dependence characteristics of the noise statistics. A second contribution of this research is the error analysis we develop for the MLSD and MAP sequence detector for ISI channels with finite memory correlated noise. he error performance of the MLSD 008 9448/00$0.00 2000 IEEE2000

292 IEEE RANSACIONS ON INFORMAION HEORY, VOL. 46, NO., JANUARY 2000 Fig.. Bloc diagram of a channel with intersymbol interference (ISI) of length and signal-dependent Gauss Marov noise with Marov memory length. he random access memory (RAM s) blocs indicate that both the autoregressive filter and the intersymbol interference are signal-dependent. is considered in Section V. We derive upper and lower bounds for the error probability and by exploiting the structure of the inverse of the covariance matrix of a Gauss Marov process we develop computationally efficient approximations to the bounds. We demonstrate the tightness of these bounds with a simulation study in realistic scenarios. he third major issue we address is the derivation of suboptimal sequence detectors. Guided by the structure of the optimal MLSD and its error performance, we introduce in Section VI new suboptimal sequence detectors and show how these and other suboptimal detectors, including the noise prediction and K-step detectors, relate to the MLSD. We illustrate their error performance and determine within a certain class of suboptimal sequence detectors the best compromise from an error performance and computational complexity perspective. Finally, Section VII concludes the wor. Notation: Before leaving this section, we introduce the basic notation used below. hroughout, we assume that the reader is familiar with the Viterbi algorithm as presented in [3]. We attempt to use the notation in [3] to the extent possible. Column vectors are denoted by underlined characters, matrices are denoted by boldface characters, and the superscript is used to denote matrix and vector transposition. If z is a discrete-time indexed sequence where denotes the time, then the column vector of sequence samples at time through 2 is denoted by z =[z ;z + ; ;z ]. he notation (ajb) N(m; C) denotes that the random vector a, conditioned on an event B, has a normal (joint Gaussian) distribution with mean m and covariance matrix C. he probability of an event A is denoted by P(A). Further notation is introduced in Section V. II. PROCESS MODEL In this section, we establish the appropriate model for ISI channels with signal-dependent correlated noise. We structure the noise so that the resulting channel has finite memory. In Section II-A, we discuss the general Marov chain model with finite memory, and in Section II-B we illustrate this model with examples that are relevant to the ISI channel with correlated noise. A. he Model Let x represent the state of a finite-state machine at time instant. he number of possible states M is finite. We shall assume that the sequence of states x is a Marov chain in the sense that the probability of the machine being in state x conditioned on all states up to the state at time 0, depends only on the state x 0, i.e., P x jx 0 0 =P(x jx 0): (3) he transition probabilities P(x jx 0) may be time-dependent, as originally presented by Forney [3]. his is assumed throughout, although not explicitly indicated. Let the transition between states x 0 and x produce an observable output random variable z, where the statistics of z may depend on the transition (x 0; x ), and time. We assume that the sequence of output variables z is a Marov sequence with a finite memory length L, whose memory is dependent only on the state transition (x 0; x ), i.e., the conditional probability density function (pdf) obeys 0 f z jz ;x0 0 = f z jz 0L 0 ;x 0; x : (4) Again, these pdf s may also depend on time, which we will not show explicitly. B. Examples Example. Channel with ISI and Data-Dependent Gauss Marov Noise: Let a be a sequence of transmitted symbols (bits). We model a channel with intersymbol interference (ISI) of length I 0 as z = y a 0I + n : (5) In (5), y(a 0I ) is the noiseless channel output dependent only on the I + latest transmitted bits. For linear channels, y(a 0I ) may be represented as the convolution of the channel response with the transmitted sequence. We consider the channel to be nonlinear. he additive noise term n is considered to be a signal-dependent Gauss Marov noise process with Marov memory length L, i.e., he vector n = b b(a 0I a 0I )=[b L (a 0I n 0L + a 0I w : (6) ); ;b (a 0I )] collects the L coefficients of an autoregressive filter whose values are dependent on the transmitted symbols a 0I, w is a zero-mean unitvariance white Gaussian noise process, and (a 0I ) is a signal-dependent standard deviation. Fig. depicts the channel described by (5) and (6). As we will derive in Section IV, (4) holds for this channel if the state is chosen as a collection of I +L consecutive transmitted symbols x = a 0I0L+. If we further have 0 P(a ja0 )=P(a ja 0I0L 0 ), then (3) holds too and the channel may be represented by a finite-state machine. If a are binary symbols, then there are 2 I+L states and 2 I+L+ transition branches in this finite-state machine. Example 2. Noncausal ISI and Noncausal Data-Dependent Gauss Marov Noise: he following is a generalization of Example to noncausal ISI and noncausal data-dependent Gauss Marov noise. z = y n = a +I +I + n (7) a +A +A n 0L +L + a+b +B w : (8) Here I 2 I, A 2 A, B 2 B, L 0, and L 2 0. For this +C + channel the state is chosen as x = a, where and +C C =min(i 0 L 2 0 L ;A 0 L 2;B 0 L 2) C 2 =max(i 2;A 2 0 L 2;B 2 0 L 2): Example is a special case of this example when I 2 = A 2 = B 2 = L 2 = 0, I = A = B = 0I, and L = L. Note that in this example the noise model is noncausal Marov, i.e., the noise at the current time may be dependent on past and future noise values. his is the appropriate model, for example, in magnetic recording, where time stands for the spatial variable, and the past and future times stand for the immediately adjacent bit-slots on a physical medium.

IEEE RANSACIONS ON INFORMAION HEORY, VOL. 46, NO., JANUARY 2000 293 Example 3. Non-Marov Observables: Often the observed process is a deterministic part corrupted by additive noise. he additive noise may not be Marov, but often it may be approximated by a Marov process if the memory length L is taen long enough [6]. In this case, the Viterbi algorithm which we present in the next section is the asymptotically optimal detector of the finite-state machine state sequence. III. VIERBI SEQUENCE DEECION A. MAP Sequence Detection We study the estimation of the state sequence x ; ;x K (K>0) given a sequence of observables z ; ;z K when (3) and (4) hold. We assume that the initial state x 0 is nown and that the initial L realizations z0l+; ;z 0 of the observable random variable are also given, where L is the Marovian memory length of the observable sequence z. he maximum a posteriori (MAP) sequence estimate of the state sequence x K =[x ; ;x K ] is the sequence ^x K =[^x ; ; ^x K ] that maximizes the joint conditional pdf, i.e., ^x K = arg max all x f x K;z Kjx 0;z 0L+ 0 : (9) As shorthand notation, denote by f (x; zji.c.) the conditional pdf on the right-hand side of (9), where i.c. stands for initial conditions. Using (3) and (4), we can factor f (x; zji.c.) as f (x; zj i.c.) = K K P x jx 0 0; z 0L+ 0 f z jz 0L+ 0 ;x0 K = = K K = P(x jx 0) f z jz0 0L 0; x : (0) = = In (0), we assumed that P(x jx 0 0; z 0L+ 0 ) = P(x jx 0 0), i.e., that the state transition probabilities are independent of the observables. Using the definition of conditional pdf s, we may further substitute f z jz 0L 0 ;x 0; x = f z0l jx 0; x f z0 0L 0; x f z 0L jx 0; x = () z =0 f z0l jx 0; x dz for the last term in (0). Maximizing (0) is the same as minimizing its negative logarithm. aing the negative logarithm of f (x; zj i.c.), we get 0 ln f (x; zji.c.) = K0 =0 where the MAP branch metrics are 3 MAP z 0L ;x 0; x = 0 ln P(x jx 0) 0 ln 3 MAP z 0L ;x 0; x (2) f z =0 f z 0L jx 0; x : z0l jx 0; x dz (3) he MAP sequence estimate is thus that sequence ^x K for which the sum of the branch metrics in (2) is minimized. B. ML Sequence Detection he maximum-lielihood (ML) sequence estimate is that sequence ^x K, for which the conditional pdf f (zjx; i.c.) =f z Kjx 0 K;z 0L+ 0 (4) is maximized. With arguments similar to those for MAP sequence detection, it follows that the ML estimate is that sequence ^x K for which the following sum of branch metrics is minimized : 0 ln f (zjx; i.c.) = he ML branch metric is K0 =0 3 ML z 0L ;x 0; x = 0 ln 3 ML z 0L ;x 0; x : (5) f z =0 f z 0L jx 0;x : z0l jx 0; x dz (6) C. Viterbi rellis Implementation he implementation of the minimization in (2) and (5) is very similar to the standard implementation described in [2] and [3]. he major difference is that the computation of the branch metrics relies now on a window of observed samples z 0L =[z K0L; ;z ], instead of on just one sample z. Depending on whether we are interested in MAP or ML sequence detection, the branch metrics are given by (3) and (6), respectively. IV. GAUSSIAN SAISICS AND BRANCH MERICS We now consider the important case in applications, [6], of channels with intersymbol interference (ISI) and signal-dependent Gauss Marov noise as described in Example in Section II and depicted in Fig.. he extension to the more general context of Example 2 is relatively easy. In the model of Example, I 0 is the ISI length, and L 0 is the Marov memory length. We also assume that a is a binary symbol sequence. A. Branch Metrics We fist define the state x. We consider the ML sequence detection and wor with (6). o find the metric 3 ML (z 0L ;x 0; x ), we must find the logarithm of the pdf f (z 0L jx 0; x ). he vector z 0L in (6) involves observables z 0L through z, whose conditional means (ideal channel outputs) are y(a 0L0I 0L ) through y(a0i ), see (5). herefore, the trellis branch t(x 0; x ) should be chosen so that it collects all inputs a 0L0I through a involved in determining the channel outputs y(a 0L0I 0L ) through y(a0i ) corresponding to the observables z 0L through z. A simple choice is to label the state by grouping the incoming bits as x = a 0I0L+ : (7) We define the vector of L +ideal channel outputs and rewrite (5) as Y (x 0; x )=[y(a 0L0I 0L z 0L ); ;y(a0l)] = Y (x 0; x )+n 0L : (8) Since n is a conditionally Gaussian process, we have (z 0L jx 0; x ) N(Y (x 0; x )C(x 0; x ))

294 IEEE RANSACIONS ON INFORMAION HEORY, VOL. 46, NO., JANUARY 2000 where C(x 0; x ) is the (L +)2 (L +)covariance matrix of n 0L conditioned on the pair of states (x 0; x ). Substituting this pdf into the expression for the ML branch metric (6), and cancelling constant terms common to all branches, we obtain the Gaussian ML branch metric 0L M ML(z ; x 0; x ) =ln detc(x 0; x ) det c(x 0; x ) + z0l 0 Y (x 0; x ) C(x 0; x ) 0 z 0L 0 Y (x 0; x ) 0 z 0L 0 0 y(x 0; x ) c(x0; x ) 0 z 0L 0 0 y(x 0; x ) : (9) Here c(x 0; x ) is the upper L2L principal minor of C(x 0; x ), i.e., C(x 0; x )= c(x 0; x ) c(x 0; x ) c(x 0; x ) c(x 0; x ) (20) and y(x 0; x ) collects the first L elements of Y (x 0; x ). In a similar fashion, the Gaussian MAP branch metric can be shown to equal M MAP z 0L ; x 0; x = M ML z 0L ; x 0; x 0 2lnP(x jx 0): (2) B. FIR Filter Branch Metric Implementation he metric given in (9) involves two vector-matrix multiplications, maing the complexity of the metric computation order O((L +) 2 ). We next exploit the structure of the covariance matrices to bring the computational complexity down to O(L +)by implementing the branch metric computation using an FIR (finite impulse response) filter. Let the covariance matrix be partitioned as in (20). Using the matrix inversion lemma [7], the inverse can be written as C(x 0; x ) 0 = c(x 0; x ) 0 0 0 0 + w c(x 0; x )w c (x 0; x ) (x 0; x ) : (22) In (22), (x 0; x ) and w c(x 0; x ) are given by the solutions of the signal-dependent Yule Waler equations [8] and can be shown to equal [9] w c (x 0; x )= (x 0; x )= det C(x 0; x ) det c(x 0; x ) = 2 0c(x 0; x ) 0 c(x 0; x ) a 0I (23) = 0b a 0I (24) where (a 0I ) and b(a 0I ) are the autoregressive coefficients from the model equation (6). Substituting (22) into (9), we rewrite the Gaussian ML branch metric as M ML 0L z ; x 0; x =ln 2 + 0b a 0I a 0I z 0L 0 Y (x 0; x ) 2 a 0I 2 : (25) he metric in (25) is obtained by filtering the observed vector z 0L through an FIR filter with coefficients w c(x 0; x ) given in (24). his filter is the inverse of the autoregressive filter in Fig. and it uncorrelates the autoregressive (AR) process n. hus we may interpret the metric (25) as first uncorrelating the noise with an FIR filter, and then applying the Euclidean (square) metric to the output of the filter. his is illustrated in Fig. 2 where the branch metric is shown to be the processed output of a branch-dependent FIR filter. Since the FIR filter coefficients 0b(a 0I ) depend on only I +symbols, and since there is a total of 2 I+L+ branches in each stage of the trellis, it turns out that there are 2 I+ distinct FIR filters in the detector, where 2 L branches may share a single filter. A special case of the above FIR filter occurs when the covariance matrix is not signal-dependent, i.e., C(x 0; x )=C. In this case, the filter w c (x 0; x ) = w c is a state-invariant tap-delay line that whitens the stationary Gauss Marov noise process. his is an intuitively pleasing result, since it is well nown that a stationary (causal) Gauss Marov AR sequence can be whitened by a time-invariant FIR filter. Notice also that, if w c is state-invariant, all branches share the same filter. We can then move the whitening filter to the front end of the Viterbi detector. his provides another explanation of why the Viterbi detector is optimal only when the channel noise is Marov. Suppose that the noise is not Marov, i.e., it is not the output of an autoregressive filter as in (6). hen to whiten this noise requires passing the observed channel outputs through an infinite impulse response (IIR) filter, thus introducing infinite-length ISI. However, with infinite-length ISI, the Viterbi detector requires an infinite number of states in order to be the optimal detector, which is clearly unrealizable. We mae a distinction here with what is usually done in Kalman Bucy filtering (KBF) [20] with Marov observation noise, see [2, Ch. ]. In this case, the (vector) observations z are modeled as z + = H x + v (26) where x is the state vector and v is a first-order Marov vector following the model v + = A v + w (27) where w is a white Gauss vector sequence. Preprocessing the observations z into the new observation z 0 = z + 0 A z (28) one can then apply standard KBF theory. Although similar, there is a fine distinction with Viterbi decoding. Due to the requirement of finite ISI, the Marov noise model (27) needs to be restricted to autoregressive processes (see (6)), while in KBF no such restriction is required. Finally, we comment on the practical implementation of the branch metric computation. As seen from (9) or (25), implementing the

IEEE RANSACIONS ON INFORMAION HEORY, VOL. 46, NO., JANUARY 2000 295 analyzing the detector performance of a simple ISI channel with finite memory Marov noise. Notation: Let "M =(9 0 M ; 900 M ) denote a length-m error event. By 9 0 M, we denote the correct path through the trellis spanning M +consecutive states x 0 ; ; x0 +M. he erroneous path is denoted by 9M 00 and spans states x 00 ; ; x00 +M. It is assumed that x 0 = x 00, x 0 +M = x 00 +M, and x 0 +m 6= x 00 +m for m M 0. Since a state is defined as x = a 0I0L+,an equivalent notation is and where and a 00I0L+m+ +m 9 0 M = a 00I0L+ +M 9 00 M = a 000I0L+ +M 6= a 000I0L+m+ +m ; for m M 0 Fig. 2. Bloc diagram of an FIR filter computing the ML branch metric for a channel with ISI of length and Gauss Marov noise with memory length. Since there are 2 branches in each stage of the trellis, the branches may be labeled by + +bits. Of these bits, only +are used to address the FIR filters. A ban of 2 different FIR filters is needed for optimal detection, while 2 branches share the same filter. ML branch metric requires nowledge of the second-order statistics C(x0; x). In practice, the second-order statistics are typically not nown and need to be estimated from the data. racing the statistics adaptively is preferable in many applications because the covariance statistics may time-drift or be nonstationary for some other reason. Adaptive covariance tracing has been widely covered in the literature, see, e.g., [8] and [22]. For a method applied in magnetic recording signal detection see [5]. Notice also that, if the involved covariance matrices are signal-dependent, and so nonstationary and non-oeplitz, many efficient AR estimation algorithms [8], [23] based on the Levinson Durbin recursion will then not wor. V. ERROR ANALYSIS We develop in this section the error analysis for the Viterbi sequence decoder when the noise in the ISI channel is signal-dependent and correlated. Our results extend the error analysis in [2] and [24] to correlated finite memory ISI channels. However, due to the channel nonlinearity and the signal-dependent noise correlation, the analysis must carefully treat the lac of symmetry. For example, in contrast with the linear channel with stationary white noise, if a 0 ; ; a0 m and a 00 ; ; a00 m are two valid binary sequences in the trellis, the probability of detecting a 0 when a 00 is sent is not the same as detecting a 00 when a 0 is sent. Also, unlie the linear channel with stationary white noise, we cannot formulate the upper bound using the flowgraph transfer function [2] because relative distances between trellis paths are not symmetric due to the asymmetric signal-dependent noise correlation. o formulate the bounds, we introduce first the necessary notation. hen, in Section V-A, we derive the binary hypothesis error probability for asymmetric Gauss Marov noise channels, followed in Section V-B by a computationally efficient approximation to this error probability. In Sections V-C and V-D, we formulate upper and lower bounds using the binary hypothesis error probability expressions. Finally, in Section V-E, we illustrate the error performance results by for all other m. a 00I0L+m+ +m = a 000I0L+m+ +m A. Binary Hypothesis Error Probability he binary hypothesis error probability P 2("M ) is the probability of detecting 9M 00 when 9M 0 is the true path. his error occurs when the accumulated metric (25) along the path 9M 0 is greater than the accumulated metric along 9M 00, i.e., P 2("M )=P M M ML m= M > M ML m= +m0l z +m ; a0+m0i0l z +m0l +m ; a00+m0i0l : (29) Notice that we can use an arbitrary in (29). For convenience, we shall use = L throughout this subsection. o evaluate the expression on the right-hand side of (29), we use the special structural properties of Gauss Marov covariance matrices [9] and rewrite (29) as P 2 ("M )=P ln det C 9 + N 0 M+L > ln det C 9 + N 00 M+L C 0 9 N 0 M+L C 0 9 N 00 M+L : (30) he matrices C 9 and C 9 are the covariance matrices of z L+M when a 00I L+M and a000i L+M are the transmitted (or written on a memory medium) binary sequences, respectively. he innovation vector N 0 L+M is defined as N 0 L+M = z L+M 0 y 0 = z L+M 0 y a 00I ; ; y a 0L+M0I L+M (3) where the values y(a 00I ) are the same as those used in the model equation (5). he vectors N 00 M+L and y 00 are defined similarly as in (3), where the superscript 0 is replaced by 00. Next, we briefly describe how to obtain the inverses C 0 9 and C0 9 of the covariance matrices from the model parameters in (6). Since the process n is conditionally Gauss Marov with Marov memory If the detector is an MAP detector rather than the ML detector, we substitute the MAP metric into (29) to get the MAP binary hypothesis error probability.

296 IEEE RANSACIONS ON INFORMAION HEORY, VOL. 46, NO., JANUARY 2000 length L, it follows from [5] that C 0 9 is an L-banded matrix. Its upper Cholesy decomposition is given by C 0 9 = U 9 D 9 U 9 : (32) In (32), according to [9] and [25], the upper triangular L-banded matrix U 9 is U 0b a 0L+0I L+ 0... U 9 = (33)... 0b a 0L+M0I L+M 0 and the diagonal matrix D 9 is D 9 =diag[d = 2 a 0L+0I L+ = 2 a 0L+M0I L+M ] : (34) he values b(a 0i0I i ) and (a 0i0I i ) in (33) and (34) are the same as those in the model equation (6). he upper triangular matrix U and the diagonal matrix D are the upper Cholesy factors of C 0 = U D U, where C is the conditional covariance matrix of z L when the sequence of transmitted (written) symbols is a 00I L. he elements of C can be obtained by rearranging a signal-dependent set of Yule Waler equations and solving a linear system of equations. Since computing C is tedious but straightforward, we refer the reader to an example provided in [26]. he Cholesy decomposition of C 0 is obtained similarly. 9 Once we have computed the Cholesy decompositions (32) of both C 0 and C0, we can rewrite the binary hypothesis error probability 9 9 for the error event " M =(9M 0 ; 9M 00 ) as P 2 (" M )=P w w > ln det 6 " +(w0m " ) 6 0 " (w0m " ) (35) Fig. 3. Binary hypothesis signal and noise constellation encountered when computing the binary hypothesis error probability. Note the asymmetric noise clouds. determined on a case-by-case basis, which ultimately may be computationally impractical. A brute-force alternative is to evaluate (35) numerically by Monte Carlo, but this is also computationally intensive for low error rate (high sisgnal-to-noise ratio (SNR)) scenarios. In the next subsection, we develop a computationally efficient Q-function approximation that leads to an accurate evaluation of (35) for high SNR scenarios where Monte Carlo numeric evaluations are not viable. B. Approximate Binary Hypothesis Error Probability Computing (35) requires integrating a zero-mean I-covariance multivariate Gaussian pdf over a region where f " (w)=w w0ln det 6" 0(w0m" ) 6 0 " (w0m " )>0: (38) Since a Gaussian function has a very fast decay, we approximate (35) by P 2 (" M ) Q(d 2 ) (39) where w is an (L + M ) 2 Gaussian random vector with E[w] =0 and E[w w ]=I. he vector m " is given by where d 2 is the point on f " (w) =0closest to the origin and m " = QD =2 9 U 9 (y00 0 y 0 )=[d 0 0] (36) Q(x) = p 2 e 0(t =2) dt: x where d = jjd =2 9 U 9 (y00 0 y 0 )jj, y 0 and y 00 are defined in (3), jj jj is the L 2 vector norm, Q is a unitary matrix (for example, the Householder reflector Q =2v v =v 2 0 I, where v = q=q + e, q = D =2 9 U 9 (y00 0 y 0 ), and e = [; 0; ; 0], see [27]), and 6 " satisfies and 6 0 " = QD 0=2 U 0 9 9 C0 9 U0 9 D0=2 Q 9 det D 9 det 6 " = : (37) det D 9 In the general case, the expression on the right-hand side of (35) cannot be simplified by integrating a 2 -lie distribution. hese simplifications are possible only in special cases, for example, when 6 " is diagonal (for a list of all special cases where a simplification is possible, see [28]). In the general case, however, evaluation of (35) involves integrating a zero-mean I-covariance multivariate Gaussian pdf in the region where the condition on the right-hand side of (35) holds. Since this is a region outlined by a quadratic form, its shape must be o determine d 2, one may use any minimization technique (Lagrange multipliers, steepest descent, etc.). A particularly fast method is to find the point d on the boundary f " (w) = 0between 0 and m " by solving a quadratic equation, see Fig. 3. For a given matrix 6 ", the point d will be between 0 and m " if d is large enough, i.e., if the SNR is large enough. Since d 2 is typically close to d, we can obtain d 2 in a few iterations, or we may simply set d 2 d cos, see Fig. 3. We prefer the iterative method since it is not too computationally expensive. C. Upper Bound he bit-error probability Pb can be upper-bounded by the union bound [2]. Since our channel is asymmetric (nonlinear with signal-dependent noise correlation) we need a union bound expression that generalizes the one in [2]. One can verify that the two expressions are equivalent for linear-symmetric channels. Let M denote the length 0 (in branches) of an error event " M = (9M ; 9M 00 ). For our binary channel, the minimum allowable error event length can be verified to be M min = I + L +, where I is the ISI length, and L is the Marov memory length of noise. he set of all allowable error events of length M is denoted by EM. Denote by b # (" M ) the number of erroneous bits

IEEE RANSACIONS ON INFORMAION HEORY, VOL. 46, NO., JANUARY 2000 297 corresponding to the error event " M. Let P(x) be the a priori probability of the trellis state x. Similarly, let P(x ;x 0) denote the conditional probability of transition from state x 0 to state x. he binary hypothesis error probability associated with the error event " M is P 2 (" M ) and can be computed by either Monte Carlo simulation or approximated by (39). he upper bound on bit-error probability is given by Pb M=M 9 P(x 0 M ) P x 0 +ijx 0 +i0 i= b # (" M)P2(" M): (40) " =(9 ; 9 )2E his bound accounts for both the ML and the MAP sequence detector, where the corresponding binary hypothesis error probability P 2 (" M ) needs to be substituted (see footnote to (29)). If all symbol sequences are equally liely, then P(x 0 )=2 0(L+I) and P(x 0 +ijx 0 +i0)==2: A practical method for evaluating the bound in (40) is to truncate the first sum to values of M < M max, where M max is a predetermined large enough constant. Notice that in (40), the binary hypothesis error probability P 2 (" M ) needs to be determined for every qualifying error event " M, which maes the bound computationally expensive if P 2(" M) is computed by Monte Carlo simulations rather than by the approximation (39). D. Lower Bound Obviously, a lower bound is obtained by picing any single term of the union bound (40). his, however, may not be a tight bound. o get a tighter bound, we modify the genie-assisted bound [24]. Modified Genie-Assisted Bound: Denote by P(") the probability of an error event. Let c (Pc) be the probability that the input sequence a 0 is such that there is a probability P Pc of confusion with some allowable sequence a 00. Even with a genie-assisted detector that has to choose only between a 0 and a 00, the probability of an error event P(") will be greater than c (Pc) Pc. Clearly, then P(") max [c(pc) Pc] : (4) 0P he difference between this formulation and the one in [24] is that we assume P Pc instead of P = Pc. In fact, the bound in (4) will never be looser (and in many cases tighter) than the one in [24], even in linear-symmetric channels. We use the result in (4) to find a lower bound for the bit-error probability. Consider only error events of length M min = I + L +. he number of these events for binary signaling is N =2 I+L+M =2 4 I+L : Label these error events as " M (); " M (2); ;" M (N), where their order is such that P 2 (" M ()) P 2 (" M (2)) P 2 (" M (N)) and P 2(" M (i)) is the binary error probability of the error event " M (i). he probability P 2 (" M (i)) can be computed either Fig. 4. Performance evaluations of detectors applied to the channel in able I. through Monte Carlo simulation or with the approximation in (39). In the binary signaling case, for every sequence 9M 0 (i) of M min + states, there exists only one erroneous sequence 9M 00 (i). We therefore have N P(9M 0 (i)) = i= i.e., 9M 0 (i) are disjoint sequences. If we choose Pc = P 2(" M (i)) for some i where i Mmin, then c (Pc) i j= P(9 0 M (i)) because there may be error events of length M > M min whose binary hypothesis error probability is greater than Pc. In the interest of eeping the bound computationally simple, we ignore these error events of length M>M min. Since for all error events of length M min, the number of erroneous bits is b # (" M (i)) =, wehave Pb c (Pc) Pc i j= P 9 0 M (j) P 2 (" M (i)) (42) for any i N. Maximizing the right-hand side of (42), we obtain a tighter lower bound Pb max in i j= P 9 0 M (j) P 2 (" M (i)) : (43) his expression provides for a lower bound for both the ML and the MAP sequence detector, where the appropriate binary hypothesis error probability would be used, see footnote to (29). When all sequences 9 0 M (i) are equiprobable, we have P(9 0 M (j)) = N = 2 40(I+L) :

298 IEEE RANSACIONS ON INFORMAION HEORY, VOL. 46, NO., JANUARY 2000 ABLE I NONLINEARLY DISORED PARIAL RESPONSE CHANNEL WIH ISI LENGH MEMORY LENGH = = AND MARKOV extremely high error rates (0 0 and above) the upper bound may fail to converge, but that is rarely a concern since at those error rates the MLSD performance can readily be evaluated through simulations. VI. SUBOPIMAL DEECORS hus the modified genie-assisted bound provides the lower bit-error probability bound Pb 2 40(I+L) max in [i P 2 ("M (i))] : (44) Since the terms P 2("M (i)) have already been computed as part of (40), no further computation is needed for the lower bound. Notice also that while (44) is always a lower bound, it may not be the tightest of the modified genie-assisted bounds because we considered only the minimum-length error events. he following example shows, however, that this bound is relatively tight. E. Performance Evaluation Example We study the performance of the maximum-lielihood sequence detector for a simple channel with ISI length I =and Marov memory length L =. he channel is given by the model equations (5) and (6), also depicted in Fig., where the values of the parameters are given in able I. he channel represented in able I can be viewed as a nonlinearly deviated 0 D partial response channel with signal-dependent correlated noise. Fig. 4 shows the performance of the ML sequence detector. he different values of the SNR shown in Fig. 4 are obtained by scaling the noiseless channel response (column denoted by y(a 0 ) in able I) while eeping the relative distance between the signal points the same. his corresponds to scaling the signal power while the noise constellation is ept the same. In Fig. 4, the solid line represents the simulated performance of the ML sequence detector (MLSD). We do not show the MLSD performance curve beyond SNR = 4 db since the Monte Carlo simulations become expensive to compute. Fig. 4 also shows the simulated performance of the Euclidean detector applied to the channel in able I. he Euclidean detector is a Viterbi detector that assumes the noise to be white. Fig. 4 shows that at the error rate of 0 05 the MLSD has a gain of 2.5 db over the Viterbi Euclidean detector. Fig. 4 also plots the bit-error probability bounds. he stars (*) in Fig. 4 show the bounds computed by evaluating the binary hypothesis error probability with Monte Carlo simulations. As the Monte Carlo simulations become too computationally intensive beyond SNR = 2 db, we only show the stars (*) for SNR 2 db. On the other hand, the approximate bounds (dashed lines in Fig. 4) computed by approximating the binary hypothesis error probability with a Q-function, see (39), can be computed at a fractional computational cost for any value of the SNR. Fig. 4 also shows that the Q-function approximation is extremely close to the bounds computed by Monte Carlo simulations. As expected, the bounds are tightening as the SNR increases which helps us evaluate the MLSD performance at high SNR s where Monte Carlo simulations of the MLSD are impractical. We point out that at Being the optimal detector, the maximum-lielihood sequence detector (MLSD) bounds from below the error rate of any suboptimal detector of lower complexity. Furthermore, the structure of the signal-dependent MLSD gives new insight into developing suboptimal detectors. Here we cover several suboptimal schemes and discuss their relationship to the MLSD. From Fig. 4 we see that there is a performance region between the MLSD and the Euclidean detector that can be filled with suboptimal detectors with complexities possibly lower than that of the MLSD. here are many ways to construct such suboptimal receivers. We limit ourselves to suboptimal receivers with a certain structure. We loo for suboptimal receivers that are Viterbi-lie, i.e., that operate on a trellis and use branch/path metric minimization as their detection strategy. We define the complexity of a Viterbi-lie detector as the number of multiplications it has to perform per cloc interval (symbol interval) in order to compute all its branch metrics. We ignore additions. In many realizations of the Viterbi detector this may not be an accurate representation of the detector complexity since there exist several practical finite-precision methods to simplify the multiplication operation. Nevertheless, for simplicity, we trust that the multiplications count provides at least a reasonable representation for the complexity of a detector. When I is the ISI length and L the Marov memory length, the MLSD for the binary channel model in (5) and (6) has 2 I+L states and 2 I+L+ branches. hereby, 2 I+ branches share a single FIR filter used to compute the branch metric, see Fig. 2. In Fig. 2, if we count each tap-multiplication by bi(a 0I ), each division by (a 0I ), and each squaring operation as one multiplication, the count of multiplications needed to compute all the branch metrics of the MLSD in one cloc interval is (2 L + L + )2 I+. Since the computational complexity of the Euclidean Viterbi detector is 2 I+ per cloc interval, we have that the computational complexity increase factor of the MLSD over the Euclidean detector is F MLSD (L) =2L + L +: (45) We next consider suboptimal detectors with lower computational increase factors. A. Low-Order Autoregressive (AR) Approximation (Q-AR Detection) We introduce in this subsection a new suboptimal detector that we refer to as the low-order Q-AR detector. A Gauss Marov noise process of order L is the output of an autoregressive (AR) model of order L. he inverse covariance matrix of an Lth-order Gauss Marov process is L-banded, meaning that it has 2L +nonzero diagonals, while the remaining are all-zero diagonals [5]. We obtain a suboptimal detector by using an AR approximation [6] of order Q L. Let C Q be the covariance matrix of the Qth-order AR process approximation. he matrix C 0 Q is then Q-banded. We can show that the Q-banded matrix C 0 Q that minimizes the Kullbac Leibler mean information loss due to approximating the original Lth-order AR process with a Qth-order AR process is given by the following decomposition [9]:

IEEE RANSACIONS ON INFORMAION HEORY, VOL. 46, NO., JANUARY 2000 299 and (46) are the inverses of the ban of L-AR filters used in the channel model equations (5) and (6). Since the K-step detector uses nonuniform AR approximations at different trellis stages, the number of multiplications has to be averaged over a cycle of K trellis stages to get the average computational complexity per cloc interval. he computational complexity increase factor of the K-step Viterbi detector over a Euclidean Viterbi detector is where 0 2Q+ denotes a band of 2Q+all-zero diagonals, and D 2Q+ denotes a band of 2Q +nonzero diagonals. Note that the decomposition in (46) is unique since there exists only one matrix C Q such that the Q-band of C Q equals the Q-band of C, and that C 0 Q is Q-banded. he suboptimal Viterbi detector based on the low-order AR approximation computes the branch metric using a Q-tap FIR filter instead of an L-tap FIR filter. 2 Consequently, the detector has 2 I+Q states and 2 I+Q+ branches where 2 I+ branches share a single Q-tap FIR filter. he computational complexity increase factor of the low-order Q-AR detector over the Euclidean detector is F Q-AR(Q) =2 Q + Q +: (47) Obviously, if Q = L, we have the MLSD. If Q = 0, we obtain the suboptimal detector presented in [7] and [8] for which the complexity increase factor is F Q-AR(0) = 2. B. Bloc Partitioning (K-Step Detection) In [3], Altear and Wolf present a suboptimal detector by replacing the covariance matrix C by a matrix C 0 obtained by eeping only the K 2 K bloc diagonal elements of C, i.e., C R 0... 0 R = C 0 (48) where R is a K 2 K principal minor of C. his contrasts with the Q-AR approximation of (46) where it is the inverse covariance that is banded rather than the covariance matrix itself, which is bloc-partitioned. In [3], the method is called the K-step Viterbi detector. In [3], the authors provide a metric computation method involving vectors of observed samples and covariance matrices R. Using techniques presented in Section IV-B, we can show that the K-step detector may also be implemented with low-order FIR filters (which is also a computationally less expensive implementation than in [3]). However, the order Q of the FIR filters is cyclicly time-varying. For example, the three-step detector would use bans of 2 I+ (Q =0)-FIR filters for the first stage of the trellis, (Q =)-FIR filters for the second stage, and (Q =2)-FIR filters for the third stage (for low-order signal-dependent AR processes, see [6] and [9]). In the fourth stage of the trellis, the detector goes bac to the ban of (Q =0)-FIR filters and repeats the cycle. hus generally, a K-step detector would cyclicly employ (Q = 0)-AR through (Q = K 0 )-AR approximations at different stages of the trellis. hereby, the FIR filters used to calculate the branch metrics are the exact inverses of the Q-AR filter approximations, see Section IV-B. At trellis stages where Q L, the ban of FIR filters 2 Note that the th-order filter is not simply a truncation of the optimal th-order filter. F K-step(K) = 2 K 0 K + K + ; for K L 2 2 L 0 +(K 0 L)(2 L + L 0 ) (49) K L(L +) + ; for K>L: 2K Clearly, the (K = )-step detector is the same as the (Q = 0)-AR detector, which is the detector presented in [7] and [8]. C. Bloc Sipping (S-Sip Detection) A drawbac of the K-step detector is that it reaches the MLSD performance only in the limit as K!. We propose here a new class of suboptimal detectors that reaches the MLSD performance for a finite order of suboptimality. he following equation shows the L-band of a covariance matrix C of an Lth-order Gauss Marov process partitioned in contiguous principal blocs of size (L +)2 (L +) (50) It can be shown, see for example [9], that the elements of the L-band of C determine in a unique way the covariance matrix C of the L-th order Gauss Marov process, in particular, they determine the elements of the blocs. We can conclude that the MLSD derives the branch metrics based on all covariance coefficients in this L-band of C. As an alternative, out of the (L +)2 (L +)blocs outlined on the right-hand side of (50), a suboptimal algorithm may ignore the blocs outlined by the dashed lines in (50). We refer to such a procedure as the (S =)-sip detector. In general, if we sip S blocs at a time, we have the S-sip detector. Much lie detailed in Section VI-B, the S-sip detector may be designed through cyclic implementation of nonuniform Q-AR approximations. Each cycle spans S +trellis stages and the detector cyclicly employs bans of (Q = L 0 S)-FIR, (Q = L 0 S +)-FIR,, (Q = L)-FIR filters. he ban of (Q = L)-FIR filters is given by the inverses of the actual Lth-order AR filters in (5) and (6). he computational complexity increase factor of the S-sip Viterbi detector over the Euclidean Viterbi detector is F S-sip(S) = 2L+ 0 2 L0S 2L 0 S +2 + ; for S L: S + 2 (5) he (S = 0)-sip detector is clearly the MLSD. It is also apparent that the (S = L)-sip detector is equivalent to the K-step bloc partitioning detector when K = L +, where L is the Marov memory length of the noise. D. Nonuniform AR Approximations he K-step and S-sip detectors are subclasses of Viterbi detectors that apply different Qth-order AR approximations at different stages of

300 IEEE RANSACIONS ON INFORMAION HEORY, VOL. 46, NO., JANUARY 2000 ABLE II NONLINEARLY DISORED PARIAL RESPONSE CHANNEL (PR4 CHANNEL) WIH ISI LENGH =2 AND MARKOV MEMORY LENGH =2 the trellis. One can easily thin of numerous other Viterbi-lie detectors based on this concept, where the different orders of AR approximations are not necessarily cyclicly employed. A typical scenario would be in nonstationary detection. Suppose that at some point during data transmission the data rate increases, i.e., the cloc interval (symbol interval) shrins. If the new cloc interval is too short to perform Viterbi detection based on a high-order AR approximation, the detector may resort to a less computationally expensive lower order AR approximation, still retaining the communication lin at the price of a less reliable communication. Combinations of different orders of AR approximations will lead to a large number of possible detector designs. E. Decision Feedbac Decision feedbac mechanisms can be used to lower the computational complexity of Viterbi detectors. Reduced-state sequence estimators (RSE) [29] use premature decisions to reduce the ISI length and thus lower the number of states. Noise predictors [], [2], [30] on the other hand, use premature decisions to guess the future noise sample. In the context of Marov noise models, this corresponds to lowering the Marov memory length. Hybrids between reduced-state sequence estimators and noise predictors are also possible. Furthermore, we can envision combining the nonuniform AR approximations with decision feedbac strategies to form a larger subclass of Viterbi-lie suboptimal detectors. A common feature to all decision feedbac suboptimal detectors is error propagation, see results presented within [3]. Error propagation precludes the use of decision feedbac detectors in many applications where larger blocs of errors cannot be tolerated, e.g., in bloc error-correction coding/decoding. F. Comparative Results In this subsection we provide the error performance for a number of detectors and show that the Q-AR detectors provide usually the best compromise between error rate and computational complexity. able II shows the parameters (for explanations, see (5) and (6)) of a very noisy real channel encountered in magnetic recording applications [6]. he channel in able II is a nonlinearly distorted class-4 partial response (PR4) channel with ISI length I =2and signal-dependent correlated noise with Marov memory length L =3. Fig. 5 shows for this channel the error rate versus the computational complexity increase factor F for a number of detectors: the MLSD, the Euclidean Viterbi decoder, and several suboptimal Q-AR, K-step, and S-sip detectors. he computational complexity increase factors are normalized with respect to that of the Euclidean detector, whose operating point in Fig. 5 corresponds then to the value on the horizontal axis. We do not consider decision feedbac suboptimal detectors since they suffer from error propagation. he error rates were computed by Monte Carlo simulations with 0 6 bits. he results confirm, as mentioned above, that the Fig. 5. Performance versus computational complexity. Shown are the operating points of the Euclidean detector, the MLSD, and different orders of the suboptimal -AR, -sip, and -step detectors. he detectors are applied to the channel given in able II. suboptimal detectors based on low-order AR approximations (Q-AR detectors) provide the best complexity-versus-error-rate tradeoff. he K-step detector is a reasonable alternative to the Q-AR detector only for low complexities and high error rates. On the other hand, the S-sip detector is a reasonable alternative to the Q-AR detector only for high complexities and low error rates. VII. SUMMARY We studied the maximum-lielihood sequence detector (MLSD) and the MAP sequence detector for intersymbol interference (ISI) channels with correlated noise whose statistics are possibly signal-dependent. We describe three major contributions to the problem. First, by modeling the correlated noise as a Marov process of finite order, we derive the optimal MLSD and MAP sequence detector as extensions of the Viterbi decoder for ISI channels with finite memory noise. Second, we provide an error performance study of the MLSD and MAP sequence detector, deriving tight upper and lower bounds for the probability of error. Finally, we present several classes of suboptimal sequence detectors, including new algorithms, and compare their computational complexity and their error performance. he ey to deriving a finite implementation for the optimal sequence detector is our modeling the correlated noise as a finite-order Marov process, see Section II. We showed in Section III that, as long as the memory of the noise is Marov, we can extend the Viterbi algorithm