Some notes about signals, orthogonal polynomials and linear algebra

Size: px
Start display at page:

Download "Some notes about signals, orthogonal polynomials and linear algebra"

Transcription

1 Some notes about signals, orthogonal polynomials and linear algebra Adhemar Bultheel Report TW 180, November 1992 Revised February 1993 n Katholieke Universiteit Leuven Department of Computer Science Celestijnenlaan 200A B-3001 Heverlee (Belgium)

2 Some notes about signals, orthogonal polynomials and linear algebra Adhemar Bultheel Report TW 180, November 1992 Revised February 1993 Department of Computer Science, KULeuven Abstract These short notes contain an introduction to filtering of deterministic and stochastic signals The connection with algorithms from classical complex analysis (the Schur algorithm) and with the recurrence relation (Szegő) for polynomials orthogonal with respect to a measure supported on the unit circle are both given The interpretation of these algorithms in terms of linear algebra lead to fast algorithms for structured matrices It is explained how these algorithms can be generalized to situations where the signal is not stationary The corresponding notion of matrices with low displacement rank is introduced The notes are organized like lecture notes and contain several exercises that form an essential part of the text Solutions are provided at the end Keywords : signals, orthogonal polynomials, linear algebra, displacement rank MSC : Primary : 93E11 Secondary : 65F05, 41A20

3 Some notes about signals, orthogonal polynomials and linear algebra Adhemar Bultheel November 23, Discrete signals and digital filters Let s(t) be a continuous time signal that is sampled to obtain a discrete time signal or time series s(nt ) T is the sampling period and 1/T is the sampling frequency We shall abbreviate s(nt ) as s n Denote by l the space of complex double sequences l = {s = (s n ) : s n C, n Z} Note: In practical applications, a signal is real valued, but considering complex signals does not really complicate the analysis On the contrary, we shall make use of complex analysis and considering complex signals seems to be quite natural On the other hand, it is easier to leave away the complex conjugate bar if you want to specialize to real signals than to introduce it at appropriate places if you want to use results from complex analysis Let l D and l R be two subspaces of l and T : l D l R a linear operator As a special case we can define the shift operator Z by A transformation T is called shift-invariant if (Zs) n = s n 1, s l, n Z ZT = T Z Note that this implies immediately Z m T = T Z m, m Z A digital filter is a linear shift invariant transformation A signal s = (s n ) = (δ n0 ) = δ is called unit pulse (at time n = 0) and h = T δ for this pulse is called the (unit) impulse response of the filter T Consider the normed spaces l p = {s l : s p = ( s k p ) 1/p < }, k Z 0 < p < l = {s l : s = sup s k < } k Z The convolution of two signals u, h l is defined as u h = s, with s = (s n ) given by s n = m Z u m h n m, n Z if these sums converge 1

4 Note that the impulse response h completely characterizes the filter T s = T ( m s m (Z m u)) = s m (T Z m u) = m Z m T u = m s m (Z m h) = m s m h n m = (s h) n Exercise 11 Let h l 1, then the sums in the convolution u h converge if u l and T u = u h = h u, u l, h l 1 defines a digital filter T : l l T is a continuous transformation and h is the unit impulse response T δ = h A filter T is called BIBO (bounded input, bounded output) stable if the sequence T u is bounded whenever u is bounded (in some l p sense) Sometimes u 2 2 for u l 2 is called the energy of the signal u A signal u l is called causal if u k = 0 for k < 0 A filter is called causal if it maps causal signals into causal signals Exercise 12 Prove that T is causal iff u k = v k for k < m (T u) k = (T v) k for k < m Exercise 13 A filter T : l l defined by T u = h u and h l 1 with h causal is a causal filter Prove this The z-transform U(z) of a signal u is defined as the formal series U(z) = n Z u n z n Note: In the engineering literature this is usually defined with z 1 instead of z We use here the mathematical convention The space of signals l is often called the time domain, while the space of all the z-transforms is called the frequency domain Note that the z-transform of a convolution is the product of the z-transforms For example let h l 1, u l and y = T u = h u, then Y (z) = H(z)U(z) If h l 1, then H(z) converges for z = 1 to some function If h l 1 and causal, then H(z) converges for z 1 to a (complex) function analytic in z < 1 Another useful observation The shift operator Z in the time domain (Zu) n = (u n 1 ) corresponds in the frequency domain to a multiplication with z u n 1 z n = u n z n+1 = z u n z n = zu(z) n n n We shall abuse the notation z for Z, also in the time domain Thus from now on zu n = u n 1 Thus z represents the (backward) shift or time delay operator 2

5 Exercise 14 Let u = (u k ) be a given signal with samples of the form I I u k = α i e jωik = λ 0 + λ i cos(ω i k + ϕ i ), i= I i=1 k Z for 1 I <, α 0 = λ 0 0, ω 0 = 0 and λ i > 0 for 1 i I, ω i = ω i R, ϕ i = ϕ i = ϕ i R, α i = ᾱ i = (1/2) λ i e jϕ i Let T : l l be defined by T u = h u, h l 1 Then (T u) k = I α i H(e jω i ) e jω ik i= I = λ 0 H(1) + I λ i H(e jω i ) cos (ω i k ϕ i + arg H(e jω i )) The autocorrelation function r = (r n ) for u l 2 is defined by i=1 r n = k u n+k ū k, r = (r n ), n = 0, ±1, ±2, and r 0 = u k 2 is called the total energy or average power of the signal u Note 0 < r 0 < if u l 2 Exercise 15 Prove r n = r n Define R u (z) as the z-transform of the autocorrelation function: R u (z) = n r nz n The function R u (e jω ) defined on the unit circle is called the energy spectrum or power spectrum of u If (r n ) l 1, then R u (z) is a function defined on z = 1 and its Fourier coefficients are given by r n = 1 2π π π e jnω R u (e jω )dω The autocorrelation function (r n ) and the energy spectrum form a Fourier transform pair Note that R u (z) = n ( k u n+k ū k )z n = i,k u i ū k z i k = ( i u i z i ) ( k ū k z k ) with U = U(1/ z) If u l 2, = U(z) U (z) R u (e jω ) = U(e jω ) 2 Indeed, U (e jω ) = U(e jω ) From Fourier analysis we borrow the following isomorphy relation : u = (u n ) l p U(e jω ) = n u n e jnω L p, 1 p 3

6 where L p is the normed space { L p = f : f p = } 1 π 1/p f(e jω ) dω p < 2π π and Parseval s identity u p = U p holds If u l p is a causal signal in l p, then U(z) = k=0 u kz k is analytic in z < 1 and if it belongs to L p, then we say that it is an element from the Hardy space H p H p = {f L p and f analytic in z < 1} 2 Stochastic processes and signals Due to components in the signal such as measurement noise, a signal is often considered as a stochastic entity A sequence s = (s n ) of identically distributed random variables is called a (discrete) (stochastic) signal or a discrete stochastic process In general, a stochastic variable x is a complex function on a probability space (Σ, m), that is a set Σ with a probability measure m ( Σ dm = 1) The expectation operator E is defined for a random variable s as Es = s(σ)dm(σ) = the mean of s If s = (s n ) is a stochastic process with we define L 2 to be the closure of the linear hull Σ E s n 2 <, n Z, m { c n s n : c n C, n Z, m Z} m This L 2 is a Hilbert space with inner product u, v = u(σ)v(σ)dm(σ), u, v L 2 The autocorrelation function is defined by r(k, l) = s k, s l = The autocovariance is defined as Σ = E s k s l Σ s k (σ)s l (σ)dm(σ), c(k, l) = E(s k Es k )(s l Es l ) k, l Z = E s k s l E s k Es l Es l E s k + E s k Es l = E s k s l E s k Es l Clearly, if we have a zero-mean stochastic process then r(k, l) = c(k, l) and in that case r is also called the (auto)covariance function The process s = (s n ) is called stationary if r(k, l) = r(k + m, l + m), k, l, m Z 4

7 This means that the covariance function of a stochastic process depends only on the difference of the indices k and l We denote then r m = r(k, k + m) = s m, s 0 For a stationary process, s k 2 = s 0 2 = r 0 is called its energy or average power Exercise 21 Prove that for a stationary process r m = r m and the Toeplitz forms n i,j=0 c ic j r i j are all positive semi-definite The Fourier transform R of the autocorrelation function is called the (power) spectrum of the stochastic process: R(e jω ) = r m e jmω ; r m = 1 π e jmω R(e jω )dω 2π m π Note that the spectrum has the properties real : R(e jω ) = R(e jω ) positive : R(e jω ) 0 Hence it can be used as a weight function to define another Hilbert space L 2 with inner product f, g = 1 2π π π f(e jω )g(e jω )R(e jω )dω It consists of the square integrable complex functions for which f 2 = 1 2π π In this formalism we can for example write π f(e jω ) 2 R(e jω )dω < r m = z m, 1 More generally, the mapping s n e jnω is an isomorphism (the Kolmogorov isomorphism) between the two Hilbert spaces L 2 and L 2 defined in this section Moreover, the L 2 space of the deterministic case and the L 2 space in the stochastic case are mathematically the same This shows that the deterministic and the stochastic case can be treated in a uniform mathematical framework Let H(z) be a causal filter with impulse response h = (h n ) l 1 and z-transform (transfer function of the filter) H(z) = n=0 h nz n Note that we assume h to be a deterministic sequence Apply as input to the filter a stationary stochastic process u = (u n ), then the output s = (s n ) is also a stationary process and they are related by s n = m u m h n m or S(z) = H(z)U(z) It can be shown that their power spectra are related by R s (e jω ) = R u (e jω ) H(e jω ) 2 = R u (e jω )H(e jω )H (e jω ) A stationary process w = (w n ) is called white noise if E w n w m = δ nm w 2, Ew n = 0 5

8 Hence its power spectrum is constant : R(e jω ) = w 2 Often, one considers normalized white noise, ie, one takes w = 1 Note that white noise is the stochastic equivalent of a deterministic unit pulse in the sense that they have both the same perfectly flat spectrum A signal having a flat spectrum means that it does not contain any significant information The samples are completely uncorrelated to each other The process forgets immediately everything about its past There is no strategy to be detected in the signal observed Thus it is useless to try and predict the next output after observing what has happened in the past In a modern terminology, one would say that it has a completely chaotic behaviour As in the deterministic case, we shall also for stochastic signals use z as a notation for the time delay for example n n A(z)u m = a k u m k for A(z) = a k z k 3 Linear prediction k=0 A major problem in signal processing is to predict its behaviour from the observations of the signal itself That is to say, we want to model the filter (= system) which produces the particular signal we are observing In many application, (eg EEG-signals, speech processing, ) we do not have access to the input signal and we can only observe the output From the observation of the present and the past, we want to predict the next observation by a linear relation Thus we estimate s n by ŝ n, given some observations s n 1, s n 2,, s n p where ŝ n = p a k s n k The coefficients a k are the predictor coefficients and A(z) = n k=0 a kz k, a 0 = 1 is called the predictor The error or residual for s n is e n = s n ŝ n = s n + k=0 p a k s n k = A(z)s n, a 0 = 1 We want to minimize the least squares error, ie the energy of the error signal E p = e 2 which is given by n e n 2 in the deterministic case and by E e n 2 in the stochastic case Note that E e n 2 is independent of n in the stationary case 31 Deterministic case We can describe this as a linear least squares problem with infinitely many equations Define the column vectors s = (s n ) (the signal), e = (e n ) (the prediction errors), ŝ = (ŝ n ) (the predicted signal), and a = (a i ) p i=1 (predictor coefficients) and also the p matrix S with columns zs, z 2 s,, z p s Then e = s ŝ = s + Sa Minimizing e corresponds to finding the least squares solution of the system Sa = s Its solution is given by the normal equations (S H S)a = S H s 6

9 where S H S is a p p matrix and S H s a p 1 vector These are obtained by setting or more explicitly: E a i = 0, i = 1,, p; E p = e 2 ( ) p s n k s n i a k = n n s n s n i, i = 1,, p Exercise 31 Show that the least squared prediction error is given by E p = n s n 2 + p a k ( n s n s n k ) In principle, the summation for n should range over Z Practically however, we can t work with infinite sequences or an infinite system of equations That is why we did not specify the range for the summation of n We can then take two approaches: either we consider only a moderate number of the infinite set of equations Sa = s (and just hope that it will give a meaningful solution as well) or we want to consider the infinite set of equations by all means As we shall see below, this assumes that it is possible by some device to perform infinite summations (or at least approximate them adequately) We accordingly get two different situations The latter being the simpler one (mathematically that is) The methods are known in the literature as covariance and autocorrelation method respectively This terminology is certainly misleading because it has nothing in common with the difference between correlation and covariance 311 The autocorrelation method Here the error is minimized over an infinite duration The normal equations reduce to p r i k a k = r i, 1 i p where r i = s ns n+i Thus (S H S) ik = r i k is a p p Toeplitz matrix (hermitian and positive (semi) definite) The least squared error is given by E p = r 0 + p r k a k Note: In practice, an infinite signal can be made finite by putting a window over it This means that we consider a window (w n ) which is only different from zero outside some interval, say 0 n N 1 Thus we consider s where { s sn w n = n 0 n N 1 0 otherwise The autocorrelation function is then r i (N) = N 1 i n=0 s ns n+i, i 0 Of course a more direct method is to minimize only over a finite interval This is done in the next subsection 7

10 312 The covariance method Here we minimize the prediction error over a finite interval 0 n N 1 The normal equations become p r ik a k = r i0, 1 i p where now r ik = N 1 n=0 s n i s n k The matrix of the system is no longer Toeplitz, but it is still hermitian and positive (semi) definite It has still a special structure as the product of two Toeplitz matrices: S H S Of course, for N, the covariance method reduces to the autocorrelation method 32 Stochastic case Here similar results can be obtained as for the deterministic case Minimizing gives the normal equations: E p = E e n 2 = E s n + p a k s n k 2 p a k Es n k s n i = Es n s n i, 1 i p with least squared error Now we can also distinguish two cases: 321 Stationary process E p = E s n 2 + p a k E s n s n k If the process is stationary, then the autocorrelation function r i of the process r i k = Es n k s n i is independent of n and we get the same Toeplitz system as we had before, which in the stochastic literature is known as the set of Yule-Walker equations 322 Nonstationary case The autocorrelation function r(n k, n i) = E s n k s n i does depend on n now and the system we have to solve is no longer Toeplitz However, when the process is close to stationary, the system will have a matrix that is close to Toeplitz For more details see later sections 8

11 4 The filter model 41 The autoregressive model In the previous analysis, we did not take into account what the input was However, if we want to simulate (model) the signal, we should produce it by some system (filter) which has got to have some input, otherwise nothing would happen Observe that if we filter the signal s through the finite impulse response filter (predictor) A(z) = p 0 a kz k, a 0 = 1, then the output is the error signal (e n ): or explicitly: A(z)S(z) = E(z) or A(z)s n = e n, n Z e n = p a k s n k = s n + k=0 p a k s n k = s n ŝ n If we invert this filter and feed the error e as input, we shall get the signal s reconstructed Mathematically this is simply writing S(z) = 1 A(z) E(z), which means that we have to build a filter with transfer function 1/A(z), which has an infinite impulse response Unfortunately, for practical reasons, it is impossible to apply the exact error So the real problem is to find out what input signal to apply which can replace the error e Therefore we make the following reasoning Our objective was to extract as much information as possible from the signal Thus what remains (the error) should be completely meaningless, ie it should be random noise which is as white as possible In the deterministic case, this is a unit impulse However, when the filter with transfer function 1/A(z) is driven by a unit pulse, the impulse response will always start with h 0 = 1 because A(0) = 1 This need not be true for the signal we want to model, since s 0 need not be 1 So, we need some scaling factor G We propose a modeling filter of the form G/A(z), G > 0, such that if u (impulse or white noise) is the input (with flat spectrum U 2 = 1), then the output s is a model (approximation) for the true signal s The factor G is called the gain of this filter Thus the system which models the signal looks like S (z) = G A(z) U(z) = 1 A(z) E (z), u n G A(z) s n We see that an equivalent formulation is that we use 1/A(z) as transfer function and apply the input E (z) = GU(z) Thus we can also say that we model the signal by replacing the true error e by the approximation e n = Gu n with u = (w n ) (unit) white noise in the stochastic case and u = (δ n0 ) a unit impulse in the deterministic case The filter A(z) itself is called the whitening filter because it makes e n = A(z)s n as white as possible So, our problem is to find a value of the gain G It is defined by requiring that the energy of the model s is equal to the energy of the true signal s Now, requiring that s 2 = s 2, corresponds to requiring that we have equality of the Fourier coefficients r 0 = r 0 Thus we have chosen the model p p s n = a k s n k + Gu n for s n = a k s n k + e n so that it satisfies r 0 = r 0 9

12 Exercise 41 Prove that the autocorrelation function r i of the deterministic signal (s n) satisfies p r 0 = a k r k + G2 r i = p a k r i k, i 1 (Assume the system is causal, thus that s n = 0 for n < 0) Exercise 42 Prove that if r 0 = r 0 then (use previous exercise) r i = r i for i = 1,, p (and hence also r i = r i, i = 1,, p) Warning: this exercise is very hard to solve with our current knowledge You better try again later If the previous exercises are solved, then it follows easily that G 2 = p a k r k = k=0 p a k r k = E p = e 2 (a 0 = 1) k=0 From the previous exercises, it follows that we can reformulate our modeling problem as: find a filter of the form G/A(z), A(z) = p k=0 a kz k, a 0 = 1, such that the autocorrelation function r i of the output (impulse response) matches the autocorrelation function of the signal for p i p Thus G R s (z) 2 = A(z)A (z) = r k zk and R s (z) = r k z k with r k = r k, k = p,, p This is an example of a (0/p) Laurent-Padé approximant 42 Other models The previous example of a model is just one possibility to model a signal A more general model could be a filter of the form q 1 + b l z l H(z) = G B(z) A(z) = G l=1 p 1 + a k z k This is pole-zero model When a k = 0, 1 k p, we get an all-zero model or a moving average model (MA) If b l = 0, 1 l q, we obtain an all-pole model or an autoregressive (AR) model The general case is an ARMA model When we know the transfer function of the given filter, eg, H(z), then we can try Padé approximation to find B(z) and A(z) And as we have seen, we can always assume that we know the transfer function Either the input U(z) and the output S(z) is known, and then we get H(z) = S(z)/U(z), or we do not know the input, and then we assume that the S(z) is generated by some filter for which it is the transfer function If we assume that not the transfer function, but its spectrum, ie, its autocorrelation coefficients are known, then we can try to approximate it with a more general Laurent-Padé approximant To find the transfer function itself, we need the approximating spectrum in the factored form R s (z) = ( G B(z) A(z) 10 ) ( G B ) (z) A (z)

13 which is easily obtained for the denominator, but is more difficult for the numerator One problem is that the spectrum has to be nonnegative on the unit circle The approximating spectrum which is constructed by Laurent-Padé approximation can take negative values there We shall make another suggestion to solve the ARMA case later on 5 Spectral estimation The previous solution to the prediction or modeling problem can also be used to solve a spectral estimation problem in the following sense Note that the prediction error e is related to the signal s by E(z) = A(z)S(z) where A(z) is the predictor polynomial and thus is the spectrum of the signal s given by R(e jω ) = S(e jω ) 2 = E(ejω ) 2 A(e jω ) 2 The spectrum of the approximating signal s is given by Thus So that R (e jω ) = S (e jω ) 2 = R(e jω ) R (e jω ) = E(ejω ) 2 G 2 G 2 A(e jω ) 2 E p = G 2 = 1 π π E(e jω ) 2 dω = G2 R(e jω ) 2π π 2π π R (e jω ) dω This implies that the model is chosen so as to make 1 π R(e jω ) 2π π R (e jω dω = 1 ) Thus we approximate the given spectrum R by R in the previous sense, rather than eg minimizing R R 2! We can also write it as 1 π 1 2π π R(e jω ) 1 R (e jω R(e jω )dω = 0 ) One can in fact prove (but we shall not do it) that the linear prediction method proposes actually minimizes 1 S A 2 G = 1 π 1 2π π S(e jω ) A(ejω 2 ) G R(e jω )dω At the same time we have solved a spectral factorization problem approximately because R(e jω ) = S(e jω ) 2 G A(e jω ) G A (e jω ) with (as we shall see later) one factor (G/A(z)) analytic in z < 1 and the other (G/A (z)) analytic in z > 1 11

14 6 Toeplitz systems and orthogonal polynomials An alternative form to write the normal equations (autocorrelation method/stationary case) is r 0 r 1 r p 1 E p r 1 r 0 a 1 = 0 T pa p = E p e 0 r p r 0 a n 0 This shows the importance in this context to solve Toeplitz systems Let Ĩ be the reversal matrix: with 1 s on the main antidiagonal, then a Toeplitz matrix T satisfies Ĩ T Ĩ = T T Using this property, we can rephrase the previous system as (recall T H = T here) or r 0 r 1 r p r 1 r 0 r p r 0 ĨT p ĨĨA p = ĨE pe 0 = E p e p ā p ā 1 1 = 0 0 E p T pa p = E p e p These normal equations are related to the optimization problem min e 2, which in the frequency domain translates into: find p 2 E p = min a 1,,a p 1 + a k z k = A p 2 with optimal solution the predictor polynomial A p (z) = p k=0 a kz k, a 0 = 1, and the minimum being E p = A p (z) 2 = 1 π p a k e jωk 2 R(e jω )dω 2π π k=0 = 1 π p ā p k e jωk 2 R(e jω )dω 2π π = ϕ p (z) 2 where ϕ p (z) = z p + a 1 z p ā p = z p A p (1/ z) = z p A p (z) In the sequel we shall denote this transformation as ϕ p (z) = A p(z) Note that the column A p indeed contains the coefficients of the polynomial A p(z) Note also that in this Hilbert space k=0 z j 1 π p, ϕ p = ( ā p k z k )z j R(z)dω, 2π π k=0 p 1 π = ā p k z k j R(z)dω 2π k=0 π p = ā p k r j k k=0 z = e jω 12

15 Thus the previous Toeplitz system expresses that z j, ϕ p = 0 for j = 0, 1,, p 1 = E p for j = p In other words, ϕ p is an orthogonal (monic) polynomial with minimal norm z p, ϕ p = ϕ p 2 = E p Thus finding a p-th order predictor is equivalent to finding a (monic) orthogonal polynomial in this weighted Hilbert space Exercise 61 Prove that in any Hilbert space, the polynomial monic and of minimal norm is an orthogonal one Thus min{ p 2, p Π n and monic} is obtained for p = ϕ n where ϕ n is the monic orthogonal polynomial It is uniquely defined up to a constant factor of modulus 1 7 Computation of the predictor Special methods exist that require only O(p 2 ) operations to solve structured p p systems such as Toeplitz systems A basic method is called the Levinson algorithm (1947)7 Define for p = 0, 1, the Toeplitz matrices T p by and define A p = 1 a 1 a p T such that T p ij = r i j ; i, j = 0,, p, r i = r i T p A p = E p 0 0 T = E p e 0 Furthermore set A p = ā p ā 1 1 T Then, as long as E i 1 0 it holds that E 0 = r 0 for p = 1, 2, η p 1 = r p r 1 A p 1 ρ p = η p 1 /E p 1 Ap 1 A p = ρ 0 p E p = (1 ρ p 2 )E p 1 0 A p 1 Exercise 71 Prove that this algorithm gives indeed the solutions of the systems T i A i = E i e 0 for i = 0, 1, Exercise 72 Compute the complexity of this algorithm Note that in this algorithm, to compute the prediction of order p, we also compute all the previous ones and as a by-product we also get all the prediction errors E i Because E p has to be decreasing with p, and it is non-negative 0 E p E p 1 E 0 = r 0 It thus follows that ρ i 1, and if ρ p = 1, then E p = 0, which means that the model matches the given signal perfectly 13

16 The coefficients ρ i are crucial and they are called reflection coefficients (derived from scattering theory see later) or partial correlation (PARCOR) coefficients (from their stochastic interpretation) The predictor A i is called the forward predictor and A i is the backward predictor The latter is explained as follows: Interpreting again z as a delay operator, we have e n = A i (z)s n, A i (z) = 1 z z i A i, A i = 1 a 1 a i T where A i (z) is constructed to minimize the least squared error e 2 Similarly A i (z) appearing in f n = A i (z)s n, A i (z) = 1 z z i A i, A i = ā i ā 1 1 T = s n i + ā 1 s n i ā i s n can be seen as a (backward) prediction of s n i, given s n i+1,, s n, resulting in a backward prediction error f n It is constructed so as to minimize the least squared error f 2 In a stochastic context, the processes (f n ) and (e n ) are called the backward and forward innovation processes Exercise 73 Prove that in terms of the polynomials A i (z) and A i (z), the Levinson algorithm gives the recursion A i (z) A i (z) = 1 ρi ρ i 1 z Ai 1 (z) A i 1 (z) Ai 1 (z) = ϑ i (z) A i 1 (z) With the interpretations above, we can derive the circuit of figure 1 where e n (i) means the forward prediction error for s n when using predictor A i (z) etc This is called a lattice realization of one e n (i 1) f n (i 1) e n (i) ρ i z ρ i f n (i) e n (i 1) f n (i 1) i e n (i) f n (i) Figure 1: Lattice filter section of the filter Thus for such a realization, we do not need the predictor coefficients a i explicitly, it is sufficient to know the reflection coefficients ρ i The whitening filter (which analyses the signal) can be realized with such blocks as illustrated in figure 2 The modeling filter (which synthesises the signal) has the realization of figure 3 where the i-th block has now the inverse lattice form as depicted in figure 4 Exercise 74 Check the inversion as suggested by figure 4 Note that the normalized predictors which generate e n(i) = e n(i) G i are given by the normalized recursion A i (z) A i (z) = 1 1 ρi 2 1 ρi ρ i 1 = A i(z) G i s n = A i(z)s n ; G 2 i = E i z A i 1 (z) A i 1 (z) A = θ i (z) i 1 (z) A i 1 (z) 14

17 s n e n (1) e n (2) e n (i 1) e n (i) 1 2 i f n (1) f n (2) f n (i 1) f n (i) Figure 2: Whitening filter e n (i) e n (i 1) e n (i 2) e n (1) s n i i 1 1 f n (i) f n (i 1) f n (i 2) f n (1) Figure 3: Modeling filter e n (i) f n (i) + ρ i ρ i + z e n (i 1) f n (i 1) e n (i) f n (i) i e n (i 1) f n (i 1) Figure 4: Inverse lattice section 15

18 8 Example : seismic signal processing Consider a layered medium (like the earth surface) with layers of equal thickness (See figure 5) A wave is propagating through this medium At every interface between two layers the wave is x Figure 5: A layered medium t partially reflected and partially transmitted, according to the relative physical properties of the two layers meeting there The wave equation has two particular solutions 2 u(x, t) x 2 = 1 2 u(x, t) c 2 t 2 (c = speed) u(x, t) = ẽ(t + x ) (ascending wave) c u(x, t) = f(t x ) (descending wave) c We also normalized the time axis, so that it takes half a time unit for a wave to travel through a layer Thus at integer time intervals, a wave reaches the surface To describe the scattering of this wave in the medium we use the following notations : f n(i) = descending wave, emerging from boundary i at time n f n (i) = descending wave, reaching boundary i + 1 at time n ẽ n (i) = ascending wave, emerging from boundary i + 1 at time n ẽ n(i) = ascending wave, reaching boundary i at time n These are illustrated in figure 6 We also suppose that the medium is passive and lossless, which means that no energy is added to the wave, nor that it looses energy Using z as a time delay : z 1/2 s n ( ) = s n 1 ( ), we have that in the homogeneous layer i : 2 fn (i) z 1/2 0 f = n (i) ẽ n (i) 0 z 1/2 ẽ n(i) which says that eg, signal f n (i) which emerges from boundary i at time n will reach boundary i + 1 unaltered, half a time unit later, where according to our notation, it is called f n (i) Thus f n (i) = z 1/2 f n (i) = f n 1 2 (i) The scattering matrix Σ i+1 which describes the interaction at boundary i + 1 should be unitary (orthogonal), because if f n (i + 1) fn (i) = Σ i+1 ẽ n (i) ẽ n(i + 1) 16

19 x t i layer i f (i) f (i) ẽ (i) ẽ (i) i + 1 ẽ (i + 1) f (i + 1) Figure 6: Notation of interacting waves then, no energy should be lost; which means f n (i) 2 + ẽ n(i + 1) 2 = n n n f n(i + 1) 2 + n ẽ n (i) 2 (the incoming energy equals the outgoing energy) This is guaranteed when Σ H i+1 Σ i+1 = I Thus we may choose Σ i+1 to be a Givens rotation cos θi+1 sin θ Σ i+1 = i+1 sin θ i+1 cos θ i+1 or, in the complex case ci+1 s Σ i+1 = i+1 s i+1 c i+1, c 2 i+1 + s i+1 2 = 1, c i+1 R Exercise 81 Prove that with this choice of the scattering matrix, the interaction can also be written in the form f n (i + 1) fn (i) ẽ = θ i+1 n(i + 1) ẽ n (i) with θ i+1 = 1 c i+1 1 s i+1 s i+1 1 = 1 1 ρi ρ i+1 ρ i+1 1, ρ i+1 = s i+1 The θ i+1 -matrix of the previous exercise is called a chain scattering matrix It represents a hyperbolic rotation It is unitary with respect to the indefinite metric 1 0 J = 0 1 ie, it is J-unitary, which means that θ H i+1jθ i+1 = J 17

20 This relation is (in the real case) satisfied by a matrix of the form cosh θi+1 sinh θ θ i+1 = i+1 sinh θ i+1 cosh θ i+1 Thus we may describe the complete effect of a layer, including its boundary by f n (i + 1) fn (i) f ẽ = θ n (i) i+1 n(i + 1) ẽ n (i) ẽ n(i) When switching to the time-delayed signals (waves) = θ i+1 z 1/2 0 0 z 1/2 e n(i) = z i/2 ẽ n(i) and f n(i) = z i/2 f n (i), this reduces to f n (i + 1) e n(i + 1) = 1 1 ρi ρ i+1 ρ i+1 1 z f n (i) e n(i) which corresponds to the normalized recursions for the forward and backward prediction errors This application explains the term reflection coefficient for ρ i Looking at figure 6 and the defining relation of Σ i+1, you see when we write the outgoing waves f n (i + 1) and ẽ n (i) in terms of the incoming waves ẽ n(i + 1) and f n (i) as f n(i + 1) = c i+1 fn (i) s i+1 ẽ n(i + 1) ẽ n (i) = s i+1 fn (i) + c i+1 ẽ n(i + 1) then you can check on the figure that the parts of the right hand side with a coefficient c i+1 correspond to transmitted parts and the parts with an s i+1 -coefficient to reflected parts Hence the term reflection coefficient for ρ i+1 = s i+1 The previous analysis suggests the following method to get a model for the earth surface, which is of crucial interest for eg, oil companies We apply an impulse as input at the surface (eg, an explosion) and measure the response of the (pressure) wave at the surface When the Levinson algorithm computes the reflection coefficients for the signal, it actually computes the relative density properties of the different earth layers below the surface, which may disclose the presence of a fluid (oil) or a gas bubble A problem of the above type, where you know the effect of some scattering medium to a signal, and where you want to model the medium is called an inverse scattering problem The same analysis can be applied to eg, an acoustic signal, scattered by an organ pipe or the vocal tract for a speech signal For small time intervals, a speech signal may be considered as stationary Computing the reflection coefficients results in a model for the speech-sound and they can be used to synthesize the speech again by a modeling filter Sending all the reflection coefficients (plus some extra information about pitch etc) uses much less bandwidth than sampling the speech signal and send these samples It is possible to get a considerable data reduction in this way Of course, in such applications the analysis and synthesis should be done in real time The CORDIC (coordinate rotation for digital computing) chip performs a hyperbolic rotation in a few clockcycles 9 Stability If H(z) L 2 is a filter which is causal (and has finite energy) then H(z) = k=0 h kz k represents a function analytic in z < 1 Thus a rational H(z) L 2 is stable ( H 2 ) if all its poles are in 18

21 z > 1 A rational filter is called minimal phase if all its zeros are in z > 1 For an AR model H(z) = G A n (z) = G ϕ n(z), A n(z) = ϕ n(z) = z n ϕ n (1/ z) = z n ϕ n(z) the filter is certainly minimal phase (all zeros at z = ) and it will be stable if all the zeros of A n (z) are in z > 1 or equivalently, the zeros of ϕ n (z) are all in z < 1 Exercise 91 Prove that if ϕ n (α) = 0 then ϕ n(1/ᾱ) = 0 too Since we identified ϕ n as an orthogonal polynomial, this property of its zeros is a consequence of that theory We can however give a direct proof Lemma 91 Define F n (z) = A n(z)/a n (z) with A n (z) = n a k z k Let A n (z) 0 for z 1 Then F n (z) (and zf n (z)) maps the unit disc onto itself Proof Writing A n(z) = ā n 0 i=1 (z α i), with all α i < 1, we check easily that z α i 1 ᾱ i z < 1, = 1, > 1 when z < 1, = 1, > 1 and thus, because also F n (z) = ā0 a 0 k=0 n i=1 z α i 1 ᾱ i z, F n (z) < 1, = 1, > 1 when z < 1, = 1, > 1 The same holds for zf n (z) Hence the functions are injective They are also surjective because the inverse of z w = z α 1 ᾱz is of the same form and thus also injective is w z = w + α 1 + wᾱ Theorem 92 A n (z) has all its zeros in z > 1 iff all ρ i < 1, i = 1,, n Proof This goes by induction on n We only prove the induction step Let A k (z) 0 for z < 1 and ρ k+1 < 1 By the previous lemma, zf k (z) = 1 ρ k+1 has no solution in z 1 Thus A k+1 (z) = A k (z) ρ k+1 za k (z) 0 in z 1 Conversely, if the zeros of A k are in z > 1, the product of its zeros is ( 1) k+1 ρ 1 k (a 0 = 1, a k = ρ k ) Hence ρ k < 1 Furthermore A k 1 (z) = 1 1 ρ k 2 A k(z) + ρ k A k (z) and this can only be zero if F k = 1/ρ k, which can only happen for z > 1 We may thus conclude that the Levinson algorithm generates a filter that is guaranteed to be minimal phase and stable, as long as the reflection coefficients are bounded by 1 When ρ n = 1, then E n = 0 and the approximation is exact Exercise 92 Prove that if ρ k < 1, k = 1, 2,, n 1 and ρ n = 1, then A n (z) = ηa n(z) with η = 1 and its zeros are all of modulus 1 19

22 10 Orthogonal polynomials We have seen before that the backward predictors ϕ n = A n are orthogonal polynomials wrt an inner product whose weight is the power spectrum R(e jω ) Its Fourier coefficients can be expressed as z n, 1 = r n, n Z They satisfy the recurrence relation (recall that ϕ n is monic) ϕ n (z) = zϕ n 1 (z) ρ n ϕ n 1(z), ϕ 0 = 1 or also with ϕn (z) ϕ n(z) ϑ n (z) = ϕn 1 (z) = ϑ n (z) ϕ n 1 (z) 1 ρ n ρ n 1 z Exercise 101 The polynomial ϕ n is orthogonal to span{1, z,, z n 1 } Prove that ϕ n is orthogonal to span{z,, z n } Exercise 102 Prove that the Levinson algorithm can be reformulated in terms of orthogonal polynomials as follows ϕ 0 = 1 E 0 = 1 2 = ϕ 0 2 = r 0 for i = 1, 2, η i 1 = zϕ i 1, ϕ i 1 ρ i = η i 1 /E i 1 ϕ i (z) = zϕ i 1 (z) ρ i ϕ i 1 (z) E i = E i 1 (1 ρ i 2 ) Exercise 103 Prove that det T n = E 0 E 1 E n and thus that E n = det T n / det T n 1 Exercise 104 Derive from the recurrence relation the Christoffel-Darboux type formula (1 xȳ) n k=0 ϕ k (x)ϕ k (y) E k = ϕ n(x)ϕ n(y) xȳϕ n (x)ϕ n (y) E n = ϕ n+1 (x)ϕ n+1 (y) ϕ n+1(x)ϕ n+1 (y) E n The polynomials of the second kind are defined by 1 π e jω + z ψ n (z) = 2π π e jω z ϕ n(e jω ) ϕ n (z)r(e jω )dω, n > 1 r0, n = 0 Exercise 105 Prove that ψ n is a polynomial of degree n 20

23 Exercise 106 Define Ω(z) = 1 2π π π D(t, z)r(t)dω, D(t, z) = t + z t z, t = ejω Prove that { Ω + r0, n = 0 ϕ n Ω + ψ n = z n g(z), n > 0, g(z) analytic in z < 1 ϕ nω ψ n = { Ω r0, n = 0 z n+1 h(z), n > 0, h(z) analytic in z < 1 This shows that ϕ n /ψ n is a two-point Padé approximant for Ω Exercise 107 Prove that the ψ n satisfy the recurrence relation ψ n (z) = zψ n 1 (z) + ρ n ψ n 1(z) Exercise 108 Prove that the recurrences for ϕ n and ψ n can be collected in one relation as ϕn (z) ψ n (z) ϕn 1 (z) ψ ϕ n(z) ψn(z) = ϑ n (z) n 1 (z) 1 ρ ϕ n 1 (z) ψ n 1 (z), ϑ n (z) = n z 0 ρ n and thus that the following holds Θ n (z) = ϑ n (z) ϑ 1 (z) = 1 2r 0 r0 ϕ n + ψ n r 0 ϕ n ψ n r 0 ϕ n ψ n r 0 ϕ n + ψ n which becomes particularly simple if r 0 is normalized to be 1 Exercise 109 Take determinants in the previous exercise to obtain ϕ n (z)ψ n(z) + ϕ nψ n (z) = 2r 0 n 1 (1 ρ i 2 )z n = 2E n z n (101) These polynomials were first studied by Szegő Therefore, they are called Szegő polynomials and in this context, the parameters ρ n are called Szegő coefficients They are the complex conjugates of the reflection coefficients The Levinson algorithm (or its reformulation we gave above) can thus be considered as a fast Gram-Schmidt procedure (only O(n 2 ) flops) to construct orthogonal polynomials Exercise 1010 Set x = y = z in the Christoffel-Darboux formula and derive in this way that ϕ n(z) 2 > 0 for z < 1 Thus all the zeros of ϕ n are in z 1 The zeros of ϕ n(z) can not be on z = 1, because if ϕ n(z 0 ) = 0, z 0 = 1, then ϕ n (z 0 ) = z0 nϕ n(z 0 ) = 0, and this is impossible by (101) This is an alternative proof for the stability of the filter in the AR model Note: The orthonormal polynomials are φ k = ϕ k /G k with G 2 k = E k = ϕ k 2 These satisfy the normalized recursion φn φn 1 φ = θ n (z) n φ n 1 21

24 with normalized transition matrix 1 1 ρ θ n (z) = n 1 ρn 2 ρ n 1 z = 1 1 ρn 2 ϑ n(z) We have proved that the AR filter is stable by construction However, the system which defines the polynomials may be ill conditioned (Note that we basically use the unmodified moments r n = z n, 1 ) The better the approximation of the signal, the smaller the prediction error E n, thus the smaller det T n which means that the matrix of the defining Toeplitz system comes closer and closer to a singular matrix This ill conditioning means that small errors in the computation of the moments r n, perturb the Toeplitz matrix so that it is no longer positive definite and this will imply that the ϕ n no longer have all their zeros in z < 1 Moreover rounding errors during the computation in the Levinson algorithm may propagate fast, which gives erroneous predictor coefficients and also this will affect its stability properties Thus it is advisable to check the stability of the actually computed filter This reasoning is still true in a much wider context: inverse problems (such as the inverse scattering problem) are often very ill conditioned 11 Stability test If we have to check if all the zeros of a polynomial are in z < 1, we could compute its zeros, but this is of course a costly procedure Instead, we could use the previous theory in reverse order We know that a (monic) polynomial ϕ n has all its zeros in z < 1 iff all the ρ k < 1 iff all the E k = ϕ k 2 > 0 iff all the Toeplitz matrices T k are positive definite iff det T k > 0 for all k = 1, 2,, n All this information can be extracted from ϕ n by inverting the Szegő recursion This gives the following algorithm: ρ n = ϕ n (0) for k = n, n 1,, 2 ϕ k 1(z) = ϕ k(z) + ρ k ϕ k (z) z(1 ρ k 2 ) ρ k 1 = ϕ k 1 (0) Exercise 111 Prove the previous algorithm What is its complexity? The previous algorithm is known as the Lehmer-Schur or Jury test 12 Fast (inverse) Cholesky factorization Suppose that we have computed the least squares (polynomial) predictors and the corresponding least squares errors In terms of the (monic) orthogonal polynomials ϕ k, we thus have z k, ϕ i = δ ki E k, k i = 0, 1,, n We can express this orthogonality relation in terms of matrices as follows Denote by U n the unit upper triangular matrix whose k-th column consists of the coefficients of ϕ k (and zeros below the diagonal) The orthogonality can then be expressed as r 0 r 1 r n T n U n = r 1 r 0 U n = r n r 0 22 E 0 E 1 =: F n E n

25 Because T n is hermitian, is upper triangular and therefore U H n T H n = U H n T n = F H n U H n T n U n = D n is diagonal: D n = diag(e 0,, E n ) Note also that Un H = D n Fn 1 This relation implies the upper triangular diagonal lower triangular factorization of T 1 T 1 n = U n Dn 1 Un H This is a kind of Cholesky factorization for Tn 1 We would rather expect Tn 1 to be factorized as a product of a lower triangular diagonal upper triangular But the latter can also be obtained as follows Let us as before denote by Ĩ the reversal operator (matrix) with 1 s on the anti-diagonal Then ĨT nĩ = (T n) T and thus Tn 1 = ĨT T n Ĩ Therefore T 1 n = (ĨŪnĨ)(ĨD 1 n Ĩ)(ĨU T n Ĩ) is indeed of the required form: lower triangular diagonal upper triangular Exercise 121 Check that ĨŪnĨ is the unit lower triangular matrix which contains in its k-th column the coefficients of the polynomial ϕ n k (and zeros above the diagonal) Thus we have identified the Levinson (or Szegő) algorithm as a fast (O(n 2 ) instead of O(n 3 ) flops) Cholesky algorithm for computing the triangular (Cholesky) factors of the (hermitian positive definite) inverted Toeplitz matrix Tn 1, starting from the data in the matrix T n itself Therefore, it is sometimes called a fast inverse Cholesky algorithm The natural question now is whether we can also find the Cholesky factors of T n itself? Of course T n = Un H D n Un 1 = L n D n L H n = F n Dn 1 Fn H with L n = F n D 1 n = Un H, a unit lower triangular matrix But, the question is, can we find the columns of L n (or F n ) directly, ie, without first computing U n? The answer is yes Therefore, consider the vectors (V 1k is the k-th column of U n ) V 1k = ā k ā T C (n+1) 1 V 2k = a 0 a k T C (n+1) 1 where ϕ k (z) = k i=0 a iz i = A k (z) Then we can easily check that (z denotes the down shift operator) η k E k 0 0 where and hence z 0 T n V 1k V 2k 0 1 = E k η k η k = r k+1 r 1 a 0 a k T η k = r 1 r k 1 ā k =: Q 1k Q 2k (k zeros) ā 0 T 23

26 is its complex conjugate and E k = ϕ k 2 = r k r 0 ā k ā 0 T = r 0 r k a 0 a k T These quantities are exactly the ones that showed up in the Levinson algorithm and the reflection coefficient is given by ρ k+1 = η k /E k It is now easy to verify that 1 ρ Q 1k Q 2k k+1 = Q ρ k+1 1 1k Q 2k ϑ T k+1 0 E k z 0 = T n V 1k V 2k ϑ T 0 1 k+1 = E k+1 0 = P 1,k+1 P 2,k+1 (k + 1 zeros) η k+1 The first column of the right hand side is the (k + 1)-st column of F n This gives in principle the algorithm we are looking for since we have again some η k+1 and some E k+1 which will define some ρ k+2 = η k+1 /E k+1 and another step can be performed As you may see from these calculations, it is not necessary to keep all the zeros in the algorithm to compute the reflection coefficients; ie, we may keep from the vectors P 1k P 2k only the elements (the operation z 1 now cuts off the top element of a vector) 1k 2k = P 1k P 2k z k = E k 0 η k, P 1k P 2k = Q 1k Q 2k z z It is then not difficult to see that these reduced vectors satisfy the recurrence relation z 0 1,k+1 2,k+1 = 1k 2k ϑ T 0 1 k+1 z 1 where the reflection coefficients ρ k+1 = η k /E k can be found from the elements E k and η k which are the top elements of the reduced vectors The initial conditions for this procedure are easily found to be = r 0 0 r 1 r 1 An operation count for this procedure will reveal that it requires also O(n 2 ) flops and it is thus a fast Cholesky algorithm for (hermitian positive definite) Toeplitz matrices The representation of the down-shift operation by a multiplication with z and the cutting (upshift) operator as a multiplication with z 1 makes sense in the following way Let n and define the functions (or formal series) D k (z) = D 1k (z) D 2k (z) = z k 1 1 z z 2 1k 2k 24 r n r n z 0 0 1,

27 then the previous recurrence shows that 1 ρ D 1k (z) D 2k (z) k+1 ρ k+1 1 z = D 1,k+1 (z) D 2,k+1 (z) where the operation denoted by z is now a genuine multiplication with z Note that if the D ik (z) converge in a neighborhood of z = 0 and indeed represent functions there, then 13 The Schur algorithm ρ k+1 = lim z 0 D 2k (z) D 1k (z) The functions D ik (z) introduced at the end of the previous section have a meaning Exercise 131 Define Γ k (z) = D 2k (z)/d 1k (z) Prove that the recurrence for the D ik (z) of the previous section implies the recurrence for the Γ k functions Γ k+1 (z) = 1 z Γ k (z) ρ k+1 1 ρ k+1 Γ k (z), ρ k+1 = Γ k (0), The functions Γ k are all Schur functions This means that they belong to the class We shall prove this by induction We start with the function S = {Γ : Γ 1 in z 1 and analytic in z < 1} Ω 0 (z) = r r k z k This is a Carathéodory function or positive real function, which means that it belongs to the class To prove the latter, note that for z < 1, and its real part is C = {Ω : RΩ 0 in z 1 and analytic in z < 1} Ω 0 (z) = 1 2π = 1 2π π π π π ( RΩ 0 (z) = 1 π 2π = 1 2π ) e jkω z k R(e jω )dω e jω + z e jω z R(ejω )dω π π π R ejω + z e jω z R(ejω )dω 1 z 2 e jω z 2 R(ejω )dω which is positive for z < 1 because the kernel (Poisson kernel) and the weight R(e jω ) are positive On the boundary, RΩ 0 (e jω ) = 1 r r k e jkω + r r k e jkω 2 = r k e jkω = R(e jω ) 25

28 and hence positive Because Ω 0 (z) has a causal expansion in z < 1, it is analytic This proves that Ω 0 C Next we note that Γ 0 (z) = 1 Ω 0 (z) Ω 0 (0) z Ω 0 (z) + Ω 0 (0), Ω 0(0) = r 0 > 0 The ratio satisfies Ω 0 (z) Ω 0 (0) 2 Ω 0 (z) + Ω 0 (0) = Ω 0(z) 2 Ω 0 (0)2R(Ω 0 (z)) + Ω 0 (0) 2 Ω 0 (z) 2 + Ω 0 (0)2R(Ω 0 (z)) + Ω 0 (0) 2 1 Because the numerator is zero for z = 0, we can divide out z (Schwarz lemma) and get a function Γ 0 which is thus a Schur function All the other Γ k are also Schur functions as is proved in the induction step below Note that the mapping z z α 1 αz, α < 1 is a bijection of Ĉ = C { } onto itself, which maps the unit circle onto itself, the open unit disk onto itself and the open outside of the unit disk onto itself Exercise 132 Prove this bijection property for the previous Moebius map Thus Γ k ρ k+1 1 ρ k+1 Γ k will again be a Schur function (since ρ k+1 = Γ k (0) is bounded by 1) Because this function is zero for z = 0, we can divide out z and still get a Schur function, which proves that Γ k+1 is in the class S I Schur used the recursion for Γ k to check whether a given function Γ 0 was in the class S He proved the following theorem (for this reason, the ρ k are also referred to as Schur coefficients) Theorem 131 (Schur 1917) Let Γ 0 be a given function Define recursively Γ k+1 (z) = 1 z Γ k (z) ρ k+1 1 ρ k+1 Γ k (z), ρ k+1 = Γ k (0) Then Γ 0 S iff ρ k < 1 for k = 1, 2, or ρ k < 1 for k = 1, 2,, n 1 and ρ n = 1 and ρ n+k = 0 for k = 1, 2, In the latter case, Γ 0 is a Blaschke product of degree n: Γ 0 (z) = P (z) P (z), = c n i=1 P (z) Π n 14 The general pole-zero model z α i 1 α i z, α i < 1, i = 1,, n, c = 1 If we allow a rational model of ARMA type, the problem becomes a nonlinear one (AR modelling was essentially reduced to a linear least squares problem) Hence we shall have to apply iterative methods to minimize the error However, if we have some idea about the location of the zeros, we can propose a filter of the form H(z) = G (1 ᾱ 1z) (1 ᾱ n z) A n (z) = G D n (z) 26

29 where 1/ᾱ i are the estimated zeros, with α i < 1 to make it minimal phase, Define D n (z) = A n(z) π n (z), π n(z) = n (1 ᾱ k z), A n Π n ϕ n (z) = A n(z) π n (z) then both ϕ n and D n are in the space { } p n n L n =, p n Π n, π n (z) = (1 ᾱ k z) π n of rational functions with prescribed poles (outside the closed unit disk) Exercise 141 Prove that the space L n is also spanned by the basis functions {ζ k, k = 1, 2,, n} with ζ 0 = 1, ζ k (z) = z α k, k = 1, 2,, n 1 ᾱ k z or by {B k, k = 1,, n} where Exercise 142 Prove that B 0 = ζ 0, B k = B k 1 ζ k, k = 1,, n ϕ n (z) = D n(z) := B n (z)d n (z) = B n (z)d n (1/ z) L n Note that when all α k = 0, the space L n reduces to the space Π n of polynomials and we are back in the AR case If we want for a stochastic process (s n ) to predict s 0 from its past, ie, from span{s k, k = 1, 2, } in best L 2 sense, we should find (h n ) such that 2 s 0 + h k s k is minimal Under the Kolmogorov isomorphism, this is equivalent to finding the minimum of H 2, where H(z) = 1 + h k z k and where the norm is taken in the Hilbert space with inner product Thus we look for f, g = 1 2π π π f(e jω )g(e jω )R(e jω )dω min{ H 2 : H H 2, H(0) = 1} In the AR model, this (infinite dimensional) problem was approximated by minimizing over polynomial subspaces Π n H 2 However, it is perfectly possible to minimize over any other finite dimensional subspace of H 2, for example the subspaces L n We should then solve the problems min{ f 2, f L n H 2, f(0) = 1} 27

30 To solve this problem (theoretically and practically) we need the concept of reproducing kernel Let H be a Hilbert space, then we call k(z, w) a reproducing kernel for H, if w k(z, w), f(z) = f(w), f H Exercise 143 Prove that if ϕ k, k = 0, 1, is an orthogonal basis for H, then k(z, w) = k ϕ k (z)ϕ k (w) ϕ k 2 is a reproducing kernel in H We have now the following Theorem 141 The optimization problem min{ f 2, f H, f(w) = 1} in the Hilbert space H with reproducing kernel k is solved for and the minimum is 1/k(w, w) f(z) = g w (z) = k(z, w) k(w, w) Proof It is clear that g w (w) = 1 and g w 2 = 1/k(w, w) Let f = k a kϕ k (z) be an arbitrary function in H such that f(w) = 1 (suppose ϕ k forms an orthonormal basis) Then we have to minimize f 2 = a k 2 k with the side condition k a kϕ k (w) = 1 By the Cauchy-Schwarz inequality 1 = k a k ϕ k (w) 2 ( k a k 2 )( k ϕ k (w) 2 ) which implies 1 k(w, w) = 1 k ϕ k(w) 2 k a k 2 The lower bound is indeed reached by choosing a k = ϕ k (w)/k(w, w) and this proves the theorem A possible solution method for the prediction problem is now clear: we first construct the orthonormal functions ϕ 0, ϕ 1,, ϕ n, eg, by the orthogonalization of the basis B 0, B 1,, B n and we so get the reproducing kernel n ϕ k (z)ϕ k (w) k n (z, w) = ϕ k 2 k=0 It is even possible to find a recurrence relation for the reproducing kernels directly The optimal predictor from L n is then given by H(z) = Gk n(0, 0) k n(z, 0) with k n(z, 0) = B n (z)k n (z, 0) 28

Lab 9a. Linear Predictive Coding for Speech Processing

Lab 9a. Linear Predictive Coding for Speech Processing EE275Lab October 27, 2007 Lab 9a. Linear Predictive Coding for Speech Processing Pitch Period Impulse Train Generator Voiced/Unvoiced Speech Switch Vocal Tract Parameters Time-Varying Digital Filter H(z)

More information

Advanced Digital Signal Processing -Introduction

Advanced Digital Signal Processing -Introduction Advanced Digital Signal Processing -Introduction LECTURE-2 1 AP9211- ADVANCED DIGITAL SIGNAL PROCESSING UNIT I DISCRETE RANDOM SIGNAL PROCESSING Discrete Random Processes- Ensemble Averages, Stationary

More information

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes (cont d)

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes (cont d) Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes (cont d) Electrical & Computer Engineering North Carolina State University Acknowledgment: ECE792-41 slides

More information

Statistical and Adaptive Signal Processing

Statistical and Adaptive Signal Processing r Statistical and Adaptive Signal Processing Spectral Estimation, Signal Modeling, Adaptive Filtering and Array Processing Dimitris G. Manolakis Massachusetts Institute of Technology Lincoln Laboratory

More information

Katholieke Universiteit Leuven Department of Computer Science

Katholieke Universiteit Leuven Department of Computer Science Separation of zeros of para-orthogonal rational functions A. Bultheel, P. González-Vera, E. Hendriksen, O. Njåstad Report TW 402, September 2004 Katholieke Universiteit Leuven Department of Computer Science

More information

Elementary linear algebra

Elementary linear algebra Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The

More information

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product Chapter 4 Hilbert Spaces 4.1 Inner Product Spaces Inner Product Space. A complex vector space E is called an inner product space (or a pre-hilbert space, or a unitary space) if there is a mapping (, )

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

1. Foundations of Numerics from Advanced Mathematics. Linear Algebra

1. Foundations of Numerics from Advanced Mathematics. Linear Algebra Foundations of Numerics from Advanced Mathematics Linear Algebra Linear Algebra, October 23, 22 Linear Algebra Mathematical Structures a mathematical structure consists of one or several sets and one or

More information

CHAPTER 2 RANDOM PROCESSES IN DISCRETE TIME

CHAPTER 2 RANDOM PROCESSES IN DISCRETE TIME CHAPTER 2 RANDOM PROCESSES IN DISCRETE TIME Shri Mata Vaishno Devi University, (SMVDU), 2013 Page 13 CHAPTER 2 RANDOM PROCESSES IN DISCRETE TIME When characterizing or modeling a random variable, estimates

More information

1 Math 241A-B Homework Problem List for F2015 and W2016

1 Math 241A-B Homework Problem List for F2015 and W2016 1 Math 241A-B Homework Problem List for F2015 W2016 1.1 Homework 1. Due Wednesday, October 7, 2015 Notation 1.1 Let U be any set, g be a positive function on U, Y be a normed space. For any f : U Y let

More information

1. Determine if each of the following are valid autocorrelation matrices of WSS processes. (Correlation Matrix),R c =

1. Determine if each of the following are valid autocorrelation matrices of WSS processes. (Correlation Matrix),R c = ENEE630 ADSP Part II w/ solution. Determine if each of the following are valid autocorrelation matrices of WSS processes. (Correlation Matrix) R a = 4 4 4,R b = 0 0,R c = j 0 j 0 j 0 j 0 j,r d = 0 0 0

More information

Very useful for designing and analyzing signal processing systems

Very useful for designing and analyzing signal processing systems z-transform z-transform The z-transform generalizes the Discrete-Time Fourier Transform (DTFT) for analyzing infinite-length signals and systems Very useful for designing and analyzing signal processing

More information

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v ) Section 3.2 Theorem 3.6. Let A be an m n matrix of rank r. Then r m, r n, and, by means of a finite number of elementary row and column operations, A can be transformed into the matrix ( ) Ir O D = 1 O

More information

Part IB. Further Analysis. Year

Part IB. Further Analysis. Year Year 2004 2003 2002 2001 10 2004 2/I/4E Let τ be the topology on N consisting of the empty set and all sets X N such that N \ X is finite. Let σ be the usual topology on R, and let ρ be the topology on

More information

Lecture 7: Linear Prediction

Lecture 7: Linear Prediction 1 Lecture 7: Linear Prediction Overview Dealing with three notions: PREDICTION, PREDICTOR, PREDICTION ERROR; FORWARD versus BACKWARD: Predicting the future versus (improper terminology) predicting the

More information

Considering our result for the sum and product of analytic functions, this means that for (a 0, a 1,..., a N ) C N+1, the polynomial.

Considering our result for the sum and product of analytic functions, this means that for (a 0, a 1,..., a N ) C N+1, the polynomial. Lecture 3 Usual complex functions MATH-GA 245.00 Complex Variables Polynomials. Construction f : z z is analytic on all of C since its real and imaginary parts satisfy the Cauchy-Riemann relations and

More information

Conformal maps. Lent 2019 COMPLEX METHODS G. Taylor. A star means optional and not necessarily harder.

Conformal maps. Lent 2019 COMPLEX METHODS G. Taylor. A star means optional and not necessarily harder. Lent 29 COMPLEX METHODS G. Taylor A star means optional and not necessarily harder. Conformal maps. (i) Let f(z) = az + b, with ad bc. Where in C is f conformal? cz + d (ii) Let f(z) = z +. What are the

More information

08a. Operators on Hilbert spaces. 1. Boundedness, continuity, operator norms

08a. Operators on Hilbert spaces. 1. Boundedness, continuity, operator norms (February 24, 2017) 08a. Operators on Hilbert spaces Paul Garrett garrett@math.umn.edu http://www.math.umn.edu/ garrett/ [This document is http://www.math.umn.edu/ garrett/m/real/notes 2016-17/08a-ops

More information

Rational bases for system identification

Rational bases for system identification Rational bases for system identification Adhemar Bultheel, Patrick Van gucht Department Computer Science Numerical Approximation and Linear Algebra Group (NALAG) K.U.Leuven, Belgium adhemar.bultheel.cs.kuleuven.ac.be

More information

COMPLEX ANALYSIS Spring 2014

COMPLEX ANALYSIS Spring 2014 COMPLEX ANALYSIS Spring 24 Homework 4 Solutions Exercise Do and hand in exercise, Chapter 3, p. 4. Solution. The exercise states: Show that if a

More information

Practical Spectral Estimation

Practical Spectral Estimation Digital Signal Processing/F.G. Meyer Lecture 4 Copyright 2015 François G. Meyer. All Rights Reserved. Practical Spectral Estimation 1 Introduction The goal of spectral estimation is to estimate how the

More information

Diagonalizing Matrices

Diagonalizing Matrices Diagonalizing Matrices Massoud Malek A A Let A = A k be an n n non-singular matrix and let B = A = [B, B,, B k,, B n ] Then A n A B = A A 0 0 A k [B, B,, B k,, B n ] = 0 0 = I n 0 A n Notice that A i B

More information

Chapter 8 Integral Operators

Chapter 8 Integral Operators Chapter 8 Integral Operators In our development of metrics, norms, inner products, and operator theory in Chapters 1 7 we only tangentially considered topics that involved the use of Lebesgue measure,

More information

Probability Space. J. McNames Portland State University ECE 538/638 Stochastic Signals Ver

Probability Space. J. McNames Portland State University ECE 538/638 Stochastic Signals Ver Stochastic Signals Overview Definitions Second order statistics Stationarity and ergodicity Random signal variability Power spectral density Linear systems with stationary inputs Random signal memory Correlation

More information

TRUNCATED TOEPLITZ OPERATORS ON FINITE DIMENSIONAL SPACES

TRUNCATED TOEPLITZ OPERATORS ON FINITE DIMENSIONAL SPACES TRUNCATED TOEPLITZ OPERATORS ON FINITE DIMENSIONAL SPACES JOSEPH A. CIMA, WILLIAM T. ROSS, AND WARREN R. WOGEN Abstract. In this paper, we study the matrix representations of compressions of Toeplitz operators

More information

TOEPLITZ OPERATORS. Toeplitz studied infinite matrices with NW-SE diagonals constant. f e C :

TOEPLITZ OPERATORS. Toeplitz studied infinite matrices with NW-SE diagonals constant. f e C : TOEPLITZ OPERATORS EFTON PARK 1. Introduction to Toeplitz Operators Otto Toeplitz lived from 1881-1940 in Goettingen, and it was pretty rough there, so he eventually went to Palestine and eventually contracted

More information

Chapter 9. Linear Predictive Analysis of Speech Signals 语音信号的线性预测分析

Chapter 9. Linear Predictive Analysis of Speech Signals 语音信号的线性预测分析 Chapter 9 Linear Predictive Analysis of Speech Signals 语音信号的线性预测分析 1 LPC Methods LPC methods are the most widely used in speech coding, speech synthesis, speech recognition, speaker recognition and verification

More information

Review of some mathematical tools

Review of some mathematical tools MATHEMATICAL FOUNDATIONS OF SIGNAL PROCESSING Fall 2016 Benjamín Béjar Haro, Mihailo Kolundžija, Reza Parhizkar, Adam Scholefield Teaching assistants: Golnoosh Elhami, Hanjie Pan Review of some mathematical

More information

REPRESENTATION THEORY WEEK 7

REPRESENTATION THEORY WEEK 7 REPRESENTATION THEORY WEEK 7 1. Characters of L k and S n A character of an irreducible representation of L k is a polynomial function constant on every conjugacy class. Since the set of diagonalizable

More information

MATH 722, COMPLEX ANALYSIS, SPRING 2009 PART 5

MATH 722, COMPLEX ANALYSIS, SPRING 2009 PART 5 MATH 722, COMPLEX ANALYSIS, SPRING 2009 PART 5.. The Arzela-Ascoli Theorem.. The Riemann mapping theorem Let X be a metric space, and let F be a family of continuous complex-valued functions on X. We have

More information

Nonparametric and Parametric Defined This text distinguishes between systems and the sequences (processes) that result when a WN input is applied

Nonparametric and Parametric Defined This text distinguishes between systems and the sequences (processes) that result when a WN input is applied Linear Signal Models Overview Introduction Linear nonparametric vs. parametric models Equivalent representations Spectral flatness measure PZ vs. ARMA models Wold decomposition Introduction Many researchers

More information

CHAPTER VIII HILBERT SPACES

CHAPTER VIII HILBERT SPACES CHAPTER VIII HILBERT SPACES DEFINITION Let X and Y be two complex vector spaces. A map T : X Y is called a conjugate-linear transformation if it is a reallinear transformation from X into Y, and if T (λx)

More information

Qualifying Exam Complex Analysis (Math 530) January 2019

Qualifying Exam Complex Analysis (Math 530) January 2019 Qualifying Exam Complex Analysis (Math 53) January 219 1. Let D be a domain. A function f : D C is antiholomorphic if for every z D the limit f(z + h) f(z) lim h h exists. Write f(z) = f(x + iy) = u(x,

More information

Inner product spaces. Layers of structure:

Inner product spaces. Layers of structure: Inner product spaces Layers of structure: vector space normed linear space inner product space The abstract definition of an inner product, which we will see very shortly, is simple (and by itself is pretty

More information

Statistical Signal Processing Detection, Estimation, and Time Series Analysis

Statistical Signal Processing Detection, Estimation, and Time Series Analysis Statistical Signal Processing Detection, Estimation, and Time Series Analysis Louis L. Scharf University of Colorado at Boulder with Cedric Demeure collaborating on Chapters 10 and 11 A TT ADDISON-WESLEY

More information

CHAPTER 8. Smoothing operators

CHAPTER 8. Smoothing operators CHAPTER 8 Smoothing operators Lecture 8: 13 October, 2005 Now I am heading towards the Atiyah-Singer index theorem. Most of the results proved in the process untimately reduce to properties of smoothing

More information

ADAPTIVE FILTER THEORY

ADAPTIVE FILTER THEORY ADAPTIVE FILTER THEORY Fourth Edition Simon Haykin Communications Research Laboratory McMaster University Hamilton, Ontario, Canada Front ice Hall PRENTICE HALL Upper Saddle River, New Jersey 07458 Preface

More information

X (z) = n= 1. Ã! X (z) x [n] X (z) = Z fx [n]g x [n] = Z 1 fx (z)g. r n x [n] ª e jnω

X (z) = n= 1. Ã! X (z) x [n] X (z) = Z fx [n]g x [n] = Z 1 fx (z)g. r n x [n] ª e jnω 3 The z-transform ² Two advantages with the z-transform:. The z-transform is a generalization of the Fourier transform for discrete-time signals; which encompasses a broader class of sequences. The z-transform

More information

6. Methods for Rational Spectra It is assumed that signals have rational spectra m k= m

6. Methods for Rational Spectra It is assumed that signals have rational spectra m k= m 6. Methods for Rational Spectra It is assumed that signals have rational spectra m k= m φ(ω) = γ ke jωk n k= n ρ, (23) jωk ke where γ k = γ k and ρ k = ρ k. Any continuous PSD can be approximated arbitrary

More information

A system that is both linear and time-invariant is called linear time-invariant (LTI).

A system that is both linear and time-invariant is called linear time-invariant (LTI). The Cooper Union Department of Electrical Engineering ECE111 Signal Processing & Systems Analysis Lecture Notes: Time, Frequency & Transform Domains February 28, 2012 Signals & Systems Signals are mapped

More information

Mathematical Optimisation, Chpt 2: Linear Equations and inequalities

Mathematical Optimisation, Chpt 2: Linear Equations and inequalities Mathematical Optimisation, Chpt 2: Linear Equations and inequalities Peter J.C. Dickinson p.j.c.dickinson@utwente.nl http://dickinson.website version: 12/02/18 Monday 5th February 2018 Peter J.C. Dickinson

More information

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes Electrical & Computer Engineering North Carolina State University Acknowledgment: ECE792-41 slides were adapted

More information

MORE NOTES FOR MATH 823, FALL 2007

MORE NOTES FOR MATH 823, FALL 2007 MORE NOTES FOR MATH 83, FALL 007 Prop 1.1 Prop 1. Lemma 1.3 1. The Siegel upper half space 1.1. The Siegel upper half space and its Bergman kernel. The Siegel upper half space is the domain { U n+1 z C

More information

7 Asymptotics for Meromorphic Functions

7 Asymptotics for Meromorphic Functions Lecture G jacques@ucsd.edu 7 Asymptotics for Meromorphic Functions Hadamard s Theorem gives a broad description of the exponential growth of coefficients in power series, but the notion of exponential

More information

Fourth Week: Lectures 10-12

Fourth Week: Lectures 10-12 Fourth Week: Lectures 10-12 Lecture 10 The fact that a power series p of positive radius of convergence defines a function inside its disc of convergence via substitution is something that we cannot ignore

More information

Computation of Rational Szegő-Lobatto Quadrature Formulas

Computation of Rational Szegő-Lobatto Quadrature Formulas Computation of Rational Szegő-Lobatto Quadrature Formulas Joint work with: A. Bultheel (Belgium) E. Hendriksen (The Netherlands) O. Njastad (Norway) P. GONZÁLEZ-VERA Departament of Mathematical Analysis.

More information

Real Variables # 10 : Hilbert Spaces II

Real Variables # 10 : Hilbert Spaces II randon ehring Real Variables # 0 : Hilbert Spaces II Exercise 20 For any sequence {f n } in H with f n = for all n, there exists f H and a subsequence {f nk } such that for all g H, one has lim (f n k,

More information

Linear Passive Stationary Scattering Systems with Pontryagin State Spaces

Linear Passive Stationary Scattering Systems with Pontryagin State Spaces mn header will be provided by the publisher Linear Passive Stationary Scattering Systems with Pontryagin State Spaces D. Z. Arov 1, J. Rovnyak 2, and S. M. Saprikin 1 1 Department of Physics and Mathematics,

More information

Special Topics: Data Science

Special Topics: Data Science Special Topics: Data Science L6b: Wiener Filter Victor Solo School of Electrical Engineering University of New South Wales Sydney, AUSTRALIA Topics Norbert Wiener 1894-1964 MIT Professor (Frequency Domain)

More information

Lesson 1. Optimal signalbehandling LTH. September Statistical Digital Signal Processing and Modeling, Hayes, M:

Lesson 1. Optimal signalbehandling LTH. September Statistical Digital Signal Processing and Modeling, Hayes, M: Lesson 1 Optimal Signal Processing Optimal signalbehandling LTH September 2013 Statistical Digital Signal Processing and Modeling, Hayes, M: John Wiley & Sons, 1996. ISBN 0471594318 Nedelko Grbic Mtrl

More information

Hilbert Spaces. Hilbert space is a vector space with some extra structure. We start with formal (axiomatic) definition of a vector space.

Hilbert Spaces. Hilbert space is a vector space with some extra structure. We start with formal (axiomatic) definition of a vector space. Hilbert Spaces Hilbert space is a vector space with some extra structure. We start with formal (axiomatic) definition of a vector space. Vector Space. Vector space, ν, over the field of complex numbers,

More information

III. Consequences of Cauchy s Theorem

III. Consequences of Cauchy s Theorem MTH6 Complex Analysis 2009-0 Lecture Notes c Shaun Bullett 2009 III. Consequences of Cauchy s Theorem. Cauchy s formulae. Cauchy s Integral Formula Let f be holomorphic on and everywhere inside a simple

More information

Math Linear Algebra II. 1. Inner Products and Norms

Math Linear Algebra II. 1. Inner Products and Norms Math 342 - Linear Algebra II Notes 1. Inner Products and Norms One knows from a basic introduction to vectors in R n Math 254 at OSU) that the length of a vector x = x 1 x 2... x n ) T R n, denoted x,

More information

Notation. General. Notation Description See. Sets, Functions, and Spaces. a b & a b The minimum and the maximum of a and b

Notation. General. Notation Description See. Sets, Functions, and Spaces. a b & a b The minimum and the maximum of a and b Notation General Notation Description See a b & a b The minimum and the maximum of a and b a + & a f S u The non-negative part, a 0, and non-positive part, (a 0) of a R The restriction of the function

More information

4 Derivations of the Discrete-Time Kalman Filter

4 Derivations of the Discrete-Time Kalman Filter Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof N Shimkin 4 Derivations of the Discrete-Time

More information

Time Series 3. Robert Almgren. Sept. 28, 2009

Time Series 3. Robert Almgren. Sept. 28, 2009 Time Series 3 Robert Almgren Sept. 28, 2009 Last time we discussed two main categories of linear models, and their combination. Here w t denotes a white noise: a stationary process with E w t ) = 0, E

More information

Review of Discrete-Time System

Review of Discrete-Time System Review of Discrete-Time System Electrical & Computer Engineering University of Maryland, College Park Acknowledgment: ENEE630 slides were based on class notes developed by Profs. K.J. Ray Liu and Min Wu.

More information

Functions of a Complex Variable and Integral Transforms

Functions of a Complex Variable and Integral Transforms Functions of a Complex Variable and Integral Transforms Department of Mathematics Zhou Lingjun Textbook Functions of Complex Analysis with Applications to Engineering and Science, 3rd Edition. A. D. Snider

More information

Lecture 4 - Spectral Estimation

Lecture 4 - Spectral Estimation Lecture 4 - Spectral Estimation The Discrete Fourier Transform The Discrete Fourier Transform (DFT) is the equivalent of the continuous Fourier Transform for signals known only at N instants separated

More information

Use: Analysis of systems, simple convolution, shorthand for e jw, stability. Motivation easier to write. Or X(z) = Z {x(n)}

Use: Analysis of systems, simple convolution, shorthand for e jw, stability. Motivation easier to write. Or X(z) = Z {x(n)} 1 VI. Z Transform Ch 24 Use: Analysis of systems, simple convolution, shorthand for e jw, stability. A. Definition: X(z) = x(n) z z - transforms Motivation easier to write Or Note if X(z) = Z {x(n)} z

More information

Cambridge University Press The Mathematics of Signal Processing Steven B. Damelin and Willard Miller Excerpt More information

Cambridge University Press The Mathematics of Signal Processing Steven B. Damelin and Willard Miller Excerpt More information Introduction Consider a linear system y = Φx where Φ can be taken as an m n matrix acting on Euclidean space or more generally, a linear operator on a Hilbert space. We call the vector x a signal or input,

More information

The following definition is fundamental.

The following definition is fundamental. 1. Some Basics from Linear Algebra With these notes, I will try and clarify certain topics that I only quickly mention in class. First and foremost, I will assume that you are familiar with many basic

More information

Voiced Speech. Unvoiced Speech

Voiced Speech. Unvoiced Speech Digital Speech Processing Lecture 2 Homomorphic Speech Processing General Discrete-Time Model of Speech Production p [ n] = p[ n] h [ n] Voiced Speech L h [ n] = A g[ n] v[ n] r[ n] V V V p [ n ] = u [

More information

Pseudospectra and Nonnormal Dynamical Systems

Pseudospectra and Nonnormal Dynamical Systems Pseudospectra and Nonnormal Dynamical Systems Mark Embree and Russell Carden Computational and Applied Mathematics Rice University Houston, Texas ELGERSBURG MARCH 1 Overview of the Course These lectures

More information

Numerical Linear Algebra

Numerical Linear Algebra University of Alabama at Birmingham Department of Mathematics Numerical Linear Algebra Lecture Notes for MA 660 (1997 2014) Dr Nikolai Chernov April 2014 Chapter 0 Review of Linear Algebra 0.1 Matrices

More information

RIEMANN MAPPING THEOREM

RIEMANN MAPPING THEOREM RIEMANN MAPPING THEOREM VED V. DATAR Recall that two domains are called conformally equivalent if there exists a holomorphic bijection from one to the other. This automatically implies that there is an

More information

Time Series 2. Robert Almgren. Sept. 21, 2009

Time Series 2. Robert Almgren. Sept. 21, 2009 Time Series 2 Robert Almgren Sept. 21, 2009 This week we will talk about linear time series models: AR, MA, ARMA, ARIMA, etc. First we will talk about theory and after we will talk about fitting the models

More information

Vectors in Function Spaces

Vectors in Function Spaces Jim Lambers MAT 66 Spring Semester 15-16 Lecture 18 Notes These notes correspond to Section 6.3 in the text. Vectors in Function Spaces We begin with some necessary terminology. A vector space V, also

More information

David Hilbert was old and partly deaf in the nineteen thirties. Yet being a diligent

David Hilbert was old and partly deaf in the nineteen thirties. Yet being a diligent Chapter 5 ddddd dddddd dddddddd ddddddd dddddddd ddddddd Hilbert Space The Euclidean norm is special among all norms defined in R n for being induced by the Euclidean inner product (the dot product). A

More information

Signal Analysis, Systems, Transforms

Signal Analysis, Systems, Transforms Michael J. Corinthios Signal Analysis, Systems, Transforms Engineering Book (English) August 29, 2007 Springer Contents Discrete-Time Signals and Systems......................... Introduction.............................................2

More information

Kernel Method: Data Analysis with Positive Definite Kernels

Kernel Method: Data Analysis with Positive Definite Kernels Kernel Method: Data Analysis with Positive Definite Kernels 2. Positive Definite Kernel and Reproducing Kernel Hilbert Space Kenji Fukumizu The Institute of Statistical Mathematics. Graduate University

More information

Statistical Geometry Processing Winter Semester 2011/2012

Statistical Geometry Processing Winter Semester 2011/2012 Statistical Geometry Processing Winter Semester 2011/2012 Linear Algebra, Function Spaces & Inverse Problems Vector and Function Spaces 3 Vectors vectors are arrows in space classically: 2 or 3 dim. Euclidian

More information

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces. Math 350 Fall 2011 Notes about inner product spaces In this notes we state and prove some important properties of inner product spaces. First, recall the dot product on R n : if x, y R n, say x = (x 1,...,

More information

Rational Covariance Extension for Boundary Data and Positive Real Lemma with Positive Semidefinite Matrix Solution

Rational Covariance Extension for Boundary Data and Positive Real Lemma with Positive Semidefinite Matrix Solution Preprints of the 18th FAC World Congress Milano (taly) August 28 - September 2, 2011 Rational Covariance Extension for Boundary Data and Positive Real Lemma with Positive Semidefinite Matrix Solution Y

More information

LINEAR FRACTIONAL COMPOSITION OPERATORS ON H 2

LINEAR FRACTIONAL COMPOSITION OPERATORS ON H 2 J Integral Equations and Operator Theory (988, 5 60 LINEAR FRACTIONAL COMPOSITION OPERATORS ON H 2 CARL C COWEN Abstract If ϕ is an analytic function mapping the unit disk D into itself, the composition

More information

Chapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space.

Chapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space. Chapter 1 Preliminaries The purpose of this chapter is to provide some basic background information. Linear Space Hilbert Space Basic Principles 1 2 Preliminaries Linear Space The notion of linear space

More information

MTH 2032 SemesterII

MTH 2032 SemesterII MTH 202 SemesterII 2010-11 Linear Algebra Worked Examples Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education December 28, 2011 ii Contents Table of Contents

More information

7.2 Conformal mappings

7.2 Conformal mappings 7.2 Conformal mappings Let f be an analytic function. At points where f (z) 0 such a map has the remarkable property that it is conformal. This means that angle is preserved (in the sense that any 2 smooth

More information

Applied Analysis (APPM 5440): Final exam 1:30pm 4:00pm, Dec. 14, Closed books.

Applied Analysis (APPM 5440): Final exam 1:30pm 4:00pm, Dec. 14, Closed books. Applied Analysis APPM 44: Final exam 1:3pm 4:pm, Dec. 14, 29. Closed books. Problem 1: 2p Set I = [, 1]. Prove that there is a continuous function u on I such that 1 ux 1 x sin ut 2 dt = cosx, x I. Define

More information

(v, w) = arccos( < v, w >

(v, w) = arccos( < v, w > MA322 F all206 Notes on Inner Products Notes on Chapter 6 Inner product. Given a real vector space V, an inner product is defined to be a bilinear map F : V V R such that the following holds: Commutativity:

More information

Kaczmarz algorithm in Hilbert space

Kaczmarz algorithm in Hilbert space STUDIA MATHEMATICA 169 (2) (2005) Kaczmarz algorithm in Hilbert space by Rainis Haller (Tartu) and Ryszard Szwarc (Wrocław) Abstract The aim of the Kaczmarz algorithm is to reconstruct an element in a

More information

Boundary Nevanlinna Pick interpolation problems for generalized Schur functions

Boundary Nevanlinna Pick interpolation problems for generalized Schur functions , Boundary Nevanlinna Pick interpolation problems for generalized Schur functions Vladimir Bolotnikov and Alexander Kheifets Abstract. Three boundary Nevanlinna-Pick interpolation problems at finitely

More information

1 Assignment 1: Nonlinear dynamics (due September

1 Assignment 1: Nonlinear dynamics (due September Assignment : Nonlinear dynamics (due September 4, 28). Consider the ordinary differential equation du/dt = cos(u). Sketch the equilibria and indicate by arrows the increase or decrease of the solutions.

More information

Complex Analysis, Stein and Shakarchi Meromorphic Functions and the Logarithm

Complex Analysis, Stein and Shakarchi Meromorphic Functions and the Logarithm Complex Analysis, Stein and Shakarchi Chapter 3 Meromorphic Functions and the Logarithm Yung-Hsiang Huang 217.11.5 Exercises 1. From the identity sin πz = eiπz e iπz 2i, it s easy to show its zeros are

More information

ADVANCED CALCULUS - MTH433 LECTURE 4 - FINITE AND INFINITE SETS

ADVANCED CALCULUS - MTH433 LECTURE 4 - FINITE AND INFINITE SETS ADVANCED CALCULUS - MTH433 LECTURE 4 - FINITE AND INFINITE SETS 1. Cardinal number of a set The cardinal number (or simply cardinal) of a set is a generalization of the concept of the number of elements

More information

THEODORE VORONOV DIFFERENTIABLE MANIFOLDS. Fall Last updated: November 26, (Under construction.)

THEODORE VORONOV DIFFERENTIABLE MANIFOLDS. Fall Last updated: November 26, (Under construction.) 4 Vector fields Last updated: November 26, 2009. (Under construction.) 4.1 Tangent vectors as derivations After we have introduced topological notions, we can come back to analysis on manifolds. Let M

More information

October 25, 2013 INNER PRODUCT SPACES

October 25, 2013 INNER PRODUCT SPACES October 25, 2013 INNER PRODUCT SPACES RODICA D. COSTIN Contents 1. Inner product 2 1.1. Inner product 2 1.2. Inner product spaces 4 2. Orthogonal bases 5 2.1. Existence of an orthogonal basis 7 2.2. Orthogonal

More information

Commutative Banach algebras 79

Commutative Banach algebras 79 8. Commutative Banach algebras In this chapter, we analyze commutative Banach algebras in greater detail. So we always assume that xy = yx for all x, y A here. Definition 8.1. Let A be a (commutative)

More information

Generalizations of orthogonal polynomials

Generalizations of orthogonal polynomials WOG project J. Comput. Appl. Math., Preprint 28 June 2004 Generalizations of orthogonal polynomials A. Bultheel Dept. Computer Science (NALAG), K.U.Leuven, Celestijnenlaan 200 A, B-300 Leuven, Belgium.

More information

Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014

Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014 Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014 Linear Algebra A Brief Reminder Purpose. The purpose of this document

More information

4 Hilbert spaces. The proof of the Hilbert basis theorem is not mathematics, it is theology. Camille Jordan

4 Hilbert spaces. The proof of the Hilbert basis theorem is not mathematics, it is theology. Camille Jordan The proof of the Hilbert basis theorem is not mathematics, it is theology. Camille Jordan Wir müssen wissen, wir werden wissen. David Hilbert We now continue to study a special class of Banach spaces,

More information

Analysis of Discrete-Time Systems

Analysis of Discrete-Time Systems TU Berlin Discrete-Time Control Systems 1 Analysis of Discrete-Time Systems Overview Stability Sensitivity and Robustness Controllability, Reachability, Observability, and Detectabiliy TU Berlin Discrete-Time

More information

Linear Algebra March 16, 2019

Linear Algebra March 16, 2019 Linear Algebra March 16, 2019 2 Contents 0.1 Notation................................ 4 1 Systems of linear equations, and matrices 5 1.1 Systems of linear equations..................... 5 1.2 Augmented

More information

Spectral Theory of Orthogonal Polynomials

Spectral Theory of Orthogonal Polynomials Spectral Theory of Orthogonal Polynomials Barry Simon IBM Professor of Mathematics and Theoretical Physics California Institute of Technology Pasadena, CA, U.S.A. Lecture 1: Introduction and Overview Spectral

More information

14.2 QR Factorization with Column Pivoting

14.2 QR Factorization with Column Pivoting page 531 Chapter 14 Special Topics Background Material Needed Vector and Matrix Norms (Section 25) Rounding Errors in Basic Floating Point Operations (Section 33 37) Forward Elimination and Back Substitution

More information

5.3 The Upper Half Plane

5.3 The Upper Half Plane Remark. Combining Schwarz Lemma with the map g α, we can obtain some inequalities of analytic maps f : D D. For example, if z D and w = f(z) D, then the composition h := g w f g z satisfies the condition

More information

Homework 2 Foundations of Computational Math 2 Spring 2019

Homework 2 Foundations of Computational Math 2 Spring 2019 Homework 2 Foundations of Computational Math 2 Spring 2019 Problem 2.1 (2.1.a) Suppose (v 1,λ 1 )and(v 2,λ 2 ) are eigenpairs for a matrix A C n n. Show that if λ 1 λ 2 then v 1 and v 2 are linearly independent.

More information

Some Time-Series Models

Some Time-Series Models Some Time-Series Models Outline 1. Stochastic processes and their properties 2. Stationary processes 3. Some properties of the autocorrelation function 4. Some useful models Purely random processes, random

More information

SGN Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection

SGN Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection SG 21006 Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection Ioan Tabus Department of Signal Processing Tampere University of Technology Finland 1 / 28

More information