Heinrich Voss. Dedicated to Richard S. Varga on the occasion of his 70th birthday. Abstract

Similar documents
ETNA Kent State University

NUMERICAL SOLUTION OF THE EIGENVALUE PROBLEM FOR HERMITIAN TOEPLITZ MATRICES. William F. Trench* SIAM J. Matrix Anal. Appl.

Numerical solution of the eigenvalue problem for Hermitian Toeplitz-like matrices

Eigenvalue Problems CHAPTER 1 : PRELIMINARIES

The geometric mean algorithm

Technical University Hamburg { Harburg, Section of Mathematics, to reduce the number of degrees of freedom to manageable size.

Wolfgang Mackens and Heinrich Voss. Abstract. Several methods for computing the smallest eigenvalue of a symmetric positive

A Schur based algorithm for computing the smallest eigenvalue of a symmetric positive definite Toeplitz matrix

Split Algorithms and ZW-Factorization for Toeplitz and Toeplitz-plus-Hankel Matrices

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

NORTHERN ILLINOIS UNIVERSITY

Numerical Methods in Matrix Computations

Parallel Computation of the Eigenstructure of Toeplitz-plus-Hankel matrices on Multicomputers

A HARMONIC RESTARTED ARNOLDI ALGORITHM FOR CALCULATING EIGENVALUES AND DETERMINING MULTIPLICITY

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

An Arnoldi Method for Nonlinear Symmetric Eigenvalue Problems

Keeping σ fixed for several steps, iterating on µ and neglecting the remainder in the Lagrange interpolation one obtains. θ = λ j λ j 1 λ j σ, (2.

Arnoldi Methods in SLEPc

problem Au = u by constructing an orthonormal basis V k = [v 1 ; : : : ; v k ], at each k th iteration step, and then nding an approximation for the e

1. Introduction. In this paper we consider the large and sparse eigenvalue problem. Ax = λx (1.1) T (λ)x = 0 (1.2)

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Exponentials of Symmetric Matrices through Tridiagonal Reductions

Numerical solution of the inverse eigenvalue problem for real symmetric Toeplitz matrices

Numerical Methods I Eigenvalue Problems

APPLIED NUMERICAL LINEAR ALGEBRA

Department of. Computer Science. Functional Implementations of. Eigensolver. December 15, Colorado State University

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.

Introduction to Iterative Solvers of Linear Systems

The restarted QR-algorithm for eigenvalue computation of structured matrices

Iterative methods for symmetric eigenvalue problems

EECS 275 Matrix Computation

AN INVERSE EIGENVALUE PROBLEM AND AN ASSOCIATED APPROXIMATION PROBLEM FOR GENERALIZED K-CENTROHERMITIAN MATRICES

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems

Matrix Computations and Semiseparable Matrices

Solving the Inverse Toeplitz Eigenproblem Using ScaLAPACK and MPI *

A Jacobi Davidson Method for Nonlinear Eigenproblems

LARGE SPARSE EIGENVALUE PROBLEMS

The Reconstruction of An Hermitian Toeplitz Matrices with Prescribed Eigenpairs

Exponential Decomposition and Hankel Matrix

Toeplitz-circulant Preconditioners for Toeplitz Systems and Their Applications to Queueing Networks with Batch Arrivals Raymond H. Chan Wai-Ki Ching y

Computation of Smallest Eigenvalues using Spectral Schur Complements

OPTIMAL SCALING FOR P -NORMS AND COMPONENTWISE DISTANCE TO SINGULARITY

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems

Index. for generalized eigenvalue problem, butterfly form, 211

Algorithms to solve block Toeplitz systems and. least-squares problems by transforming to Cauchy-like. matrices

Approximate Inverse-Free Preconditioners for Toeplitz Matrices

Toeplitz matrices. Niranjan U N. May 12, NITK, Surathkal. Definition Toeplitz theory Computational aspects References

A Jacobi Davidson-type projection method for nonlinear eigenvalue problems

Exercise Set 7.2. Skills

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Universiteit-Utrecht. Department. of Mathematics. Jacobi-Davidson algorithms for various. eigenproblems. - A working document -

Inverse Eigenvalue Problems for Checkerboard Toeplitz Matrices

The Lanczos and conjugate gradient algorithms

Solving Regularized Total Least Squares Problems

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Diagonalizing Matrices

On the Eigenstructure of Hermitian Toeplitz Matrices with Prescribed Eigenpairs

EIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems

Reduced Synchronization Overhead on. December 3, Abstract. The standard formulation of the conjugate gradient algorithm involves

STABILITY OF INVARIANT SUBSPACES OF COMMUTING MATRICES We obtain some further results for pairs of commuting matrices. We show that a pair of commutin

Foundations of Matrix Analysis

Parallel Singular Value Decomposition. Jiaxing Tan

Math 108b: Notes on the Spectral Theorem

Lecture 10 - Eigenvalues problem

Interlacing Inequalities for Totally Nonnegative Matrices

A lower bound for the Laplacian eigenvalues of a graph proof of a conjecture by Guo

The Best Circulant Preconditioners for Hermitian Toeplitz Systems II: The Multiple-Zero Case Raymond H. Chan Michael K. Ng y Andy M. Yip z Abstract In

Structured Krylov Subspace Methods for Eigenproblems with Spectral Symmetries

arxiv: v1 [hep-lat] 2 May 2012

Matrix Algebra: Summary

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis

Spectra of Adjacency and Laplacian Matrices

Scientific Computing: An Introductory Survey

Block Bidiagonal Decomposition and Least Squares Problems

On the Ritz values of normal matrices

Efficient and Accurate Rectangular Window Subspace Tracking

Application of Lanczos and Schur vectors in structural dynamics

Bounding the Spectrum of Large Hermitian Matrices

Chapter 6. Algebraic eigenvalue problems Introduction Introduction 113. Das also war des Pudels Kern!

Eigen-solving via reduction to DPR1 matrices

0 Introduction Let the matrix A C nn be given by all its n entries The problem is to compute the product Ab of a matrix A by input vector b C n Using

ALGORITHMS FOR CENTROSYMMETRIC AND SKEW-CENTROSYMMETRIC MATRICES. Iyad T. Abu-Jeib

6. Iterative Methods for Linear Systems. The stepwise approach to the solution...

RITZ VALUE BOUNDS THAT EXPLOIT QUASI-SPARSITY

Comparing iterative methods to compute the overlap Dirac operator at nonzero chemical potential

Solving Large Nonlinear Sparse Systems

Positive Denite Matrix. Ya Yan Lu 1. Department of Mathematics. City University of Hong Kong. Kowloon, Hong Kong. Abstract

Total least squares. Gérard MEURANT. October, 2008

G1110 & 852G1 Numerical Linear Algebra

A JACOBI-DAVIDSON ITERATION METHOD FOR LINEAR EIGENVALUE PROBLEMS. GERARD L.G. SLEIJPEN y AND HENK A. VAN DER VORST y

1 Singular Value Decomposition and Principal Component

Majorization for Changes in Ritz Values and Canonical Angles Between Subspaces (Part I and Part II)

On group inverse of singular Toeplitz matrices

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

Approximating the matrix exponential of an advection-diffusion operator using the incomplete orthogonalization method

The parallel computation of the smallest eigenpair of an. acoustic problem with damping. Martin B. van Gijzen and Femke A. Raeven.

A Levinson algorithm based on an isometric transformation of Durbin`s

NORMS ON SPACE OF MATRICES

Krylov Subspaces. Lab 1. The Arnoldi Iteration

Transcription:

A Symmetry Exploiting Lanczos Method for Symmetric Toeplitz Matrices Heinrich Voss Technical University Hamburg{Harburg, Section of Mathematics, D{27 Hamburg, Federal Republic of Germany, e-mail: voss tu-harburg.de Dedicated to Richard S. Varga on the occasion of his 7th birthday Abstract Several methods for computing the smallest eigenvalues of a symmetric and positive denite Toeplitz matrix T have been studied in the literature. The most of them share the disadvantage that they do not reect symmetry properties of the corresponding eigenvector. In this note we present a Lanczos method which approximates simultaneously the odd and the even spectrum of T at the same cost as the classical Lanczos approach. Keywords: Toeplitz matrix, eigenvalue problem, Lanczos method, symmetry AMS-classication: 65F5 Introduction Several approaches have been reported in the literature for computing the smallest eigenvalue of a real symmetric, positive denite Toeplitz matrix. This problem is of considerable interest in signal processing. Given the covariance sequence of the observed data, Pisarenko [4] suggested a method which determines the sinusoidal frequencies from the eigenvector of the covariance matrix associated with its minimum eigenvalue. ybenko and Van Loan [3] presented an algorithm which is a combination of bisection and Newton's method for the secular equation. Replacing Newton's method by a root nding method based on rational Hermitian interpolation of the secular equation Mackens and the present author in [] and [2] improved this approach substantially. Hu and Kung [8] considered a safeguarded inverse iteration with Rayleigh quotient shifts, and Huckle [9], [] studied the spectral transformation method.

Trench [5] and Noor and Morgera [3] generalized the method of ybenko andvan Loan to the computation of the complete spectrum. A disadvantage of all of these approaches is that neither of them exploits symmetry properties of the eigenvectors. Let J n =( i n+;i ) i j= ::: n denote the (n n) ipmatrix with ones in its secondary diagonal and zeros elsewhere. A vector x 2 IR n is called symmetric ifx = J n x, and it is called skew-symmetric (or anti-symmetric) if x = ;J n x.itiswell known (cf. Andrew [], antoni and utler [2]) even for the larger class of symmetric and centrosymmetric matrices that every eigenspace of a matrix n in this class (i.e. c n+;i n+;j = c ij for every i j = ::: n) has a basis of vectors which are either symmetricorskew-symmetric, and that there exists a basis of IR n consisting of dn=2e symmetric andbn=2c skew-symmetric eigenvectors of n.we call an eigenvalue of n even and odd, respectively, if there exists a corresponding eigenvector which is symmetric and skew-symmetric, respectively. In [8] and [9] we improved the methods of [] and [2] for computing the smallest eigenvalue of a symmetric Toeplitz matrix T n using even and odd secular equations and hence exploiting the symmetry properties of the eigenvectors of T n. Andrew [], antoni and utler [2] and Trench [6] (the latter for symmetrictoeplitz matrices only) took advantage of symmetry properties to reduce the eigenvalue problem for n to two eigenproblems of half the dimension. For a symmetric Toeplitz matrix T n 2 IR n, T n =(t ji;jj ), their approach is as follows: If n =2m is even then T n may be rewritten as T n = T m ~ H m ~H T m T m! ~ Hm = t m t m+ ::: t n; t m; t m ::: t n;2... A : t t 2 ::: t m If x = J n x is a symmetric eigenvector of T n then it holds that x =[~x J m ~x], ~x := x( : m) 2 IR m,and T n x = T m ~x + ~ H m J m ~x ~H T m~x + T m J m ~x m! = (T m + ~ H m J m )~x ( ~ H T mj m + T m )J m ~x! = ~x J m ~x The second equation ( ~ H T m J m + T m )J m ~x = J m ~x of () is equivalent to (J m ~ H T m + J mt m J m )~x =(~ Hm J m + T m )~x = ~x! = x: () and therefore x is a symmetric eigenvalue of T n if and only if ~x = x( : m) isan eigenvector of (T m + ~ H m J m )~x = ~x: (2) Similarly x is a skew-symmetric eigenvector of T n if and only if ~x = x( : m) isan eigenvector of (T m ; ~ Hm J m )~x = ~x: (3) 2

Hence the eigenvalue problem T n x = x has been reduced to two eigenvalue problems of dimension m = n=2. If n = 2m + is odd then in an analogous way the eigenvalue problem can be reduced to two eigenproblems of dimensions m and m +. Andrew [] and antoni and utler [2] derived (2) and (3) and the corresponding equations for odd dimension n but they did not specify methods for the solution of these problems. Trench [6] applied the methods from [5] to problems (2) and (3) taking advantage of the fact that H m := ~ Hm J m is a Hankel matrix and employing fast algorithms for the solution of linear systems with Toeplitz-plus-Hankel matrices (cf. Heinig, Jankowski and Rost [7]). In this note we consider a symmetry exploiting variant of the Lanczos method. It is well known that the accuracy of an approximation to an eigenvalue obtained by the Lanczos method is inuenced mainly by the relative separation of from the other eigenvalues of T n. The bigger the separation is the better is the approximation. Therefore to compute the smallest eigenvalue of T n (or the lower part of the spectrum of T n ) it is often advantageous to apply the inverted Lanczos method, i.e. to determine the approximation from the projection of the eigenvalue problem T n ; x = ; x to the Krylov space K k (T n ; u) := span fu T n ; u ::: T n ;k+ ug (cf. [5], [9]). Obviously, if the initial vector u is symmetric or skew-symmetric then the whole Krylov space K k (T n ; u) belongs to the same symmetry class (and the same statement holds for K k (T n u)orthevector spaces constructed in a rational Krylov method or in Davidson's method. Hence, the following considerations hold for these methods, too). For the restriction of T n to the space of symmetric vectors and to skew-symmetric vectors the gaps between consecutive eigenvalues usually will be bigger than for the original matrix T n.thus, if the symmetry class of the principal eigenvector is known in advance then the computation of the smallest eigenvalue by the inverted Lanczos method can be accelerated by choosing the initial vector u in the appropriate symmetry class. However, the symmetry class of the principal eigenvector is known in advance only for a small class of Toeplitz matrices. The following result was given by Trench [7]: Theorem : Let T n =(t ji;jj ) i j= ::: n t j := Z F () cos(j) d j = 2 ::: n; where F : ( )! IR is nonincreasing and F (+) =: M>m:= F (;): Then for every n the matrix T n has n distinct eigenvalues in (m M), its even and odd spectra are interlaced, and its largest eigenvalue is even. In general the smallest odd and even eigenvalues of T n have to be determined and the smaller one of them is the requested principal eigenvalue of T n. In Section 2 we specify an algorithm which simultaneously executes the inverted Lanczos method for a symmetric and for a skew-symmetric initial vector. It is im- 3

portant to note that in every step only one linear system with system matrix T n has to be solved. Moreover, for symmetry reasons the level--operations in the Lanczos method have to be performed only for the upper half of the vectors. Therefore, if s 2 IR n is symmetric and a 2 IR n is skew-symmetric then the method produces orthonormal bases of the Krylov spaces K k (T n ; s)andk k (T n ; a) essentially at the same cost as the Lanczos method for T n and a general initial vector u. In Section 3 we report on extensive numerical examples which demonstrate the favourable behaviour of the method. 2 A symmetry exploiting Lanczos method Let T n := (t ji;jj ) i j= ::: n be a symmetric and positive denite Toeplitz matrix. The key to the simultaneous Lanczos methods with symmetric andskew-symmetric initial vectors is the following simple fact: If v 2 IR n solves the linear system T n v = w then the symmetric part v s := :5(v + J n v)ofv solves the linear system T n v s =:5(w + J n w)=:w s, and the skew-symmetric partv a := v ; v s solves the linear system T n v a = w ; w s. This observation leads to the following basic form of the Symmetry Exploiting Lanczos Iteration Let p = Jp 6=,q = ;Jq 6=begiven initial vectors. Set = = p = q = p = p =kp k q = q =kq k for k = 2 ::: until convergence do w = p k + q k solve T n v = w for v v s =:5 (v + J n v) k = vs p k v s = v s ; k p k ; k; p k; k = kv s k p k+ = v s = k end v a =:5 (v ; J n v) k = v a q k v a = v a ; k q k ; k; q k; k = kv a k q k+ = v a = k The iteration on the left is exactly the Lanczos method for T n with the symmetric initial vector p,andontherightwehave the Lanczos method with the skew-symmetric initial vector q. Selectiveoreven full re-orthogonalization could be incorporated on either side. However, since weareinterested in the smallest eigenvalue of T n only and since for the inverted Lanczos method this eigenvalue can be expected to converge rst re-orthogonalization is not necessary. This consideration is supported by the large number of examples that we considered and on which we report in Section 3. The symmetry exploiting Lanczos iteration simultaneously determines orthogonal 4

matrices P k := [p p 2 ::: p k ]andq k := [q q 2 ::: q k ]andtwo tridiagonal matrices S k :=... k; k; k A and A k :=... k; k; k where S k = Pk T T n ;P k and A k = Q T k T n ;Q k is the orthogonal projection of T n ; to the Krylov space K k (T n ; p )andk k (T n ; q ), respectively. Hence the eigenvalues of S k and A k are the Ritz values approximating the even and odd eigenvalues of T ; If s =: = s is the maximal eigenvalue of S k with corresponding normalized eigenvector y then it is well known (cf. [4]) that there exists an eigenvalue =: = of T n ; such that j s ; j j k y k j (4) i.e. j ; s j j s k y k j =: s : (5) Likewise if a =: = a is the maximal eigenvalue of A k and z denotes a normalized eigenvector corresponding to a then there exists an eigenvalue =: = of T n ; such that j ; a j j a k z k j =: a : (6) Hence, each of the intervals [ s ; s s + s ], s := s s =(; s ), and [ a ; a a + a ], a := a a =( ; a ), contains at least one eigenvalue of T n. In all of our examples := minf s ; s a ; a g was a lower bound of the minimum eigenvalue of T n. However, this is not necessarily the case depending on the minimum of the angles between the eigenvector of T n corresponding to the minimum eigenvalue and the initial vectors p and q.to be on the save side one may determine the Schur parameters of T n ; I n by Durbin's algorithm after the method terminated. Each of the vectors v s and p k on the right of the algorithm is symmetric and each of the vectors v a and q k is skew-symmetric. Hence, the level--operations in the body of the symmetry exploiting Lanczos iteration can be performed for the upper half of the vectors only, and the following procedure results. Let [ t], t 2 IR n;, be the rst column of the symmetric and positive denite Toeplitz matrix T n of even dimension n = 2m. The following MATLA function [la,k]=seli(t,m,tol) determines an approximation to the smallest eigenvalue of T n such that the relative error is smaller than tol. k is the dimension of the Krylov space that is needed to obtain this accuracy. AMATLA function v=solve_toeplitz(t,n,w) which yields the solution v of T n v = w is used. A n. 5

function [la,k]=seli(t,m,tol) p=rand(m,) q=rand(m,) p=p/sqrt(2*p'*p) q=q/sqrt(2*q'*q) err= k= while err > tol k=k+ w(:m)=q+p w(m+:n)=flipud(p-q) v=solve_toeplitz(t,n,w) vs=.5*(z(:m)+flipud(z(m+:n))) va=z(:m)-vs ms(k,k)=2*vs'*p ma(k,k)=2*va'*q vs=vs-ms(k,k)*p va=va-ma(k,k)*q if k> vs=vs-ms(k,k-)*p_old va=va-ma(k,k-)*q_old end [evs,las]=eig(ms) [eva,laa]=eig(ma) [las,is]=sort(diag(las)) [laa,ia]=sort(diag(laa)) ms(k,k+)=sqrt(2*vs'*vs) ma(k,k+)=sqrt(2*va'*va) ms(k+,k)=ms(k,k+) ma(k+,k)=ma(k,k+) if laa(k) < las(k) err=abs(ms(k+,k)*evs(k,is(k)))/las(k) else err=abs(ma(k+,k)*eva(k,ia(k)))/laa(k) end p_old=p q_old=q p=vs/ms(k+,k) q=va/ma(k+,k) end la=min(/las(k),/laa(k)) The modications for Toeplitz matrices of odd dimension n = 2m + are obvious. 3 Numerical onsiderations The most costly step in the algorithms of Section 2 is the solution of the linear system T n v = w: (7) (7) can be solved eciently in one of the following two ways. Durbin's algorithm for the corresponding Yule-Walker system supplies a decomposition LT n L T = D where L is a lower triangular matrix and D is a diagonal matrix. Hence, in every iteration step the solution of equation (7) requires 2n 2 ops. This method for solving system (7) is called Levinson-Durbin algorithm. For large dimensions n equation (7) can be solved using the Gohberg-Semencul formula for the inverse T ; n (cf. [6]) T ; n = ; y T t (GGT ; HH T ) (8) 6

Table. Average number of ops and iteration steps dim Lanczos method SELI ops steps ops steps 32 3:68E4 (:3E5) 7:53 2:78E4 (8:6E4) 5:73 64 :2E5 (4:39E5) 8: 8:E4 (3:39E5) 6:5 28 3:27E5 (9:39E5) 7:95 2:5E5 (7:4E5) 5:89 256 :28E6 (2:7E6) 8:35 9:67E5 (:62E6) 6:7 52 5:2E6 (5:E6) 8:73 3:83E6 (3:77E6) 6:6 24 2:6E7 (:25E7) 9:24 :58E7 (9:38E6) 6:45 where G := ::: y ::: y 2 y :::. y n; y n;2 y n;3 ::: A and H := ::: y n; ::: y n;2 y n; :::. y y 2 y 3 ::: A are Toeplitz matrices and y denotes the solution of the Yule-Walker system T n; y = t. The advantages associated with equation (8) are at hand. Firstly, the representation of the inverse of T n requires only n storage elements. Secondly, the matrices G, G T, H and H T are Toeplitz matrices, and hence the solution T n ; w can be calculated in only O(n log n) ops using the fast Fourier transform. Experiments show that for n 52 this approach is actually more ecient than the Levinson-Durbin algorithm. To test the improvement of the symmetry exploiting Lanczos iteration upon the ordinary Lanczos method we considered Toeplitz matrices T = nx j= j T 2j (9) where is chosen such that the diagonal of T is normalized to, j and j are uniformly distributed random numbers in the interval [ ] and T = (cos((i ; j))) i j= ::: n (cf. ybenko and Van Loan [3]). For each of the dimensions n =32 64 28 256 52 and 24 we considered test examples. Table contains the average number of ops and the average dimension k of the Krylov spaces needed to determine an approximation to the smallest eigenvalue with relative error less than ;6.Herewe solved the linear systems using the Levinson-Durbin algorithm. In parenthesis we added the average number of ops using the Gohberg-Semencul formula. 7

Huckle [9] suggested to use the error estimate corresponding to (4) after a small number of steps of the inverted Lanczos process to obtain a good lower bound of the smallest eigenvalue of T n and to continue with the spectral transformation Lanczos method with this shift, i.e. the Lanczos method for (T n ; I n ) ;. The example above demonstrates that after a very small number of Lanczos steps a good approximation of the smallest eigenvalue is derived. Hence, the variant of Huckle usually will not pay. 4 oncluding remarks In this note we presented a symmetry exploiting variant of the inverted Lanczos method to determine the smallest eigenvalue of a real symmetric and positive denite Toeplitz matrix which improves the classical Lanczos method considerably. In the numerical examples we solved Toeplitz systems by the Levinson-Durbin algorithm or by the Gohberg-Semencul formula. Superfast Toeplitz solvers can be incorporated as well reducing the complexity of the method to O(n(log n) 2 ). Acknowledgement The author wishes to thank William Trench for bringing some references to his attention. References [] A.L. Andrew, Eigenvectors of certain matrices. Lin. Alg. Appl. 7 (973), pp. 5 62 [2] A. antoni and P. utler, Eigenvalues and eigenvectors of symmetric centrosymmetric matrices. Lin. Alg. Appl. 3 (976), pp. 275 288 [3] G. ybenko and. Van Loan, omputing the minimum eigenvalue of a symmetric positive denite Toeplitz matrix. SIAM J. Sci. Stat. omput. 7 (986), pp. 23 3 [4] J.W. Demmel, Applied Numerical Linear Algebra. SIAM, Philadelphia 997 [5] T. Ericsson and A. Ruhe, The spectral transformation Lanczos method for the numerical solution of large sparse generalized symmetric eigenvalue problems. Math. omp. 35 (98), pp. 25 268 [6] I.. Gohberg and A.A. Semencul, On the inversion of nite Toeplitz matrices and their continuous analogs. Mat. Issled. 2 (972), pp. 2{233. [7] G. Heinig, P. Jankowski and K. Rost, Fast inversion algorithms of Toeplitz{ plus{hankel matrices. Numer. Math. 52 (988), pp. 665 682 [8] Y.H. Hu and S.{Y. Kung, Toeplitz eigensystem solver. IEEE Trans. Acoustics, Speech, Signal Processing 33 (985), pp. 264 27 8

[9] T. Huckle, omputing the minimum eigenvalue of a symmetric positive denite Toeplitz matrix with spectral transformation Lanczos method. In Numerical Treatment of Eigenvalue Problems, Vol. 5, J. Albrecht, L. ollatz, P. Hagedorn, W. Velte (eds.), irkhauser Verlag, asel 99, pp. 9 5 [] T. Huckle, irculant and skewcirculant matrices for solving Toeplitz matrices. SIAM J. Matr. Anal. Appl. 3 (992), pp. 767 777 [] W. Mackens and H. Voss, The minimum eigenvalue of a symmetric positive denite Toeplitz matrix and rational Hermitian interpolation. SIAM J. Matr. Anal. Appl. 8 (997), pp. 52 534 [2] W. Mackens and H. Voss, A projection method for computing the minimum eigenvalue of a symmetric positive denite Toeplitz matrix. Lin. Alg. Appl. 275 { 276 (998), pp. 4 { 45 [3] F. Noor and S.D. Morgera, Recursive and iterative algorithms for computing eigenvalues of Hermitian Toeplitz matrices. IEEE Trans. Signal Processing 4 (993), pp. 272 28 [4] V.F. Pisarenko, The retrieval of harmonics from a covariance function. Geophys. J. R. astr. Soc. 33 (973), pp. 347 366 [5] W.F. Trench, Numerical solution of the eigenvalue problem for Hermitian Toeplitz matrices. SIAM J. Matr. Anal. Appl. (989), pp. 35 46 [6] W.F. Trench, Numerical solution of the eigenvalue problem for eciently structured Hermitian matrices. Lin. Alg. Appl. 54-56 (99), pp. 45 432 [7] W.F. Trench, Interlacement of the even and odd spectra of real symmetric Toeplitz matrices. Lin. Alg. Appl. 95 (993), pp. 59 68 [8] H. Voss, Symmetric schemes for computing the minimal eigenvalue of a symmetric Toeplitz matrix. Lin. Alg. Appl. 287 (999), pp. 359 37 [9] H. Voss, ounds for the minimum eigenvalue of a symmetric Toeplitz matrix. Electr. Trans. Numer. Anal. 8 (999), pp. 27 38 9