Displacement structure approach to discrete-trigonometric-transform based preconditioners of G.Strang type and of T.Chan type

Similar documents
Algorithms to solve block Toeplitz systems and. least-squares problems by transforming to Cauchy-like. matrices

A Comrade-Matrix-Based Derivation of the Eight Versions of Fast Cosine and Sine Transforms

Foundations of Matrix Analysis

Lecture Notes for Inf-Mat 3350/4350, Tom Lyche

Algebra C Numerical Linear Algebra Sample Exam Problems

Math Linear Algebra II. 1. Inner Products and Norms

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

MTH Linear Algebra. Study Guide. Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Ranks of Hadamard Matrices and Equivalence of Sylvester Hadamard and Pseudo-Noise Matrices

6.4 Krylov Subspaces and Conjugate Gradients

Numerical Methods I Non-Square and Sparse Linear Systems

Numerical Methods in Matrix Computations

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

Numerical Methods - Numerical Linear Algebra

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

4.2. ORTHOGONALITY 161

Linear Algebra: Matrix Eigenvalue Problems

MAT Linear Algebra Collection of sample exams

Matrix Theory. A.Holst, V.Ufnarovski

The goal of this chapter is to study linear systems of ordinary differential equations: dt,..., dx ) T

NORMS ON SPACE OF MATRICES

Math 471 (Numerical methods) Chapter 3 (second half). System of equations

A Traub like algorithm for Hessenberg-quasiseparable-Vandermonde matrices of arbitrary order

Introduction. Chapter One

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Linear Algebra Massoud Malek

1. General Vector Spaces

8. Diagonalization.

Split Algorithms and ZW-Factorization for Toeplitz and Toeplitz-plus-Hankel Matrices

Preconditioners for ill conditioned (block) Toeplitz systems: facts a

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

arxiv: v5 [math.na] 16 Nov 2017

Midterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015

13-2 Text: 28-30; AB: 1.3.3, 3.2.3, 3.4.2, 3.5, 3.6.2; GvL Eigen2

Elementary linear algebra

INTEGER POWERS OF ANTI-BIDIAGONAL HANKEL MATRICES

MAT 2037 LINEAR ALGEBRA I web:

ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3

Stabilization and Acceleration of Algebraic Multigrid Method

Review of some mathematical tools

Math Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012.

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

Stat 159/259: Linear Algebra Notes

Algebra II. Paulius Drungilas and Jonas Jankauskas

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

Duke University, Department of Electrical and Computer Engineering Optimization for Scientists and Engineers c Alex Bronstein, 2014

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 9

Lecture4 INF-MAT : 5. Fast Direct Solution of Large Linear Systems

Linear Algebra March 16, 2019

ELEMENTARY LINEAR ALGEBRA

Equality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same.

Jim Lambers MAT 610 Summer Session Lecture 2 Notes

Classifications of recurrence relations via subclasses of (H, m) quasiseparable matrices

Linear Algebra. Session 12

Math113: Linear Algebra. Beifang Chen

In particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with

w T 1 w T 2. w T n 0 if i j 1 if i = j

On Spectral Factorization and Riccati Equations for Time-Varying Systems in Discrete Time

Fundamentals of Engineering Analysis (650163)

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Fundamentals of Linear Algebra. Marcel B. Finan Arkansas Tech University c All Rights Reserved

Numerical Methods I Solving Square Linear Systems: GEM and LU factorization

Mathematical Methods wk 2: Linear Operators

1 Linear Algebra Problems

256 Summary. D n f(x j ) = f j+n f j n 2n x. j n=1. α m n = 2( 1) n (m!) 2 (m n)!(m + n)!. PPW = 2π k x 2 N + 1. i=0?d i,j. N/2} N + 1-dim.

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Class notes: Approximation

L. Vandenberghe EE133A (Spring 2017) 3. Matrices. notation and terminology. matrix operations. linear and affine functions.

Math 307 Learning Goals. March 23, 2010

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

Numerical Analysis Lecture Notes

Tikhonov Regularization of Large Symmetric Problems

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

NOTES ON LINEAR ALGEBRA. 1. Determinants

October 25, 2013 INNER PRODUCT SPACES

Numerical Linear Algebra

HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION)

Chapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space.

Reconstruction of sparse Legendre and Gegenbauer expansions

arxiv: v1 [math.na] 5 May 2011

Vectors in Function Spaces

Supplementary Notes on Linear Algebra

Chapter 7 Iterative Techniques in Matrix Algebra

Lipschitz stability of canonical Jordan bases of H-selfadjoint matrices under structure-preserving perturbations

Dot Products. K. Behrend. April 3, Abstract A short review of some basic facts on the dot product. Projections. The spectral theorem.

Approximation theory

Orthogonalization and least squares methods

Applied Linear Algebra

CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3

Math 408 Advanced Linear Algebra

A matrix over a field F is a rectangular array of elements from F. The symbol

Linear algebra. S. Richard

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

MTH 2032 Semester II

SPRING 2006 PRELIMINARY EXAMINATION SOLUTIONS

Chapter 2 The Group U(1) and its Representations

Lecture 2: Linear Algebra Review

Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm

Transcription:

Displacement structure approach to discrete-trigonometric-transform based preconditioners of GStrang type and of TChan type Thomas Kailath Information System Laboratory, Stanford University, Stanford CA 9435 455, USA and Vadim Olshevsky Department of Mathematics, University of Connecticut 96 Auditorium Road, U-9, Storrs, CT 669 Email: olshevsky@mathuconnedu Web page: http://wwwmathuconnedu/ olshevsky Abstract In this paper we use a displacement structure approach to design a class of new preconditioners for the conjugate gradient method applied to the solution of large Toeplitz linear equations Explicit formulas are suggested for the Strang-type and for the TChan-type preconditioners belonging to any of eight classes of matrices diagonalized by the corresponding discrete cosine or sine transforms Under the standard Wiener class assumption the clustering property is established for all of these preconditioners, guaranteeing rapid convergence of the preconditioned conjugate gradient method All the computations related to the new preconditioners can be done in real arithmetic and to fully exploit this advantageous property one has to suggest a fast real-arithmetic algorithm for multiplication of a Toeplitz matrix by a vector It turns out that the obtained formulas for the Strang-type preconditioners allow a number of representations for Toeplitz matrices leading to a wide variety of real-arithmetic multiplication algorithms based on any of eight discrete cosine or sine transforms Recently, transformations of Toeplitz matrices to Vandermonde-like or Cauchy-like matrices have been found to be useful in developing accurate direct methods for Toeplitz linear equations In this paper we suggest to further extend the range of the transformation approach by exploring it for iterative methods; this technique allowed us to reduce the complexity of each iteration of the preconditioned conjugate gradient method We conclude the paper with a suggestion on how to exploit the displacement structure to efficiently organize numerical experiments in a numerically reliable way Introduction PCGM for Toeplitz linear equations We consider the solution of a large linear system of equations A m x = b whose coefficient matrix A m is an m m leading submatrix of a semi-infinite real symmetric This work was supported in part by the NSF Awards CCR 98 and CCR 458

Toeplitz matrix of the form [ A = a a a a 3 a a a a ] a i j = a a a a, () a 3 a a a usually associated with the corresponding generating function a(z) = k= a k z k Gaussian elimination ignores any special structure, thus requiring O(m 3 ) arithmetic operations to solve A m x = b There is a number of fast Toeplitz solvers all taking advantage of the structure () to significantly reduce the number of operations For example, the classical Schur and Levinson algorithms, see, eg, [K87], and references therein, each require only O(m ) operations per system Moreover, there are even super-fast Toeplitz solvers with a smaller complexity of O(m log m) operations The numerical stability of such direct Toeplitz solvers is discussed in a number of recent papers, see, eg, [O3] and the references therein They reveal that although fast algorithms were believed to be unstable by their very nature, there are methods to obtain a simultaneously fast and accurate solution Along with considerable current efforts to develop and to stabilize direct methods, the preconditioned conjugate gradient method (PCGM) for solving Toeplitz linear systems has garnered much attention This is a well-known iterative procedure which computes at each iteration step two inner products of length m and one multiplication of the coefficient matrix by a vector, thus requiring O(m log m) operations per iteration The number of iterations depends upon the clustering of the spectrum of the A m, and if the latter is clustered around having a small number of outliers, then PCGM will converge rapidly, see, eg, [GL89] Classical results on the eigenvalue distribution of Toeplitz matrices (see, eg, [GS84]) indicate that we cannot expect, in general, any clustering, and the convergence of the method will be slow This disadvantage motivated GStrang to propose the use of a certain circulant matrix P to reduce the number of iterations The idea was to apply the algorithm to a preconditioned system P Ax = P b () where the preconditioner P should satisfy the following three requirements Property The complexity of the construction of P should be small, not exceeding O(m log m) operations Property A linear system with P should be solved in O(m log m) operations Property 3 The spectrum of P A m should be clustered around, more precisely the following holds : For any ε> there exist integers N and s such that for any m>n, at most s eigenvalues of P A lie outside the interval [ ε, +ε] Summarizing, if a preconditioner satisfying the above properties -3 can be constructed, then the complexity of the PCGM will be reduced to only O(m log m) operations, which will be even less than the complexity of superfast direct methods The first (now well-known) proposed preconditioners of GStrang [S86] and of TChan [C88] were circulant matrices, defined respectively by S(A m )= circ (a,a,a,,a,a ),

C(A m )= circ (a, m m a + m a m, m m a + m a m,, m a m + m m a ) Here circ (r) denotes a circulant matrix specified by its first row r For these two preconditioners the first property holds by their construction, and since circulant matrices are diagonalized by the discrete Fourier transform (DFT) matrix F, the second property is also immediately satisfied Moreover, for the case when the generating function a(z) = k= a k z k is a function from the Wiener class, positive on the unit circle, the 3rd property for the GStrang and TChan preconditioners was established in [C89], [CS89] and in [CY9], resp A recent survey [CN96] gives a fairly comprehensive review of these and related results, and describes many other preconditioners, including those of RChan, ETyrtyshnikov, TKu and CKuo, THuckle, and others (A thorough theoretical and numerical comparison of all different preconditioners is one of the directions of current research, indicating that the question of which preconditioner is better may have different answers depending upon the particular classes of Toeplitz systems, and their generating functions, see, eg, [TS96], [T95], [CN96]) Along with many favorable properties of circulant preconditioners, they unfortunately require complex arithmetic (for computing FFT s), even for real symmetric Toeplitz matrices To overcome this disadvantage, DBini and FDi Benedetto [BB9] proposed non-circulant analogs of the GStrang and of T Chan preconditioners, belonging to the so-called τ-class (introduced in [BC83] as the class of all matrices diagonalized by the (real) discrete sine I transform (DST-I) matrix) DBini and FDi Benedetto established for their preconditioners the properties -3 under the Wiener class assumption In this paper we continue the work started in [S86], [C88], [BB9], and give a systematic account of Strang-type and of TChan-type preconditioners belonging to the classes of matrices diagonalized by other real trigonometric transforms We consider four discrete cosine and four discrete sine transforms, and refer to a recent paper [OOW3, O4] for a systematic matrix presentation of the fast algorithms for the all eight transforms For each of these eight cases we derive explicit formulas for the Strang-type and the TChan-type preconditioners and establish for them the above properties -3 (under the standard Wiener class assumption) This problem could perhaps be solved directly, but we have found that an interpretation in terms of displacement structure [KKM79] and of partially reconstructible matrices [KO95a] often allows us to simplify many arguments We believe that the displacement structure approach (systematically exposed in this contribution) will be useful in addressing other problems related to preconditioning, and a recent work [CNP94], [H95] supports this anticipation Displacement structure approach We next use the results of [KO95a] to briefly give an interpretation of the classical GStrang and TChan circulant preconditioners in terms of partially reconstructible matrices This technique will be further extended in the main text below The displacement structure approach initiated by [KKM79] is based on introducing in a linear space of all m m matrices a suitable displacement operator ( ) :R m m R m m of the form (R) =R FRF T, or (R) =F T R RF (3) A matrix R is said to have -displacement structure, if it is mapped to a low-rank matrix (R) Since a low-rank matrix can be described by a small number of parameters, a representation of a matrix by its image (R) often leads to interesting results, and is useful for the design of many fast algorithms This approach has been found to be useful for studying many different patterns of structure (for example, Toeplitz T =[T i j ], Vandermonde V =[x j i ], Cauchy C =[ x i y j ], etc) by specifying for each of them an appropriate displacement operator For example, Toeplitz-like matrices are defined as having displacement structure with respect to the choice Z (R) =R Z RZ T, (4) 3

where Z = circ (,,, ) The motivation for the above definition can be inferred from the easily verified fact that the constant-diagonal structure of a Toeplitz matrix A in() implies that the rank of the matrix???? Z (A) =? does not exceed two If the rank of Z (A) is bigger than two but still sufficiently small, A is called Toeplitz-like Although the latter definition of Toeplitz-like matrices has already used by several authors, it is slightly different from the standard one, Z (R) =R Z RZ T where Z is the lower shift matrix The crucial difference is that Z clearly has a nontrivial kernel, so the image Z (R) no longer contains all the information on R Such matrices R have been called partially reconstructible in [KO95a], and systematically studied there In the Toeplitz-like case Ker Z coincides with the subspace of all circulant matrices in R m m, so we can observe that the GStrang and TChan preconditioners are both chosen from Ker Z 3 A proposal : HQ -kernel preconditioner The above displacement operator Z is not the only one associated with the class of Toeplitz matrices We propose to apply the above interpretation, and develop the analogs of GStrang and TChan preconditioners in the kernels of several other related displacement operators of the form HQ (R) =HQR T RH Q (5) Moreover, we shall specify eight matrices H Q for which the kernel of the corresponding displacement operator (5) coincides with the subspace of matrices diagonalized by any one of the eight known versions of discrete cosine/sine transforms For each of these cases we write down the formulas for the corresponding Strang-type preconditioner and TChan-type preconditioner Under the standard Wiener class assumption we establish for these new preconditioners the properties -3 4 Fast real-arithmetic multiplication of a Toeplitz matrix by a vector As was mentioned above, each iteration of the PCGM involves a multiplication of the coefficient matrix A and a preconditioner P by vectors All the computations related to the new preconditioners can be done in real arithmetic However, the standard technique for the multiplication of a Toeplitz matrix by a vector is based on the FFT, thus requiring complex arithmetic We show that in each of the considered cases the new formulas for the Strang-type preconditioners allow us an embedding of an m m matrix A into a larger m m matrix, which is diagonalized by the corresponding (real) discrete cosine/sine transform matrix This observation allows us to suggest a variety of new O(m log m) real-arithmetic algorithms for the multiplication of a Toeplitz matrix by a vector, using any of eight versions of discrete cosine or sine transforms For the DST-I case such an algorithm was suggested earlier in [BK95] 5 Transformations and further reduction of the cost of one iteration Toeplitz-like matrices display just one kind of displacement structure, see (4) A different displacement operator (R) =D x R RD y (with diagonal D x =diag(x,,x n ) and D y =diag(y,,y n )) has been used to define the class of and Cauchy-like matrices Clearly, for a Cauchy matrix C =[ x i y j ] the rank of (C) =D x C CD x =[]is one Hence, if the latter rank is bigger than one but small, the matrix is called Cauchy-like It was observed in many places (cf with [P9] [GO94a], [He95a] among others) that matrices with displacement structure can be transformed from one class to another In particular (cf [GO94a]) the fact 4

that Z = F DF where F is the normalized DFT matrix, allows us to transform a Toeplitz-like matrix R into a Cauchy-like matrix FAF Since Cauchy-like matrices allow introducing pivoting into fast Gaussian elimination algorithms (cf [GO94b], Alg 7 (partial pivoting) and Alg 6 (symmetric pivoting)), this idea has been found to be useful to numerically reliable direct methods for solving Toeplitz linear equations, see, eg, [He95a]; for the first accurate algorithms of this kind see [GKO95], [KO95a], as well as [KO95b], [KO94], [BKO94], [SB95], [Gu95], [Gu96] In this paper we suggest to exploit the transformation technique for iterative methods, and to replace a preconditioned system (P A)x = P b by a transformed system (FP F )(FAF )(Fx) =(FP )b (6) Two advantages of the latter equation are in that (a) the transformed preconditioner FP F is a diagonal matrix, so the cost of computing a matrix vector product for it is just linear; (b) it can be shown that a Cauchy-like matrix FAF ) can be multiplied by a vector with exactly the same complexity as for the initial Toeplitz matrix A Hence (6) allows us to use only four FFT s at each iteration (cf with [H94]) In this paper we propose the exact counterpart (T P T T )(T AT T )(T x) =(T P )b (7) of such a technique for all eight discrete cosine/sine transforms T Again, in all sixteen considered cases (a) each of the transformed preconditioners T P T is a diagonal matrix, and (b) a Cauchy-like matrix T AT T can be multiplied by a vector with exactly the same complexity as for the initial Toeplitz matrix A, ie, four cosine of sine transforms So, all preconditioners, while dramatically reducing the number of iterations, do not increase the cost of a single iteration Partially reconstructible matrices We shall address the problem of constructing discrete-transform based preconditioners in the second part of the paper, and start here with necessary definitions and related facts on displacement structure and partially reconstructible matrices Let us consider a displacement operator and recall the following standard definitions {F,A} (R) =F R R A () A number α = rank {F,A} (R) is called the {F,A} -displacement rank of R (A matrix R is said to have a {F,A} -displacement structure if α is small compared to the size) A pair of rectangular n α matrices {G, B} in any possible factorization is called a {F,A} -generator of R; {F,A} (R) =F R R A = G B T, () If the matrices F and A have no common eigenvalues, {F,A} is invertible, so its generator contains a complete information on R For our purposes in this paper it will be necessary to consider another case where the displacement operator F (R) =F T R R F (3) clearly has a nontrivial kernel Such R have been called partially reconstructible in [KO95a], because now only part of the information on R is contained in {G, B} Following [KO95a] we shall refer to a triple {G, J, R K } as a F -generator of R, where the latter three matrices are defined as follows 5

Since F (R T )= F (R), we can write F (R) =F T R R F = G J G T, with J T = J R α α (4) Further, let us decompose R = R K + R K with respect to R n n = K K (5) where K = Ker F and the orthogonality in R n n is defined using the inner product <A,B>=tr(B A), A,B R n n, (6) with tr(a) denoting the sum of all diagonal entries of A, or, equivalently, the sum of eigenvalues of A Note that the latter inner product induces the Frobenius norm in R n n Clearly, now all the information on R is contained in the newly defined generator, {G, J, R K } 3 Polynomial Hankel-like matrices 3 Polynomial Hankel-like matrices In this paper we exploit a special displacement operator with the upper Hessenberg matrix HQ (R) =H T Q R R H Q = GJG T, (3) H Q = a a a,n a a a,n a a,n a n,n a n,n (3) The latter has been called in [MB79] a confederate matrix of the associated system of polynomials Q = {Q (x),q (x),, Q n (x)} defined by x Q k (x) =a k,k Q k (x)+a k,k Q k (x)+ + a,k Q (x) (33) We shall refer to matrices having low HQ -displacement rank as polynomial Hankel-like matrices, anexplanation for this nomenclature will be offered in Sec 34 after presenting the following example 3 Example Classical Hankel and Hankel-like matrices For the simplest polynomial system P = {,x,x,,x n,q n (x)}, its confederate matrix trivially reduces to the companion matrix H P = q q n q q n q n q n of the Q n (x) =q n x n + [ + q x + q Now, it is immediate to check that the shift-invariance-property of a Hankel matrix, R = implies that h i+j ] i,j n, [ HP (R) =HP T R RH P = e n g ] [ ] [ e n g ] T, (34) 6

where e n is the last coordinate vector and g = h n h n + H q n q q n q n Briefly, the HP -displacement rank of an arbitrary Hankel matrix does not exceed two Hence matrices with small HP -displacement rank (not just α ) are referred to as Hankel-like matrices 33 Diagonalization of confederate matrices To explain the name polynomial Hankel-like matrices we shall need the following result, which will be widely used in what follows It can be easily checked by direct multiplication (cf with [MB79]) that the confederate matrix is diagonalized by the polynomial Vandermonde matrix V Q, H Q = VQ D xv Q, (35) where Q (x ) Q (x ) Q n (x ) Q (x ) Q (x ) Q n (x ) V Q = D x =diag(x,x,, x n ) (36) Q (x n ) Q (x n ) Q n (x n ) Here {x k } are the zeros of Q n (x), which in our application here always will be n distinct numbers, so we shall impose this restriction throughout the paper 34 Change of basis Since H P = VP D xv P (see, eg, (35)), we have H Q = SPQ H P S PQ, with S PQ = VP V Q (37) Now, using (37) and (34) one sees that the HQ -displacement rank of SPQ T RS PQ (with Hankel R) also does not exceed We refer to such SPQ T RS PQ as polynomial Hankel matrices (or Hankel matrices represented in the polynomial basis Q), because the similarity matrix S PQ = can be easily [ shown s i,j ] i,j n to be an upper triangular matrix with the entries being the coefficients of Q k (x) = k i= s i+,k+ x k Therefore the more general matrices having low HQ -displacement rank (not just α ) are called polynomial Hankel-like matrices Since (3) has a nontrivial kernel, such R are partially reconstructible Our next goal is to describe the kernel of HQ 4 Transformation to Cauchy-like matrices and the kernel of HQ Recently, transformation of structured matrices from one class to another has been found to be useful to design for them many efficient algorithms In this paper we exploit an approach of [KO95a] for transformation of partially reconstructible matrices to transform polynomial Hankel-like matrices into Cauchy-like matrices, defined as having low displacement rank with respect to the simplest displacement operator Dx (R) =D x R RD x (4) with a diagonal matrix D x In fact, (35) immediately implies the following statement Proposition 4 Let R be a polynomial Hankel-like matrix in (3), given by its HQ -generator {G, J, R K }, and W Q denotes an arbitrary invertible diagonal matrix Then WQ T V Q T RVQ W Q is a Cauchy-like matrix with a Dx -generator {W T Q V T Q G, J, W T Q V T Q R KV Q W Q } (4) 7

Since the kernel of Dx is easy to describe (it is the subspace of all diagonal matrices), the above proposition implies the next statement Proposition 4 Let H Q,V Q and D x be defined by (3) and (36), where we assume that the diagonal entries of D x are m different numbers The kernel of HQ ( ) in (3) has the form K =span{(h T Q) k (V T Q W QV Q ), k =,,n } (43) where W Q is an arbitrary invertible diagonal matrix Finally, by replacing in (43) powers (HQ T )k by Q k (HQ T ), and using (4) we obtain the following statement Corollary 43 A matrix R K= Ker HQ given by can be diagonalized as follows : W T Q R = V T Q n k= RVQ r k Q k (H T Q) (V T Q W QV Q ), W Q = r(x ) r(xn) where the diagonal entries are computed via a polynomial Vandermonde transform r(x ) r r(x n ) = V Q r n Here we may note that the idea of displacement is to replace operations on the n entries of an n n structured matrix by manipulation on a smaller number O(n) of parameters The results of Sec 3 and 4 are based on the displacement equation (3), which describes R by the entries of H Q and {G, J, R K } In the general situation matrix H Q itself involves O(n ) parameters, so such a representation is no longer efficient In the next section we specialize the results of Sec 3 and 4, and list eight cases for which the above approach is beneficial 5 Orthonormal polynomials and discrete trigonometric transforms 5 Orthonormal polynomials Examination of the propositions in the previous section indicate that the kernel of HQ will have the simplest form in the case when there is a diagonal matrix W Q such that the matrix T Q = W Q V Q is orthonormal, see, eg, (43) It is easy to see that the latter condition is satisfied when the polynomials in {Q k (x)} are orthonormal with respect to the discrete inner product <p(x),q(x) >= n p(x k )q(x k )wk, k= where the nodes {x k } are the zeros of Q n (x), and the weights w k are diagonal entries of W Q Moreover, in this case the polynomials {Q k (x)} satisfy three-term recurrence relations so their confederate matrix reduces to the corresponding Jacobi (ie, symmetric tridiagonal) matrix 8

5 Discrete cosine and sine transforms Recall that our aim in this paper is to construct preconditioners diagonalized by discrete cosine or sine transform matrices, formally defined in the Table, where { k =,N η k = otherwise Table Discrete trigonometric transforms Discrete transform Inverse transform [ ] DCT-I CN I = N η k η N k η j η N j cos kjπ N N [C I k,j= N ] =[CN I ]T = CN I [ ] N DCT-II CN II = N η k cos k(j+)π [C II N k,j= N ] =[CN II]T = CN III [ ] N DCT-III CN III = N η j cos (k+)jπ [C III N k,j= N ] =[CN III]T = CN II [ ] N DCT-IV CN IV = N cos (k+)(j+)π [C IV 4N k,j= N ] =[CN IV ]T = CN IV [ ] DST-I SN I = N+ sin kj N N+ π [S I k,j= N ] =[SN I ]T = SN I [ ] N DST-II SN II = N η k sin k(j ) N π [S II k,j= N ] =[SN II]T = SN III [ ] N DST-III SN III = N η j sin (k )j N π [S III k,j= N ] =[SN III]T = SN II [ ] N DST-IV SN IV = N sin (k )(j ) 4N π [SN IV ] =[SN IV ]T = SN IV k,j= Note that we use a slightly different definition for the DCT-I, ensuring that now all the discrete transform matrices in Table are orthogonal The modified DCT-I can be transformed into the regular one by just appropriate scaling Now, using the fact that the Chebyshev polynomials of the first and second kind, sin((k + ) arccos x) T k (x) =cos(karccos x), U k (x) = (5) sin(arccos x) are essentially cosines and sines, we obtain that all the discrete transform matrices T Q in Table can be seen as the orthogonal matrices W Q V Q defined by orthonormal (Chebyshev-like) polynomials {Q k (x)} k= n } specified in Table, and the weight matrices W Q specified in Tables 3 We therefore adopt the designation T Q = W Q V Q for all eight transform matrices in Table by associated them with the corresponding polynomial systems Q Table First n polynomials {Q, Q, Q n Q n } DCT-I { T, T,, T n, T n } DCT-II {U, U U,, U n U n } DCT-III { T, T,, T n } DCT-IV {U, U U,, U n U n } DST-I {U, U,, U n } DST-II {U, U + U,, U n + U n } DST-III {U, U,, U n, U n } DST-IV {U, U + U,, U n + U n } 9

For each of the eight systems {Q k (x)} n k= the above Table lists the first n polynomials To specify V Q we have to also define the nodes {x k } n k=, or, equivalently, the last polynomial Q n(x), which is done in the second part of Table Table Continuation The last polynomial Q n (x) Q n zeros of Q n DCT-I xt n T n {cos( kπ N )}N DCT-II U n U n + U n {cos( kπ N )}N DCT-III T n {cos( (k+)π N )}N ) DCT-IV T n {cos( (k+)π N )}N ) DST-I U n {cos( kπ N+ )}N DST-II U n +U n + U n {cos( kπ N )}N DST-III T n {cos( (k )π N )}N DST-IV T n cos( (k )π N )}N Table 3 DCT-I CN I = W Q V Q with W Q = N diag(,,,, ) DCT-II CN II = W Q V Q with W Q = N diag(, cos( π (N )π N ),,cos( N )) DCT-III CN III = W Q V Q with W Q = N I DCT-IV CN IV = W Q V Q with W Q = N diag(cos( π π (N )π 4N ), cos(3 4N ),cos( DST-I SN I = W Q V Q with W Q = N+ diag(sin( π Nπ N+ ),,sin( N+ )) DST-II DST-III DST-IV S II N = W Q V Q with W Q = S III N = W Q V Q with W Q = S IV N = W Q V Q with W Q = N diag(sin( π N N diag(sin( π N 4N )) (N )π ),,sin( N ), sin( π )) ), sin( 3π N ),, sin( (N )π N )) N diag(sin( π 3π (N )π 4N ), sin( 4N ),,sin( 4N )) Finally, we specify the corresponding confederate matrices H Q DCT-I DCT-II DCT-III DCT-IV Table 4 H Q = tridiag H Q = tridiag H Q = tridiag H Q = tridiag

DST-I DST-II DST-III DST-IV Table 4 Continuation H Q = tridiag H Q = tridiag H Q = tridiag H Q = tridiag All the proofs are straightforward and based on the well-known recurrence relations T (x) =, T = xt (x), T k (x) =xt k (x) T k (x) U (x) =, U =xu (x), U k (x) =xu k (x) U k (x) In the second part of the paper we shall use the eight displacement operators of the form HQ (R) =H Q R RH Q with eight Jacobi matrices listed in Table 4 to design real-arithmetic discrete cosine/sine transform based preconditioners The expressions for the preconditioners will be obtained by using certain auxiliary formulas which it is convenient to present in the next section 6 Strang-type preconditioners 6 Definition of Strang-type preconditioners Now we apply the technique developed in the first part of the paper to design a family of preconditioners for Toeplitz matrices Consider a generating function of the form a(x) = a k z k, a k = a k R, k= which we assume (a) to be from the Wiener class, ie, a k, k= and (b) to have positive values on the unit circle, f(z) > z =

As is well-known, these two conditions guarantee that all leading submatrices A m (m =,,) of the associated infinite real symmetric Toeplitz matrix a a a a 3 a a a a [ ] A = a i j = a a a a a 3 a a a are positive definite : A m > Our next goal is to construct for A m a good preconditioner S Q (A m ) from the class K Q = Ker HQ, where H Q is one of the eight matrices in Table 4 The preconditioners are defined by S Q (A) = where the coefficients {r k } are listed in Table 5 m j= r j Q j (H Q ), (6) Table 5 Definition of the Strang-type preconditioners The coefficients of the decomposition (6) of S Q (A) r r r r m 3 r m r m DCT-I a a a a m 3 a m a m DCT-II a + a a + a a + a 3 a m 3 + a m a m + a m a m DCT-III a a a a m 3 a m a m DCT-IV a + a a + a a + a 3 a m 3 + a m a m + a m a m DST-I a a a a 3 a a 4 a m 3 a m a m a m DST-II a a a a a a 3 a m 3 a m a m a m DST-III a a a a 3 a a 4 a m 3 a m a m a m am DST-IV a a a a a a 3 a m 3 a m a m a m a m Note that the DST-I-based preconditioner was designed earlier in [BB9] and it was also discussed in [BK95] [H95] 6 Properties of the Strang-type preconditioners The motivation behind the definitions of Table 5 is that they imply several properties (shared with the classical Strang preconditioner) that are listed next Property 4 For any ε> there exist M > so that for m>mthe spectrum of S Q (A m ) lies in the interval [min z = a(z) ε, max z = a(z)+ε] Property 5 All Strang-type preconditioners S Q (A) defined in Table 5 are positive definite matrices for sufficiently large m Property 6 For all Strang-type preconditioners in Table 5 we have that S Q (A) and S Q (A) are uniformly bounded independently of m Properties 5 and 6 are easily deduced from the property 4, so we need to establish only the latter In order to prove that the Strang-type preconditioners all have the property 4 we need to specify corollary 43 for each of the eight cases, and to obtain the description of the spectrum of S Q (A m ) This is done in the next statement

Corollary 6 Let H Q be one of the matrices in Table 4, K Q = Ker HQ, and let R = n k= be a decomposition of R K Q with respect to a basis r k Q k (H Q ) (6) {Q k (H Q ),k =,,, n } (63) in K Q Then T Q RTQ T =diag(λ,,λ n ), (64) where T Q is the corresponding discrete transform from Table, and λ r = WQ Q λ n r n, (65) where W Q is the corresponding weight matrix from Table 3 Specifying the latter corollary to each of the eight cases, we obtain that the eigenvalues of S Q (A m ) will have the form shown in Table 6 DCT-I DCT-II DCT-III DCT-IV DST-I DST-II DST-III DST-IV Table 6 Eigenvalues {λ,,λ m } of S Q (A) λ k = a + m j= a j cos (k )jπ m = a m (z k ) where z k = e k m πi λ k = a + m j= a j cos (k )jπ m = a m (z k ) where z k = e k m πi λ k = a + m j= a j cos (k )jπ m = a m (z k ) where z k = e k m πi λ k = a + m j= a j cos (k )jπ m = a m (z k ) where z k = e k m πi λ k = a + m j= a j cos kjπ m+ = a m (z k ) where z k = e k m+ πi λ k = a + m j= a j cos kjπ m = a m (z k ) where z k = e k m πi λ k = a + m j= a j cos (k )jπ λ k = a + m j= a j cos (k )jπ m = a m (z k ) where z k = e k m m = a m (z k ) where z k = e k m πi πi (In Table 6 we list expressions for all eight cases, because we shall use them in our arguments below) Thus, the eigenvalues {λ k } of S Q (A) are the values of a truncated function a m (z) = m+ k= (m+) at certain points on the unit circle, specified in Table 6 Since a(x) is in the Wiener class, a m (x) is its approximation, so the property 4 holds true, and further, it implies the properties 5 and 6 63 Terminology: Strang-type preconditioners Even though the formula for the classical Strang preconditioner S(A m )=a I +a Z +a Z ++a Z n +a Z n, where Z = circ (,,, ) (66) looks like (6) with H Q replaced by Z and with O j (x) replaced by x j, but it is not immediately clear how strong is the analogy Moreover, (66) has a wrap-around property of {a k } that is missing in the definition given in Table 5 a k z k 3

However, the wrap-around is a not a crucial property of the original Strang preconditioner It is needed only to cope with the non-symmetry of Z and to make the matrix S(A m ) symmetric The matrices H Q are already symmetric, and the wrap-around is not needed in Table 5 However, S Q (A) and S(A) share a number of crucial properties, eg, the trick with computing the spectrum via results like in the corollary 6, deriving the properties 4,5,6, and establishing the equivalence to the Chan-type preconditioners We think these analogies justify the name Strang-type for the new preconditioners Summarizing, in this section we presented explicit formulas for Strang-type preconditioners, and proved for them the properties 4-6 Moreover, the properties - stated in the Introduction are also trivially satisfied It remains only to establish the property 3, crucial for the rapid convergence of the PCGM This property will be proved in Sec 8 below, using another description of the new preconditioners given next 7 The new preconditioners are Toeplitz-plus-Hankel-like matrices 7 The classical GStrang (Toeplitz-plus-Toeplitz) preconditioner We called new preconditioners as GStrang-like preconditioners, a justification for this nomenclature is offered next In [S86] GStrang proposed a circulant preconditioner, a a a a a a a a a a S(A) = a a a a a a a a a a a a a [ obtained by copying first [m/] diagonals of A = a i j ] In fact, this preconditioner can be seen as a Toeplitz-plus-Toeplitz matrix, S(A) =A + T, [ where A is the given Toeplitz matrix, and the first column of a second Toeplitz term T is given by a a m In fact, many favorable properties of S(A) can be explained by the fact that the entries of the central diagonals of A now occupy corner positions in T In case when the generating function a(z) = k= a k z k is from the Wiener class, only first few coefficients are large, implying that T = A lr + A sn is a sum of a low rank and a small norm matrices, a property implying the usefulness of S(A) =A + A lr + A sn (7) as a preconditioner for A It turns out that all eight Strang-type preconditioners S Q (A), considered above are Toeplitz-plus-Hankel-like matrices (formally defined below), a fact allowing us to use the above lowrank-small-norm-perturbation argument to prove the favorable properties of S Q (A) as preconditioners for A 7 Toeplitz-plus-Hankel-like matrices Recall that matrices R with a low HQ -displacement rank have been called polynomial Hankel matrices Since all eight matrices H Q in Table 4 correspond to the Chebyshev-like polynomial systems listed in Table, we could refer to such R as Chebyshev-Hankel matrices It can be checked, however, that [ if H Q ] is defined as, for example, [ in the ] line DST-I of Table 4, then for any sum T + H of a Toeplitz T = t i j and a Hankel H = h i+j matrices we have HQ (T + H) = (7) 4

( t t m t t m t m t t m t + h h m 3 h m h m h h m+ ) h m 3 h m h m h m h m+ h m In our terminology, the HQ -displacement rank of T + H does not exceed 4 This fact was observed and used in [HJR88] and [GK89] to develop fast algorithms for inversion of Toeplitz-plus-Hankel matrices In [GKO95] we introduced (and suggested, for the first time, fast algorithms for) the more general class of Toeplitz-plus-Hankel-like matrices, defined as having low (not just α 4) HQ -displacement rank Clearly (cf with [GKO95]), the other choices for H Q in Table 4 can be used to define the same class of Toeplitzplus-Hankel-like matrices (the actual displacement rank may vary, depending upon a particular H Q, but it remains low) Summarizing, there are two nomenclatures (ie Chebyshev Hankel-like and Toeplitz-plus- Hankel-like matrices) for the same class of structured matrices 73 Toeplitz-plus-Hankel-like representations for S Q (A) Since all the preconditioners S Q (A) in Table 5 belong to the kernel of the corresponding HQ ( ), they clearly belong to the above class of Toeplitzplus-Hankel-like matrices In fact, each of them can even be represented as S Q (A) =A + H + B, (73) where A is the given Toeplitz matrix, H is a certain Hankel matrix, and B is a certain border matrix, having nonzero entries only in its first and last rows and columns DCT-I DCT-II DCT-III DCT-IV Table 7 A Hankel part and a border part of S Q (A) H B a a a m a m ( a a m ) a m a a a a m a m a a a m a m a a a m a m a a a a a m a a 3 a m a m a 3 a a m a a a a a m a m a a a m a m a m a m a a a a m a a 3 a m a m a 3 a a m a a ( ) a m a m a a a a a m a m a a m a m A reader should be warned that such a specific representation is not valid for arbitrary Toeplitz-plus-Hankel-like matrices 5

DST-I Table 7 Continuation H H a a m a m a m B B DST-II a m a a a a m a a 3 a m a m a 3 a DST-III DST-IV a m a a a a m a m a m a m a a a m a m a a a a a m a a 3 a m a m a 3 a a m a a + a m a m a a m a m a ( + )a The proof of the fact that S Q (A) has the form (73) where A is the given Toeplitz matrix, and H, B are specified in Table 7, is based on the following observations The fact that A + H + B K Q = Ker HQ can be easily checked by inspection CDi Fiore and PZellini showed in [DFZ95] that Q k (H Q ) e = Q e k+, (74) and used this fact to derive inversion formulas for structured matrices As was mentioned in Sec 4, 6, matrices {Q k (H Q )} form a basis in Ker HQ ( because of the nice property (74) we suggest to call it Di Fiore-Zellini basis), and since the first column of the matrix A + H + B coincides with the first column of S Q (A) given in Table 5, the representations (73) follow Recall that we started this section saying that all eight S Q (A) are the (discrete-trigonometric-transform) analogs of the Strang preconditioner S Indeed, the results in Table 7 show that the form (73) for S Q (A) is similar to that in (7) for S(A) Indeed, both the Strang circulant preconditioner S(A) and all of S Q (A) are constructed by adding to the given Toeplitz matrix A a matrix, in which the entries of the central diagonals of A now occupy the corner locations This fact is used next to prove the useful properties of S Q (A) 8 Clustering of the spectrum of S Q (A) A Here we establish the crucial property 3 for all eight GStrang-like preconditioners S Q (A) under the standard Wiener class assumption For the preconditioner S Q (A) corresponding to the DST-I, this property was established in [BB9] and below we adapt their arguments for the other seven S Q (A) Since S Q (A) A = I 6

S Q (A) (H +B) with H, B specified in Table 6, it is sufficient to show that the spectrum of S Q (A) (H + B) is clustered around Let ε> be fixed, choose N such that k=n+ a k <ε Then we can split H + B = A lr + A sn, (8) by taking out in A lr antidiagonals of H + B with the entries a,a,,a N Then the -norm of the second matrix in (8) can be bounded by ε (cf with [BB9]) Hence by the Cauchy interlace theorem, the eigenvalues of (H + B) are clustered around zero, except at most s = rank A lr outliers Applying the Courant-Fischer theorem to the matrix S Q (A) (H + B), we obtain implying that the Property 3 holds λ k {(S Q (A) (H + B)} < λ k{h + B} min z = f(z) 9 TChan-type preconditioners Here we specify another family of preconditioners C Q (A), defined by C Q (A) A F = min R K Q R A F, ie, the optimal Frobenius-norm approximants of A in K Q = Ker HQ (for the circulant case, H Q = Z, such preconditioner was proposed by TChan [C88]) Recall that we designed all Strang-type preconditioners S Q (A) using a representation of the form (6) It turns out that the same basis is convenient for writing down the formulas also for the TChan-type preconditioners C Q (A), and especially for the analysis of the clustering property in Sec 9 In order to obtain the coefficients in m C Q (A) = r k Q k (H Q ), (9) we solve a linear system of equations k= r k C Q (A) A F = (k =,,,m) (9) To solve (9) we found the entries of matrices {Q k (H Q )} in all eight cases as follows The entries of the first column of each Q k (H Q ) are given by (74) The entries of the other columns can be recursively computed using the fact that Q k (H Q ) K Q = Ker HQ For example, for n =5we have : Q (H Q )=Q I (with Q = for DCT-I and DCT-III, and Q = in the other six cases), Q (H Q )=H Q (shown in Table 4), and Q (H Q ),Q 3 (H Q ),are given in Table 8 7

Table 8 First-column bases Q (H Q ) Q 3 (H Q ) Q 4 (H Q ) DCT-I DCT-II DCT-III DCT-IV DST-I DST-II DST-III DST-IV The idea of using similar bases was also used for DST-I and discrete Hartley transforms in [BB9], [BF93], [BK95] 8

Using a particular form of Q k (H Q ) shown in Table 8, we obtain [ from ] (9) that the desired coefficients in (9) are then obtained from the given Toeplitz matrix A = a i j by r r m = G Q a a m, (93) where the matrix G Q has the simple structure shown in the Table 9 Table 9 Definition of the TChan-type preconditioner C Q (A m ) Matrix G Q for (93) DCT-I G = (m ) diag {,,,, }(D + E + L + U), with the terms specified by (94), (95), (96), (97) DCT-II m m (m )(m ) (m ) 4 m (m ) (m )(m ) 4 m (m ) 4 4 (m ) 4 m (m )(m ) (m ) 4 (m ) m (m )(m ) m DCT-III m (m + ) (m + 3) ( +) m m m m m DCT-IV m D = diag ((m ), (m ) + (m 3)(m 3), (m ) + (m 3)(m 4), (94), (m ) + (m 3), (m ) + (m 3), (m ), (m 3)), (a recursion for the, 3,,m,m entries is apparent) E = [ toeplitz (,,,,,, ] [ ) diag (,,,,, ] ) (95) 9

Here we follow the MATLAB notations, where toeplitz (c, r) denotes the Toeplitz matrix with the first column c and the first row r toeplitz (c) denotes the symmetric Toeplitz matrix with the first column c [ ] [ ] L = toeplitz (,,,,,,,,,,,, ) (96) [ diag ( [ U = toeplitz (,, [ diag ( ],, 3,, (m ),, ) ] [,,,,,,,,,,, (m 4), (m 3),, 4,,, ] ) (97) ] ) Table 9 Continuation Definition of the TChan-type preconditioner C Q (A m ) Matrix G Q for (93) DST-I m+ m + (m ) m + (m 3) m 5 4 3 DST-II m m (m )(m ) (m ) ( ) m 4 ( ) m m (m ) (m )(m ) m (m ) 4 4 (m ) m (m )(m ) (m ) ( ) m ( ) m 4 (m ) m (m )(m ) DST-III m m ( 3+ ) m + m 3+ ( + ) ( + ) + + DST-IV m m (m ) m (m ) m In [CNW96]a method is proposed to find the optimal sine transform based preconditioner when we have the corresponding basis

Formula (93) along with the data in Table 9 show that all TChan-type preconditioners trivially satisfy the properties, In the next section we establish for them the property 3 Clustering of eigenvalues of C Q (A) A Clustering Here we establish for all eight preconditioners C Q (A) the property 3 under the Wiener class assumption, by showing that the spectra of S Q (A) and C Q (A) are asymptotically the same (for the classical GStrang and TChan circulant preconditioners this property was established in [C89]) Proposition Let A m be a finite section of a semi-infinite Toeplitz matrix, {Q k } m k= be one of the eight polynomial systems in Table, and let S Q (A),C Q (A) R n n denote the corresponding Strang-type and the TChan-type preconditioners, respectively Then where denotes the spectral norm in R m m lim C Q(A) S Q (A) =, m Proof Since both C Q (A) and S Q (A) belong to K Q, they are both diagonalized by the corresponding discrete-trigonometric transform matrix T Q, see, eg, Proposition 6 Since T Q is an orthogonal matrix (ie, one of the eight orthogonal matrices displayed in Table ), we have essentially to establish the convergence to zero of the eigenvalues of S Q (A) C Q (A): lim λ k(s Q (A) C Q (A)) = m (k =,,, m) Again, by Proposition 6 the eigenvalues of S Q (A) and C Q (A) can be obtained from the coefficients {r k } in the representation (6) for these matrices by using (65) As shown in Sec 9, for the C Q (A) these coefficients are given by r r m a = G Q a m, () where the matrices G Q are listed in Table 9 For convenience we next rewrite the results of Table 5 in a similar manner Moreover, the coefficients in the representation S Q (A) = m k= r kc Q (HQ T ) are obtained by r a = R Q, () r m where matrices R Q are specified in Table a m

Table Definition of S Q (A m ) The matrix R Q in () DCT-I DST-I DCT-II DST-II DCT-III DST-III DCT-IV DST-IV By comparing (65), () and () we have : λ (S Q (A) C Q (A)) λ m (S Q (A) C Q (A)) = V Q (R Q G Q ) a a m, (3) where the matrices G Q and R Q are displayed in Tables 9 and, respectively Recall that not all of the eight matrices V Q have uniformly bounded entries, see, eg, Table Therefore it is more convenient to rewrite (3) as λ (S Q (A) C Q (A)) λ m (S Q (A) C Q (A)) =(V Q R Q ) (I R Q G Q)D (D a a m ), (4)

where D = diag ( m, m m,, m, ) Now we can prove the statement of the proposition, ie that the entries on the left-hand side of (4) tend to zero by making three following observations for three factors on the right-hand side of (4) Left factor The entries of the matrix V Q R Q are uniformly bounded independently of m Middle factor The column sums of the matrix (I R Q G Q)D have uniformly bounded column sums 3 Right factor If f(z) = k= a k z k is from the Wiener class, then ε > N > such that M >N we have M k M a k <ε k= The first observation can be deduced from the comparison of () and the Table 6, showing that V Q R Q is a cosine matrix The assertion in the second observation is deduced from the particular form of matrices R Q and G Q displayed in Tables 9 and The arguments are immediate in the cases DCT-III, DCT-IV, DST-III, DST-IV, and they are not much more involved in the case of DST-I In the cases of DCT-I, DCT-II and DST-II one has to split G Q into three parts : bidiagonal, upper and lower triangular, and for each of the corresponding parts of (I RQ G Q)D the statement is easily deduced The third observation is immediate (cf with [C89]) Thus, Proposition is now proved, and it implies the property 3 using standard arguments (cf with [C89]) Transformation-to-Cauchy-like and the Tyrtyshnikov property In the previous section we proved the properties -3 for C Q (A m ) The properties 4 and 6 (formulated in Sec 6) also follow from Prop Here we show that property 5 also holds, and, moreover, we prove that for any of eight TChan-type preconditioners the following Tyrtyshnikov property holds independently of m λ min (A m ) λ min (C Q (A m )) λ max (C Q (A m )) λ max (A m ) (5) For the circulant TChan preconditioner such a property was proved in [T9] and [CJY9] To prove (5) we shall use our definitions in Sec Observe that since the Frobenius norm in R m m generates the inner product (6), we have that the HQ -generator of A m is given by {G Q,J,C Q (A m )} in H Q A m A m H Q = G Q JG T Q The particular form of the m 4 matrix G Q and 4 4 matrix J is not relevant at the moment (for each H Q of Table 4 they can be easily written down as, for example, in (7) It is important that the corresponding TChan-type preconditioner describes the kernel component of A m, ie, the third matrix in its HQ -generator Furthermore, specifying the transformation-to-cauchy-like prop 4 to our settings here we obtain the results displayed in the Table Table Transformation-to-Cauchy-like Toeplitz Cauchy-like Matrix A m T Q A m TQ T G Q T Q G Q Generator J J C Q (A m ) T Q C Q (A m ) TQ T 3

The inequality (5) now follows from the following two observations First, since T Q is orthogonal, the spectra of A m and T Q A m TQ T and of C Q(A m ) and T Q C Q (A m )TQ T are, respectively, the same Second, since the Frobenius norm is unitary-equivalent, we have that the diagonal matrix T Q C Q (A m ) TQ T is the optimal Frobenius-norm diagonal approximants of the Cauchy-like matrix T Q A m TQ T In other words, T Q C Q (A m ) TQ T is simply a diagonal part of T QA m TQ T, implying (5) The above arguments indicate that there is a close connection between finding an optimal Frobeniusnorm approximant of a Toeplitz matrix, and transformations of Toeplitz matrices to Cauchy-like matrices Such transformations were closely studied in several recent papers For example, in [O93b], [O95] and [He95b] direct (ie, based on the computation on the matrix entries, without explicit use of displacement operators) transformations were considered Their results can be applied to obtain TChan-type preconditioners for the transforms II and III In this paper explicit formulas are obtained not only for TChan-type, but also for Strang-type preconditioners Moreover, in obtaining TChan-type preconditioners we follow [GO94a], [GKO95], [KO95a], [KO94], [KO95b]), and explore a different approach to transformations to Cauchy-like matrices The crucial point here is to introduce an appropriate displacement operator, HQ, where H Q is diagonalized by a unitary matrix New transformation formulas (systematically obtained here for all eight cases) require only one discrete trigonometric transform to compute the diagonal part of a Cauchy-like matrices, as compared to such transforms in [O93b], [O95], [He95b] Furthermore, the concept of partially reconstructible matrices suggested to use the definition of HQ -generator, given in [KO95a] This allowed us to obtain a unified descriptions for the both Strang-type and TChan-type preconditioners, given in Tables 9 and, resp These new formulas allowed us to establish in Sec 9 the result on close asymptotic behavior of both classes of preconditioners, and to prove the crucial clustering property for C Q (A m ) Real-arithmetic algorithms for multiplication of a Toeplitz matrix by a vector Embedding Real symmetric Toeplitz matrices In the first part of the paper we developed two families of Strangtype and of TChan-type preconditioners for real symmetric Toeplitz matrices, and established the properties -3, guaranteeing a convergence for the PCGM (under the Wiener class assumption) In this and the next sections we address the question of how to efficiently organize the iteration process itself First observe that all the computations with new preconditioners (ie, their construction, and then solving the associated linear systems) can be done in real arithmetic To fully exploit this advantageous property we have to specify an efficient real-arithmetic algorithm for multiplication of a Toeplitz matrix by a vector (the standard technique is based on the FFT, assuming complex arithmetic) In this section we observe that the explicit formulas obtained for the Strang-type readily suggest such algorithms for all eight cases These algorithms can be derived in two following steps First, we embed an m m Toeplitz matrix A m into a larger m m Toeplitz matrix A m by padding its first column with m zeros Secondly, we construct for A m the Strang-type preconditioner S Q (A m ) As was shown in Sec 7, this preconditioner admits a Toeplitz-plus-Hankel-plus-border representation, S Q (A m )=A m + H + B, with the Hankel part H and the border part B displayed in Table 7 Now, taking into account the banded structure of A m, one sees that in all eight cases the Hankel and the border part do not affect the central 4

part of S Q (A m ) In other words, our initial matrix A m is a submatrix of S Q (A m ) This observation allows us to use any of eight DCT s or DST s to multiply a real symmetric Toeplitz matrices by a vector in only two discrete trigonometric transforms of the order m (one more such transform is needed to compute the diagonal form for S Q (A)) For the case DST-I such an algorithm was proposed earlier by Boman and Koltracht in [BK95] Although this is beyond our needs in the present paper, note that the formulas for the Strang-type preconditioners allow us to multiply in the same way Toeplitz-plus-Hankel matrices by a vector These algorithms are analogs of the well-known embedding-into-circulant (complex arithmetic) multiplication algorithm There is another well-known (complex arithmetic) method for multiplying a Toeplitz matrix A m by a vector, based on the decomposition of A m into a sum of circulant and skew-circulant matrices Realarithmetic discrete-trigonometric transform based analogs of this algorithm are offered in Sec However, before presenting them it is worth to note that the formulas for the Strang-type preconditioners also allow a multiplication of a non-symmetric Toeplitz by a vector Non-symmetric Toeplitz [ matrices ] If A is a non-symmetric Toeplitz matrix, it can be embedded B A into a symmetric Toeplitz matric A and the latter can be used to multiply A by vectors B Another way to compute the matrix-vector product Decomposition In this section we describe alternative algorithms to multiply a real symmetric Toeplitz matrix by a vector These new algorithms are based on the formulas, which are counterparts of the well-know decomposition of a Toeplitz matrix into a sum of a circulant and a skew-circulant matrices For example, from the Toeplitzplus-Hankel-plus-border decompositions in Table 7 it immediately follows that, A m = (S C(A m )+S S (A m )), A m = (S C4(A m )+S S4 (A m )) () where we denote by C,C,C3,C4,S,S,S3,S4 the corresponding polynomial systems Q of Table Since the preconditioners S Q (A m ) are diagonalized by the corresponding transform matrices T Q, each of these formulas clearly allow us to multiply A m by a vector in just four discrete trigonometric transforms (with two more transforms needed only once to diagonalize S Q (A)) 3 Transformation-to-Cauchy-like approach for the PCGM As was detailed in Sec 4, polynomial Hankel-like matrices (this class includes Toeplitz-like matrices) can be transformed into Cauchy-like matrices In several recent papers this idea has been found to be useful for design of accurate direct methods for solving Toeplitz linear systems In this section we suggests an application of this technique to PCGM for Toeplitz matrices More specifically, instead of applying PCGM to the preconditioned system we suggest to apply it to the transformed system S Q (A m ) A m x = b, (3) (T Q S Q (A m ) T T Q ) (T Q A m T T Q )(T Q x)=t Q b where the preconditioner is transformed to the diagonal matrix (T Q S Q (A m ) TQ T ), and the Toeplitz matrix A m is transformed into a Cauchy-like matrix T Q A m TQ T Since a diagonal linear system can be solved in m operations, such transformation saves us two discrete transforms per iteration, if we can multiply the Cauchy-like matrix T Q A m TQ T by a vector with exactly the same complexity as the initial matrix A m 5