Cosine transform preconditioners for high resolution image reconstruction

Size: px
Start display at page:

Download "Cosine transform preconditioners for high resolution image reconstruction"

Transcription

1 Linear Algebra and its Applications 36 (000) Cosine transform preconditioners for high resolution image reconstruction Michael K. Ng a,,, Raymond H. Chan b,,tonyf.chan c,3, Andy M. Yip a a Department of Mathematics, The University of Hong Kong, Pokfulam Road, Hong Kong b Department of Mathematics, The Chinese University of Hong Kong, Shatin, Hong Kong c Department of Mathematics, University of California at Los Angeles, 405 Hilgard Avenue, Los Angeles, CA , USA Received March 999; accepted 5 December 999 Dedicated to Prof. Robert Plemmons in celebration of his 60th birthday Submitted by J. Nagy Abstract This paper studies the application of preconditioned conjugate gradient methods in high resolution image reconstruction problems. We consider reconstructing high resolution images from multiple undersampled, shifted, degraded frames with subpixel displacement errors. The resulting blurring matrices are spatially variant. The classical Tikhonov regularization and the Neumann boundary condition are used in the reconstruction process. The preconditioners are derived by taking the cosine transform approximation of the blurring matrices. We prove that when the L or H norm regularization functional is used, the spectra of the preconditioned normal systems are clustered around for sufficiently small subpixel displacement errors. Conjugate gradient methods will hence converge very quickly when applied to solving these preconditioned normal equations. Numerical examples are given to illustrate the fast convergence. 000 Elsevier Science Inc. All rights reserved. Keywords: Toeplitz matrix; Discrete cosine transform; Image reconstruction; Neumann boundary condition Corresponding author. Tel.: ; fax: addresses: mng@maths.hku.hk (M.K. Ng), rchan@math.cuhk.edu.hk (R.H. Chan), chan@math.ucla.edu (T.F. Chan), mhyipa@hkusua.hku.hk (A.M. Yip). Research supported in part by Hong Kong Research Grants Council Grant No. HKU 747/99P and HKU CRCG Grant No Research supported in part by Hong Kong Research Grants Council Grant No. CUHK 407/97P and CUHK DAG Grant No Research supported by ONR under grant N and NSF under grant DMS /00/$ - see front matter 000 Elsevier Science Inc. All rights reserved. PII:S (99)0074-8

2 90 M.K. Ng et al. / Linear Algebra and its Applications 36 (000) Introduction Due to hardware limitations, imaging systems often provide us with only multiple low resolution images. However, in many applications, a high resolution image is desired. For example, the resolution of the pictures of the ground taken from a satellite is relatively low and retrieving details on the ground becomes impossible. Increasing the image resolution by using digital signal processing techniques [4,,6,8,0,] is therefore of great interest. We consider the reconstruction of a high resolution image from multiple undersampled, shifted, degraded and noisy images. Multiple undersampled images are often obtained by using multiple identical image sensors shifted from each other by subpixel displacements. The reconstruction of high resolution images can be modeled as solving Hf = g, () where g is the observed high resolution image formed from the low resolution images, f the desired high resolution image and H is the reconstruction operator. If all the low resolution images are shifted from each other with exactly half-pixel displacements, H will be a spatially invariant operator. However, displacement errors may be present in practice, and the resulting operator H becomes spatially variant. Since the systems are ill-conditioned and generally not positive definite, we solve them by using a minimization and regularization technique: } { Hf g + αr(f ). () min f Here R(f ) is a functional which measures the regularity of f and the regularization parameter α is to control the degree of regularity of the solution. In this paper, we will use the L and H regularization functionals f and Lf,whereLis the first order differential operator. Owing to the blurring (convolution) process, the boundary values of g are not completely determined by the original image f inside the scene. They are also affected by the values of f outside the scene. Thus in solving f from (), we need some assumptions on the values of f outside the scene. These assumptions are called boundary conditions. Bose and Boo [4] used the traditional choice of imposing the zero boundary condition outside the scene, i.e., assuming a dark background outside the scene in the image reconstruction. However, when this assumption is not satisfied by the images, ringing effects will occur at the boundary of the reconstructed images. The problem is more severe if the images are reconstructed from a large sensor array since the number of pixel values of the image affected by the sensor array increases. In this paper, we will use the Neumann boundary condition on the image, i.e., we assume that the scene immediately outside is a reflection of the original scene at the boundary. The Neumann boundary condition has been studied in image restoration [,3,5] and in image compression [4,9]. Our experimental results in [6] have shown that the Neumann image model gives better reconstructed high resolution

3 M.K. Ng et al. / Linear Algebra and its Applications 36 (000) images than that under the zero or periodic boundary conditions. In [6], we also proposed to use cosine transform preconditioners to precondition the resulting linear systems and preliminary numerical results have shown that these preconditioners are effective. The main aim of this paper is to analyze the convergence rate of these systems. We prove that when the L or H norm regularization functional is used, the spectra of the preconditioned systems are clustered around for sufficiently small displacement errors. The outline of the paper is as follows. In Section, we give a mathematical formulation of the problem. A brief introduction on the cosine transform preconditioners and the convergence analysis will be given in Section 3. In Section 4, numerical results are presented to demonstrate the effectiveness of the cosine transform preconditioners.. The mathematical model We begin with a brief introduction of the mathematical model in high resolution image reconstruction. Details can be seen in [4]. Consider a sensor array with L L sensors, each sensor has N N sensing elements (pixels) and the size of each sensing element is T T.Ouraimis to reconstruct an image of resolution M M,whereM = L N and M = L N. To maintain the aspect ratio of the reconstructed image, we consider the case where L = L = L only. For simplicity, we assume that L is an even number in the following discussion. In order to have enough information to resolve the high resolution image, there are subpixel displacements between the sensors. In the ideal case, the sensors are shifted from each other by a value proportional to T /L T /L. However, in practice, there can be small perturbations around these ideal subpixel locations due to imperfection of the mechanical imaging system. Thus, for l,l = 0,,...,L with (l,l )/= (0, 0), the horizontal and vertical displacements d x l l and d y l l of the [l,l ]th sensor array with respect to the [0, 0]th reference sensor array are given by d x l l = T ) (l + ɛ x l l L and d y l l = T ) (l + ɛ y. l l L Here ɛl x l and ɛ y l l denote, respectively, the normalized horizontal and vertical displacement errors. We remark that the parameters ɛl x l and ɛ y l l can be obtained by manufacturers during camera calibration. We assume that ɛl x l < and ɛ y l l <. For if not, the low resolution images observed from two different sensor arrays will be overlapped so much that the reconstruction of the high resolution image is rendered impossible.

4 9 M.K. Ng et al. / Linear Algebra and its Applications 36 (000) Let f be the original scene. Then the observed low resolution image g l l for the (l,l )th sensor is modeled by ( ) T n + +d y l l g l l [n,n ]= T (n ) +d y l l ( ) T n + +d x l l T (n )+d x l l f(x,x ) dx dx +η l l [n,n ] (3) for n =,...,N and n =,...,N.Hereη l l is the noise corresponding to the (l,l )th sensor. We intersperse the low resolution images to form an M M image by assigning g[l(n ) + l,l(n ) + l ]=g l l [n,n ]. (4) Here g is an M M image and is called the observed high resolution image.fig. shows the method of forming a 4 4imageg with a sensor array, where each g ij has a sensing elements, i.e. L =, M = M = 4, and N = N =. Using a column by column ordering for g, we obtain g = Hf + η,whereh is a spatially variant operator [4]. Since H is ill-conditioned due to the averaging of the pixel values in the image model in (3), the classical Tikhonov regularization is used and the minimization problem () is solved. In this paper, we use the regularization functionals R(f ) = f and R(f ) = Lf, (5) where L is the first order differential operator. Fig.. Construction of the observed high resolution image.

5 .. Image boundary M.K. Ng et al. / Linear Algebra and its Applications 36 (000) The continuous image model in (3) can be discretized by the rectangular rule and approximated by a discrete image model. Because of the blurring process (cf. (3)), the boundary values of g are also affected by the values of f outside the scene. Thus in solving f from (), we need some assumptions on the values of f outside the scene. Bose and Boo [4] imposed the zero boundary condition outside the scene, i.e., assuming a dark background outside the scene in the image reconstruction. Let g and f be, respectively, the discretization of g and f using a column by column ordering. Under the zero boundary condition, the blurring matrix corresponding to the (l,l )th sensor can be written as H l l (ɛ) = H x l l (ɛ) H y l l (ɛ), where H x l l (ɛ) is an M M banded Toeplitz matrix with bandwidth L and h x+ l l H x l l (ɛ) = h x+ l l L h x l l h x l l h x± l l = ± ɛx l l. The M M banded blurring matrix H y l l (ɛ) is defined similarly. We note that ringing effects will occur at the boundary of the reconstructed images if f is indeed not zero close to the boundary, see for instance Fig. 3 in Section 4. The problem is more severe if the image is reconstructed from a large sensor array since the number of pixel values of the image affected by the sensor array increases. In [6], we proposed to use the Neumann boundary condition on the image. It assumes that the scene immediately outside is a reflection of the original scene at the boundary. Our numerical results have shown that the Neumann boundary condition gives better reconstructed high resolution images than that by the zero or periodic boundary conditions. Under the Neumann boundary condition, the blurring matrices are still banded matrices with bandwidth L, but there are entries added to the upper left-part and the lower right-part of the matrices (see the second matrix in (6)). The resulting matrices, denoted by H x l l (ɛ) and H y l l (ɛ), have a Toeplitz-plus- Hankel structure

6 94 M.K. Ng et al. / Linear Algebra and its Applications 36 (000) h x+ l l H x l l (ɛ) = h x+ l l L h x l l h x l l h x l l h x+ l l L h x... l l h x+ l l (6) and H y l l (ɛ) is defined similarly. The blurring matrix corresponding to the (l,l )th sensor under the Neumann boundary condition is given by H l l (ɛ) = H x l l (ɛ) H y l l (ɛ). The blurring matrix for the whole sensor array is made up of blurring matrices from each sensor: L L H L (ɛ) = D l l H l l (ɛ). (7) l =0 l =0 Here D l l are diagonal matrices with diagonal elements equal to if the corresponding component of g comes from the (l,l )th sensor and zero otherwise, see [4] for more details. With the Tikhonov regularization, our discretization problem becomes (H L (ɛ) t H L (ɛ) + αr)f = H L (ɛ) t g, (8) where R is the discretization matrix corresponding to the regularization functional R(f ) in (5). 3. Cosine transform based preconditioners The linear system (8) will be solved by using the preconditioned conjugate gradient method. In this section, we construct the cosine transform preconditioner of H L (ɛ) which exploits the banded and block structures of the matrix. Let C n be the n n discrete cosine transform matrix, i.e., the (i, j)th entry of C n is given by δi n ( (i )(j )π cos n ), i, j n,

7 M.K. Ng et al. / Linear Algebra and its Applications 36 (000) where δ ij is the Kronecker delta. Note that the matrix vector product C n z can be computed in O(n log n) operations for any vector z, see [3, pp. 59,60]. For an m m block matrix B with the size of each block equal to n n, the cosine transform preconditioner c(b) of B is defined to be the matrix (C m C n )Λ(C m C n ) that minimizes (C m C n )Λ(C m C n ) B F in the Frobenius norm, see [8]. Here Λ is any diagonal matrix. Clearly, the cost of computing c(b) y for any vector y is O(mn log mn) operations. For banded matrices in (7), which have (L ) non-zero diagonals and are of size M M M M, the cost of constructing c(h L (ɛ)) is of O(L M M ) operations only, see [7]. 3.. Spatially invariant case When there are no subpixel displacement errors, i.e., when all ɛl x,l = ɛ y l,l = 0, the matrices H x l l (0) and also H y l l (0) are the same for all l and l. We will denote them simply by H x L and Hy L. We claim that in this case, the blurring matrix H L H L (0) = H x L Hy L can always be diagonalized by the discrete cosine transform matrix. We begin with L =. The blurring matrix H = H x Hy,whereHx is an M M tridiagonal matrix given by H x = = and H y is an M M matrix with the same structure. It is easy to see that in this case, the matrices H x and Hy can be diagonalized by C M and C M respectively, see the basis given in [,3] for the class of matrices that can be diagonalized by the cosine transform matrix. Thus H can be diagonalized by C M C M. Next we observe that the blurring matrix is ill-conditioned. Lemma. Under the Neumann boundary condition, the M M matrix H x can be diagonalized by the discrete cosine transform matrix and its eigenvalues are given by

8 96 M.K. Ng et al. / Linear Algebra and its Applications 36 (000) λ j (H x ) = cos ( (j )π M ), j M. (9) In particular, the condition number κ(h x ) of the matrix Hx satisfies κ(h x ) O(M ). (0) Proof. The formula for the eigenvalues can be derived easily using the basis given in [,3] for the class of matrices that can be diagonalized by the cosine transform matrix. Since λ max (H x ) = and λ min (H x ) = cos ( (M )π M ) sin ( πm ) π the estimate of the condition number is then given by (0). M It follows from Lemma that the condition number of the matrix H (= H x Hy ) is of O(M M ). The matrix is very ill-conditioned. For L>, we have the following theorem. Theorem. Under the Neumann boundary condition, the matrix H x L can be diagonalized by the discrete cosine transform matrix and its eigenvalues are given by λ i (H x L ) = 4 ( ) ( ) (i )π (i )π L cos p L, i M, () M M where ( ) (i )π p L = M L/4 ( (i )(j )π cos j= M ), L = 4k for some positive integer k, (L )/4 + ( ) (i )jπ cos, M j= otherwise. Proof. We first establish a relationship between the matrices H x L and Hx.From(6), for L>, we have L/4 S j H x, L = 4k for some positive integer k, L H x L = j= (3) (L )/4 S j H x L, otherwise, j=0 ()

9 M.K. Ng et al. / Linear Algebra and its Applications 36 (000) where S 0 is the M M identity matrix and S k = Toeplitz(e k+ ) + Hankel(e k ), k M. Here Toeplitz(e k )isthem M symmetric Toeplitz matrix with the kth unit vector e k as the first column, and Hankel(e k )isthem M Hankel matrix with e k as the first column and e k in the reverse order as the last column. We remark that the Toeplitz part in S k can be interpreted as the decomposition of the discrete blurring function [0.5,,...,,...,, 0.5] into the sum of the elementary discrete blurring function [0.5,, 0.5] with different shifts. For example, for L = 4, we have [,,,, ] [ =,, ] [, 0, 0 + 0, 0,,, ], where the two terms on the right-hand side together give the Toeplitz part in S.For L = 6, we have [,,,,,, ] [ =,, ], 0, 0, 0, 0 + [0, 0,,, ], 0, 0 + [ 0, 0, 0, 0,,, where the first and the third terms on the right-hand side together give the Toeplitz part in S while the middle term gives the Toeplitz part of S 0. Because we are considering the Neumann boundary condition, entries outside the blurring matrix H x L are flipped into the matrix (cf. (6)). This is done by means of the Hankel part of S k. Thus the resulting shift matrices are given by S k and we obtain (3). Since {S k } M k=0 is exactly a basis for the space containing all matrices that can be diagonalized by C M, see [3], it follows that the matrix H x L can be diagonalized by the discrete cosine transform matrix. We also note that the eigenvalues of S k are given by ( (i )kπ λ i (S k ) = cos M ], ), i M, see for instance [3]. Using (3) and (9), the eigenvalues of H x L are given in (). Theorem states that the matrices H L (= H x L Hy L ) are also very ill-conditioned and their condition numbers are at least of order M M (cf. ()). We remark that some of these matrices may even be singular. For instance, when L = 4andM = M = 64, λ 33 (H x 4 ) = 0. Thus a regularization procedure such as (8) should be imposed to obtain a reasonable estimate for the original image in the high resolution reconstruction. In this paper, we consider the L and H norm regularization functionals in (8). Correspondingly, we are required to solve the following linear systems: (H t L H L + αi)f = H t L g or (Ht L H L + αl t L)f = H t Lg, (4)

10 98 M.K. Ng et al. / Linear Algebra and its Applications 36 (000) where α>0, I is the identity matrix and L t L is the discrete Laplacian matrix with the Neumann boundary condition. We note that L t L can be diagonalized by the discrete cosine transform matrix, see for instance [5]. Thus if we use the Neumann boundary condition for both the blurring matrix H L and the regularization operator L t L, then the coefficient matrix in (4) can be diagonalized by the discrete cosine transform matrix and hence its inversion can be done in three two-dimensional fast cosine transforms (one for finding the eigenvalues of the coefficient matrix, two for transforming the right-hand side and the solution vector, see for instance [7]). Thus the total cost of solving the system is of O(M M log M M ) operations. We remark that for the zero boundary condition, discrete sine transform matrices can diagonalize Toeplitz matrices with at most three bands (e.g., H ) but not dense Toeplitz matrices in general (e.g., H 4 ), see for instance [9]. Therefore, in general, we have to solve large block-toeplitz Toeplitz-block systems. The fastest direct Toeplitz solvers require O(M M ) operations, see []. The systems can also be solved by the preconditioned conjugate gradient method with some suitable preconditioners, see [8]. We note however that the cost per iteration is at least four two-dimensional fast Fourier transforms. Thus we see that the cost of using the Neumann boundary condition is significantly lower than that of using the zero boundary condition. 3.. Spatially variant case When there are subpixel displacement errors, the blurring matrix H L (ɛ) has the same banded structure as that of H L, but with some entries slightly perturbed. It is a near block-toeplitz Toeplitz-block matrix, but, it can no longer be diagonalized by the cosine transform matrix. Therefore we solve the corresponding linear system by the preconditioned conjugate gradient method. We will use the cosine transform preconditioner c(h L (ɛ)) of H L (ɛ) as the preconditioner. Below we study the convergence rate of the preconditioned conjugate gradient method for solving the linear systems [c(h L (ɛ)) t c(h L (ɛ)) + αi] [H L (ɛ) t H L (ɛ) + αi]f = H L (ɛ) t g (5) and [c(h L (ɛ)) t c(h L (ɛ)) + αl t L] [H L (ɛ) t H L (ɛ) + αl t L]f = H L (ɛ) t g, (6) where α is a positive constant. We prove that the spectra of the preconditioned normal systems are clustered around for sufficiently small subpixel displacement errors. Hence when the conjugate gradient method is applied to solving the preconditioned systems (5) and (6), we expect fast convergence. Our numerical results in Section 4 show that the cosine transform preconditioners can indeed speed up the convergence of the method. We begin the proof with the following lemma. Lemma. Let ɛ = max 0 l,l L {ɛl x l,ɛ y l l }. Then for all M and M, we have H L (ɛ) H L 4ɛ and c(h L (ɛ)) H L 4ɛ. (7)

11 M.K. Ng et al. / Linear Algebra and its Applications 36 (000) Proof. From (7), each row or column of H L (ɛ) and H L differ in at most 4L entries and each entry is bounded by ɛ /L. It follows that H L (ɛ) H L 4ɛ and H L (ɛ) H L 4ɛ. Hence the first inequality in (7) follows by using.forthe second inequality, we first note that by Theorem, c(h L ) = H L. Hence we have c(h L (ɛ)) H L = c(h L (ɛ) H L ) H L (ɛ) H L, where the last inequality follows from c( ), see []. Lemma 3. Let ɛ = max 0 l,l L {ɛ x l l,ɛ y l l }.Then H L (ɛ) t H L (ɛ) c(h L (ɛ)) t c(h L (ɛ)) <d L (ɛ ), where d L ( ) is a function independent of M and M and lim ɛ 0 d L (ɛ ) = 0. Proof. We note that H L (ɛ) t H L (ɛ) c(h L (ɛ)) t c(h L (ɛ)) H L (ɛ) t [H L (ɛ) c(h L (ɛ))] + [H L (ɛ) t c(h L (ɛ)) t ]c(h L (ɛ)). By Theorem, H L is bounded above by a constant independent of M and M. Hence by (7), H L (ɛ) t and c(h L (ɛ)) are also bounded above by some constants independent of M and M. Moreover, by (7) again, H L (ɛ) c(h L (ɛ)) and H L (ɛ) t c(h L (ɛ)) t are less than 8ɛ. The result therefore follows. Using the above lemmas, we can analyze the convergence rate of the preconditioned systems (5) and (6). Theorem. Let ɛ = max 0 l,l L {ɛ x l l,ɛ y l l }.Ifɛ is sufficiently small, then the spectra of the preconditioned matrices [c(h L (ɛ)) t c(h L (ɛ)) + αi] [H L (ɛ) t H L (ɛ) + αi] are clustered around and their smallest eigenvalues are bounded away from 0 by a positive constant independent of M and M. Proof. We just note that [c(h L (ɛ)) t c(h L (ɛ)) + αi] α and [c(h L (ɛ)) t c(h L (ɛ)) + αi] [H L (ɛ) t H L (ɛ) + αi] = I +[c(h L (ɛ)) t c(h L (ɛ)) + αi] [H L (ɛ) t H L (ɛ) c(h L (ɛ)) t c(h L (ɛ))]. Hence the result follows by applying Lemma 3.

12 00 M.K. Ng et al. / Linear Algebra and its Applications 36 (000) Theorem 3. Let ɛ = max 0 l,l L {ɛ x l l,ɛ y l l }.Ifɛ is sufficiently small, then the spectra of the preconditioned matrices [c(h L (ɛ)) t c(h L (ɛ)) + αl t L] [H L (ɛ) t H L (ɛ) + αl t L] are clustered around and their smallest eigenvalues are uniformly bounded away from 0 by a positive constant independent of M and M. Proof. Since [c(h L (ɛ)) t c(h L (ɛ)) + αl t L] [H L (ɛ) t H L (ɛ) + αl t L] = I +[c(h L (ɛ)) t c(h L (ɛ)) + αl t L] [H L (ɛ) t H L (ɛ) c(h L (ɛ)) t c(h L (ɛ))], it suffices to show that [c(h L (ɛ)) t c(h L (ɛ)) + αl t L] is bounded above by a constant independent of M and M.Sinceλ min (A) + λ min (B) λ min (A + B) for any Hermitian matrices A and B (see [0, Theorem 8..5, p. 396]), we have [c(h L (ɛ)) t c(h L (ɛ)) + αl t L] = λ min (c(h L (ɛ)) t c(h L (ɛ)) + αl t L) λ min (H t L H L + αl t L) + λ min (c(h L (ɛ)) t c(h L (ɛ)) H t L H L) λ min (H t L H L + αl t L) c(h L (ɛ)) t c(h L (ɛ)) H t L H. (8) L Since the matrix H t L H L + αl t L can be diagonalized by the two-dimensional discrete cosine transform matrix, we can estimate the smallest eigenvalue of this matrix. We first note that λ (i )M +j (L t L) = 4sin ( (i )π M ) ( ) (j )π + 4sin M for i M and j M, see [5]. By using () and the fact that H L = H x L Hy L, we obtain ( ) 4 4 ( ) ( ) (i )π (j )π λ (i )M +j (H t L H L) = cos 4 cos 4 L M M ( ) ( ) (i )π (j )π pl pl (0) M for i M and j M,wherep L ( ) is defined in (). Clearly the function sin (x/) is zero at x = 0 and positive in (0,π], whereas the function cos 4 (x/)pl (x) /4 atx = 0 and is nonnegative in [0,π]. Thus we see that the function M (9)

13 M.K. Ng et al. / Linear Algebra and its Applications 36 (000) α sin ( x ) ( + 4α sin y ) + ( 4 L ) 4 cos 4 ( x ) ( cos 4 y ) pl (x)p L (y) is positive for all x and y in [0,π]. It follows from (9) and (0) that the matrix H t L H L + αl t L is positive definite and its smallest eigenvalue is bounded away from 0 by a positive constant independent of M and M. In view of Lemma 3, the righthand side of (8) is therefore bounded by a positive constant independent of M and M for sufficiently small ɛ. Thus we conclude that the preconditioned conjugate gradient method applied to (5) and (6) with α>0 will converge superlinearly for sufficiently small displacement errors, see for instance [8]. Since H L (ɛ) has only (L ) non-zero diagonals, the matrix vector product H L (ɛ)x can be done in O(L M M ). Thus the cost per each iteration is O(M M log M M + L M M ) operations, see [0, p. 59]. Hence the total cost for finding the high resolution image vector is of O(M M log M M + L M M ) operations. 4. Numerical examples In this section, we illustrate the effectiveness of using cosine transform preconditioners for solving high resolution image reconstruction problems. The original image is shown in Fig. (a). The conjugate gradient method is employed to solving the preconditioned systems (5) and (6). The stopping criteria is r (j) / r (0) < 0 6,wherer (j) is the normal equations residual after j iterations. In the tests, the parameters ɛ x l l and ɛ y l l are random values chosen between / and /. A Gaussian white noise with signal-to-noise ratio of 30 db is added to the low resolution images. Tables and show the numbers of iterations required for convergence for L = and 4, respectively. In the tables, cos, cir or no signify that the cosine transform preconditioner, the level- circulant preconditioner [8] or no preconditioner is used, Fig.. The original image (a), a low resolution image (b), and the observed high resolution image (c).

14 0 M.K. Ng et al. / Linear Algebra and its Applications 36 (000) Table α M cos cir no cos cir no cos cir no (a) Number of iterations for L =, where the L norm regularization is used (b) Number of iterations for L =, where the H norm regularization is used Table α M cos cir no cos cir no cos cir no (a) Number of iterations for L = 4, where the L norm regularization is used (b) Number of iterations for L = 4, where the H norm regularization is used respectively. We see from the tables that the cosine transform preconditioner converges much faster than the circulant preconditioners for different M and α, where M(= M = M ) is the size of the reconstructed image and α is the regularization parameter. Also the convergence rate is independent of M for fixed α as predicted by Theorems and 3. Next we show the reconstructed images from four 8 8 low resolution images, i.e., a sensor array is used. One of the low resolution images is shown in Fig. (b). The observed high resolution image g isshowninfig.(c). We tried the Neumann, zero and periodic boundary conditions to reconstruct the high resolution images. Fig. 3 shows the reconstructed images. The optimal regularization parameter α is chosen such that it minimizes the relative error of the reconstructed image f r (α) to the original image f, i.e., it minimizes f f r (α) / f.bycomparing the (a) (c) in Fig. 3, it is clear that the trees in the image are much better reconstructed under the Neumann boundary condition than that under the zero and

15 M.K. Ng et al. / Linear Algebra and its Applications 36 (000) Fig. 3. Reconstructed image using the Neumann boundary condition (a), the zero boundary condition (b) and the periodic boundary condition (c). periodic boundary conditions. We also see that the boundary artifacts under the Neumann boundary condition are less prominent than that under the other two boundary conditions. References [] M. Banham, A. Katsaggelos, Digital image restoration, IEEE Signal Processing Magazine, March 997, pp [] E. Boman, Fast algorithms for Toeplitz equations, Ph.D. thesis, University of Connecticut, Storrs, CT, October 993. [3] E. Boman, I. Koltracht, Fast transform based preconditioners for toeplitz equations, SIAM J. Matrix Anal. Appl. 6 (995) [4] N. Bose, K. Boo, High-resolution image reconstruction with multisensors, Internat. J. Imaging Systems Technol. 9 (998) [5] M. Buckley, Fast computation of a discretized thin-plate smoothing spline for image data, Biometrika 8 (994) [6] R. Chan, T. Chan, M. Ng, W. Tang, C. Wong, Preconditioned iterative methods for high-resolution image reconstruction with multisensors, in: F. Luk (Ed.), Proceedings of the SPIE Symposium on Advanced Signal Processing: Algorithms, Architectures, and Implementations, vol. 346, San Diego CA, July, 998. [7] R. Chan, T. Chan, C. Wong, Cosine transform based preconditioners for total variation minimization problems in image processing, in: S. Margenov, P. Vassilevski (Eds.), Iterative Methods in Linear Algebra, II, V3, IMACS Series in Computational and Applied Mathematics, Proceedings of the Second IMACS International Symposium on Iterative Methods in Linear Algebra, Bulgaria, June, 995, pp [8] R. Chan, M. Ng, Conjugate gradient method for toeplitz system, SIAM Review 38 (996) [9] R. Chan, M. Ng, C. Wong, Sine transform based preconditioners for symmetric Toeplitz systems, Linear Algebra Appl. 3 (996) [0] G. Golub, C. Van Loan, Matrix Computations, third ed., Johns Hopkins University Press, Baltimore, MD, 996. [] N. Kalouptsidis, G. Carayannis, D. Manolakis, Fast algorithms for block Toeplitz matrices with Toeplitz entries, Signal Processing 6 (984) [] E. Kaltenbacher, R. Hardie, High resolution infrared image reconstruction using multiple, low resolution, aliased frames, in: Proceedings of the IEEE 996 National Aerospace and Electronic Conference on NAECON, vol., 996, pp

16 04 M.K. Ng et al. / Linear Algebra and its Applications 36 (000) [3] R. Lagendijk, J. Biemond, Iterative identification and restoration of images, Kluwer Academic Publishers, Dordrecht, 99. [4] J. Lim, Two-dimensional Signal and Image Processing, Prentice-Hall, Englewood Cliffs, NJ, 990. [5] F. Luk, D. Vandevoorde, Reducing boundary distortion in image restoration, in: Proceedings of the SPIE 96, Advanced Signal Processing Algorithms, Architectures and Implementations VI, 994. [6] S. Kim, N. Bose, H. Valenzuela, Recursive reconstruction of high resolution image from noisy undersampled multiframes, IEEE Trans. Acoust. Speech, and Signal Process. 38 (6) (990) [7] M. Ng, R. Chan, W. Tang, A fast algorithm for deblurring models with Neumann boundary conditions, Res. Rept , Department of Mathematics, The Chinese University of Hong Kong, SIAM J. Sci. Comput. (to appear). [8] R. Schultz, R. Stevenson, Extraction of high-resolution frames from video sequences, IEEE Trans. Image Process. 5 (6) (996) [9] G. Strang, The discrete cosine transform, SIAM Review 4 (999) [0] A. Tekalp, M. Ozkan, M. Sezan, High-resolution image reconstruction from lower-resolution image sequences and space-varying image restoration, in: Proceedings of the IEEE International Conference on Acoustic, Speech, and Signal Processing, III, San Francisco, CA, March 99, pp [] R. Tsai, T. Huang, Multiframe image restoration and registration, Adv. Comput. Vision and Image Process. (984)

Preconditioning regularized least squares problems arising from high-resolution image reconstruction from low-resolution frames

Preconditioning regularized least squares problems arising from high-resolution image reconstruction from low-resolution frames Linear Algebra and its Applications 39 (2004) 49 68 wwwelseviercom/locate/laa Preconditioning regularized least squares problems arising from high-resolution image reconstruction from low-resolution frames

More information

SIAM Journal on Scientific Computing, 1999, v. 21 n. 3, p

SIAM Journal on Scientific Computing, 1999, v. 21 n. 3, p Title A Fast Algorithm for Deblurring Models with Neumann Boundary Conditions Author(s) Ng, MKP; Chan, RH; Tang, WC Citation SIAM Journal on Scientific Computing, 1999, v 21 n 3, p 851-866 Issued Date

More information

The Best Circulant Preconditioners for Hermitian Toeplitz Systems II: The Multiple-Zero Case Raymond H. Chan Michael K. Ng y Andy M. Yip z Abstract In

The Best Circulant Preconditioners for Hermitian Toeplitz Systems II: The Multiple-Zero Case Raymond H. Chan Michael K. Ng y Andy M. Yip z Abstract In The Best Circulant Preconditioners for Hermitian Toeplitz Systems II: The Multiple-ero Case Raymond H. Chan Michael K. Ng y Andy M. Yip z Abstract In [0, 4], circulant-type preconditioners have been proposed

More information

Krylov subspace iterative methods for nonsymmetric discrete ill-posed problems in image restoration

Krylov subspace iterative methods for nonsymmetric discrete ill-posed problems in image restoration Krylov subspace iterative methods for nonsymmetric discrete ill-posed problems in image restoration D. Calvetti a, B. Lewis b and L. Reichel c a Department of Mathematics, Case Western Reserve University,

More information

Blind image restoration as a convex optimization problem

Blind image restoration as a convex optimization problem Int. J. Simul. Multidisci.Des. Optim. 4, 33-38 (2010) c ASMDO 2010 DOI: 10.1051/ijsmdo/ 2010005 Available online at: http://www.ijsmdo.org Blind image restoration as a convex optimization problem A. Bouhamidi

More information

Toeplitz-circulant Preconditioners for Toeplitz Systems and Their Applications to Queueing Networks with Batch Arrivals Raymond H. Chan Wai-Ki Ching y

Toeplitz-circulant Preconditioners for Toeplitz Systems and Their Applications to Queueing Networks with Batch Arrivals Raymond H. Chan Wai-Ki Ching y Toeplitz-circulant Preconditioners for Toeplitz Systems and Their Applications to Queueing Networks with Batch Arrivals Raymond H. Chan Wai-Ki Ching y November 4, 994 Abstract The preconditioned conjugate

More information

Preconditioning. Noisy, Ill-Conditioned Linear Systems

Preconditioning. Noisy, Ill-Conditioned Linear Systems Preconditioning Noisy, Ill-Conditioned Linear Systems James G. Nagy Emory University Atlanta, GA Outline 1. The Basic Problem 2. Regularization / Iterative Methods 3. Preconditioning 4. Example: Image

More information

Preconditioning. Noisy, Ill-Conditioned Linear Systems

Preconditioning. Noisy, Ill-Conditioned Linear Systems Preconditioning Noisy, Ill-Conditioned Linear Systems James G. Nagy Emory University Atlanta, GA Outline 1. The Basic Problem 2. Regularization / Iterative Methods 3. Preconditioning 4. Example: Image

More information

Preconditioning of elliptic problems by approximation in the transform domain

Preconditioning of elliptic problems by approximation in the transform domain TR-CS-97-2 Preconditioning of elliptic problems by approximation in the transform domain Michael K. Ng July 997 Joint Computer Science Technical Report Series Department of Computer Science Faculty of

More information

arxiv: v1 [math.na] 15 Jun 2009

arxiv: v1 [math.na] 15 Jun 2009 Noname manuscript No. (will be inserted by the editor) Fast transforms for high order boundary conditions Marco Donatelli arxiv:0906.2704v1 [math.na] 15 Jun 2009 the date of receipt and acceptance should

More information

Simultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors

Simultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors Simultaneous Multi-frame MAP Super-Resolution Video Enhancement using Spatio-temporal Priors Sean Borman and Robert L. Stevenson Department of Electrical Engineering, University of Notre Dame Notre Dame,

More information

1 Introduction The conjugate gradient (CG) method is an iterative method for solving Hermitian positive denite systems Ax = b, see for instance Golub

1 Introduction The conjugate gradient (CG) method is an iterative method for solving Hermitian positive denite systems Ax = b, see for instance Golub Circulant Preconditioned Toeplitz Least Squares Iterations Raymond H. Chan, James G. Nagy y and Robert J. Plemmons z November 1991, Revised April 1992 Abstract We consider the solution of least squares

More information

In [7], Vogel introduced the \lagged diusivity xed point iteration", which we denote by FP, to solve the system (6). If A u k, H and L u k denote resp

In [7], Vogel introduced the \lagged diusivity xed point iteration, which we denote by FP, to solve the system (6). If A u k, H and L u k denote resp Cosine Transform Based Preconditioners for Total Variation Deblurring Raymond H. Chan, Tony F. Chan, Chiu-Kwong Wong Abstract Image reconstruction is a mathematically illposed problem and regularization

More information

Mathematics and Computer Science

Mathematics and Computer Science Technical Report TR-2004-012 Kronecker Product Approximation for Three-Dimensional Imaging Applications by MIsha Kilmer, James Nagy Mathematics and Computer Science EMORY UNIVERSITY Kronecker Product Approximation

More information

A fast algorithm of two-level banded Toeplitz systems of linear equations with application to image restoration

A fast algorithm of two-level banded Toeplitz systems of linear equations with application to image restoration NTMSCI 5, No. 2, 277-283 (2017) 277 New Trends in Mathematical Sciences http://dx.doi.org/ A fast algorithm of two-level banded Toeplitz systems of linear equations with application to image restoration

More information

General Properties for Determining Power Loss and Efficiency of Passive Multi-Port Microwave Networks

General Properties for Determining Power Loss and Efficiency of Passive Multi-Port Microwave Networks University of Massachusetts Amherst From the SelectedWorks of Ramakrishna Janaswamy 015 General Properties for Determining Power Loss and Efficiency of Passive Multi-Port Microwave Networks Ramakrishna

More information

A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations

A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations Jin Yun Yuan Plamen Y. Yalamov Abstract A method is presented to make a given matrix strictly diagonally dominant

More information

Numerical Linear Algebra and. Image Restoration

Numerical Linear Algebra and. Image Restoration Numerical Linear Algebra and Image Restoration Maui High Performance Computing Center Wednesday, October 8, 2003 James G. Nagy Emory University Atlanta, GA Thanks to: AFOSR, Dave Tyler, Stuart Jefferies,

More information

The Solvability Conditions for the Inverse Eigenvalue Problem of Hermitian and Generalized Skew-Hamiltonian Matrices and Its Approximation

The Solvability Conditions for the Inverse Eigenvalue Problem of Hermitian and Generalized Skew-Hamiltonian Matrices and Its Approximation The Solvability Conditions for the Inverse Eigenvalue Problem of Hermitian and Generalized Skew-Hamiltonian Matrices and Its Approximation Zheng-jian Bai Abstract In this paper, we first consider the inverse

More information

Modified Gauss Seidel type methods and Jacobi type methods for Z-matrices

Modified Gauss Seidel type methods and Jacobi type methods for Z-matrices Linear Algebra and its Applications 7 (2) 227 24 www.elsevier.com/locate/laa Modified Gauss Seidel type methods and Jacobi type methods for Z-matrices Wen Li a,, Weiwei Sun b a Department of Mathematics,

More information

On the Local Convergence of an Iterative Approach for Inverse Singular Value Problems

On the Local Convergence of an Iterative Approach for Inverse Singular Value Problems On the Local Convergence of an Iterative Approach for Inverse Singular Value Problems Zheng-jian Bai Benedetta Morini Shu-fang Xu Abstract The purpose of this paper is to provide the convergence theory

More information

Tikhonov Regularization of Large Symmetric Problems

Tikhonov Regularization of Large Symmetric Problems NUMERICAL LINEAR ALGEBRA WITH APPLICATIONS Numer. Linear Algebra Appl. 2000; 00:1 11 [Version: 2000/03/22 v1.0] Tihonov Regularization of Large Symmetric Problems D. Calvetti 1, L. Reichel 2 and A. Shuibi

More information

Computer Vision & Digital Image Processing

Computer Vision & Digital Image Processing Computer Vision & Digital Image Processing Image Restoration and Reconstruction I Dr. D. J. Jackson Lecture 11-1 Image restoration Restoration is an objective process that attempts to recover an image

More information

The restarted QR-algorithm for eigenvalue computation of structured matrices

The restarted QR-algorithm for eigenvalue computation of structured matrices Journal of Computational and Applied Mathematics 149 (2002) 415 422 www.elsevier.com/locate/cam The restarted QR-algorithm for eigenvalue computation of structured matrices Daniela Calvetti a; 1, Sun-Mi

More information

RITZ VALUE BOUNDS THAT EXPLOIT QUASI-SPARSITY

RITZ VALUE BOUNDS THAT EXPLOIT QUASI-SPARSITY RITZ VALUE BOUNDS THAT EXPLOIT QUASI-SPARSITY ILSE C.F. IPSEN Abstract. Absolute and relative perturbation bounds for Ritz values of complex square matrices are presented. The bounds exploit quasi-sparsity

More information

Numerical solution of the eigenvalue problem for Hermitian Toeplitz-like matrices

Numerical solution of the eigenvalue problem for Hermitian Toeplitz-like matrices TR-CS-9-1 Numerical solution of the eigenvalue problem for Hermitian Toeplitz-like matrices Michael K Ng and William F Trench July 199 Joint Computer Science Technical Report Series Department of Computer

More information

Structured Linear Algebra Problems in Adaptive Optics Imaging

Structured Linear Algebra Problems in Adaptive Optics Imaging Structured Linear Algebra Problems in Adaptive Optics Imaging Johnathan M. Bardsley, Sarah Knepper, and James Nagy Abstract A main problem in adaptive optics is to reconstruct the phase spectrum given

More information

Atmospheric turbulence restoration by diffeomorphic image registration and blind deconvolution

Atmospheric turbulence restoration by diffeomorphic image registration and blind deconvolution Atmospheric turbulence restoration by diffeomorphic image registration and blind deconvolution Jérôme Gilles, Tristan Dagobert, Carlo De Franchis DGA/CEP - EORD department, 16bis rue Prieur de la Côte

More information

THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR

THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR WEN LI AND MICHAEL K. NG Abstract. In this paper, we study the perturbation bound for the spectral radius of an m th - order n-dimensional

More information

A fast randomized algorithm for overdetermined linear least-squares regression

A fast randomized algorithm for overdetermined linear least-squares regression A fast randomized algorithm for overdetermined linear least-squares regression Vladimir Rokhlin and Mark Tygert Technical Report YALEU/DCS/TR-1403 April 28, 2008 Abstract We introduce a randomized algorithm

More information

arxiv: v1 [math.na] 3 Jan 2019

arxiv: v1 [math.na] 3 Jan 2019 STRUCTURED FISTA FOR IMAGE RESTORATION ZIXUAN CHEN, JAMES G. NAGY, YUANZHE XI, AND BO YU arxiv:9.93v [math.na] 3 Jan 29 Abstract. In this paper, we propose an efficient numerical scheme for solving some

More information

MATH 590: Meshfree Methods

MATH 590: Meshfree Methods MATH 590: Meshfree Methods Chapter 34: Improving the Condition Number of the Interpolation Matrix Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Fall 2010 fasshauer@iit.edu

More information

Interval solutions for interval algebraic equations

Interval solutions for interval algebraic equations Mathematics and Computers in Simulation 66 (2004) 207 217 Interval solutions for interval algebraic equations B.T. Polyak, S.A. Nazin Institute of Control Sciences, Russian Academy of Sciences, 65 Profsoyuznaya

More information

Approximate Inverse-Free Preconditioners for Toeplitz Matrices

Approximate Inverse-Free Preconditioners for Toeplitz Matrices Approximate Inverse-Free Preconditioners for Toeplitz Matrices You-Wei Wen Wai-Ki Ching Michael K. Ng Abstract In this paper, we propose approximate inverse-free preconditioners for solving Toeplitz systems.

More information

Fast Angular Synchronization for Phase Retrieval via Incomplete Information

Fast Angular Synchronization for Phase Retrieval via Incomplete Information Fast Angular Synchronization for Phase Retrieval via Incomplete Information Aditya Viswanathan a and Mark Iwen b a Department of Mathematics, Michigan State University; b Department of Mathematics & Department

More information

The inverse of a tridiagonal matrix

The inverse of a tridiagonal matrix Linear Algebra and its Applications 325 (2001) 109 139 www.elsevier.com/locate/laa The inverse of a tridiagonal matrix Ranjan K. Mallik Department of Electrical Engineering, Indian Institute of Technology,

More information

1. Introduction. We consider the system of saddle point linear systems

1. Introduction. We consider the system of saddle point linear systems VALIDATED SOLUTIONS OF SADDLE POINT LINEAR SYSTEMS TAKUMA KIMURA AND XIAOJUN CHEN Abstract. We propose a fast verification method for saddle point linear systems where the (, block is singular. The proposed

More information

A fast nonstationary preconditioning strategy for ill-posed problems, with application to image deblurring

A fast nonstationary preconditioning strategy for ill-posed problems, with application to image deblurring A fast nonstationary preconditioning strategy for ill-posed problems, with application to image deblurring Marco Donatelli Dept. of Science and High Tecnology U. Insubria (Italy) Joint work with M. Hanke

More information

A Note on Eigenvalues of Perturbed Hermitian Matrices

A Note on Eigenvalues of Perturbed Hermitian Matrices A Note on Eigenvalues of Perturbed Hermitian Matrices Chi-Kwong Li Ren-Cang Li July 2004 Let ( H1 E A = E H 2 Abstract and à = ( H1 H 2 be Hermitian matrices with eigenvalues λ 1 λ k and λ 1 λ k, respectively.

More information

LINEARIZED BREGMAN ITERATIONS FOR FRAME-BASED IMAGE DEBLURRING

LINEARIZED BREGMAN ITERATIONS FOR FRAME-BASED IMAGE DEBLURRING LINEARIZED BREGMAN ITERATIONS FOR FRAME-BASED IMAGE DEBLURRING JIAN-FENG CAI, STANLEY OSHER, AND ZUOWEI SHEN Abstract. Real images usually have sparse approximations under some tight frame systems derived

More information

Iterative Krylov Subspace Methods for Sparse Reconstruction

Iterative Krylov Subspace Methods for Sparse Reconstruction Iterative Krylov Subspace Methods for Sparse Reconstruction James Nagy Mathematics and Computer Science Emory University Atlanta, GA USA Joint work with Silvia Gazzola University of Padova, Italy Outline

More information

Key words. Boundary conditions, fast transforms, matrix algebras and Toeplitz matrices, Tikhonov regularization, regularizing iterative methods.

Key words. Boundary conditions, fast transforms, matrix algebras and Toeplitz matrices, Tikhonov regularization, regularizing iterative methods. REGULARIZATION OF IMAGE RESTORATION PROBLEMS WITH ANTI-REFLECTIVE BOUNDARY CONDITIONS MARCO DONATELLI, CLAUDIO ESTATICO, AND STEFANO SERRA-CAPIZZANO Abstract. Anti-reflective boundary conditions have been

More information

BLOCK DIAGONAL AND SCHUR COMPLEMENT PRECONDITIONERS FOR BLOCK TOEPLITZ SYSTEMS WITH SMALL SIZE BLOCKS

BLOCK DIAGONAL AND SCHUR COMPLEMENT PRECONDITIONERS FOR BLOCK TOEPLITZ SYSTEMS WITH SMALL SIZE BLOCKS BLOCK DIAGONAL AND SCHUR COMPLEMENT PRECONDITIONERS FOR BLOCK TOEPLITZ SYSTEMS WITH SMALL SIZE BLOCKS WAI-KI CHING, MICHAEL K NG, AND YOU-WEI WEN Abstract In this paper we consider the solution of Hermitian

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Key words. conjugate gradients, normwise backward error, incremental norm estimation. Proceedings of ALGORITMY 2016 pp. 323 332 ON ERROR ESTIMATION IN THE CONJUGATE GRADIENT METHOD: NORMWISE BACKWARD ERROR PETR TICHÝ Abstract. Using an idea of Duff and Vömel [BIT, 42 (2002), pp. 300 322

More information

Tikhonov Regularization for Weighted Total Least Squares Problems

Tikhonov Regularization for Weighted Total Least Squares Problems Tikhonov Regularization for Weighted Total Least Squares Problems Yimin Wei Naimin Zhang Michael K. Ng Wei Xu Abstract In this paper, we study and analyze the regularized weighted total least squares (RWTLS)

More information

Linear Algebra and its Applications

Linear Algebra and its Applications Linear Algebra and its Applications 430 (2009) 532 543 Contents lists available at ScienceDirect Linear Algebra and its Applications journal homepage: wwwelseviercom/locate/laa Computing tight upper bounds

More information

Iterative Methods for Smooth Objective Functions

Iterative Methods for Smooth Objective Functions Optimization Iterative Methods for Smooth Objective Functions Quadratic Objective Functions Stationary Iterative Methods (first/second order) Steepest Descent Method Landweber/Projected Landweber Methods

More information

Mathematical Beer Goggles or The Mathematics of Image Processing

Mathematical Beer Goggles or The Mathematics of Image Processing How Mathematical Beer Goggles or The Mathematics of Image Processing Department of Mathematical Sciences University of Bath Postgraduate Seminar Series University of Bath 12th February 2008 1 How 2 How

More information

Phase-Correlation Motion Estimation Yi Liang

Phase-Correlation Motion Estimation Yi Liang EE 392J Final Project Abstract Phase-Correlation Motion Estimation Yi Liang yiliang@stanford.edu Phase-correlation motion estimation is studied and implemented in this work, with its performance, efficiency

More information

Signal Identification Using a Least L 1 Norm Algorithm

Signal Identification Using a Least L 1 Norm Algorithm Optimization and Engineering, 1, 51 65, 2000 c 2000 Kluwer Academic Publishers. Manufactured in The Netherlands. Signal Identification Using a Least L 1 Norm Algorithm J. BEN ROSEN Department of Computer

More information

For δa E, this motivates the definition of the Bauer-Skeel condition number ([2], [3], [14], [15])

For δa E, this motivates the definition of the Bauer-Skeel condition number ([2], [3], [14], [15]) LAA 278, pp.2-32, 998 STRUCTURED PERTURBATIONS AND SYMMETRIC MATRICES SIEGFRIED M. RUMP Abstract. For a given n by n matrix the ratio between the componentwise distance to the nearest singular matrix and

More information

Inverse Problems in Image Processing

Inverse Problems in Image Processing H D Inverse Problems in Image Processing Ramesh Neelamani (Neelsh) Committee: Profs. R. Baraniuk, R. Nowak, M. Orchard, S. Cox June 2003 Inverse Problems Data estimation from inadequate/noisy observations

More information

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH V. FABER, J. LIESEN, AND P. TICHÝ Abstract. Numerous algorithms in numerical linear algebra are based on the reduction of a given matrix

More information

Digital Image Processing

Digital Image Processing Digital Image Processing 2D SYSTEMS & PRELIMINARIES Hamid R. Rabiee Fall 2015 Outline 2 Two Dimensional Fourier & Z-transform Toeplitz & Circulant Matrices Orthogonal & Unitary Matrices Block Matrices

More information

8 The SVD Applied to Signal and Image Deblurring

8 The SVD Applied to Signal and Image Deblurring 8 The SVD Applied to Signal and Image Deblurring We will discuss the restoration of one-dimensional signals and two-dimensional gray-scale images that have been contaminated by blur and noise. After an

More information

On the Convergence of the Inverses of Toeplitz Matrices and Its Applications

On the Convergence of the Inverses of Toeplitz Matrices and Its Applications 180 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 49, NO. 1, JANUARY 2003 On the Convergence of the Inverses of Toeplitz Matrices Its Applications Feng-Wen Sun, Member, IEEE, Yimin Jiang, Member, IEEE,

More information

8 The SVD Applied to Signal and Image Deblurring

8 The SVD Applied to Signal and Image Deblurring 8 The SVD Applied to Signal and Image Deblurring We will discuss the restoration of one-dimensional signals and two-dimensional gray-scale images that have been contaminated by blur and noise. After an

More information

ETNA Kent State University

ETNA Kent State University Electronic Transactions on Numerical Analysis. Volume 31, pp. 204-220, 2008. Copyright 2008,. ISSN 1068-9613. ETNA NOISE PROPAGATION IN REGULARIZING ITERATIONS FOR IMAGE DEBLURRING PER CHRISTIAN HANSEN

More information

Exponential Decomposition and Hankel Matrix

Exponential Decomposition and Hankel Matrix Exponential Decomposition and Hankel Matrix Franklin T Luk Department of Computer Science and Engineering, Chinese University of Hong Kong, Shatin, NT, Hong Kong luk@csecuhkeduhk Sanzheng Qiao Department

More information

An estimation of the spectral radius of a product of block matrices

An estimation of the spectral radius of a product of block matrices Linear Algebra and its Applications 379 (2004) 267 275 wwwelseviercom/locate/laa An estimation of the spectral radius of a product of block matrices Mei-Qin Chen a Xiezhang Li b a Department of Mathematics

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information

Efficient and Accurate Rectangular Window Subspace Tracking

Efficient and Accurate Rectangular Window Subspace Tracking Efficient and Accurate Rectangular Window Subspace Tracking Timothy M. Toolan and Donald W. Tufts Dept. of Electrical Engineering, University of Rhode Island, Kingston, RI 88 USA toolan@ele.uri.edu, tufts@ele.uri.edu

More information

OUTLINE 1. Introduction 1.1 Notation 1.2 Special matrices 2. Gaussian Elimination 2.1 Vector and matrix norms 2.2 Finite precision arithmetic 2.3 Fact

OUTLINE 1. Introduction 1.1 Notation 1.2 Special matrices 2. Gaussian Elimination 2.1 Vector and matrix norms 2.2 Finite precision arithmetic 2.3 Fact Computational Linear Algebra Course: (MATH: 6800, CSCI: 6800) Semester: Fall 1998 Instructors: { Joseph E. Flaherty, aherje@cs.rpi.edu { Franklin T. Luk, luk@cs.rpi.edu { Wesley Turner, turnerw@cs.rpi.edu

More information

6 The SVD Applied to Signal and Image Deblurring

6 The SVD Applied to Signal and Image Deblurring 6 The SVD Applied to Signal and Image Deblurring We will discuss the restoration of one-dimensional signals and two-dimensional gray-scale images that have been contaminated by blur and noise. After an

More information

Interlacing Inequalities for Totally Nonnegative Matrices

Interlacing Inequalities for Totally Nonnegative Matrices Interlacing Inequalities for Totally Nonnegative Matrices Chi-Kwong Li and Roy Mathias October 26, 2004 Dedicated to Professor T. Ando on the occasion of his 70th birthday. Abstract Suppose λ 1 λ n 0 are

More information

EFFECTS OF ILL-CONDITIONED DATA ON LEAST SQUARES ADAPTIVE FILTERS. Gary A. Ybarra and S.T. Alexander

EFFECTS OF ILL-CONDITIONED DATA ON LEAST SQUARES ADAPTIVE FILTERS. Gary A. Ybarra and S.T. Alexander EFFECTS OF ILL-CONDITIONED DATA ON LEAST SQUARES ADAPTIVE FILTERS Gary A. Ybarra and S.T. Alexander Center for Communications and Signal Processing Electrical and Computer Engineering Department North

More information

Scientific Computing

Scientific Computing Scientific Computing Direct solution methods Martin van Gijzen Delft University of Technology October 3, 2018 1 Program October 3 Matrix norms LU decomposition Basic algorithm Cost Stability Pivoting Pivoting

More information

This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and

This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and education use, including for instruction at the authors institution

More information

AMS classification scheme numbers: 65F10, 65F15, 65Y20

AMS classification scheme numbers: 65F10, 65F15, 65Y20 Improved image deblurring with anti-reflective boundary conditions and re-blurring (This is a preprint of an article published in Inverse Problems, 22 (06) pp. 35-53.) M. Donatelli, C. Estatico, A. Martinelli,

More information

A fast and well-conditioned spectral method: The US method

A fast and well-conditioned spectral method: The US method A fast and well-conditioned spectral method: The US method Alex Townsend University of Oxford Leslie Fox Prize, 24th of June 23 Partially based on: S. Olver & T., A fast and well-conditioned spectral method,

More information

6.4 Krylov Subspaces and Conjugate Gradients

6.4 Krylov Subspaces and Conjugate Gradients 6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P

More information

Efficient boundary element analysis of periodic sound scatterers

Efficient boundary element analysis of periodic sound scatterers Boundary Element and Meshless Methods in Acoustics and Vibrations: Paper ICA2016-418 Efficient boundary element analysis of periodic sound scatterers M. Karimi, P. Croaker, N. Kessissoglou 1 School of

More information

Golub-Kahan iterative bidiagonalization and determining the noise level in the data

Golub-Kahan iterative bidiagonalization and determining the noise level in the data Golub-Kahan iterative bidiagonalization and determining the noise level in the data Iveta Hnětynková,, Martin Plešinger,, Zdeněk Strakoš, * Charles University, Prague ** Academy of Sciences of the Czech

More information

Integer Least Squares: Sphere Decoding and the LLL Algorithm

Integer Least Squares: Sphere Decoding and the LLL Algorithm Integer Least Squares: Sphere Decoding and the LLL Algorithm Sanzheng Qiao Department of Computing and Software McMaster University 28 Main St. West Hamilton Ontario L8S 4L7 Canada. ABSTRACT This paper

More information

Superlinear convergence for PCG using band plus algebra preconditioners for Toeplitz systems

Superlinear convergence for PCG using band plus algebra preconditioners for Toeplitz systems Computers and Mathematics with Applications 56 (008) 155 170 Contents lists available at ScienceDirect Computers and Mathematics with Applications journal homepage: www.elsevier.com/locate/camwa Superlinear

More information

Introduction. Chapter One

Introduction. Chapter One Chapter One Introduction The aim of this book is to describe and explain the beautiful mathematical relationships between matrices, moments, orthogonal polynomials, quadrature rules and the Lanczos and

More information

Main matrix factorizations

Main matrix factorizations Main matrix factorizations A P L U P permutation matrix, L lower triangular, U upper triangular Key use: Solve square linear system Ax b. A Q R Q unitary, R upper triangular Key use: Solve square or overdetrmined

More information

The Discrete Kalman Filtering of a Class of Dynamic Multiscale Systems

The Discrete Kalman Filtering of a Class of Dynamic Multiscale Systems 668 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II: ANALOG AND DIGITAL SIGNAL PROCESSING, VOL 49, NO 10, OCTOBER 2002 The Discrete Kalman Filtering of a Class of Dynamic Multiscale Systems Lei Zhang, Quan

More information

Dealing with edge effects in least-squares image deconvolution problems

Dealing with edge effects in least-squares image deconvolution problems Astronomy & Astrophysics manuscript no. bc May 11, 05 (DOI: will be inserted by hand later) Dealing with edge effects in least-squares image deconvolution problems R. Vio 1 J. Bardsley 2, M. Donatelli

More information

We first repeat some well known facts about condition numbers for normwise and componentwise perturbations. Consider the matrix

We first repeat some well known facts about condition numbers for normwise and componentwise perturbations. Consider the matrix BIT 39(1), pp. 143 151, 1999 ILL-CONDITIONEDNESS NEEDS NOT BE COMPONENTWISE NEAR TO ILL-POSEDNESS FOR LEAST SQUARES PROBLEMS SIEGFRIED M. RUMP Abstract. The condition number of a problem measures the sensitivity

More information

Sensitivity analysis of the differential matrix Riccati equation based on the associated linear differential system

Sensitivity analysis of the differential matrix Riccati equation based on the associated linear differential system Advances in Computational Mathematics 7 (1997) 295 31 295 Sensitivity analysis of the differential matrix Riccati equation based on the associated linear differential system Mihail Konstantinov a and Vera

More information

Structured Condition Numbers of Symmetric Algebraic Riccati Equations

Structured Condition Numbers of Symmetric Algebraic Riccati Equations Proceedings of the 2 nd International Conference of Control Dynamic Systems and Robotics Ottawa Ontario Canada May 7-8 2015 Paper No. 183 Structured Condition Numbers of Symmetric Algebraic Riccati Equations

More information

The effect on the algebraic connectivity of a tree by grafting or collapsing of edges

The effect on the algebraic connectivity of a tree by grafting or collapsing of edges Available online at www.sciencedirect.com Linear Algebra and its Applications 428 (2008) 855 864 www.elsevier.com/locate/laa The effect on the algebraic connectivity of a tree by grafting or collapsing

More information

Regularization methods for large-scale, ill-posed, linear, discrete, inverse problems

Regularization methods for large-scale, ill-posed, linear, discrete, inverse problems Regularization methods for large-scale, ill-posed, linear, discrete, inverse problems Silvia Gazzola Dipartimento di Matematica - Università di Padova January 10, 2012 Seminario ex-studenti 2 Silvia Gazzola

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline

More information

An iterative multigrid regularization method for Toeplitz discrete ill-posed problems

An iterative multigrid regularization method for Toeplitz discrete ill-posed problems NUMERICAL MATHEMATICS: Theory, Methods and Applications Numer. Math. Theor. Meth. Appl., Vol. xx, No. x, pp. 1-18 (200x) An iterative multigrid regularization method for Toeplitz discrete ill-posed problems

More information

c 2008 Society for Industrial and Applied Mathematics

c 2008 Society for Industrial and Applied Mathematics SIAM J. SCI. COMPUT. Vol. 30, No. 3, pp. 1430 1458 c 8 Society for Industrial and Applied Mathematics IMPROVEMENT OF SPACE-INVARIANT IMAGE DEBLURRING BY PRECONDITIONED LANDWEBER ITERATIONS PAOLA BRIANZI,

More information

Non-negative matrix factorization with fixed row and column sums

Non-negative matrix factorization with fixed row and column sums Available online at www.sciencedirect.com Linear Algebra and its Applications 9 (8) 5 www.elsevier.com/locate/laa Non-negative matrix factorization with fixed row and column sums Ngoc-Diep Ho, Paul Van

More information

Uniform Boundedness of a Preconditioned Normal Matrix Used in Interior Point Methods

Uniform Boundedness of a Preconditioned Normal Matrix Used in Interior Point Methods Uniform Boundedness of a Preconditioned Normal Matrix Used in Interior Point Methods Renato D. C. Monteiro Jerome W. O Neal Takashi Tsuchiya March 31, 2003 (Revised: December 3, 2003) Abstract Solving

More information

On the application of different numerical methods to obtain null-spaces of polynomial matrices. Part 1: block Toeplitz algorithms.

On the application of different numerical methods to obtain null-spaces of polynomial matrices. Part 1: block Toeplitz algorithms. On the application of different numerical methods to obtain null-spaces of polynomial matrices. Part 1: block Toeplitz algorithms. J.C. Zúñiga and D. Henrion Abstract Four different algorithms are designed

More information

Error estimates for the ESPRIT algorithm

Error estimates for the ESPRIT algorithm Error estimates for the ESPRIT algorithm Daniel Potts Manfred Tasche Let z j := e f j j = 1,..., M) with f j [ ϕ, 0] + i [ π, π) and small ϕ 0 be distinct nodes. With complex coefficients c j 0, we consider

More information

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2. APPENDIX A Background Mathematics A. Linear Algebra A.. Vector algebra Let x denote the n-dimensional column vector with components 0 x x 2 B C @. A x n Definition 6 (scalar product). The scalar product

More information

Arnoldi-Tikhonov regularization methods

Arnoldi-Tikhonov regularization methods Arnoldi-Tikhonov regularization methods Bryan Lewis a, Lothar Reichel b,,1 a Rocketcalc LLC, 100 W. Crain Ave., Kent, OH 44240, USA. b Department of Mathematical Sciences, Kent State University, Kent,

More information

Using Hankel structured low-rank approximation for sparse signal recovery

Using Hankel structured low-rank approximation for sparse signal recovery Using Hankel structured low-rank approximation for sparse signal recovery Ivan Markovsky 1 and Pier Luigi Dragotti 2 Department ELEC Vrije Universiteit Brussel (VUB) Pleinlaan 2, Building K, B-1050 Brussels,

More information

A Note on Inverse Iteration

A Note on Inverse Iteration A Note on Inverse Iteration Klaus Neymeyr Universität Rostock, Fachbereich Mathematik, Universitätsplatz 1, 18051 Rostock, Germany; SUMMARY Inverse iteration, if applied to a symmetric positive definite

More information

The Global Krylov subspace methods and Tikhonov regularization for image restoration

The Global Krylov subspace methods and Tikhonov regularization for image restoration The Global Krylov subspace methods and Tikhonov regularization for image restoration Abderrahman BOUHAMIDI (joint work with Khalide Jbilou) Université du Littoral Côte d Opale LMPA, CALAIS-FRANCE bouhamidi@lmpa.univ-littoral.fr

More information

On solving linear systems arising from Shishkin mesh discretizations

On solving linear systems arising from Shishkin mesh discretizations On solving linear systems arising from Shishkin mesh discretizations Petr Tichý Faculty of Mathematics and Physics, Charles University joint work with Carlos Echeverría, Jörg Liesen, and Daniel Szyld October

More information

arxiv: v1 [math.na] 1 Sep 2018

arxiv: v1 [math.na] 1 Sep 2018 On the perturbation of an L -orthogonal projection Xuefeng Xu arxiv:18090000v1 [mathna] 1 Sep 018 September 5 018 Abstract The L -orthogonal projection is an important mathematical tool in scientific computing

More information

On Solving Large Algebraic. Riccati Matrix Equations

On Solving Large Algebraic. Riccati Matrix Equations International Mathematical Forum, 5, 2010, no. 33, 1637-1644 On Solving Large Algebraic Riccati Matrix Equations Amer Kaabi Department of Basic Science Khoramshahr Marine Science and Technology University

More information