Lecture 1: Numerical Issues from Inverse Problems (Parameter Estimation, Regularization Theory, and Parallel Algorithms)

Size: px
Start display at page:

Download "Lecture 1: Numerical Issues from Inverse Problems (Parameter Estimation, Regularization Theory, and Parallel Algorithms)"

Transcription

1 Lecture 1: Numerical Issues from Inverse Problems (Parameter Estimation, Regularization Theory, and Parallel Algorithms) Youzuo Lin 1 Joint work with: Rosemary A. Renaut 2 Brendt Wohlberg 1 Hongbin Guo 2 1. Los Alamos National Laboratory 2. School of Mathematical and Statistical Sciences, Arizona State University Graduate Student Workshop on Inverse Problem, 2016

2 Outline 1 Introduction: Inverse Problem and Regularization 2 Topic 1: Multisplitting for Regularized Least Squares Regularization Parallelization Multiple Right Hand Side Problem 3 Topic 2: Total-Variation Regularization Numerical Methods for Total Variation Optimal Parameter Selection, UPRE Approach 4 Topic 3: Projected Krylov Subspace Solvers on GPU Fine-Grained Parallelism Model for Krylov Subspace Solvers on GPU Optimizations of Krylov Subspace Algorithms on GPU Numerical Results 5 References Inverse Theory Inverse Problems Workshop 1 / 64

3 Outline 1 Introduction: Inverse Problem and Regularization 2 Topic 1: Multisplitting for Regularized Least Squares Regularization Parallelization Multiple Right Hand Side Problem 3 Topic 2: Total-Variation Regularization Numerical Methods for Total Variation Optimal Parameter Selection, UPRE Approach 4 Topic 3: Projected Krylov Subspace Solvers on GPU Fine-Grained Parallelism Model for Krylov Subspace Solvers on GPU Optimizations of Krylov Subspace Algorithms on GPU Numerical Results 5 References Inverse Theory Inverse Problems Workshop 2 / 64

4 Inverse Problems and Ill-posedness General Linear Inverse Problem b = Af + η, Measured data b and system (or transform) matrix A Input Data f Inverse Theory Inverse Problems Workshop 2 / 64

5 Inverse Problems and Ill-posedness General Linear Inverse Problem b = Af + η, Measured data b and system (or transform) matrix A Input Data f Image Restoration - Deconvolution Transform matrix A is blurring matrix, η is the noise Inverse Theory Inverse Problems Workshop 2 / 64

6 Inverse Problems and Ill-posedness General Linear Inverse Problem b = Af + η, Measured data b and system (or transform) matrix A Input Data f Image Restoration - Deconvolution Transform matrix A is blurring matrix, η is the noise Image Reconstruction - Tomography Transform matrix A is Radon transform, η is the noise Inverse Theory Inverse Problems Workshop 2 / 64

7 Inverse Problems and Ill-posedness General Linear Inverse Problem b = Af + η, Measured data b and system (or transform) matrix A Input Data f Image Restoration - Deconvolution Transform matrix A is blurring matrix, η is the noise Image Reconstruction - Tomography Transform matrix A is Radon transform, η is the noise Inverse Theory Inverse Problems Workshop 2 / 64

8 Direct Inverse Filtering How about f = A 1 b? Inverse Theory Inverse Problems Workshop 3 / 64

9 Direct Inverse Filtering How about f = A 1 b? Figure: Left: Blurred and Noisy Image with SNR of (db), Right: Restored Image by Direct Inversion with SNR of (db) Inverse Theory Inverse Problems Workshop 3 / 64

10 Constrained Least Squares and Regularization Model Constrained Least Squares: min f {J(f)}, subject to Af b 2 2 = σ2. Regularization Model: { } min Af b 2 2 +λ2 J(f), λ > 0, f Af b 2 2 : Data Fidelity Term J(f) : Regularization Term λ : Regularization Parameter Inverse Theory Inverse Problems Workshop 4 / 64

11 Specific Regularization Tikhonov Regularization: When J(f) is a quadratic functional f T L T Lf, { } min Af b 2 2 +λ2 Lf 2 2, f Total Variation (TV) Regularization: When J(f) is a nonquadratic functional like Total Variation, { } min Af b 2 2 +λ2 f 1, f Others: Learning Based Regularization, Compressive Sensing etc. Inverse Theory Inverse Problems Workshop 5 / 64

12 Outline 1 Introduction: Inverse Problem and Regularization 2 Topic 1: Multisplitting for Regularized Least Squares Regularization Parallelization Multiple Right Hand Side Problem 3 Topic 2: Total-Variation Regularization Numerical Methods for Total Variation Optimal Parameter Selection, UPRE Approach 4 Topic 3: Projected Krylov Subspace Solvers on GPU Fine-Grained Parallelism Model for Krylov Subspace Solvers on GPU Optimizations of Krylov Subspace Algorithms on GPU Numerical Results 5 References Inverse Theory Inverse Problems Workshop 6 / 64

13 Topic 1: Multisplitting for Regularized Least Squares [Renaut, Lin and Guo, 2009] Regularization Parallelization Bottleneck for Applications Formulation-Multisplitting and Domain Decomposition Local Solver Selection and Computational Cost Update Scheme Multiple Right-Hand Sides Problem A Second Look at Multisplitting Hybrid Regularization with Multisplitting and Multiple Right Hand Sides Computational Cost Analysis Application in Image Reconstruction and Restoration Numerical Results Inverse Theory Inverse Problems Workshop 6 / 64

14 Tikhonov Solution and Memory Bottleneck Solving for the TK regularization is equivalent to solving for a large scale linear system: (A T A + λ 2 L T L)f = A T b. What about RAM requirement? In the positron emission tomography (PET) image reconstruction example in single precision, for an image of size , with projection angles from 0 o to 180 o with a 2 o increment, the required RAM for storing A is:: GB! What can we do about this? We apply Domain Decomposition idea to our problem. Inverse Theory Inverse Problems Workshop 7 / 64

15 Formulation - Multisplitting Linear multisplitting (LMS): Given a matrix A R n n and a collection of matrices M (i), N (i), E (i) R n n, i = 1,..., p satisfying 1 A = M (i) N (i) for each i, i = 1,..., p. 2 M (i) is regular (nonsingular), i = 1,..., p. 3 E (i) is a non-negative diagonal matrix, i = 1,..., p and p i=1 E (i) = I. Then the collection of triples (M (i), N (i), E (i) ), i = 1,..., p is called a multisplitting of A and the LMS method is defined by the iteration: x (k+1) = p E (i) (M (i) ) 1 (N (i) x (k) + fb). i=1 Inverse Theory Inverse Problems Workshop 8 / 64

16 Formulation - Domain Decomposition Domain decomposition is a particular method of multisplitting in which the submatrices M (i) are defined to be consistent with a partitioning of the domain: x = (x T 1, x T 2,..., x T p ) T. Inverse Theory Inverse Problems Workshop 9 / 64

17 Formulation - Domain Decomposition Domain decomposition is a particular method of multisplitting in which the submatrices M (i) are defined to be consistent with a partitioning of the domain: x = (x T 1, x T 2,..., x T p ) T. eg: Different Splitting Schemes Inverse Theory Inverse Problems Workshop 9 / 64

18 Formulation - Operator Split [Renaut 1998] Corresponding to different splitting of the image x, we will have splitting on kernel operator A as, A = (A 1, A 2,..., A p ), Inverse Theory Inverse Problems Workshop 10 / 64

19 Formulation - Operator Split [Renaut 1998] Corresponding to different splitting of the image x, we will have splitting on kernel operator A as, A = (A 1, A 2,..., A p ), The linear system Ax = b is replaced with the split systems A i y i = bi (x), where b i (x) = b j i A jx j = b Ax + A i x i. Inverse Theory Inverse Problems Workshop 10 / 64

20 Formulation - Local Problem Local Problem - Solving Ax = b turns into solving p subproblems, min A i y i b i (x) 2, 1 i p. y R n i Inverse Theory Inverse Problems Workshop 11 / 64

21 Formulation - Local Problem Local Problem - Solving Ax = b turns into solving p subproblems, min A i y i b i (x) 2, 1 i p. y R n i Update Scheme - The global solution update from local solutions at step K to step K + 1 is given by, x (k+1) = p i=1 τ (k+1) i (x local ) (k+1) i, where (x local ) (k+1) i = ((x (k) 1 )T,..., (x (k) i 1 )T, (y (k+1) i ) T, (x (k) i+1 )T,..., (x (k) p ) T ) T Inverse Theory Inverse Problems Workshop 11 / 64

22 Formulation - Regularized Multisplitting Tikhonov Regularization: } min { Ax b 2 x 2 +λ2 Lx 2 2, λ > 0, Inverse Theory Inverse Problems Workshop 12 / 64

23 Formulation - Regularized Multisplitting Tikhonov Regularization: } min { Ax b 2 x 2 +λ2 Lx 2 2, λ > 0, equivalent to, ( ) min A x LΛ ( ) b 2, 0 2 where Λ is block diagonal with entries λi ni. Inverse Theory Inverse Problems Workshop 12 / 64

24 Formulation - Regularized Multisplitting Similarly, we will have splitting on the operator as, ( ) (( ) ( ) ( )) A A1 A2 Ap =. LΛ λ 1 L 1 λ 2 L 2 λ p L p Inverse Theory Inverse Problems Workshop 13 / 64

25 Formulation - Regularized Multisplitting Similarly, we will have splitting on the operator as, ( ) (( ) ( ) ( )) A A1 A2 Ap =. LΛ λ 1 L 1 λ 2 L 2 λ p L p i th block equation of the normal equations is given by p (A T i A i + λ 2 i LT i L i )y i = A T i b i (x) λ i λ j L T i L j x j. j=1,j i Inverse Theory Inverse Problems Workshop 13 / 64

26 Local Solver Selection and Computational Cost How to select the local solver? Inverse Theory Inverse Problems Workshop 14 / 64

27 Local Solver Selection and Computational Cost How to select the local solver? Computation Cost and Memory Cost Inverse Theory Inverse Problems Workshop 14 / 64

28 Local Solver Selection and Computational Cost How to select the local solver? Computation Cost and Memory Cost Iterative Solver: Conjugate Gradient for Least Squares (CGLS) Inverse Theory Inverse Problems Workshop 14 / 64

29 Local Solver Selection and Computational Cost How to select the local solver? Computation Cost and Memory Cost Iterative Solver: Conjugate Gradient for Least Squares (CGLS) Computation Cost: C CGLS 2n 2 i m + K (2(k i + 1)n 2 i + 10k i n i ) Memory Cost: Matrix vector multiplication, only need to store matrix A i Inverse Theory Inverse Problems Workshop 14 / 64

30 Local Solver Selection and Computational Cost How to select the local solver? Computation Cost and Memory Cost Iterative Solver: Conjugate Gradient for Least Squares (CGLS) Computation Cost: C CGLS 2n 2 i m + K (2(k i + 1)n 2 i + 10k i n i ) Memory Cost: Matrix vector multiplication, only need to store matrix A i Solver: CGLS seems to be reasonable at this point! Inverse Theory Inverse Problems Workshop 14 / 64

31 Four Different Update Schemes Block Jacobian Update Set τ (k+1) i = τ i = 1, so x (k+1) = ((y (k) 1 )T, (y (k) 2 )T,..., (y (k) i 1 )T, (y (k+1) i ) T, (y (k) i+1 )T,..., (y (k) p ) T ) T. Inverse Theory Inverse Problems Workshop 15 / 64

32 Four Different Update Schemes Block Jacobian Update Set τ (k+1) i = τ i = 1, so x (k+1) = ((y (k) 1 )T, (y (k) 2 )T,..., (y (k) i 1 )T, (y (k+1) i ) T, (y (k) i+1 )T,..., (y (k) p ) T ) T. Convex Update Set τ (k+1) i = τ i = 1 p, x (k+1) = x (k) + 1 p (y(k+1) x (k) ). Inverse Theory Inverse Problems Workshop 15 / 64

33 Four Different Update Schemes Block Jacobian Update Set τ (k+1) i = τ i = 1, so x (k+1) = ((y (k) 1 )T, (y (k) 2 )T,..., (y (k) i 1 )T, (y (k+1) i ) T, (y (k) i+1 )T,..., (y (k) p ) T ) T. Convex Update Set τ (k+1) i = τ i = 1 p, x (k+1) = x (k) + 1 p (y(k+1) x (k) ). Local Update x (k+1) = (x local ) (k+1) i. Inverse Theory Inverse Problems Workshop 15 / 64

34 Four Different Update Schemes Block Jacobian Update Set τ (k+1) i = τ i = 1, so x (k+1) = ((y (k) 1 )T, (y (k) 2 )T,..., (y (k) i 1 )T, (y (k+1) i ) T, (y (k) i+1 )T,..., (y (k) p ) T ) T. Convex Update Set τ (k+1) i = τ i = 1 p, x (k+1) = x (k) + 1 p (y(k+1) x (k) ). Local Update x (k+1) = (x local ) (k+1) i. Optimal τ Update min τ r(x (k+1) ), where r(x (k+1) ) = r(x (k) ) p i=1 ( ) τ (k+1) Ai i δ (k+1) λ i L i. i Inverse Theory Inverse Problems Workshop 15 / 64

35 A Second Look at the Subproblem The i th block equation of the normal equations is given by p (A T i A i + λ 2 i LT i L i )y i = A T i b i (x) λ i λ j L T i L j x j. j=1,j i The system matrix is unchanged; Inverse Theory Inverse Problems Workshop 16 / 64

36 A Second Look at the Subproblem Rewrite the right hand side (RHS) as, p b i (x (k+1) ) = b i (x (k) ) where B (k+1) j = A j (x (k+1) j x k j overlapped problem. j=1,j i ), OB(k+1) i τ (k+1) j B (k+1) j + OB (k+1) is the update of the i, The new RHS is updated based on the previous RHS; Inverse Theory Inverse Problems Workshop 16 / 64

37 Linear System with Multiple RHS Multiple RHS Problem Setup AX = B = [b (1),..., b (s) ], where A R m n, X R n s and B R m s. Inverse Theory Inverse Problems Workshop 17 / 64

38 Linear System with Multiple RHS Multiple RHS Problem Setup AX = B = [b (1),..., b (s) ], where A R m n, X R n s and B R m s. What about solving s systems independently? Direct Method (LU): Major Computational Cost O(n 3 ) + s O(n 2 ); Iterative Method (CGLS): Major Computational Cost s O(n 2 ); Inverse Theory Inverse Problems Workshop 17 / 64

39 Linear System with Multiple RHS Multiple RHS Problem Setup AX = B = [b (1),..., b (s) ], where A R m n, X R n s and B R m s. What about solving s systems independently? Direct Method (LU): Major Computational Cost O(n 3 ) + s O(n 2 ); Iterative Method (CGLS): Major Computational Cost s O(n 2 ); Several algorithms have been proposed to speed up solving the above system. Inverse Theory Inverse Problems Workshop 17 / 64

40 Efficient Iterative Algorithm for Solving Multiple RHS If b (1),..., b (s) are random or completely orthogonal to each other, no way to speed up the algorithm! Inverse Theory Inverse Problems Workshop 18 / 64

41 Efficient Iterative Algorithm for Solving Multiple RHS If b (1),..., b (s) are random or completely orthogonal to each other, no way to speed up the algorithm! If b (1),..., b (s) are close to each other and share information Galerkin-Projection based Krylov subspace methods are proposed for an efficient solver. We choose CG as the specific Krylov subspace method proposed in [Chan and Wan, 1997], because it is well-supported theoretically. Inverse Theory Inverse Problems Workshop 18 / 64

42 Basic Idea of Projected CG Algorithm for Solving MRHS Step 1: Solve the seed system. Select the seed; Apply the CGLS algorithm for solving seed system and save the conjugate directions for Step 2; Inverse Theory Inverse Problems Workshop 19 / 64

43 Basic Idea of Projected CG Algorithm for Solving MRHS Step 1: Solve the seed system. Select the seed; Apply the CGLS algorithm for solving seed system and save the conjugate directions for Step 2; Step 2: Solve the remaining system by projection Use the conjugate directions from the seed systems to span the Krylov subspace Project the systems onto the Krylov subspace for an approximated solution Inverse Theory Inverse Problems Workshop 19 / 64

44 Basic Idea of Projected CG Algorithm for Solving MRHS Step 1: Solve the seed system. Select the seed; Apply the CGLS algorithm for solving seed system and save the conjugate directions for Step 2; Step 2: Solve the remaining system by projection Use the conjugate directions from the seed systems to span the Krylov subspace Project the systems onto the Krylov subspace for an approximated solution Step 3: Restart the seed and Refine the solution A new seed might be selected from the remaining systems to restart the procedure. Refine the approximation, if the accuracy is insufficient. Inverse Theory Inverse Problems Workshop 19 / 64

45 Basic Idea of Projected CG Algorithm for Solving MRHS Step 1: Solve the seed system. Select the seed; Apply the CGLS algorithm for solving seed system and save the conjugate directions for Step 2; Step 2: Solve the remaining system by projection Use the conjugate directions from the seed systems to span the Krylov subspace Project the systems onto the Krylov subspace for an approximated solution Step 3: Restart the seed and Refine the solution A new seed might be selected from the remaining systems to restart the procedure. Refine the approximation, if the accuracy is insufficient. Inverse Theory Inverse Problems Workshop 19 / 64

46 Adapt to Our Problem: Solving Each Subproblem from RLSMS, Scheme 1 Scheme 1 (Project/Restart): Treat the i th subproblem at all iteration as a Multiple RHS with s = 2, i.e. at iteration k = 1, solve the subproblem as seed, for iteration k > 1, solve it as projection and refine it as necessary; Inverse Theory Inverse Problems Workshop 20 / 64

47 Adapt to Our Problem: Solving Each Subproblem from RLSMS, Scheme 1 Scheme 1 (Project/Restart): Treat the i th subproblem at all iteration as a Multiple RHS with s = 2, i.e. at iteration k = 1, solve the subproblem as seed, for iteration k > 1, solve it as projection and refine it as necessary; Comment: Straight forward adaptation, but usually the accuracy of approximation remaining after projection is not enough; Need extra steps for further refinement; Inverse Theory Inverse Problems Workshop 20 / 64

48 Adapt to Our Problem: Solving Each Subproblem from RLSMS, Scheme 2 Scheme 2 (Project/Augment): Based on seed selection as in Scheme 1, in the refinement stages of solving the remaining systems, stores the newly generated conjugate directions with those from solving seeds; Inverse Theory Inverse Problems Workshop 21 / 64

49 Adapt to Our Problem: Solving Each Subproblem from RLSMS, Scheme 2 Scheme 2 (Project/Augment): Based on seed selection as in Scheme 1, in the refinement stages of solving the remaining systems, stores the newly generated conjugate directions with those from solving seeds; Comment: The motivation of this change is again to improve the Krylov subspace. By updating the conjugate directions, we will be able to add in new information which is not from solving the seed system. Numerical tests indicate this scheme works best with large scale problems! Inverse Theory Inverse Problems Workshop 21 / 64

50 Computational Cost Analysis The computation cost of projected CGLS consists of two other components: one is the cost for solving the seed system, which is the cost of the CGLS; the other is the cost for solving the remaining systems, which includes the projection and the few more steps of CGLS as needed: C ProjCG 2(K + 2k i + 2)mn i + (6(K 1)k i + 10k i + 1)n i + µ CG. Inverse Theory Inverse Problems Workshop 22 / 64

51 1-D Phillips, Restoration Result Figure: Restoration results using different methods for the sample of Phillips size of , noise level 6%. Inverse Theory Inverse Problems Workshop 23 / 64

52 1-D Phillips, Global and Local Iteration Result Figure: Maximum numbers of inner iterations against outer iteration count. Inverse Theory Inverse Problems Workshop 24 / 64

53 2-D Seismic Reconstruction, Phantom Size Figure: SNR (Without Splitting): db, SNR (CGLS/Zero): db, SNR (Project/Restart): db, SNR (Project/Augment): db. Inverse Theory Inverse Problems Workshop 25 / 64

54 2-D Seismic Reconstruction, Global and Local Iteration Result Figure: Maximum numbers of inner iterations against outer iteration count. Inverse Theory Inverse Problems Workshop 26 / 64

55 Outline 1 Introduction: Inverse Problem and Regularization 2 Topic 1: Multisplitting for Regularized Least Squares Regularization Parallelization Multiple Right Hand Side Problem 3 Topic 2: Total-Variation Regularization Numerical Methods for Total Variation Optimal Parameter Selection, UPRE Approach 4 Topic 3: Projected Krylov Subspace Solvers on GPU Fine-Grained Parallelism Model for Krylov Subspace Solvers on GPU Optimizations of Krylov Subspace Algorithms on GPU Numerical Results 5 References Inverse Theory Inverse Problems Workshop 27 / 64

56 Topic 2: Total-Variation Regularization Numerical Methods for Total Variation Iteratively Reweighted Norm Algorithm Lagged Diffusivity Fixed Point Iteration Algorithm Optimal Regularization Parameter Selection, Trace Approximation UPRE Approach Numerical Results Inverse Theory Inverse Problems Workshop 27 / 64

57 Iteratively Reweighted Norm Algorithm Approximate L1 norm by weighted L2 norm, TV regularization becomes: { } 1 min f 2 Af b 2 2 +λ 2 W 1/2 r Df 2 2, ( ) (( A min f { 1 2 I 0 0 W 1/2 r λd ) f ( )) } b 0 2 2, where W r = diag(2((d x f ) 2 + (D y f ) 2 ) 1 2 ). This minimization problem can be solved by CG solver. Ref: B. Wohlberg, P. Rodrguez, An Iteratively Reweighted Norm Algorithm for Minimization of Total Variation Functionals, IEEE Signal Processing Letters, 2007 Inverse Theory Inverse Problems Workshop 28 / 64

58 Lagged Diffusivity Fixed Point Iteration Algorithm Approximate L1 norm by, ψ(f ) = ( df dx + β2 ) Gradient = A T (Af b) + λl(f )f, Approximate Hessian = A T A + λl(f ), where L(f ) = xd T diag(ψ (f ))D. Based on the Quasi-Newton idea, we solve for f iteratively by, f k+1 = f k (A T A + λl(f k )) 1 Gradient, which is called as Lagged Diffusivity Fixed Point Iteration. Ref: C.R. Vogel, M. E. Oman, Iterative Methods for Total Variation, SIAM J. on Sci. Comp, 17(1996) Inverse Theory Inverse Problems Workshop 29 / 64

59 Image Deblurring Comparison using TV and Tikhonov Figure: Left: Original Clean Image, Right: Blurred and Noisy Image Inverse Theory Inverse Problems Workshop 30 / 64

60 Image Deblurring Comparison using TV and Tikhonov Figure: Left: TK based Deblurred Result, Right: TV Based Deblurred Result Inverse Theory Inverse Problems Workshop 31 / 64

61 Problem Description-Simple Example Inverse Theory Inverse Problems Workshop 32 / 64

62 Problem Description-Simple Example Inverse Theory Inverse Problems Workshop 32 / 64

63 Problem Description-Simple Example Inverse Theory Inverse Problems Workshop 32 / 64

64 Problem Description-Simple Example Inverse Theory Inverse Problems Workshop 32 / 64

65 Problem Description-Simple Example Inverse Theory Inverse Problems Workshop 32 / 64

66 Problem Description-Simple Example Inverse Theory Inverse Problems Workshop 32 / 64

67 Problem Description-Simple Example Inverse Theory Inverse Problems Workshop 32 / 64

68 Problem Description-Simple Example Inverse Theory Inverse Problems Workshop 32 / 64

69 Problem Description-Simple Example Inverse Theory Inverse Problems Workshop 32 / 64

70 Problem Description-Simple Example Inverse Theory Inverse Problems Workshop 32 / 64

71 Problem Description-Simple Example Inverse Theory Inverse Problems Workshop 32 / 64

72 Problem Description-Simple Example Inverse Theory Inverse Problems Workshop 32 / 64

73 Problem Description-Simple Example Inverse Theory Inverse Problems Workshop 32 / 64

74 Fundamental Idea of UPRE Derived from the Mean Squared Error (MSE) of the predictive error, 1 n P λ 2 = 1 n Af λ Af true 2, where f λ is the regularized solution, f true is the true solution. To avoid using f true, we define a functional which is the unbiased estimator of the above, and name it UPRE (Unbiased Predictive Risk Estimator). λ opt = arg min UPRE(λ) λ Inverse Theory Inverse Problems Workshop 33 / 64

75 UPRE for Tikhonov Vogel in [1] proved that, because the solution for Tikhonov Regularization is linearly dependent on the right hand side of (2), i.e., f TK = R TK (b + η), where R TK = (A T A + λi) 1 A T, we will have UPRE Tikhonov (λ) = E( 1 n P λ 2 ) = 1 n r λ 2 + 2σ2 n trace(a λ) σ 2, where r λ = Af λ b, A λ = A(A T A + λi) 1 A T, and σ is the variance of the noise. Inverse Theory Inverse Problems Workshop 34 / 64

76 UPRE for TV Difficulty 1: Nonquadratic penalty functional in TV term. Hint: Linearization Approximation! In Lagged Diffusivity approach, f TV = R TV (b + η), where R TV = (A T A + λ 2 L(f k)) 1 A T the influence matrix, A λ = A(A T A + λ 2 L(f k)) 1 A T. A λ can be proved to be symmetric, which will give us the same derivation of the UPRE functional for TV. Inverse Theory Inverse Problems Workshop 35 / 64

77 UPRE for TV Difficulty 2: How to compute trace(a λ ) = trace(a(a T A + λ 2 L(f k)) 1 A T ) Hint: Krylov Space Method! Basic Idea : trace(f (M)) E(u T f (M)u), where u is a discrete multivariate random variable, where each entry takes the values -1 and +1 with probability 0.5, and the matrix M is symmetric positive definite (SPD). Inverse Theory Inverse Problems Workshop 36 / 64

78 UPRE for TV Golub [2] pointed out that, E(u T f (M)u) = k ω i f (θ i ), i=1 where θ i are the eigenvalues of M, and ω i are squares of the first components of the normalized eigenvectors of M. Start the Lanczos procedure on M, and find out the most significant eigenvectors until accuracy is satisfied. Inverse Theory Inverse Problems Workshop 37 / 64

79 Trace Approximation Results Figure: Left: Accurate Trace Vs. Estimated Trace, Right: Time Cost for Computing Accurate Trace Vs. Time Cost for Computing Estimated Trace Inverse Theory Inverse Problems Workshop 38 / 64

80 TV Based Deblurring Using Optimal and Random Parameters Figure: Left: Use Optimal Parameter, Right: Use Random Parameter Inverse Theory Inverse Problems Workshop 39 / 64

81 More General Results for Deblurring Figure: Plot of Optimal SNR and Estimated SNR versus original noisy SNR on satellite image. Inverse Theory Inverse Problems Workshop 40 / 64

82 More General Results for Denoising Figure: Plot of Optimal SNR and Estimated SNR versus original noisy SNR on satellite image. Ref: G. Gilboa, N. Sochen, Y. Y. Zeevi, Estimation of optimal PDE-based denoising in the SNR sense, IEEE Transactions on Image Processing 15(8), Inverse Theory Inverse Problems Workshop 41 / 64

83 Conclusions Tikhonov Regularization: Pros : easy to implement, cheap cost; Cons : the result is usually smoothed out; Total Variation (TV) Regularization: Pros : Preserve image edges provided with reasonable parameter, piecewise constant Cons : Expensive cost Inverse Theory Inverse Problems Workshop 42 / 64

84 Outline 1 Introduction: Inverse Problem and Regularization 2 Topic 1: Multisplitting for Regularized Least Squares Regularization Parallelization Multiple Right Hand Side Problem 3 Topic 2: Total-Variation Regularization Numerical Methods for Total Variation Optimal Parameter Selection, UPRE Approach 4 Topic 3: Projected Krylov Subspace Solvers on GPU Fine-Grained Parallelism Model for Krylov Subspace Solvers on GPU Optimizations of Krylov Subspace Algorithms on GPU Numerical Results 5 References Inverse Theory Inverse Problems Workshop 43 / 64

85 Topic 3: Projected Krylov Subspace Solvers on GPU [Lin and Renaut, 2010] Problem Description and Krylov Subspace Solvers 3-D Medical Image Reconstruction Inverse Linear Systems with Multiple Right-hand Sides Conjugate Gradient Algorithm Fine-Grained Parallelism Model for Krylov Subspace Solvers on GPU GPU Structure BLAS Performance on GPU CG on GPU (ver 0) Optimizations of Krylov Subspace Algorithms on GPU Projected CG Modified Projected CG Numerical Results Inverse Theory Inverse Problems Workshop 43 / 64

86 3-D Medical Image Reconstruction Figure: Image Slices from a 3-D Scan Inverse Theory Inverse Problems Workshop 44 / 64

87 Inverse Linear Systems with Multiple Right-hand Sides Multiple Right-hand Sides System: Ax = B = [b (1),... b (s) ], where b i stands for the i th measurement with respect to the i th slice as in a 3-D medical image. Tikhonov regularization is applied to the above systems. Regularization parameter is assumed to be close enough to the optimal value. Inverse Theory Inverse Problems Workshop 45 / 64

88 Conjugate Gradient Algorithm k = 0, r 0 = b Ax 0 ; while ( r k 2 > TOL b 2 ) k = k + 1; if (k == 1) p 1 = r 0 ; else β k =< r k 1, r k 1 > / < r k 2, r k 2 >; p k = r k 1 + β k p k 1 ; end; α k =< r k 1, r k 1 > / < p k, Ap k >; x k = x k 1 + α k p k ; r k = r k 1 α k Ap k ; end (while). Inverse Theory Inverse Problems Workshop 46 / 64

89 GPU Structure - NVIDIA Tesla C1060 Figure: NVIDIA Tesla C1060 Inverse Theory Inverse Problems Workshop 47 / 64

90 GPU Structure - NVIDIA Tesla C1060 Streaming Multiprocessors (SM): 30 Streaming Processors (SP): 8 Register: 16 K per SM Shared Memory: 16 KB per SM Constant Memory: 64 KB Global Memory: 4 GB Bandwidth: 102 GB/Sec GFLOPS (PEAK) Single Precision: 933 Double Precision: 78 Figure: Structure of NVIDIA Tesla C1060 Inverse Theory Inverse Problems Workshop 48 / 64

91 BLAS Performance on GPU - Matrix Vector Multiplication Testing Environment Operating System: Ubuntu, 64 bit CPU (Host): Intel Quad Core Xeon E5506, 2.13 GHz GPU (Device): NVIDIA Tesla C1060 Programming on Host: Matlab using Single / Multi Core(s) Programming on Device: CUDA + CUBLAS, Matlab + Jacket Inverse Theory Inverse Problems Workshop 49 / 64

92 BLAS Performance on GPU - Matrix Vector Multiplication Testing matrices are selected from Matrix Market Benchmark Prob. Size bcsstm bus 1138 bcsstm bcsstk s1rmt3m bcsstk Inverse Theory Inverse Problems Workshop 50 / 64

93 BLAS Performance on GPU - Matrix Vector Multiplication Figure: Time Cost Figure: Speedup Ratio Inverse Theory Inverse Problems Workshop 51 / 64

94 BLAS Performance on GPU - Matrix Vector Multiplication Figure: Relative Error Inverse Theory Inverse Problems Workshop 52 / 64

95 BLAS Performance on GPU - BLAS 1, BLAS 2 and BLAS 3 Figure: GFLOPS Comparison Comment: BLAS 3 outperforms BLAS 1 and BLAS 2 significantly. Inverse Theory Inverse Problems Workshop 53 / 64

96 CG on GPU (ver 0) A synthetic problem setup using Chan s example [Chan and Wan, 1997] A = diag(1,..., n), n = b i (t) = sin(t + i t), i = 1,..., n, and t = 2π/100. TOL = Results: CG on Host: sec on a single core or sec on a multi-core. CG on Device: 5.49 sec. Inverse Theory Inverse Problems Workshop 54 / 64

97 CG on GPU (ver 0) - A closer look Comment: Mat-Vec is too expensive! Total Time: 5.49 sec; Mat-Vec: 3.98 sec (72.5%); Dot-Prod: 0.57 sec (10.4%); SAXPY: 0.16 sec (2.9%); Inverse Theory Inverse Problems Workshop 55 / 64

98 Projected CG (PrCG) - Algorithm By applying Lanczos-Galerkin projection method to CG, we have the PrCG [Chan and Wan, 1997]. i = 0; r (q,1) i = b (q) Ax (q,1) i ; for i = 1 to k α (q,1) i x (q,1) i r (q,1) = < p(1) i, r (q,1) i > < p (1) i, Ap (1) i > ; = x (q,1) i i+1 = r(q,1) i + α (q,1) i p (1) i ; α (q,1) i Ap (1) i ; end (for); % Restart the system if needed; Inverse Theory Inverse Problems Workshop 56 / 64

99 Projected CG (PrCG) Total Time: 1.91 sec; Mat-Vec: 1.01 sec (52.8%); Dot-Prod: 0.45 sec (23.6%); SAXPY: 0.29 sec (15.2%); Comment: SAXPY and Dot-Prod are becoming relatively expensive. Inverse Theory Inverse Problems Workshop 57 / 64

100 Augmented Projected CG (APrCG) Using the orthogonality properties that < p (1) j, Ap (1) i > = 0, j i, and < p (1) j, r (q,1) i > = 0, j < i, take the dot product with p (1) on both sides of r (q,1) i+1 = b(q) k α (q,1) i Ap (1) i i=1 i α (q,1) i, we can rewrite α i as, = < p(1) i, b (q) > < p (1) i, Ap (1) i >. Inverse Theory Inverse Problems Workshop 58 / 64

101 Augmented Projected CG (APrCG) - Algorithm Let P k = [p 1,..., p k ] and AP k = [Ap 1,..., Ap k ]. i = 0; r (q,1) i = b (q) Ax (q,1) i ; α =< P k, r (q,1) 0 >./diagvec(< P k, AP k >); Λ = diag(α); x (q,1) = x (q,1) 0 + sum(p k Λ); r (q,1) = r (q,1) 0 sum(ap k Λ); Inverse Theory Inverse Problems Workshop 59 / 64

102 Augmented Projected CG (APrCG) Figure: PrCG Figure: APrCG Inverse Theory Inverse Problems Workshop 60 / 64

103 Numerical Test / 3D Shepp-Logan Phantom Figure: Slice Show of A 3D Shepp-Logan Phantom Inverse Theory Inverse Problems Workshop 61 / 64

104 3D Shepp-Logan Phantom - Results Slice CG PrCG APrCG Cost SNR Cost SNR Cost SNR ImpRate % % % % Table: Reconstruction of Four Consecutive Slices from 65 to 68. The 64 th slice is selected as seed. Inverse Theory Inverse Problems Workshop 62 / 64

105 3D Shepp-Logan Phantom - Results Figure: Reconstruction Results on Host and Device, slice index = 66. Inverse Theory Inverse Problems Workshop 63 / 64

106 Outline 1 Introduction: Inverse Problem and Regularization 2 Topic 1: Multisplitting for Regularized Least Squares Regularization Parallelization Multiple Right Hand Side Problem 3 Topic 2: Total-Variation Regularization Numerical Methods for Total Variation Optimal Parameter Selection, UPRE Approach 4 Topic 3: Projected Krylov Subspace Solvers on GPU Fine-Grained Parallelism Model for Krylov Subspace Solvers on GPU Optimizations of Krylov Subspace Algorithms on GPU Numerical Results 5 References Inverse Theory Inverse Problems Workshop 64 / 64

107 References 1 R. A. Renaut, Y. Lin, and H. Guo. Solving Least Squares Problems by Using a Regularized Multisplitting Method. Numerical Linear Algebra with applications, vol. 19, issue. 4, Y. Lin, B. Wohlberg, and H. Guo. UPRE method for total variation parameter selection. Signal processing, vol. 90, no. 8, Inverse Theory Inverse Problems Workshop 64 / 64

UPRE Method for Total Variation Parameter Selection

UPRE Method for Total Variation Parameter Selection UPRE Method for Total Variation Parameter Selection Youzuo Lin School of Mathematical and Statistical Sciences, Arizona State University, Tempe, AZ 85287 USA. Brendt Wohlberg 1, T-5, Los Alamos National

More information

Towards Improved Sensitivity in Feature Extraction from Signals: one and two dimensional

Towards Improved Sensitivity in Feature Extraction from Signals: one and two dimensional Towards Improved Sensitivity in Feature Extraction from Signals: one and two dimensional Rosemary Renaut, Hongbin Guo, Jodi Mead and Wolfgang Stefan Supported by Arizona Alzheimer s Research Center and

More information

SIGNAL AND IMAGE RESTORATION: SOLVING

SIGNAL AND IMAGE RESTORATION: SOLVING 1 / 55 SIGNAL AND IMAGE RESTORATION: SOLVING ILL-POSED INVERSE PROBLEMS - ESTIMATING PARAMETERS Rosemary Renaut http://math.asu.edu/ rosie CORNELL MAY 10, 2013 2 / 55 Outline Background Parameter Estimation

More information

Regularization Parameter Estimation for Least Squares: A Newton method using the χ 2 -distribution

Regularization Parameter Estimation for Least Squares: A Newton method using the χ 2 -distribution Regularization Parameter Estimation for Least Squares: A Newton method using the χ 2 -distribution Rosemary Renaut, Jodi Mead Arizona State and Boise State September 2007 Renaut and Mead (ASU/Boise) Scalar

More information

Scaled gradient projection methods in image deblurring and denoising

Scaled gradient projection methods in image deblurring and denoising Scaled gradient projection methods in image deblurring and denoising Mario Bertero 1 Patrizia Boccacci 1 Silvia Bonettini 2 Riccardo Zanella 3 Luca Zanni 3 1 Dipartmento di Matematica, Università di Genova

More information

The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation

The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation Rosemary Renaut Collaborators: Jodi Mead and Iveta Hnetynkova DEPARTMENT OF MATHEMATICS

More information

Newton s Method for Estimating the Regularization Parameter for Least Squares: Using the Chi-curve

Newton s Method for Estimating the Regularization Parameter for Least Squares: Using the Chi-curve Newton s Method for Estimating the Regularization Parameter for Least Squares: Using the Chi-curve Rosemary Renaut, Jodi Mead Arizona State and Boise State Copper Mountain Conference on Iterative Methods

More information

The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation

The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation Rosemary Renaut DEPARTMENT OF MATHEMATICS AND STATISTICS Prague 2008 MATHEMATICS AND STATISTICS

More information

Block AIR Methods. For Multicore and GPU. Per Christian Hansen Hans Henrik B. Sørensen. Technical University of Denmark

Block AIR Methods. For Multicore and GPU. Per Christian Hansen Hans Henrik B. Sørensen. Technical University of Denmark Block AIR Methods For Multicore and GPU Per Christian Hansen Hans Henrik B. Sørensen Technical University of Denmark Model Problem and Notation Parallel-beam 3D tomography exact solution exact data noise

More information

Linear Inverse Problems

Linear Inverse Problems Linear Inverse Problems Ajinkya Kadu Utrecht University, The Netherlands February 26, 2018 Outline Introduction Least-squares Reconstruction Methods Examples Summary Introduction 2 What are inverse problems?

More information

A Newton-Galerkin-ADI Method for Large-Scale Algebraic Riccati Equations

A Newton-Galerkin-ADI Method for Large-Scale Algebraic Riccati Equations A Newton-Galerkin-ADI Method for Large-Scale Algebraic Riccati Equations Peter Benner Max-Planck-Institute for Dynamics of Complex Technical Systems Computational Methods in Systems and Control Theory

More information

Statistically-Based Regularization Parameter Estimation for Large Scale Problems

Statistically-Based Regularization Parameter Estimation for Large Scale Problems Statistically-Based Regularization Parameter Estimation for Large Scale Problems Rosemary Renaut Joint work with Jodi Mead and Iveta Hnetynkova March 1, 2010 National Science Foundation: Division of Computational

More information

Compressive Imaging by Generalized Total Variation Minimization

Compressive Imaging by Generalized Total Variation Minimization 1 / 23 Compressive Imaging by Generalized Total Variation Minimization Jie Yan and Wu-Sheng Lu Department of Electrical and Computer Engineering University of Victoria, Victoria, BC, Canada APCCAS 2014,

More information

Inverse problems Total Variation Regularization Mark van Kraaij Casa seminar 23 May 2007 Technische Universiteit Eindh ove n University of Technology

Inverse problems Total Variation Regularization Mark van Kraaij Casa seminar 23 May 2007 Technische Universiteit Eindh ove n University of Technology Inverse problems Total Variation Regularization Mark van Kraaij Casa seminar 23 May 27 Introduction Fredholm first kind integral equation of convolution type in one space dimension: g(x) = 1 k(x x )f(x

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline

More information

M.A. Botchev. September 5, 2014

M.A. Botchev. September 5, 2014 Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev

More information

A Hybrid LSQR Regularization Parameter Estimation Algorithm for Large Scale Problems

A Hybrid LSQR Regularization Parameter Estimation Algorithm for Large Scale Problems A Hybrid LSQR Regularization Parameter Estimation Algorithm for Large Scale Problems Rosemary Renaut Joint work with Jodi Mead and Iveta Hnetynkova SIAM Annual Meeting July 10, 2009 National Science Foundation:

More information

Algebraic Multigrid as Solvers and as Preconditioner

Algebraic Multigrid as Solvers and as Preconditioner Ò Algebraic Multigrid as Solvers and as Preconditioner Domenico Lahaye domenico.lahaye@cs.kuleuven.ac.be http://www.cs.kuleuven.ac.be/ domenico/ Department of Computer Science Katholieke Universiteit Leuven

More information

Chapter 7 Iterative Techniques in Matrix Algebra

Chapter 7 Iterative Techniques in Matrix Algebra Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition

More information

An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84

An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84 An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84 Introduction Almost all numerical methods for solving PDEs will at some point be reduced to solving A

More information

10. Multi-objective least squares

10. Multi-objective least squares L Vandenberghe ECE133A (Winter 2018) 10 Multi-objective least squares multi-objective least squares regularized data fitting control estimation and inversion 10-1 Multi-objective least squares we have

More information

Computational Linear Algebra

Computational Linear Algebra Computational Linear Algebra PD Dr. rer. nat. habil. Ralf Peter Mundani Computation in Engineering / BGU Scientific Computing in Computer Science / INF Winter Term 2017/18 Part 3: Iterative Methods PD

More information

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying I.2 Quadratic Eigenvalue Problems 1 Introduction The quadratic eigenvalue problem QEP is to find scalars λ and nonzero vectors u satisfying where Qλx = 0, 1.1 Qλ = λ 2 M + λd + K, M, D and K are given

More information

DUAL REGULARIZED TOTAL LEAST SQUARES SOLUTION FROM TWO-PARAMETER TRUST-REGION ALGORITHM. Geunseop Lee

DUAL REGULARIZED TOTAL LEAST SQUARES SOLUTION FROM TWO-PARAMETER TRUST-REGION ALGORITHM. Geunseop Lee J. Korean Math. Soc. 0 (0), No. 0, pp. 1 0 https://doi.org/10.4134/jkms.j160152 pissn: 0304-9914 / eissn: 2234-3008 DUAL REGULARIZED TOTAL LEAST SQUARES SOLUTION FROM TWO-PARAMETER TRUST-REGION ALGORITHM

More information

Accelerated Dual Gradient-Based Methods for Total Variation Image Denoising/Deblurring Problems (and other Inverse Problems)

Accelerated Dual Gradient-Based Methods for Total Variation Image Denoising/Deblurring Problems (and other Inverse Problems) Accelerated Dual Gradient-Based Methods for Total Variation Image Denoising/Deblurring Problems (and other Inverse Problems) Donghwan Kim and Jeffrey A. Fessler EECS Department, University of Michigan

More information

Incomplete Cholesky preconditioners that exploit the low-rank property

Incomplete Cholesky preconditioners that exploit the low-rank property anapov@ulb.ac.be ; http://homepages.ulb.ac.be/ anapov/ 1 / 35 Incomplete Cholesky preconditioners that exploit the low-rank property (theory and practice) Artem Napov Service de Métrologie Nucléaire, Université

More information

Enhanced Compressive Sensing and More

Enhanced Compressive Sensing and More Enhanced Compressive Sensing and More Yin Zhang Department of Computational and Applied Mathematics Rice University, Houston, Texas, U.S.A. Nonlinear Approximation Techniques Using L1 Texas A & M University

More information

Accelerating linear algebra computations with hybrid GPU-multicore systems.

Accelerating linear algebra computations with hybrid GPU-multicore systems. Accelerating linear algebra computations with hybrid GPU-multicore systems. Marc Baboulin INRIA/Université Paris-Sud joint work with Jack Dongarra (University of Tennessee and Oak Ridge National Laboratory)

More information

Mini-project in scientific computing

Mini-project in scientific computing Mini-project in scientific computing Eran Treister Computer Science Department, Ben-Gurion University of the Negev, Israel. March 7, 2018 1 / 30 Scientific computing Involves the solution of large computational

More information

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A. AMSC/CMSC 661 Scientific Computing II Spring 2005 Solution of Sparse Linear Systems Part 2: Iterative methods Dianne P. O Leary c 2005 Solving Sparse Linear Systems: Iterative methods The plan: Iterative

More information

A memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration

A memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration A memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration E. Chouzenoux, A. Jezierska, J.-C. Pesquet and H. Talbot Université Paris-Est Lab. d Informatique Gaspard

More information

Algorithm for Sparse Approximate Inverse Preconditioners in the Conjugate Gradient Method

Algorithm for Sparse Approximate Inverse Preconditioners in the Conjugate Gradient Method Algorithm for Sparse Approximate Inverse Preconditioners in the Conjugate Gradient Method Ilya B. Labutin A.A. Trofimuk Institute of Petroleum Geology and Geophysics SB RAS, 3, acad. Koptyug Ave., Novosibirsk

More information

A fast nonstationary preconditioning strategy for ill-posed problems, with application to image deblurring

A fast nonstationary preconditioning strategy for ill-posed problems, with application to image deblurring A fast nonstationary preconditioning strategy for ill-posed problems, with application to image deblurring Marco Donatelli Dept. of Science and High Tecnology U. Insubria (Italy) Joint work with M. Hanke

More information

Statistical Geometry Processing Winter Semester 2011/2012

Statistical Geometry Processing Winter Semester 2011/2012 Statistical Geometry Processing Winter Semester 2011/2012 Linear Algebra, Function Spaces & Inverse Problems Vector and Function Spaces 3 Vectors vectors are arrows in space classically: 2 or 3 dim. Euclidian

More information

Conjugate Gradient algorithm. Storage: fixed, independent of number of steps.

Conjugate Gradient algorithm. Storage: fixed, independent of number of steps. Conjugate Gradient algorithm Need: A symmetric positive definite; Cost: 1 matrix-vector product per step; Storage: fixed, independent of number of steps. The CG method minimizes the A norm of the error,

More information

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009) Iterative methods for Linear System of Equations Joint Advanced Student School (JASS-2009) Course #2: Numerical Simulation - from Models to Software Introduction In numerical simulation, Partial Differential

More information

Regularization methods for large-scale, ill-posed, linear, discrete, inverse problems

Regularization methods for large-scale, ill-posed, linear, discrete, inverse problems Regularization methods for large-scale, ill-posed, linear, discrete, inverse problems Silvia Gazzola Dipartimento di Matematica - Università di Padova January 10, 2012 Seminario ex-studenti 2 Silvia Gazzola

More information

Golub-Kahan iterative bidiagonalization and determining the noise level in the data

Golub-Kahan iterative bidiagonalization and determining the noise level in the data Golub-Kahan iterative bidiagonalization and determining the noise level in the data Iveta Hnětynková,, Martin Plešinger,, Zdeněk Strakoš, * Charles University, Prague ** Academy of Sciences of the Czech

More information

Newton s method for obtaining the Regularization Parameter for Regularized Least Squares

Newton s method for obtaining the Regularization Parameter for Regularized Least Squares Newton s method for obtaining the Regularization Parameter for Regularized Least Squares Rosemary Renaut, Jodi Mead Arizona State and Boise State International Linear Algebra Society, Cancun Renaut and

More information

Karhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques

Karhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques Institut für Numerische Mathematik und Optimierung Karhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques Oliver Ernst Computational Methods with Applications Harrachov, CR,

More information

FEM and sparse linear system solving

FEM and sparse linear system solving FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich

More information

Linear Algebra Review

Linear Algebra Review Chapter 1 Linear Algebra Review It is assumed that you have had a course in linear algebra, and are familiar with matrix multiplication, eigenvectors, etc. I will review some of these terms here, but quite

More information

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Key words. conjugate gradients, normwise backward error, incremental norm estimation. Proceedings of ALGORITMY 2016 pp. 323 332 ON ERROR ESTIMATION IN THE CONJUGATE GRADIENT METHOD: NORMWISE BACKWARD ERROR PETR TICHÝ Abstract. Using an idea of Duff and Vömel [BIT, 42 (2002), pp. 300 322

More information

Inverse Ill Posed Problems in Image Processing

Inverse Ill Posed Problems in Image Processing Inverse Ill Posed Problems in Image Processing Image Deblurring I. Hnětynková 1,M.Plešinger 2,Z.Strakoš 3 hnetynko@karlin.mff.cuni.cz, martin.plesinger@tul.cz, strakos@cs.cas.cz 1,3 Faculty of Mathematics

More information

Image restoration: numerical optimisation

Image restoration: numerical optimisation Image restoration: numerical optimisation Short and partial presentation Jean-François Giovannelli Groupe Signal Image Laboratoire de l Intégration du Matériau au Système Univ. Bordeaux CNRS BINP / 6 Context

More information

On Lagrange multipliers of trust region subproblems

On Lagrange multipliers of trust region subproblems On Lagrange multipliers of trust region subproblems Ladislav Lukšan, Ctirad Matonoha, Jan Vlček Institute of Computer Science AS CR, Prague Applied Linear Algebra April 28-30, 2008 Novi Sad, Serbia Outline

More information

A Computational Framework for Total Variation-Regularized Positron Emission Tomography

A Computational Framework for Total Variation-Regularized Positron Emission Tomography Numerical Algorithms manuscript No. (will be inserted by the editor) A Computational Framework for Total Variation-Regularized Positron Emission Tomography Johnathan M. Bardsley John Goldes Received: date

More information

On nonstationary preconditioned iterative regularization methods for image deblurring

On nonstationary preconditioned iterative regularization methods for image deblurring On nonstationary preconditioned iterative regularization methods for image deblurring Alessandro Buccini joint work with Prof. Marco Donatelli University of Insubria Department of Science and High Technology

More information

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems Linear algebra issues in Interior Point methods for bound-constrained least-squares problems Stefania Bellavia Dipartimento di Energetica S. Stecco Università degli Studi di Firenze Joint work with Jacek

More information

COMPUTATIONAL ISSUES RELATING TO INVERSION OF PRACTICAL DATA: WHERE IS THE UNCERTAINTY? CAN WE SOLVE Ax = b?

COMPUTATIONAL ISSUES RELATING TO INVERSION OF PRACTICAL DATA: WHERE IS THE UNCERTAINTY? CAN WE SOLVE Ax = b? COMPUTATIONAL ISSUES RELATING TO INVERSION OF PRACTICAL DATA: WHERE IS THE UNCERTAINTY? CAN WE SOLVE Ax = b? Rosemary Renaut http://math.asu.edu/ rosie BRIDGING THE GAP? OCT 2, 2012 Discussion Yuen: Solve

More information

GPU acceleration of Newton s method for large systems of polynomial equations in double double and quad double arithmetic

GPU acceleration of Newton s method for large systems of polynomial equations in double double and quad double arithmetic GPU acceleration of Newton s method for large systems of polynomial equations in double double and quad double arithmetic Jan Verschelde joint work with Xiangcheng Yu University of Illinois at Chicago

More information

Krylov subspace iterative methods for nonsymmetric discrete ill-posed problems in image restoration

Krylov subspace iterative methods for nonsymmetric discrete ill-posed problems in image restoration Krylov subspace iterative methods for nonsymmetric discrete ill-posed problems in image restoration D. Calvetti a, B. Lewis b and L. Reichel c a Department of Mathematics, Case Western Reserve University,

More information

Iterative Krylov Subspace Methods for Sparse Reconstruction

Iterative Krylov Subspace Methods for Sparse Reconstruction Iterative Krylov Subspace Methods for Sparse Reconstruction James Nagy Mathematics and Computer Science Emory University Atlanta, GA USA Joint work with Silvia Gazzola University of Padova, Italy Outline

More information

Laplace-distributed increments, the Laplace prior, and edge-preserving regularization

Laplace-distributed increments, the Laplace prior, and edge-preserving regularization J. Inverse Ill-Posed Probl.? (????), 1 15 DOI 1.1515/JIIP.????.??? de Gruyter???? Laplace-distributed increments, the Laplace prior, and edge-preserving regularization Johnathan M. Bardsley Abstract. For

More information

Parallel programming practices for the solution of Sparse Linear Systems (motivated by computational physics and graphics)

Parallel programming practices for the solution of Sparse Linear Systems (motivated by computational physics and graphics) Parallel programming practices for the solution of Sparse Linear Systems (motivated by computational physics and graphics) Eftychios Sifakis CS758 Guest Lecture - 19 Sept 2012 Introduction Linear systems

More information

A Bregman alternating direction method of multipliers for sparse probabilistic Boolean network problem

A Bregman alternating direction method of multipliers for sparse probabilistic Boolean network problem A Bregman alternating direction method of multipliers for sparse probabilistic Boolean network problem Kangkang Deng, Zheng Peng Abstract: The main task of genetic regulatory networks is to construct a

More information

Conjugate Gradient (CG) Method

Conjugate Gradient (CG) Method Conjugate Gradient (CG) Method by K. Ozawa 1 Introduction In the series of this lecture, I will introduce the conjugate gradient method, which solves efficiently large scale sparse linear simultaneous

More information

Non-negative Quadratic Programming Total Variation Regularization for Poisson Vector-Valued Image Restoration

Non-negative Quadratic Programming Total Variation Regularization for Poisson Vector-Valued Image Restoration University of New Mexico UNM Digital Repository Electrical & Computer Engineering Technical Reports Engineering Publications 5-10-2011 Non-negative Quadratic Programming Total Variation Regularization

More information

SPARSE SIGNAL RESTORATION. 1. Introduction

SPARSE SIGNAL RESTORATION. 1. Introduction SPARSE SIGNAL RESTORATION IVAN W. SELESNICK 1. Introduction These notes describe an approach for the restoration of degraded signals using sparsity. This approach, which has become quite popular, is useful

More information

COMPUTATIONAL METHODS IN MRI: MATHEMATICS

COMPUTATIONAL METHODS IN MRI: MATHEMATICS COMPUTATIONAL METHODS IN MATHEMATICS Imaging Sciences-KCL November 20, 2008 OUTLINE 1 MATRICES AND LINEAR TRANSFORMS: FORWARD OUTLINE 1 MATRICES AND LINEAR TRANSFORMS: FORWARD 2 LINEAR SYSTEMS: INVERSE

More information

arxiv: v1 [hep-lat] 2 May 2012

arxiv: v1 [hep-lat] 2 May 2012 A CG Method for Multiple Right Hand Sides and Multiple Shifts in Lattice QCD Calculations arxiv:1205.0359v1 [hep-lat] 2 May 2012 Fachbereich C, Mathematik und Naturwissenschaften, Bergische Universität

More information

Iterative Methods for Ill-Posed Problems

Iterative Methods for Ill-Posed Problems Iterative Methods for Ill-Posed Problems Based on joint work with: Serena Morigi Fiorella Sgallari Andriy Shyshkov Salt Lake City, May, 2007 Outline: Inverse and ill-posed problems Tikhonov regularization

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 1: Course Overview; Matrix Multiplication Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information

Iterative methods for Linear System

Iterative methods for Linear System Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and

More information

Computing least squares condition numbers on hybrid multicore/gpu systems

Computing least squares condition numbers on hybrid multicore/gpu systems Computing least squares condition numbers on hybrid multicore/gpu systems M. Baboulin and J. Dongarra and R. Lacroix Abstract This paper presents an efficient computation for least squares conditioning

More information

Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota

Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota SIAM CSE Boston - March 1, 2013 First: Joint work with Ruipeng Li Work

More information

CME342 Parallel Methods in Numerical Analysis. Matrix Computation: Iterative Methods II. Sparse Matrix-vector Multiplication.

CME342 Parallel Methods in Numerical Analysis. Matrix Computation: Iterative Methods II. Sparse Matrix-vector Multiplication. CME342 Parallel Methods in Numerical Analysis Matrix Computation: Iterative Methods II Outline: CG & its parallelization. Sparse Matrix-vector Multiplication. 1 Basic iterative methods: Ax = b r = b Ax

More information

Solving PDEs with CUDA Jonathan Cohen

Solving PDEs with CUDA Jonathan Cohen Solving PDEs with CUDA Jonathan Cohen jocohen@nvidia.com NVIDIA Research PDEs (Partial Differential Equations) Big topic Some common strategies Focus on one type of PDE in this talk Poisson Equation Linear

More information

Jacobi-Based Eigenvalue Solver on GPU. Lung-Sheng Chien, NVIDIA

Jacobi-Based Eigenvalue Solver on GPU. Lung-Sheng Chien, NVIDIA Jacobi-Based Eigenvalue Solver on GPU Lung-Sheng Chien, NVIDIA lchien@nvidia.com Outline Symmetric eigenvalue solver Experiment Applications Conclusions Symmetric eigenvalue solver The standard form is

More information

1 Conjugate gradients

1 Conjugate gradients Notes for 2016-11-18 1 Conjugate gradients We now turn to the method of conjugate gradients (CG), perhaps the best known of the Krylov subspace solvers. The CG iteration can be characterized as the iteration

More information

Geometric Modeling Summer Semester 2010 Mathematical Tools (1)

Geometric Modeling Summer Semester 2010 Mathematical Tools (1) Geometric Modeling Summer Semester 2010 Mathematical Tools (1) Recap: Linear Algebra Today... Topics: Mathematical Background Linear algebra Analysis & differential geometry Numerical techniques Geometric

More information

2 Regularized Image Reconstruction for Compressive Imaging and Beyond

2 Regularized Image Reconstruction for Compressive Imaging and Beyond EE 367 / CS 448I Computational Imaging and Display Notes: Compressive Imaging and Regularized Image Reconstruction (lecture ) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement

More information

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

MATH 304 Linear Algebra Lecture 34: Review for Test 2. MATH 304 Linear Algebra Lecture 34: Review for Test 2. Topics for Test 2 Linear transformations (Leon 4.1 4.3) Matrix transformations Matrix of a linear mapping Similar matrices Orthogonality (Leon 5.1

More information

AM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods

AM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods AM 205: lecture 19 Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods Quasi-Newton Methods General form of quasi-newton methods: x k+1 = x k α

More information

J.I. Aliaga 1 M. Bollhöfer 2 A.F. Martín 1 E.S. Quintana-Ortí 1. March, 2009

J.I. Aliaga 1 M. Bollhöfer 2 A.F. Martín 1 E.S. Quintana-Ortí 1. March, 2009 Parallel Preconditioning of Linear Systems based on ILUPACK for Multithreaded Architectures J.I. Aliaga M. Bollhöfer 2 A.F. Martín E.S. Quintana-Ortí Deparment of Computer Science and Engineering, Univ.

More information

On Lagrange multipliers of trust-region subproblems

On Lagrange multipliers of trust-region subproblems On Lagrange multipliers of trust-region subproblems Ladislav Lukšan, Ctirad Matonoha, Jan Vlček Institute of Computer Science AS CR, Prague Programy a algoritmy numerické matematiky 14 1.- 6. června 2008

More information

Introduction to Iterative Solvers of Linear Systems

Introduction to Iterative Solvers of Linear Systems Introduction to Iterative Solvers of Linear Systems SFB Training Event January 2012 Prof. Dr. Andreas Frommer Typeset by Lukas Krämer, Simon-Wolfgang Mages and Rudolf Rödl 1 Classes of Matrices and their

More information

A Dual Formulation of the TV-Stokes Algorithm for Image Denoising

A Dual Formulation of the TV-Stokes Algorithm for Image Denoising A Dual Formulation of the TV-Stokes Algorithm for Image Denoising Christoffer A. Elo, Alexander Malyshev, and Talal Rahman Department of Mathematics, University of Bergen, Johannes Bruns gate 12, 5007

More information

An Accelerated Block-Parallel Newton Method via Overlapped Partitioning

An Accelerated Block-Parallel Newton Method via Overlapped Partitioning An Accelerated Block-Parallel Newton Method via Overlapped Partitioning Yurong Chen Lab. of Parallel Computing, Institute of Software, CAS (http://www.rdcps.ac.cn/~ychen/english.htm) Summary. This paper

More information

GPU accelerated Arnoldi solver for small batched matrix

GPU accelerated Arnoldi solver for small batched matrix 15. 09. 22 GPU accelerated Arnoldi solver for small batched matrix Samsung Advanced Institute of Technology Hyung-Jin Kim Contents - Eigen value problems - Solution - Arnoldi Algorithm - Target - CUDA

More information

Review problems for MA 54, Fall 2004.

Review problems for MA 54, Fall 2004. Review problems for MA 54, Fall 2004. Below are the review problems for the final. They are mostly homework problems, or very similar. If you are comfortable doing these problems, you should be fine on

More information

A Limited Memory, Quasi-Newton Preconditioner. for Nonnegatively Constrained Image. Reconstruction

A Limited Memory, Quasi-Newton Preconditioner. for Nonnegatively Constrained Image. Reconstruction A Limited Memory, Quasi-Newton Preconditioner for Nonnegatively Constrained Image Reconstruction Johnathan M. Bardsley Department of Mathematical Sciences, The University of Montana, Missoula, MT 59812-864

More information

Linear Solvers. Andrew Hazel

Linear Solvers. Andrew Hazel Linear Solvers Andrew Hazel Introduction Thus far we have talked about the formulation and discretisation of physical problems...... and stopped when we got to a discrete linear system of equations. Introduction

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 21: Sensitivity of Eigenvalues and Eigenvectors; Conjugate Gradient Method Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis

More information

Conjugate gradient acceleration of non-linear smoothing filters Iterated edge-preserving smoothing

Conjugate gradient acceleration of non-linear smoothing filters Iterated edge-preserving smoothing Cambridge, Massachusetts Conjugate gradient acceleration of non-linear smoothing filters Iterated edge-preserving smoothing Andrew Knyazev (knyazev@merl.com) (speaker) Alexander Malyshev (malyshev@merl.com)

More information

EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6)

EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6) EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement to the material discussed in

More information

Multivariate Statistical Analysis

Multivariate Statistical Analysis Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 4 for Applied Multivariate Analysis Outline 1 Eigen values and eigen vectors Characteristic equation Some properties of eigendecompositions

More information

Iterative Methods for Linear Systems of Equations

Iterative Methods for Linear Systems of Equations Iterative Methods for Linear Systems of Equations Projection methods (3) ITMAN PhD-course DTU 20-10-08 till 24-10-08 Martin van Gijzen 1 Delft University of Technology Overview day 4 Bi-Lanczos method

More information

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = 30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can

More information

Hybrid Analog-Digital Solution of Nonlinear Partial Differential Equations

Hybrid Analog-Digital Solution of Nonlinear Partial Differential Equations Hybrid Analog-Digital Solution of Nonlinear Partial Differential Equations Yipeng Huang, Ning Guo, Kyle Mandli, Mingoo Seok, Yannis Tsividis, Simha Sethumadhavan Columbia University Hybrid Analog-Digital

More information

PARALLEL VARIABLE DISTRIBUTION FOR TOTAL LEAST SQUARES

PARALLEL VARIABLE DISTRIBUTION FOR TOTAL LEAST SQUARES PARALLEL VARIABLE DISTRIBUTION FOR TOTAL LEAST SQUARES HONGBIN GUO AND ROSEMARY A. RENAUT Abstract. A novel parallel method for determining an approximate total least squares (TLS) solution is introduced.

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information

Proximal Newton Method. Zico Kolter (notes by Ryan Tibshirani) Convex Optimization

Proximal Newton Method. Zico Kolter (notes by Ryan Tibshirani) Convex Optimization Proximal Newton Method Zico Kolter (notes by Ryan Tibshirani) Convex Optimization 10-725 Consider the problem Last time: quasi-newton methods min x f(x) with f convex, twice differentiable, dom(f) = R

More information

MARCH 24-27, 2014 SAN JOSE, CA

MARCH 24-27, 2014 SAN JOSE, CA MARCH 24-27, 2014 SAN JOSE, CA Sparse HPC on modern architectures Important scientific applications rely on sparse linear algebra HPCG a new benchmark proposal to complement Top500 (HPL) To solve A x =

More information

A Riemannian Framework for Denoising Diffusion Tensor Images

A Riemannian Framework for Denoising Diffusion Tensor Images A Riemannian Framework for Denoising Diffusion Tensor Images Manasi Datar No Institute Given Abstract. Diffusion Tensor Imaging (DTI) is a relatively new imaging modality that has been extensively used

More information

Convex Optimization Algorithms for Machine Learning in 10 Slides

Convex Optimization Algorithms for Machine Learning in 10 Slides Convex Optimization Algorithms for Machine Learning in 10 Slides Presenter: Jul. 15. 2015 Outline 1 Quadratic Problem Linear System 2 Smooth Problem Newton-CG 3 Composite Problem Proximal-Newton-CD 4 Non-smooth,

More information

Numerical Methods for Solving Large Scale Eigenvalue Problems

Numerical Methods for Solving Large Scale Eigenvalue Problems Peter Arbenz Computer Science Department, ETH Zürich E-mail: arbenz@inf.ethz.ch arge scale eigenvalue problems, Lecture 2, February 28, 2018 1/46 Numerical Methods for Solving Large Scale Eigenvalue Problems

More information

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian FE661 - Statistical Methods for Financial Engineering 2. Linear algebra Jitkomut Songsiri matrices and vectors linear equations range and nullspace of matrices function of vectors, gradient and Hessian

More information

Seismic imaging and optimal transport

Seismic imaging and optimal transport Seismic imaging and optimal transport Bjorn Engquist In collaboration with Brittany Froese, Sergey Fomel and Yunan Yang Brenier60, Calculus of Variations and Optimal Transportation, Paris, January 10-13,

More information

Projection methods to solve SDP

Projection methods to solve SDP Projection methods to solve SDP Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria F. Rendl, Oberwolfach Seminar, May 2010 p.1/32 Overview Augmented Primal-Dual Method

More information