On the regularization properties of some spectral gradient methods

Size: px
Start display at page:

Download "On the regularization properties of some spectral gradient methods"

Transcription

1 On the regularization properties of some spectral gradient methods Daniela di Serafino Department of Mathematics and Physics, Second University of Naples contributions from R. De Asmundis, G. Landi, W.W. Hager, G. Toraldo, M. Viola, H. Zhang PING (Inverse Problems in Geophysics) GNCS Project Opening Workshop Florence, April 6, 2016

2 Outline 1 Linear discrete inverse problems and gradient methods 2 Recent spectral gradient methods for QP: A and C 3 Regularization properties of A and C 4 Extension to bound-constrained QP 5 Possible applications in solving nonlinear inverse problems Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

3 Linear discrete inverse problems and gradient methods Linear discrete inverse problem b = Ax + n, A R p n, n R p, x R n, p n A and b known data, A ill-conditioned, with singular values decaying to zero, and full rank n unknown, representing perturbations in the data x unknown, representing the object to be recovered Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

4 Linear discrete inverse problems and gradient methods Linear discrete inverse problem b = Ax + n, A R p n, n R p, x R n, p n A and b known data, A ill-conditioned, with singular values decaying to zero, and full rank n unknown, representing perturbations in the data x unknown, representing the object to be recovered Reformulation as linear least squares problem: Exact least squares solution: x = A b = n i=1 minimize b Ax 2 2 u T i b n u T i n v i = x true + v i σ i σ i x R n 1 A = UΣV T, U = [u 1,..., u p] R p p, V = [v 1,..., v n] R n n, Σ = diag(σ 1,..., σ n) R p n i=1 useless, because the noise is amplified! Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

5 Linear discrete inverse problems and gradient methods Filter factors and iterative regularization Regularization by filter factors: x reg = n i=1 u T i b φ i v i σ i choose φ i 1 to preserve the components of the solution corresponding to large σ i s, and φ i 0 to filter out the components corresponding to small σ i s Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

6 Linear discrete inverse problems and gradient methods Filter factors and iterative regularization Regularization by filter factors: x reg = n i=1 u T i b φ i v i σ i choose φ i 1 to preserve the components of the solution corresponding to large σ i s, and φ i 0 to filter out the components corresponding to small σ i s Iterative regularization methods, with a suitable early stop, can provide useful regularized solutions x reg Widely investigated classical iterative methods (see, e.g., [Hanke 95; Engl, Hanke & Neubauer 96; Nagy & Palmer 05]): Landweber and Steepest Descent (): very slow but stable convergence, rarely used in practice unless they are coupled with ad hoc preconditioners CG (CGLS, LSQR): fast in reducing the error, but too sensitive to stopping criteria (an early or late stopping may significantly deteriorate the solution) Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

7 Linear discrete inverse problems and gradient methods Gradient methods for convex quadratic problems General framework choose x 0 R n ; k = 0 while (not stop cond) do g k = Qx c compute a suitable steplength α k x k+1 = x k α k g k k = k + 1 end while QP: minimize f (x) 1 x R n 2 xt Qx c T x old origins [Cauchy 1847; Akaike 1959; Forsythe 1968] long considered bad and ineffective because of slow convergence rate and oscillatory behaviour Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

8 Linear discrete inverse problems and gradient methods Gradient methods for convex quadratic problems General framework choose x 0 R n ; k = 0 while (not stop cond) do g k = Qx c compute a suitable steplength α k x k+1 = x k α k g k k = k + 1 end while QP: minimize f (x) 1 x R n 2 xt Qx c T x old origins [Cauchy 1847; Akaike 1959; Forsythe 1968] long considered bad and ineffective because of slow convergence rate and oscillatory behaviour Starting from [Barzilai & Borwein 88], several more efficient gradient methods have been developed, with steplengths related to Hessian spectral properties [Friedlander, Martínez, Molina & Raydan 99; Dai & Yuan 03, 05; Fletcher 05, 12; Dai, Hager, Schittowski & Zhang 06; Yuan 06, 08; Frassoldati, Zanni & Zanghirati 08; De Asmundis, ds, Riccio & Toraldo 13; De Asmundis, ds, Hager, Toraldo & Zhang 14; Gonzaga & Schneider 15] interest in the use of the new gradient methods as regularization methods [Ascher, van den Doel, Huang & Svaiter 09; Cornelio, Porta, Prato & Zanni 13; De Asmundis, ds & Landi 16] Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

9 Linear discrete inverse problems and gradient methods Analysis of gradient methods (for linear least squares) g k = A T (Ax k b), k = 0, 1, 2,... Write g k in terms of the SVD of A: if g 0 = n i=1 µ0 i v i, then n k g k = µ k i v i, µ k i = µ 0 i (1 α j σi 2 ) i=1 j=0 Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

10 Linear discrete inverse problems and gradient methods Analysis of gradient methods (for linear least squares) g k = A T (Ax k b), k = 0, 1, 2,... Write g k in terms of the SVD of A: if g 0 = n i=1 µ0 i v i, then n k g k = µ k i v i, µ k i = µ 0 i (1 α j σi 2 ) i=1 if at the k-th iteration µ k i j=0 = 0 for some i, then µ l i = 0 for l > k µ k i = 0 iff µ 0 i = 0 or α j = 1/σi 2 for some j k α k 1 µ k+1 i << µ k i σi 2 = µ k+1 r < µ k r if r > i µ r k+1 > µ k r if r < i and λ r > 2σi 2 Non-restrictive assumptions: σ 1 > σ 2 > > σ n, µ 0 1 0, µ0 n 0 Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

11 Recent spectral gradient methods for QP: A and C A framework for building fast gradient methods A new steplength selection rule α k = { α k if mod(k, h + m) < h ᾱ s otherwise, with s = max{i k : mod(i, h + m) = h} h 2 α k classical (Cauchy) steplength ᾱ s special steplength with spectral properties In other words: make h consecutive exact line searches and then compute a different steplength, to be kept constant and applied in m consecutive gradient iterations Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

12 Recent spectral gradient methods for QP: A and C A method [De Asmundis, ds, Riccio & Toraldo 13] Set ᾱ s = α s, where α s = ( 1 α s ) 1 αs Let {x k } be the sequence of iterates generated by the method applied to the least squares problem, starting from any point x 0. Then lim α k = k 1 σ σ2 n Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

13 Recent spectral gradient methods for QP: A and C A method [De Asmundis, ds, Riccio & Toraldo 13] Set ᾱ s = α s, where α s = ( 1 α s ) 1 αs Let {x k } be the sequence of iterates generated by the method applied to the least squares problem, starting from any point x 0. Then lim α k = k 1 σ σ2 n A ( with Alignment) combines the tendency of to choose its search direction in span{v 1, v n } the tendency of the gradient method with α k = 1/(σ σ2 n) to align the search direction with v n, R-linear conv., but significant improvement of practical convergence speed over Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

14 Recent spectral gradient methods for QP: A and C C method [De Asmundis, ds, Hager, Toraldo & Zhang 14] Set ᾱ s equal to the Yuan steplength [Yuan 06] αs Y = 2 ( 1 1 ) 2 g s αs 1 αs ( α s 1 g ) s 1 αs αs 1 Let {x k } be the sequence generated by the method applied to the least squares problem, starting from any point x 0. Then lim k αy k = 1 σ1 2. Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

15 Recent spectral gradient methods for QP: A and C C method [De Asmundis, ds, Hager, Toraldo & Zhang 14] Set ᾱ s equal to the Yuan steplength [Yuan 06] αs Y = 2 ( 1 1 ) 2 g s αs 1 αs ( α s 1 g ) s 1 αs αs 1 Let {x k } be the sequence generated by the method applied to the least squares problem, starting from any point x 0. Then lim k αy k = 1 σ1 2. C ( with Constant Yuan steps) uses a finite sequence of Cauchy steps in order to force the search in span{v 1, v n} and to get a suitable approximation of 1/σ 2 1 applies this approximation in multiple steps in order to drive toward zero the component of the gradient along v 1 R-linear convergence, but significant improvement of practical convergence speed over Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

16 Recent spectral gradient methods for QP: A and C Some remarks If σ 1 σ n, then 1/(σ1 2 + σ2 n) 1/σ1 2 and A fosters the elimination of the component of g k corresponding to σ 1 In the ideal case where the component of g k along v 1 is completely removed, the problem size decreases by 1, and A and C tend to drive toward zero the component of g k along v 2. The same reasoning applies to v i for i > 2 Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

17 Recent spectral gradient methods for QP: A and C Some remarks If σ 1 σ n, then 1/(σ1 2 + σ2 n) 1/σ1 2 and A fosters the elimination of the component of g k corresponding to σ 1 In the ideal case where the component of g k along v 1 is completely removed, the problem size decreases by 1, and A and C tend to drive toward zero the component of g k along v 2. The same reasoning applies to v i for i > 2 In general A and C are non-monotone. However, for small values of m, such as m = 2, 3, 4, A and C show monotonicity in practice if h is sufficiently large when very low accuracy is required, as in the regularization of inverse ill-posed problems, h is not required to be too large (h = 2, 3 and m = 2 is a good combination) Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

18 Regularization properties of A and C Filter factors of A and C [De Asmundis, ds & Landi 16] Filter factors of gradient methods x k+1 = ( n k ( 1 1 αr σi 2 ) ) i=1 r=0 } {{ } φ k+1 i u T i b σ i v i, x 0 = 0 The better α r approximates 1/σi 2 for some r, the closer φ k+1 i will be to 1; multiple values of α r close to 1/σi 2 push φ k+1 i to quickly go toward 1 1/α r σi 2 φ k i 0 the tendency of A and C to push toward zero the components of the gradient, according to the decreasing order of the singular values, translates into the approximation of the most significant components of the solution Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

19 Regularization properties of A and C Comparison of filter factors heat problem from Regularization Tools [Hansen 94], size(a) = 64 64, cond(a) 10 28, Gaussian noise, noise level 0.01, A/C with h = 3 and m = 2 φ i heat k=5 A C CG φ i heat k=5 A C DY BB i heat k=10 A C CG i heat k=10 A C DY BB φ i 0 φ i i i Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

20 Regularization properties of A and C Comparison of filter factors (cont d) heat k=40 A C CG heat k=40 A C DY BB φ i 0 φ i i i Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

21 Regularization properties of A and C Comparison of filter factors (cont d) and relative errors heat k=40 A C CG heat k=40 A C DY BB φ i 0 φ i i i heat heat relative error A C CG relative error (log scale) 10 0 A C DY BB iteration iteration Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

22 Regularization properties of A and C Experiments on image restoration problems: paralleltomo parallel-beam tomography AIR Tools [Hansen & Saxild-Hansen 12] img size = 50 50, 36 angles (0 179 ), 75 parallel rays, cond(a) 10 15, h = 3, m = 2 relative error (log scale) A C DY CG paralleltomo (nl=0.025) relative error (log scale) A C DY CG paralleltomo (nl=0.075) iteration iteration Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

23 Regularization properties of A and C Experiments on image restoration problems: paralleltomo parallel-beam tomography AIR Tools [Hansen & Saxild-Hansen 12] img size = 50 50, 36 angles (0 179 ), 75 parallel rays, cond(a) 10 15, h = 3, m = 2 relative error (log scale) A C DY CG paralleltomo (nl=0.025) relative error (log scale) A C DY CG paralleltomo (nl=0.075) iteration iteration original A CG nl = Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

24 Regularization properties of A and C Experiments on image restoration problems: peppers image deblurring problem, image size = Gaussian PSF, noise level nl = 0.01, cond(a) 10 18, h = 3, m = 2 relative error (log scale) 1 A C DY CG peppers (nl=0.01) original iteration noisy & blurred A CG Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

25 Extension to bound-constrained QP Extending A and C to bound-constrained problems BCQP: minimize f (x) = 1 2 xt Qx c T x s. t. x Ω, Ω = {x : l x u} Q R n n symmetric (positive definite), l {R { }} n, u {R {+ }} n General framework x 0 R n ; k = 0 while (not stop cond) do g k = Qx k c compute a suitable steplength α k x k+1 = P Ω (x k α k g k ) k = k + 1 end while Gradient Projection (GP) methods [Goldstein, 1964; Levitin & Polyak, 1966; Calamai & Moré, 1987] P Ω (x) = argmin{ x z : z Ω} Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

26 Extension to bound-constrained QP Extending A and C to bound-constrained problems BCQP: minimize f (x) = 1 2 xt Qx c T x s. t. x Ω, Ω = {x : l x u} Q R n n symmetric (positive definite), l {R { }} n, u {R {+ }} n General framework x 0 R n ; k = 0 while (not stop cond) do g k = Qx k c compute a suitable steplength α k x k+1 = P Ω (x k α k g k ) k = k + 1 end while Gradient Projection (GP) methods [Goldstein, 1964; Levitin & Polyak, 1966; Calamai & Moré, 1987] P Ω (x) = argmin{ x z : z Ω} The spectral properties of A and C are not preserved! Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

27 Extension to bound-constrained QP Two-phase GP algorithm: basics active set at x: A(x) = {i : x i = l i opp. x i = u i } f i (x), x i (l i, u i ) projected gradient at x: ( Ω f (x)) i = min{ f i (x), 0}, x i = l i max{ f i (x), 0}, x i = u i x, x Ω binding set at x: B(x) = {i : (x i = l i and f i (x) 0) or (x i = u i e f i (x) 0)} x nondegenerate stationary point: f i (x ) 0 i A(x ) Identification of the active constraints at the solution [Calamai & Moré, 1987]: if {x k } converges to nondegenerate x Ω and { Ω f (x k ) } converges to 0, then A(x k ) = A(x ) for all sufficiently large k Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

28 Extension to bound-constrained QP Two-phase GP algorithm: basics active set at x: A(x) = {i : x i = l i opp. x i = u i } f i (x), x i (l i, u i ) projected gradient at x: ( Ω f (x)) i = min{ f i (x), 0}, x i = l i max{ f i (x), 0}, x i = u i x, x Ω binding set at x: B(x) = {i : (x i = l i and f i (x) 0) or (x i = u i e f i (x) 0)} x nondegenerate stationary point: f i (x ) 0 i A(x ) Identification of the active constraints at the solution [Calamai & Moré, 1987]: if {x k } converges to nondegenerate x Ω and { Ω f (x k ) } converges to 0, then A(x k ) = A(x ) for all sufficiently large k Basic idea (as in [Moré & Toraldo, 1991]): use a GP method to select a candidate active set use A/C to explore the face of Ω identified by GP (unconstr. subprob.) = preserve the spectral properties of the new gradient methods Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

29 Extension to bound-constrained QP Two-phase GP algorithm [ds, Toraldo, Viola, work in progress] Sketch of the algorithm x 0 R n ; k = 0 while (not stop cond) do apply a GP method to BCQP: starting from y 0 = x k, generate {y j } until cond1 is satisfied x k = y jk, where y jk = last y j apply A/C to min{f k (d) f ( x k + d) : d i = 0 i A( x k )}: starting from d 0 = 0, generate {d j } until cond2 is satisfied x k+1 = P Ω ( x k + α k d rk ), with d rk = last d k and α k computed by a projected search if A(x k+1 ) = B(x k+1 ), then continue with A/C end while cond1: A(y j ) = A(y j 1 ) or f (y j 1 ) f (y j ) η 2 max{f (y l 1 ) f (y l ), 1 l < j } cond2: f k (d j 1 ) f k (d j ) η 1 max{f k (d l 1 ) f k (d l ), 1 l < j } Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

30 Extension to bound-constrained QP Two-phase GP algorithm: convergence Projected search along f (x k ) e d k : generate a sequence of trial steplengths such that [ ] α (l+1) k γ 1 α (l) k, γ 2α (l) k, 0 < γ 1 < γ 2 < 1, α (0) k > 0 α k = α (r) k satisfying an Armijo-like condition for f Convergence: if Q is spd and x is the solution of BCQP, then any sequence {x k } generated by the two-phase GP algorithm is such that either x k = x after a finite number of iterations or x k x Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

31 Extension to bound-constrained QP Some numerical experiments random Q with n = 10 4 and varying κ(q); bounds: β x i β, β = 1, 5, 9 x 0 = 0; stop crit. Ω f (x k ) 10 5 f (x 0) ; C with h = m = 4 (Matlab) # MAT-VET PRODUCTS κ(q) η 1 η 2 10% active constr. 50% active constr. 90% active constr. GPC GPCG GPC GPCG GPC GPCG GPC competitive with GPCG, especially on the most difficult problems GPC less sensitive to η 1 ed η 2 than GPCG Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

32 Possible applications in solving nonlinear inverse problems Exploiting gradient methods for QP/BCQP in nonlinear inverse problems 1 (h, b) = data, m(x, h) = model function, x = parameters to be estimated, r(x) = b m(x, h) = error in model prediction minimize x R n f (x), f (x) = 1 2 r(x) 2 2 = m i=1 r 2 i (x) Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

33 Possible applications in solving nonlinear inverse problems Exploiting gradient methods for QP/BCQP in nonlinear inverse problems 1 (h, b) = data, m(x, h) = model function, x = parameters to be estimated, r(x) = b m(x, h) = error in model prediction minimize x R n f (x), f (x) = 1 2 r(x) 2 2 = m i=1 r 2 i (x) Regularized Gauss-Newton method x 0 R n ; k = 0 while (not stop cond) do compute J k Jacobian of r at x k compute d k regularized solution of minimize d R n J k d + r(x k ) 2 2 compute α k by a suitable line search x k+1 = x k + α k d k k = k + 1 end while Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

34 Possible applications in solving nonlinear inverse problems Exploiting gradient methods for QP/BCQP in nonlinear inverse problems 1 (cont d) Compute a regularized solution of minimize d R n J k d + r(x k ) 2 2 [Deidda, Fenu & Rodriguez, 2014]: compute a TSVD solution (or a TGSVD one, in order to introduce a regularization matrix) Possible alternative: use A/C to compute a regularized solution Less sensitive to the estimate of the noise norm Easy to use in a matrix-free regime Effective? Efficient? Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

35 Possible applications in solving nonlinear inverse problems Exploiting gradient methods for QP/BCQP in nonlinear inverse problems 2 minimize f (u) f fit (u) + λf reg (u) s. t. u 0 f fit (u) = KL(Au, b) Kullback-Leibler divergence f reg (u) = TV (u) or f reg (u) = W u 1 (frame-based regularization) Solve the problem by combining Iteratively Reweighted Norm approach [Wolke & Schwetlick, 1988; Rodriguex & Wohlberg, 2009] Weighted Least Squares approximation of KL fidelity term [Shen, Yin, Zhang, 2015] [Work in progress (just started), with G. Landi] Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

36 Possible applications in solving nonlinear inverse problems Exploiting gradient methods for QP/BCQP in nonlinear inverse problems 2 (cont d) Algorithm (sketch) u 0 R n ; k = 0 while (not stop cond) do 1. compute fk fit(u) quadratic approx of f fit (u) (using u k ) 2. compute f reg k (u) quadratic approx of f reg (u) (using u k ) 3. compute u k+1 argmin u 0 fk fit reg (u) + λfk (u) 4. k = k + 1 end while 1. Weighted Least Squares approximation 2. Iteratively Reweighted Norm approach 3. Two-phase GP algorithm Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

37 Possible applications in solving nonlinear inverse problems CAN WE EFFICIENTLY EXPLOIT SPECTRAL GRADIENT METHODS IN THE PING PROJECT? Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24

First order methods in optimization: a second life?

First order methods in optimization: a second life? First order methods in optimization: a second life? Gerardo Toraldo Dipartimento di Matematica e Applicazioni "Renato Caccioppoli", Università di Napoli Federico II toraldo@unina.it Joint work with Daniela

More information

On spectral properties of steepest descent methods

On spectral properties of steepest descent methods ON SPECTRAL PROPERTIES OF STEEPEST DESCENT METHODS of 20 On spectral properties of steepest descent methods ROBERTA DE ASMUNDIS Department of Statistical Sciences, University of Rome La Sapienza, Piazzale

More information

Preconditioning. Noisy, Ill-Conditioned Linear Systems

Preconditioning. Noisy, Ill-Conditioned Linear Systems Preconditioning Noisy, Ill-Conditioned Linear Systems James G. Nagy Emory University Atlanta, GA Outline 1. The Basic Problem 2. Regularization / Iterative Methods 3. Preconditioning 4. Example: Image

More information

Preconditioning. Noisy, Ill-Conditioned Linear Systems

Preconditioning. Noisy, Ill-Conditioned Linear Systems Preconditioning Noisy, Ill-Conditioned Linear Systems James G. Nagy Emory University Atlanta, GA Outline 1. The Basic Problem 2. Regularization / Iterative Methods 3. Preconditioning 4. Example: Image

More information

Scaled gradient projection methods in image deblurring and denoising

Scaled gradient projection methods in image deblurring and denoising Scaled gradient projection methods in image deblurring and denoising Mario Bertero 1 Patrizia Boccacci 1 Silvia Bonettini 2 Riccardo Zanella 3 Luca Zanni 3 1 Dipartmento di Matematica, Università di Genova

More information

Accelerated Gradient Methods for Constrained Image Deblurring

Accelerated Gradient Methods for Constrained Image Deblurring Accelerated Gradient Methods for Constrained Image Deblurring S Bonettini 1, R Zanella 2, L Zanni 2, M Bertero 3 1 Dipartimento di Matematica, Università di Ferrara, Via Saragat 1, Building B, I-44100

More information

Regularization methods for large-scale, ill-posed, linear, discrete, inverse problems

Regularization methods for large-scale, ill-posed, linear, discrete, inverse problems Regularization methods for large-scale, ill-posed, linear, discrete, inverse problems Silvia Gazzola Dipartimento di Matematica - Università di Padova January 10, 2012 Seminario ex-studenti 2 Silvia Gazzola

More information

Gradient methods exploiting spectral properties

Gradient methods exploiting spectral properties Noname manuscript No. (will be inserted by the editor) Gradient methods exploiting spectral properties Yaui Huang Yu-Hong Dai Xin-Wei Liu Received: date / Accepted: date Abstract A new stepsize is derived

More information

On the steepest descent algorithm for quadratic functions

On the steepest descent algorithm for quadratic functions On the steepest descent algorithm for quadratic functions Clóvis C. Gonzaga Ruana M. Schneider July 9, 205 Abstract The steepest descent algorithm with exact line searches (Cauchy algorithm) is inefficient,

More information

Nonlinear Optimization: What s important?

Nonlinear Optimization: What s important? Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global

More information

Numerical Methods for Separable Nonlinear Inverse Problems with Constraint and Low Rank

Numerical Methods for Separable Nonlinear Inverse Problems with Constraint and Low Rank Numerical Methods for Separable Nonlinear Inverse Problems with Constraint and Low Rank Taewon Cho Thesis submitted to the Faculty of the Virginia Polytechnic Institute and State University in partial

More information

Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem

Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem min{f (x) : x R n }. The iterative algorithms that we will consider are of the form x k+1 = x k + t k d k, k = 0, 1,...

More information

Math 164: Optimization Barzilai-Borwein Method

Math 164: Optimization Barzilai-Borwein Method Math 164: Optimization Barzilai-Borwein Method Instructor: Wotao Yin Department of Mathematics, UCLA Spring 2015 online discussions on piazza.com Main features of the Barzilai-Borwein (BB) method The BB

More information

On the steplength selection in gradient methods for unconstrained optimization

On the steplength selection in gradient methods for unconstrained optimization On the steplength selection in gradient methods for unconstrained optimization Daniela di Serafino a,, Valeria Ruggiero b, Gerardo Toraldo c, Luca Zanni d a Department of Mathematics and Physics, University

More information

Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem

Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem min{f (x) : x R n }. The iterative algorithms that we will consider are of the form x k+1 = x k + t k d k, k = 0, 1,...

More information

A fast nonstationary preconditioning strategy for ill-posed problems, with application to image deblurring

A fast nonstationary preconditioning strategy for ill-posed problems, with application to image deblurring A fast nonstationary preconditioning strategy for ill-posed problems, with application to image deblurring Marco Donatelli Dept. of Science and High Tecnology U. Insubria (Italy) Joint work with M. Hanke

More information

A Globally Optimal Iterative Algorithm to Solve an Ill-Posed Linear System

A Globally Optimal Iterative Algorithm to Solve an Ill-Posed Linear System Copyright 2012 Tech Science Press CMES, vol.84, no.4, pp.383-403, 2012 A Globally Optimal Iterative Algorithm to Solve an Ill-Posed Linear System Chein-Shan Liu 1 Abstract: An iterative algorithm based

More information

SIGNAL AND IMAGE RESTORATION: SOLVING

SIGNAL AND IMAGE RESTORATION: SOLVING 1 / 55 SIGNAL AND IMAGE RESTORATION: SOLVING ILL-POSED INVERSE PROBLEMS - ESTIMATING PARAMETERS Rosemary Renaut http://math.asu.edu/ rosie CORNELL MAY 10, 2013 2 / 55 Outline Background Parameter Estimation

More information

Iterative Methods for Smooth Objective Functions

Iterative Methods for Smooth Objective Functions Optimization Iterative Methods for Smooth Objective Functions Quadratic Objective Functions Stationary Iterative Methods (first/second order) Steepest Descent Method Landweber/Projected Landweber Methods

More information

ACQUIRE: an inexact iteratively reweighted norm approach for TV-based Poisson image restoration

ACQUIRE: an inexact iteratively reweighted norm approach for TV-based Poisson image restoration : an inexact iteratively reweighted norm approach for TV-based Poisson image restoration Daniela di Serafino Germana Landi Marco Viola February 6, 2019 Abstract We propose a method, called, for the solution

More information

Algorithms for Constrained Optimization

Algorithms for Constrained Optimization 1 / 42 Algorithms for Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University April 19, 2015 2 / 42 Outline 1. Convergence 2. Sequential quadratic

More information

Step-size Estimation for Unconstrained Optimization Methods

Step-size Estimation for Unconstrained Optimization Methods Volume 24, N. 3, pp. 399 416, 2005 Copyright 2005 SBMAC ISSN 0101-8205 www.scielo.br/cam Step-size Estimation for Unconstrained Optimization Methods ZHEN-JUN SHI 1,2 and JIE SHEN 3 1 College of Operations

More information

Regularization Parameter Estimation for Least Squares: A Newton method using the χ 2 -distribution

Regularization Parameter Estimation for Least Squares: A Newton method using the χ 2 -distribution Regularization Parameter Estimation for Least Squares: A Newton method using the χ 2 -distribution Rosemary Renaut, Jodi Mead Arizona State and Boise State September 2007 Renaut and Mead (ASU/Boise) Scalar

More information

The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation

The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation Rosemary Renaut Collaborators: Jodi Mead and Iveta Hnetynkova DEPARTMENT OF MATHEMATICS

More information

The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation

The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation Rosemary Renaut DEPARTMENT OF MATHEMATICS AND STATISTICS Prague 2008 MATHEMATICS AND STATISTICS

More information

Linear Inverse Problems

Linear Inverse Problems Linear Inverse Problems Ajinkya Kadu Utrecht University, The Netherlands February 26, 2018 Outline Introduction Least-squares Reconstruction Methods Examples Summary Introduction 2 What are inverse problems?

More information

ITERATIVE REGULARIZATION WITH MINIMUM-RESIDUAL METHODS

ITERATIVE REGULARIZATION WITH MINIMUM-RESIDUAL METHODS BIT Numerical Mathematics 6-3835/3/431-1 $16. 23, Vol. 43, No. 1, pp. 1 18 c Kluwer Academic Publishers ITERATIVE REGULARIZATION WITH MINIMUM-RESIDUAL METHODS T. K. JENSEN and P. C. HANSEN Informatics

More information

E5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization

E5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization E5295/5B5749 Convex optimization with engineering applications Lecture 8 Smooth convex unconstrained and equality-constrained minimization A. Forsgren, KTH 1 Lecture 8 Convex optimization 2006/2007 Unconstrained

More information

Mathematical Beer Goggles or The Mathematics of Image Processing

Mathematical Beer Goggles or The Mathematics of Image Processing How Mathematical Beer Goggles or The Mathematics of Image Processing Department of Mathematical Sciences University of Bath Postgraduate Seminar Series University of Bath 12th February 2008 1 How 2 How

More information

Inverse problems Total Variation Regularization Mark van Kraaij Casa seminar 23 May 2007 Technische Universiteit Eindh ove n University of Technology

Inverse problems Total Variation Regularization Mark van Kraaij Casa seminar 23 May 2007 Technische Universiteit Eindh ove n University of Technology Inverse problems Total Variation Regularization Mark van Kraaij Casa seminar 23 May 27 Introduction Fredholm first kind integral equation of convolution type in one space dimension: g(x) = 1 k(x x )f(x

More information

Spectral Projected Gradient Methods

Spectral Projected Gradient Methods Spectral Projected Gradient Methods E. G. Birgin J. M. Martínez M. Raydan January 17, 2007 Keywords: Spectral Projected Gradient Methods, projected gradients, nonmonotone line search, large scale problems,

More information

8 Numerical methods for unconstrained problems

8 Numerical methods for unconstrained problems 8 Numerical methods for unconstrained problems Optimization is one of the important fields in numerical computation, beside solving differential equations and linear systems. We can see that these fields

More information

Newton s Method for Estimating the Regularization Parameter for Least Squares: Using the Chi-curve

Newton s Method for Estimating the Regularization Parameter for Least Squares: Using the Chi-curve Newton s Method for Estimating the Regularization Parameter for Least Squares: Using the Chi-curve Rosemary Renaut, Jodi Mead Arizona State and Boise State Copper Mountain Conference on Iterative Methods

More information

An l 2 -l q regularization method for large discrete ill-posed problems

An l 2 -l q regularization method for large discrete ill-posed problems Noname manuscript No. (will be inserted by the editor) An l -l q regularization method for large discrete ill-posed problems Alessandro Buccini Lothar Reichel the date of receipt and acceptance should

More information

Adaptive two-point stepsize gradient algorithm

Adaptive two-point stepsize gradient algorithm Numerical Algorithms 27: 377 385, 2001. 2001 Kluwer Academic Publishers. Printed in the Netherlands. Adaptive two-point stepsize gradient algorithm Yu-Hong Dai and Hongchao Zhang State Key Laboratory of

More information

Statistically-Based Regularization Parameter Estimation for Large Scale Problems

Statistically-Based Regularization Parameter Estimation for Large Scale Problems Statistically-Based Regularization Parameter Estimation for Large Scale Problems Rosemary Renaut Joint work with Jodi Mead and Iveta Hnetynkova March 1, 2010 National Science Foundation: Division of Computational

More information

Optimization and Root Finding. Kurt Hornik

Optimization and Root Finding. Kurt Hornik Optimization and Root Finding Kurt Hornik Basics Root finding and unconstrained smooth optimization are closely related: Solving ƒ () = 0 can be accomplished via minimizing ƒ () 2 Slide 2 Basics Root finding

More information

An Evolving Gradient Resampling Method for Machine Learning. Jorge Nocedal

An Evolving Gradient Resampling Method for Machine Learning. Jorge Nocedal An Evolving Gradient Resampling Method for Machine Learning Jorge Nocedal Northwestern University NIPS, Montreal 2015 1 Collaborators Figen Oztoprak Stefan Solntsev Richard Byrd 2 Outline 1. How to improve

More information

Inverse Ill Posed Problems in Image Processing

Inverse Ill Posed Problems in Image Processing Inverse Ill Posed Problems in Image Processing Image Deblurring I. Hnětynková 1,M.Plešinger 2,Z.Strakoš 3 hnetynko@karlin.mff.cuni.cz, martin.plesinger@tul.cz, strakos@cs.cas.cz 1,3 Faculty of Mathematics

More information

A derivative-free nonmonotone line search and its application to the spectral residual method

A derivative-free nonmonotone line search and its application to the spectral residual method IMA Journal of Numerical Analysis (2009) 29, 814 825 doi:10.1093/imanum/drn019 Advance Access publication on November 14, 2008 A derivative-free nonmonotone line search and its application to the spectral

More information

7.3 The Jacobi and Gauss-Siedel Iterative Techniques. Problem: To solve Ax = b for A R n n. Methodology: Iteratively approximate solution x. No GEPP.

7.3 The Jacobi and Gauss-Siedel Iterative Techniques. Problem: To solve Ax = b for A R n n. Methodology: Iteratively approximate solution x. No GEPP. 7.3 The Jacobi and Gauss-Siedel Iterative Techniques Problem: To solve Ax = b for A R n n. Methodology: Iteratively approximate solution x. No GEPP. 7.3 The Jacobi and Gauss-Siedel Iterative Techniques

More information

Numerisches Rechnen. (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang. Institut für Geometrie und Praktische Mathematik RWTH Aachen

Numerisches Rechnen. (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang. Institut für Geometrie und Praktische Mathematik RWTH Aachen Numerisches Rechnen (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang Institut für Geometrie und Praktische Mathematik RWTH Aachen Wintersemester 2011/12 IGPM, RWTH Aachen Numerisches Rechnen

More information

Image restoration: numerical optimisation

Image restoration: numerical optimisation Image restoration: numerical optimisation Short and partial presentation Jean-François Giovannelli Groupe Signal Image Laboratoire de l Intégration du Matériau au Système Univ. Bordeaux CNRS BINP / 6 Context

More information

Programming, numerics and optimization

Programming, numerics and optimization Programming, numerics and optimization Lecture C-3: Unconstrained optimization II Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428

More information

Newton s method for obtaining the Regularization Parameter for Regularized Least Squares

Newton s method for obtaining the Regularization Parameter for Regularized Least Squares Newton s method for obtaining the Regularization Parameter for Regularized Least Squares Rosemary Renaut, Jodi Mead Arizona State and Boise State International Linear Algebra Society, Cancun Renaut and

More information

On efficiency of nonmonotone Armijo-type line searches

On efficiency of nonmonotone Armijo-type line searches Noname manuscript No. (will be inserted by the editor On efficiency of nonmonotone Armijo-type line searches Masoud Ahookhosh Susan Ghaderi Abstract Monotonicity and nonmonotonicity play a key role in

More information

Zero-Order Methods for the Optimization of Noisy Functions. Jorge Nocedal

Zero-Order Methods for the Optimization of Noisy Functions. Jorge Nocedal Zero-Order Methods for the Optimization of Noisy Functions Jorge Nocedal Northwestern University Simons Institute, October 2017 1 Collaborators Albert Berahas Northwestern University Richard Byrd University

More information

Due Giorni di Algebra Lineare Numerica (2GALN) Febbraio 2016, Como. Iterative regularization in variable exponent Lebesgue spaces

Due Giorni di Algebra Lineare Numerica (2GALN) Febbraio 2016, Como. Iterative regularization in variable exponent Lebesgue spaces Due Giorni di Algebra Lineare Numerica (2GALN) 16 17 Febbraio 2016, Como Iterative regularization in variable exponent Lebesgue spaces Claudio Estatico 1 Joint work with: Brigida Bonino 1, Fabio Di Benedetto

More information

Numerical Optimization: Basic Concepts and Algorithms

Numerical Optimization: Basic Concepts and Algorithms May 27th 2015 Numerical Optimization: Basic Concepts and Algorithms R. Duvigneau R. Duvigneau - Numerical Optimization: Basic Concepts and Algorithms 1 Outline Some basic concepts in optimization Some

More information

Numerical Linear Algebra and. Image Restoration

Numerical Linear Algebra and. Image Restoration Numerical Linear Algebra and Image Restoration Maui High Performance Computing Center Wednesday, October 8, 2003 James G. Nagy Emory University Atlanta, GA Thanks to: AFOSR, Dave Tyler, Stuart Jefferies,

More information

Optimization 2. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng

Optimization 2. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng Optimization 2 CS5240 Theoretical Foundations in Multimedia Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (NUS) Optimization 2 1 / 38

More information

Iterative Methods for Ill-Posed Problems

Iterative Methods for Ill-Posed Problems Iterative Methods for Ill-Posed Problems Based on joint work with: Serena Morigi Fiorella Sgallari Andriy Shyshkov Salt Lake City, May, 2007 Outline: Inverse and ill-posed problems Tikhonov regularization

More information

Lecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University

Lecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University Lecture Note 7: Iterative methods for solving linear systems Xiaoqun Zhang Shanghai Jiao Tong University Last updated: December 24, 2014 1.1 Review on linear algebra Norms of vectors and matrices vector

More information

Coordinate Update Algorithm Short Course Proximal Operators and Algorithms

Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 36 Why proximal? Newton s method: for C 2 -smooth, unconstrained problems allow

More information

1. Introduction. We develop an active set method for the box constrained optimization

1. Introduction. We develop an active set method for the box constrained optimization SIAM J. OPTIM. Vol. 17, No. 2, pp. 526 557 c 2006 Society for Industrial and Applied Mathematics A NEW ACTIVE SET ALGORITHM FOR BOX CONSTRAINED OPTIMIZATION WILLIAM W. HAGER AND HONGCHAO ZHANG Abstract.

More information

TMA4180 Solutions to recommended exercises in Chapter 3 of N&W

TMA4180 Solutions to recommended exercises in Chapter 3 of N&W TMA480 Solutions to recommended exercises in Chapter 3 of N&W Exercise 3. The steepest descent and Newtons method with the bactracing algorithm is implemented in rosenbroc_newton.m. With initial point

More information

Numerical solutions of nonlinear systems of equations

Numerical solutions of nonlinear systems of equations Numerical solutions of nonlinear systems of equations Tsung-Ming Huang Department of Mathematics National Taiwan Normal University, Taiwan E-mail: min@math.ntnu.edu.tw August 28, 2011 Outline 1 Fixed points

More information

Residual iterative schemes for largescale linear systems

Residual iterative schemes for largescale linear systems Universidad Central de Venezuela Facultad de Ciencias Escuela de Computación Lecturas en Ciencias de la Computación ISSN 1316-6239 Residual iterative schemes for largescale linear systems William La Cruz

More information

Optimization Methods. Lecture 18: Optimality Conditions and. Gradient Methods. for Unconstrained Optimization

Optimization Methods. Lecture 18: Optimality Conditions and. Gradient Methods. for Unconstrained Optimization 5.93 Optimization Methods Lecture 8: Optimality Conditions and Gradient Methods for Unconstrained Optimization Outline. Necessary and sucient optimality conditions Slide. Gradient m e t h o d s 3. The

More information

On nonstationary preconditioned iterative regularization methods for image deblurring

On nonstationary preconditioned iterative regularization methods for image deblurring On nonstationary preconditioned iterative regularization methods for image deblurring Alessandro Buccini joint work with Prof. Marco Donatelli University of Insubria Department of Science and High Technology

More information

Least-Squares Fitting of Model Parameters to Experimental Data

Least-Squares Fitting of Model Parameters to Experimental Data Least-Squares Fitting of Model Parameters to Experimental Data Div. of Mathematical Sciences, Dept of Engineering Sciences and Mathematics, LTU, room E193 Outline of talk What about Science and Scientific

More information

Dept. of Mathematics, University of Dundee, Dundee DD1 4HN, Scotland, UK.

Dept. of Mathematics, University of Dundee, Dundee DD1 4HN, Scotland, UK. A LIMITED MEMORY STEEPEST DESCENT METHOD ROGER FLETCHER Dept. of Mathematics, University of Dundee, Dundee DD1 4HN, Scotland, UK. e-mail fletcher@uk.ac.dundee.mcs Edinburgh Research Group in Optimization

More information

, b = 0. (2) 1 2 The eigenvectors of A corresponding to the eigenvalues λ 1 = 1, λ 2 = 3 are

, b = 0. (2) 1 2 The eigenvectors of A corresponding to the eigenvalues λ 1 = 1, λ 2 = 3 are Quadratic forms We consider the quadratic function f : R 2 R defined by f(x) = 2 xt Ax b T x with x = (x, x 2 ) T, () where A R 2 2 is symmetric and b R 2. We will see that, depending on the eigenvalues

More information

Algorithms for constrained local optimization

Algorithms for constrained local optimization Algorithms for constrained local optimization Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Algorithms for constrained local optimization p. Feasible direction methods Algorithms for constrained

More information

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30 Optimization Escuela de Ingeniería Informática de Oviedo (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30 Unconstrained optimization Outline 1 Unconstrained optimization 2 Constrained

More information

Lecture 3: Linesearch methods (continued). Steepest descent methods

Lecture 3: Linesearch methods (continued). Steepest descent methods Lecture 3: Linesearch methods (continued). Steepest descent methods Coralia Cartis, Mathematical Institute, University of Oxford C6.2/B2: Continuous Optimization Lecture 3: Linesearch methods (continued).

More information

Convex Optimization Algorithms for Machine Learning in 10 Slides

Convex Optimization Algorithms for Machine Learning in 10 Slides Convex Optimization Algorithms for Machine Learning in 10 Slides Presenter: Jul. 15. 2015 Outline 1 Quadratic Problem Linear System 2 Smooth Problem Newton-CG 3 Composite Problem Proximal-Newton-CD 4 Non-smooth,

More information

Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2

Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2 Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2 Coralia Cartis, University of Oxford INFOMM CDT: Modelling, Analysis and Computation of Continuous Real-World Problems Methods

More information

Proximal Newton Method. Zico Kolter (notes by Ryan Tibshirani) Convex Optimization

Proximal Newton Method. Zico Kolter (notes by Ryan Tibshirani) Convex Optimization Proximal Newton Method Zico Kolter (notes by Ryan Tibshirani) Convex Optimization 10-725 Consider the problem Last time: quasi-newton methods min x f(x) with f convex, twice differentiable, dom(f) = R

More information

PDE-Constrained and Nonsmooth Optimization

PDE-Constrained and Nonsmooth Optimization Frank E. Curtis October 1, 2009 Outline PDE-Constrained Optimization Introduction Newton s method Inexactness Results Summary and future work Nonsmooth Optimization Sequential quadratic programming (SQP)

More information

Modifications of Steepest Descent Method and Conjugate Gradient Method Against Noise for Ill-posed Linear Systems

Modifications of Steepest Descent Method and Conjugate Gradient Method Against Noise for Ill-posed Linear Systems Available online at www.ispacs.com/cna Volume 2012, Year 2012 Article ID cna-00115, 24 pages doi: 10.5899/2012/cna-00115 Research Article Modifications of Steepest Descent Method and Conjugate Gradient

More information

Proximal Newton Method. Ryan Tibshirani Convex Optimization /36-725

Proximal Newton Method. Ryan Tibshirani Convex Optimization /36-725 Proximal Newton Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: primal-dual interior-point method Given the problem min x subject to f(x) h i (x) 0, i = 1,... m Ax = b where f, h

More information

Chapter 2. Optimization. Gradients, convexity, and ALS

Chapter 2. Optimization. Gradients, convexity, and ALS Chapter 2 Optimization Gradients, convexity, and ALS Contents Background Gradient descent Stochastic gradient descent Newton s method Alternating least squares KKT conditions 2 Motivation We can solve

More information

Steepest Descent. Juan C. Meza 1. Lawrence Berkeley National Laboratory Berkeley, California 94720

Steepest Descent. Juan C. Meza 1. Lawrence Berkeley National Laboratory Berkeley, California 94720 Steepest Descent Juan C. Meza Lawrence Berkeley National Laboratory Berkeley, California 94720 Abstract The steepest descent method has a rich history and is one of the simplest and best known methods

More information

A Limited Memory, Quasi-Newton Preconditioner. for Nonnegatively Constrained Image. Reconstruction

A Limited Memory, Quasi-Newton Preconditioner. for Nonnegatively Constrained Image. Reconstruction A Limited Memory, Quasi-Newton Preconditioner for Nonnegatively Constrained Image Reconstruction Johnathan M. Bardsley Department of Mathematical Sciences, The University of Montana, Missoula, MT 59812-864

More information

Matrix Derivatives and Descent Optimization Methods

Matrix Derivatives and Descent Optimization Methods Matrix Derivatives and Descent Optimization Methods 1 Qiang Ning Department of Electrical and Computer Engineering Beckman Institute for Advanced Science and Techonology University of Illinois at Urbana-Champaign

More information

A non-monotonic method for large-scale nonnegative least squares

A non-monotonic method for large-scale nonnegative least squares A non-monotonic method for large-scale nonnegative least squares Dongmin Kim Suvrit Sra Inderjit S. Dhillon May 26, 2010 Abstract We present a new algorithm for nonnegative least-squares (NNLS). Our algorithm

More information

Progetto di Ricerca GNCS 2016 PING Problemi Inversi in Geofisica Firenze, 6 aprile Regularized nonconvex minimization for image restoration

Progetto di Ricerca GNCS 2016 PING Problemi Inversi in Geofisica Firenze, 6 aprile Regularized nonconvex minimization for image restoration Progetto di Ricerca GNCS 2016 PING Problemi Inversi in Geofisica Firenze, 6 aprile 2016 Regularized nonconvex minimization for image restoration Claudio Estatico Joint work with: Fabio Di Benedetto, Flavia

More information

Nonlinear Programming

Nonlinear Programming Nonlinear Programming Kees Roos e-mail: C.Roos@ewi.tudelft.nl URL: http://www.isa.ewi.tudelft.nl/ roos LNMB Course De Uithof, Utrecht February 6 - May 8, A.D. 2006 Optimization Group 1 Outline for week

More information

Unconstrained optimization

Unconstrained optimization Chapter 4 Unconstrained optimization An unconstrained optimization problem takes the form min x Rnf(x) (4.1) for a target functional (also called objective function) f : R n R. In this chapter and throughout

More information

Constrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method.

Constrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method. Optimization Unconstrained optimization One-dimensional Multi-dimensional Newton s method Basic Newton Gauss- Newton Quasi- Newton Descent methods Gradient descent Conjugate gradient Constrained optimization

More information

SPARSE SIGNAL RESTORATION. 1. Introduction

SPARSE SIGNAL RESTORATION. 1. Introduction SPARSE SIGNAL RESTORATION IVAN W. SELESNICK 1. Introduction These notes describe an approach for the restoration of degraded signals using sparsity. This approach, which has become quite popular, is useful

More information

Computational Methods. Least Squares Approximation/Optimization

Computational Methods. Least Squares Approximation/Optimization Computational Methods Least Squares Approximation/Optimization Manfred Huber 2011 1 Least Squares Least squares methods are aimed at finding approximate solutions when no precise solution exists Find the

More information

4TE3/6TE3. Algorithms for. Continuous Optimization

4TE3/6TE3. Algorithms for. Continuous Optimization 4TE3/6TE3 Algorithms for Continuous Optimization (Algorithms for Constrained Nonlinear Optimization Problems) Tamás TERLAKY Computing and Software McMaster University Hamilton, November 2005 terlaky@mcmaster.ca

More information

January 29, Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière January 29, 2014 Hestenes Stiefel 1 / 13

January 29, Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière January 29, 2014 Hestenes Stiefel 1 / 13 Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière Hestenes Stiefel January 29, 2014 Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière January 29, 2014 Hestenes

More information

Outline Introduction: Problem Description Diculties Algebraic Structure: Algebraic Varieties Rank Decient Toeplitz Matrices Constructing Lower Rank St

Outline Introduction: Problem Description Diculties Algebraic Structure: Algebraic Varieties Rank Decient Toeplitz Matrices Constructing Lower Rank St Structured Lower Rank Approximation by Moody T. Chu (NCSU) joint with Robert E. Funderlic (NCSU) and Robert J. Plemmons (Wake Forest) March 5, 1998 Outline Introduction: Problem Description Diculties Algebraic

More information

IPAM Summer School Optimization methods for machine learning. Jorge Nocedal

IPAM Summer School Optimization methods for machine learning. Jorge Nocedal IPAM Summer School 2012 Tutorial on Optimization methods for machine learning Jorge Nocedal Northwestern University Overview 1. We discuss some characteristics of optimization problems arising in deep

More information

Chapter 7 Iterative Techniques in Matrix Algebra

Chapter 7 Iterative Techniques in Matrix Algebra Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition

More information

Numerical Methods for Parameter Estimation in Poisson Data Inversion

Numerical Methods for Parameter Estimation in Poisson Data Inversion DOI 10.1007/s10851-014-0553-9 Numerical Methods for Parameter Estimation in Poisson Data Inversion Luca Zanni Alessandro Benfenati Mario Bertero Valeria Ruggiero Received: 28 June 2014 / Accepted: 11 December

More information

An interior-point gradient method for large-scale totally nonnegative least squares problems

An interior-point gradient method for large-scale totally nonnegative least squares problems An interior-point gradient method for large-scale totally nonnegative least squares problems Michael Merritt and Yin Zhang Technical Report TR04-08 Department of Computational and Applied Mathematics Rice

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3

CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3 CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3 Felix Kwok February 27, 2004 Written Problems 1. (Heath E3.10) Let B be an n n matrix, and assume that B is both

More information

Chapter 4. Unconstrained optimization

Chapter 4. Unconstrained optimization Chapter 4. Unconstrained optimization Version: 28-10-2012 Material: (for details see) Chapter 11 in [FKS] (pp.251-276) A reference e.g. L.11.2 refers to the corresponding Lemma in the book [FKS] PDF-file

More information

1. Introduction. We analyze a trust region version of Newton s method for the optimization problem

1. Introduction. We analyze a trust region version of Newton s method for the optimization problem SIAM J. OPTIM. Vol. 9, No. 4, pp. 1100 1127 c 1999 Society for Industrial and Applied Mathematics NEWTON S METHOD FOR LARGE BOUND-CONSTRAINED OPTIMIZATION PROBLEMS CHIH-JEN LIN AND JORGE J. MORÉ To John

More information

Research Article Vector Extrapolation Based Landweber Method for Discrete Ill-Posed Problems

Research Article Vector Extrapolation Based Landweber Method for Discrete Ill-Posed Problems Hindawi Mathematical Problems in Engineering Volume 2017, Article ID 1375716, 8 pages https://doi.org/10.1155/2017/1375716 Research Article Vector Extrapolation Based Landweber Method for Discrete Ill-Posed

More information

10. Unconstrained minimization

10. Unconstrained minimization Convex Optimization Boyd & Vandenberghe 10. Unconstrained minimization terminology and assumptions gradient descent method steepest descent method Newton s method self-concordant functions implementation

More information

CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares

CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares Robert Bridson October 29, 2008 1 Hessian Problems in Newton Last time we fixed one of plain Newton s problems by introducing line search

More information

1 Newton s Method. Suppose we want to solve: x R. At x = x, f (x) can be approximated by:

1 Newton s Method. Suppose we want to solve: x R. At x = x, f (x) can be approximated by: Newton s Method Suppose we want to solve: (P:) min f (x) At x = x, f (x) can be approximated by: n x R. f (x) h(x) := f ( x)+ f ( x) T (x x)+ (x x) t H ( x)(x x), 2 which is the quadratic Taylor expansion

More information

Parameter Identification by Iterative Constrained Regularization

Parameter Identification by Iterative Constrained Regularization Journal of Physics: Conference Series PAPER OPEN ACCESS Parameter Identification by Iterative Constrained Regularization To cite this article: Fabiana Zama 2015 J. Phys.: Conf. Ser. 657 012002 View the

More information

GRADIENT = STEEPEST DESCENT

GRADIENT = STEEPEST DESCENT GRADIENT METHODS GRADIENT = STEEPEST DESCENT Convex Function Iso-contours gradient 0.5 0.4 4 2 0 8 0.3 0.2 0. 0 0. negative gradient 6 0.2 4 0.3 2.5 0.5 0 0.5 0.5 0 0.5 0.4 0.5.5 0.5 0 0.5 GRADIENT DESCENT

More information