On the regularization properties of some spectral gradient methods
|
|
- Arlene Douglas
- 5 years ago
- Views:
Transcription
1 On the regularization properties of some spectral gradient methods Daniela di Serafino Department of Mathematics and Physics, Second University of Naples contributions from R. De Asmundis, G. Landi, W.W. Hager, G. Toraldo, M. Viola, H. Zhang PING (Inverse Problems in Geophysics) GNCS Project Opening Workshop Florence, April 6, 2016
2 Outline 1 Linear discrete inverse problems and gradient methods 2 Recent spectral gradient methods for QP: A and C 3 Regularization properties of A and C 4 Extension to bound-constrained QP 5 Possible applications in solving nonlinear inverse problems Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
3 Linear discrete inverse problems and gradient methods Linear discrete inverse problem b = Ax + n, A R p n, n R p, x R n, p n A and b known data, A ill-conditioned, with singular values decaying to zero, and full rank n unknown, representing perturbations in the data x unknown, representing the object to be recovered Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
4 Linear discrete inverse problems and gradient methods Linear discrete inverse problem b = Ax + n, A R p n, n R p, x R n, p n A and b known data, A ill-conditioned, with singular values decaying to zero, and full rank n unknown, representing perturbations in the data x unknown, representing the object to be recovered Reformulation as linear least squares problem: Exact least squares solution: x = A b = n i=1 minimize b Ax 2 2 u T i b n u T i n v i = x true + v i σ i σ i x R n 1 A = UΣV T, U = [u 1,..., u p] R p p, V = [v 1,..., v n] R n n, Σ = diag(σ 1,..., σ n) R p n i=1 useless, because the noise is amplified! Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
5 Linear discrete inverse problems and gradient methods Filter factors and iterative regularization Regularization by filter factors: x reg = n i=1 u T i b φ i v i σ i choose φ i 1 to preserve the components of the solution corresponding to large σ i s, and φ i 0 to filter out the components corresponding to small σ i s Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
6 Linear discrete inverse problems and gradient methods Filter factors and iterative regularization Regularization by filter factors: x reg = n i=1 u T i b φ i v i σ i choose φ i 1 to preserve the components of the solution corresponding to large σ i s, and φ i 0 to filter out the components corresponding to small σ i s Iterative regularization methods, with a suitable early stop, can provide useful regularized solutions x reg Widely investigated classical iterative methods (see, e.g., [Hanke 95; Engl, Hanke & Neubauer 96; Nagy & Palmer 05]): Landweber and Steepest Descent (): very slow but stable convergence, rarely used in practice unless they are coupled with ad hoc preconditioners CG (CGLS, LSQR): fast in reducing the error, but too sensitive to stopping criteria (an early or late stopping may significantly deteriorate the solution) Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
7 Linear discrete inverse problems and gradient methods Gradient methods for convex quadratic problems General framework choose x 0 R n ; k = 0 while (not stop cond) do g k = Qx c compute a suitable steplength α k x k+1 = x k α k g k k = k + 1 end while QP: minimize f (x) 1 x R n 2 xt Qx c T x old origins [Cauchy 1847; Akaike 1959; Forsythe 1968] long considered bad and ineffective because of slow convergence rate and oscillatory behaviour Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
8 Linear discrete inverse problems and gradient methods Gradient methods for convex quadratic problems General framework choose x 0 R n ; k = 0 while (not stop cond) do g k = Qx c compute a suitable steplength α k x k+1 = x k α k g k k = k + 1 end while QP: minimize f (x) 1 x R n 2 xt Qx c T x old origins [Cauchy 1847; Akaike 1959; Forsythe 1968] long considered bad and ineffective because of slow convergence rate and oscillatory behaviour Starting from [Barzilai & Borwein 88], several more efficient gradient methods have been developed, with steplengths related to Hessian spectral properties [Friedlander, Martínez, Molina & Raydan 99; Dai & Yuan 03, 05; Fletcher 05, 12; Dai, Hager, Schittowski & Zhang 06; Yuan 06, 08; Frassoldati, Zanni & Zanghirati 08; De Asmundis, ds, Riccio & Toraldo 13; De Asmundis, ds, Hager, Toraldo & Zhang 14; Gonzaga & Schneider 15] interest in the use of the new gradient methods as regularization methods [Ascher, van den Doel, Huang & Svaiter 09; Cornelio, Porta, Prato & Zanni 13; De Asmundis, ds & Landi 16] Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
9 Linear discrete inverse problems and gradient methods Analysis of gradient methods (for linear least squares) g k = A T (Ax k b), k = 0, 1, 2,... Write g k in terms of the SVD of A: if g 0 = n i=1 µ0 i v i, then n k g k = µ k i v i, µ k i = µ 0 i (1 α j σi 2 ) i=1 j=0 Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
10 Linear discrete inverse problems and gradient methods Analysis of gradient methods (for linear least squares) g k = A T (Ax k b), k = 0, 1, 2,... Write g k in terms of the SVD of A: if g 0 = n i=1 µ0 i v i, then n k g k = µ k i v i, µ k i = µ 0 i (1 α j σi 2 ) i=1 if at the k-th iteration µ k i j=0 = 0 for some i, then µ l i = 0 for l > k µ k i = 0 iff µ 0 i = 0 or α j = 1/σi 2 for some j k α k 1 µ k+1 i << µ k i σi 2 = µ k+1 r < µ k r if r > i µ r k+1 > µ k r if r < i and λ r > 2σi 2 Non-restrictive assumptions: σ 1 > σ 2 > > σ n, µ 0 1 0, µ0 n 0 Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
11 Recent spectral gradient methods for QP: A and C A framework for building fast gradient methods A new steplength selection rule α k = { α k if mod(k, h + m) < h ᾱ s otherwise, with s = max{i k : mod(i, h + m) = h} h 2 α k classical (Cauchy) steplength ᾱ s special steplength with spectral properties In other words: make h consecutive exact line searches and then compute a different steplength, to be kept constant and applied in m consecutive gradient iterations Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
12 Recent spectral gradient methods for QP: A and C A method [De Asmundis, ds, Riccio & Toraldo 13] Set ᾱ s = α s, where α s = ( 1 α s ) 1 αs Let {x k } be the sequence of iterates generated by the method applied to the least squares problem, starting from any point x 0. Then lim α k = k 1 σ σ2 n Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
13 Recent spectral gradient methods for QP: A and C A method [De Asmundis, ds, Riccio & Toraldo 13] Set ᾱ s = α s, where α s = ( 1 α s ) 1 αs Let {x k } be the sequence of iterates generated by the method applied to the least squares problem, starting from any point x 0. Then lim α k = k 1 σ σ2 n A ( with Alignment) combines the tendency of to choose its search direction in span{v 1, v n } the tendency of the gradient method with α k = 1/(σ σ2 n) to align the search direction with v n, R-linear conv., but significant improvement of practical convergence speed over Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
14 Recent spectral gradient methods for QP: A and C C method [De Asmundis, ds, Hager, Toraldo & Zhang 14] Set ᾱ s equal to the Yuan steplength [Yuan 06] αs Y = 2 ( 1 1 ) 2 g s αs 1 αs ( α s 1 g ) s 1 αs αs 1 Let {x k } be the sequence generated by the method applied to the least squares problem, starting from any point x 0. Then lim k αy k = 1 σ1 2. Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
15 Recent spectral gradient methods for QP: A and C C method [De Asmundis, ds, Hager, Toraldo & Zhang 14] Set ᾱ s equal to the Yuan steplength [Yuan 06] αs Y = 2 ( 1 1 ) 2 g s αs 1 αs ( α s 1 g ) s 1 αs αs 1 Let {x k } be the sequence generated by the method applied to the least squares problem, starting from any point x 0. Then lim k αy k = 1 σ1 2. C ( with Constant Yuan steps) uses a finite sequence of Cauchy steps in order to force the search in span{v 1, v n} and to get a suitable approximation of 1/σ 2 1 applies this approximation in multiple steps in order to drive toward zero the component of the gradient along v 1 R-linear convergence, but significant improvement of practical convergence speed over Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
16 Recent spectral gradient methods for QP: A and C Some remarks If σ 1 σ n, then 1/(σ1 2 + σ2 n) 1/σ1 2 and A fosters the elimination of the component of g k corresponding to σ 1 In the ideal case where the component of g k along v 1 is completely removed, the problem size decreases by 1, and A and C tend to drive toward zero the component of g k along v 2. The same reasoning applies to v i for i > 2 Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
17 Recent spectral gradient methods for QP: A and C Some remarks If σ 1 σ n, then 1/(σ1 2 + σ2 n) 1/σ1 2 and A fosters the elimination of the component of g k corresponding to σ 1 In the ideal case where the component of g k along v 1 is completely removed, the problem size decreases by 1, and A and C tend to drive toward zero the component of g k along v 2. The same reasoning applies to v i for i > 2 In general A and C are non-monotone. However, for small values of m, such as m = 2, 3, 4, A and C show monotonicity in practice if h is sufficiently large when very low accuracy is required, as in the regularization of inverse ill-posed problems, h is not required to be too large (h = 2, 3 and m = 2 is a good combination) Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
18 Regularization properties of A and C Filter factors of A and C [De Asmundis, ds & Landi 16] Filter factors of gradient methods x k+1 = ( n k ( 1 1 αr σi 2 ) ) i=1 r=0 } {{ } φ k+1 i u T i b σ i v i, x 0 = 0 The better α r approximates 1/σi 2 for some r, the closer φ k+1 i will be to 1; multiple values of α r close to 1/σi 2 push φ k+1 i to quickly go toward 1 1/α r σi 2 φ k i 0 the tendency of A and C to push toward zero the components of the gradient, according to the decreasing order of the singular values, translates into the approximation of the most significant components of the solution Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
19 Regularization properties of A and C Comparison of filter factors heat problem from Regularization Tools [Hansen 94], size(a) = 64 64, cond(a) 10 28, Gaussian noise, noise level 0.01, A/C with h = 3 and m = 2 φ i heat k=5 A C CG φ i heat k=5 A C DY BB i heat k=10 A C CG i heat k=10 A C DY BB φ i 0 φ i i i Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
20 Regularization properties of A and C Comparison of filter factors (cont d) heat k=40 A C CG heat k=40 A C DY BB φ i 0 φ i i i Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
21 Regularization properties of A and C Comparison of filter factors (cont d) and relative errors heat k=40 A C CG heat k=40 A C DY BB φ i 0 φ i i i heat heat relative error A C CG relative error (log scale) 10 0 A C DY BB iteration iteration Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
22 Regularization properties of A and C Experiments on image restoration problems: paralleltomo parallel-beam tomography AIR Tools [Hansen & Saxild-Hansen 12] img size = 50 50, 36 angles (0 179 ), 75 parallel rays, cond(a) 10 15, h = 3, m = 2 relative error (log scale) A C DY CG paralleltomo (nl=0.025) relative error (log scale) A C DY CG paralleltomo (nl=0.075) iteration iteration Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
23 Regularization properties of A and C Experiments on image restoration problems: paralleltomo parallel-beam tomography AIR Tools [Hansen & Saxild-Hansen 12] img size = 50 50, 36 angles (0 179 ), 75 parallel rays, cond(a) 10 15, h = 3, m = 2 relative error (log scale) A C DY CG paralleltomo (nl=0.025) relative error (log scale) A C DY CG paralleltomo (nl=0.075) iteration iteration original A CG nl = Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
24 Regularization properties of A and C Experiments on image restoration problems: peppers image deblurring problem, image size = Gaussian PSF, noise level nl = 0.01, cond(a) 10 18, h = 3, m = 2 relative error (log scale) 1 A C DY CG peppers (nl=0.01) original iteration noisy & blurred A CG Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
25 Extension to bound-constrained QP Extending A and C to bound-constrained problems BCQP: minimize f (x) = 1 2 xt Qx c T x s. t. x Ω, Ω = {x : l x u} Q R n n symmetric (positive definite), l {R { }} n, u {R {+ }} n General framework x 0 R n ; k = 0 while (not stop cond) do g k = Qx k c compute a suitable steplength α k x k+1 = P Ω (x k α k g k ) k = k + 1 end while Gradient Projection (GP) methods [Goldstein, 1964; Levitin & Polyak, 1966; Calamai & Moré, 1987] P Ω (x) = argmin{ x z : z Ω} Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
26 Extension to bound-constrained QP Extending A and C to bound-constrained problems BCQP: minimize f (x) = 1 2 xt Qx c T x s. t. x Ω, Ω = {x : l x u} Q R n n symmetric (positive definite), l {R { }} n, u {R {+ }} n General framework x 0 R n ; k = 0 while (not stop cond) do g k = Qx k c compute a suitable steplength α k x k+1 = P Ω (x k α k g k ) k = k + 1 end while Gradient Projection (GP) methods [Goldstein, 1964; Levitin & Polyak, 1966; Calamai & Moré, 1987] P Ω (x) = argmin{ x z : z Ω} The spectral properties of A and C are not preserved! Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
27 Extension to bound-constrained QP Two-phase GP algorithm: basics active set at x: A(x) = {i : x i = l i opp. x i = u i } f i (x), x i (l i, u i ) projected gradient at x: ( Ω f (x)) i = min{ f i (x), 0}, x i = l i max{ f i (x), 0}, x i = u i x, x Ω binding set at x: B(x) = {i : (x i = l i and f i (x) 0) or (x i = u i e f i (x) 0)} x nondegenerate stationary point: f i (x ) 0 i A(x ) Identification of the active constraints at the solution [Calamai & Moré, 1987]: if {x k } converges to nondegenerate x Ω and { Ω f (x k ) } converges to 0, then A(x k ) = A(x ) for all sufficiently large k Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
28 Extension to bound-constrained QP Two-phase GP algorithm: basics active set at x: A(x) = {i : x i = l i opp. x i = u i } f i (x), x i (l i, u i ) projected gradient at x: ( Ω f (x)) i = min{ f i (x), 0}, x i = l i max{ f i (x), 0}, x i = u i x, x Ω binding set at x: B(x) = {i : (x i = l i and f i (x) 0) or (x i = u i e f i (x) 0)} x nondegenerate stationary point: f i (x ) 0 i A(x ) Identification of the active constraints at the solution [Calamai & Moré, 1987]: if {x k } converges to nondegenerate x Ω and { Ω f (x k ) } converges to 0, then A(x k ) = A(x ) for all sufficiently large k Basic idea (as in [Moré & Toraldo, 1991]): use a GP method to select a candidate active set use A/C to explore the face of Ω identified by GP (unconstr. subprob.) = preserve the spectral properties of the new gradient methods Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
29 Extension to bound-constrained QP Two-phase GP algorithm [ds, Toraldo, Viola, work in progress] Sketch of the algorithm x 0 R n ; k = 0 while (not stop cond) do apply a GP method to BCQP: starting from y 0 = x k, generate {y j } until cond1 is satisfied x k = y jk, where y jk = last y j apply A/C to min{f k (d) f ( x k + d) : d i = 0 i A( x k )}: starting from d 0 = 0, generate {d j } until cond2 is satisfied x k+1 = P Ω ( x k + α k d rk ), with d rk = last d k and α k computed by a projected search if A(x k+1 ) = B(x k+1 ), then continue with A/C end while cond1: A(y j ) = A(y j 1 ) or f (y j 1 ) f (y j ) η 2 max{f (y l 1 ) f (y l ), 1 l < j } cond2: f k (d j 1 ) f k (d j ) η 1 max{f k (d l 1 ) f k (d l ), 1 l < j } Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
30 Extension to bound-constrained QP Two-phase GP algorithm: convergence Projected search along f (x k ) e d k : generate a sequence of trial steplengths such that [ ] α (l+1) k γ 1 α (l) k, γ 2α (l) k, 0 < γ 1 < γ 2 < 1, α (0) k > 0 α k = α (r) k satisfying an Armijo-like condition for f Convergence: if Q is spd and x is the solution of BCQP, then any sequence {x k } generated by the two-phase GP algorithm is such that either x k = x after a finite number of iterations or x k x Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
31 Extension to bound-constrained QP Some numerical experiments random Q with n = 10 4 and varying κ(q); bounds: β x i β, β = 1, 5, 9 x 0 = 0; stop crit. Ω f (x k ) 10 5 f (x 0) ; C with h = m = 4 (Matlab) # MAT-VET PRODUCTS κ(q) η 1 η 2 10% active constr. 50% active constr. 90% active constr. GPC GPCG GPC GPCG GPC GPCG GPC competitive with GPCG, especially on the most difficult problems GPC less sensitive to η 1 ed η 2 than GPCG Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
32 Possible applications in solving nonlinear inverse problems Exploiting gradient methods for QP/BCQP in nonlinear inverse problems 1 (h, b) = data, m(x, h) = model function, x = parameters to be estimated, r(x) = b m(x, h) = error in model prediction minimize x R n f (x), f (x) = 1 2 r(x) 2 2 = m i=1 r 2 i (x) Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
33 Possible applications in solving nonlinear inverse problems Exploiting gradient methods for QP/BCQP in nonlinear inverse problems 1 (h, b) = data, m(x, h) = model function, x = parameters to be estimated, r(x) = b m(x, h) = error in model prediction minimize x R n f (x), f (x) = 1 2 r(x) 2 2 = m i=1 r 2 i (x) Regularized Gauss-Newton method x 0 R n ; k = 0 while (not stop cond) do compute J k Jacobian of r at x k compute d k regularized solution of minimize d R n J k d + r(x k ) 2 2 compute α k by a suitable line search x k+1 = x k + α k d k k = k + 1 end while Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
34 Possible applications in solving nonlinear inverse problems Exploiting gradient methods for QP/BCQP in nonlinear inverse problems 1 (cont d) Compute a regularized solution of minimize d R n J k d + r(x k ) 2 2 [Deidda, Fenu & Rodriguez, 2014]: compute a TSVD solution (or a TGSVD one, in order to introduce a regularization matrix) Possible alternative: use A/C to compute a regularized solution Less sensitive to the estimate of the noise norm Easy to use in a matrix-free regime Effective? Efficient? Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
35 Possible applications in solving nonlinear inverse problems Exploiting gradient methods for QP/BCQP in nonlinear inverse problems 2 minimize f (u) f fit (u) + λf reg (u) s. t. u 0 f fit (u) = KL(Au, b) Kullback-Leibler divergence f reg (u) = TV (u) or f reg (u) = W u 1 (frame-based regularization) Solve the problem by combining Iteratively Reweighted Norm approach [Wolke & Schwetlick, 1988; Rodriguex & Wohlberg, 2009] Weighted Least Squares approximation of KL fidelity term [Shen, Yin, Zhang, 2015] [Work in progress (just started), with G. Landi] Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
36 Possible applications in solving nonlinear inverse problems Exploiting gradient methods for QP/BCQP in nonlinear inverse problems 2 (cont d) Algorithm (sketch) u 0 R n ; k = 0 while (not stop cond) do 1. compute fk fit(u) quadratic approx of f fit (u) (using u k ) 2. compute f reg k (u) quadratic approx of f reg (u) (using u k ) 3. compute u k+1 argmin u 0 fk fit reg (u) + λfk (u) 4. k = k + 1 end while 1. Weighted Least Squares approximation 2. Iteratively Reweighted Norm approach 3. Two-phase GP algorithm Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
37 Possible applications in solving nonlinear inverse problems CAN WE EFFICIENTLY EXPLOIT SPECTRAL GRADIENT METHODS IN THE PING PROJECT? Daniela di Serafino (II Univ. Naples) Regulariz. properties of gradient methods PING Workshop, April 6, / 24
First order methods in optimization: a second life?
First order methods in optimization: a second life? Gerardo Toraldo Dipartimento di Matematica e Applicazioni "Renato Caccioppoli", Università di Napoli Federico II toraldo@unina.it Joint work with Daniela
More informationOn spectral properties of steepest descent methods
ON SPECTRAL PROPERTIES OF STEEPEST DESCENT METHODS of 20 On spectral properties of steepest descent methods ROBERTA DE ASMUNDIS Department of Statistical Sciences, University of Rome La Sapienza, Piazzale
More informationPreconditioning. Noisy, Ill-Conditioned Linear Systems
Preconditioning Noisy, Ill-Conditioned Linear Systems James G. Nagy Emory University Atlanta, GA Outline 1. The Basic Problem 2. Regularization / Iterative Methods 3. Preconditioning 4. Example: Image
More informationPreconditioning. Noisy, Ill-Conditioned Linear Systems
Preconditioning Noisy, Ill-Conditioned Linear Systems James G. Nagy Emory University Atlanta, GA Outline 1. The Basic Problem 2. Regularization / Iterative Methods 3. Preconditioning 4. Example: Image
More informationScaled gradient projection methods in image deblurring and denoising
Scaled gradient projection methods in image deblurring and denoising Mario Bertero 1 Patrizia Boccacci 1 Silvia Bonettini 2 Riccardo Zanella 3 Luca Zanni 3 1 Dipartmento di Matematica, Università di Genova
More informationAccelerated Gradient Methods for Constrained Image Deblurring
Accelerated Gradient Methods for Constrained Image Deblurring S Bonettini 1, R Zanella 2, L Zanni 2, M Bertero 3 1 Dipartimento di Matematica, Università di Ferrara, Via Saragat 1, Building B, I-44100
More informationRegularization methods for large-scale, ill-posed, linear, discrete, inverse problems
Regularization methods for large-scale, ill-posed, linear, discrete, inverse problems Silvia Gazzola Dipartimento di Matematica - Università di Padova January 10, 2012 Seminario ex-studenti 2 Silvia Gazzola
More informationGradient methods exploiting spectral properties
Noname manuscript No. (will be inserted by the editor) Gradient methods exploiting spectral properties Yaui Huang Yu-Hong Dai Xin-Wei Liu Received: date / Accepted: date Abstract A new stepsize is derived
More informationOn the steepest descent algorithm for quadratic functions
On the steepest descent algorithm for quadratic functions Clóvis C. Gonzaga Ruana M. Schneider July 9, 205 Abstract The steepest descent algorithm with exact line searches (Cauchy algorithm) is inefficient,
More informationNonlinear Optimization: What s important?
Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global
More informationNumerical Methods for Separable Nonlinear Inverse Problems with Constraint and Low Rank
Numerical Methods for Separable Nonlinear Inverse Problems with Constraint and Low Rank Taewon Cho Thesis submitted to the Faculty of the Virginia Polytechnic Institute and State University in partial
More informationLecture 4 - The Gradient Method Objective: find an optimal solution of the problem
Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem min{f (x) : x R n }. The iterative algorithms that we will consider are of the form x k+1 = x k + t k d k, k = 0, 1,...
More informationMath 164: Optimization Barzilai-Borwein Method
Math 164: Optimization Barzilai-Borwein Method Instructor: Wotao Yin Department of Mathematics, UCLA Spring 2015 online discussions on piazza.com Main features of the Barzilai-Borwein (BB) method The BB
More informationOn the steplength selection in gradient methods for unconstrained optimization
On the steplength selection in gradient methods for unconstrained optimization Daniela di Serafino a,, Valeria Ruggiero b, Gerardo Toraldo c, Luca Zanni d a Department of Mathematics and Physics, University
More informationLecture 4 - The Gradient Method Objective: find an optimal solution of the problem
Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem min{f (x) : x R n }. The iterative algorithms that we will consider are of the form x k+1 = x k + t k d k, k = 0, 1,...
More informationA fast nonstationary preconditioning strategy for ill-posed problems, with application to image deblurring
A fast nonstationary preconditioning strategy for ill-posed problems, with application to image deblurring Marco Donatelli Dept. of Science and High Tecnology U. Insubria (Italy) Joint work with M. Hanke
More informationA Globally Optimal Iterative Algorithm to Solve an Ill-Posed Linear System
Copyright 2012 Tech Science Press CMES, vol.84, no.4, pp.383-403, 2012 A Globally Optimal Iterative Algorithm to Solve an Ill-Posed Linear System Chein-Shan Liu 1 Abstract: An iterative algorithm based
More informationSIGNAL AND IMAGE RESTORATION: SOLVING
1 / 55 SIGNAL AND IMAGE RESTORATION: SOLVING ILL-POSED INVERSE PROBLEMS - ESTIMATING PARAMETERS Rosemary Renaut http://math.asu.edu/ rosie CORNELL MAY 10, 2013 2 / 55 Outline Background Parameter Estimation
More informationIterative Methods for Smooth Objective Functions
Optimization Iterative Methods for Smooth Objective Functions Quadratic Objective Functions Stationary Iterative Methods (first/second order) Steepest Descent Method Landweber/Projected Landweber Methods
More informationACQUIRE: an inexact iteratively reweighted norm approach for TV-based Poisson image restoration
: an inexact iteratively reweighted norm approach for TV-based Poisson image restoration Daniela di Serafino Germana Landi Marco Viola February 6, 2019 Abstract We propose a method, called, for the solution
More informationAlgorithms for Constrained Optimization
1 / 42 Algorithms for Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University April 19, 2015 2 / 42 Outline 1. Convergence 2. Sequential quadratic
More informationStep-size Estimation for Unconstrained Optimization Methods
Volume 24, N. 3, pp. 399 416, 2005 Copyright 2005 SBMAC ISSN 0101-8205 www.scielo.br/cam Step-size Estimation for Unconstrained Optimization Methods ZHEN-JUN SHI 1,2 and JIE SHEN 3 1 College of Operations
More informationRegularization Parameter Estimation for Least Squares: A Newton method using the χ 2 -distribution
Regularization Parameter Estimation for Least Squares: A Newton method using the χ 2 -distribution Rosemary Renaut, Jodi Mead Arizona State and Boise State September 2007 Renaut and Mead (ASU/Boise) Scalar
More informationThe Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation
The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation Rosemary Renaut Collaborators: Jodi Mead and Iveta Hnetynkova DEPARTMENT OF MATHEMATICS
More informationThe Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation
The Chi-squared Distribution of the Regularized Least Squares Functional for Regularization Parameter Estimation Rosemary Renaut DEPARTMENT OF MATHEMATICS AND STATISTICS Prague 2008 MATHEMATICS AND STATISTICS
More informationLinear Inverse Problems
Linear Inverse Problems Ajinkya Kadu Utrecht University, The Netherlands February 26, 2018 Outline Introduction Least-squares Reconstruction Methods Examples Summary Introduction 2 What are inverse problems?
More informationITERATIVE REGULARIZATION WITH MINIMUM-RESIDUAL METHODS
BIT Numerical Mathematics 6-3835/3/431-1 $16. 23, Vol. 43, No. 1, pp. 1 18 c Kluwer Academic Publishers ITERATIVE REGULARIZATION WITH MINIMUM-RESIDUAL METHODS T. K. JENSEN and P. C. HANSEN Informatics
More informationE5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization
E5295/5B5749 Convex optimization with engineering applications Lecture 8 Smooth convex unconstrained and equality-constrained minimization A. Forsgren, KTH 1 Lecture 8 Convex optimization 2006/2007 Unconstrained
More informationMathematical Beer Goggles or The Mathematics of Image Processing
How Mathematical Beer Goggles or The Mathematics of Image Processing Department of Mathematical Sciences University of Bath Postgraduate Seminar Series University of Bath 12th February 2008 1 How 2 How
More informationInverse problems Total Variation Regularization Mark van Kraaij Casa seminar 23 May 2007 Technische Universiteit Eindh ove n University of Technology
Inverse problems Total Variation Regularization Mark van Kraaij Casa seminar 23 May 27 Introduction Fredholm first kind integral equation of convolution type in one space dimension: g(x) = 1 k(x x )f(x
More informationSpectral Projected Gradient Methods
Spectral Projected Gradient Methods E. G. Birgin J. M. Martínez M. Raydan January 17, 2007 Keywords: Spectral Projected Gradient Methods, projected gradients, nonmonotone line search, large scale problems,
More information8 Numerical methods for unconstrained problems
8 Numerical methods for unconstrained problems Optimization is one of the important fields in numerical computation, beside solving differential equations and linear systems. We can see that these fields
More informationNewton s Method for Estimating the Regularization Parameter for Least Squares: Using the Chi-curve
Newton s Method for Estimating the Regularization Parameter for Least Squares: Using the Chi-curve Rosemary Renaut, Jodi Mead Arizona State and Boise State Copper Mountain Conference on Iterative Methods
More informationAn l 2 -l q regularization method for large discrete ill-posed problems
Noname manuscript No. (will be inserted by the editor) An l -l q regularization method for large discrete ill-posed problems Alessandro Buccini Lothar Reichel the date of receipt and acceptance should
More informationAdaptive two-point stepsize gradient algorithm
Numerical Algorithms 27: 377 385, 2001. 2001 Kluwer Academic Publishers. Printed in the Netherlands. Adaptive two-point stepsize gradient algorithm Yu-Hong Dai and Hongchao Zhang State Key Laboratory of
More informationStatistically-Based Regularization Parameter Estimation for Large Scale Problems
Statistically-Based Regularization Parameter Estimation for Large Scale Problems Rosemary Renaut Joint work with Jodi Mead and Iveta Hnetynkova March 1, 2010 National Science Foundation: Division of Computational
More informationOptimization and Root Finding. Kurt Hornik
Optimization and Root Finding Kurt Hornik Basics Root finding and unconstrained smooth optimization are closely related: Solving ƒ () = 0 can be accomplished via minimizing ƒ () 2 Slide 2 Basics Root finding
More informationAn Evolving Gradient Resampling Method for Machine Learning. Jorge Nocedal
An Evolving Gradient Resampling Method for Machine Learning Jorge Nocedal Northwestern University NIPS, Montreal 2015 1 Collaborators Figen Oztoprak Stefan Solntsev Richard Byrd 2 Outline 1. How to improve
More informationInverse Ill Posed Problems in Image Processing
Inverse Ill Posed Problems in Image Processing Image Deblurring I. Hnětynková 1,M.Plešinger 2,Z.Strakoš 3 hnetynko@karlin.mff.cuni.cz, martin.plesinger@tul.cz, strakos@cs.cas.cz 1,3 Faculty of Mathematics
More informationA derivative-free nonmonotone line search and its application to the spectral residual method
IMA Journal of Numerical Analysis (2009) 29, 814 825 doi:10.1093/imanum/drn019 Advance Access publication on November 14, 2008 A derivative-free nonmonotone line search and its application to the spectral
More information7.3 The Jacobi and Gauss-Siedel Iterative Techniques. Problem: To solve Ax = b for A R n n. Methodology: Iteratively approximate solution x. No GEPP.
7.3 The Jacobi and Gauss-Siedel Iterative Techniques Problem: To solve Ax = b for A R n n. Methodology: Iteratively approximate solution x. No GEPP. 7.3 The Jacobi and Gauss-Siedel Iterative Techniques
More informationNumerisches Rechnen. (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang. Institut für Geometrie und Praktische Mathematik RWTH Aachen
Numerisches Rechnen (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang Institut für Geometrie und Praktische Mathematik RWTH Aachen Wintersemester 2011/12 IGPM, RWTH Aachen Numerisches Rechnen
More informationImage restoration: numerical optimisation
Image restoration: numerical optimisation Short and partial presentation Jean-François Giovannelli Groupe Signal Image Laboratoire de l Intégration du Matériau au Système Univ. Bordeaux CNRS BINP / 6 Context
More informationProgramming, numerics and optimization
Programming, numerics and optimization Lecture C-3: Unconstrained optimization II Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428
More informationNewton s method for obtaining the Regularization Parameter for Regularized Least Squares
Newton s method for obtaining the Regularization Parameter for Regularized Least Squares Rosemary Renaut, Jodi Mead Arizona State and Boise State International Linear Algebra Society, Cancun Renaut and
More informationOn efficiency of nonmonotone Armijo-type line searches
Noname manuscript No. (will be inserted by the editor On efficiency of nonmonotone Armijo-type line searches Masoud Ahookhosh Susan Ghaderi Abstract Monotonicity and nonmonotonicity play a key role in
More informationZero-Order Methods for the Optimization of Noisy Functions. Jorge Nocedal
Zero-Order Methods for the Optimization of Noisy Functions Jorge Nocedal Northwestern University Simons Institute, October 2017 1 Collaborators Albert Berahas Northwestern University Richard Byrd University
More informationDue Giorni di Algebra Lineare Numerica (2GALN) Febbraio 2016, Como. Iterative regularization in variable exponent Lebesgue spaces
Due Giorni di Algebra Lineare Numerica (2GALN) 16 17 Febbraio 2016, Como Iterative regularization in variable exponent Lebesgue spaces Claudio Estatico 1 Joint work with: Brigida Bonino 1, Fabio Di Benedetto
More informationNumerical Optimization: Basic Concepts and Algorithms
May 27th 2015 Numerical Optimization: Basic Concepts and Algorithms R. Duvigneau R. Duvigneau - Numerical Optimization: Basic Concepts and Algorithms 1 Outline Some basic concepts in optimization Some
More informationNumerical Linear Algebra and. Image Restoration
Numerical Linear Algebra and Image Restoration Maui High Performance Computing Center Wednesday, October 8, 2003 James G. Nagy Emory University Atlanta, GA Thanks to: AFOSR, Dave Tyler, Stuart Jefferies,
More informationOptimization 2. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng
Optimization 2 CS5240 Theoretical Foundations in Multimedia Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (NUS) Optimization 2 1 / 38
More informationIterative Methods for Ill-Posed Problems
Iterative Methods for Ill-Posed Problems Based on joint work with: Serena Morigi Fiorella Sgallari Andriy Shyshkov Salt Lake City, May, 2007 Outline: Inverse and ill-posed problems Tikhonov regularization
More informationLecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University
Lecture Note 7: Iterative methods for solving linear systems Xiaoqun Zhang Shanghai Jiao Tong University Last updated: December 24, 2014 1.1 Review on linear algebra Norms of vectors and matrices vector
More informationCoordinate Update Algorithm Short Course Proximal Operators and Algorithms
Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 36 Why proximal? Newton s method: for C 2 -smooth, unconstrained problems allow
More information1. Introduction. We develop an active set method for the box constrained optimization
SIAM J. OPTIM. Vol. 17, No. 2, pp. 526 557 c 2006 Society for Industrial and Applied Mathematics A NEW ACTIVE SET ALGORITHM FOR BOX CONSTRAINED OPTIMIZATION WILLIAM W. HAGER AND HONGCHAO ZHANG Abstract.
More informationTMA4180 Solutions to recommended exercises in Chapter 3 of N&W
TMA480 Solutions to recommended exercises in Chapter 3 of N&W Exercise 3. The steepest descent and Newtons method with the bactracing algorithm is implemented in rosenbroc_newton.m. With initial point
More informationNumerical solutions of nonlinear systems of equations
Numerical solutions of nonlinear systems of equations Tsung-Ming Huang Department of Mathematics National Taiwan Normal University, Taiwan E-mail: min@math.ntnu.edu.tw August 28, 2011 Outline 1 Fixed points
More informationResidual iterative schemes for largescale linear systems
Universidad Central de Venezuela Facultad de Ciencias Escuela de Computación Lecturas en Ciencias de la Computación ISSN 1316-6239 Residual iterative schemes for largescale linear systems William La Cruz
More informationOptimization Methods. Lecture 18: Optimality Conditions and. Gradient Methods. for Unconstrained Optimization
5.93 Optimization Methods Lecture 8: Optimality Conditions and Gradient Methods for Unconstrained Optimization Outline. Necessary and sucient optimality conditions Slide. Gradient m e t h o d s 3. The
More informationOn nonstationary preconditioned iterative regularization methods for image deblurring
On nonstationary preconditioned iterative regularization methods for image deblurring Alessandro Buccini joint work with Prof. Marco Donatelli University of Insubria Department of Science and High Technology
More informationLeast-Squares Fitting of Model Parameters to Experimental Data
Least-Squares Fitting of Model Parameters to Experimental Data Div. of Mathematical Sciences, Dept of Engineering Sciences and Mathematics, LTU, room E193 Outline of talk What about Science and Scientific
More informationDept. of Mathematics, University of Dundee, Dundee DD1 4HN, Scotland, UK.
A LIMITED MEMORY STEEPEST DESCENT METHOD ROGER FLETCHER Dept. of Mathematics, University of Dundee, Dundee DD1 4HN, Scotland, UK. e-mail fletcher@uk.ac.dundee.mcs Edinburgh Research Group in Optimization
More information, b = 0. (2) 1 2 The eigenvectors of A corresponding to the eigenvalues λ 1 = 1, λ 2 = 3 are
Quadratic forms We consider the quadratic function f : R 2 R defined by f(x) = 2 xt Ax b T x with x = (x, x 2 ) T, () where A R 2 2 is symmetric and b R 2. We will see that, depending on the eigenvalues
More informationAlgorithms for constrained local optimization
Algorithms for constrained local optimization Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Algorithms for constrained local optimization p. Feasible direction methods Algorithms for constrained
More informationOptimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30
Optimization Escuela de Ingeniería Informática de Oviedo (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30 Unconstrained optimization Outline 1 Unconstrained optimization 2 Constrained
More informationLecture 3: Linesearch methods (continued). Steepest descent methods
Lecture 3: Linesearch methods (continued). Steepest descent methods Coralia Cartis, Mathematical Institute, University of Oxford C6.2/B2: Continuous Optimization Lecture 3: Linesearch methods (continued).
More informationConvex Optimization Algorithms for Machine Learning in 10 Slides
Convex Optimization Algorithms for Machine Learning in 10 Slides Presenter: Jul. 15. 2015 Outline 1 Quadratic Problem Linear System 2 Smooth Problem Newton-CG 3 Composite Problem Proximal-Newton-CD 4 Non-smooth,
More informationMethods for Unconstrained Optimization Numerical Optimization Lectures 1-2
Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2 Coralia Cartis, University of Oxford INFOMM CDT: Modelling, Analysis and Computation of Continuous Real-World Problems Methods
More informationProximal Newton Method. Zico Kolter (notes by Ryan Tibshirani) Convex Optimization
Proximal Newton Method Zico Kolter (notes by Ryan Tibshirani) Convex Optimization 10-725 Consider the problem Last time: quasi-newton methods min x f(x) with f convex, twice differentiable, dom(f) = R
More informationPDE-Constrained and Nonsmooth Optimization
Frank E. Curtis October 1, 2009 Outline PDE-Constrained Optimization Introduction Newton s method Inexactness Results Summary and future work Nonsmooth Optimization Sequential quadratic programming (SQP)
More informationModifications of Steepest Descent Method and Conjugate Gradient Method Against Noise for Ill-posed Linear Systems
Available online at www.ispacs.com/cna Volume 2012, Year 2012 Article ID cna-00115, 24 pages doi: 10.5899/2012/cna-00115 Research Article Modifications of Steepest Descent Method and Conjugate Gradient
More informationProximal Newton Method. Ryan Tibshirani Convex Optimization /36-725
Proximal Newton Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: primal-dual interior-point method Given the problem min x subject to f(x) h i (x) 0, i = 1,... m Ax = b where f, h
More informationChapter 2. Optimization. Gradients, convexity, and ALS
Chapter 2 Optimization Gradients, convexity, and ALS Contents Background Gradient descent Stochastic gradient descent Newton s method Alternating least squares KKT conditions 2 Motivation We can solve
More informationSteepest Descent. Juan C. Meza 1. Lawrence Berkeley National Laboratory Berkeley, California 94720
Steepest Descent Juan C. Meza Lawrence Berkeley National Laboratory Berkeley, California 94720 Abstract The steepest descent method has a rich history and is one of the simplest and best known methods
More informationA Limited Memory, Quasi-Newton Preconditioner. for Nonnegatively Constrained Image. Reconstruction
A Limited Memory, Quasi-Newton Preconditioner for Nonnegatively Constrained Image Reconstruction Johnathan M. Bardsley Department of Mathematical Sciences, The University of Montana, Missoula, MT 59812-864
More informationMatrix Derivatives and Descent Optimization Methods
Matrix Derivatives and Descent Optimization Methods 1 Qiang Ning Department of Electrical and Computer Engineering Beckman Institute for Advanced Science and Techonology University of Illinois at Urbana-Champaign
More informationA non-monotonic method for large-scale nonnegative least squares
A non-monotonic method for large-scale nonnegative least squares Dongmin Kim Suvrit Sra Inderjit S. Dhillon May 26, 2010 Abstract We present a new algorithm for nonnegative least-squares (NNLS). Our algorithm
More informationProgetto di Ricerca GNCS 2016 PING Problemi Inversi in Geofisica Firenze, 6 aprile Regularized nonconvex minimization for image restoration
Progetto di Ricerca GNCS 2016 PING Problemi Inversi in Geofisica Firenze, 6 aprile 2016 Regularized nonconvex minimization for image restoration Claudio Estatico Joint work with: Fabio Di Benedetto, Flavia
More informationNonlinear Programming
Nonlinear Programming Kees Roos e-mail: C.Roos@ewi.tudelft.nl URL: http://www.isa.ewi.tudelft.nl/ roos LNMB Course De Uithof, Utrecht February 6 - May 8, A.D. 2006 Optimization Group 1 Outline for week
More informationUnconstrained optimization
Chapter 4 Unconstrained optimization An unconstrained optimization problem takes the form min x Rnf(x) (4.1) for a target functional (also called objective function) f : R n R. In this chapter and throughout
More informationConstrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method.
Optimization Unconstrained optimization One-dimensional Multi-dimensional Newton s method Basic Newton Gauss- Newton Quasi- Newton Descent methods Gradient descent Conjugate gradient Constrained optimization
More informationSPARSE SIGNAL RESTORATION. 1. Introduction
SPARSE SIGNAL RESTORATION IVAN W. SELESNICK 1. Introduction These notes describe an approach for the restoration of degraded signals using sparsity. This approach, which has become quite popular, is useful
More informationComputational Methods. Least Squares Approximation/Optimization
Computational Methods Least Squares Approximation/Optimization Manfred Huber 2011 1 Least Squares Least squares methods are aimed at finding approximate solutions when no precise solution exists Find the
More information4TE3/6TE3. Algorithms for. Continuous Optimization
4TE3/6TE3 Algorithms for Continuous Optimization (Algorithms for Constrained Nonlinear Optimization Problems) Tamás TERLAKY Computing and Software McMaster University Hamilton, November 2005 terlaky@mcmaster.ca
More informationJanuary 29, Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière January 29, 2014 Hestenes Stiefel 1 / 13
Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière Hestenes Stiefel January 29, 2014 Non-linear conjugate gradient method(s): Fletcher Reeves Polak Ribière January 29, 2014 Hestenes
More informationOutline Introduction: Problem Description Diculties Algebraic Structure: Algebraic Varieties Rank Decient Toeplitz Matrices Constructing Lower Rank St
Structured Lower Rank Approximation by Moody T. Chu (NCSU) joint with Robert E. Funderlic (NCSU) and Robert J. Plemmons (Wake Forest) March 5, 1998 Outline Introduction: Problem Description Diculties Algebraic
More informationIPAM Summer School Optimization methods for machine learning. Jorge Nocedal
IPAM Summer School 2012 Tutorial on Optimization methods for machine learning Jorge Nocedal Northwestern University Overview 1. We discuss some characteristics of optimization problems arising in deep
More informationChapter 7 Iterative Techniques in Matrix Algebra
Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition
More informationNumerical Methods for Parameter Estimation in Poisson Data Inversion
DOI 10.1007/s10851-014-0553-9 Numerical Methods for Parameter Estimation in Poisson Data Inversion Luca Zanni Alessandro Benfenati Mario Bertero Valeria Ruggiero Received: 28 June 2014 / Accepted: 11 December
More informationAn interior-point gradient method for large-scale totally nonnegative least squares problems
An interior-point gradient method for large-scale totally nonnegative least squares problems Michael Merritt and Yin Zhang Technical Report TR04-08 Department of Computational and Applied Mathematics Rice
More informationNumerical Methods in Matrix Computations
Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices
More informationCS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3
CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3 Felix Kwok February 27, 2004 Written Problems 1. (Heath E3.10) Let B be an n n matrix, and assume that B is both
More informationChapter 4. Unconstrained optimization
Chapter 4. Unconstrained optimization Version: 28-10-2012 Material: (for details see) Chapter 11 in [FKS] (pp.251-276) A reference e.g. L.11.2 refers to the corresponding Lemma in the book [FKS] PDF-file
More information1. Introduction. We analyze a trust region version of Newton s method for the optimization problem
SIAM J. OPTIM. Vol. 9, No. 4, pp. 1100 1127 c 1999 Society for Industrial and Applied Mathematics NEWTON S METHOD FOR LARGE BOUND-CONSTRAINED OPTIMIZATION PROBLEMS CHIH-JEN LIN AND JORGE J. MORÉ To John
More informationResearch Article Vector Extrapolation Based Landweber Method for Discrete Ill-Posed Problems
Hindawi Mathematical Problems in Engineering Volume 2017, Article ID 1375716, 8 pages https://doi.org/10.1155/2017/1375716 Research Article Vector Extrapolation Based Landweber Method for Discrete Ill-Posed
More information10. Unconstrained minimization
Convex Optimization Boyd & Vandenberghe 10. Unconstrained minimization terminology and assumptions gradient descent method steepest descent method Newton s method self-concordant functions implementation
More informationCS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares
CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares Robert Bridson October 29, 2008 1 Hessian Problems in Newton Last time we fixed one of plain Newton s problems by introducing line search
More information1 Newton s Method. Suppose we want to solve: x R. At x = x, f (x) can be approximated by:
Newton s Method Suppose we want to solve: (P:) min f (x) At x = x, f (x) can be approximated by: n x R. f (x) h(x) := f ( x)+ f ( x) T (x x)+ (x x) t H ( x)(x x), 2 which is the quadratic Taylor expansion
More informationParameter Identification by Iterative Constrained Regularization
Journal of Physics: Conference Series PAPER OPEN ACCESS Parameter Identification by Iterative Constrained Regularization To cite this article: Fabiana Zama 2015 J. Phys.: Conf. Ser. 657 012002 View the
More informationGRADIENT = STEEPEST DESCENT
GRADIENT METHODS GRADIENT = STEEPEST DESCENT Convex Function Iso-contours gradient 0.5 0.4 4 2 0 8 0.3 0.2 0. 0 0. negative gradient 6 0.2 4 0.3 2.5 0.5 0 0.5 0.5 0 0.5 0.4 0.5.5 0.5 0 0.5 GRADIENT DESCENT
More information