Stochas(c Dual Ascent Linear Systems, Quasi-Newton Updates and Matrix Inversion

Size: px
Start display at page:

Download "Stochas(c Dual Ascent Linear Systems, Quasi-Newton Updates and Matrix Inversion"

Transcription

1 Stochas(c Dual Ascent Linear Systems, Quasi-Newton Updates and Matrix Inversion Peter Richtárik (joint work with Robert M. Gower) University of Edinburgh Oberwolfach, March 8, 2016

2 Part I Stochas(c Dual Ascent for Linear Systems

3 Robert Mansel Gower (Edinburgh) Robert Mansel Gower and P.R. Randomized Itera.ve Methods for Linear Systems SIAM Journal on Matrix Analysis and Applica5ons 36(4): , 2015 [GR 15a] Robert Mansel Gower and P.R. Stochas.c Dual Ascent for Solving Linear Systems arxiv: , 2015 [GR 15b]

4 The Problem

5 The Problem: Solve a Linear System n 2 R n m Ax = b m Assump.on 1 The system is consistent (i.e., has a solu(on)

6 Op(miza(on Formula(on Primal Problem B 0 minimize P (x) := 1 2 kx ck2 B subject to Ax = b A 2 R m n x 2 R n 1 2 (x c)> B(x c) Dual Problem Unconstrained non-strongly concave quadra(c maximiza(on problem maximize D(y) :=(b Ac) > 1 y 2 ka> yk 2 B 1 subject to y 2 R m

7 Dual Correspondence Lemma Lemma (GR 15b) (Any) dual op(mal point A ne mapping from R m to R n x(y) :=c + B 1 A > y Primal op(mal point D(y ) D(y) = 1 2 kx(y) x k 2 B Dual error (in func(on values) Primal error (in distance)

8 New Algorithm: Stochas(c Dual Ascent (SDA)

9 Stochas(c Dual Ascent A random m matrix drawn i.i.d. in each itera(on S D y t+1 = y t + S t m m m Moore-Penrose pseudo-inverse of a small matrix t := arg min 2Q t k k 2 Q t := arg max D(y t + S ) t = S > AB 1 A > S S > b A c + B 1 A > y t

10 Primal Method = Linear Image of the Dual Method x t := x(y t )=c + B 1 A > y t Corresponding primal iterates Dual iterates produced by SDA

11 Main Assump(on Assump.on 2 The matrix E S D h S S > AB 1 A > S S >i is nonsingular H

12 Complexity := 1 + min B 1/2 A > E[H]AB 1/2 of SDA U 0 = 1 2 kx0 x k 2 B Theorem (GR 15b) Primal iterates: E 1 2 kxt x k 2 B apple t U 0 GR 15a Residual: E[kAx t bk B ] apple t/2 kak B p 2 U0 Dual error: E[OPT D(y t )] apple t U 0 Primal error: E[P (x t ) OPT] apple t U 0 +2 t/2p OPT U 0 Duality gap: E[P (x t ) D(y t )] apple 2 t U 0 +2 t/2p OPT U 0

13 The Rate: Lower and Upper Bounds Rank(S > A)=dim(Range(B 1 A > S)) = Tr(B 1 Z) Theorem [RG 15ab] 0 apple 1 Rank(S > A) Rank(A) apple < 1 Insight: apple 1 always < 1 if Assumption 2 holds Insight: The lower bound is good when: i) the dimension of the search space in the constrain and approximate viewpoint is large, ii) the rank of A is small

14 The Primal Iterates: 6 Equivalent Viewpoints x t := x(y t )=c + B 1 A > y t Corresponding primal iterates Dual iterates produced by SDA

15 1. Relaxa(on Viewpoint Sketch and Project x t+1 = arg min x2r n kx xt k 2 B subject to S > Ax = S > b S = iden(ty matrix convergence in 1 step

16 2. Approxima(on Viewpoint Constrain and Approximate x t+1 = arg min x2r n kx x k 2 B subject to x = x t + B 1 A > S is free

17 3. Geometric Viewpoint Random Intersect x t (1) x + Null(S > A) x (2) x t+1 x t + Range(B 1 A > S) (1) x t+1 = arg min kx x x t k B subject to S > Ax = S > b (2) x t+1 = arg min kx x x k B subject to x = x t + B 1 A > S {x t+1 } = x + Null(S > A) \ x t + Range(B 1 A > S)

18 4. Algebraic Viewpoint Random Linear Solve x t+1 = solution in x of the linear system S > Ax = S > b x = x t + B 1 A > S Unknown Unknown

19 5. Algebraic Viewpoint Random Update Random Update Vector x t+1 = x t B 1 A > S(S > AB 1 A > S) S > (Ax t b) Moore-Penrose pseudo-inverse

20 6. Analy(c Viewpoint Random Fixed Point Z := A > S(S > AB 1 A > S) S > A x t+1 x = (I B 1 Z)(x t x ) Random Itera(on Matrix (B 1 Z) 2 = B 1 Z (I B 1 Z) 2 = I B 1 Z B 1 Z projects orthogonally onto Range(B 1 A > S) I B 1 Z projects orthogonally onto Null(S > A)

21 Special Case: Randomized Kaczmarz Method

22 Randomized Kaczmarz (RK) Method M. S. Kaczmarz. Angenaherte Auflosung von Systemen linearer Gleichungen, Bulle5n Interna5onal de l Académie Polonaise des Sciences et des LeKres. Classe des Sciences Mathéma5ques et Naturelles. Série A, Sciences Mathéma5ques 35, pp , 1937 Kaczmarz method (1937) T. Strohmer and R. Vershynin. A Randomized Kaczmarz Algorithm with Exponen.al Convergence. Journal of Fourier Analysis and Applica5ons 15(2), pp , 2009 Randomized Kaczmarz method (2009) RK arises as a special case for parameters B, S set as follows: B = I S = e i =(0,...,0, 1, 0,...,0) with probability p i x t+1 = x t A i: x t b i ka i: k 2 2 (A i: ) T RK was analyzed for p i = ka i:k 2 kak 2 F

23 RK: Deriva(on and Rate General Method x t+1 = x t B 1 A T S (S T AB 1 A T S) S T (Ax t b) Special Choice of Parameters P(S = e i )=p i Complexity Rate B = I x t+1 = x t A i:xt b i S = e i ka i: k 2 (A i: ) T 2 p i = ka i:k 2 kak 2 F E kx t x k 2 min A T A 2 apple 1 kak 2 F! t kx 0 x k 2 2

24 RK: Further Reading D. Needell. Randomized Kaczmarz solver for noisy linear systems. BIT 50 (2), pp , 2010 D. Needell and J. Tropp. Paved with good inten.ons: analyzis of a randomized block Kaczmarz method. Linear Algebra and its Applica5ons 441, pp , 2012 D. Needell, N. Srebro and R. Ward. Stochas.c gradient descent, weighted sampling and the randomized Kaczmarz algorithm. Mathema5cal Programming, 2015 (arxiv: ) A. Ramdas. Rows vs Columns for Linear Systems of Equa.ons Randomized Kaczmarz or Coordinate Descent? arxiv: , 2014

25 Special Case: Gaussian Descent

26 Gaussian Descent General Method x t+1 = x t B 1 A T S (S T AB 1 A T S) S T (Ax t b) Special Choice of Parameters S N(0, ) x t+1 = x t S T (Ax t b) S T AB 1 A T S B 1 A T S Posi(ve definite covariance matrix Complexity Rate E kx t x k 2 B apple t kx 0 x k 2 B

27 x t+1 = x t h t B 1/2 := B 1/2 A T S N(0, ) x := B 1/2 A T AB 1/2 x 0

28 Gaussian Descent: The Rate Lemma [GR 15] E apple T k k Tr( ) 1 1 n apple apple 1 2 min( ) Tr( ) This follows from the general lower

29 Gaussian Descent: Further Reading Yurii Nesterov. Random gradient-free minimiza.on of convex func.ons. CORE Discussion Paper # 2011/1, 2011 S. U. S(tch, C. L. Muller and G. Gartner. Op.miza.on of convex func.ons with random pursuit. SIAM Journal on Op(miza(on 23 (2), pp , 2014 S. U. S(tch. Convex op.miza.on with random pursuit. PhD Thesis, ETH Zurich, 2014

30 Special Case: Randomized Coordinate Descent

31 Randomized Coordinate Descent (RCD) A. S. Lewis and D. Leventhal. Randomized methods for linear constraints: convergence rates and condi.oning. Mathema5cs of OR 35(3), , 2010 (arxiv: ) min f(x) = 1 x2r n 2 xt Ax b T x Assume: Posi(ve definite S = e i =(0,...,0, 1, 0,...,0) with probability p i RCD (2008) RCD arises as a special case for parameters B, S set as follows: B = A x = A 1 b Recall: In RK we had B = I x t+1 = x t (A i: ) T x t b i A ii e i RCD was analyzed for p i = A ii Tr(A)

32 RCD: Deriva(on and Rate General Method x t+1 = x t B 1 A T S (S T AB 1 A T S) S T (Ax t b) Special Choice of Parameters B = A P(S = e i )=p i Complexity Rate S = e i x t+1 = x t (A i: ) T x t b i A ii e i p i = A ii Tr(A) E kx t x k 2 A apple 1 t min(a) kx 0 x k 2 A Tr(A)

33 RCD: Standard Op(miza(on Form Yurii Nesterov. Efficiency of coordinate descent methods on huge-scale op.miza.on problems. SIAM J. on Op5miza5on, 22(2): , 2012 (CORE Discussion Paper 2010/2) Nesterov considered the problem: min f(x) x2r n Convex and smooth Nesterov assumed that the following inequality holds for all x, h and i: Given a current iterate x, choosing h by minimizing the RHS gives: Nesterov s RCD method: x t+1 = x t 1 L i r i f(x t )e i f(x + he i ) apple f(x)+r i f(x)h + L i 2 h2 f(x) = 1 2 xt Ax b T x ) L i = A ii r i f(x) =(A i: ) T x We recover RCD as we have seen it: x t+1 = x t (A i: ) T x t b i A ii e i b i

34 Special Case: Randomized Newton Method

35 Randomized Newton (RN) Z. Qu, PR, M. Takáč and O. Fercoq. Stochas.c Dual Newton Ascent for Empirical Risk Minimiza.on. arxiv: , 2015 min f(x) = 1 x2r n 2 xt Ax b T x Assume: Posi(ve definite RN arises as a special case for parameters B, S set as follows: B = A x = A 1 b S = I :C with probability p C p C 0 8C {1,...,n} X C {1,...,n} p C =1 RCD is special case with p C =0whenever C 6= 1

36 RN: Deriva(on General Method x t+1 = x t B 1 A T S (S T AB 1 A T S) S T (Ax t b) Special Choice of Parameters B = A S = I :C with probability p C x t+1 = x t I :C ((I :C ) T AI :C ) 1 (I :C ) T (Ax t b) This method minimizes f exactly in a random subspace spanned by the coordinates belonging to C

37 e 7 C = {2, 7} C =2 x t+1 x t e 2

38 Summary: Linear Systems SDA: A new class of randomized op(miza(on algorithms Extremely versa(le Works for almost any random S Get several exis(ng algorithms in special cases (RK, RCD, RN, RBK) Get many new algorithms in special cases Linear convergence despite lack of strong concavity RK in the primal = RCD in the dual Did not talk about: Randomized gossip Distributed variant Op(mal sampling via SDP Experiments

39

40 Part II Stochas(c Dual Ascent for Matrix Inversion

41 Robert Mansel Gower (Edinburgh) Robert Mansel Gower and P.R. Randomized Quasi-Newton Methods are Linearly Convergent Matrix Inversion Algorithms arxiv: , 2016

42 The Problem: Invert a Matrix n 2 R n n Iden(ty matrix n AX = I Assump.on 1 Matrix A is inver(ble

43 Inver(ng Symmetric Matrices

44 1. Sketch and Project kxk F (B) := p Tr(X > BXB) X t+1 = arg min kx Xt X2Rn n k 2 F (B) subject to S > AX = S >, X = X > Quasi-Newton updates are of this form: S = determinis(c column vector We get randomized block version of quasi-newton updates! Randomized quasi-newton updates are linearly convergent matrix inversion methods Interpreta(on: Gaussian Inference (Henning, 2015) Donald Goldfarb. A Family of Variable-Metric Methods Derived by Varia.onal Means. Mathema5cs of Computa5on 24(109), 1970

45 Gaussian Inference Philipp Henning Probabilis.c Interpreta.on of Linear Solvers SIAM Journal on Op5miza5on 25(1): , 2015 X k+1 The new iterate can be interpreted as the mean of a posterior distribu(on under a Gaussian prior with mean and noiseless (and random) linear observa(on of X k A 1

46 Randomized QN Updates B Equation Method I AX = I Powel-Symmetric-Broyden (PSB) A 1 XA 1 = I Davidon-Fletcher-Powell (DFP) A AX = I Broyden-Fletcher-Goldfarb-Shanno (BFGS) All these QN methods arise as special cases of our framework All are linearly convergent, with explicit convergence rates We also recover non-symmetric updates such as Bad Broyden and Good Broyden We get block versions We get randomized versions of new QN updates

47 2. Constrain and Approximate X t+1 = arg min kx A X2Rn n 1 k 2 F (B) s.t. X = X t + S > AB 1 + B 1 A > S > 2 R n is free Randomized BFGS: B = A, =1 X t+1 = arg New formula.on even for standard QN methods min X2R n n kx A 1 k 2 F (A) = kax Ik2 F s.t. X = X t + S > + S > 2 R n is free RBFGS performs best symmetric rank-2 update

48 4. Random Update H = S(S > AB 1 A > S) S > X t+1 = X t (X t A I)HAB 1 + B 1 AH(AX t I)(AHAB 1 I) 6. Random Fixed Point X t+1 A 1 = (I B 1 A > HA)(X t A 1 )(I AHA > B 1 )

49 Complexity / Convergence Theorem [GR 16] kmk B := kb 1/2 MB 1/2 k 2 1 E X t A 1 B apple t kx 0 A 1 k B 2 E[H] 0 < 1 E h kx t A 1 k 2 F (B) i apple t kx 0 A 1 k 2 F (B)

50 Summary: Matrix Inversion Block version of QN updates New points of view (constrain and approximate,...) New link between QN and approx. inverse precondi(oning First (me randomized QN updates are proposed First stochas(c method for matrix inversion (with complexity bounds)? Linear convergence under weak assump(ons Did not talk about: Nonsymmetric variants Theore(cal bounds for discretely distributed S Adap(ve randomized BFGS Limited memory and factored implementa(ons Experiments (Newton-Schultz; MinRes) Use in empirical risk minimiza(on (Gower, Goldfarb & R. 16) Extension to the computa(on of a pseudoinverse

51 The End

arxiv: v2 [math.na] 28 Jan 2016

arxiv: v2 [math.na] 28 Jan 2016 Stochastic Dual Ascent for Solving Linear Systems Robert M. Gower and Peter Richtárik arxiv:1512.06890v2 [math.na 28 Jan 2016 School of Mathematics University of Edinburgh United Kingdom December 21, 2015

More information

arxiv: v3 [math.na] 23 Mar 2016

arxiv: v3 [math.na] 23 Mar 2016 Randomized Quasi-Newton Updates are Linearly Convergent Matrix Inversion Algorithms Robert M. Gower and Peter Richtárik arxiv:602.0768v3 [math.na] 23 Mar 206 School of Mathematics University of Edinburgh

More information

Bellman s Curse of Dimensionality

Bellman s Curse of Dimensionality Bellman s Curse of Dimensionality n- dimensional state space Number of states grows exponen

More information

Randomized Quasi-Newton Updates are Linearly Convergent Matrix Inversion Algorithms

Randomized Quasi-Newton Updates are Linearly Convergent Matrix Inversion Algorithms Randomized Quasi-Newton Updates are Linearly Convergent Matrix Inversion Algorithms Robert M. Gower and Peter Richtárik School of Mathematics University of Edinburgh United Kingdom ebruary 4, 06 Abstract

More information

Sketch and Project: Randomized Iterative Methods for Linear Systems and Inverting Matrices

Sketch and Project: Randomized Iterative Methods for Linear Systems and Inverting Matrices Sketch and Project: Randomized Iterative Methods for Linear Systems and Inverting Matrices arxiv:1612.06013v1 [math.na] 19 Dec 2016 Robert Mansel Gower Supervisor: Dr. Peter Richtárik Thesis submitted

More information

A Sampling Kaczmarz-Motzkin Algorithm for Linear Feasibility

A Sampling Kaczmarz-Motzkin Algorithm for Linear Feasibility A Sampling Kaczmarz-Motzkin Algorithm for Linear Feasibility Jamie Haddock Graduate Group in Applied Mathematics, Department of Mathematics, University of California, Davis Copper Mountain Conference on

More information

Lecture 3: Minimizing Large Sums. Peter Richtárik

Lecture 3: Minimizing Large Sums. Peter Richtárik Lecture 3: Minimizing Large Sums Peter Richtárik Graduate School in Systems, Op@miza@on, Control and Networks Belgium 2015 Mo@va@on: Machine Learning & Empirical Risk Minimiza@on Training Linear Predictors

More information

On the exponential convergence of. the Kaczmarz algorithm

On the exponential convergence of. the Kaczmarz algorithm On the exponential convergence of the Kaczmarz algorithm Liang Dai and Thomas B. Schön Department of Information Technology, Uppsala University, arxiv:4.407v [cs.sy] 0 Mar 05 75 05 Uppsala, Sweden. E-mail:

More information

Convergence Rates for Greedy Kaczmarz Algorithms

Convergence Rates for Greedy Kaczmarz Algorithms onvergence Rates for Greedy Kaczmarz Algorithms Julie Nutini 1, Behrooz Sepehry 1, Alim Virani 1, Issam Laradji 1, Mark Schmidt 1, Hoyt Koepke 2 1 niversity of British olumbia, 2 Dato Abstract We discuss

More information

5 Quasi-Newton Methods

5 Quasi-Newton Methods Unconstrained Convex Optimization 26 5 Quasi-Newton Methods If the Hessian is unavailable... Notation: H = Hessian matrix. B is the approximation of H. C is the approximation of H 1. Problem: Solve min

More information

Edinburgh Research Explorer

Edinburgh Research Explorer Edinburgh Research Explorer Randomized Iterative Methods for Linear Systems Citation for published version: Gower, R & Richtarik, P 2015, 'Randomized Iterative Methods for Linear Systems' SIAM Journal

More information

Stochastic Reformulations of Linear Systems: Algorithms and Convergence Theory

Stochastic Reformulations of Linear Systems: Algorithms and Convergence Theory Stochastic Reformulations of Linear Systems: Algorithms and Convergence Theory Peter Richtárik Martin Takáč June 4, 2017 Abstract We develop a family of reformulations of an arbitrary consistent linear

More information

Quasi-Newton methods for minimization

Quasi-Newton methods for minimization Quasi-Newton methods for minimization Lectures for PHD course on Numerical optimization Enrico Bertolazzi DIMS Universitá di Trento November 21 December 14, 2011 Quasi-Newton methods for minimization 1

More information

Least Square Es?ma?on, Filtering, and Predic?on: ECE 5/639 Sta?s?cal Signal Processing II: Linear Es?ma?on

Least Square Es?ma?on, Filtering, and Predic?on: ECE 5/639 Sta?s?cal Signal Processing II: Linear Es?ma?on Least Square Es?ma?on, Filtering, and Predic?on: Sta?s?cal Signal Processing II: Linear Es?ma?on Eric Wan, Ph.D. Fall 2015 1 Mo?va?ons If the second-order sta?s?cs are known, the op?mum es?mator is given

More information

Convex Optimization CMU-10725

Convex Optimization CMU-10725 Convex Optimization CMU-10725 Quasi Newton Methods Barnabás Póczos & Ryan Tibshirani Quasi Newton Methods 2 Outline Modified Newton Method Rank one correction of the inverse Rank two correction of the

More information

Randomized Iterative Methods for Linear Systems

Randomized Iterative Methods for Linear Systems Randomized Iterative Methods for Linear Systems Robert M. Gower Peter Richtárik June 16, 015 Abstract We develop a novel, fundamental and surprisingly simple randomized iterative method for solving consistent

More information

SGD and Randomized projection algorithms for overdetermined linear systems

SGD and Randomized projection algorithms for overdetermined linear systems SGD and Randomized projection algorithms for overdetermined linear systems Deanna Needell Claremont McKenna College IPAM, Feb. 25, 2014 Includes joint work with Eldar, Ward, Tropp, Srebro-Ward Setup Setup

More information

Newton s Method. Ryan Tibshirani Convex Optimization /36-725

Newton s Method. Ryan Tibshirani Convex Optimization /36-725 Newton s Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, Properties and examples: f (y) = max x

More information

Lecture 14: October 17

Lecture 14: October 17 1-725/36-725: Convex Optimization Fall 218 Lecture 14: October 17 Lecturer: Lecturer: Ryan Tibshirani Scribes: Pengsheng Guo, Xian Zhou Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:

More information

Randomized projection algorithms for overdetermined linear systems

Randomized projection algorithms for overdetermined linear systems Randomized projection algorithms for overdetermined linear systems Deanna Needell Claremont McKenna College ISMP, Berlin 2012 Setup Setup Let Ax = b be an overdetermined, standardized, full rank system

More information

Acceleration of Randomized Kaczmarz Method

Acceleration of Randomized Kaczmarz Method Acceleration of Randomized Kaczmarz Method Deanna Needell [Joint work with Y. Eldar] Stanford University BIRS Banff, March 2011 Problem Background Setup Setup Let Ax = b be an overdetermined consistent

More information

Optimization II: Unconstrained Multivariable

Optimization II: Unconstrained Multivariable Optimization II: Unconstrained Multivariable CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Justin Solomon CS 205A: Mathematical Methods Optimization II: Unconstrained Multivariable 1

More information

Nonlinear Optimization: What s important?

Nonlinear Optimization: What s important? Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global

More information

Stochastic Gradient Descent, Weighted Sampling, and the Randomized Kaczmarz algorithm

Stochastic Gradient Descent, Weighted Sampling, and the Randomized Kaczmarz algorithm Stochastic Gradient Descent, Weighted Sampling, and the Randomized Kaczmarz algorithm Deanna Needell Department of Mathematical Sciences Claremont McKenna College Claremont CA 97 dneedell@cmc.edu Nathan

More information

arxiv: v2 [math.na] 11 May 2017

arxiv: v2 [math.na] 11 May 2017 Rows vs. Columns: Randomized Kaczmarz or Gauss-Seidel for Ridge Regression arxiv:1507.05844v2 [math.na] 11 May 2017 Ahmed Hefny Machine Learning Department Carnegie Mellon University Deanna Needell Department

More information

Nonlinear Programming

Nonlinear Programming Nonlinear Programming Kees Roos e-mail: C.Roos@ewi.tudelft.nl URL: http://www.isa.ewi.tudelft.nl/ roos LNMB Course De Uithof, Utrecht February 6 - May 8, A.D. 2006 Optimization Group 1 Outline for week

More information

MATH 4211/6211 Optimization Quasi-Newton Method

MATH 4211/6211 Optimization Quasi-Newton Method MATH 4211/6211 Optimization Quasi-Newton Method Xiaojing Ye Department of Mathematics & Statistics Georgia State University Xiaojing Ye, Math & Stat, Georgia State University 0 Quasi-Newton Method Motivation:

More information

Optimization II: Unconstrained Multivariable

Optimization II: Unconstrained Multivariable Optimization II: Unconstrained Multivariable CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Doug James (and Justin Solomon) CS 205A: Mathematical Methods Optimization II: Unconstrained

More information

Quasi-Newton Methods. Javier Peña Convex Optimization /36-725

Quasi-Newton Methods. Javier Peña Convex Optimization /36-725 Quasi-Newton Methods Javier Peña Convex Optimization 10-725/36-725 Last time: primal-dual interior-point methods Consider the problem min x subject to f(x) Ax = b h(x) 0 Assume f, h 1,..., h m are convex

More information

Chapter 4. Unconstrained optimization

Chapter 4. Unconstrained optimization Chapter 4. Unconstrained optimization Version: 28-10-2012 Material: (for details see) Chapter 11 in [FKS] (pp.251-276) A reference e.g. L.11.2 refers to the corresponding Lemma in the book [FKS] PDF-file

More information

Pseudospectral Methods For Op2mal Control. Jus2n Ruths March 27, 2009

Pseudospectral Methods For Op2mal Control. Jus2n Ruths March 27, 2009 Pseudospectral Methods For Op2mal Control Jus2n Ruths March 27, 2009 Introduc2on Pseudospectral methods arose to find solu2ons to Par2al Differen2al Equa2ons Recently adapted for Op2mal Control Key Ideas

More information

Programming, numerics and optimization

Programming, numerics and optimization Programming, numerics and optimization Lecture C-3: Unconstrained optimization II Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428

More information

Determinant maximization with linear. S. Boyd, L. Vandenberghe, S.-P. Wu. Information Systems Laboratory. Stanford University

Determinant maximization with linear. S. Boyd, L. Vandenberghe, S.-P. Wu. Information Systems Laboratory. Stanford University Determinant maximization with linear matrix inequality constraints S. Boyd, L. Vandenberghe, S.-P. Wu Information Systems Laboratory Stanford University SCCM Seminar 5 February 1996 1 MAXDET problem denition

More information

Quasi-Newton Methods

Quasi-Newton Methods Newton s Method Pros and Cons Quasi-Newton Methods MA 348 Kurt Bryan Newton s method has some very nice properties: It s extremely fast, at least once it gets near the minimum, and with the simple modifications

More information

Methods that avoid calculating the Hessian. Nonlinear Optimization; Steepest Descent, Quasi-Newton. Steepest Descent

Methods that avoid calculating the Hessian. Nonlinear Optimization; Steepest Descent, Quasi-Newton. Steepest Descent Nonlinear Optimization Steepest Descent and Niclas Börlin Department of Computing Science Umeå University niclas.borlin@cs.umu.se A disadvantage with the Newton method is that the Hessian has to be derived

More information

2. Quasi-Newton methods

2. Quasi-Newton methods L. Vandenberghe EE236C (Spring 2016) 2. Quasi-Newton methods variable metric methods quasi-newton methods BFGS update limited-memory quasi-newton methods 2-1 Newton method for unconstrained minimization

More information

Randomized Coordinate Descent Methods on Optimization Problems with Linearly Coupled Constraints

Randomized Coordinate Descent Methods on Optimization Problems with Linearly Coupled Constraints Randomized Coordinate Descent Methods on Optimization Problems with Linearly Coupled Constraints By I. Necoara, Y. Nesterov, and F. Glineur Lijun Xu Optimization Group Meeting November 27, 2012 Outline

More information

Op#mal convex op#miza#on under Tsybakov noise through connec#ons to ac#ve learning

Op#mal convex op#miza#on under Tsybakov noise through connec#ons to ac#ve learning Op#mal convex op#miza#on under Tsybakov noise through connec#ons to ac#ve learning Aar$ Singh Joint work with: Aaditya Ramdas Connec#ons between convex op#miza#on and ac#ve learning (a formal reduc#on)

More information

Optimization and Root Finding. Kurt Hornik

Optimization and Root Finding. Kurt Hornik Optimization and Root Finding Kurt Hornik Basics Root finding and unconstrained smooth optimization are closely related: Solving ƒ () = 0 can be accomplished via minimizing ƒ () 2 Slide 2 Basics Root finding

More information

Deriva'on of The Kalman Filter. Fred DePiero CalPoly State University EE 525 Stochas'c Processes

Deriva'on of The Kalman Filter. Fred DePiero CalPoly State University EE 525 Stochas'c Processes Deriva'on of The Kalman Filter Fred DePiero CalPoly State University EE 525 Stochas'c Processes KF Uses State Predic'ons KF es'mates the state of a system Example Measure: posi'on State: [ posi'on velocity

More information

Quasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb Shanno February 6, / 25 (BFG. Limited memory BFGS (L-BFGS)

Quasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb Shanno February 6, / 25 (BFG. Limited memory BFGS (L-BFGS) Quasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb Shanno (BFGS) Limited memory BFGS (L-BFGS) February 6, 2014 Quasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb

More information

Lecture 18: November Review on Primal-dual interior-poit methods

Lecture 18: November Review on Primal-dual interior-poit methods 10-725/36-725: Convex Optimization Fall 2016 Lecturer: Lecturer: Javier Pena Lecture 18: November 2 Scribes: Scribes: Yizhu Lin, Pan Liu Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:

More information

Convergence Properties of the Randomized Extended Gauss-Seidel and Kaczmarz Methods

Convergence Properties of the Randomized Extended Gauss-Seidel and Kaczmarz Methods Claremont Colleges Scholarship @ Claremont CMC Faculty Publications and Research CMC Faculty Scholarship 11-1-015 Convergence Properties of the Randomized Extended Gauss-Seidel and Kaczmarz Methods Anna

More information

Optimisation in Higher Dimensions

Optimisation in Higher Dimensions CHAPTER 6 Optimisation in Higher Dimensions Beyond optimisation in 1D, we will study two directions. First, the equivalent in nth dimension, x R n such that f(x ) f(x) for all x R n. Second, constrained

More information

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 3. Gradient Method

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 3. Gradient Method Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 3 Gradient Method Shiqian Ma, MAT-258A: Numerical Optimization 2 3.1. Gradient method Classical gradient method: to minimize a differentiable convex

More information

Randomized Coordinate Descent with Arbitrary Sampling: Algorithms and Complexity

Randomized Coordinate Descent with Arbitrary Sampling: Algorithms and Complexity Randomized Coordinate Descent with Arbitrary Sampling: Algorithms and Complexity Zheng Qu University of Hong Kong CAM, 23-26 Aug 2016 Hong Kong based on joint work with Peter Richtarik and Dominique Cisba(University

More information

Quasi-Newton Methods

Quasi-Newton Methods Quasi-Newton Methods Werner C. Rheinboldt These are excerpts of material relating to the boos [OR00 and [Rhe98 and of write-ups prepared for courses held at the University of Pittsburgh. Some further references

More information

March 8, 2010 MATH 408 FINAL EXAM SAMPLE

March 8, 2010 MATH 408 FINAL EXAM SAMPLE March 8, 200 MATH 408 FINAL EXAM SAMPLE EXAM OUTLINE The final exam for this course takes place in the regular course classroom (MEB 238) on Monday, March 2, 8:30-0:20 am. You may bring two-sided 8 page

More information

Learning Theory of Randomized Kaczmarz Algorithm

Learning Theory of Randomized Kaczmarz Algorithm Journal of Machine Learning Research 16 015 3341-3365 Submitted 6/14; Revised 4/15; Published 1/15 Junhong Lin Ding-Xuan Zhou Department of Mathematics City University of Hong Kong 83 Tat Chee Avenue Kowloon,

More information

WHY DUALITY? Gradient descent Newton s method Quasi-newton Conjugate gradients. No constraints. Non-differentiable ???? Constrained problems? ????

WHY DUALITY? Gradient descent Newton s method Quasi-newton Conjugate gradients. No constraints. Non-differentiable ???? Constrained problems? ???? DUALITY WHY DUALITY? No constraints f(x) Non-differentiable f(x) Gradient descent Newton s method Quasi-newton Conjugate gradients etc???? Constrained problems? f(x) subject to g(x) apple 0???? h(x) =0

More information

FAST DISTRIBUTED COORDINATE DESCENT FOR NON-STRONGLY CONVEX LOSSES. Olivier Fercoq Zheng Qu Peter Richtárik Martin Takáč

FAST DISTRIBUTED COORDINATE DESCENT FOR NON-STRONGLY CONVEX LOSSES. Olivier Fercoq Zheng Qu Peter Richtárik Martin Takáč FAST DISTRIBUTED COORDINATE DESCENT FOR NON-STRONGLY CONVEX LOSSES Olivier Fercoq Zheng Qu Peter Richtárik Martin Takáč School of Mathematics, University of Edinburgh, Edinburgh, EH9 3JZ, United Kingdom

More information

Sparse Covariance Selection using Semidefinite Programming

Sparse Covariance Selection using Semidefinite Programming Sparse Covariance Selection using Semidefinite Programming A. d Aspremont ORFE, Princeton University Joint work with O. Banerjee, L. El Ghaoui & G. Natsoulis, U.C. Berkeley & Iconix Pharmaceuticals Support

More information

Quasi-Newton Methods. Zico Kolter (notes by Ryan Tibshirani, Javier Peña, Zico Kolter) Convex Optimization

Quasi-Newton Methods. Zico Kolter (notes by Ryan Tibshirani, Javier Peña, Zico Kolter) Convex Optimization Quasi-Newton Methods Zico Kolter (notes by Ryan Tibshirani, Javier Peña, Zico Kolter) Convex Optimization 10-725 Last time: primal-dual interior-point methods Given the problem min x f(x) subject to h(x)

More information

Stochastic Optimization Algorithms Beyond SG

Stochastic Optimization Algorithms Beyond SG Stochastic Optimization Algorithms Beyond SG Frank E. Curtis 1, Lehigh University involving joint work with Léon Bottou, Facebook AI Research Jorge Nocedal, Northwestern University Optimization Methods

More information

Algorithms for Constrained Optimization

Algorithms for Constrained Optimization 1 / 42 Algorithms for Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University April 19, 2015 2 / 42 Outline 1. Convergence 2. Sequential quadratic

More information

Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2

Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2 Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2 Coralia Cartis, University of Oxford INFOMM CDT: Modelling, Analysis and Computation of Continuous Real-World Problems Methods

More information

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications Professor M. Chiang Electrical Engineering Department, Princeton University March

More information

Numerical Methods in Physics

Numerical Methods in Physics Numerical Methods in Physics Numerische Methoden in der Physik, 515.421. Instructor: Ass. Prof. Dr. Lilia Boeri Room: PH 03 090 Tel: +43-316- 873 8191 Email Address: l.boeri@tugraz.at Room: TDK Seminarraum

More information

Randomized Kaczmarz Nick Freris EPFL

Randomized Kaczmarz Nick Freris EPFL Randomized Kaczmarz Nick Freris EPFL (Joint work with A. Zouzias) Outline Randomized Kaczmarz algorithm for linear systems Consistent (noiseless) Inconsistent (noisy) Optimal de-noising Convergence analysis

More information

Tsybakov noise adap/ve margin- based ac/ve learning

Tsybakov noise adap/ve margin- based ac/ve learning Tsybakov noise adap/ve margin- based ac/ve learning Aar$ Singh A. Nico Habermann Associate Professor NIPS workshop on Learning Faster from Easy Data II Dec 11, 2015 Passive Learning Ac/ve Learning (X j,?)

More information

The speed of Shor s R-algorithm

The speed of Shor s R-algorithm IMA Journal of Numerical Analysis 2008) 28, 711 720 doi:10.1093/imanum/drn008 Advance Access publication on September 12, 2008 The speed of Shor s R-algorithm J. V. BURKE Department of Mathematics, University

More information

Optimization: Nonlinear Optimization without Constraints. Nonlinear Optimization without Constraints 1 / 23

Optimization: Nonlinear Optimization without Constraints. Nonlinear Optimization without Constraints 1 / 23 Optimization: Nonlinear Optimization without Constraints Nonlinear Optimization without Constraints 1 / 23 Nonlinear optimization without constraints Unconstrained minimization min x f(x) where f(x) is

More information

Mathematical Optimisation, Chpt 2: Linear Equations and inequalities

Mathematical Optimisation, Chpt 2: Linear Equations and inequalities Mathematical Optimisation, Chpt 2: Linear Equations and inequalities Peter J.C. Dickinson p.j.c.dickinson@utwente.nl http://dickinson.website version: 12/02/18 Monday 5th February 2018 Peter J.C. Dickinson

More information

Machine Learning Lecture Notes

Machine Learning Lecture Notes Machine Learning Lecture Notes Predrag Radivojac February 2, 205 Given a data set D = {(x i,y i )} n the objective is to learn the relationship between features and the target. We usually start by hypothesizing

More information

CS6750: Cryptography and Communica7on Security

CS6750: Cryptography and Communica7on Security CS6750: Cryptography and Communica7on Security Class 6: Simple Number Theory Dr. Erik- Oliver Blass Plan 1. Role of number theory in cryptography 2. Classical problems in computa7onal number theory 3.

More information

Stochastic Quasi-Newton Methods

Stochastic Quasi-Newton Methods Stochastic Quasi-Newton Methods Donald Goldfarb Department of IEOR Columbia University UCLA Distinguished Lecture Series May 17-19, 2016 1 / 35 Outline Stochastic Approximation Stochastic Gradient Descent

More information

Convex Optimization. Problem set 2. Due Monday April 26th

Convex Optimization. Problem set 2. Due Monday April 26th Convex Optimization Problem set 2 Due Monday April 26th 1 Gradient Decent without Line-search In this problem we will consider gradient descent with predetermined step sizes. That is, instead of determining

More information

Unconstrained optimization

Unconstrained optimization Chapter 4 Unconstrained optimization An unconstrained optimization problem takes the form min x Rnf(x) (4.1) for a target functional (also called objective function) f : R n R. In this chapter and throughout

More information

On Nesterov s Random Coordinate Descent Algorithms - Continued

On Nesterov s Random Coordinate Descent Algorithms - Continued On Nesterov s Random Coordinate Descent Algorithms - Continued Zheng Xu University of Texas At Arlington February 20, 2015 1 Revisit Random Coordinate Descent The Random Coordinate Descent Upper and Lower

More information

Input: System of inequalities or equalities over the reals R. Output: Value for variables that minimizes cost function

Input: System of inequalities or equalities over the reals R. Output: Value for variables that minimizes cost function Linear programming Input: System of inequalities or equalities over the reals R A linear cost function Output: Value for variables that minimizes cost function Example: Minimize 6x+4y Subject to 3x + 2y

More information

Written Examination

Written Examination Division of Scientific Computing Department of Information Technology Uppsala University Optimization Written Examination 202-2-20 Time: 4:00-9:00 Allowed Tools: Pocket Calculator, one A4 paper with notes

More information

Linear and non-linear programming

Linear and non-linear programming Linear and non-linear programming Benjamin Recht March 11, 2005 The Gameplan Constrained Optimization Convexity Duality Applications/Taxonomy 1 Constrained Optimization minimize f(x) subject to g j (x)

More information

Fundamentals of Data Assimila1on

Fundamentals of Data Assimila1on 014 GSI Community Tutorial NCAR Foothills Campus, Boulder, CO July 14-16, 014 Fundamentals of Data Assimila1on Milija Zupanski Cooperative Institute for Research in the Atmosphere Colorado State University

More information

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE CONVEX ANALYSIS AND DUALITY Basic concepts of convex analysis Basic concepts of convex optimization Geometric duality framework - MC/MC Constrained optimization

More information

COMP 562: Introduction to Machine Learning

COMP 562: Introduction to Machine Learning COMP 562: Introduction to Machine Learning Lecture 20 : Support Vector Machines, Kernels Mahmoud Mostapha 1 Department of Computer Science University of North Carolina at Chapel Hill mahmoudm@cs.unc.edu

More information

On Nesterov s Random Coordinate Descent Algorithms

On Nesterov s Random Coordinate Descent Algorithms On Nesterov s Random Coordinate Descent Algorithms Zheng Xu University of Texas At Arlington February 19, 2015 1 Introduction Full-Gradient Descent Coordinate Descent 2 Random Coordinate Descent Algorithm

More information

Comparative study of Optimization methods for Unconstrained Multivariable Nonlinear Programming Problems

Comparative study of Optimization methods for Unconstrained Multivariable Nonlinear Programming Problems International Journal of Scientific and Research Publications, Volume 3, Issue 10, October 013 1 ISSN 50-3153 Comparative study of Optimization methods for Unconstrained Multivariable Nonlinear Programming

More information

Reduced Models for Process Simula2on and Op2miza2on

Reduced Models for Process Simula2on and Op2miza2on Reduced Models for Process Simulaon and Opmizaon Yidong Lang, Lorenz T. Biegler and David Miller ESI annual meeng March, 0 Models are mapping Equaon set or Module simulators Input space Reduced model Surrogate

More information

The Dual Lattice, Integer Linear Systems and Hermite Normal Form

The Dual Lattice, Integer Linear Systems and Hermite Normal Form New York University, Fall 2013 Lattices, Convexity & Algorithms Lecture 2 The Dual Lattice, Integer Linear Systems and Hermite Normal Form Lecturers: D. Dadush, O. Regev Scribe: D. Dadush 1 Dual Lattice

More information

Mini-Batch Primal and Dual Methods for SVMs

Mini-Batch Primal and Dual Methods for SVMs Mini-Batch Primal and Dual Methods for SVMs Peter Richtárik School of Mathematics The University of Edinburgh Coauthors: M. Takáč (Edinburgh), A. Bijral and N. Srebro (both TTI at Chicago) arxiv:1303.2314

More information

Chapter 2. Optimization. Gradients, convexity, and ALS

Chapter 2. Optimization. Gradients, convexity, and ALS Chapter 2 Optimization Gradients, convexity, and ALS Contents Background Gradient descent Stochastic gradient descent Newton s method Alternating least squares KKT conditions 2 Motivation We can solve

More information

Agenda. 1 Duality for LP. 2 Theorem of alternatives. 3 Conic Duality. 4 Dual cones. 5 Geometric view of cone programs. 6 Conic duality theorem

Agenda. 1 Duality for LP. 2 Theorem of alternatives. 3 Conic Duality. 4 Dual cones. 5 Geometric view of cone programs. 6 Conic duality theorem Agenda 1 Duality for LP 2 Theorem of alternatives 3 Conic Duality 4 Dual cones 5 Geometric view of cone programs 6 Conic duality theorem 7 Examples Lower bounds on LPs By eliminating variables (if needed)

More information

Methodological Foundations of Biomedical Informatics (BMSC-GA 4449) Optimization

Methodological Foundations of Biomedical Informatics (BMSC-GA 4449) Optimization Methodological Foundations of Biomedical Informatics (BMSCGA 4449) Optimization Op#miza#on A set of techniques for finding the values of variables at which the objec#ve func#on amains its cri#cal (minimal

More information

10-725/36-725: Convex Optimization Prerequisite Topics

10-725/36-725: Convex Optimization Prerequisite Topics 10-725/36-725: Convex Optimization Prerequisite Topics February 3, 2015 This is meant to be a brief, informal refresher of some topics that will form building blocks in this course. The content of the

More information

Outline Introduction: Problem Description Diculties Algebraic Structure: Algebraic Varieties Rank Decient Toeplitz Matrices Constructing Lower Rank St

Outline Introduction: Problem Description Diculties Algebraic Structure: Algebraic Varieties Rank Decient Toeplitz Matrices Constructing Lower Rank St Structured Lower Rank Approximation by Moody T. Chu (NCSU) joint with Robert E. Funderlic (NCSU) and Robert J. Plemmons (Wake Forest) March 5, 1998 Outline Introduction: Problem Description Diculties Algebraic

More information

Lecture 7 Unconstrained nonlinear programming

Lecture 7 Unconstrained nonlinear programming Lecture 7 Unconstrained nonlinear programming Weinan E 1,2 and Tiejun Li 2 1 Department of Mathematics, Princeton University, weinan@princeton.edu 2 School of Mathematical Sciences, Peking University,

More information

NONSMOOTH VARIANTS OF POWELL S BFGS CONVERGENCE THEOREM

NONSMOOTH VARIANTS OF POWELL S BFGS CONVERGENCE THEOREM NONSMOOTH VARIANTS OF POWELL S BFGS CONVERGENCE THEOREM JIAYI GUO AND A.S. LEWIS Abstract. The popular BFGS quasi-newton minimization algorithm under reasonable conditions converges globally on smooth

More information

Fundamentals of Data Assimila1on

Fundamentals of Data Assimila1on 2015 GSI Community Tutorial NCAR Foothills Campus, Boulder, CO August 11-14, 2015 Fundamentals of Data Assimila1on Milija Zupanski Cooperative Institute for Research in the Atmosphere Colorado State University

More information

ORIE 6326: Convex Optimization. Quasi-Newton Methods

ORIE 6326: Convex Optimization. Quasi-Newton Methods ORIE 6326: Convex Optimization Quasi-Newton Methods Professor Udell Operations Research and Information Engineering Cornell April 10, 2017 Slides on steepest descent and analysis of Newton s method adapted

More information

Numerical Methods I: Orthogonalization and Newton s method

Numerical Methods I: Orthogonalization and Newton s method 1/42 Numerical Methods I: Orthogonalization and Newton s method Georg Stadler Courant Institute, NYU stadler@cims.nyu.edu October 5, 2017 Overview 2/42 Linear least squares and orthogonalization methods

More information

LMI Methods in Optimal and Robust Control

LMI Methods in Optimal and Robust Control LMI Methods in Optimal and Robust Control Matthew M. Peet Arizona State University Lecture 02: Optimization (Convex and Otherwise) What is Optimization? An Optimization Problem has 3 parts. x F f(x) :

More information

Parallel Coordinate Optimization

Parallel Coordinate Optimization 1 / 38 Parallel Coordinate Optimization Julie Nutini MLRG - Spring Term March 6 th, 2018 2 / 38 Contours of a function F : IR 2 IR. Goal: Find the minimizer of F. Coordinate Descent in 2D Contours of a

More information

More First-Order Optimization Algorithms

More First-Order Optimization Algorithms More First-Order Optimization Algorithms Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye Chapters 3, 8, 3 The SDM

More information

Dual and primal-dual methods

Dual and primal-dual methods ELE 538B: Large-Scale Optimization for Data Science Dual and primal-dual methods Yuxin Chen Princeton University, Spring 2018 Outline Dual proximal gradient method Primal-dual proximal gradient method

More information

Improving L-BFGS Initialization for Trust-Region Methods in Deep Learning

Improving L-BFGS Initialization for Trust-Region Methods in Deep Learning Improving L-BFGS Initialization for Trust-Region Methods in Deep Learning Jacob Rafati http://rafati.net jrafatiheravi@ucmerced.edu Ph.D. Candidate, Electrical Engineering and Computer Science University

More information

DS-GA 1002 Lecture notes 10 November 23, Linear models

DS-GA 1002 Lecture notes 10 November 23, Linear models DS-GA 2 Lecture notes November 23, 2 Linear functions Linear models A linear model encodes the assumption that two quantities are linearly related. Mathematically, this is characterized using linear functions.

More information

Reduced-Hessian Methods for Constrained Optimization

Reduced-Hessian Methods for Constrained Optimization Reduced-Hessian Methods for Constrained Optimization Philip E. Gill University of California, San Diego Joint work with: Michael Ferry & Elizabeth Wong 11th US & Mexico Workshop on Optimization and its

More information

Homework 5. Convex Optimization /36-725

Homework 5. Convex Optimization /36-725 Homework 5 Convex Optimization 10-725/36-725 Due Tuesday November 22 at 5:30pm submitted to Christoph Dann in Gates 8013 (Remember to a submit separate writeup for each problem, with your name at the top)

More information

Chapter 6 - Orthogonality

Chapter 6 - Orthogonality Chapter 6 - Orthogonality Maggie Myers Robert A. van de Geijn The University of Texas at Austin Orthogonality Fall 2009 http://z.cs.utexas.edu/wiki/pla.wiki/ 1 Orthogonal Vectors and Subspaces http://z.cs.utexas.edu/wiki/pla.wiki/

More information

Random coordinate descent algorithms for. huge-scale optimization problems. Ion Necoara

Random coordinate descent algorithms for. huge-scale optimization problems. Ion Necoara Random coordinate descent algorithms for huge-scale optimization problems Ion Necoara Automatic Control and Systems Engineering Depart. 1 Acknowledgement Collaboration with Y. Nesterov, F. Glineur ( Univ.

More information