Stochas(c Dual Ascent Linear Systems, Quasi-Newton Updates and Matrix Inversion
|
|
- Harry Cody Washington
- 5 years ago
- Views:
Transcription
1 Stochas(c Dual Ascent Linear Systems, Quasi-Newton Updates and Matrix Inversion Peter Richtárik (joint work with Robert M. Gower) University of Edinburgh Oberwolfach, March 8, 2016
2 Part I Stochas(c Dual Ascent for Linear Systems
3 Robert Mansel Gower (Edinburgh) Robert Mansel Gower and P.R. Randomized Itera.ve Methods for Linear Systems SIAM Journal on Matrix Analysis and Applica5ons 36(4): , 2015 [GR 15a] Robert Mansel Gower and P.R. Stochas.c Dual Ascent for Solving Linear Systems arxiv: , 2015 [GR 15b]
4 The Problem
5 The Problem: Solve a Linear System n 2 R n m Ax = b m Assump.on 1 The system is consistent (i.e., has a solu(on)
6 Op(miza(on Formula(on Primal Problem B 0 minimize P (x) := 1 2 kx ck2 B subject to Ax = b A 2 R m n x 2 R n 1 2 (x c)> B(x c) Dual Problem Unconstrained non-strongly concave quadra(c maximiza(on problem maximize D(y) :=(b Ac) > 1 y 2 ka> yk 2 B 1 subject to y 2 R m
7 Dual Correspondence Lemma Lemma (GR 15b) (Any) dual op(mal point A ne mapping from R m to R n x(y) :=c + B 1 A > y Primal op(mal point D(y ) D(y) = 1 2 kx(y) x k 2 B Dual error (in func(on values) Primal error (in distance)
8 New Algorithm: Stochas(c Dual Ascent (SDA)
9 Stochas(c Dual Ascent A random m matrix drawn i.i.d. in each itera(on S D y t+1 = y t + S t m m m Moore-Penrose pseudo-inverse of a small matrix t := arg min 2Q t k k 2 Q t := arg max D(y t + S ) t = S > AB 1 A > S S > b A c + B 1 A > y t
10 Primal Method = Linear Image of the Dual Method x t := x(y t )=c + B 1 A > y t Corresponding primal iterates Dual iterates produced by SDA
11 Main Assump(on Assump.on 2 The matrix E S D h S S > AB 1 A > S S >i is nonsingular H
12 Complexity := 1 + min B 1/2 A > E[H]AB 1/2 of SDA U 0 = 1 2 kx0 x k 2 B Theorem (GR 15b) Primal iterates: E 1 2 kxt x k 2 B apple t U 0 GR 15a Residual: E[kAx t bk B ] apple t/2 kak B p 2 U0 Dual error: E[OPT D(y t )] apple t U 0 Primal error: E[P (x t ) OPT] apple t U 0 +2 t/2p OPT U 0 Duality gap: E[P (x t ) D(y t )] apple 2 t U 0 +2 t/2p OPT U 0
13 The Rate: Lower and Upper Bounds Rank(S > A)=dim(Range(B 1 A > S)) = Tr(B 1 Z) Theorem [RG 15ab] 0 apple 1 Rank(S > A) Rank(A) apple < 1 Insight: apple 1 always < 1 if Assumption 2 holds Insight: The lower bound is good when: i) the dimension of the search space in the constrain and approximate viewpoint is large, ii) the rank of A is small
14 The Primal Iterates: 6 Equivalent Viewpoints x t := x(y t )=c + B 1 A > y t Corresponding primal iterates Dual iterates produced by SDA
15 1. Relaxa(on Viewpoint Sketch and Project x t+1 = arg min x2r n kx xt k 2 B subject to S > Ax = S > b S = iden(ty matrix convergence in 1 step
16 2. Approxima(on Viewpoint Constrain and Approximate x t+1 = arg min x2r n kx x k 2 B subject to x = x t + B 1 A > S is free
17 3. Geometric Viewpoint Random Intersect x t (1) x + Null(S > A) x (2) x t+1 x t + Range(B 1 A > S) (1) x t+1 = arg min kx x x t k B subject to S > Ax = S > b (2) x t+1 = arg min kx x x k B subject to x = x t + B 1 A > S {x t+1 } = x + Null(S > A) \ x t + Range(B 1 A > S)
18 4. Algebraic Viewpoint Random Linear Solve x t+1 = solution in x of the linear system S > Ax = S > b x = x t + B 1 A > S Unknown Unknown
19 5. Algebraic Viewpoint Random Update Random Update Vector x t+1 = x t B 1 A > S(S > AB 1 A > S) S > (Ax t b) Moore-Penrose pseudo-inverse
20 6. Analy(c Viewpoint Random Fixed Point Z := A > S(S > AB 1 A > S) S > A x t+1 x = (I B 1 Z)(x t x ) Random Itera(on Matrix (B 1 Z) 2 = B 1 Z (I B 1 Z) 2 = I B 1 Z B 1 Z projects orthogonally onto Range(B 1 A > S) I B 1 Z projects orthogonally onto Null(S > A)
21 Special Case: Randomized Kaczmarz Method
22 Randomized Kaczmarz (RK) Method M. S. Kaczmarz. Angenaherte Auflosung von Systemen linearer Gleichungen, Bulle5n Interna5onal de l Académie Polonaise des Sciences et des LeKres. Classe des Sciences Mathéma5ques et Naturelles. Série A, Sciences Mathéma5ques 35, pp , 1937 Kaczmarz method (1937) T. Strohmer and R. Vershynin. A Randomized Kaczmarz Algorithm with Exponen.al Convergence. Journal of Fourier Analysis and Applica5ons 15(2), pp , 2009 Randomized Kaczmarz method (2009) RK arises as a special case for parameters B, S set as follows: B = I S = e i =(0,...,0, 1, 0,...,0) with probability p i x t+1 = x t A i: x t b i ka i: k 2 2 (A i: ) T RK was analyzed for p i = ka i:k 2 kak 2 F
23 RK: Deriva(on and Rate General Method x t+1 = x t B 1 A T S (S T AB 1 A T S) S T (Ax t b) Special Choice of Parameters P(S = e i )=p i Complexity Rate B = I x t+1 = x t A i:xt b i S = e i ka i: k 2 (A i: ) T 2 p i = ka i:k 2 kak 2 F E kx t x k 2 min A T A 2 apple 1 kak 2 F! t kx 0 x k 2 2
24 RK: Further Reading D. Needell. Randomized Kaczmarz solver for noisy linear systems. BIT 50 (2), pp , 2010 D. Needell and J. Tropp. Paved with good inten.ons: analyzis of a randomized block Kaczmarz method. Linear Algebra and its Applica5ons 441, pp , 2012 D. Needell, N. Srebro and R. Ward. Stochas.c gradient descent, weighted sampling and the randomized Kaczmarz algorithm. Mathema5cal Programming, 2015 (arxiv: ) A. Ramdas. Rows vs Columns for Linear Systems of Equa.ons Randomized Kaczmarz or Coordinate Descent? arxiv: , 2014
25 Special Case: Gaussian Descent
26 Gaussian Descent General Method x t+1 = x t B 1 A T S (S T AB 1 A T S) S T (Ax t b) Special Choice of Parameters S N(0, ) x t+1 = x t S T (Ax t b) S T AB 1 A T S B 1 A T S Posi(ve definite covariance matrix Complexity Rate E kx t x k 2 B apple t kx 0 x k 2 B
27 x t+1 = x t h t B 1/2 := B 1/2 A T S N(0, ) x := B 1/2 A T AB 1/2 x 0
28 Gaussian Descent: The Rate Lemma [GR 15] E apple T k k Tr( ) 1 1 n apple apple 1 2 min( ) Tr( ) This follows from the general lower
29 Gaussian Descent: Further Reading Yurii Nesterov. Random gradient-free minimiza.on of convex func.ons. CORE Discussion Paper # 2011/1, 2011 S. U. S(tch, C. L. Muller and G. Gartner. Op.miza.on of convex func.ons with random pursuit. SIAM Journal on Op(miza(on 23 (2), pp , 2014 S. U. S(tch. Convex op.miza.on with random pursuit. PhD Thesis, ETH Zurich, 2014
30 Special Case: Randomized Coordinate Descent
31 Randomized Coordinate Descent (RCD) A. S. Lewis and D. Leventhal. Randomized methods for linear constraints: convergence rates and condi.oning. Mathema5cs of OR 35(3), , 2010 (arxiv: ) min f(x) = 1 x2r n 2 xt Ax b T x Assume: Posi(ve definite S = e i =(0,...,0, 1, 0,...,0) with probability p i RCD (2008) RCD arises as a special case for parameters B, S set as follows: B = A x = A 1 b Recall: In RK we had B = I x t+1 = x t (A i: ) T x t b i A ii e i RCD was analyzed for p i = A ii Tr(A)
32 RCD: Deriva(on and Rate General Method x t+1 = x t B 1 A T S (S T AB 1 A T S) S T (Ax t b) Special Choice of Parameters B = A P(S = e i )=p i Complexity Rate S = e i x t+1 = x t (A i: ) T x t b i A ii e i p i = A ii Tr(A) E kx t x k 2 A apple 1 t min(a) kx 0 x k 2 A Tr(A)
33 RCD: Standard Op(miza(on Form Yurii Nesterov. Efficiency of coordinate descent methods on huge-scale op.miza.on problems. SIAM J. on Op5miza5on, 22(2): , 2012 (CORE Discussion Paper 2010/2) Nesterov considered the problem: min f(x) x2r n Convex and smooth Nesterov assumed that the following inequality holds for all x, h and i: Given a current iterate x, choosing h by minimizing the RHS gives: Nesterov s RCD method: x t+1 = x t 1 L i r i f(x t )e i f(x + he i ) apple f(x)+r i f(x)h + L i 2 h2 f(x) = 1 2 xt Ax b T x ) L i = A ii r i f(x) =(A i: ) T x We recover RCD as we have seen it: x t+1 = x t (A i: ) T x t b i A ii e i b i
34 Special Case: Randomized Newton Method
35 Randomized Newton (RN) Z. Qu, PR, M. Takáč and O. Fercoq. Stochas.c Dual Newton Ascent for Empirical Risk Minimiza.on. arxiv: , 2015 min f(x) = 1 x2r n 2 xt Ax b T x Assume: Posi(ve definite RN arises as a special case for parameters B, S set as follows: B = A x = A 1 b S = I :C with probability p C p C 0 8C {1,...,n} X C {1,...,n} p C =1 RCD is special case with p C =0whenever C 6= 1
36 RN: Deriva(on General Method x t+1 = x t B 1 A T S (S T AB 1 A T S) S T (Ax t b) Special Choice of Parameters B = A S = I :C with probability p C x t+1 = x t I :C ((I :C ) T AI :C ) 1 (I :C ) T (Ax t b) This method minimizes f exactly in a random subspace spanned by the coordinates belonging to C
37 e 7 C = {2, 7} C =2 x t+1 x t e 2
38 Summary: Linear Systems SDA: A new class of randomized op(miza(on algorithms Extremely versa(le Works for almost any random S Get several exis(ng algorithms in special cases (RK, RCD, RN, RBK) Get many new algorithms in special cases Linear convergence despite lack of strong concavity RK in the primal = RCD in the dual Did not talk about: Randomized gossip Distributed variant Op(mal sampling via SDP Experiments
39
40 Part II Stochas(c Dual Ascent for Matrix Inversion
41 Robert Mansel Gower (Edinburgh) Robert Mansel Gower and P.R. Randomized Quasi-Newton Methods are Linearly Convergent Matrix Inversion Algorithms arxiv: , 2016
42 The Problem: Invert a Matrix n 2 R n n Iden(ty matrix n AX = I Assump.on 1 Matrix A is inver(ble
43 Inver(ng Symmetric Matrices
44 1. Sketch and Project kxk F (B) := p Tr(X > BXB) X t+1 = arg min kx Xt X2Rn n k 2 F (B) subject to S > AX = S >, X = X > Quasi-Newton updates are of this form: S = determinis(c column vector We get randomized block version of quasi-newton updates! Randomized quasi-newton updates are linearly convergent matrix inversion methods Interpreta(on: Gaussian Inference (Henning, 2015) Donald Goldfarb. A Family of Variable-Metric Methods Derived by Varia.onal Means. Mathema5cs of Computa5on 24(109), 1970
45 Gaussian Inference Philipp Henning Probabilis.c Interpreta.on of Linear Solvers SIAM Journal on Op5miza5on 25(1): , 2015 X k+1 The new iterate can be interpreted as the mean of a posterior distribu(on under a Gaussian prior with mean and noiseless (and random) linear observa(on of X k A 1
46 Randomized QN Updates B Equation Method I AX = I Powel-Symmetric-Broyden (PSB) A 1 XA 1 = I Davidon-Fletcher-Powell (DFP) A AX = I Broyden-Fletcher-Goldfarb-Shanno (BFGS) All these QN methods arise as special cases of our framework All are linearly convergent, with explicit convergence rates We also recover non-symmetric updates such as Bad Broyden and Good Broyden We get block versions We get randomized versions of new QN updates
47 2. Constrain and Approximate X t+1 = arg min kx A X2Rn n 1 k 2 F (B) s.t. X = X t + S > AB 1 + B 1 A > S > 2 R n is free Randomized BFGS: B = A, =1 X t+1 = arg New formula.on even for standard QN methods min X2R n n kx A 1 k 2 F (A) = kax Ik2 F s.t. X = X t + S > + S > 2 R n is free RBFGS performs best symmetric rank-2 update
48 4. Random Update H = S(S > AB 1 A > S) S > X t+1 = X t (X t A I)HAB 1 + B 1 AH(AX t I)(AHAB 1 I) 6. Random Fixed Point X t+1 A 1 = (I B 1 A > HA)(X t A 1 )(I AHA > B 1 )
49 Complexity / Convergence Theorem [GR 16] kmk B := kb 1/2 MB 1/2 k 2 1 E X t A 1 B apple t kx 0 A 1 k B 2 E[H] 0 < 1 E h kx t A 1 k 2 F (B) i apple t kx 0 A 1 k 2 F (B)
50 Summary: Matrix Inversion Block version of QN updates New points of view (constrain and approximate,...) New link between QN and approx. inverse precondi(oning First (me randomized QN updates are proposed First stochas(c method for matrix inversion (with complexity bounds)? Linear convergence under weak assump(ons Did not talk about: Nonsymmetric variants Theore(cal bounds for discretely distributed S Adap(ve randomized BFGS Limited memory and factored implementa(ons Experiments (Newton-Schultz; MinRes) Use in empirical risk minimiza(on (Gower, Goldfarb & R. 16) Extension to the computa(on of a pseudoinverse
51 The End
arxiv: v2 [math.na] 28 Jan 2016
Stochastic Dual Ascent for Solving Linear Systems Robert M. Gower and Peter Richtárik arxiv:1512.06890v2 [math.na 28 Jan 2016 School of Mathematics University of Edinburgh United Kingdom December 21, 2015
More informationarxiv: v3 [math.na] 23 Mar 2016
Randomized Quasi-Newton Updates are Linearly Convergent Matrix Inversion Algorithms Robert M. Gower and Peter Richtárik arxiv:602.0768v3 [math.na] 23 Mar 206 School of Mathematics University of Edinburgh
More informationBellman s Curse of Dimensionality
Bellman s Curse of Dimensionality n- dimensional state space Number of states grows exponen
More informationRandomized Quasi-Newton Updates are Linearly Convergent Matrix Inversion Algorithms
Randomized Quasi-Newton Updates are Linearly Convergent Matrix Inversion Algorithms Robert M. Gower and Peter Richtárik School of Mathematics University of Edinburgh United Kingdom ebruary 4, 06 Abstract
More informationSketch and Project: Randomized Iterative Methods for Linear Systems and Inverting Matrices
Sketch and Project: Randomized Iterative Methods for Linear Systems and Inverting Matrices arxiv:1612.06013v1 [math.na] 19 Dec 2016 Robert Mansel Gower Supervisor: Dr. Peter Richtárik Thesis submitted
More informationA Sampling Kaczmarz-Motzkin Algorithm for Linear Feasibility
A Sampling Kaczmarz-Motzkin Algorithm for Linear Feasibility Jamie Haddock Graduate Group in Applied Mathematics, Department of Mathematics, University of California, Davis Copper Mountain Conference on
More informationLecture 3: Minimizing Large Sums. Peter Richtárik
Lecture 3: Minimizing Large Sums Peter Richtárik Graduate School in Systems, Op@miza@on, Control and Networks Belgium 2015 Mo@va@on: Machine Learning & Empirical Risk Minimiza@on Training Linear Predictors
More informationOn the exponential convergence of. the Kaczmarz algorithm
On the exponential convergence of the Kaczmarz algorithm Liang Dai and Thomas B. Schön Department of Information Technology, Uppsala University, arxiv:4.407v [cs.sy] 0 Mar 05 75 05 Uppsala, Sweden. E-mail:
More informationConvergence Rates for Greedy Kaczmarz Algorithms
onvergence Rates for Greedy Kaczmarz Algorithms Julie Nutini 1, Behrooz Sepehry 1, Alim Virani 1, Issam Laradji 1, Mark Schmidt 1, Hoyt Koepke 2 1 niversity of British olumbia, 2 Dato Abstract We discuss
More information5 Quasi-Newton Methods
Unconstrained Convex Optimization 26 5 Quasi-Newton Methods If the Hessian is unavailable... Notation: H = Hessian matrix. B is the approximation of H. C is the approximation of H 1. Problem: Solve min
More informationEdinburgh Research Explorer
Edinburgh Research Explorer Randomized Iterative Methods for Linear Systems Citation for published version: Gower, R & Richtarik, P 2015, 'Randomized Iterative Methods for Linear Systems' SIAM Journal
More informationStochastic Reformulations of Linear Systems: Algorithms and Convergence Theory
Stochastic Reformulations of Linear Systems: Algorithms and Convergence Theory Peter Richtárik Martin Takáč June 4, 2017 Abstract We develop a family of reformulations of an arbitrary consistent linear
More informationQuasi-Newton methods for minimization
Quasi-Newton methods for minimization Lectures for PHD course on Numerical optimization Enrico Bertolazzi DIMS Universitá di Trento November 21 December 14, 2011 Quasi-Newton methods for minimization 1
More informationLeast Square Es?ma?on, Filtering, and Predic?on: ECE 5/639 Sta?s?cal Signal Processing II: Linear Es?ma?on
Least Square Es?ma?on, Filtering, and Predic?on: Sta?s?cal Signal Processing II: Linear Es?ma?on Eric Wan, Ph.D. Fall 2015 1 Mo?va?ons If the second-order sta?s?cs are known, the op?mum es?mator is given
More informationConvex Optimization CMU-10725
Convex Optimization CMU-10725 Quasi Newton Methods Barnabás Póczos & Ryan Tibshirani Quasi Newton Methods 2 Outline Modified Newton Method Rank one correction of the inverse Rank two correction of the
More informationRandomized Iterative Methods for Linear Systems
Randomized Iterative Methods for Linear Systems Robert M. Gower Peter Richtárik June 16, 015 Abstract We develop a novel, fundamental and surprisingly simple randomized iterative method for solving consistent
More informationSGD and Randomized projection algorithms for overdetermined linear systems
SGD and Randomized projection algorithms for overdetermined linear systems Deanna Needell Claremont McKenna College IPAM, Feb. 25, 2014 Includes joint work with Eldar, Ward, Tropp, Srebro-Ward Setup Setup
More informationNewton s Method. Ryan Tibshirani Convex Optimization /36-725
Newton s Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, Properties and examples: f (y) = max x
More informationLecture 14: October 17
1-725/36-725: Convex Optimization Fall 218 Lecture 14: October 17 Lecturer: Lecturer: Ryan Tibshirani Scribes: Pengsheng Guo, Xian Zhou Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:
More informationRandomized projection algorithms for overdetermined linear systems
Randomized projection algorithms for overdetermined linear systems Deanna Needell Claremont McKenna College ISMP, Berlin 2012 Setup Setup Let Ax = b be an overdetermined, standardized, full rank system
More informationAcceleration of Randomized Kaczmarz Method
Acceleration of Randomized Kaczmarz Method Deanna Needell [Joint work with Y. Eldar] Stanford University BIRS Banff, March 2011 Problem Background Setup Setup Let Ax = b be an overdetermined consistent
More informationOptimization II: Unconstrained Multivariable
Optimization II: Unconstrained Multivariable CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Justin Solomon CS 205A: Mathematical Methods Optimization II: Unconstrained Multivariable 1
More informationNonlinear Optimization: What s important?
Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global
More informationStochastic Gradient Descent, Weighted Sampling, and the Randomized Kaczmarz algorithm
Stochastic Gradient Descent, Weighted Sampling, and the Randomized Kaczmarz algorithm Deanna Needell Department of Mathematical Sciences Claremont McKenna College Claremont CA 97 dneedell@cmc.edu Nathan
More informationarxiv: v2 [math.na] 11 May 2017
Rows vs. Columns: Randomized Kaczmarz or Gauss-Seidel for Ridge Regression arxiv:1507.05844v2 [math.na] 11 May 2017 Ahmed Hefny Machine Learning Department Carnegie Mellon University Deanna Needell Department
More informationNonlinear Programming
Nonlinear Programming Kees Roos e-mail: C.Roos@ewi.tudelft.nl URL: http://www.isa.ewi.tudelft.nl/ roos LNMB Course De Uithof, Utrecht February 6 - May 8, A.D. 2006 Optimization Group 1 Outline for week
More informationMATH 4211/6211 Optimization Quasi-Newton Method
MATH 4211/6211 Optimization Quasi-Newton Method Xiaojing Ye Department of Mathematics & Statistics Georgia State University Xiaojing Ye, Math & Stat, Georgia State University 0 Quasi-Newton Method Motivation:
More informationOptimization II: Unconstrained Multivariable
Optimization II: Unconstrained Multivariable CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Doug James (and Justin Solomon) CS 205A: Mathematical Methods Optimization II: Unconstrained
More informationQuasi-Newton Methods. Javier Peña Convex Optimization /36-725
Quasi-Newton Methods Javier Peña Convex Optimization 10-725/36-725 Last time: primal-dual interior-point methods Consider the problem min x subject to f(x) Ax = b h(x) 0 Assume f, h 1,..., h m are convex
More informationChapter 4. Unconstrained optimization
Chapter 4. Unconstrained optimization Version: 28-10-2012 Material: (for details see) Chapter 11 in [FKS] (pp.251-276) A reference e.g. L.11.2 refers to the corresponding Lemma in the book [FKS] PDF-file
More informationPseudospectral Methods For Op2mal Control. Jus2n Ruths March 27, 2009
Pseudospectral Methods For Op2mal Control Jus2n Ruths March 27, 2009 Introduc2on Pseudospectral methods arose to find solu2ons to Par2al Differen2al Equa2ons Recently adapted for Op2mal Control Key Ideas
More informationProgramming, numerics and optimization
Programming, numerics and optimization Lecture C-3: Unconstrained optimization II Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428
More informationDeterminant maximization with linear. S. Boyd, L. Vandenberghe, S.-P. Wu. Information Systems Laboratory. Stanford University
Determinant maximization with linear matrix inequality constraints S. Boyd, L. Vandenberghe, S.-P. Wu Information Systems Laboratory Stanford University SCCM Seminar 5 February 1996 1 MAXDET problem denition
More informationQuasi-Newton Methods
Newton s Method Pros and Cons Quasi-Newton Methods MA 348 Kurt Bryan Newton s method has some very nice properties: It s extremely fast, at least once it gets near the minimum, and with the simple modifications
More informationMethods that avoid calculating the Hessian. Nonlinear Optimization; Steepest Descent, Quasi-Newton. Steepest Descent
Nonlinear Optimization Steepest Descent and Niclas Börlin Department of Computing Science Umeå University niclas.borlin@cs.umu.se A disadvantage with the Newton method is that the Hessian has to be derived
More information2. Quasi-Newton methods
L. Vandenberghe EE236C (Spring 2016) 2. Quasi-Newton methods variable metric methods quasi-newton methods BFGS update limited-memory quasi-newton methods 2-1 Newton method for unconstrained minimization
More informationRandomized Coordinate Descent Methods on Optimization Problems with Linearly Coupled Constraints
Randomized Coordinate Descent Methods on Optimization Problems with Linearly Coupled Constraints By I. Necoara, Y. Nesterov, and F. Glineur Lijun Xu Optimization Group Meeting November 27, 2012 Outline
More informationOp#mal convex op#miza#on under Tsybakov noise through connec#ons to ac#ve learning
Op#mal convex op#miza#on under Tsybakov noise through connec#ons to ac#ve learning Aar$ Singh Joint work with: Aaditya Ramdas Connec#ons between convex op#miza#on and ac#ve learning (a formal reduc#on)
More informationOptimization and Root Finding. Kurt Hornik
Optimization and Root Finding Kurt Hornik Basics Root finding and unconstrained smooth optimization are closely related: Solving ƒ () = 0 can be accomplished via minimizing ƒ () 2 Slide 2 Basics Root finding
More informationDeriva'on of The Kalman Filter. Fred DePiero CalPoly State University EE 525 Stochas'c Processes
Deriva'on of The Kalman Filter Fred DePiero CalPoly State University EE 525 Stochas'c Processes KF Uses State Predic'ons KF es'mates the state of a system Example Measure: posi'on State: [ posi'on velocity
More informationQuasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb Shanno February 6, / 25 (BFG. Limited memory BFGS (L-BFGS)
Quasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb Shanno (BFGS) Limited memory BFGS (L-BFGS) February 6, 2014 Quasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb
More informationLecture 18: November Review on Primal-dual interior-poit methods
10-725/36-725: Convex Optimization Fall 2016 Lecturer: Lecturer: Javier Pena Lecture 18: November 2 Scribes: Scribes: Yizhu Lin, Pan Liu Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:
More informationConvergence Properties of the Randomized Extended Gauss-Seidel and Kaczmarz Methods
Claremont Colleges Scholarship @ Claremont CMC Faculty Publications and Research CMC Faculty Scholarship 11-1-015 Convergence Properties of the Randomized Extended Gauss-Seidel and Kaczmarz Methods Anna
More informationOptimisation in Higher Dimensions
CHAPTER 6 Optimisation in Higher Dimensions Beyond optimisation in 1D, we will study two directions. First, the equivalent in nth dimension, x R n such that f(x ) f(x) for all x R n. Second, constrained
More informationShiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 3. Gradient Method
Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 3 Gradient Method Shiqian Ma, MAT-258A: Numerical Optimization 2 3.1. Gradient method Classical gradient method: to minimize a differentiable convex
More informationRandomized Coordinate Descent with Arbitrary Sampling: Algorithms and Complexity
Randomized Coordinate Descent with Arbitrary Sampling: Algorithms and Complexity Zheng Qu University of Hong Kong CAM, 23-26 Aug 2016 Hong Kong based on joint work with Peter Richtarik and Dominique Cisba(University
More informationQuasi-Newton Methods
Quasi-Newton Methods Werner C. Rheinboldt These are excerpts of material relating to the boos [OR00 and [Rhe98 and of write-ups prepared for courses held at the University of Pittsburgh. Some further references
More informationMarch 8, 2010 MATH 408 FINAL EXAM SAMPLE
March 8, 200 MATH 408 FINAL EXAM SAMPLE EXAM OUTLINE The final exam for this course takes place in the regular course classroom (MEB 238) on Monday, March 2, 8:30-0:20 am. You may bring two-sided 8 page
More informationLearning Theory of Randomized Kaczmarz Algorithm
Journal of Machine Learning Research 16 015 3341-3365 Submitted 6/14; Revised 4/15; Published 1/15 Junhong Lin Ding-Xuan Zhou Department of Mathematics City University of Hong Kong 83 Tat Chee Avenue Kowloon,
More informationWHY DUALITY? Gradient descent Newton s method Quasi-newton Conjugate gradients. No constraints. Non-differentiable ???? Constrained problems? ????
DUALITY WHY DUALITY? No constraints f(x) Non-differentiable f(x) Gradient descent Newton s method Quasi-newton Conjugate gradients etc???? Constrained problems? f(x) subject to g(x) apple 0???? h(x) =0
More informationFAST DISTRIBUTED COORDINATE DESCENT FOR NON-STRONGLY CONVEX LOSSES. Olivier Fercoq Zheng Qu Peter Richtárik Martin Takáč
FAST DISTRIBUTED COORDINATE DESCENT FOR NON-STRONGLY CONVEX LOSSES Olivier Fercoq Zheng Qu Peter Richtárik Martin Takáč School of Mathematics, University of Edinburgh, Edinburgh, EH9 3JZ, United Kingdom
More informationSparse Covariance Selection using Semidefinite Programming
Sparse Covariance Selection using Semidefinite Programming A. d Aspremont ORFE, Princeton University Joint work with O. Banerjee, L. El Ghaoui & G. Natsoulis, U.C. Berkeley & Iconix Pharmaceuticals Support
More informationQuasi-Newton Methods. Zico Kolter (notes by Ryan Tibshirani, Javier Peña, Zico Kolter) Convex Optimization
Quasi-Newton Methods Zico Kolter (notes by Ryan Tibshirani, Javier Peña, Zico Kolter) Convex Optimization 10-725 Last time: primal-dual interior-point methods Given the problem min x f(x) subject to h(x)
More informationStochastic Optimization Algorithms Beyond SG
Stochastic Optimization Algorithms Beyond SG Frank E. Curtis 1, Lehigh University involving joint work with Léon Bottou, Facebook AI Research Jorge Nocedal, Northwestern University Optimization Methods
More informationAlgorithms for Constrained Optimization
1 / 42 Algorithms for Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University April 19, 2015 2 / 42 Outline 1. Convergence 2. Sequential quadratic
More informationMethods for Unconstrained Optimization Numerical Optimization Lectures 1-2
Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2 Coralia Cartis, University of Oxford INFOMM CDT: Modelling, Analysis and Computation of Continuous Real-World Problems Methods
More informationELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications
ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications Professor M. Chiang Electrical Engineering Department, Princeton University March
More informationNumerical Methods in Physics
Numerical Methods in Physics Numerische Methoden in der Physik, 515.421. Instructor: Ass. Prof. Dr. Lilia Boeri Room: PH 03 090 Tel: +43-316- 873 8191 Email Address: l.boeri@tugraz.at Room: TDK Seminarraum
More informationRandomized Kaczmarz Nick Freris EPFL
Randomized Kaczmarz Nick Freris EPFL (Joint work with A. Zouzias) Outline Randomized Kaczmarz algorithm for linear systems Consistent (noiseless) Inconsistent (noisy) Optimal de-noising Convergence analysis
More informationTsybakov noise adap/ve margin- based ac/ve learning
Tsybakov noise adap/ve margin- based ac/ve learning Aar$ Singh A. Nico Habermann Associate Professor NIPS workshop on Learning Faster from Easy Data II Dec 11, 2015 Passive Learning Ac/ve Learning (X j,?)
More informationThe speed of Shor s R-algorithm
IMA Journal of Numerical Analysis 2008) 28, 711 720 doi:10.1093/imanum/drn008 Advance Access publication on September 12, 2008 The speed of Shor s R-algorithm J. V. BURKE Department of Mathematics, University
More informationOptimization: Nonlinear Optimization without Constraints. Nonlinear Optimization without Constraints 1 / 23
Optimization: Nonlinear Optimization without Constraints Nonlinear Optimization without Constraints 1 / 23 Nonlinear optimization without constraints Unconstrained minimization min x f(x) where f(x) is
More informationMathematical Optimisation, Chpt 2: Linear Equations and inequalities
Mathematical Optimisation, Chpt 2: Linear Equations and inequalities Peter J.C. Dickinson p.j.c.dickinson@utwente.nl http://dickinson.website version: 12/02/18 Monday 5th February 2018 Peter J.C. Dickinson
More informationMachine Learning Lecture Notes
Machine Learning Lecture Notes Predrag Radivojac February 2, 205 Given a data set D = {(x i,y i )} n the objective is to learn the relationship between features and the target. We usually start by hypothesizing
More informationCS6750: Cryptography and Communica7on Security
CS6750: Cryptography and Communica7on Security Class 6: Simple Number Theory Dr. Erik- Oliver Blass Plan 1. Role of number theory in cryptography 2. Classical problems in computa7onal number theory 3.
More informationStochastic Quasi-Newton Methods
Stochastic Quasi-Newton Methods Donald Goldfarb Department of IEOR Columbia University UCLA Distinguished Lecture Series May 17-19, 2016 1 / 35 Outline Stochastic Approximation Stochastic Gradient Descent
More informationConvex Optimization. Problem set 2. Due Monday April 26th
Convex Optimization Problem set 2 Due Monday April 26th 1 Gradient Decent without Line-search In this problem we will consider gradient descent with predetermined step sizes. That is, instead of determining
More informationUnconstrained optimization
Chapter 4 Unconstrained optimization An unconstrained optimization problem takes the form min x Rnf(x) (4.1) for a target functional (also called objective function) f : R n R. In this chapter and throughout
More informationOn Nesterov s Random Coordinate Descent Algorithms - Continued
On Nesterov s Random Coordinate Descent Algorithms - Continued Zheng Xu University of Texas At Arlington February 20, 2015 1 Revisit Random Coordinate Descent The Random Coordinate Descent Upper and Lower
More informationInput: System of inequalities or equalities over the reals R. Output: Value for variables that minimizes cost function
Linear programming Input: System of inequalities or equalities over the reals R A linear cost function Output: Value for variables that minimizes cost function Example: Minimize 6x+4y Subject to 3x + 2y
More informationWritten Examination
Division of Scientific Computing Department of Information Technology Uppsala University Optimization Written Examination 202-2-20 Time: 4:00-9:00 Allowed Tools: Pocket Calculator, one A4 paper with notes
More informationLinear and non-linear programming
Linear and non-linear programming Benjamin Recht March 11, 2005 The Gameplan Constrained Optimization Convexity Duality Applications/Taxonomy 1 Constrained Optimization minimize f(x) subject to g j (x)
More informationFundamentals of Data Assimila1on
014 GSI Community Tutorial NCAR Foothills Campus, Boulder, CO July 14-16, 014 Fundamentals of Data Assimila1on Milija Zupanski Cooperative Institute for Research in the Atmosphere Colorado State University
More informationLECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE
LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE CONVEX ANALYSIS AND DUALITY Basic concepts of convex analysis Basic concepts of convex optimization Geometric duality framework - MC/MC Constrained optimization
More informationCOMP 562: Introduction to Machine Learning
COMP 562: Introduction to Machine Learning Lecture 20 : Support Vector Machines, Kernels Mahmoud Mostapha 1 Department of Computer Science University of North Carolina at Chapel Hill mahmoudm@cs.unc.edu
More informationOn Nesterov s Random Coordinate Descent Algorithms
On Nesterov s Random Coordinate Descent Algorithms Zheng Xu University of Texas At Arlington February 19, 2015 1 Introduction Full-Gradient Descent Coordinate Descent 2 Random Coordinate Descent Algorithm
More informationComparative study of Optimization methods for Unconstrained Multivariable Nonlinear Programming Problems
International Journal of Scientific and Research Publications, Volume 3, Issue 10, October 013 1 ISSN 50-3153 Comparative study of Optimization methods for Unconstrained Multivariable Nonlinear Programming
More informationReduced Models for Process Simula2on and Op2miza2on
Reduced Models for Process Simulaon and Opmizaon Yidong Lang, Lorenz T. Biegler and David Miller ESI annual meeng March, 0 Models are mapping Equaon set or Module simulators Input space Reduced model Surrogate
More informationThe Dual Lattice, Integer Linear Systems and Hermite Normal Form
New York University, Fall 2013 Lattices, Convexity & Algorithms Lecture 2 The Dual Lattice, Integer Linear Systems and Hermite Normal Form Lecturers: D. Dadush, O. Regev Scribe: D. Dadush 1 Dual Lattice
More informationMini-Batch Primal and Dual Methods for SVMs
Mini-Batch Primal and Dual Methods for SVMs Peter Richtárik School of Mathematics The University of Edinburgh Coauthors: M. Takáč (Edinburgh), A. Bijral and N. Srebro (both TTI at Chicago) arxiv:1303.2314
More informationChapter 2. Optimization. Gradients, convexity, and ALS
Chapter 2 Optimization Gradients, convexity, and ALS Contents Background Gradient descent Stochastic gradient descent Newton s method Alternating least squares KKT conditions 2 Motivation We can solve
More informationAgenda. 1 Duality for LP. 2 Theorem of alternatives. 3 Conic Duality. 4 Dual cones. 5 Geometric view of cone programs. 6 Conic duality theorem
Agenda 1 Duality for LP 2 Theorem of alternatives 3 Conic Duality 4 Dual cones 5 Geometric view of cone programs 6 Conic duality theorem 7 Examples Lower bounds on LPs By eliminating variables (if needed)
More informationMethodological Foundations of Biomedical Informatics (BMSC-GA 4449) Optimization
Methodological Foundations of Biomedical Informatics (BMSCGA 4449) Optimization Op#miza#on A set of techniques for finding the values of variables at which the objec#ve func#on amains its cri#cal (minimal
More information10-725/36-725: Convex Optimization Prerequisite Topics
10-725/36-725: Convex Optimization Prerequisite Topics February 3, 2015 This is meant to be a brief, informal refresher of some topics that will form building blocks in this course. The content of the
More informationOutline Introduction: Problem Description Diculties Algebraic Structure: Algebraic Varieties Rank Decient Toeplitz Matrices Constructing Lower Rank St
Structured Lower Rank Approximation by Moody T. Chu (NCSU) joint with Robert E. Funderlic (NCSU) and Robert J. Plemmons (Wake Forest) March 5, 1998 Outline Introduction: Problem Description Diculties Algebraic
More informationLecture 7 Unconstrained nonlinear programming
Lecture 7 Unconstrained nonlinear programming Weinan E 1,2 and Tiejun Li 2 1 Department of Mathematics, Princeton University, weinan@princeton.edu 2 School of Mathematical Sciences, Peking University,
More informationNONSMOOTH VARIANTS OF POWELL S BFGS CONVERGENCE THEOREM
NONSMOOTH VARIANTS OF POWELL S BFGS CONVERGENCE THEOREM JIAYI GUO AND A.S. LEWIS Abstract. The popular BFGS quasi-newton minimization algorithm under reasonable conditions converges globally on smooth
More informationFundamentals of Data Assimila1on
2015 GSI Community Tutorial NCAR Foothills Campus, Boulder, CO August 11-14, 2015 Fundamentals of Data Assimila1on Milija Zupanski Cooperative Institute for Research in the Atmosphere Colorado State University
More informationORIE 6326: Convex Optimization. Quasi-Newton Methods
ORIE 6326: Convex Optimization Quasi-Newton Methods Professor Udell Operations Research and Information Engineering Cornell April 10, 2017 Slides on steepest descent and analysis of Newton s method adapted
More informationNumerical Methods I: Orthogonalization and Newton s method
1/42 Numerical Methods I: Orthogonalization and Newton s method Georg Stadler Courant Institute, NYU stadler@cims.nyu.edu October 5, 2017 Overview 2/42 Linear least squares and orthogonalization methods
More informationLMI Methods in Optimal and Robust Control
LMI Methods in Optimal and Robust Control Matthew M. Peet Arizona State University Lecture 02: Optimization (Convex and Otherwise) What is Optimization? An Optimization Problem has 3 parts. x F f(x) :
More informationParallel Coordinate Optimization
1 / 38 Parallel Coordinate Optimization Julie Nutini MLRG - Spring Term March 6 th, 2018 2 / 38 Contours of a function F : IR 2 IR. Goal: Find the minimizer of F. Coordinate Descent in 2D Contours of a
More informationMore First-Order Optimization Algorithms
More First-Order Optimization Algorithms Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/ yyye Chapters 3, 8, 3 The SDM
More informationDual and primal-dual methods
ELE 538B: Large-Scale Optimization for Data Science Dual and primal-dual methods Yuxin Chen Princeton University, Spring 2018 Outline Dual proximal gradient method Primal-dual proximal gradient method
More informationImproving L-BFGS Initialization for Trust-Region Methods in Deep Learning
Improving L-BFGS Initialization for Trust-Region Methods in Deep Learning Jacob Rafati http://rafati.net jrafatiheravi@ucmerced.edu Ph.D. Candidate, Electrical Engineering and Computer Science University
More informationDS-GA 1002 Lecture notes 10 November 23, Linear models
DS-GA 2 Lecture notes November 23, 2 Linear functions Linear models A linear model encodes the assumption that two quantities are linearly related. Mathematically, this is characterized using linear functions.
More informationReduced-Hessian Methods for Constrained Optimization
Reduced-Hessian Methods for Constrained Optimization Philip E. Gill University of California, San Diego Joint work with: Michael Ferry & Elizabeth Wong 11th US & Mexico Workshop on Optimization and its
More informationHomework 5. Convex Optimization /36-725
Homework 5 Convex Optimization 10-725/36-725 Due Tuesday November 22 at 5:30pm submitted to Christoph Dann in Gates 8013 (Remember to a submit separate writeup for each problem, with your name at the top)
More informationChapter 6 - Orthogonality
Chapter 6 - Orthogonality Maggie Myers Robert A. van de Geijn The University of Texas at Austin Orthogonality Fall 2009 http://z.cs.utexas.edu/wiki/pla.wiki/ 1 Orthogonal Vectors and Subspaces http://z.cs.utexas.edu/wiki/pla.wiki/
More informationRandom coordinate descent algorithms for. huge-scale optimization problems. Ion Necoara
Random coordinate descent algorithms for huge-scale optimization problems Ion Necoara Automatic Control and Systems Engineering Depart. 1 Acknowledgement Collaboration with Y. Nesterov, F. Glineur ( Univ.
More information