Operators. Pontus Giselsson

Size: px
Start display at page:

Download "Operators. Pontus Giselsson"

Transcription

1 Operators Pontus Giselsson 1

2 Today s lecture forward step operators gradient step operators (subclass of forward step operators) resolvents proimal operators (subclass of resolvents) reflected resolvents reflected proimal operators (or proimal reflectors) 2

3 Forward step operator suppose that T : R n R n is single-valued the forward step operator is (Id γt ) 3

4 Cocoercivity suppose that T is 1 β -cocoercive with β = 1 2 then Id γt with γ = 3 is α-averaged, decide α: γβ γβ 1 γβ γt γt 0.5 Id γt T is 2-cocoercive γt is 2 3 -cocoercive (Id γt ) is 3 4 -averaged generally: suppose γ (0, 2 β ) 1 then: β -cocoercivity of T γβ 2 -averagedness of (Id γt ) 4

5 Iterating the forward step operator since 1 β -cocoercivity of T γβ 2 -averagedness of (Id γt ) iterating k+1 = (Id γt ) k converges to fied-point (if eists) (if γ (0, 2 β )) 5

6 Lipschitz continuity suppose that T is 1 2-Lipschitz and γ = 1 motivate that Id γt is not averaged γt 0.5 γt 0.5 Id γt 1.5 cannot make Id γt nonepansive independent of γ iterating forward step of Lipschitz T not guaranteed to converge 6

7 Gradient step operator suppose that f : R n R is differentiable the forward step becomes the gradient step operator of f I γ f if f (proper closed and) conve, then 1 β -cocoercivity of f β-lipschitz continuity of f β-smoothness of f if f is β-smooth, the gradient method converges for γ (0, 2 β ) k+1 = (Id γ f) k (since f 1 β -cocoercive) 7

8 Stronger properties assume f is β-smooth and σ-strongly conve then γf is γβ-smooth and γσ-strongly conve then γf γσ 2 2 is γ(β σ)-smooth or γ f γσid is 1 γ(β σ) -cocoercive (Id γ f) is δ-lipschitz, decide δ 0 γ(β σ) γσ γβ 1 γβ 1 γσ γ f γσid γ f Id γ f (Id γt ) is ma(γβ 1, 1 γσ)-lipschitz contractive if 1 γσ < 1 and γβ 1 < 1, i.e., γ (0, 2 β ) gradient method k+1 = (Id γt ) k then converges linearly optimal γ (center circle) given by γ = 2 β+σ δ = β/σ 1 β/σ+1 8

9 Summary: Gradient step operator σ > 0 : 1 γβ 1 γσ 1 γβ σ = 0 : Id γ f γ (0, 2 β ) contractive optimal γ = 2 β+σ factor (= γβ 1 = 1 γσ) β/σ 1 β/σ+1 Id γ f γ = 2α/β, α (0, 1) 1 γβ = 1 2α (Id γ f) α-averaged 9

10 Resolvent resolvent J A : D R n to monotone operator is defined as J A = (Id + A) 1 due to Minty, if A maimally monotone, then D = R n (dom(id + A 1 ) = ran(id + A) = R n iff A maimally monotone) this is important for algorithms involving resolvent we will consider resolvents to maimally monotone operators 10

11 Properties of resolvent assume A is σ-strongly monotone (σ = 0 implies monotone) Id + A is (1 + σ)-strongly monotone A Ay + ( y), y σ y 2 + y 2 = (1 + σ) y 2 properties of J A = (Id + A) 1? J A = (Id + A) 1 is (1 + σ)-cocoercive 1 1+σ 0 y σ = 0: J A is 1 2-averaged (or 1-cocoercive or firmly nonepansive) 1 σ > 0: J A is 1+σ -contractive (iteration of the resolvent converges to a fied-point, if eists) note: resolvent is single-valued 11

12 Lipschitz continuity assume A is β-lipschitz continuous, then 2 J A J A y, y y 2 + (1 β 2 ) J A J A y 2 (besides being 1-cocoercive) proof sketch: A + βid is 1 2β -cocoercive 0 y dotted: A Ay gray: (βid + A) (βid + A)y using βid = Id + (β 1)Id, the definition of a cocoercive operator, and the definition of the inverse, gives the result 12

13 Graphical representation assume A is 1-Lipschitz continuous then (besides being 1-cocoercive) J A is 1 2-strongly monotone 0 y 1-cocoercivity: J A J A y in dotted region 1 2 -strong monotonicity: J A J A y to the right of dashed line 13

14 Lipschitz continuity and strong monotonicity let A be 1-Lipschitz and σ-strongly monotone (with 0 σ < 1) σ-strong monotonicity of A (1 + σ)-cocoercivity of J A 1-Lipschitz continuity of A 1 2 -strong monotonicity of J A intersect regions to find region when both properties are present 0 y J A J A y ends up in gray region (σ = 1 9 and β = 1 in figure) 14

15 Proimal operators assume f is proper closed and conve then f maimally monotone let A = f, then: { J A (z) = argmin f() z 2} =: pro f (z) where pro f is called pro operator proof: = pro f (z) if and only if 0 f() + z z f() + z (Id + f) = (Id + f) 1 z 15

16 Proimal operator characterization the proimal operator satisfies pro f = h where h = f why? h is proper closed and conve, and h = f + Id therefore h = ( h) 1 = ( f + Id) 1 = J f can this be used to derive tighter properties of J f? 16

17 Proimal operator properties we have pro f (z) = h (z) where h = f recall equivalent dual properties (i) f is σ-strongly conve (ii) f is σ-strongly monotone (iii) f is σ-cocoercive (iv) f is 1 -Lipschitz continuous σ (v) f is 1 -smooth σ assume f is σ-strongly conve h is (1 + σ)-strongly conve h = pro f is (1 + σ)-cocoercive (same as in general case) 17

18 Lipschitz continuity let h = f , i.e. pro f = h assume that f is β-smooth and σ-strongly conve 0 σ β then h is (β + 1)-smooth and (σ + 1)-strongly conve therefore h is 1 1+β and h 1 2(1+β) 2 is ( 1 1+σ 1 finally h 1 1+β Id is strongly conve and 1+σ -smooth 1+β )-smooth 1+σ 1 1+β -cocoercive (if β > σ) 0 y h 1 1+β Id = pro f 1 1+β Id inside dashed circle h = pro f in gray area (shift by 1 1+β Id) (figure has β = 17 3 and σ = 3 17 ) 18

19 Comparison assume A is a maimal monotone operator and that f is PCC assume that A and f are 1-Lipschitz J A and pro f end up in darker and lighter gray area respectively 0 y conclusion: under Lipschitz assumptions, the resolvent of subdifferentials are confined to smaller regions 19

20 Proimal operator for separable functions consider a separable function g() = n i=1 g i( i ) the pro is also separable: pro g (z) = argmin{g() z 2 } n n = argmin{ g i ( i ) ( i z i ) 2 } = i=1 i=1 argmin 1 {g 1 ( 1 ) ( 1 z 1 ) 2 }... argmin n {g n ( n ) ( n z n ) 2 } cheap evaluation good to have in algorithms 20

21 Separability and compositions assume that g is separable, i.e., g() = n i=1 g i( i ) let h = g L where L is arbitrary linear operator the pro becomes pro h (z) = argmin{h() z 2 } = argmin{g(l) z 2 } separability is lost in general 21

22 Moreau s identity the following relation holds between the pro of f and f when f scaled by γ, we have pro f + pro f = Id pro γf + pro (γf) = pro γf + γpro γ 1 f γ 1 Id = Id when f composed with L, we have where pro γ(f L) (z) = z γl µ µ Argmin{f (µ) + γ 2 L µ γ 1 z 2 } µ (assuming that the Argmin is nonempty) these identities are very useful! 22

23 Reflected resolvent the reflected resolvent R A to a monotone operator A is defined as R A := 2J A Id it gives the reflection point (therefore its name) R A J A if A = f then reflected proimal operator given by R f = 2pro f Id (sometimes denoted rpro f or R f ) 23

24 Properties of reflected resolvent in the general case, A monotone reflected resolvent R A is β-lipschitz, what is β? β = 1, i.e., R A is nonepansive proof : 1. J A J Ay within dashed region (since J A 1-cocoercive in general case) 2. 2J A J Ay within dotted region (multiply by 2) 3. (2J A Id) (2J A Id)y = (2J A 2J Ay) ( y) in gray area (shift by Id) 0 y 24

25 Further properties of reflected resolvent properties of R A obtained by multiplying resolvent (J A ) area by 2 (radially) and shifting with Id eamples: A = f is β-smooth and σ-strongly monotone 0 y 0 y 0 y β =, σ = 1 4 β = 4, σ = 0 β = 4, σ = 1 4 left: negatively averaged, middle: averaged, right: contractive (fairly easy to visualize, can be harder to prove) 25

26 More properties of reflected proimal operator assume f is σ-strongly monotone and β-lipschitz then pro γf 1 1+γβ Id is γσ 1 -cocoercive (if β > σ) 1+γβ ( ) it can be shown that R γf is ma 1 γσ 1+γσ, γβ 1 1+γβ -contractive fir γf contraction factor optimized for γ = 1 β/σ 1 (gives a contraction factor of ) β/σ+1 σβ 26

27 A reflected resolvent identity assume that A is a maimally monotone operator and γ (0, ) then R γa (Id + γa) = Id γa proof R γa (Id + γa) = 2(Id + γa) 1 (Id + γa) (Id + γa) = 2Id (Id + γa) = (Id γa) where second step holds since (Id + γa) 1 = J γa is single-valued 27

CSC 576: Gradient Descent Algorithms

CSC 576: Gradient Descent Algorithms CSC 576: Gradient Descent Algorithms Ji Liu Department of Computer Sciences, University of Rochester December 22, 205 Introduction The gradient descent algorithm is one of the most popular optimization

More information

Math 273a: Optimization Lagrange Duality

Math 273a: Optimization Lagrange Duality Math 273a: Optimization Lagrange Duality Instructor: Wotao Yin Department of Mathematics, UCLA Winter 2015 online discussions on piazza.com Gradient descent / forward Euler assume function f is proper

More information

Diagonal Scaling in Douglas-Rachford Splitting and ADMM

Diagonal Scaling in Douglas-Rachford Splitting and ADMM 53rd IEEE Conference on Decision and Control December 15-17, 2014. Los Angeles, California, USA Diagonal Scaling in Douglas-Rachford Splitting and ADMM Pontus Giselsson and Stephen Boyd Abstract The performance

More information

Convex Functions. Pontus Giselsson

Convex Functions. Pontus Giselsson Convex Functions Pontus Giselsson 1 Today s lecture lower semicontinuity, closure, convex hull convexity preserving operations precomposition with affine mapping infimal convolution image function supremum

More information

Convex Optimization Notes

Convex Optimization Notes Convex Optimization Notes Jonathan Siegel January 2017 1 Convex Analysis This section is devoted to the study of convex functions f : B R {+ } and convex sets U B, for B a Banach space. The case of B =

More information

Dual Methods. Lecturer: Ryan Tibshirani Convex Optimization /36-725

Dual Methods. Lecturer: Ryan Tibshirani Convex Optimization /36-725 Dual Methods Lecturer: Ryan Tibshirani Conve Optimization 10-725/36-725 1 Last time: proimal Newton method Consider the problem min g() + h() where g, h are conve, g is twice differentiable, and h is simple.

More information

Lecture 23: November 19

Lecture 23: November 19 10-725/36-725: Conve Optimization Fall 2018 Lecturer: Ryan Tibshirani Lecture 23: November 19 Scribes: Charvi Rastogi, George Stoica, Shuo Li Charvi Rastogi: 23.1-23.4.2, George Stoica: 23.4.3-23.8, Shuo

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course otes for EE7C (Spring 018): Conve Optimization and Approimation Instructor: Moritz Hardt Email: hardt+ee7c@berkeley.edu Graduate Instructor: Ma Simchowitz Email: msimchow+ee7c@berkeley.edu October

More information

Dual Ascent. Ryan Tibshirani Convex Optimization

Dual Ascent. Ryan Tibshirani Convex Optimization Dual Ascent Ryan Tibshirani Conve Optimization 10-725 Last time: coordinate descent Consider the problem min f() where f() = g() + n i=1 h i( i ), with g conve and differentiable and each h i conve. Coordinate

More information

Second order forward-backward dynamical systems for monotone inclusion problems

Second order forward-backward dynamical systems for monotone inclusion problems Second order forward-backward dynamical systems for monotone inclusion problems Radu Ioan Boţ Ernö Robert Csetnek March 6, 25 Abstract. We begin by considering second order dynamical systems of the from

More information

9.3 Path independence

9.3 Path independence 66 CHAPTER 9. ONE DIMENSIONAL INTEGRALS IN SEVERAL VARIABLES 9.3 Path independence Note:??? lectures 9.3.1 Path independent integrals Let U R n be a set and ω a one-form defined on U, The integral of ω

More information

Monotone Operator Splitting Methods in Signal and Image Recovery

Monotone Operator Splitting Methods in Signal and Image Recovery Monotone Operator Splitting Methods in Signal and Image Recovery P.L. Combettes 1, J.-C. Pesquet 2, and N. Pustelnik 3 2 Univ. Pierre et Marie Curie, Paris 6 LJLL CNRS UMR 7598 2 Univ. Paris-Est LIGM CNRS

More information

Introduction to Alternating Direction Method of Multipliers

Introduction to Alternating Direction Method of Multipliers Introduction to Alternating Direction Method of Multipliers Yale Chang Machine Learning Group Meeting September 29, 2016 Yale Chang (Machine Learning Group Meeting) Introduction to Alternating Direction

More information

Optimization and Optimal Control in Banach Spaces

Optimization and Optimal Control in Banach Spaces Optimization and Optimal Control in Banach Spaces Bernhard Schmitzer October 19, 2017 1 Convex non-smooth optimization with proximal operators Remark 1.1 (Motivation). Convex optimization: easier to solve,

More information

The FTRL Algorithm with Strongly Convex Regularizers

The FTRL Algorithm with Strongly Convex Regularizers CSE599s, Spring 202, Online Learning Lecture 8-04/9/202 The FTRL Algorithm with Strongly Convex Regularizers Lecturer: Brandan McMahan Scribe: Tamara Bonaci Introduction In the last lecture, we talked

More information

3.3.1 Linear functions yet again and dot product In 2D, a homogenous linear scalar function takes the general form:

3.3.1 Linear functions yet again and dot product In 2D, a homogenous linear scalar function takes the general form: 3.3 Gradient Vector and Jacobian Matri 3 3.3 Gradient Vector and Jacobian Matri Overview: Differentiable functions have a local linear approimation. Near a given point, local changes are determined by

More information

Dual Proximal Gradient Method

Dual Proximal Gradient Method Dual Proximal Gradient Method http://bicmr.pku.edu.cn/~wenzw/opt-2016-fall.html Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes Outline 2/19 1 proximal gradient method

More information

Coordinate Update Algorithm Short Course Proximal Operators and Algorithms

Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 36 Why proximal? Newton s method: for C 2 -smooth, unconstrained problems allow

More information

Lecture 14: Newton s Method

Lecture 14: Newton s Method 10-725/36-725: Conve Optimization Fall 2016 Lecturer: Javier Pena Lecture 14: Newton s ethod Scribes: Varun Joshi, Xuan Li Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes

More information

In applications, we encounter many constrained optimization problems. Examples Basis pursuit: exact sparse recovery problem

In applications, we encounter many constrained optimization problems. Examples Basis pursuit: exact sparse recovery problem 1 Conve Analsis Main references: Vandenberghe UCLA): EECS236C - Optimiation methods for large scale sstems, http://www.seas.ucla.edu/ vandenbe/ee236c.html Parikh and Bod, Proimal algorithms, slides and

More information

Coordinate Update Algorithm Short Course Operator Splitting

Coordinate Update Algorithm Short Course Operator Splitting Coordinate Update Algorithm Short Course Operator Splitting Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 25 Operator splitting pipeline 1. Formulate a problem as 0 A(x) + B(x) with monotone operators

More information

Math 273a: Optimization Overview of First-Order Optimization Algorithms

Math 273a: Optimization Overview of First-Order Optimization Algorithms Math 273a: Optimization Overview of First-Order Optimization Algorithms Wotao Yin Department of Mathematics, UCLA online discussions on piazza.com 1 / 9 Typical flow of numerical optimization Optimization

More information

Math 273a: Optimization Subgradient Methods

Math 273a: Optimization Subgradient Methods Math 273a: Optimization Subgradient Methods Instructor: Wotao Yin Department of Mathematics, UCLA Fall 2015 online discussions on piazza.com Nonsmooth convex function Recall: For ˉx R n, f(ˉx) := {g R

More information

Coordinate Update Algorithm Short Course Subgradients and Subgradient Methods

Coordinate Update Algorithm Short Course Subgradients and Subgradient Methods Coordinate Update Algorithm Short Course Subgradients and Subgradient Methods Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 30 Notation f : H R { } is a closed proper convex function domf := {x R n

More information

Convex Optimization. Ofer Meshi. Lecture 6: Lower Bounds Constrained Optimization

Convex Optimization. Ofer Meshi. Lecture 6: Lower Bounds Constrained Optimization Convex Optimization Ofer Meshi Lecture 6: Lower Bounds Constrained Optimization Lower Bounds Some upper bounds: #iter μ 2 M #iter 2 M #iter L L μ 2 Oracle/ops GD κ log 1/ε M x # ε L # x # L # ε # με f

More information

The resolvent average of monotone operators: dominant and recessive properties

The resolvent average of monotone operators: dominant and recessive properties The resolvent average of monotone operators: dominant and recessive properties Sedi Bartz, Heinz H. Bauschke, Sarah M. Moffat, and Xianfu Wang September 30, 2015 (first revision) December 22, 2015 (second

More information

5 Systems of Equations

5 Systems of Equations Systems of Equations Concepts: Solutions to Systems of Equations-Graphically and Algebraically Solving Systems - Substitution Method Solving Systems - Elimination Method Using -Dimensional Graphs to Approximate

More information

A Dykstra-like algorithm for two monotone operators

A Dykstra-like algorithm for two monotone operators A Dykstra-like algorithm for two monotone operators Heinz H. Bauschke and Patrick L. Combettes Abstract Dykstra s algorithm employs the projectors onto two closed convex sets in a Hilbert space to construct

More information

Proximal methods. S. Villa. October 7, 2014

Proximal methods. S. Villa. October 7, 2014 Proximal methods S. Villa October 7, 2014 1 Review of the basics Often machine learning problems require the solution of minimization problems. For instance, the ERM algorithm requires to solve a problem

More information

Bregman Divergence and Mirror Descent

Bregman Divergence and Mirror Descent Bregman Divergence and Mirror Descent Bregman Divergence Motivation Generalize squared Euclidean distance to a class of distances that all share similar properties Lots of applications in machine learning,

More information

Monotonicity and Restart in Fast Gradient Methods

Monotonicity and Restart in Fast Gradient Methods 53rd IEEE Conference on Decision and Control December 5-7, 204. Los Angeles, California, USA Monotonicity and Restart in Fast Gradient Methods Pontus Giselsson and Stephen Boyd Abstract Fast gradient methods

More information

Convex Optimization. (EE227A: UC Berkeley) Lecture 15. Suvrit Sra. (Gradient methods III) 12 March, 2013

Convex Optimization. (EE227A: UC Berkeley) Lecture 15. Suvrit Sra. (Gradient methods III) 12 March, 2013 Convex Optimization (EE227A: UC Berkeley) Lecture 15 (Gradient methods III) 12 March, 2013 Suvrit Sra Optimal gradient methods 2 / 27 Optimal gradient methods We saw following efficiency estimates for

More information

Convergence of Fixed-Point Iterations

Convergence of Fixed-Point Iterations Convergence of Fixed-Point Iterations Instructor: Wotao Yin (UCLA Math) July 2016 1 / 30 Why study fixed-point iterations? Abstract many existing algorithms in optimization, numerical linear algebra, and

More information

Graph Convergence for H(, )-co-accretive Mapping with over-relaxed Proximal Point Method for Solving a Generalized Variational Inclusion Problem

Graph Convergence for H(, )-co-accretive Mapping with over-relaxed Proximal Point Method for Solving a Generalized Variational Inclusion Problem Iranian Journal of Mathematical Sciences and Informatics Vol. 12, No. 1 (2017), pp 35-46 DOI: 10.7508/ijmsi.2017.01.004 Graph Convergence for H(, )-co-accretive Mapping with over-relaxed Proximal Point

More information

Lecture 23: Conditional Gradient Method

Lecture 23: Conditional Gradient Method 10-725/36-725: Conve Optimization Spring 2015 Lecture 23: Conditional Gradient Method Lecturer: Ryan Tibshirani Scribes: Shichao Yang,Diyi Yang,Zhanpeng Fang Note: LaTeX template courtesy of UC Berkeley

More information

Lecture notes for Analysis of Algorithms : Markov decision processes

Lecture notes for Analysis of Algorithms : Markov decision processes Lecture notes for Analysis of Algorithms : Markov decision processes Lecturer: Thomas Dueholm Hansen June 6, 013 Abstract We give an introduction to infinite-horizon Markov decision processes (MDPs) with

More information

12.10 Lagrange Multipliers

12.10 Lagrange Multipliers .0 Lagrange Multipliers In the last two sections we were often solving problems involving maimizing or minimizing a function f subject to a 'constraint' equation g. For eample, we minimized the cost of

More information

1 Kernel methods & optimization

1 Kernel methods & optimization Machine Learning Class Notes 9-26-13 Prof. David Sontag 1 Kernel methods & optimization One eample of a kernel that is frequently used in practice and which allows for highly non-linear discriminant functions

More information

ADMM for monotone operators: convergence analysis and rates

ADMM for monotone operators: convergence analysis and rates ADMM for monotone operators: convergence analysis and rates Radu Ioan Boţ Ernö Robert Csetne May 4, 07 Abstract. We propose in this paper a unifying scheme for several algorithms from the literature dedicated

More information

10-725/36-725: Convex Optimization Spring Lecture 21: April 6

10-725/36-725: Convex Optimization Spring Lecture 21: April 6 10-725/36-725: Conve Optimization Spring 2015 Lecturer: Ryan Tibshirani Lecture 21: April 6 Scribes: Chiqun Zhang, Hanqi Cheng, Waleed Ammar Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:

More information

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008 Lecture 8 Plus properties, merit functions and gap functions September 28, 2008 Outline Plus-properties and F-uniqueness Equation reformulations of VI/CPs Merit functions Gap merit functions FP-I book:

More information

Approaching monotone inclusion problems via second order dynamical systems with linear and anisotropic damping

Approaching monotone inclusion problems via second order dynamical systems with linear and anisotropic damping March 0, 206 3:4 WSPC Proceedings - 9in x 6in secondorderanisotropicdamping206030 page Approaching monotone inclusion problems via second order dynamical systems with linear and anisotropic damping Radu

More information

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.18

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.18 XVIII - 1 Contraction Methods for Convex Optimization and Monotone Variational Inequalities No18 Linearized alternating direction method with Gaussian back substitution for separable convex optimization

More information

CONVERGENCE AND STABILITY OF A REGULARIZATION METHOD FOR MAXIMAL MONOTONE INCLUSIONS AND ITS APPLICATIONS TO CONVEX OPTIMIZATION

CONVERGENCE AND STABILITY OF A REGULARIZATION METHOD FOR MAXIMAL MONOTONE INCLUSIONS AND ITS APPLICATIONS TO CONVEX OPTIMIZATION Variational Analysis and Appls. F. Giannessi and A. Maugeri, Eds. Kluwer Acad. Publ., Dordrecht, 2004 CONVERGENCE AND STABILITY OF A REGULARIZATION METHOD FOR MAXIMAL MONOTONE INCLUSIONS AND ITS APPLICATIONS

More information

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16 XVI - 1 Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16 A slightly changed ADMM for convex optimization with three separable operators Bingsheng He Department of

More information

Unconstrained minimization of smooth functions

Unconstrained minimization of smooth functions Unconstrained minimization of smooth functions We want to solve min x R N f(x), where f is convex. In this section, we will assume that f is differentiable (so its gradient exists at every point), and

More information

Lecture 6: September 12

Lecture 6: September 12 10-725: Optimization Fall 2013 Lecture 6: September 12 Lecturer: Ryan Tibshirani Scribes: Micol Marchetti-Bowick Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have not

More information

Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem

Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem Charles Byrne (Charles Byrne@uml.edu) http://faculty.uml.edu/cbyrne/cbyrne.html Department of Mathematical Sciences

More information

On the Iteration Complexity of Some Projection Methods for Monotone Linear Variational Inequalities

On the Iteration Complexity of Some Projection Methods for Monotone Linear Variational Inequalities On the Iteration Complexity of Some Projection Methods for Monotone Linear Variational Inequalities Caihua Chen Xiaoling Fu Bingsheng He Xiaoming Yuan January 13, 2015 Abstract. Projection type methods

More information

Vector fields Lecture 2

Vector fields Lecture 2 Vector fields Lecture 2 Let U be an open subset of R n and v a vector field on U. We ll say that v is complete if, for every p U, there exists an integral curve, γ : R U with γ(0) = p, i.e., for every

More information

This theorem guarantees solutions to many problems you will encounter. exists, then f ( c)

This theorem guarantees solutions to many problems you will encounter. exists, then f ( c) Maimum and Minimum Values Etreme Value Theorem If f () is continuous on the closed interval [a, b], then f () achieves both a global (absolute) maimum and global minimum at some numbers c and d in [a,

More information

COMS 4771 Introduction to Machine Learning. Nakul Verma

COMS 4771 Introduction to Machine Learning. Nakul Verma COMS 4771 Introduction to Machine Learning Nakul Verma Announcements HW1 due next lecture Project details are available decide on the group and topic by Thursday Last time Generative vs. Discriminative

More information

Conditional Gradient (Frank-Wolfe) Method

Conditional Gradient (Frank-Wolfe) Method Conditional Gradient (Frank-Wolfe) Method Lecturer: Aarti Singh Co-instructor: Pradeep Ravikumar Convex Optimization 10-725/36-725 1 Outline Today: Conditional gradient method Convergence analysis Properties

More information

Subgradient Method. Ryan Tibshirani Convex Optimization

Subgradient Method. Ryan Tibshirani Convex Optimization Subgradient Method Ryan Tibshirani Convex Optimization 10-725 Consider the problem Last last time: gradient descent min x f(x) for f convex and differentiable, dom(f) = R n. Gradient descent: choose initial

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning Proximal-Gradient Mark Schmidt University of British Columbia Winter 2018 Admin Auditting/registration forms: Pick up after class today. Assignment 1: 2 late days to hand in

More information

Lecture 24 November 27

Lecture 24 November 27 EE 381V: Large Scale Optimization Fall 01 Lecture 4 November 7 Lecturer: Caramanis & Sanghavi Scribe: Jahshan Bhatti and Ken Pesyna 4.1 Mirror Descent Earlier, we motivated mirror descent as a way to improve

More information

17 Solution of Nonlinear Systems

17 Solution of Nonlinear Systems 17 Solution of Nonlinear Systems We now discuss the solution of systems of nonlinear equations. An important ingredient will be the multivariate Taylor theorem. Theorem 17.1 Let D = {x 1, x 2,..., x m

More information

A user s guide to Lojasiewicz/KL inequalities

A user s guide to Lojasiewicz/KL inequalities Other A user s guide to Lojasiewicz/KL inequalities Toulouse School of Economics, Université Toulouse I SLRA, Grenoble, 2015 Motivations behind KL f : R n R smooth ẋ(t) = f (x(t)) or x k+1 = x k λ k f

More information

Operator Splitting for Parallel and Distributed Optimization

Operator Splitting for Parallel and Distributed Optimization Operator Splitting for Parallel and Distributed Optimization Wotao Yin (UCLA Math) Shanghai Tech, SSDS 15 June 23, 2015 URL: alturl.com/2z7tv 1 / 60 What is splitting? Sun-Tzu: (400 BC) Caesar: divide-n-conquer

More information

CLASS NOTES Computational Methods for Engineering Applications I Spring 2015

CLASS NOTES Computational Methods for Engineering Applications I Spring 2015 CLASS NOTES Computational Methods for Engineering Applications I Spring 2015 Petros Koumoutsakos Gerardo Tauriello (Last update: July 2, 2015) IMPORTANT DISCLAIMERS 1. REFERENCES: Much of the material

More information

Dual methods and ADMM. Barnabas Poczos & Ryan Tibshirani Convex Optimization /36-725

Dual methods and ADMM. Barnabas Poczos & Ryan Tibshirani Convex Optimization /36-725 Dual methods and ADMM Barnabas Poczos & Ryan Tibshirani Convex Optimization 10-725/36-725 1 Given f : R n R, the function is called its conjugate Recall conjugate functions f (y) = max x R n yt x f(x)

More information

Math 273a: Optimization Subgradients of convex functions

Math 273a: Optimization Subgradients of convex functions Math 273a: Optimization Subgradients of convex functions Made by: Damek Davis Edited by Wotao Yin Department of Mathematics, UCLA Fall 2015 online discussions on piazza.com 1 / 42 Subgradients Assumptions

More information

Exercise Sketch these lines and find their intersection.

Exercise Sketch these lines and find their intersection. These are brief notes for the lecture on Friday August 21, 2009: they are not complete, but they are a guide to what I want to say today. They are not guaranteed to be correct. 1. Solving systems of linear

More information

Randomized Coordinate Descent Methods on Optimization Problems with Linearly Coupled Constraints

Randomized Coordinate Descent Methods on Optimization Problems with Linearly Coupled Constraints Randomized Coordinate Descent Methods on Optimization Problems with Linearly Coupled Constraints By I. Necoara, Y. Nesterov, and F. Glineur Lijun Xu Optimization Group Meeting November 27, 2012 Outline

More information

Lecture 1: Background on Convex Analysis

Lecture 1: Background on Convex Analysis Lecture 1: Background on Convex Analysis John Duchi PCMI 2016 Outline I Convex sets 1.1 Definitions and examples 2.2 Basic properties 3.3 Projections onto convex sets 4.4 Separating and supporting hyperplanes

More information

Firmly Nonexpansive Mappings and Maximally Monotone Operators: Correspondence and Duality

Firmly Nonexpansive Mappings and Maximally Monotone Operators: Correspondence and Duality Firmly Nonexpansive Mappings and Maximally Monotone Operators: Correspondence and Duality Heinz H. Bauschke, Sarah M. Moffat, and Xianfu Wang June 8, 2011 Abstract The notion of a firmly nonexpansive mapping

More information

Convex Optimization Lecture 6: KKT Conditions, and applications

Convex Optimization Lecture 6: KKT Conditions, and applications Convex Optimization Lecture 6: KKT Conditions, and applications Dr. Michel Baes, IFOR / ETH Zürich Quick recall of last week s lecture Various aspects of convexity: The set of minimizers is convex. Convex

More information

Lecture 5: September 12

Lecture 5: September 12 10-725/36-725: Convex Optimization Fall 2015 Lecture 5: September 12 Lecturer: Lecturer: Ryan Tibshirani Scribes: Scribes: Barun Patra and Tyler Vuong Note: LaTeX template courtesy of UC Berkeley EECS

More information

A Primal-dual Three-operator Splitting Scheme

A Primal-dual Three-operator Splitting Scheme Noname manuscript No. (will be inserted by the editor) A Primal-dual Three-operator Splitting Scheme Ming Yan Received: date / Accepted: date Abstract In this paper, we propose a new primal-dual algorithm

More information

Newton s Method. Ryan Tibshirani Convex Optimization /36-725

Newton s Method. Ryan Tibshirani Convex Optimization /36-725 Newton s Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, Properties and examples: f (y) = max x

More information

Fixed Point Theorem and Sequences in One or Two Dimensions

Fixed Point Theorem and Sequences in One or Two Dimensions Fied Point Theorem and Sequences in One or Two Dimensions Dr. Wei-Chi Yang Let us consider a recursive sequence of n+ = n + sin n and the initial value can be an real number. Then we would like to ask

More information

/ / MET Day 000 NC1^ INRTL MNVR I E E PRE SLEEP K PRE SLEEP R E

/ / MET Day 000 NC1^ INRTL MNVR I E E PRE SLEEP K PRE SLEEP R E 05//0 5:26:04 09/6/0 (259) 6 7 8 9 20 2 22 2 09/7 0 02 0 000/00 0 02 0 04 05 06 07 08 09 0 2 ay 000 ^ 0 X Y / / / / ( %/ ) 2 /0 2 ( ) ^ 4 / Y/ 2 4 5 6 7 8 9 2 X ^ X % 2 // 09/7/0 (260) ay 000 02 05//0

More information

A Unified Approach to Proximal Algorithms using Bregman Distance

A Unified Approach to Proximal Algorithms using Bregman Distance A Unified Approach to Proximal Algorithms using Bregman Distance Yi Zhou a,, Yingbin Liang a, Lixin Shen b a Department of Electrical Engineering and Computer Science, Syracuse University b Department

More information

Economics 205 Exercises

Economics 205 Exercises Economics 05 Eercises Prof. Watson, Fall 006 (Includes eaminations through Fall 003) Part 1: Basic Analysis 1. Using ε and δ, write in formal terms the meaning of lim a f() = c, where f : R R.. Write the

More information

Auxiliary-Function Methods in Optimization

Auxiliary-Function Methods in Optimization Auxiliary-Function Methods in Optimization Charles Byrne (Charles Byrne@uml.edu) http://faculty.uml.edu/cbyrne/cbyrne.html Department of Mathematical Sciences University of Massachusetts Lowell Lowell,

More information

PSY 310: Sensory and Perceptual Processes 1

PSY 310: Sensory and Perceptual Processes 1 Fourier Analysis PSY 310 Greg Francis Lecture 08 It s all waves! Representation of information One of the big issues in perception (and cognition in general) is identifying how the brain represents information

More information

1 Introduction and preliminaries

1 Introduction and preliminaries Proximal Methods for a Class of Relaxed Nonlinear Variational Inclusions Abdellatif Moudafi Université des Antilles et de la Guyane, Grimaag B.P. 7209, 97275 Schoelcher, Martinique abdellatif.moudafi@martinique.univ-ag.fr

More information

Newton s Method. Javier Peña Convex Optimization /36-725

Newton s Method. Javier Peña Convex Optimization /36-725 Newton s Method Javier Peña Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, f ( (y) = max y T x f(x) ) x Properties and

More information

Master 2 MathBigData. 3 novembre CMAP - Ecole Polytechnique

Master 2 MathBigData. 3 novembre CMAP - Ecole Polytechnique Master 2 MathBigData S. Gaïffas 1 3 novembre 2014 1 CMAP - Ecole Polytechnique 1 Supervised learning recap Introduction Loss functions, linearity 2 Penalization Introduction Ridge Sparsity Lasso 3 Some

More information

Gradient Descent. Ryan Tibshirani Convex Optimization /36-725

Gradient Descent. Ryan Tibshirani Convex Optimization /36-725 Gradient Descent Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: canonical convex programs Linear program (LP): takes the form min x subject to c T x Gx h Ax = b Quadratic program (QP): like

More information

Solution Sheet 1.4 Questions 26-31

Solution Sheet 1.4 Questions 26-31 Solution Sheet 1.4 Questions 26-31 26. Using the Limit Rules evaluate i) ii) iii) 3 2 +4+1 0 2 +4+3, 3 2 +4+1 2 +4+3, 3 2 +4+1 1 2 +4+3. Note When using a Limit Rule you must write down which Rule you

More information

M. Marques Alves Marina Geremia. November 30, 2017

M. Marques Alves Marina Geremia. November 30, 2017 Iteration complexity of an inexact Douglas-Rachford method and of a Douglas-Rachford-Tseng s F-B four-operator splitting method for solving monotone inclusions M. Marques Alves Marina Geremia November

More information

arxiv: v5 [math.oc] 12 Apr 2016

arxiv: v5 [math.oc] 12 Apr 2016 arxiv:40.8479v5 [math.oc] 2 Apr 206 Linear Convergence and Metric Selection for Douglas-Rachford Splitting and ADMM Pontus Giselsson and Stephen Boyd April 3, 206 Abstract Recently, several convergence

More information

Gradient descent. Barnabas Poczos & Ryan Tibshirani Convex Optimization /36-725

Gradient descent. Barnabas Poczos & Ryan Tibshirani Convex Optimization /36-725 Gradient descent Barnabas Poczos & Ryan Tibshirani Convex Optimization 10-725/36-725 1 Gradient descent First consider unconstrained minimization of f : R n R, convex and differentiable. We want to solve

More information

Sequential Unconstrained Minimization: A Survey

Sequential Unconstrained Minimization: A Survey Sequential Unconstrained Minimization: A Survey Charles L. Byrne February 21, 2013 Abstract The problem is to minimize a function f : X (, ], over a non-empty subset C of X, where X is an arbitrary set.

More information

Topic 5.2: Introduction to Vector Fields

Topic 5.2: Introduction to Vector Fields Math 75 Notes Topic 5.: Introduction to Vector Fields Tetbook Section: 16.1 From the Toolbo (what you need from previous classes): Know what a vector is. Be able to sketch a vector using its component

More information

Fourier Analysis PSY 310 Greg Francis. Lecture 08. Representation of information

Fourier Analysis PSY 310 Greg Francis. Lecture 08. Representation of information Fourier Analysis PSY 310 Greg Francis Lecture 08 It s all waves! Representation of information One of the big issues in perception (and cognition in general) is identifying how the brain represents information

More information

CONVERSION OF COORDINATES BETWEEN FRAMES

CONVERSION OF COORDINATES BETWEEN FRAMES ECS 178 Course Notes CONVERSION OF COORDINATES BETWEEN FRAMES Kenneth I. Joy Institute for Data Analysis and Visualization Department of Computer Science University of California, Davis Overview Frames

More information

Lecture 14: October 17

Lecture 14: October 17 1-725/36-725: Convex Optimization Fall 218 Lecture 14: October 17 Lecturer: Lecturer: Ryan Tibshirani Scribes: Pengsheng Guo, Xian Zhou Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:

More information

Primal-dual algorithms for the sum of two and three functions 1

Primal-dual algorithms for the sum of two and three functions 1 Primal-dual algorithms for the sum of two and three functions 1 Ming Yan Michigan State University, CMSE/Mathematics 1 This works is partially supported by NSF. optimization problems for primal-dual algorithms

More information

ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS

ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS MATHEMATICS OF OPERATIONS RESEARCH Vol. 28, No. 4, November 2003, pp. 677 692 Printed in U.S.A. ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS ALEXANDER SHAPIRO We discuss in this paper a class of nonsmooth

More information

About Split Proximal Algorithms for the Q-Lasso

About Split Proximal Algorithms for the Q-Lasso Thai Journal of Mathematics Volume 5 (207) Number : 7 http://thaijmath.in.cmu.ac.th ISSN 686-0209 About Split Proximal Algorithms for the Q-Lasso Abdellatif Moudafi Aix Marseille Université, CNRS-L.S.I.S

More information

Limits and the derivative function. Limits and the derivative function

Limits and the derivative function. Limits and the derivative function The Velocity Problem A particle is moving in a straight line. t is the time that has passed from the start of motion (which corresponds to t = 0) s(t) is the distance from the particle to the initial position

More information

Variational inequalities for fixed point problems of quasi-nonexpansive operators 1. Rafał Zalas 2

Variational inequalities for fixed point problems of quasi-nonexpansive operators 1. Rafał Zalas 2 University of Zielona Góra Faculty of Mathematics, Computer Science and Econometrics Summary of the Ph.D. thesis Variational inequalities for fixed point problems of quasi-nonexpansive operators 1 by Rafał

More information

arxiv: v2 [math.oc] 26 May 2018

arxiv: v2 [math.oc] 26 May 2018 Noname manuscript No. (will be inserted by the editor) A successive difference-of-conve approimation method for a class of nonconve nonsmooth optimization problems Tianiang Liu Ting Kei Pong Akiko Takeda

More information

Electromagnetic Theory Prof. D. K. Ghosh Department of Physics Indian Institute of Technology, Bombay

Electromagnetic Theory Prof. D. K. Ghosh Department of Physics Indian Institute of Technology, Bombay Electromagnetic Theory Prof. D. K. Ghosh Department of Physics Indian Institute of Technology, Bombay Lecture -1 Element of vector calculus: Scalar Field and its Gradient This is going to be about one

More information

6. Proximal gradient method

6. Proximal gradient method L. Vandenberghe EE236C (Spring 2016) 6. Proximal gradient method motivation proximal mapping proximal gradient method with fixed step size proximal gradient method with line search 6-1 Proximal mapping

More information

Lecture 5: Gradient Descent. 5.1 Unconstrained minimization problems and Gradient descent

Lecture 5: Gradient Descent. 5.1 Unconstrained minimization problems and Gradient descent 10-725/36-725: Convex Optimization Spring 2015 Lecturer: Ryan Tibshirani Lecture 5: Gradient Descent Scribes: Loc Do,2,3 Disclaimer: These notes have not been subjected to the usual scrutiny reserved for

More information

On the convergence rate of a forward-backward type primal-dual splitting algorithm for convex optimization problems

On the convergence rate of a forward-backward type primal-dual splitting algorithm for convex optimization problems On the convergence rate of a forward-backward type primal-dual splitting algorithm for convex optimization problems Radu Ioan Boţ Ernö Robert Csetnek August 5, 014 Abstract. In this paper we analyze the

More information

A general iterative algorithm for equilibrium problems and strict pseudo-contractions in Hilbert spaces

A general iterative algorithm for equilibrium problems and strict pseudo-contractions in Hilbert spaces A general iterative algorithm for equilibrium problems and strict pseudo-contractions in Hilbert spaces MING TIAN College of Science Civil Aviation University of China Tianjin 300300, China P. R. CHINA

More information