Operators. Pontus Giselsson
|
|
- Kristian Byrd
- 5 years ago
- Views:
Transcription
1 Operators Pontus Giselsson 1
2 Today s lecture forward step operators gradient step operators (subclass of forward step operators) resolvents proimal operators (subclass of resolvents) reflected resolvents reflected proimal operators (or proimal reflectors) 2
3 Forward step operator suppose that T : R n R n is single-valued the forward step operator is (Id γt ) 3
4 Cocoercivity suppose that T is 1 β -cocoercive with β = 1 2 then Id γt with γ = 3 is α-averaged, decide α: γβ γβ 1 γβ γt γt 0.5 Id γt T is 2-cocoercive γt is 2 3 -cocoercive (Id γt ) is 3 4 -averaged generally: suppose γ (0, 2 β ) 1 then: β -cocoercivity of T γβ 2 -averagedness of (Id γt ) 4
5 Iterating the forward step operator since 1 β -cocoercivity of T γβ 2 -averagedness of (Id γt ) iterating k+1 = (Id γt ) k converges to fied-point (if eists) (if γ (0, 2 β )) 5
6 Lipschitz continuity suppose that T is 1 2-Lipschitz and γ = 1 motivate that Id γt is not averaged γt 0.5 γt 0.5 Id γt 1.5 cannot make Id γt nonepansive independent of γ iterating forward step of Lipschitz T not guaranteed to converge 6
7 Gradient step operator suppose that f : R n R is differentiable the forward step becomes the gradient step operator of f I γ f if f (proper closed and) conve, then 1 β -cocoercivity of f β-lipschitz continuity of f β-smoothness of f if f is β-smooth, the gradient method converges for γ (0, 2 β ) k+1 = (Id γ f) k (since f 1 β -cocoercive) 7
8 Stronger properties assume f is β-smooth and σ-strongly conve then γf is γβ-smooth and γσ-strongly conve then γf γσ 2 2 is γ(β σ)-smooth or γ f γσid is 1 γ(β σ) -cocoercive (Id γ f) is δ-lipschitz, decide δ 0 γ(β σ) γσ γβ 1 γβ 1 γσ γ f γσid γ f Id γ f (Id γt ) is ma(γβ 1, 1 γσ)-lipschitz contractive if 1 γσ < 1 and γβ 1 < 1, i.e., γ (0, 2 β ) gradient method k+1 = (Id γt ) k then converges linearly optimal γ (center circle) given by γ = 2 β+σ δ = β/σ 1 β/σ+1 8
9 Summary: Gradient step operator σ > 0 : 1 γβ 1 γσ 1 γβ σ = 0 : Id γ f γ (0, 2 β ) contractive optimal γ = 2 β+σ factor (= γβ 1 = 1 γσ) β/σ 1 β/σ+1 Id γ f γ = 2α/β, α (0, 1) 1 γβ = 1 2α (Id γ f) α-averaged 9
10 Resolvent resolvent J A : D R n to monotone operator is defined as J A = (Id + A) 1 due to Minty, if A maimally monotone, then D = R n (dom(id + A 1 ) = ran(id + A) = R n iff A maimally monotone) this is important for algorithms involving resolvent we will consider resolvents to maimally monotone operators 10
11 Properties of resolvent assume A is σ-strongly monotone (σ = 0 implies monotone) Id + A is (1 + σ)-strongly monotone A Ay + ( y), y σ y 2 + y 2 = (1 + σ) y 2 properties of J A = (Id + A) 1? J A = (Id + A) 1 is (1 + σ)-cocoercive 1 1+σ 0 y σ = 0: J A is 1 2-averaged (or 1-cocoercive or firmly nonepansive) 1 σ > 0: J A is 1+σ -contractive (iteration of the resolvent converges to a fied-point, if eists) note: resolvent is single-valued 11
12 Lipschitz continuity assume A is β-lipschitz continuous, then 2 J A J A y, y y 2 + (1 β 2 ) J A J A y 2 (besides being 1-cocoercive) proof sketch: A + βid is 1 2β -cocoercive 0 y dotted: A Ay gray: (βid + A) (βid + A)y using βid = Id + (β 1)Id, the definition of a cocoercive operator, and the definition of the inverse, gives the result 12
13 Graphical representation assume A is 1-Lipschitz continuous then (besides being 1-cocoercive) J A is 1 2-strongly monotone 0 y 1-cocoercivity: J A J A y in dotted region 1 2 -strong monotonicity: J A J A y to the right of dashed line 13
14 Lipschitz continuity and strong monotonicity let A be 1-Lipschitz and σ-strongly monotone (with 0 σ < 1) σ-strong monotonicity of A (1 + σ)-cocoercivity of J A 1-Lipschitz continuity of A 1 2 -strong monotonicity of J A intersect regions to find region when both properties are present 0 y J A J A y ends up in gray region (σ = 1 9 and β = 1 in figure) 14
15 Proimal operators assume f is proper closed and conve then f maimally monotone let A = f, then: { J A (z) = argmin f() z 2} =: pro f (z) where pro f is called pro operator proof: = pro f (z) if and only if 0 f() + z z f() + z (Id + f) = (Id + f) 1 z 15
16 Proimal operator characterization the proimal operator satisfies pro f = h where h = f why? h is proper closed and conve, and h = f + Id therefore h = ( h) 1 = ( f + Id) 1 = J f can this be used to derive tighter properties of J f? 16
17 Proimal operator properties we have pro f (z) = h (z) where h = f recall equivalent dual properties (i) f is σ-strongly conve (ii) f is σ-strongly monotone (iii) f is σ-cocoercive (iv) f is 1 -Lipschitz continuous σ (v) f is 1 -smooth σ assume f is σ-strongly conve h is (1 + σ)-strongly conve h = pro f is (1 + σ)-cocoercive (same as in general case) 17
18 Lipschitz continuity let h = f , i.e. pro f = h assume that f is β-smooth and σ-strongly conve 0 σ β then h is (β + 1)-smooth and (σ + 1)-strongly conve therefore h is 1 1+β and h 1 2(1+β) 2 is ( 1 1+σ 1 finally h 1 1+β Id is strongly conve and 1+σ -smooth 1+β )-smooth 1+σ 1 1+β -cocoercive (if β > σ) 0 y h 1 1+β Id = pro f 1 1+β Id inside dashed circle h = pro f in gray area (shift by 1 1+β Id) (figure has β = 17 3 and σ = 3 17 ) 18
19 Comparison assume A is a maimal monotone operator and that f is PCC assume that A and f are 1-Lipschitz J A and pro f end up in darker and lighter gray area respectively 0 y conclusion: under Lipschitz assumptions, the resolvent of subdifferentials are confined to smaller regions 19
20 Proimal operator for separable functions consider a separable function g() = n i=1 g i( i ) the pro is also separable: pro g (z) = argmin{g() z 2 } n n = argmin{ g i ( i ) ( i z i ) 2 } = i=1 i=1 argmin 1 {g 1 ( 1 ) ( 1 z 1 ) 2 }... argmin n {g n ( n ) ( n z n ) 2 } cheap evaluation good to have in algorithms 20
21 Separability and compositions assume that g is separable, i.e., g() = n i=1 g i( i ) let h = g L where L is arbitrary linear operator the pro becomes pro h (z) = argmin{h() z 2 } = argmin{g(l) z 2 } separability is lost in general 21
22 Moreau s identity the following relation holds between the pro of f and f when f scaled by γ, we have pro f + pro f = Id pro γf + pro (γf) = pro γf + γpro γ 1 f γ 1 Id = Id when f composed with L, we have where pro γ(f L) (z) = z γl µ µ Argmin{f (µ) + γ 2 L µ γ 1 z 2 } µ (assuming that the Argmin is nonempty) these identities are very useful! 22
23 Reflected resolvent the reflected resolvent R A to a monotone operator A is defined as R A := 2J A Id it gives the reflection point (therefore its name) R A J A if A = f then reflected proimal operator given by R f = 2pro f Id (sometimes denoted rpro f or R f ) 23
24 Properties of reflected resolvent in the general case, A monotone reflected resolvent R A is β-lipschitz, what is β? β = 1, i.e., R A is nonepansive proof : 1. J A J Ay within dashed region (since J A 1-cocoercive in general case) 2. 2J A J Ay within dotted region (multiply by 2) 3. (2J A Id) (2J A Id)y = (2J A 2J Ay) ( y) in gray area (shift by Id) 0 y 24
25 Further properties of reflected resolvent properties of R A obtained by multiplying resolvent (J A ) area by 2 (radially) and shifting with Id eamples: A = f is β-smooth and σ-strongly monotone 0 y 0 y 0 y β =, σ = 1 4 β = 4, σ = 0 β = 4, σ = 1 4 left: negatively averaged, middle: averaged, right: contractive (fairly easy to visualize, can be harder to prove) 25
26 More properties of reflected proimal operator assume f is σ-strongly monotone and β-lipschitz then pro γf 1 1+γβ Id is γσ 1 -cocoercive (if β > σ) 1+γβ ( ) it can be shown that R γf is ma 1 γσ 1+γσ, γβ 1 1+γβ -contractive fir γf contraction factor optimized for γ = 1 β/σ 1 (gives a contraction factor of ) β/σ+1 σβ 26
27 A reflected resolvent identity assume that A is a maimally monotone operator and γ (0, ) then R γa (Id + γa) = Id γa proof R γa (Id + γa) = 2(Id + γa) 1 (Id + γa) (Id + γa) = 2Id (Id + γa) = (Id γa) where second step holds since (Id + γa) 1 = J γa is single-valued 27
CSC 576: Gradient Descent Algorithms
CSC 576: Gradient Descent Algorithms Ji Liu Department of Computer Sciences, University of Rochester December 22, 205 Introduction The gradient descent algorithm is one of the most popular optimization
More informationMath 273a: Optimization Lagrange Duality
Math 273a: Optimization Lagrange Duality Instructor: Wotao Yin Department of Mathematics, UCLA Winter 2015 online discussions on piazza.com Gradient descent / forward Euler assume function f is proper
More informationDiagonal Scaling in Douglas-Rachford Splitting and ADMM
53rd IEEE Conference on Decision and Control December 15-17, 2014. Los Angeles, California, USA Diagonal Scaling in Douglas-Rachford Splitting and ADMM Pontus Giselsson and Stephen Boyd Abstract The performance
More informationConvex Functions. Pontus Giselsson
Convex Functions Pontus Giselsson 1 Today s lecture lower semicontinuity, closure, convex hull convexity preserving operations precomposition with affine mapping infimal convolution image function supremum
More informationConvex Optimization Notes
Convex Optimization Notes Jonathan Siegel January 2017 1 Convex Analysis This section is devoted to the study of convex functions f : B R {+ } and convex sets U B, for B a Banach space. The case of B =
More informationDual Methods. Lecturer: Ryan Tibshirani Convex Optimization /36-725
Dual Methods Lecturer: Ryan Tibshirani Conve Optimization 10-725/36-725 1 Last time: proimal Newton method Consider the problem min g() + h() where g, h are conve, g is twice differentiable, and h is simple.
More informationLecture 23: November 19
10-725/36-725: Conve Optimization Fall 2018 Lecturer: Ryan Tibshirani Lecture 23: November 19 Scribes: Charvi Rastogi, George Stoica, Shuo Li Charvi Rastogi: 23.1-23.4.2, George Stoica: 23.4.3-23.8, Shuo
More informationCourse Notes for EE227C (Spring 2018): Convex Optimization and Approximation
Course otes for EE7C (Spring 018): Conve Optimization and Approimation Instructor: Moritz Hardt Email: hardt+ee7c@berkeley.edu Graduate Instructor: Ma Simchowitz Email: msimchow+ee7c@berkeley.edu October
More informationDual Ascent. Ryan Tibshirani Convex Optimization
Dual Ascent Ryan Tibshirani Conve Optimization 10-725 Last time: coordinate descent Consider the problem min f() where f() = g() + n i=1 h i( i ), with g conve and differentiable and each h i conve. Coordinate
More informationSecond order forward-backward dynamical systems for monotone inclusion problems
Second order forward-backward dynamical systems for monotone inclusion problems Radu Ioan Boţ Ernö Robert Csetnek March 6, 25 Abstract. We begin by considering second order dynamical systems of the from
More information9.3 Path independence
66 CHAPTER 9. ONE DIMENSIONAL INTEGRALS IN SEVERAL VARIABLES 9.3 Path independence Note:??? lectures 9.3.1 Path independent integrals Let U R n be a set and ω a one-form defined on U, The integral of ω
More informationMonotone Operator Splitting Methods in Signal and Image Recovery
Monotone Operator Splitting Methods in Signal and Image Recovery P.L. Combettes 1, J.-C. Pesquet 2, and N. Pustelnik 3 2 Univ. Pierre et Marie Curie, Paris 6 LJLL CNRS UMR 7598 2 Univ. Paris-Est LIGM CNRS
More informationIntroduction to Alternating Direction Method of Multipliers
Introduction to Alternating Direction Method of Multipliers Yale Chang Machine Learning Group Meeting September 29, 2016 Yale Chang (Machine Learning Group Meeting) Introduction to Alternating Direction
More informationOptimization and Optimal Control in Banach Spaces
Optimization and Optimal Control in Banach Spaces Bernhard Schmitzer October 19, 2017 1 Convex non-smooth optimization with proximal operators Remark 1.1 (Motivation). Convex optimization: easier to solve,
More informationThe FTRL Algorithm with Strongly Convex Regularizers
CSE599s, Spring 202, Online Learning Lecture 8-04/9/202 The FTRL Algorithm with Strongly Convex Regularizers Lecturer: Brandan McMahan Scribe: Tamara Bonaci Introduction In the last lecture, we talked
More information3.3.1 Linear functions yet again and dot product In 2D, a homogenous linear scalar function takes the general form:
3.3 Gradient Vector and Jacobian Matri 3 3.3 Gradient Vector and Jacobian Matri Overview: Differentiable functions have a local linear approimation. Near a given point, local changes are determined by
More informationDual Proximal Gradient Method
Dual Proximal Gradient Method http://bicmr.pku.edu.cn/~wenzw/opt-2016-fall.html Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes Outline 2/19 1 proximal gradient method
More informationCoordinate Update Algorithm Short Course Proximal Operators and Algorithms
Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 36 Why proximal? Newton s method: for C 2 -smooth, unconstrained problems allow
More informationLecture 14: Newton s Method
10-725/36-725: Conve Optimization Fall 2016 Lecturer: Javier Pena Lecture 14: Newton s ethod Scribes: Varun Joshi, Xuan Li Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes
More informationIn applications, we encounter many constrained optimization problems. Examples Basis pursuit: exact sparse recovery problem
1 Conve Analsis Main references: Vandenberghe UCLA): EECS236C - Optimiation methods for large scale sstems, http://www.seas.ucla.edu/ vandenbe/ee236c.html Parikh and Bod, Proimal algorithms, slides and
More informationCoordinate Update Algorithm Short Course Operator Splitting
Coordinate Update Algorithm Short Course Operator Splitting Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 25 Operator splitting pipeline 1. Formulate a problem as 0 A(x) + B(x) with monotone operators
More informationMath 273a: Optimization Overview of First-Order Optimization Algorithms
Math 273a: Optimization Overview of First-Order Optimization Algorithms Wotao Yin Department of Mathematics, UCLA online discussions on piazza.com 1 / 9 Typical flow of numerical optimization Optimization
More informationMath 273a: Optimization Subgradient Methods
Math 273a: Optimization Subgradient Methods Instructor: Wotao Yin Department of Mathematics, UCLA Fall 2015 online discussions on piazza.com Nonsmooth convex function Recall: For ˉx R n, f(ˉx) := {g R
More informationCoordinate Update Algorithm Short Course Subgradients and Subgradient Methods
Coordinate Update Algorithm Short Course Subgradients and Subgradient Methods Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 30 Notation f : H R { } is a closed proper convex function domf := {x R n
More informationConvex Optimization. Ofer Meshi. Lecture 6: Lower Bounds Constrained Optimization
Convex Optimization Ofer Meshi Lecture 6: Lower Bounds Constrained Optimization Lower Bounds Some upper bounds: #iter μ 2 M #iter 2 M #iter L L μ 2 Oracle/ops GD κ log 1/ε M x # ε L # x # L # ε # με f
More informationThe resolvent average of monotone operators: dominant and recessive properties
The resolvent average of monotone operators: dominant and recessive properties Sedi Bartz, Heinz H. Bauschke, Sarah M. Moffat, and Xianfu Wang September 30, 2015 (first revision) December 22, 2015 (second
More information5 Systems of Equations
Systems of Equations Concepts: Solutions to Systems of Equations-Graphically and Algebraically Solving Systems - Substitution Method Solving Systems - Elimination Method Using -Dimensional Graphs to Approximate
More informationA Dykstra-like algorithm for two monotone operators
A Dykstra-like algorithm for two monotone operators Heinz H. Bauschke and Patrick L. Combettes Abstract Dykstra s algorithm employs the projectors onto two closed convex sets in a Hilbert space to construct
More informationProximal methods. S. Villa. October 7, 2014
Proximal methods S. Villa October 7, 2014 1 Review of the basics Often machine learning problems require the solution of minimization problems. For instance, the ERM algorithm requires to solve a problem
More informationBregman Divergence and Mirror Descent
Bregman Divergence and Mirror Descent Bregman Divergence Motivation Generalize squared Euclidean distance to a class of distances that all share similar properties Lots of applications in machine learning,
More informationMonotonicity and Restart in Fast Gradient Methods
53rd IEEE Conference on Decision and Control December 5-7, 204. Los Angeles, California, USA Monotonicity and Restart in Fast Gradient Methods Pontus Giselsson and Stephen Boyd Abstract Fast gradient methods
More informationConvex Optimization. (EE227A: UC Berkeley) Lecture 15. Suvrit Sra. (Gradient methods III) 12 March, 2013
Convex Optimization (EE227A: UC Berkeley) Lecture 15 (Gradient methods III) 12 March, 2013 Suvrit Sra Optimal gradient methods 2 / 27 Optimal gradient methods We saw following efficiency estimates for
More informationConvergence of Fixed-Point Iterations
Convergence of Fixed-Point Iterations Instructor: Wotao Yin (UCLA Math) July 2016 1 / 30 Why study fixed-point iterations? Abstract many existing algorithms in optimization, numerical linear algebra, and
More informationGraph Convergence for H(, )-co-accretive Mapping with over-relaxed Proximal Point Method for Solving a Generalized Variational Inclusion Problem
Iranian Journal of Mathematical Sciences and Informatics Vol. 12, No. 1 (2017), pp 35-46 DOI: 10.7508/ijmsi.2017.01.004 Graph Convergence for H(, )-co-accretive Mapping with over-relaxed Proximal Point
More informationLecture 23: Conditional Gradient Method
10-725/36-725: Conve Optimization Spring 2015 Lecture 23: Conditional Gradient Method Lecturer: Ryan Tibshirani Scribes: Shichao Yang,Diyi Yang,Zhanpeng Fang Note: LaTeX template courtesy of UC Berkeley
More informationLecture notes for Analysis of Algorithms : Markov decision processes
Lecture notes for Analysis of Algorithms : Markov decision processes Lecturer: Thomas Dueholm Hansen June 6, 013 Abstract We give an introduction to infinite-horizon Markov decision processes (MDPs) with
More information12.10 Lagrange Multipliers
.0 Lagrange Multipliers In the last two sections we were often solving problems involving maimizing or minimizing a function f subject to a 'constraint' equation g. For eample, we minimized the cost of
More information1 Kernel methods & optimization
Machine Learning Class Notes 9-26-13 Prof. David Sontag 1 Kernel methods & optimization One eample of a kernel that is frequently used in practice and which allows for highly non-linear discriminant functions
More informationADMM for monotone operators: convergence analysis and rates
ADMM for monotone operators: convergence analysis and rates Radu Ioan Boţ Ernö Robert Csetne May 4, 07 Abstract. We propose in this paper a unifying scheme for several algorithms from the literature dedicated
More information10-725/36-725: Convex Optimization Spring Lecture 21: April 6
10-725/36-725: Conve Optimization Spring 2015 Lecturer: Ryan Tibshirani Lecture 21: April 6 Scribes: Chiqun Zhang, Hanqi Cheng, Waleed Ammar Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:
More informationLecture 8 Plus properties, merit functions and gap functions. September 28, 2008
Lecture 8 Plus properties, merit functions and gap functions September 28, 2008 Outline Plus-properties and F-uniqueness Equation reformulations of VI/CPs Merit functions Gap merit functions FP-I book:
More informationApproaching monotone inclusion problems via second order dynamical systems with linear and anisotropic damping
March 0, 206 3:4 WSPC Proceedings - 9in x 6in secondorderanisotropicdamping206030 page Approaching monotone inclusion problems via second order dynamical systems with linear and anisotropic damping Radu
More informationContraction Methods for Convex Optimization and Monotone Variational Inequalities No.18
XVIII - 1 Contraction Methods for Convex Optimization and Monotone Variational Inequalities No18 Linearized alternating direction method with Gaussian back substitution for separable convex optimization
More informationCONVERGENCE AND STABILITY OF A REGULARIZATION METHOD FOR MAXIMAL MONOTONE INCLUSIONS AND ITS APPLICATIONS TO CONVEX OPTIMIZATION
Variational Analysis and Appls. F. Giannessi and A. Maugeri, Eds. Kluwer Acad. Publ., Dordrecht, 2004 CONVERGENCE AND STABILITY OF A REGULARIZATION METHOD FOR MAXIMAL MONOTONE INCLUSIONS AND ITS APPLICATIONS
More informationContraction Methods for Convex Optimization and Monotone Variational Inequalities No.16
XVI - 1 Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16 A slightly changed ADMM for convex optimization with three separable operators Bingsheng He Department of
More informationUnconstrained minimization of smooth functions
Unconstrained minimization of smooth functions We want to solve min x R N f(x), where f is convex. In this section, we will assume that f is differentiable (so its gradient exists at every point), and
More informationLecture 6: September 12
10-725: Optimization Fall 2013 Lecture 6: September 12 Lecturer: Ryan Tibshirani Scribes: Micol Marchetti-Bowick Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have not
More informationIterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem
Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem Charles Byrne (Charles Byrne@uml.edu) http://faculty.uml.edu/cbyrne/cbyrne.html Department of Mathematical Sciences
More informationOn the Iteration Complexity of Some Projection Methods for Monotone Linear Variational Inequalities
On the Iteration Complexity of Some Projection Methods for Monotone Linear Variational Inequalities Caihua Chen Xiaoling Fu Bingsheng He Xiaoming Yuan January 13, 2015 Abstract. Projection type methods
More informationVector fields Lecture 2
Vector fields Lecture 2 Let U be an open subset of R n and v a vector field on U. We ll say that v is complete if, for every p U, there exists an integral curve, γ : R U with γ(0) = p, i.e., for every
More informationThis theorem guarantees solutions to many problems you will encounter. exists, then f ( c)
Maimum and Minimum Values Etreme Value Theorem If f () is continuous on the closed interval [a, b], then f () achieves both a global (absolute) maimum and global minimum at some numbers c and d in [a,
More informationCOMS 4771 Introduction to Machine Learning. Nakul Verma
COMS 4771 Introduction to Machine Learning Nakul Verma Announcements HW1 due next lecture Project details are available decide on the group and topic by Thursday Last time Generative vs. Discriminative
More informationConditional Gradient (Frank-Wolfe) Method
Conditional Gradient (Frank-Wolfe) Method Lecturer: Aarti Singh Co-instructor: Pradeep Ravikumar Convex Optimization 10-725/36-725 1 Outline Today: Conditional gradient method Convergence analysis Properties
More informationSubgradient Method. Ryan Tibshirani Convex Optimization
Subgradient Method Ryan Tibshirani Convex Optimization 10-725 Consider the problem Last last time: gradient descent min x f(x) for f convex and differentiable, dom(f) = R n. Gradient descent: choose initial
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning Proximal-Gradient Mark Schmidt University of British Columbia Winter 2018 Admin Auditting/registration forms: Pick up after class today. Assignment 1: 2 late days to hand in
More informationLecture 24 November 27
EE 381V: Large Scale Optimization Fall 01 Lecture 4 November 7 Lecturer: Caramanis & Sanghavi Scribe: Jahshan Bhatti and Ken Pesyna 4.1 Mirror Descent Earlier, we motivated mirror descent as a way to improve
More information17 Solution of Nonlinear Systems
17 Solution of Nonlinear Systems We now discuss the solution of systems of nonlinear equations. An important ingredient will be the multivariate Taylor theorem. Theorem 17.1 Let D = {x 1, x 2,..., x m
More informationA user s guide to Lojasiewicz/KL inequalities
Other A user s guide to Lojasiewicz/KL inequalities Toulouse School of Economics, Université Toulouse I SLRA, Grenoble, 2015 Motivations behind KL f : R n R smooth ẋ(t) = f (x(t)) or x k+1 = x k λ k f
More informationOperator Splitting for Parallel and Distributed Optimization
Operator Splitting for Parallel and Distributed Optimization Wotao Yin (UCLA Math) Shanghai Tech, SSDS 15 June 23, 2015 URL: alturl.com/2z7tv 1 / 60 What is splitting? Sun-Tzu: (400 BC) Caesar: divide-n-conquer
More informationCLASS NOTES Computational Methods for Engineering Applications I Spring 2015
CLASS NOTES Computational Methods for Engineering Applications I Spring 2015 Petros Koumoutsakos Gerardo Tauriello (Last update: July 2, 2015) IMPORTANT DISCLAIMERS 1. REFERENCES: Much of the material
More informationDual methods and ADMM. Barnabas Poczos & Ryan Tibshirani Convex Optimization /36-725
Dual methods and ADMM Barnabas Poczos & Ryan Tibshirani Convex Optimization 10-725/36-725 1 Given f : R n R, the function is called its conjugate Recall conjugate functions f (y) = max x R n yt x f(x)
More informationMath 273a: Optimization Subgradients of convex functions
Math 273a: Optimization Subgradients of convex functions Made by: Damek Davis Edited by Wotao Yin Department of Mathematics, UCLA Fall 2015 online discussions on piazza.com 1 / 42 Subgradients Assumptions
More informationExercise Sketch these lines and find their intersection.
These are brief notes for the lecture on Friday August 21, 2009: they are not complete, but they are a guide to what I want to say today. They are not guaranteed to be correct. 1. Solving systems of linear
More informationRandomized Coordinate Descent Methods on Optimization Problems with Linearly Coupled Constraints
Randomized Coordinate Descent Methods on Optimization Problems with Linearly Coupled Constraints By I. Necoara, Y. Nesterov, and F. Glineur Lijun Xu Optimization Group Meeting November 27, 2012 Outline
More informationLecture 1: Background on Convex Analysis
Lecture 1: Background on Convex Analysis John Duchi PCMI 2016 Outline I Convex sets 1.1 Definitions and examples 2.2 Basic properties 3.3 Projections onto convex sets 4.4 Separating and supporting hyperplanes
More informationFirmly Nonexpansive Mappings and Maximally Monotone Operators: Correspondence and Duality
Firmly Nonexpansive Mappings and Maximally Monotone Operators: Correspondence and Duality Heinz H. Bauschke, Sarah M. Moffat, and Xianfu Wang June 8, 2011 Abstract The notion of a firmly nonexpansive mapping
More informationConvex Optimization Lecture 6: KKT Conditions, and applications
Convex Optimization Lecture 6: KKT Conditions, and applications Dr. Michel Baes, IFOR / ETH Zürich Quick recall of last week s lecture Various aspects of convexity: The set of minimizers is convex. Convex
More informationLecture 5: September 12
10-725/36-725: Convex Optimization Fall 2015 Lecture 5: September 12 Lecturer: Lecturer: Ryan Tibshirani Scribes: Scribes: Barun Patra and Tyler Vuong Note: LaTeX template courtesy of UC Berkeley EECS
More informationA Primal-dual Three-operator Splitting Scheme
Noname manuscript No. (will be inserted by the editor) A Primal-dual Three-operator Splitting Scheme Ming Yan Received: date / Accepted: date Abstract In this paper, we propose a new primal-dual algorithm
More informationNewton s Method. Ryan Tibshirani Convex Optimization /36-725
Newton s Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, Properties and examples: f (y) = max x
More informationFixed Point Theorem and Sequences in One or Two Dimensions
Fied Point Theorem and Sequences in One or Two Dimensions Dr. Wei-Chi Yang Let us consider a recursive sequence of n+ = n + sin n and the initial value can be an real number. Then we would like to ask
More information/ / MET Day 000 NC1^ INRTL MNVR I E E PRE SLEEP K PRE SLEEP R E
05//0 5:26:04 09/6/0 (259) 6 7 8 9 20 2 22 2 09/7 0 02 0 000/00 0 02 0 04 05 06 07 08 09 0 2 ay 000 ^ 0 X Y / / / / ( %/ ) 2 /0 2 ( ) ^ 4 / Y/ 2 4 5 6 7 8 9 2 X ^ X % 2 // 09/7/0 (260) ay 000 02 05//0
More informationA Unified Approach to Proximal Algorithms using Bregman Distance
A Unified Approach to Proximal Algorithms using Bregman Distance Yi Zhou a,, Yingbin Liang a, Lixin Shen b a Department of Electrical Engineering and Computer Science, Syracuse University b Department
More informationEconomics 205 Exercises
Economics 05 Eercises Prof. Watson, Fall 006 (Includes eaminations through Fall 003) Part 1: Basic Analysis 1. Using ε and δ, write in formal terms the meaning of lim a f() = c, where f : R R.. Write the
More informationAuxiliary-Function Methods in Optimization
Auxiliary-Function Methods in Optimization Charles Byrne (Charles Byrne@uml.edu) http://faculty.uml.edu/cbyrne/cbyrne.html Department of Mathematical Sciences University of Massachusetts Lowell Lowell,
More informationPSY 310: Sensory and Perceptual Processes 1
Fourier Analysis PSY 310 Greg Francis Lecture 08 It s all waves! Representation of information One of the big issues in perception (and cognition in general) is identifying how the brain represents information
More information1 Introduction and preliminaries
Proximal Methods for a Class of Relaxed Nonlinear Variational Inclusions Abdellatif Moudafi Université des Antilles et de la Guyane, Grimaag B.P. 7209, 97275 Schoelcher, Martinique abdellatif.moudafi@martinique.univ-ag.fr
More informationNewton s Method. Javier Peña Convex Optimization /36-725
Newton s Method Javier Peña Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, f ( (y) = max y T x f(x) ) x Properties and
More informationMaster 2 MathBigData. 3 novembre CMAP - Ecole Polytechnique
Master 2 MathBigData S. Gaïffas 1 3 novembre 2014 1 CMAP - Ecole Polytechnique 1 Supervised learning recap Introduction Loss functions, linearity 2 Penalization Introduction Ridge Sparsity Lasso 3 Some
More informationGradient Descent. Ryan Tibshirani Convex Optimization /36-725
Gradient Descent Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: canonical convex programs Linear program (LP): takes the form min x subject to c T x Gx h Ax = b Quadratic program (QP): like
More informationSolution Sheet 1.4 Questions 26-31
Solution Sheet 1.4 Questions 26-31 26. Using the Limit Rules evaluate i) ii) iii) 3 2 +4+1 0 2 +4+3, 3 2 +4+1 2 +4+3, 3 2 +4+1 1 2 +4+3. Note When using a Limit Rule you must write down which Rule you
More informationM. Marques Alves Marina Geremia. November 30, 2017
Iteration complexity of an inexact Douglas-Rachford method and of a Douglas-Rachford-Tseng s F-B four-operator splitting method for solving monotone inclusions M. Marques Alves Marina Geremia November
More informationarxiv: v5 [math.oc] 12 Apr 2016
arxiv:40.8479v5 [math.oc] 2 Apr 206 Linear Convergence and Metric Selection for Douglas-Rachford Splitting and ADMM Pontus Giselsson and Stephen Boyd April 3, 206 Abstract Recently, several convergence
More informationGradient descent. Barnabas Poczos & Ryan Tibshirani Convex Optimization /36-725
Gradient descent Barnabas Poczos & Ryan Tibshirani Convex Optimization 10-725/36-725 1 Gradient descent First consider unconstrained minimization of f : R n R, convex and differentiable. We want to solve
More informationSequential Unconstrained Minimization: A Survey
Sequential Unconstrained Minimization: A Survey Charles L. Byrne February 21, 2013 Abstract The problem is to minimize a function f : X (, ], over a non-empty subset C of X, where X is an arbitrary set.
More informationTopic 5.2: Introduction to Vector Fields
Math 75 Notes Topic 5.: Introduction to Vector Fields Tetbook Section: 16.1 From the Toolbo (what you need from previous classes): Know what a vector is. Be able to sketch a vector using its component
More informationFourier Analysis PSY 310 Greg Francis. Lecture 08. Representation of information
Fourier Analysis PSY 310 Greg Francis Lecture 08 It s all waves! Representation of information One of the big issues in perception (and cognition in general) is identifying how the brain represents information
More informationCONVERSION OF COORDINATES BETWEEN FRAMES
ECS 178 Course Notes CONVERSION OF COORDINATES BETWEEN FRAMES Kenneth I. Joy Institute for Data Analysis and Visualization Department of Computer Science University of California, Davis Overview Frames
More informationLecture 14: October 17
1-725/36-725: Convex Optimization Fall 218 Lecture 14: October 17 Lecturer: Lecturer: Ryan Tibshirani Scribes: Pengsheng Guo, Xian Zhou Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:
More informationPrimal-dual algorithms for the sum of two and three functions 1
Primal-dual algorithms for the sum of two and three functions 1 Ming Yan Michigan State University, CMSE/Mathematics 1 This works is partially supported by NSF. optimization problems for primal-dual algorithms
More informationON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS
MATHEMATICS OF OPERATIONS RESEARCH Vol. 28, No. 4, November 2003, pp. 677 692 Printed in U.S.A. ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS ALEXANDER SHAPIRO We discuss in this paper a class of nonsmooth
More informationAbout Split Proximal Algorithms for the Q-Lasso
Thai Journal of Mathematics Volume 5 (207) Number : 7 http://thaijmath.in.cmu.ac.th ISSN 686-0209 About Split Proximal Algorithms for the Q-Lasso Abdellatif Moudafi Aix Marseille Université, CNRS-L.S.I.S
More informationLimits and the derivative function. Limits and the derivative function
The Velocity Problem A particle is moving in a straight line. t is the time that has passed from the start of motion (which corresponds to t = 0) s(t) is the distance from the particle to the initial position
More informationVariational inequalities for fixed point problems of quasi-nonexpansive operators 1. Rafał Zalas 2
University of Zielona Góra Faculty of Mathematics, Computer Science and Econometrics Summary of the Ph.D. thesis Variational inequalities for fixed point problems of quasi-nonexpansive operators 1 by Rafał
More informationarxiv: v2 [math.oc] 26 May 2018
Noname manuscript No. (will be inserted by the editor) A successive difference-of-conve approimation method for a class of nonconve nonsmooth optimization problems Tianiang Liu Ting Kei Pong Akiko Takeda
More informationElectromagnetic Theory Prof. D. K. Ghosh Department of Physics Indian Institute of Technology, Bombay
Electromagnetic Theory Prof. D. K. Ghosh Department of Physics Indian Institute of Technology, Bombay Lecture -1 Element of vector calculus: Scalar Field and its Gradient This is going to be about one
More information6. Proximal gradient method
L. Vandenberghe EE236C (Spring 2016) 6. Proximal gradient method motivation proximal mapping proximal gradient method with fixed step size proximal gradient method with line search 6-1 Proximal mapping
More informationLecture 5: Gradient Descent. 5.1 Unconstrained minimization problems and Gradient descent
10-725/36-725: Convex Optimization Spring 2015 Lecturer: Ryan Tibshirani Lecture 5: Gradient Descent Scribes: Loc Do,2,3 Disclaimer: These notes have not been subjected to the usual scrutiny reserved for
More informationOn the convergence rate of a forward-backward type primal-dual splitting algorithm for convex optimization problems
On the convergence rate of a forward-backward type primal-dual splitting algorithm for convex optimization problems Radu Ioan Boţ Ernö Robert Csetnek August 5, 014 Abstract. In this paper we analyze the
More informationA general iterative algorithm for equilibrium problems and strict pseudo-contractions in Hilbert spaces
A general iterative algorithm for equilibrium problems and strict pseudo-contractions in Hilbert spaces MING TIAN College of Science Civil Aviation University of China Tianjin 300300, China P. R. CHINA
More information