A General Framework for a Class of Primal-Dual Algorithms for TV Minimization

Size: px
Start display at page:

Download "A General Framework for a Class of Primal-Dual Algorithms for TV Minimization"

Transcription

1 A General Framework for a Class of Primal-Dual Algorithms for TV Minimization Ernie Esser UCLA 1

2 Outline A Model Convex Minimization Problem Main Idea Behind the Primal Dual Hybrid Gradient (PDHG) Method Application to TV Minimization Problems Equivalent Primal, Dual and Saddle Point Formulations Moreau Decomposition Connections Between PDHG and Related Algorithms Proximal Forward Backward Splitting (PFBS) Alternating Minimization Algorithm (AMA) Split Bregman Alternating Direction Method of Multipliers (ADMM) Douglas Rachford Split Inexact Uzawa (SIU) Modified PDHG Application to Convex Problems with Separable Structure Projection Interpretations Example Applications Constrained TV Deblurring Compressive Sensing 2

3 A Model Convex Minimization Problem min J(Au) + H(u) (P) u Rm J, H closed proper convex H : R m (, ] J : R n (, ] A R n m We will also assume there exists an optimal solution u to (P) 3

4 Let J =, an arbitrary norm on R n PDHG Main Idea Dual norm p := max w 1 w, p, = So Au = Au = max p 1 p, Au Rewrite (P) as saddle point problem, min u max p 1 p, Au + H(u) PDHG alternates dual and primal steps of the form p k+1 = arg max p 1 p, Auk 1 2δ k p p k 2 2 u k+1 = arg min u R m pk+1, Au + H(u) + 1 2α k u u k 2 2 for appropriate parameters α k and δ k. Ref: M. ZHU, AND T. F. CHAN, An Efficient Primal-Dual Hybrid Gradient Algorithm for Total Variation Image Restoration, UCLA CAM Report [08-34], May

5 Application to TV Minimization Problems Discretize u TV using forward differences and assuming Neumann BC u TV = M r r=1 M c c=1 (D + c u r,c ) 2 + (D + r u r,c ) 2 Vectorize M r M c matrix by stacking columns, (r, c) element of matrix (c 1)M r + r element of vector Define a directed grid-shaped graph with m = M r M c nodes corresponding to elements (r, c). Index the m nodes by (c 1)M r + r and the e = 2M r Mc M r M c edges arbitrarily. 3 3 example: 5

6 Definition of D Define D R e m to be the edge-node adjacency matrix for the graph. For each edge η with endpoint indices (i, j), i < j, define: 1 for k = i, D η,k = 1 for k = j, 0 for k i, j. Interpret D as discretization of gradient To identify the edges used in each forward difference, also define E R e m such that { 1 if D η,k = 1, E η,k = 0 otherwise. 6

7 Definition of J Can use E to define norm E on R e w E = = = E T (w 2 ) i=1 1 m ( ) E T (w 2 ) i m w i 2 i=1 where w i is the vector of edge values for directed edges coming out of node i. For TV regularization, let J = E and A = D. Then J(Au) = Du E = u TV Note that dual norm p E = E T (p 2 ) 7

8 Nonlocal TV Extension By redefining the graph, the same notation extends to nonlocal TV regularization. node j Double edges in general: node i edge e ij edge e ji For a general graph: D NL e ij,i = 1 D NL e ji,i = 1 Define E NL in same way E was defined earlier. D NL e ij,j = 1 D NL e ji,j = 1 By defining a diagonal matrix W of nonnegative edge weights and letting A = WD NL, we can write u NLTV = Au E NL 8

9 PDHG for Unconstrained TV Deblurring min u TV + λ u R m 2 Ku f 2 2 min max p, Du + λ u R m p X 2 Ku f 2 2 where X = {p R e : p E 1} Apply PDHG: p k+1 = arg max p X p, Duk 1 2λτ k p p k 2 2 u k+1 = arg min u R m pk+1, Du + λ 2 Ku f λ(1 θ k) 2θ k u u k 2 2 where θ and τ are related to α and δ by θ k = λα k 1 + α k λ τ k = δ k λ 9

10 PDHG Iterations for TV Deblurring The u k+1 and p k+1 updates have closed form solutions: p k+1 = Π X ( p k + τ k λdu k) u k+1 = ( (1 θ k ) + θ k K T K ) 1 ( (1 θ k )u k + θ k (K T f 1 ) λ DT p k+1 ) where Π X is the orthogonal projection onto X defined by Π X (q) = arg min p p X q 2 2 q = ( ) E max ET (q 2 ), 1 10

11 Saddle Point Form via Legendre Transform Legendre transform of J is defined by J (p) = sup w, p J(w) w Since J = J for arbitrary closed proper convex J, we can define analogous saddle point version of (P) even when J is not a norm. J(Au) = J (Au) = sup p, Au J (p) p min u sup p min u F P (u) := J(Au) + H(u) (P) L PD (u, p) := p, Au J (p) + H(u) (PD) 11

12 Special Case When J Is a Norm When J(w) = w, J (p) is the indicator function for the unit ball in the dual norm J (p) = sup w, p w w { 0 if w, p w w = otherwise { 0 if sup = w 1 w, p 1 otherwise { 0 if p 1 (by dual norm definition) = otherwise 12

13 Dual Problem and Strong Duality The dual problem is max F D(p) p R n (D) where the dual functional F D (p) is a concave function defined by F D (p) = inf u R m L PD(u, p) = inf u R m p, Au J (p)+h(u) = J (p) H ( A T p) Having assumed u is an optimal solution to (P), it follows that there exists an optimal solution p to (D) Strong duality holds, meaning F P (u ) = F D (p ) u solves (P) and p solves (D) iff (u, p ) is saddle point of L PD Ref: R. T. ROCKAFELLAR, Convex Analysis, Princeton University Press, Princeton, NJ,

14 Split Primal Saddle Point Formulation Introduce the constraint w = Au in (P) and form the Lagrangian L P (u, w, p) = J(w) + H(u) + p, Au w The corresponding saddle point problem is max p R n inf L P(u, w, p) u R m,w R n (SPP) If (u, w, p ) is a saddle point for (SPP), then (u, p ) is a saddle point for (PD) Although p was introduced in L P as a Lagrange multiplier for the constraint Au = w, it has the same interpretation as the dual variable p in (PD). 14

15 Split Dual Saddle Point Formulation Introducing the constraint y = A T p in (D) and form the corresponding Lagrangian L D (p, y, u) = J (p) + H (y) + u, A T p y Obtain yet another saddle point problem, max u R m inf L D(p, y, u) p R n,y R m (SPD) If (p, y, u ) is a saddle point for (SPD), then (u, p ) is a saddle point for (PD). Although u was introduced in (SPD) as a Lagrange multiplier for the constraint y = A T p, it has the same interpretation as the primal variable u in (P). 15

16 (P) min u F P (u) F P (u) = J(Au) + H(u) (D) max p F D (p) F D (p) = J (p) H ( A T p) (PD) min u sup p L PD (u, p) L PD (u, p) = p, Au J (p) + H(u) (SP P ) max p inf u,w L P (u, w, p) L P (u, w, p) = J(w) + H(u) + p, Au w (SP D ) max u inf p,y L D (p, y, u) L D (p, y, u) = J (p) + H (y) + u, A T p y AMA on (SP P ) PFBS on (D) PFBS on (P) AMA on (SP D ) + 1 u 2α uk p 2δ pk 2 2 Relaxed AMA Relaxed AMA on (SP P ) on (SP D ) + δ Au 2 w α 2 AT p + y 2 2 ADMM on (SP P ) Douglas Rachford on (D) Primal-Dual Proximal Point on (PD) = PDHG Douglas Rachford on (P) ADMM on (SP D ) + 1 u 2 uk,( 1 α δat A)(u u k ) p k+1 u k + 1 p 2 pk,( 1 δ αaat )(p p k ) 2p k+1 p k 2u k u k 1 Split Inexact Uzawa on (SP P ) PDHGMp PDHGMu Split Inexact Uzawa on (SP D ) Legend: (P): Primal (D): Dual (PD): Primal-Dual (SP P ): Split Primal (SP D ): Split Dual AMA: Alternating Minimization Algorithm (4.2.1) PFBS: Proximal Forward Backward Splitting (4.2.1) ADMM: Alternating Direction Method of Multipliers (4.2.2) PDHG: Primal Dual Hybrid Gradient (4.2) PDHGM: Modified PDHG (4.2.3) Bold: Well Understood Convergence Properties 16

17 Some Convex Optimization References D. BERTSEKAS, Constrained Optimization and Lagrange Multiplier Methods, Athena Scientific, D. BERTSEKAS, Nonlinear Programming, Athena Scientific, Second Edition D. BERTSEKAS AND J. TSITSIKLIS, Parallel and Distributed Computation, Prentice Hall, S. BOYD AND L. VANDENBERGHE, Convex Analysis, Cambridge University Press, I. EKELAND AND R. TEMAM, Convex Analysis and Variational Problems, SIAM, Classics in Applied Mathematics, 28, R. T. ROCKAFELLAR, Convex Analysis, Princeton University Press, Princeton, NJ, R. T. ROCKAFELLAR, AND R. WETS, Variational Analysis, Springer, Vol. 317,

18 Moreau Decomposition Moreau s decomposition will be a main tool in demonstrating connections between the algorithms that follow. Let f R m, J a closed proper convex function on R n, and A R n m. Then: f = arg min u R m J(Au)+ 1 2α u f 2 2 +αat arg min p R n J (p)+ α 2 AT p f α 2 2 J. J. MOREAU, Proximité et dualité dans un espace hilbertien, Bull. Soc. Math. France, 93, 1965, pp

19 PDHG on (PD) p k+1 = arg min p J (p) + p, Au k + 1 2δ k p p k 2 2 u k+1 = arg min u H(u) + A T p k+1, u + 1 2α k u u k 2 2 Initialization: p 0 and u 0 arbitrary Parameters: α k, δ k > 0 Interesting special cases: α k = or δ k = Example: α k = case corresponds to PFBS on (D) Ref: E. Esser, X. Zhang, and T. F. Chan, A General Framework for a Class of First Order Primal-Dual Algorithms for TV Minimization, UCLA CAM Report [09-67], August

20 PFBS on (D) PFBS alternates a gradient descent step with a proximal step: p k+1 = arg min p R n J (p) + 1 2δ k p (p k + δ k Au k+1 ) 2 2, where u k+1 = H ( A T p k ). Since u k+1 = H ( A T p k ) A T p k H(u k+1 ), which is equivalent to u k+1 = arg min u R m H(u) + AT p k, u, PFBS on (D) can be rewritten as u k+1 = arg min u R m H(u) + AT p k, u p k+1 = arg min p R n J (p) + p, Au k δ k p p k

21 Convergence of PFBS Convergence requires H differentiable, (H ( A T p)) Lipschitz continuous with Lipschitz constant equal to 1 β, and 0 < inf δ k supδ k < 2β. {p k } converges to a solution of (D) {u k } converges to the unique solution of (P) Note that when J is a norm, PFBS on (D) can be interpreted as a gradient projection algorithm. Ref: P. COMBETTES, AND W. WAJS, Signal Recovery by Proximal Forward-Backward Splitting, Multiscale Model. Simul.,

22 AMA on Split Primal AMA applied to (SPP) alternately minimizes first the Lagrangian L P (u, w, p) with respect to u and then the augmented Lagrangian L P + δ k 2 Au w 2 2 with respect to w before updating the Lagrange multiplier p. u k+1 = arg min u R m H(u) + AT p k, u w k+1 = arg min w R n J(w) pk, w + δ k 2 Auk+1 w 2 2 p k+1 = p k + δ k (Au k+1 w k+1 ) Convergence: Requires same assumptions as for PFBS, which translates here to H being strongly convex with modulus β A 2 2 AMA on (SPP) is actually equivalent to PFBS on (D) Ref: P. TSENG, Applications of a Splitting Algorithm to Decomposition in Convex Programming and Variational Inequalities, SIAM J. Control Optim., Vol. 29, No. 1, 1991, pp

23 Equivalence by Moreau Decomposition PFBS on (D) and AMA on (SPP) have the same first step, namely u k+1 = arg min u R m H(u) + AT p k, u, so it suffices to show the last two steps of AMA (combined into one), p k+1 = (p k + δ k Au k+1 ) δ k arg min w is equivalent to the second step of PFBS, J(w) + δ k 2 w (pk + δ k Au k+1 ) δ k 2 2, p k+1 = arg min p J (p) + 1 2δ k p (p k + δ k Au k+1 ) 2 2. This is true by direct application of Moreau decomposition. 23

24 Analogous δ k = Case of PDHG PFBS on (P), p k+1 = arg min p R n J (p) + Au k, p u k+1 = arg min u R m H(u) + u, AT p k α k u u k 2 2 is analogously equivalent to AMA applied to (SPD), p k+1 = arg min p R m J (p) + Au k, p y k+1 = arg min y R m H (y) u k, y + α k 2 y + AT p k u k+1 = u k + α k ( A T p k+1 y k+1 ) 24

25 Connection to PDHG PFBS on (P) with additional proximal penalty, p k+1 = arg min p R n J (p) + Au k, p + 1 2δ k p p k 2 2 u k+1 = arg min u R m H(u) + u, AT p k α k u u k 2 2 AMA applied to (SPD) with first step relaxed by same additional penalty, p k+1 = arg min p R m J (p) + Au k, p + 1 2δ k p p k 2 2 y k+1 = arg min y R m H (y) u k, y + α k 2 y + AT p k u k+1 = u k + α k ( A T p k+1 y k+1 ) PFBS on (D) and AMA on (SPP) are connected to PDHG in the analogous way. 25

26 Connection to ADMM and Split Bregman Add augmented Lagrangian penalty to first step of AMA to get ADMM. ADMM on (SPP) ADMM on (SPD) u k+1 = arg min u R m H(u) + AT p k, u + δ k 2 Au wk 2 2 w k+1 = arg min w R n J(w) pk, w + δ k 2 w Auk p k+1 = p k + δ k (Au k+1 w k+1 ) p k+1 = arg min p R m J (p) + Au k, p + α k 2 AT p + y k 2 2 y k+1 = arg min y R m H (y) u k, y + α k 2 y + AT p k u k+1 = u k + α k ( A T p k+1 y k+1 ) Ref: D. BERTSEKAS, AND J. TSITSIKLIS, Parallel and Distributed Computation, Prentice Hall, Ref: T. GOLDSTEIN AND S. OSHER, The Split Bregman Algorithm for L1 Regularized Problems, UCLA CAM Report [08-29], April

27 Convergence of ADMM For (u k, w k, p k ) satisfying ADMM iteration iterations for (SPP) to converge to saddle point of L P (u, w, p), require H(u) + Au 2 2 to be strictly convex and δ k = δ > 0 For (p k, y k, u k ) satisfying ADMM iteration iterations for (SPD) to converge to saddle point of L D (p, y, u), require J (p) + A T p 2 2 to be strictly convex and α k = α > 0 Minimization steps can be computed inexactly as long as sum of errors is finite Ref: J. ECKSTEIN, AND D. BERTSEKAS, On the Douglas-Rachford splitting method and the proximal point algorithm for maximal monotone operators, Math. Program. 55, 1992, pp

28 Equivalence to Douglas Rachford Splitting Can apply Moreau decomposition twice along with an appropriate change of variables to show ADMM on (SPP) or (SPD) is equivalent to Douglas Rachford Splitting on (D) and (P) resp. Example: Douglas Rachford splitting on (D) with t k = p k + δ k w k : t k+1 = arg min q H ( A T q) + 1 2δ k q + t k 2p k t k p k p k+1 = arg min p J (p) + 1 2δ k p t k Ref: S. SETZER, Split Bregman Algorithm, Douglas-Rachford Splitting and Frame Shrinkage, ssetzer/pub/setzer_fba_fbs_frames08.pdf, Ref: J. ECKSTEIN, Splitting Methods for Monotone Operators with Applications to Parallel Optimization, Ph. D. Thesis, Massachusetts Institute of Technology, Dept. of Civil Engineering,

29 (P) min u F P (u) F P (u) = J(Au) + H(u) (D) max p F D (p) F D (p) = J (p) H ( A T p) (PD) min u sup p L PD (u, p) L PD (u, p) = p, Au J (p) + H(u) (SP P ) max p inf u,w L P (u, w, p) L P (u, w, p) = J(w) + H(u) + p, Au w (SP D ) max u inf p,y L D (p, y, u) L D (p, y, u) = J (p) + H (y) + u, A T p y AMA on (SP P ) PFBS on (D) PFBS on (P) AMA on (SP D ) + 1 u 2α uk p 2δ pk 2 2 Relaxed AMA Relaxed AMA on (SP P ) on (SP D ) + δ Au 2 w α 2 AT p + y 2 2 ADMM on (SP P ) Douglas Rachford on (D) Primal-Dual Proximal Point on (PD) = PDHG Douglas Rachford on (P) ADMM on (SP D ) + 1 u 2 uk,( 1 α δat A)(u u k ) p k+1 u k + 1 p 2 pk,( 1 δ αaat )(p p k ) 2p k+1 p k 2u k u k 1 Split Inexact Uzawa on (SP P ) PDHGMp PDHGMu Split Inexact Uzawa on (SP D ) Legend: (P): Primal (D): Dual (PD): Primal-Dual (SP P ): Split Primal (SP D ): Split Dual AMA: Alternating Minimization Algorithm (4.2.1) PFBS: Proximal Forward Backward Splitting (4.2.1) ADMM: Alternating Direction Method of Multipliers (4.2.2) PDHG: Primal Dual Hybrid Gradient (4.2) PDHGM: Modified PDHG (4.2.3) Bold: Well Understood Convergence Properties 29

30 Decoupling Variables One can add additional proximal-like penalties to ADMM iterations and obtain a more explicit algorithm that still converges. Given a step of the ADMM algorithm of the form u k+1 = arg min u J(u) + δ 2 Ku fk 2 2, modify the objective functional by adding 1 2 ( ) 1 u u k, α δkt K (u u k ), where α is chosen such that 0 < α < 1 δ K T K. Modified update is given by u k+1 = arg min u J(u) + 1 2α u uk + αδk T (Ku k f k ) 2. Ref: X. ZHANG, M. BURGER, X. BRESSON, AND S. OSHER, Bregmanized Nonlocal Regularization for Deconvolution and Sparse Reconstruction, UCLA CAM Report [09-03]

31 Split Inexact Uzawa Method Consider adding 1 2 p pk, ( 1 δ k α k AA T )(p p k ) to the first step of ADMM on (SPD), with 0 < δ k < 1 α k A 2. Split Inexact Uzawa on (SPD) p k+1 = arg min p R n J (p) + 1 2δ k p p k δ k Au k + α k δ k A(A T p k + y k ) 2 2 y k+1 = arg min y R m H (y) u k, y + α k 2 y + AT p k u k+1 = u k + α k ( A T p k+1 y k+1 ) Note: In general we could similarly modify both minimization steps in ADMM, but by only modifying the first step we can obtain an interesting PDHG-like interpretation Ref: X. ZHANG, A Unified Primal-Dual Algorithm Based on l 1 and Bregman Iteration, (Private Communication), April

32 Convergence of SIU on (SPD) Convergence requires α k = α > 0 δ k = δ > 0 0 < δ < 1 α A 2 Then if p is optimal for (D) and y = A T p, A T p k + y k 2 0 J (p k ) J (p ) H (y k ) H (y ) All convergent subsequences of (p k, y k, u k ) converge to a saddle point of L D Ref: X. ZHANG, A Unified Primal-Dual Algorithm Based on l 1 and Bregman Iteration, (Private Communication), April

33 Equivalence to Modified PDHG (PDHGMu) From the multiplier update, y k = uk 1 α k 1 uk α k 1 A T p k. Substitute this into the first step of SIU on (SPD) and apply Moreau decomposition to the remaining steps to obtain PDHGMu: ( p k+1 = arg min J (p) + p, A (1 + α k )u k α ) k u k 1 p R n α k 1 α k 1 u k+1 = arg min u R m H(u) + AT p k+1, u + 1 2α k u u k δ k p p k 2 2 When α k = α, PDHGMu is just PDHG with u k replaced by 2u k u k 1 Ref: T. POCK, D. CREMERS, H. BISCHOF, AND A. CHAMBOLLE, An Algorithm for Minimizing the Mumford-Shah Functional, ICCV,

34 Analogous SIU on (SPP) and PDHGMp Split Inexact Uzawa applied to (SPP): u k+1 = arg min u R m H(u) + 1 2α k u u k + α k A T p k + δ k α k A T (Au k w k ) 2 2 w k+1 = arg min w R n J(w) pk, w + δ k 2 Auk+1 w 2 2 p k+1 = p k + δ k (Au k+1 w k+1 ) PDHGMp: u k+1 = arg min u R m H(u) + A T ( (1 + δ k δ k 1 )p k p k+1 = arg min p R n J (p) p, Au k δ k p p k 2 2 δ ) k p k 1, u + 1 u u k 2 2 δ k 1 2α k 34

35 Modified PDHG Pros and Cons Pros: As simple as PDHG Explicit: avoids solving linear systems in ADMM Convergence theory for SIU methods applies Requires no extra variables beyond original primal and dual variables Cons: Stability restriction may require small step size parameters Dynamic step size schemes not theoretically justified 35

36 (P) min u F P (u) F P (u) = J(Au) + H(u) (D) max p F D (p) F D (p) = J (p) H ( A T p) (PD) min u sup p L PD (u, p) L PD (u, p) = p, Au J (p) + H(u) (SP P ) max p inf u,w L P (u, w, p) L P (u, w, p) = J(w) + H(u) + p, Au w (SP D ) max u inf p,y L D (p, y, u) L D (p, y, u) = J (p) + H (y) + u, A T p y AMA on (SP P ) PFBS on (D) PFBS on (P) AMA on (SP D ) + 1 u 2α uk p 2δ pk 2 2 Relaxed AMA Relaxed AMA on (SP P ) on (SP D ) + δ Au 2 w α 2 AT p + y 2 2 ADMM on (SP P ) Douglas Rachford on (D) Primal-Dual Proximal Point on (PD) = PDHG Douglas Rachford on (P) ADMM on (SP D ) + 1 u 2 uk,( 1 α δat A)(u u k ) p k+1 u k + 1 p 2 pk,( 1 δ αaat )(p p k ) 2p k+1 p k 2u k u k 1 Split Inexact Uzawa on (SP P ) PDHGMp PDHGMu Split Inexact Uzawa on (SP D ) Legend: (P): Primal (D): Dual (PD): Primal-Dual (SP P ): Split Primal (SP D ): Split Dual AMA: Alternating Minimization Algorithm (4.2.1) PFBS: Proximal Forward Backward Splitting (4.2.1) ADMM: Alternating Direction Method of Multipliers (4.2.2) PDHG: Primal Dual Hybrid Gradient (4.2) PDHGM: Modified PDHG (4.2.3) Bold: Well Understood Convergence Properties 36

37 Convex Problems with Separable Structure Many seemingly more complicated problems can be written in the form (P). Example: N φ i (B i A i u + b i ) + H(u) i=1 = N J i (A i u) + H(u) i=1 = J(Au) + H(u), where A = A 1. and J i(z i ) = φ i (B i z i + b i ). Let p = p 1.. Then A N p N J (p) = N Ji (p i ) i=1 37

38 Decoupling Minimization Steps with PDHG Applying PDHGMp to min u N i=1 φ i(b i A i u + b i ) + H(u) yields: u k+1 = arg min u H(u) + 1 2α p k+1 i = arg min p i J i (p i ) + 1 2δ ( u u k α N i=1 pi ( p k i + δa i u k+1) 2 2 A T i (2p k i p k 1 i )) 2 2 i = 1,..., N Need 0 < αδ < 1 A 2 for stability 38

39 Convex Constraints as Indicator Functions Given a constraint of the form u S where S is convex, we can enforce the constraint by adding to the objective functional the indicator function for S, g S (u) = { 0 if u S otherwise In this way, even constrained minimization problems can be written in the form of (P). If PDHG or a related method is used to decouple the terms of the functional into separate minimization steps, the step involving g S can be interpreted as an orthogonal projection onto S. arg min u g S (u) + 1 2α u f 2 2 = Π S (f) 39

40 Effective Operator Splitting Modified PDHG and related algorithms are most effective for problems with separable structure where the decoupled minimization steps are easy to solve. Main examples are when H, J i OR H, J i are Indicator functions for convex sets that are easy to project onto l 2, l or l 1 -like norms l 2 norm squared Any of the above composed with linear operators 40

41 Soft Thresholding for 1 When J(u) = u 1, define soft thresholding operator S α (f) by S α (f) = arg minj(u) + 1 u 2α u f 2 2 { f α sign(f) if f > α = 0 otherwise Can use Moreau decomposition to reinterpret S α (f) in terms of a projection S α (f) = f α arg min p J (p) + α 2 = f απ {p: p 1}( f α ) = f Π {p: p α}(f) p f α 2 2 where Π(p) = p max( p,1) is orthogonal projection onto {p : p 1}. 41

42 Soft Thresholding for E When J(u) = u TV = Du E, problems arise of the form S α (z) = arg min w w E + 1 2α w z 2 2, which via the Moreau decomposition can be again written as a projection ( z S α (z) = z απ X α = z Π αx (z), ) where X = {p : p E 1} p Π X (p) = ( ) E max ET (p 2 ), 1 42

43 Consider TV denoising Not All Projections Are So Easy Moreau decomposition u = arg min u Du E + λ 2 u f 2 2 u = f 1 λ DT arg min p X 1 2λ DT p λf 2 2 Let z = D T p and define K = {z : z = D T p and p X}. Then u = f Π K λ but the projection is nontrivial to compute. (f), Ref: A. CHAMBOLLE, An Algorithm for Total Variation Minimization and Applications, J. Math. Imaging Vision, Vol. 20, 2004, pp

44 A Few More Easy Examples Using g S to denote the indicator function for the set S: J(u) = u 2 J (p) = g {p: p 2 1} J(u) = 1 2α u 2 2 J (p) = α 2 p 2 2 J(u) = u J (p) = g {p: p 1 1} J(u) = max(u) J (p) = g {p:p 0 and p 1 =1} Note: Although there s no simple formula for projecting a vector onto the l 1 unit ball (or its positive face) in R n, this can be computed with O(n log n) complexity. 44

45 Example: Constrained TV Deblurring can be rewritten as min u min Ku f 2 ɛ u TV Du E + g T (Ku), where g T is the indicator function for T = {z : z f 2 ɛ} In order to treat both D and K explicitly, let where A = H(u) = 0 and J(Au) = J 1 (Du) + J 2 (Ku), [ ] D. K Write the dual variable as p = [ p 1 p 2 ] and apply PDHGMp. 45

46 PDHGMp for Constrained TV Deblurring u k+1 = u k α k ( D T (2p k 1 p k 1 1 ) + K T (2p k 2 p k 1 2 ) ) p k+1 1 = Π X ( p k 1 + δ k Du k+1) p k+1 2 = p k 2 + δ k Ku k+1 δ k Π T ( p k 2 δ k + Ku k+1 where Π T is defined by ), Π T (z) = f + z f ( ). max z f 2 ɛ, 1 Note that λ in the unconstrained problem, min u u TV + λ 2 Ku f 2 2, is related to ɛ by λ = p 2 2ɛ. 46

47 Deblurring Numerical Result K convolution operator for normalized Gaussian blur with Std. dev. 3 h clean image f = Kh + η u u* PDHG PDHGMp η zero mean Gaussian noise Std. dev. 1 ɛ = iterations α =.2, δ =.55 Original, blurry/noisy and image recovered from 300 PDHGMp iterations 47

48 Compressive Sensing Example where K = RΓΨ T min u u 1 s.t. Ku f 2 ɛ, R is row selector (selects 10% low frequency DCT measurements) Γ is DCT transform Ψ is orthogonal Haar wavelet transform Similar to constrained deblurring example, except here KK T = I and K = K T, so it is reasonable to satisfy the constraint at each iteration. 48

49 Two PDHGMu Formulations 1. J(u) = u 1, A = I, H(u) = g S (u), S = {u : Ku f 2 ɛ} Π S (u) = (I K T K)u + K T f + (stays on constraint set) Ku f ( ) max Ku f 2 ɛ, J(u) = g T (Ku), H(u) = u 1, T = {z : z f 2 ɛ} Π T (z) = f + z f ( ) max z f 2 ɛ, 1 (simpler projection, but doesn t stay on constraint set) 49

50 Numerical Results for Constrained l 1 Original, damaged and benchmark recovered image PDHGMu PDHGRMu Ku f 2 ε 10 4 u 1 u* iterations Constraint versus iterations for PDHGMu (α = 1, δ = 1) iterations l 1 Norm Comparison (α = 1, δ = 1) 50

A GENERAL FRAMEWORK FOR A CLASS OF FIRST ORDER PRIMAL-DUAL ALGORITHMS FOR TV MINIMIZATION

A GENERAL FRAMEWORK FOR A CLASS OF FIRST ORDER PRIMAL-DUAL ALGORITHMS FOR TV MINIMIZATION A GENERAL FRAMEWORK FOR A CLASS OF FIRST ORDER PRIMAL-DUAL ALGORITHMS FOR TV MINIMIZATION ERNIE ESSER XIAOQUN ZHANG TONY CHAN Abstract. We generalize the primal-dual hybrid gradient (PDHG) algorithm proposed

More information

A GENERAL FRAMEWORK FOR A CLASS OF FIRST ORDER PRIMAL-DUAL ALGORITHMS FOR CONVEX OPTIMIZATION IN IMAGING SCIENCE

A GENERAL FRAMEWORK FOR A CLASS OF FIRST ORDER PRIMAL-DUAL ALGORITHMS FOR CONVEX OPTIMIZATION IN IMAGING SCIENCE A GENERAL FRAMEWORK FOR A CLASS OF FIRST ORDER PRIMAL-DUAL ALGORITHMS FOR CONVEX OPTIMIZATION IN IMAGING SCIENCE ERNIE ESSER XIAOQUN ZHANG TONY CHAN Abstract. We generalize the primal-dual hybrid gradient

More information

A General Framework for a Class of First Order Primal Dual Algorithms for Convex Optimization in Imaging Science

A General Framework for a Class of First Order Primal Dual Algorithms for Convex Optimization in Imaging Science A General Framework for a Class of First Order Primal Dual Algorithms for Convex Optimization in Imaging Science Ernie Esser (UCI) Georgia Tech Applied and Computational Mathematics Seminar 11-8-2010 Joint

More information

Primal Dual Methods in Imaging Science

Primal Dual Methods in Imaging Science Primal Dual Methods in Imaging Science Ernie Esser UC Irvine Joint work with Xiaoqun Zhang (Shanghai Jiaotong University) and Tony Chan (HKUST) ICIAM 2011 1 Outline Motivation to Study Primal Dual Methods

More information

Primal-dual fixed point algorithms for separable minimization problems and their applications in imaging

Primal-dual fixed point algorithms for separable minimization problems and their applications in imaging 1/38 Primal-dual fixed point algorithms for separable minimization problems and their applications in imaging Xiaoqun Zhang Department of Mathematics and Institute of Natural Sciences Shanghai Jiao Tong

More information

Bregman Iterative Methods, Lagrangian Connections, Dual Interpretations, and Applications

Bregman Iterative Methods, Lagrangian Connections, Dual Interpretations, and Applications Bregman Iterative Methods, Lagrangian Connections, Dal Interpretations, and Applications Ernie Esser UCLA 6-30-09 1 Otline Bregman Iteration Overview Method for Constrained Optimization Compare to Denoising

More information

Solving DC Programs that Promote Group 1-Sparsity

Solving DC Programs that Promote Group 1-Sparsity Solving DC Programs that Promote Group 1-Sparsity Ernie Esser Contains joint work with Xiaoqun Zhang, Yifei Lou and Jack Xin SIAM Conference on Imaging Science Hong Kong Baptist University May 14 2014

More information

A Convex Model for Image Registration

A Convex Model for Image Registration A Convex Model for Image Registration Ernie Esser January 010 Abstract Variational methods for image registration generally involve minimizing a nonconvex functional with respect to the unknown displacement

More information

Adaptive Primal Dual Optimization for Image Processing and Learning

Adaptive Primal Dual Optimization for Image Processing and Learning Adaptive Primal Dual Optimization for Image Processing and Learning Tom Goldstein Rice University tag7@rice.edu Ernie Esser University of British Columbia eesser@eos.ubc.ca Richard Baraniuk Rice University

More information

Variational Image Restoration

Variational Image Restoration Variational Image Restoration Yuling Jiao yljiaostatistics@znufe.edu.cn School of and Statistics and Mathematics ZNUFE Dec 30, 2014 Outline 1 1 Classical Variational Restoration Models and Algorithms 1.1

More information

Dual and primal-dual methods

Dual and primal-dual methods ELE 538B: Large-Scale Optimization for Data Science Dual and primal-dual methods Yuxin Chen Princeton University, Spring 2018 Outline Dual proximal gradient method Primal-dual proximal gradient method

More information

Sparse Optimization Lecture: Dual Methods, Part I

Sparse Optimization Lecture: Dual Methods, Part I Sparse Optimization Lecture: Dual Methods, Part I Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know dual (sub)gradient iteration augmented l 1 iteration

More information

Self-dual Smooth Approximations of Convex Functions via the Proximal Average

Self-dual Smooth Approximations of Convex Functions via the Proximal Average Chapter Self-dual Smooth Approximations of Convex Functions via the Proximal Average Heinz H. Bauschke, Sarah M. Moffat, and Xianfu Wang Abstract The proximal average of two convex functions has proven

More information

A Primal-dual Three-operator Splitting Scheme

A Primal-dual Three-operator Splitting Scheme Noname manuscript No. (will be inserted by the editor) A Primal-dual Three-operator Splitting Scheme Ming Yan Received: date / Accepted: date Abstract In this paper, we propose a new primal-dual algorithm

More information

Primal Dual Algorithms for Convex Optimization in Imaging Science

Primal Dual Algorithms for Convex Optimization in Imaging Science Primal Dual Algorithms for Convex Optimization in Imaging Science Ernie Esser UC Irvine MAIPCV2011 Winter School, Tohoku University 11-25-2011 1 Variational Models (1) A powerful modeling tool in image

More information

Math 273a: Optimization Overview of First-Order Optimization Algorithms

Math 273a: Optimization Overview of First-Order Optimization Algorithms Math 273a: Optimization Overview of First-Order Optimization Algorithms Wotao Yin Department of Mathematics, UCLA online discussions on piazza.com 1 / 9 Typical flow of numerical optimization Optimization

More information

Coordinate Update Algorithm Short Course Operator Splitting

Coordinate Update Algorithm Short Course Operator Splitting Coordinate Update Algorithm Short Course Operator Splitting Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 25 Operator splitting pipeline 1. Formulate a problem as 0 A(x) + B(x) with monotone operators

More information

A Dykstra-like algorithm for two monotone operators

A Dykstra-like algorithm for two monotone operators A Dykstra-like algorithm for two monotone operators Heinz H. Bauschke and Patrick L. Combettes Abstract Dykstra s algorithm employs the projectors onto two closed convex sets in a Hilbert space to construct

More information

Splitting Techniques in the Face of Huge Problem Sizes: Block-Coordinate and Block-Iterative Approaches

Splitting Techniques in the Face of Huge Problem Sizes: Block-Coordinate and Block-Iterative Approaches Splitting Techniques in the Face of Huge Problem Sizes: Block-Coordinate and Block-Iterative Approaches Patrick L. Combettes joint work with J.-C. Pesquet) Laboratoire Jacques-Louis Lions Faculté de Mathématiques

More information

Some Inexact Hybrid Proximal Augmented Lagrangian Algorithms

Some Inexact Hybrid Proximal Augmented Lagrangian Algorithms Some Inexact Hybrid Proximal Augmented Lagrangian Algorithms Carlos Humes Jr. a, Benar F. Svaiter b, Paulo J. S. Silva a, a Dept. of Computer Science, University of São Paulo, Brazil Email: {humes,rsilva}@ime.usp.br

More information

First-order methods for structured nonsmooth optimization

First-order methods for structured nonsmooth optimization First-order methods for structured nonsmooth optimization Sangwoon Yun Department of Mathematics Education Sungkyunkwan University Oct 19, 2016 Center for Mathematical Analysis & Computation, Yonsei University

More information

Accelerated Dual Gradient-Based Methods for Total Variation Image Denoising/Deblurring Problems (and other Inverse Problems)

Accelerated Dual Gradient-Based Methods for Total Variation Image Denoising/Deblurring Problems (and other Inverse Problems) Accelerated Dual Gradient-Based Methods for Total Variation Image Denoising/Deblurring Problems (and other Inverse Problems) Donghwan Kim and Jeffrey A. Fessler EECS Department, University of Michigan

More information

On convergence rate of the Douglas-Rachford operator splitting method

On convergence rate of the Douglas-Rachford operator splitting method On convergence rate of the Douglas-Rachford operator splitting method Bingsheng He and Xiaoming Yuan 2 Abstract. This note provides a simple proof on a O(/k) convergence rate for the Douglas- Rachford

More information

Dual methods for the minimization of the total variation

Dual methods for the minimization of the total variation 1 / 30 Dual methods for the minimization of the total variation Rémy Abergel supervisor Lionel Moisan MAP5 - CNRS UMR 8145 Different Learning Seminar, LTCI Thursday 21st April 2016 2 / 30 Plan 1 Introduction

More information

On the equivalence of the primal-dual hybrid gradient method and Douglas Rachford splitting

On the equivalence of the primal-dual hybrid gradient method and Douglas Rachford splitting Mathematical Programming manuscript No. (will be inserted by the editor) On the equivalence of the primal-dual hybrid gradient method and Douglas Rachford splitting Daniel O Connor Lieven Vandenberghe

More information

Contraction Methods for Convex Optimization and monotone variational inequalities No.12

Contraction Methods for Convex Optimization and monotone variational inequalities No.12 XII - 1 Contraction Methods for Convex Optimization and monotone variational inequalities No.12 Linearized alternating direction methods of multipliers for separable convex programming Bingsheng He Department

More information

Primal-dual algorithms for the sum of two and three functions 1

Primal-dual algorithms for the sum of two and three functions 1 Primal-dual algorithms for the sum of two and three functions 1 Ming Yan Michigan State University, CMSE/Mathematics 1 This works is partially supported by NSF. optimization problems for primal-dual algorithms

More information

An Algorithmic Framework of Generalized Primal-Dual Hybrid Gradient Methods for Saddle Point Problems

An Algorithmic Framework of Generalized Primal-Dual Hybrid Gradient Methods for Saddle Point Problems An Algorithmic Framework of Generalized Primal-Dual Hybrid Gradient Methods for Saddle Point Problems Bingsheng He Feng Ma 2 Xiaoming Yuan 3 January 30, 206 Abstract. The primal-dual hybrid gradient method

More information

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.11

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.11 XI - 1 Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.11 Alternating direction methods of multipliers for separable convex programming Bingsheng He Department of Mathematics

More information

Asymmetric Cheeger cut and application to multi-class unsupervised clustering

Asymmetric Cheeger cut and application to multi-class unsupervised clustering Asymmetric Cheeger cut and application to multi-class unsupervised clustering Xavier Bresson Thomas Laurent April 8, 0 Abstract Cheeger cut has recently been shown to provide excellent classification results

More information

Inexact Alternating Direction Method of Multipliers for Separable Convex Optimization

Inexact Alternating Direction Method of Multipliers for Separable Convex Optimization Inexact Alternating Direction Method of Multipliers for Separable Convex Optimization Hongchao Zhang hozhang@math.lsu.edu Department of Mathematics Center for Computation and Technology Louisiana State

More information

The Alternating Direction Method of Multipliers

The Alternating Direction Method of Multipliers The Alternating Direction Method of Multipliers With Adaptive Step Size Selection Peter Sutor, Jr. Project Advisor: Professor Tom Goldstein December 2, 2015 1 / 25 Background The Dual Problem Consider

More information

About Split Proximal Algorithms for the Q-Lasso

About Split Proximal Algorithms for the Q-Lasso Thai Journal of Mathematics Volume 5 (207) Number : 7 http://thaijmath.in.cmu.ac.th ISSN 686-0209 About Split Proximal Algorithms for the Q-Lasso Abdellatif Moudafi Aix Marseille Université, CNRS-L.S.I.S

More information

A Linearly Convergent First-order Algorithm for Total Variation Minimization in Image Processing

A Linearly Convergent First-order Algorithm for Total Variation Minimization in Image Processing A Linearly Convergent First-order Algorithm for Total Variation Minimization in Image Processing Cong D. Dang Kaiyu Dai Guanghui Lan October 9, 0 Abstract We introduce a new formulation for total variation

More information

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 9 Alternating Direction Method of Multipliers Shiqian Ma, MAT-258A: Numerical Optimization 2 Separable convex optimization a special case is min f(x)

More information

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16 XVI - 1 Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16 A slightly changed ADMM for convex optimization with three separable operators Bingsheng He Department of

More information

9. Dual decomposition and dual algorithms

9. Dual decomposition and dual algorithms EE 546, Univ of Washington, Spring 2016 9. Dual decomposition and dual algorithms dual gradient ascent example: network rate control dual decomposition and the proximal gradient method examples with simple

More information

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization /

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization / Uses of duality Geoff Gordon & Ryan Tibshirani Optimization 10-725 / 36-725 1 Remember conjugate functions Given f : R n R, the function is called its conjugate f (y) = max x R n yt x f(x) Conjugates appear

More information

Gauge optimization and duality

Gauge optimization and duality 1 / 54 Gauge optimization and duality Junfeng Yang Department of Mathematics Nanjing University Joint with Shiqian Ma, CUHK September, 2015 2 / 54 Outline Introduction Duality Lagrange duality Fenchel

More information

Operator Splitting for Parallel and Distributed Optimization

Operator Splitting for Parallel and Distributed Optimization Operator Splitting for Parallel and Distributed Optimization Wotao Yin (UCLA Math) Shanghai Tech, SSDS 15 June 23, 2015 URL: alturl.com/2z7tv 1 / 60 What is splitting? Sun-Tzu: (400 BC) Caesar: divide-n-conquer

More information

FAST ALTERNATING DIRECTION OPTIMIZATION METHODS

FAST ALTERNATING DIRECTION OPTIMIZATION METHODS FAST ALTERNATING DIRECTION OPTIMIZATION METHODS TOM GOLDSTEIN, BRENDAN O DONOGHUE, SIMON SETZER, AND RICHARD BARANIUK Abstract. Alternating direction methods are a common tool for general mathematical

More information

You should be able to...

You should be able to... Lecture Outline Gradient Projection Algorithm Constant Step Length, Varying Step Length, Diminishing Step Length Complexity Issues Gradient Projection With Exploration Projection Solving QPs: active set

More information

ADMM algorithm for demosaicking deblurring denoising

ADMM algorithm for demosaicking deblurring denoising ADMM algorithm for demosaicking deblurring denoising DANIELE GRAZIANI MORPHEME CNRS/UNS I3s 2000 route des Lucioles BP 121 93 06903 SOPHIA ANTIPOLIS CEDEX, FRANCE e.mail:graziani@i3s.unice.fr LAURE BLANC-FÉRAUD

More information

Dual Proximal Gradient Method

Dual Proximal Gradient Method Dual Proximal Gradient Method http://bicmr.pku.edu.cn/~wenzw/opt-2016-fall.html Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes Outline 2/19 1 proximal gradient method

More information

A Unified Primal-Dual Algorithm Framework Based on Bregman Iteration

A Unified Primal-Dual Algorithm Framework Based on Bregman Iteration J Sci Comput (2011) 46: 20 46 DOI 10.1007/s10915-010-9408-8 A Unified Primal-Dual Algorithm Framework Based on Bregman Iteration Xiaoqun Zhang Martin Burger Stanley Osher Received: 23 November 2009 / Revised:

More information

Accelerated primal-dual methods for linearly constrained convex problems

Accelerated primal-dual methods for linearly constrained convex problems Accelerated primal-dual methods for linearly constrained convex problems Yangyang Xu SIAM Conference on Optimization May 24, 2017 1 / 23 Accelerated proximal gradient For convex composite problem: minimize

More information

On the equivalence of the primal-dual hybrid gradient method and Douglas-Rachford splitting

On the equivalence of the primal-dual hybrid gradient method and Douglas-Rachford splitting On the equivalence of the primal-dual hybrid gradient method and Douglas-Rachford splitting Daniel O Connor Lieven Vandenberghe September 27, 2017 Abstract The primal-dual hybrid gradient (PDHG) algorithm

More information

Splitting Enables Overcoming the Curse of Dimensionality

Splitting Enables Overcoming the Curse of Dimensionality Splitting Enables Overcoming the Curse of Dimensionality Jerome Darbon Stanley Osher In this short note we briefly outline a new and remarkably fast algorithm for solving a a large class of high dimensional

More information

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization MATHEMATICS OF OPERATIONS RESEARCH Vol. 29, No. 3, August 2004, pp. 479 491 issn 0364-765X eissn 1526-5471 04 2903 0479 informs doi 10.1287/moor.1040.0103 2004 INFORMS Some Properties of the Augmented

More information

Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem

Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem Charles Byrne (Charles Byrne@uml.edu) http://faculty.uml.edu/cbyrne/cbyrne.html Department of Mathematical Sciences

More information

arxiv: v4 [math.oc] 29 Jan 2018

arxiv: v4 [math.oc] 29 Jan 2018 Noname manuscript No. (will be inserted by the editor A new primal-dual algorithm for minimizing the sum of three functions with a linear operator Ming Yan arxiv:1611.09805v4 [math.oc] 29 Jan 2018 Received:

More information

Lecture: Algorithms for Compressed Sensing

Lecture: Algorithms for Compressed Sensing 1/56 Lecture: Algorithms for Compressed Sensing Zaiwen Wen Beijing International Center For Mathematical Research Peking University http://bicmr.pku.edu.cn/~wenzw/bigdata2017.html wenzw@pku.edu.cn Acknowledgement:

More information

Coordinate Update Algorithm Short Course Proximal Operators and Algorithms

Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 36 Why proximal? Newton s method: for C 2 -smooth, unconstrained problems allow

More information

Recent developments on sparse representation

Recent developments on sparse representation Recent developments on sparse representation Zeng Tieyong Department of Mathematics, Hong Kong Baptist University Email: zeng@hkbu.edu.hk Hong Kong Baptist University Dec. 8, 2008 First Previous Next Last

More information

Dual Decomposition.

Dual Decomposition. 1/34 Dual Decomposition http://bicmr.pku.edu.cn/~wenzw/opt-2017-fall.html Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes Outline 2/34 1 Conjugate function 2 introduction:

More information

Constrained Optimization and Lagrangian Duality

Constrained Optimization and Lagrangian Duality CIS 520: Machine Learning Oct 02, 2017 Constrained Optimization and Lagrangian Duality Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may

More information

Adaptive Primal-Dual Hybrid Gradient Methods for Saddle-Point Problems

Adaptive Primal-Dual Hybrid Gradient Methods for Saddle-Point Problems Adaptive Primal-Dual Hybrid Gradient Methods for Saddle-Point Problems Tom Goldstein, Min Li, Xiaoming Yuan, Ernie Esser, Richard Baraniuk arxiv:3050546v2 [mathna] 24 Mar 205 Abstract The Primal-Dual hybrid

More information

A Unified Approach to Proximal Algorithms using Bregman Distance

A Unified Approach to Proximal Algorithms using Bregman Distance A Unified Approach to Proximal Algorithms using Bregman Distance Yi Zhou a,, Yingbin Liang a, Lixin Shen b a Department of Electrical Engineering and Computer Science, Syracuse University b Department

More information

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions International Journal of Control Vol. 00, No. 00, January 2007, 1 10 Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions I-JENG WANG and JAMES C.

More information

The Multi-Path Utility Maximization Problem

The Multi-Path Utility Maximization Problem The Multi-Path Utility Maximization Problem Xiaojun Lin and Ness B. Shroff School of Electrical and Computer Engineering Purdue University, West Lafayette, IN 47906 {linx,shroff}@ecn.purdue.edu Abstract

More information

A Tutorial on Primal-Dual Algorithm

A Tutorial on Primal-Dual Algorithm A Tutorial on Primal-Dual Algorithm Shenlong Wang University of Toronto March 31, 2016 1 / 34 Energy minimization MAP Inference for MRFs Typical energies consist of a regularization term and a data term.

More information

EE 546, Univ of Washington, Spring Proximal mapping. introduction. review of conjugate functions. proximal mapping. Proximal mapping 6 1

EE 546, Univ of Washington, Spring Proximal mapping. introduction. review of conjugate functions. proximal mapping. Proximal mapping 6 1 EE 546, Univ of Washington, Spring 2012 6. Proximal mapping introduction review of conjugate functions proximal mapping Proximal mapping 6 1 Proximal mapping the proximal mapping (prox-operator) of a convex

More information

Splitting methods for decomposing separable convex programs

Splitting methods for decomposing separable convex programs Splitting methods for decomposing separable convex programs Philippe Mahey LIMOS - ISIMA - Université Blaise Pascal PGMO, ENSTA 2013 October 4, 2013 1 / 30 Plan 1 Max Monotone Operators Proximal techniques

More information

Lecture Note 5: Semidefinite Programming for Stability Analysis

Lecture Note 5: Semidefinite Programming for Stability Analysis ECE7850: Hybrid Systems:Theory and Applications Lecture Note 5: Semidefinite Programming for Stability Analysis Wei Zhang Assistant Professor Department of Electrical and Computer Engineering Ohio State

More information

Investigating the Influence of Box-Constraints on the Solution of a Total Variation Model via an Efficient Primal-Dual Method

Investigating the Influence of Box-Constraints on the Solution of a Total Variation Model via an Efficient Primal-Dual Method Article Investigating the Influence of Box-Constraints on the Solution of a Total Variation Model via an Efficient Primal-Dual Method Andreas Langer Department of Mathematics, University of Stuttgart,

More information

A First Order Primal-Dual Algorithm for Nonconvex T V q Regularization

A First Order Primal-Dual Algorithm for Nonconvex T V q Regularization A First Order Primal-Dual Algorithm for Nonconvex T V q Regularization Thomas Möllenhoff, Evgeny Strekalovskiy, and Daniel Cremers TU Munich, Germany Abstract. We propose an efficient first order primal-dual

More information

Primal-dual Subgradient Method for Convex Problems with Functional Constraints

Primal-dual Subgradient Method for Convex Problems with Functional Constraints Primal-dual Subgradient Method for Convex Problems with Functional Constraints Yurii Nesterov, CORE/INMA (UCL) Workshop on embedded optimization EMBOPT2014 September 9, 2014 (Lucca) Yu. Nesterov Primal-dual

More information

Variable Metric Forward-Backward Algorithm

Variable Metric Forward-Backward Algorithm Variable Metric Forward-Backward Algorithm 1/37 Variable Metric Forward-Backward Algorithm for minimizing the sum of a differentiable function and a convex function E. Chouzenoux in collaboration with

More information

arxiv: v1 [cs.na] 29 Nov 2017

arxiv: v1 [cs.na] 29 Nov 2017 A fast nonconvex Compressed Sensing algorithm for highly low-sampled MR images reconstruction D. Lazzaro 1, E. Loli Piccolomini 1 and F. Zama 1 1 Department of Mathematics, University of Bologna, arxiv:1711.11075v1

More information

ACCELERATED FIRST-ORDER PRIMAL-DUAL PROXIMAL METHODS FOR LINEARLY CONSTRAINED COMPOSITE CONVEX PROGRAMMING

ACCELERATED FIRST-ORDER PRIMAL-DUAL PROXIMAL METHODS FOR LINEARLY CONSTRAINED COMPOSITE CONVEX PROGRAMMING ACCELERATED FIRST-ORDER PRIMAL-DUAL PROXIMAL METHODS FOR LINEARLY CONSTRAINED COMPOSITE CONVEX PROGRAMMING YANGYANG XU Abstract. Motivated by big data applications, first-order methods have been extremely

More information

arxiv: v1 [math.oc] 20 Jun 2014

arxiv: v1 [math.oc] 20 Jun 2014 A forward-backward view of some primal-dual optimization methods in image recovery arxiv:1406.5439v1 [math.oc] 20 Jun 2014 P. L. Combettes, 1 L. Condat, 2 J.-C. Pesquet, 3 and B. C. Vũ 4 1 Sorbonne Universités

More information

Convex Optimization and l 1 -minimization

Convex Optimization and l 1 -minimization Convex Optimization and l 1 -minimization Sangwoon Yun Computational Sciences Korea Institute for Advanced Study December 11, 2009 2009 NIMS Thematic Winter School Outline I. Convex Optimization II. l

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course otes for EE7C (Spring 018): Conve Optimization and Approimation Instructor: Moritz Hardt Email: hardt+ee7c@berkeley.edu Graduate Instructor: Ma Simchowitz Email: msimchow+ee7c@berkeley.edu October

More information

Tight Rates and Equivalence Results of Operator Splitting Schemes

Tight Rates and Equivalence Results of Operator Splitting Schemes Tight Rates and Equivalence Results of Operator Splitting Schemes Wotao Yin (UCLA Math) Workshop on Optimization for Modern Computing Joint w Damek Davis and Ming Yan UCLA CAM 14-51, 14-58, and 14-59 1

More information

Distributed Optimization via Alternating Direction Method of Multipliers

Distributed Optimization via Alternating Direction Method of Multipliers Distributed Optimization via Alternating Direction Method of Multipliers Stephen Boyd, Neal Parikh, Eric Chu, Borja Peleato Stanford University ITMANET, Stanford, January 2011 Outline precursors dual decomposition

More information

Signal Processing and Networks Optimization Part VI: Duality

Signal Processing and Networks Optimization Part VI: Duality Signal Processing and Networks Optimization Part VI: Duality Pierre Borgnat 1, Jean-Christophe Pesquet 2, Nelly Pustelnik 1 1 ENS Lyon Laboratoire de Physique CNRS UMR 5672 pierre.borgnat@ens-lyon.fr,

More information

Sequential Unconstrained Minimization: A Survey

Sequential Unconstrained Minimization: A Survey Sequential Unconstrained Minimization: A Survey Charles L. Byrne February 21, 2013 Abstract The problem is to minimize a function f : X (, ], over a non-empty subset C of X, where X is an arbitrary set.

More information

Adaptive Corrected Procedure for TVL1 Image Deblurring under Impulsive Noise

Adaptive Corrected Procedure for TVL1 Image Deblurring under Impulsive Noise Adaptive Corrected Procedure for TVL1 Image Deblurring under Impulsive Noise Minru Bai(x T) College of Mathematics and Econometrics Hunan University Joint work with Xiongjun Zhang, Qianqian Shao June 30,

More information

Convergence of Fixed-Point Iterations

Convergence of Fixed-Point Iterations Convergence of Fixed-Point Iterations Instructor: Wotao Yin (UCLA Math) July 2016 1 / 30 Why study fixed-point iterations? Abstract many existing algorithms in optimization, numerical linear algebra, and

More information

Enhanced Fritz John Optimality Conditions and Sensitivity Analysis

Enhanced Fritz John Optimality Conditions and Sensitivity Analysis Enhanced Fritz John Optimality Conditions and Sensitivity Analysis Dimitri P. Bertsekas Laboratory for Information and Decision Systems Massachusetts Institute of Technology March 2016 1 / 27 Constrained

More information

ALADIN An Algorithm for Distributed Non-Convex Optimization and Control

ALADIN An Algorithm for Distributed Non-Convex Optimization and Control ALADIN An Algorithm for Distributed Non-Convex Optimization and Control Boris Houska, Yuning Jiang, Janick Frasch, Rien Quirynen, Dimitris Kouzoupis, Moritz Diehl ShanghaiTech University, University of

More information

Brøndsted-Rockafellar property of subdifferentials of prox-bounded functions. Marc Lassonde Université des Antilles et de la Guyane

Brøndsted-Rockafellar property of subdifferentials of prox-bounded functions. Marc Lassonde Université des Antilles et de la Guyane Conference ADGO 2013 October 16, 2013 Brøndsted-Rockafellar property of subdifferentials of prox-bounded functions Marc Lassonde Université des Antilles et de la Guyane Playa Blanca, Tongoy, Chile SUBDIFFERENTIAL

More information

SOR- and Jacobi-type Iterative Methods for Solving l 1 -l 2 Problems by Way of Fenchel Duality 1

SOR- and Jacobi-type Iterative Methods for Solving l 1 -l 2 Problems by Way of Fenchel Duality 1 SOR- and Jacobi-type Iterative Methods for Solving l 1 -l 2 Problems by Way of Fenchel Duality 1 Masao Fukushima 2 July 17 2010; revised February 4 2011 Abstract We present an SOR-type algorithm and a

More information

LINEARIZED BREGMAN ITERATIONS FOR FRAME-BASED IMAGE DEBLURRING

LINEARIZED BREGMAN ITERATIONS FOR FRAME-BASED IMAGE DEBLURRING LINEARIZED BREGMAN ITERATIONS FOR FRAME-BASED IMAGE DEBLURRING JIAN-FENG CAI, STANLEY OSHER, AND ZUOWEI SHEN Abstract. Real images usually have sparse approximations under some tight frame systems derived

More information

Linear & nonlinear classifiers

Linear & nonlinear classifiers Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1396 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1396 1 / 44 Table

More information

Convex Hodge Decomposition and Regularization of Image Flows

Convex Hodge Decomposition and Regularization of Image Flows Convex Hodge Decomposition and Regularization of Image Flows Jing Yuan, Christoph Schnörr, Gabriele Steidl April 14, 2008 Abstract The total variation (TV) measure is a key concept in the field of variational

More information

Introduction to Alternating Direction Method of Multipliers

Introduction to Alternating Direction Method of Multipliers Introduction to Alternating Direction Method of Multipliers Yale Chang Machine Learning Group Meeting September 29, 2016 Yale Chang (Machine Learning Group Meeting) Introduction to Alternating Direction

More information

EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6)

EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6) EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement to the material discussed in

More information

Journal of Convex Analysis (accepted for publication) A HYBRID PROJECTION PROXIMAL POINT ALGORITHM. M. V. Solodov and B. F.

Journal of Convex Analysis (accepted for publication) A HYBRID PROJECTION PROXIMAL POINT ALGORITHM. M. V. Solodov and B. F. Journal of Convex Analysis (accepted for publication) A HYBRID PROJECTION PROXIMAL POINT ALGORITHM M. V. Solodov and B. F. Svaiter January 27, 1997 (Revised August 24, 1998) ABSTRACT We propose a modification

More information

On the interior of the simplex, we have the Hessian of d(x), Hd(x) is diagonal with ith. µd(w) + w T c. minimize. subject to w T 1 = 1,

On the interior of the simplex, we have the Hessian of d(x), Hd(x) is diagonal with ith. µd(w) + w T c. minimize. subject to w T 1 = 1, Math 30 Winter 05 Solution to Homework 3. Recognizing the convexity of g(x) := x log x, from Jensen s inequality we get d(x) n x + + x n n log x + + x n n where the equality is attained only at x = (/n,...,

More information

Recent Developments of Alternating Direction Method of Multipliers with Multi-Block Variables

Recent Developments of Alternating Direction Method of Multipliers with Multi-Block Variables Recent Developments of Alternating Direction Method of Multipliers with Multi-Block Variables Department of Systems Engineering and Engineering Management The Chinese University of Hong Kong 2014 Workshop

More information

Implicit Fixed-point Proximity Framework for Optimization Problems and Its Applications

Implicit Fixed-point Proximity Framework for Optimization Problems and Its Applications Syracuse University SURFACE Dissertations - ALL SURFACE June 2018 Implicit Fixed-point Proximity Framework for Optimization Problems and Its Applications Xiaoxia Liu Syracuse University Follow this and

More information

A Parallel Block-Coordinate Approach for Primal-Dual Splitting with Arbitrary Random Block Selection

A Parallel Block-Coordinate Approach for Primal-Dual Splitting with Arbitrary Random Block Selection EUSIPCO 2015 1/19 A Parallel Block-Coordinate Approach for Primal-Dual Splitting with Arbitrary Random Block Selection Jean-Christophe Pesquet Laboratoire d Informatique Gaspard Monge - CNRS Univ. Paris-Est

More information

Sparse Covariance Selection using Semidefinite Programming

Sparse Covariance Selection using Semidefinite Programming Sparse Covariance Selection using Semidefinite Programming A. d Aspremont ORFE, Princeton University Joint work with O. Banerjee, L. El Ghaoui & G. Natsoulis, U.C. Berkeley & Iconix Pharmaceuticals Support

More information

Lecture 24 November 27

Lecture 24 November 27 EE 381V: Large Scale Optimization Fall 01 Lecture 4 November 7 Lecturer: Caramanis & Sanghavi Scribe: Jahshan Bhatti and Ken Pesyna 4.1 Mirror Descent Earlier, we motivated mirror descent as a way to improve

More information

Merit functions and error bounds for generalized variational inequalities

Merit functions and error bounds for generalized variational inequalities J. Math. Anal. Appl. 287 2003) 405 414 www.elsevier.com/locate/jmaa Merit functions and error bounds for generalized variational inequalities M.V. Solodov 1 Instituto de Matemática Pura e Aplicada, Estrada

More information

A Primal-Dual Method for Total Variation-Based. Wavelet Domain Inpainting

A Primal-Dual Method for Total Variation-Based. Wavelet Domain Inpainting A Primal-Dual Method for Total Variation-Based 1 Wavelet Domain Inpainting You-Wei Wen, Raymond H. Chan, Andy M. Yip Abstract Loss of information in a wavelet domain can occur during storage or transmission

More information

On the convergence rate of a forward-backward type primal-dual splitting algorithm for convex optimization problems

On the convergence rate of a forward-backward type primal-dual splitting algorithm for convex optimization problems On the convergence rate of a forward-backward type primal-dual splitting algorithm for convex optimization problems Radu Ioan Boţ Ernö Robert Csetnek August 5, 014 Abstract. In this paper we analyze the

More information

I P IANO : I NERTIAL P ROXIMAL A LGORITHM FOR N ON -C ONVEX O PTIMIZATION

I P IANO : I NERTIAL P ROXIMAL A LGORITHM FOR N ON -C ONVEX O PTIMIZATION I P IANO : I NERTIAL P ROXIMAL A LGORITHM FOR N ON -C ONVEX O PTIMIZATION Peter Ochs University of Freiburg Germany 17.01.2017 joint work with: Thomas Brox and Thomas Pock c 2017 Peter Ochs ipiano c 1

More information

Image restoration: numerical optimisation

Image restoration: numerical optimisation Image restoration: numerical optimisation Short and partial presentation Jean-François Giovannelli Groupe Signal Image Laboratoire de l Intégration du Matériau au Système Univ. Bordeaux CNRS BINP / 6 Context

More information