An Algorithmic Framework of Generalized Primal-Dual Hybrid Gradient Methods for Saddle Point Problems

Similar documents
On convergence rate of the Douglas-Rachford operator splitting method

On the Iteration Complexity of Some Projection Methods for Monotone Linear Variational Inequalities

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16

The Direct Extension of ADMM for Multi-block Convex Minimization Problems is Not Necessarily Convergent

On the O(1/t) convergence rate of the projection and contraction methods for variational inequalities with Lipschitz continuous monotone operators

Optimal Linearized Alternating Direction Method of Multipliers for Convex Programming 1

Linearized Alternating Direction Method of Multipliers via Positive-Indefinite Proximal Regularization for Convex Programming.

Adaptive Primal Dual Optimization for Image Processing and Learning

Contraction Methods for Convex Optimization and monotone variational inequalities No.12

A relaxed customized proximal point algorithm for separable convex programming

Proximal-like contraction methods for monotone variational inequalities in a unified framework

Improving an ADMM-like Splitting Method via Positive-Indefinite Proximal Regularization for Three-Block Separable Convex Minimization

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.11

Generalized ADMM with Optimal Indefinite Proximal Term for Linearly Constrained Convex Optimization

Application of the Strictly Contractive Peaceman-Rachford Splitting Method to Multi-block Separable Convex Programming

On the acceleration of augmented Lagrangian method for linearly constrained optimization

A Primal-dual Three-operator Splitting Scheme

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.18

arxiv: v4 [math.oc] 29 Jan 2018

A General Framework for a Class of Primal-Dual Algorithms for TV Minimization

Solving DC Programs that Promote Group 1-Sparsity

A GENERAL FRAMEWORK FOR A CLASS OF FIRST ORDER PRIMAL-DUAL ALGORITHMS FOR TV MINIMIZATION

Accelerated primal-dual methods for linearly constrained convex problems

Recent Developments of Alternating Direction Method of Multipliers with Multi-Block Variables

On the Convergence and O(1/N) Complexity of a Class of Nonlinear Proximal Point Algorithms for Monotonic Variational Inequalities

A GENERAL FRAMEWORK FOR A CLASS OF FIRST ORDER PRIMAL-DUAL ALGORITHMS FOR CONVEX OPTIMIZATION IN IMAGING SCIENCE

A Unified Approach to Proximal Algorithms using Bregman Distance

Adaptive Primal-Dual Hybrid Gradient Methods for Saddle-Point Problems

A New Primal Dual Algorithm for Minimizing the Sum of Three Functions with a Linear Operator

On Stochastic Primal-Dual Hybrid Gradient Approach for Compositely Regularized Minimization

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers

ALTERNATING DIRECTION METHOD OF MULTIPLIERS WITH VARIABLE STEP SIZES

The Direct Extension of ADMM for Multi-block Convex Minimization Problems is Not Necessarily Convergent

On the equivalence of the primal-dual hybrid gradient method and Douglas Rachford splitting

ACCELERATED FIRST-ORDER PRIMAL-DUAL PROXIMAL METHODS FOR LINEARLY CONSTRAINED COMPOSITE CONVEX PROGRAMMING

Coordinate Update Algorithm Short Course Operator Splitting

Generalized Alternating Direction Method of Multipliers: New Theoretical Insights and Applications

A GENERAL INERTIAL PROXIMAL POINT METHOD FOR MIXED VARIATIONAL INEQUALITY PROBLEM

A Tutorial on Primal-Dual Algorithm

Proximal splitting methods on convex problems with a quadratic term: Relax!

Splitting Enables Overcoming the Curse of Dimensionality

ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE. Sangho Kum and Gue Myung Lee. 1. Introduction

Inexact Alternating-Direction-Based Contraction Methods for Separable Linearly Constrained Convex Optimization

Convex Hodge Decomposition of Image Flows

Proximal ADMM with larger step size for two-block separable convex programming and its application to the correlation matrices calibrating problems

Primal Dual Methods in Imaging Science

Coordinate Update Algorithm Short Course Proximal Operators and Algorithms

Accelerated Dual Gradient-Based Methods for Total Variation Image Denoising/Deblurring Problems (and other Inverse Problems)

Spectral gradient projection method for solving nonlinear monotone equations

Convergence of a Class of Stationary Iterative Methods for Saddle Point Problems

Dual methods for the minimization of the total variation

A Linearly Convergent First-order Algorithm for Total Variation Minimization in Image Processing

About Split Proximal Algorithms for the Q-Lasso

Research Article Solving the Matrix Nearness Problem in the Maximum Norm by Applying a Projection and Contraction Method

Iterative common solutions of fixed point and variational inequality problems

On the equivalence of the primal-dual hybrid gradient method and Douglas-Rachford splitting

Dual and primal-dual methods

A Relaxed Explicit Extragradient-Like Method for Solving Generalized Mixed Equilibria, Variational Inequalities and Constrained Convex Minimization

Strong Convergence Theorem by a Hybrid Extragradient-like Approximation Method for Variational Inequalities and Fixed Point Problems

Convergence rate of inexact proximal point methods with relative error criteria for convex optimization

Splitting Techniques in the Face of Huge Problem Sizes: Block-Coordinate and Block-Iterative Approaches

Journal of Convex Analysis (accepted for publication) A HYBRID PROJECTION PROXIMAL POINT ALGORITHM. M. V. Solodov and B. F.

Primal-dual algorithms for the sum of two and three functions 1

Primal-dual fixed point algorithms for separable minimization problems and their applications in imaging

Convergence rate estimates for the gradient differential inclusion

A double projection method for solving variational inequalities without monotonicity

Sparse Optimization Lecture: Dual Methods, Part I

Math 273a: Optimization Overview of First-Order Optimization Algorithms

An inexact subgradient algorithm for Equilibrium Problems

arxiv: v1 [math.oc] 28 Jan 2015

Research Article Strong Convergence of a Projected Gradient Method

Self Equivalence of the Alternating Direction Method of Multipliers

Lecture 24 November 27

Convex Hodge Decomposition and Regularization of Image Flows

STRONG CONVERGENCE OF AN ITERATIVE METHOD FOR VARIATIONAL INEQUALITY PROBLEMS AND FIXED POINT PROBLEMS

Iteration-complexity of first-order penalty methods for convex programming

Inexact Alternating Direction Method of Multipliers for Separable Convex Optimization

Tight Rates and Equivalence Results of Operator Splitting Schemes

On nonexpansive and accretive operators in Banach spaces

An Algorithm for Solving Triple Hierarchical Pseudomonotone Variational Inequalities

Extensions of Korpelevich s Extragradient Method for the Variational Inequality Problem in Euclidean Space

Distributed and Consensus Optimization for Non-smooth Image Reconstruction

SOR- and Jacobi-type Iterative Methods for Solving l 1 -l 2 Problems by Way of Fenchel Duality 1

AUGMENTED LAGRANGIAN METHODS FOR CONVEX OPTIMIZATION

SOME EXTRAGRADIENT METHODS FOR NONCONVEX QUASI VARIATIONAL INEQUALITIES

Variational Image Restoration

A Bregman alternating direction method of multipliers for sparse probabilistic Boolean network problem

A NEW ITERATIVE METHOD FOR THE SPLIT COMMON FIXED POINT PROBLEM IN HILBERT SPACES. Fenghui Wang

Complexity of the relaxed Peaceman-Rachford splitting method for the sum of two maximal strongly monotone operators

arxiv: v1 [math.oc] 23 May 2017

A first-order primal-dual algorithm with linesearch

496 B.S. HE, S.L. WANG AND H. YANG where w = x y 0 A ; Q(w) f(x) AT g(y) B T Ax + By b A ; W = X Y R r : (5) Problem (4)-(5) is denoted as MVI

Supplemental Figures: Results for Various Color-image Completion

Primal-dual coordinate descent A Coordinate Descent Primal-Dual Algorithm with Large Step Size and Possibly Non-Separable Functions

A convergence result for an Outer Approximation Scheme

A derivative-free nonmonotone line search and its application to the spectral residual method

On an iterative algorithm for variational inequalities in. Banach space

An accelerated non-euclidean hybrid proximal extragradient-type algorithm for convex-concave saddle-point problems

Lagrangian-Conic Relaxations, Part I: A Unified Framework and Its Applications to Quadratic Optimization Problems

An Accelerated Hybrid Proximal Extragradient Method for Convex Optimization and its Implications to Second-Order Methods

Transcription:

An Algorithmic Framework of Generalized Primal-Dual Hybrid Gradient Methods for Saddle Point Problems Bingsheng He Feng Ma 2 Xiaoming Yuan 3 January 30, 206 Abstract. The primal-dual hybrid gradient method (PDHG originates from the Arrow-Hurwicz method, and it has been widely used to solve saddle point problems, particularly in image processing areas. With the introduction of a combination parameter, Chambolle and Pock proposed a generalized PDHG scheme with both theoretical and numerical advantages. It has been analyzed that except for the special case where the combination parameter is, the PDHG cannot be casted to the proximal point algorithm framework due to the lack of symmetry in the matrix associated with the proximal regularization terms. The PDHG scheme is asymmetric also in the sense that one variable is updated twice while the other only updated once at each iteration. These asymmetry features also explain why more theoretical issues remain challenging for generalized PDHG schemes; for example, the worst-case convergence rate of PDHG measured by the iteration complexity in a nonergodic sense is still missing. In this paper, we further consider how to generalize the PDHG and propose an algorithmic framework of generalized PDHG schemes for saddle point problems. This algorithmic framework allows the output of the PDHG subroutine to be further updated by correction steps with constant step sizes. We investigate the restriction onto these step sizes and conduct the convergence analysis for the algorithmic framework. The algorithmic framework turns out to include some existing PDHG schemes as special cases; and it immediately yields a class of new generalized PDHG schemes by choosing different step sizes for the correction steps. In particular, a completely symmetric PDHG scheme with the golden-ratio step sizes is included. Theoretically, an advantage of the algorithmic framework is that the worst-case convergence rate measured by the iteration complexity in both the ergodic and nonergodic senses can be established. Keywords. Primal-dual hybrid gradient method, saddle point problem, convex programming, image restoration, convergence rate, variational inequalities. Introduction We consider the saddle point problem min max Φ(x, y := θ (x y T (Ax b θ 2 (y, (. x X y Y Department of Mathematics, South University of Science and Technology of China, Shenzhen, 58055, China, and Department of Mathematics, Nanjing University, Nanjing, 20093, China. This author was supported by the NSFC Grant 4756. Email: hebma@nju.edu.cn 2 College of Communications Engineering, PLA University of Science and Technology, Nanjing, 20007, China. Email: mafengnju@gmail.com 3 Department of Mathematics, Hong Kong Baptist University, Hong Kong. This author was supported by the General Research Fund from Hong Kong Research Grants Council: HKBU 230055. Email: xmyuan@hkbu.edu.hk

where A R m n, b R m, X R n and Y R m are closed convex sets, θ : R n R and θ 2 : R m R are convex but not necessarily smooth functions. The solution set of (. is assumed to be nonempty throughout our discussion. The model (. captures a variety of applications in different areas. For examples, finding a saddle point of the the Lagrangian function of the canonical convex minimization model with linear equality or inequality constraints is a special case of (.. Moreover, a number of variational image restoration problems with the total variation (TV regularization (see [24] can be reformulated as special cases of (., see details in, e.g., [4, 26, 27, 28]. Since the work [28], the primal-dual hybrid gradient (PDHG method has attracted much attention from the image processing field and it is still being widely used for many applications. We start the discussion of PDHG with the scheme proposed in [4]: x k+ = arg min{φ(x, y k + r 2 x xk 2 x X }, x k = x k+ + τ(x k+ x k, (.2 y k+ = arg max{φ( x k, y s 2 y yk 2 y Y}, where τ [0, ] is a combination parameter, r > 0 and s > 0 are proximal parameters of the regularization terms. The scheme (.2 splits the coupled term y T Ax in (. and treats the functions θ and θ 2 individually; the resulting subproblems are usually easier than the original problem (.. In [4], it was shown that the PDHG scheme (.2 is closely related to the extrapolational gradient methods in [9, 23], the Douglas-Rachford splitting method in [8, 20] and the alternating direction method of multipliers (ADMM in []. In particular, we refer to [9, 25] for the equivalence between a special case of (.2 and a linearized version of the ADMM. As analyzed in [4], the convergence of (.2 can be guaranteed under the condition rs > A T A. (.3 When τ = 0, the scheme (.2 reduces to the PDHG scheme in [28] which is indeed the Arrow- Hurwicz method in []. In addition to the numerical advantages shown in, e.g., [4, 5], the theoretical significance of extending τ = 0 to τ [0, ] was demonstrated in [4]. That is, if τ = 0, then the scheme (.2 could be divergent even if r and s are fixed at very large values; thus the condition (.3 is not sufficient to ensure the convergence of (.2. Note that it could be numerically beneficial to tune the parameters r and s as shown in, e.g., [5, 27]; and it is still possible to investigate the convergence of the PDHG scheme (.2 with adaptively-adjusting proximal parameters, see, e.g., [3, 9, 2]. Here, to expose our main idea more clearly and to avoid heavy notation, we only focus on the case where both r and s are constant and they satisfy the condition in (.3 throughout our discussion. In [5], we showed that the special case of (.2 with τ = is an application of the proximal point algorithm (PPA in [2], and thus the acceleration scheme in [3] can be immediately used to accelerate the PDHG scheme (see Algorithm 4 in [5]. Its numerical efficiency has also been verified therein. This PPA revisit has been further studied in [22], in which a preconditioning version of the PDHG scheme (.2 was proposed. When τ 4, it is shown in [5] (see also (2.7b that the matrix associated with the proximal regularization terms in (.2 is not symmetric and thus the scheme (.2 cannot be casted to an application of the PPA. The acceleration techniques in [3] are thus not applicable to (.2. But the convergence can be guaranteed if the output of the PDHG subroutine (.2 is further corrected by some correction steps (See Algorithms and 2 in [5]. The step sizes of these correction steps, however, must be calculated iteratively and they usually require expensive computation (e.g., multiplications 4 We analyzed the case τ [, ] in [5]; but for simplicity we only focus on τ [0, ] in this paper. 2

of matrices or matrices and vectors if high dimensional variables are considered. It is thus natural to ask whether or not we can further correct the output of (.2 even when τ while the step sizes of the correction steps are iteration-independent (more precisely, constants? Moreover, it is easy to notice that the PDHG scheme (.2 is not symmetric also in the sense that the variable x is updated twice while y is updated only once at each iteration. So one more question is whether or not we can update y also right after the PDHG subroutine and modify (.2 as a completely symmetric version? To answer these questions, we propose the following unified algorithmic framework that allows the variables to be updated both, either, or neither, after the PDHG subroutine (.2: (Algorithmic Framework x k = arg min{φ(x, y k + r 2 x xk 2 x X }, x k = x k + τ( x k x k, ỹ k = arg max{φ( x k, y s 2 y yk 2 y Y}, { x k+ = x k α(x k x k, y k+ = y k β(y k ỹ k, (.4a (.4b where the output of the PDHG subroutine (.2 is denoted by ( x k, ỹ k, α > 0 and β > 0 are step sizes of the correction steps to further update the variables. This new algorithmic framework is thus a combination of the PDHG subroutine (.2 with a correction step. We also call ( x k, ỹ k generated by (.4a a predictor and the new iterate (x k+, y k+ updated by (.4b a corrector. To see the generality of (.4, when α = β =, the algorithmic framework (.4 reduces to the generalized PDHG (.2 in [4]; when τ = and α = β (0, 2, the accelerated version of PDHG from the PPA revisit (Algorithm 4 in [5] is recovered. Moreover, if α =, we obtain a symmetric generalized PDHG (see (5.9 that updates both the variables twice at each iteration and even a completely symmetric version with the golden-ratio step sizes (see (5.2. Therefore, based on this algorithmic framework, some existing PDHG schemes can be recovered and a class of new generalized PDHG schemes can be easily proposed by choosing different values of α, β and τ. We will focus on the case where τ [0, ] and investigate the restriction onto these constants so that the convergence of the algorithmic framework can be ensured. The analysis is conducted in a unified manner. In addition, the mentioned asymmetry features of the PDHG scheme (.2 may also be the reason why some critical theoretical issues remain challenging. For example, the worst-case convergence rate measured by the iteration complexity in a nonergodic sense is still open, even when θ or θ 2 is strongly convex, as analyzed in [4]. We will show that for the generalized PDHG algorithmic framework (.4, the worst-case convergence rate in both the ergodic and nonergodic senses can be derived when the parameters are appropriately restricted. This is a theoretical advantage of the algorithmic framework (.4. Finally, we would mention that there are a rich set of literature discussing how to extend the PDHG with more sophisticated analysis to more complicated saddle point models or to more abstract spaces, e.g., [5, 6, 7]. But in this paper we concentrate on the canonical saddle point model (. in a finite dimensional space and the representative PDHG scheme (.2 to present our idea of algorithmic design clearly. The rest of this paper is organized as follows. In Section 2, we summarize some understandings on the model (. and the algorithmic framework (.4 from the variational inequality perspective. Then, we prove the convergence of (.4 in Section 3 and derive its worst-case convergence rate measured by the iteration complexity in Section 4. In Section 5, we elaborate on the conditions that can ensure the convergence of (.4; some specific PDHG schemes are 3

thus yielded based on the algorithmic framework (.4. Finally, we draw some conclusions in Section 6. 2 Variational Inequality Understanding In this section, we provide the variational inequality (VI reformulation of the model (. and characterize the algorithmic framework (.4 via a VI. These VI understandings are the basis of our analysis to be conducted. 2. Variational Inequality Reformulation of (. We first show that the saddle point problem (. can be written as a VI problem. More specifically, if (x, y X Y is a solution point of the saddle point problem (., then we have Φ y Y (x, y Φ(x, y Φ x X (x, y. (2. Obviously, the second inequality in (2. implies that x X, θ (x θ (x + (x x T ( A T y 0, x X ; and the first one in (2. implies that y Y, θ 2 (y θ 2 (y + (y y T (Ax b 0, y Y. Therefore, finding a solution point (x, y of (. is equivalent to solving the VI problem: Find u = (x, y such that VI(Ω, F, θ u Ω, θ(u θ(u + (u u T F (u 0, u Ω, (2.2a where u = ( x y ( A, θ(u = θ (x + θ 2 (y, F (u = T y Ax b and Ω = X Y. (2.2b We denote by Ω the set of all solution points of VI(Ω, F, θ (2.2. Notice that Ω is convex (see Theorem 2.3.5 in [0] or Theorem 2. in [6]. Clearly, for the mapping F given in (2.2b, we have (u v T (F (u F (v = 0, u, v Ω. (2.3 Moreover, under the condition (.3, the matrix ( rin A T A si m (2.4 is positive definite. The positive definiteness of this matrix plays a significant role in analyzing the convergence of the PDHG scheme (.2, see e.g. [4, 9, 5, 27]. 4

2.2 Variational Inequality Characterization of (.4 Then, we rewrite the algorithmic framework (.4 also in the VI form. Note that the PDHG subroutine (.4a can be written as x k = arg min { θ (x (y k T (Ax b + r 2 x xk 2 x X }, (2.5a ỹ k = arg max { θ 2 (y + y T (A x k b + s 2 y yk 2 } y Y. Thus, the optimality conditions of (2.5a and (2.5b are and x k X, θ (x θ ( x k + (x x k T ( A T y k + r( x k x k 0, x X, ỹ k Y, θ 2 (y θ 2 (ỹ k + (y ỹ k T ( (A x k b + s(ỹ k y k 0, y Y, respectively. Recall x k = x k + τ( x k x k. We get the following VI containing only (x k, y k and ( x k, ỹ k : (2.5b ( x k, ỹ k X Y, (θ (x θ ( x k + (θ 2 (y θ 2 (ỹ k ( x x k T {( AT ỹ + k ( r( x y ỹ k A x k + k x k + A T (ỹ k y k } b τa( x k x k + s(ỹ k y k 0, (x, y X Y. (2.6 Using the notations in (2.2b, the PDHG subroutine (.4a can be written compactly as follows. PDHG subroutine Step (Prediction: ũ k Ω, θ(u θ(ũ k + (u ũ k T {F (ũ k + Q(ũ k u k } 0, u Ω, (2.7a where ( rin A Q = T τa si m (2.7b Accordingly, the correction step (.4b can be rewritten as the following compact form. Correction Step where u k+ = u k M(u k ũ k, ( αin 0 M = 0 βi m (2.8a. (2.8b 3 Convergence In this section, we investigate the convergence of the algorithmic framework (.4. Recall that (.4 can be rewritten as (2.7-(2.8. We prove the convergence of (.4 under the condition (Convergence Condition { H := QM 0, (3.a G := Q T + Q M T HM 0, (3.b where the matrices Q and M are given in (2.7b and (2.8b, respectively. Thus, it suffices to check the positive definiteness of two matrices to ensure the convergence of the algorithmic 5

framework (.4. The condition (3. also implies some specific restrictions onto the involved parameters in (.4 to be discussed in Section 5. Because of the VI reformulation (2.2 of the model (., we are interested in using the VI characterization (2.7-(2.8 to discern how accurate an iterate generated by the algorithmic framework (.4 is to a solution point of VI(Ω, F, θ. This analysis is summarized in the following theorem. Theorem 3.. Let {u k } be the sequence generated by the algorithmic framework (.4. Under the condition (3., we have θ(u θ(ũ k + (u ũ k T F (ũ k 2( u u k+ 2 H u u k 2 H + 2 uk ũ k 2 G, u Ω. (3.2 Proof. It follows from (3.a that Q = HM. Recall that M(u k ũ k = (u k u k+ in (2.8a. Thus, (2.7a can be written as Applying the identity θ(u θ(ũ k + (u ũ k T F (ũ k (u ũ k T H(u k u k+, u Ω. (3.3 (a b T H(c d = 2 { a d 2 H a c 2 H} + 2 { c b 2 H d b 2 H} to the right-hand side of (3.3 with we obtain a = u, b = ũ k, c = u k, and d = u k+, (u ũ k T H(u k u k+ = u u 2( k+ 2 H u u k 2 H + 2 ( uk ũ k 2 H u k+ ũ k 2 H. (3.4 For the last term of the right-hand side of (3.4, we have u k ũ k 2 H u k+ ũ k 2 H = u k ũ k 2 H (u k ũ k (u k u k+ 2 H (3.a = u k ũ k 2 H (u k ũ k M(u k ũ k 2 H = 2(u k ũ k T HM(u k ũ k (u k ũ k T M T HM(u k ũ k = (u k ũ k T (Q T + Q M T HM(u k ũ k (3.b = u k ũ k 2 G. (3.5 Substituting (3.4 and (3.5 into (3.3, we obtain the assertion (3.2. With the assertion in Theorem 3., we can show that the sequence {u k } generated by the algorithmic framework (.4 is strictly contractive with respect to the solution set Ω. We prove this property in the following theorem. Theorem 3.2. Let {u k } be the sequence generated by the algorithmic framework (.4. Under the condition (3., we have u k+ u 2 H u k u 2 H u k ũ k 2 G, u Ω. (3.6 6

Proof. Setting u = u in (3.2, we get u k u 2 H u k+ u 2 H u k ũ k 2 G + 2{θ(ũ k θ(u + (ũ k u T F (ũ k }. (3.7 Then, using (2.3 and the optimality of u, we have and thus θ(ũ k θ(u + (ũ k u T F (ũ k = θ(ũ k θ(u + (ũ k u T F (u 0 u k u 2 H u k+ u 2 H u k ũ k 2 G. (3.8 The assertion (3.6 follows directly. Theorem 3.2 shows that the sequence {u k } is Fèjer monotone and the convergence of {u k } to a u Ω in H-norm is immediately implied, see, e.g., [2]. 4 Convergence Rate In this section, we establish the worst-case O(/t convergence rate measured by the iteration complexity for the algorithmic framework (.4, where t is the iteration counter. We discuss the convergence rates in both the ergodic and nonergodic senses. 4. Convergence Rate in the Ergodic Sense First, let us characterize an approximate solution of VI(Ω, F, θ (2.2. Because of (2.3, we have (u u T F (u = (u u T F (u. Thus, (2.2a can be rewritten as u Ω, θ(u θ(u + (u u T F (u 0, u Ω. (4. For given ϵ > 0, ũ Ω is called an ϵ-approximate solution of VI(Ω, F, θ if it satisfies where θ(u θ(ũ + (u ũ T F (u ϵ, u D (ũ, (4.2 D (ũ := {u Ω u ũ }. We refer the reader to [8] (see (2.5 therein for the rational of defining an approximate solution onto the set D (ũ. In the following we show that for given ϵ > 0, based on t iterations of the algorithmic framework (.4, we can find ũ Ω such that ũ Ω and sup u D (ũ { θ(ũ θ(u + (ũ u T F (u } ϵ, (4.3 i.e., an ϵ-approximate solution of VI(Ω, F, θ. Theorem 3. will be used in the analysis; similar techniques can be found in [6]. Theorem 4.. Let {u k } be the sequence generated by the algorithmic framework (.4 under the conditions (.3 and (3.. For any integer t > 0, let ũ t be defined as Then we have ũ t Ω, ũ t = t + t ũ k. (4.4 k=0 θ(ũ t θ(u + (ũ t u T F (u 2(t + u u0 2 H, u Ω. (4.5 7

Proof. First, we know ũ k Ω for all k 0. Because of the convexity of X and Y, it follows from (4.4 that ũ t Ω. Using (2.3 and (3.2, we have θ(u θ(ũ k + (u ũ k T F (u + 2 u uk 2 H 2 u uk+ 2 H, u Ω. (4.6 Summarizing the inequality (4.6 over k = 0,,..., t, we obtain (t + θ(u t k=0 ( θ(ũ k + (t + u Since ũ t is defined in (4.4, we have t + t k=0 t θ(ũ k θ(u + (ũ t u T F (u k=0 Further, because of the convexity of θ(u, we get θ(ũ t t + ũ k T F (u + 2 u u0 2 H 0, u Ω. 2(t + u u0 2 H, u Ω. (4.7 t θ(ũ k. Substituting it into (4.7, the assertion of this theorem follows directly. The conclusion (4.5 indicates that ũ t defined in (4.4 is an approximate solution of VI(Ω, F, θ with an accuracy of O(/t. Since ũ t is given by the average of t iterations of (.4, the worst-case O(/t convergence rate in the ergodic sense is established for the the algorithmic framework (.4. Finally, we would remark that the condition of G 0 in (3. can be relaxed to G 0 for deriving the convergence rate result in this subsection. k=0 4.2 Convergence Rate in a Nonergodic Sense In this subsection, we derive the worst-case O(/t convergence rate in a nonergodic sense for the algorithmic framework (.4. Note that a nonergodic worst-case convergence rate is generally stronger than its ergodic counterparts. As mentioned, it seems still open whether or not the PDHG scheme (.2 has any nonergodic worst-case convergence rate. We first prove a lemma. Lemma 4.2. Let {u k } be the sequence generated by the algorithmic framework (.4. Under the condition (3., we have (u k ũ k T M T HM{(u k ũ k (u k+ ũ k+ } 2 (uk ũ k (u k+ ũ k+ 2 (Q T +Q. (4.8 Proof. Setting u = ũ k+ in (2.7a, we get θ(ũ k+ θ(ũ k + (ũ k+ ũ k T F (ũ k (ũ k+ ũ k T Q(u k ũ k. (4.9 Note that (2.7a is also true for k := k +. Thus, it holds that θ(u θ(ũ k+ + (u ũ k+ T F (ũ k+ (u ũ k+ T Q(u k+ ũ k+, u Ω. Then, setting u = ũ k in the above inequality, we obtain θ(ũ k θ(ũ k+ + (ũ k ũ k+ T F (ũ k+ (ũ k ũ k+ Q(u k+ ũ k+. (4.0 8

Combining (4.9 and (4.0, and using the monotonicity of F, we have Adding the term (ũ k ũ k+ T Q{(u k ũ k (u k+ ũ k+ } 0. (4. {(u k ũ k (u k+ ũ k+ } T Q{(u k ũ k (u k+ ũ k+ } to both sides of (4., and using v T Qv = 2 vt (Q T + Qv, we obtain (u k u k+ T Q{(u k ũ k (u k+ ũ k+ } 2 (uk ũ k (u k+ ũ k+ 2 (Q T +Q. Substituting (u k u k+ = M(u k ũ k into the left-hand side of the last inequality and using Q = HM, we obtain (4.8 and the lemma is proved. Now, we establish the worst-case O(/t convergence rate in a nonergodic sense for the algorithmic framework (.4. Similar techniques can be found in [7]. Theorem 4.3. Let {u k } be the sequence generated by the algorithmic framework (.4. Under the condition (3., then we have where c 0 > 0 is a constant. M(u t ũ t 2 H (t + c 0 u 0 u 2 H, (4.2 Proof. First, setting a = M(u k ũ k and b = M(u k+ ũ k+ in the identity we obtain a 2 H b 2 H = 2a T H(a b a b 2 H, M(u k ũ k 2 H M(u k+ ũ k+ 2 H = 2(u k ũ k T M T HM[(u k ũ k (u k+ ũ k+ ] M[(u k ũ k (u k+ ũ k+ ] 2 H. Inserting (4.8 into the first term of the right-hand side of the last equality, we obtain M(u k ũ k 2 H M(u k+ ũ k+ 2 H (u k ũ k (u k+ ũ k+ 2 (Q T +Q M[(uk ũ k (u k+ ũ k+ ] 2 H = (u k ũ k (u k+ ũ k+ 2 G 0. The last inequality holds because the matrix (Q T + Q M T HM = G and G 0. We thus have M(u k+ ũ k+ H M(u k ũ k H, k > 0. (4.3 Recall that the matric M is positive definite. Thus, the sequence { M(u k ũ k 2 H } is monotonically non-increasing. Then, it follows from G 0 and Theorem 3.2 there is a constant c 0 > 0 such that u k+ u 2 H u k u 2 H c 0 M(u k ũ k 2 H, u Ω. (4.4 Furthermore, it follows from (4.4 that c 0 M(u k ũ k 2 H u 0 u 2 H, u Ω. (4.5 k=0 9

Therefore, we have (t + M(u t ũ t 2 H t M(u k ũ k 2 H. (4.6 k=0 The assertion (4.2 follows from (4.5 and (4.6 immediately. Let d := inf{ u 0 u H u Ω }. Then, for any given ϵ > 0, Theorem 4.3 shows that the algorithmic framework (.4 needs at most d 2 /c 0 ϵ iterations to ensure that M(u k ũ k 2 H ϵ. Recall that u k is a solution of VI(Ω, F, θ if M(u k ũ k 2 H = 0 (see (2.7a and due to Q = HM. A worst-case O(/t convergence rate in a nonergodic sense is thus established for the algorithmic framework (.4. 5 How to Ensure the Condition (3. In previous sections, we have analyzed the convergence for the algorithmic framework (.4 under the condition (3.. Now, we discuss how to appropriately choose the parameters τ, α and β to ensure the condition (3.. Different choices of these parameters also specify the algorithmic framework (.4 as a class of specific PDHG schemes for the saddle point model (.. 5. General Study Recall the definitions of Q and M in (2.7b and (2.8b, respectively. Let us take a closer look at the matrices H and G defined in (3.. First, we have ( H = QM rin A = T τa si m ( α I n 0 0 β I m ( r α = I n τ α A β AT β si m. Thus, we set β = α τ to ensure that H is symmetric. With this restriction, we have ( r α H = I ( τ n α AT τ α si = rin τa T. m α τa τsi m τ α A Then, it is clear that under the condition (.3, we can choose τ (0, ] to ensure the positive definiteness of H. Overall, to ensure the positive definiteness and symmetry of H, we pose the restriction: τ (0, ] and β = α τ. (5. Second, let us deal with the matric G in (3.. Since HM = Q, we have Q T + Q M T HM = Q T + Q M T Q = ( ( ( 2rIn ( + τa T αin 0 rin A T α ( + τa 2sI m 0 τ I m τa si m = ( ( 2rIn ( + τa T αrin αa T ( + τa 2sI m αa α τ si m = ( (2 αrin ( + τ αa T ( + τ αa (2 α τ si. m (5.2 0

Therefore, we need to ensure ( (2 αrin ( + τ αa T G = ( + τ αa (2 α τ si m 0. (5.3 We present a strategy to ensure (5.3 in the following theorem. Theorem 5.. For given τ (0, ], under the condition (.3, the assertion (5.3 holds when τ = and α (0, 2, (5.4 or τ (0,, and α ( + τ τ. (5.5 Proof. First, if τ =, it follows from (5.2 that ( rin A T G = (2 α A si m. Under the condition (.3, we have G 0 for all α (0, 2. Now, we consider τ (0,. For G 0, we need to ensure (2 α > 0, (2 α τ > 0 and (2 α(2 α τ rs > ( + τ α2 A T A. Thus, under the condition (.3, we need to guarantee (2 α > 0, (2 α > 0, (5.6a τ and (2 α(2 α τ ( + τ α2. (5.6b By a manipulation, the inequality (5.6b is equivalent to ( τ α 2 + 2 (( + τ ( + τ α + ( 4 ( + τ 2 0, and thus ( τ α 2 + 2 τ τ τ Multiplying the positive factor, we get τ Because the small root of the equation ( τ 2 α + (3 + τ( τ 0. α 2 2( + τα + τ(3 + τ 0. α 2 2( + τα + τ(3 + τ = 0 is ( + τ τ, the condition (5.6b is satisfied when (5.5 holds. In the following we show that (5.6a is satisfied when (5.5 holds. It follows from (5.5 directly α < 2 for all τ (0,. In addition, because ( τ τ < 0, τ (0,,

we have ( + τ τ < 2τ, τ (0,. Consequently, using (5.5, we get α ( + τ τ τ τ < 2, τ (0,. The proof is complete. In the following theorem, we summarize the restriction onto the involved parameters that can ensure the convergence of the algorithmic framework (.4. Theorem 5.2. Under the condition (.3, if the parameters τ, α and β in (.4 are chosen such that either τ =, and α = β (0, 2, (5.7a or τ (0,, α + τ t and β = α τ, (5.7b then the matrices Q and M defined respectively in (2.7b and (2.8b satisfy the convergence condition (3.. Thus, the sequence generated by the algorithmic framework (.4 is convergent to a solution point of VI(Ω, F, θ. 5.2 Special Cases In this subsection, we specify the general restriction posed in Theorem 5.2 on the parameters of the algorithmic framework (.4 and discuss some special cases of this framework. 5.2. Case I: τ = According to (5.7a, we can take α = β (0, 2 and thus the algorithmic framework (.4 can be written as x k = arg min{φ(x, y k + r 2 x xk 2 x X }, x k = 2 x k x k, (5.8a ỹ k = arg max{φ( x k, y s 2 y yk 2 y Y}. ( ( x k+ x k = y k+ y k ( x α k x k y k ỹ k. (5.8b This is exactly Algorithm 4 in [5] which can be regarded as an accelerated version of (.2 using the technique in [3]. Clearly, choosing α = β = in (5.8 recovers the special case of (.2 with τ =. 5.2.2 Case II: α = Now let α = be fixed. In this case, the algorithmic framework (.4 can be written as x k+ = arg min{φ(x, y k + r 2 x xk 2 x X }, x k = x k+ + τ(x k+ x k, ỹ k = arg max{φ( x k, y s 2 y yk 2 y Y}, y k+ = y k + β(ỹ k y k, (5.9 2

which is symmetric in the sense that it updates both the variables x and y twice at each iteration. For this special case, the restriction in Theorem 5.2 can be further specified. We present it in the following theorem. Theorem 5.3. If α = ; the parameters β and τ satisfy [ 5 ] τ,, and β = 2 τ, (5.0 then the conditions (5.7a or (5.7b is satisfied and thus the PDHG scheme (5.9 is convergent. Proof. If α = and τ =, it follows from (5.0 that the condition (5.7a is satisfied. Now we consider α = and τ (0,. In this case the condition (5.7b becomes τ (0, and + τ τ, and thus τ (0, and τ 2 + τ 0. The above condition is satisfied for all τ [ 5 2,. The proof is complete. Therefore, with the restriction in (5.0, the PDHG scheme (5.9 can be written as where τ x k+ = arg min{φ(x, y k + r 2 x xk 2 x X }, x k = x k+ + τ(x k+ x k, ỹ k = arg max{φ( x k, y s 2 y yk 2 y Y}, y k+ = y k + τ (ỹk y k, [ 5 ] 2,. Indeed, we can further consider choosing (5. Then, we have and it follows from (5. that 5 τ 0 =. 2 τ 0 = + τ 0, y k + τ 0 (ỹ k y k = y k + ( + τ 0 (ỹ k y k = ỹ k + τ 0 (ỹ k y k. Thus, the scheme (5. can be further specified as x k+ = arg min{φ(x, y k + r 2 x xk 2 x X }, x k = x k+ + τ 0 (x k+ x k, ỹ k = arg max{φ( x k, y s 2 y yk 2 y Y}, y k+ = ỹ k + τ 0 (ỹ k y k, (5.2 where τ 0 = 5 2. It is a completely symmetric PDHG scheme in the sense that both the variables are updated twice at each iteration and the additional update steps both use the golden-ratio step sizes. 3

5.2.3 Case III: β = We can also fix β =. In this case, the algorithmic framework (.4 reduces to x k = arg min{φ(x, y k + r 2 x xk 2 x X }, x k = x k + τ( x k x k, y k+ = arg max{φ( x k, y s 2 y yk 2 y Y}. (5.3a x k+ = x k α(x k x k. (5.3b For this case, the restriction on the involved parameters in Theorem 5.2 can be further specified as the following theorem. Theorem 5.4. If β = ; the parameters α and τ satisfy then the convergence conditions (5.3 are satisfied. α = τ (0, ], (5.4 Proof. First, in this case, the condition (5. is satisfied. The matrix G in (5.3 becomes ( (2 αrin A T G =, A si m which is positive definite for all α (0, ] under the assumption (.3. Therefore, considering the restriction in (5.4, we can specify the scheme (5.3 as x k = arg min{φ(x, y k + r 2 x xk 2 x X }, x k = x k + τ( x k x k, (5.5a y k+ = arg max{φ( x k, y s 2 y yk 2 y Y}. where τ ( 0, ]. 6 Conclusions x k+ = x k τ(x k x k. (5.5b We propose an algorithmic framework of generalized primal-dual hybrid gradient (PDHG methods for saddle point problems and study its convergence. This algorithmic framework includes some existing PDHG schemes as special cases, and it can yield a class of new generalized PDHG schemes. It also possesses some theoretical advantages such as the worst-case convergence rate measured by the iteration complexity in a nonergodic sense. Our analysis provides a unified perspective to the study of some PDHG schemes for saddle point problems. It is interesting to know if our analysis could be extended to some special nonconvex cases with strong application backgrounds in image processing such as some semiconvex variational image restoration models. We leave it as a further research topic. 4

References [] K. J. Arrow, L. Hurwicz and H. Uzawa, Studies in linear and non-linear programming, With contributions by H. B. Chenery, S. M. Johnson, S. Karlin, T. Marschak, R. M. Solow. Stanford Mathematical Studies in the Social Science, Vol. II. Stanford University Press, Stanford, Calif., 958. [2] E. Blum and W. Oettli, Mathematische Optimierung Grundlagen und Verfahren. Ökonometrie und Unternehmensforschung, Springer-Verlag, Berlin-Heidelberg-New York, 975. [3] S. Bonettini and V. Ruggiero, On the convergence of primal-dual hybrid gradient algorithms for total variation image restoration, J. Math. Imaging Vision, 44 (202, pp. 236-253. [4] A. Chambolle and T. Pock, A first-order primal-dual algorithms for convex problem with applications to imaging, J. Math. Imaging Vision, 40 (20, pp. 20-45. [5] A. Chambolle and T. Pock, On the ergodic convergence rates of a first-order primal-dual algorithm, Math. Program., to appear. [6] A. Chambolle and T. Pock, A remark on accelerated block coordinate descent for computing the proximity operators of a sum of convex functions, SMAI J. Comput. Math., to appear. [7] L. Condat, A primaldual splitting method for convex optimization involving Lipschitzian, proximable and linear composite terms, J. Optim. Theory Appli., 58 (203, pp. 460-479. [8] J. Douglas and H. H. Rachford, On the numerical solution of the heat conduction problem in 2 and 3 space variables, Trans. Amer. Math. Soc., 82(956, pp. 42-439. [9] E. Esser, X. Zhang and T. F. Chan, A general framework for a class of first order primaldual algorithms for TV minimization, SIAM J. Imaging Sci., 3 (200, pp. 05-046. [0] F. Facchinei and J. S. Pang, Finite-Dimensional Variational Inequalities and Complementarity Problems, Vol. I. Springer Series in Operations Research, Springer Verlag, New York, 2003. [] R. Glowinski and A. Marrocco, Approximation par éléments finis d ordre un et résolution par pénalisation-dualité d une classe de problèmes non linéaires, R.A.I.R.O., R2, 975, pp. 4-76. [2] T. Goldstein, M. Li, X. M. Yuan, E. Esser and R. Baraniuk, Adaptive primal-dual hybrid gradient methods for saddle-point problems, NIPS, 205. [3] E. G. Gol shtein and N. V. Tret yakov, Modified Lagrangian in convex programming and their generalizations, Math. Program. Study, 0(979, pp. 86-97. [4] B. S. He, Y. F. You and X. M. Yuan, On the convergence of primal-dual hybrid gradient algorithm, SIAM J. Imag. Sci., 7 (204, pp. 2526-2537. [5] B. S. He and X. M. Yuan, Convergence analysis of primal-dual algorithms for a saddlepoint problem: From contraction perspective, SIAM J. Imag. Sci., 5 (202, pp. 9-49. [6] B. S. He and X. M. Yuan, On the O(/n convergence rate of Douglas-Rachford alternating direction method, SIAM J. Numer. Anal., 50 (202, pp. 700-709. 5

[7] B. S. He and X.M. Yuan, On non-ergodic convergence rate of Douglas-Rachford alternating directions method of multipliers, Numerische Mathematik, 30 (205, pp. 567-577. [8] Y. E. Nesterov, Gradient methods for minimizing composite objective function, Math. Prog., Ser. B, 40 (203, pp. 25-6. [9] G. M. Korpelevich, The extragradient method for finding saddle points and other problems, Ekonomika i Matematchskie Metody, 2 (976, pp. 747-756. [20] P. L. Lions, B. Mercier, splitting algorithms for the sum of two nonlinear operators, SIAM J. Numer. Anal., 6(979, pp. 964-979. [2] B. Martinet, Regularization d inequations variationelles par approximations successives, Revue Francaise d Informatique et de Recherche Opérationelle, 4(970, pp. 54-59. [22] T. Pock and A. Chambolle, Diagonal preconditioning for first order primal-dual algorithms in convex optimization, IEEE International Conference on Computer Vision, 20, pp. 762-769. [23] L. D. Popov, A modification of the Arrow-Hurwitz method of search for saddle points, Mat. Zametki, 28(5 (980, pp. 777-784. [24] L. Rudin, S. Osher and E. Fatemi, Nonlinear total variation based noise removal algorithms, Physica D, 60 (992, pp. 227-238. [25] R. Shefi, Rate of convergence analysis for convex nonsmooth optimization algorithms, PhD Thesis, Tel Aviv University, Israel, 205. [26] P. Weiss, L. Blanc-Feraud and G. Aubert, Efficient schemes for total variation minimization under constraints in image processing, SIAM J. Sci. Comput., 3(3 (2009, pp. 2047-2080. [27] X. Zhang, M. Burger and S. Osher, A unified primal-dual algorithm framework based on Bregman iteration, J. Sci. Comput., 46 ( (200, pp. 20-46. [28] M. Zhu and T. F. Chan, An efficient primal-dual hybrid gradient algorithm for total variation image restoration, CAM Report 08-34, UCLA, Los Angeles, CA, 2008. 6