A Proximal Alternating Direction Method for Semi-Definite Rank Minimization (Supplementary Material)

Size: px
Start display at page:

Download "A Proximal Alternating Direction Method for Semi-Definite Rank Minimization (Supplementary Material)"

Transcription

1 A Proximal Alternating Direction Method for Semi-Definite Rank Minimization (Supplementary Material) Ganzhao Yuan and Bernard Ghanem King Abdullah University of Science and Technology (KAUST), Saudi Arabia The supplementary material is organized as follows. Section 1 presents the details of our proofs. Section 2 presents the convergence analysis of the proposed MPEC-based proximal ADM algorithm. Section 3 presents the classical ADM for solving the subproblem. Finally, Section presents additional experimental results. 1 Proofs Lemma 1. [Semidefinite Embedding Lemma (Fazel, Hindi, and Boyd 23)] Let R R m n be a given matrix. Then (R) r if and only if there exist matrices S = S T R m m and T = T T R n n such that [ ] S R R T, (S) + (T) 2r. (1) T Remark. We remark that the semidefinite optimization problem discussed in this paper is very general. To illustrate this point, we consider the following optimization problem: min g(r) + λ (R) (2) R By the Semidefinite Embedding Lemma, we have the following equivalent semi-definite optimization problem : min g(r) + 1 ([ ]) [ ] S S R R 2, s.t. T R T. T Since [ ] S R R T T is positive semidefinite, by the Schur complement condition 1, it holds that S and T. Then the variable matrix [ ] S T T is also positive semidefinite. Using the MPEC reformulation, we have the following equivalent optimization problem: min g(r) + 1 R,S,T,V 2 tr(v), [ ] [ ] S S R s.t. I V, =, T R T, I V. T Clearly, the equivalent optimization problem in Eq (2) can be solved by our proposed PADM algorithm. Copyright c 2, Association for the Advancement of Artificial Intelligence ( All rights reserved. 1 Schur complement condition implies that: X ( A B B T C ) A, S = C B T A B Here, we prove the variational formulation of the NP-hard function. Lemma 2. For any given positive semidefinite matrix X R n n, it holds that (X) = min tr(i V), s.t. V, X =, V I (3) V and the unique optimal solution of the minimization problem in Eq (3) is given by V = Udiag(sign(σ))U T, where X = Udiag(σ)U T is the eigenvalue decomposition of X, sign is the standard signum function 2. Proof. First of all, we let X and V be arbitrary feasible solutions with their eigenvalue decomposition given by X = Udiag(σ)U T and V = Sdiag(δ)S T, respectively. It always holds that: X, V = Udiag( σ)diag( δ)s T 2 F The second equality is achieved only when V and X are simultaneously unitarily diagonalizable, i.e. both V and X share the same spectral decomposition. Therefore, the feasible set for V in Eq (3) must be contained in the set {V V = Udiag(v)U T, v 1, v R r }. Using the fact that tr(v) = v, 1, tr(x) = σ, 1 and V, X = Udiag(v)U T, Udiag(σ)U T = v, σ, Eq (3) reduces to the vector case: (X) = min v 1, 1 v, s.t. v, σ =, v 1 () Since v and σ, Eq () is equivalent to Eq (): (X) = min v 1, 1 v, s.t. v σ =, v 1 () where denotes the Hadamard product (also known as the entrywise product). Note that for all i [r] when σ i =, v i = 1 will be achieved by minimization, when σ i, v i = will be enforced by the constraint. Since the objective function in Eq () is linear, minimization is always achieved at the boundaries of the feasible solution space. Thus, v = 1 sign(σ). Finally, we have V = Udiag(1 sign(σ))u T. We thus complete the proof of this lemma. 1, x < ; 2 sign(x) =, x = ; 1, x >.

2 The following lemma shows how to compute the generalized Singular Value Thresholding (SVT) operator which is involved in V-subproblem in our PADM Algorithm. Lemma 3. Assume that W has the SVD decomposition that W = Udiag(σ)U T. The optimal solution of the following optimization problem arg min V 1 2 V W 2 F + I (V) () can be computed as Udiag(min(1, max(, σ)))t. Here I is an indicator function{ of the convex set {V, V V I} with I (V)., otherwise. Proof. The proof of this lemma is very natural. For completeness, we present our proof here. For notation convenience, we use σ and z to denote the singular values of W and V, respectively. We naturally derive the following inequalities: 1 2 V W 2 F + I (V) = 1 2 ( z 2 + σ 2 2 V, W ) + I Θ (z) 1 2 ( z 2 + σ 2 2 z, σ ) + I Θ (z) = 1 2 z σ 2 + I Θ (z) From the von Veumann s trace inequality, the solution set of () must be contained in the set {V V = Udiag(v )U T }, where v is given by v 1 = arg min z 2 z σ 2 + I Θ (z) (7) Since the optimization problem in Eq(7) is decomposable, a simple computation yields that the solution can be computed in closed form as: v = min(1, max(, σ)). Therefore, V = Udiag(min(1, max(, σ)))u T. We thus complete the proof of this lemma. 2 Convergence Analysis The global convergence of ADM for convex problems was given by He and Yuan in (He and Yuan 2) under an elegant variational inequality framework. However, since our MPEC optimization problem is non-convex, the convergence analysis for ADM needs additional conditions. In non-convex optimization, convergence to a stationary point (local minimum) is the best convergence property that we can hope for. Under boundedness condition, we show that the sequence generated by the proximal ADM converges to a KKT point. For the ease of discussions, we define: and u {X, V}, s {X, V, π} () Ω {X X κi}, {V V I} (9) First of all, we present the first-order KKT conditions of the MPEC reformulation optimization problem. Based on the augmented Lagrangian function of the MPEC reformulation, we naturally derive the following KKT conditions of the optimization problem for {X, V, π }: I Ω (X ) + A T g(a(x ) b) + πv I (V ) λi + πx = V, X whose existence can be guaranteed by Robinson s constraint qualification (Rockafellar, Wets, and Wets 199). First of all, we prove the subgradient lower bound for the iterates gap by the following lemma. Lemma. Assume that π k is bounded for all k, then there exists a constant ϖ > such that the following inequality holds: L(s k+1 ) ϖ s k+1 s k (1) Proof. By the optimal condition of the X-subproblem and V-subproblem, we have: D(X k+1 X k ) + A T g(a(x k+1 ) b) + π k V k + α V k, X k+1 V k + I Ω (X k+1 ) (11) E(V k+1 V k ) λi + π k X k+1 +α V k+1, X k+1 X k+1 + I (V k+1 ) () By the definition of L we have that L X (s k+1 ) = A T g(ax k+1 b) + π k+1 V k+1 + α V k+1, X k+1 V k+1 + I Ω (X k+1 ) = (π k + α V k, X k+1 )V k + (π k+1 + α V k+1, X k+1 )V k+1 D(X k+1 X k ) = (π k + α V k+1, X k+1 + α V k V k+1, X k+1 )V k +(π k+1 + α V k+1, X k+1 )V k+1 D(X k+1 X k ) = (π k+1 + α V k V k+1, X k+1 )V k + (2π k+1 π k )V k+1 D(X k+1 X k ) = (α V k V k+1, X k+1 )V k + (π k+1 π k )V k+1 +π k+1 (V k+1 V k ) D(X k+1 X k ) The first step uses the definition of L X (s k+1 ), the second step uses Eq (11), the third step uses V k + V k+1 V k+1 = V k, the fourth step uses the multiplier update rule for π. Assume that π k+1 is bounded by a constant ρ that π k+1 ρ. We have: L X (s k+1 ) F α V k V k+1, X k+1 V k + (π k+1 π k )V k+1 + π k+1 (V k+1 V k ) + D(X k+1 X k ) 2nκα V k V k+1 F + 2 π k+1 π k ρ V k V k+1 F + D X k+1 X k F (13)

3 Similarly, we have L V (s k+1 ) = I (V k+1 ) λi + π k+1 X k+1 +α V k+1, X k+1 X k+1 = (π k+1 π k )X k+1 E(V k+1 V k ) L π (s k+1 ) = I V k+1, X k+1 = 1 α (πk+1 π k ) Then we derive the following inequalities: L V (s k+1 ) F κ π k π k+1 + E V k+1 V k F () L π (s k+1 ) 1 α πk+1 π k () Combining Eqs (13-), we conclude that there exists ϖ > such that the following inequality holds L(s) F ϖ s s k. Thus, we complete the proof of this lemma. The following lemma is useful in our convergence analysis. Lemma. Assume that π k is bounded for all k, then we have the following inequality: s k s k+1 2 < + () k= In particular the sequence s k s k+1 is asymptotic regular, namely s k s k+1 as k. Moreover any cluster point of s k is a stationary point of L. Proof. Due to the initialization and the update rule of π, we conclude that π k is nonnegative and monotone non-decreasing. Moreover, as k, we have: X k+1, V k+1 =. This can be proved by contradiction. Suppose that X k+1, V k+1, then π k = + as k. This contradicts our assumption that π k is bounded. Therefore, we conclude that as k + it holds that X k+1, V k+1 =, π i+1 π i < + (17) π i+1 π i 2 < + On the other hand, we naturally derive the following inequalities: L(X k+1, V k+1 ; π k+1 ) = L(X k+1, V k+1 ; π k ) + π k+1 π k, V k+1, X k+1 = L(X k+1, V k+1 ; π k ) + 1 α πk+1 π k 2 L(X k, V k+1 ; π k ) 2 Xk+1 X k α πk+1 π k 2 L(X k, V k ; π k ) 2 Xk+1 X k 2 2 Vk+1 V k α πk+1 π k 2 (1) The first step uses the definition of L; the second step uses update rule of the Lagrangian multiplier π; the third and fourth step use the -strongly convexity of L with respect to X and v, respectively. We define C = L(X, V ; π ) L(X k+1, V k+1 ; π k+1 )+ 1 α k πi+1 π i 2. Clearly, by the boundedness of X k, V k and π k, both L(X k+1, V k+1 ; π k+1 ) and C are bounded. Summing Eq (1) over i = 1, 2..., k, we have: 2 u i+1 u i 2 C (19) Therefore, combining Eq(17) and Eq (19), we have + k=1 sk+1 s k 2 < + ; in particular s k+1 s k. By Eq (1), we have that: L(s k+1 ) ϖ s k+1 s k (2) which implies that any cluster point of s k is a stationary point of L. We complete the proof of this lemma. Remarks: Lemma states that any cluster point is the KKT point. Strictly speaking, this result does not imply the convergence of the algorithm. This is because the boundedness of k= sk s k+1 2 does not imply that the sequence s k is convergent 3. In what follows, we aim to prove stronger result in Theorem 1. Our analysis is mainly based on a recent non-convex analysis tool called Kurdyka-Łojasiewicz inequality (Attouch et al. 21; Bolte, Sabach, and Teboulle 2). One key condition of our proof requires that the Lagrangian function L(s) satisfies the so-call (KL) property in its effective domain. It is so-called the semi-algebraic function satisfy the Kurdyka- Łojasiewicz property. It is not hard to validate that the Lagrangian function L(s) is a semi-algebraic function. This is not surprising since semi-algebraic function is ubiquitous in applications. Interested readers can refer to (Xu and Yin 213) for more details. We now present the following proposition established in (Attouch et al. 21). Proposition 1. For a given semi-algebraic function L(s), for all s doml, there exists θ [, 1), η (, + ] a neighborhood S of s and a concave and continuous function ϕ(t) = c t 1 θ, t [, η) such that for all s S and satisfies L( s) (L(s), L(s) + η), the following inequality holds: (, L( s))ϕ k (L(s) L( s)) 1, s where (, L( s)) = min{ u : u L( s)}. 3 One typical counter-example is s k = k 1. Clearly, i k= sk s k+1 2 = k=1 ( 1 k )2 is bounded by π2 ; however, s k is divergent since s k = ln(k) + C e, where C e is the well-known Euler s constant. Note that semi-algebraic functions include (i) real polynomial functions, (ii) finite sums and products of semi-algebraic functions, and (iii) indicator functions of semi-algebraic sets. Using these definitions repeatedly, the graph of L(s) : {(s, t) t = L(s)} can be proved to be a semi-algebraic set. Therefore, L(s) is a semialgebraic function.

4 The following theorem establishes the convergence properties of the proposed algorithm under a boundedness condition. Theorem 1. Assume that π k is bounded for all k. Then we have the following inequality: + k= s k s k+1 < (21) Moreover, as k, Algorithm 1 converges to the first order KKT point of the MPEC reformulation optimization problem. Proof. For simplicity, we define R k = ϕ(l(s k ) L(s )) ϕ(l(s k+1 ) L(s )). We naturally derive the following inequalities: 2 uk+1 u k 2 1 α πk+1 π k 2 L(s k ) L(s k+1 ) = (L(s k ) L(s )) (L(s k+1 ) L(s )) R k ϕ k (L(s k ) L(s )) R k (, L(s k )) R k ϖ( X k X k 1 + V k V k 1 + π k π k 1 ) R k ϖ( 2 u k u k 1 + π k π k 1 ) = R k ϖ ( uk u k 1 + πk π k 1 ) (Rk ϖ) 2 + ( uk u k 1 + πk π k 1 ) 2 The first step uses Eq (1); the third step uses the concavity of ϕ such that ϕ(a) ϕ(b) ϕ k (a)(a b) for all a, b R; the fourth step uses the KL property such that (, L(s k ))ϕ k (L(s k ) L(s )) 1 as in Proposition 1; the fifth step uses Eq (1); the sixth step uses the inequality that x; y x + y 2 x; y, where ; in [ ] denotes the row-wise partitioning indicator as in Matlab; the last step uses that fact that 2ab a 2 + b 2 for all a, b R. As k, we have: 2 uk+1 u k 2 (Rk κ) 2 + ( ) 2 uk u k 1 Take the squared root of both side and use the inequality that a + b a + b for all a, b R+, we have uk+1 u k Rk κ + 32 uk u k 1 Rk κ + ( uk u k 1 u k+1 u k ) Summing the inequality above over i = 1, 2..., k, we have: u i+1 u i Z + ( u 1 u + u k+1 u k ) where Z = κ k (ϕ(l(s ) L(s )) ϕ(l(s k+1 ) L(s ))) is bounded real number. Therefore, we conclude that as k, we obtain: u i+1 u i < + () Moreover, by Eq (1) in Lemma we have L(s k+1 ) =. In other words, we have the following results: I (X k+1 ) + A T g(a(x k+1 ) b) + π k+1 V k+1 I (V k+1 ) + λi + π k+1 X k+1 = V k+1, X k+1 which imply that {X k+1, V k+1, π k+1 } is a first-order KKT point. Algorithm 2 Classical Alternating Direction Method for Solving the Convex X-Subproblem in Eq (). (S.) Initialize X, Y = R n n, y = R m. z = R m, Z = R n n. Set t =. (S.1) Solve the following (y, X)-subproblem: X t+1 = arg min J (X, y t, Y t ; Z t, z t ) (23) X (S.2) Solve the following Y-subproblem: (y t+1, Y t+1 ) = arg min J (X t+1, y, Y; Z t, z t ) (2) y, Y κi (S.3) Update the Lagrange multiplier via the following formula: Z t+1 = Z t + β(x t+1 Y t+1 ) z t+1 = z t + γ(a(x k+1 ) b y k+1 ) (S.) Set t := t + 1 and then go to Step (S.1). 3 Solving the Convex Subproblem The efficiency of Algorithm proximal ADM in Algorithm 1 relies whether the convex subproblem can be efficiently solved. In this section, we aim to solve the following semidefinite optimization subproblem involved in the proposed PADM algorithm: min g (A(X) b) + α X κi 2 B(X) 2 F + 2 X 2 F + X, C, () Our solution is naturally based on the classical ADM (He and Yuan 2; Lin, Liu, and Su 211). For completeness,

5 we include our algorithm details here. First, we introduce t- wo auxiliary vectors y R m and Y R n n to reformulate Eq () as: min g (y) + α y,x,y 2 B(X) 2 F + 2 X 2 F + X, C, s.t. A(X) b = y, X = Y, Y κi. (2) Let J β,γ : R m R n n R n n R m R n n R be the augmented Lagrangian function in Eq(2) J (y, X, Y; z, Z) = g (y) + α 2 B(X) 2 F + 2 X 2 F + X, C + z, A(X) b y + γ 2 A(X) b y 2 F + Z, X Y + β 2 X Y 2 F, s.t. Y κi z and Z are the Lagrange multipliers associated to the constraints A(Y) b y = and X Y =, respectively, and γ, β > are the penalty parameters. The detailed iteration steps of the classical ADM for Eq (2) are described in Algorithm 2. Next, we focus our attention on the solutions of subproblems (23) and (2) arising in Algorithm 2. (i) X-subproblem. The first-order optimality condition for the variable X is: αb B(X t+1 ) + γa A(X t+1 ) + ( + β)x t+1 = E where E = βy + γa (b + y) C A z Z. Solving this linear system gives: X t+1 = (αb B + γa A + ( + β)i) 1 E (27) When the dimension of the solution is high, solving this linear system may dominant the computation time. However, one can use iterative conjugate gradient to alleviate this computational burden. We remark that it is also possible to utilize linearized ADM to address this issue (He and Yuan 2; Lin, Liu, and Su 211). (ii) (y, Y)-subproblem. Variable y in Eq (2) is updated by solving the following problem: y t+1 = arg min g (y) + γ y R m 2 q y 2, with q = A(X) b+z/γ. It reduces to the Moreau proximal operator prox g ( ) that can be evaluated efficiently by our assumption in the introduction section. Variable Y in Eq (2) is updated by solving the following problem: Y t+1 β = arg min Y κi 2 Y S 2 F with S = X t+1 + Z t /β. Assume that S has the spectral decomposition that S = Vdiag(s)V T. A simple computation yields that the solution Y t+1 can be computed in closed form as: Y t+1 = Vdiag(max(, min(κ1, s)))v T. The exposition above shows that the computation required in each iteration of Algorithm 2 is insignificant. Classical ADM has excellent convergence both in theory and in practice for convex problems. The convergence of Algorithm 2 can be obtained since the Féjer monotonicity of iterative sequences {y t, X t, Y t, z t, Z t } holds due to convexity. For the proof of convergence of Algorithm 2, interested readers can refer to (He and Yuan 2) for more details. Additional Experimental Results In this section, we present some additional experimental results to demonstrate the superiority of our proposed proximal ADM algorithm. Due to page limitations, we were not able to add these results in the submission. We extend our method for sensor network localization problem in the presence of laplace noise and uniform noise. We show our results of laplace noise in Figure 1- and uniform noise in Figure 7-, these experimental results strengthen our conclusions drawn in our submission. References [Attouch et al. 21] Attouch, H.; Bolte, J.; Redont, P.; and Soubeyran, A. 21. Proximal alternating minimization and projection methods for nonconvex problems: An approach based on the kurdyka-lojasiewicz inequality. Mathematics of Operations Research (MOR) 3(2): [Bolte, Sabach, and Teboulle 2] Bolte, J.; Sabach, S.; and Teboulle, M. 2. Proximal alternating linearized minimization for nonconvex and nonsmooth problems. Mathematical Programming (1-2): [Fazel, Hindi, and Boyd 23] Fazel, M.; Hindi, H.; and Boyd, S. P. 23. Log-det heuristic for matrix minimization with applications to hankel and euclidean ance matrices. In Proceedings of the American Control Conference, volume 3, 2. 1 [He and Yuan 2] He, B., and Yuan, X. 2. On the O(1/n) convergence rate of the douglas-rachford alternating direction method. SIAM Journal on Numerical Analysis (SINUM) (2): ,, [Lin, Liu, and Su 211] Lin, Z.; Liu, R.; and Su, Z Linearized alternating direction method with adaptive penalty for low- representation. In NIPS, 2., [Rockafellar, Wets, and Wets 199] Rockafellar, R. T.; Wets, R. J.-B.; and Wets, M Variational analysis. springe, berlin, heidelberg, new york, [Xu and Yin 213] Xu, Y., and Yin, W A block coordinate descent method for regularized multiconvex optimization with applications to nonnegative tensor factorization and completion. SIAM Journal on Imaging Sciences (SIIMS) (3):

6 7.e1.3e1 3.7e1 2.e1 3.9e 3.e 2.1e 1.1e Iteration Figure 1: Asymptotic behavior on minimum- sensor network localization problem in the presence of laplace noise. We plot the values of (blue) and (red) against the number of iterations, as well as how the sensors have been located at different stages of the process (1, 2, 3,, ) (a) FM, =.2, = (b) TAM, =.21, = (c) LPAM, =.19, = (d) LDHM, =.1, = (e) PDA, =.27, = 11. Figure 2: Performance comparison on 2d data in the presence of laplace noise (f) PADM, =., = (a) FM, =.32, = (b) TAM, =.29, = (c) LPAM, =.21, = (d) LDHM, =.32, = (e) PDA, =., = 9. Figure 3: Performance comparison on 3d data in the presence of laplace noise (f) PADM, =.21, = (b) and comparisons on 2d data (c) and comparisons on 3d data (d) and comparisons on 7d data Figure : Performance comparison with varying the number of sensor u in the presence of laplace noise (a) and comparisons on 2d data (b) and comparisons on 3d data (c) and comparisons on 7d data Figure : Performance comparison with varying the s in the presence of laplace noise (a) and comparisons on 2d data (b) and comparisons on 3d data (c) and comparisons on 7d data Figure : Performance comparison with varying the r in the presence of laplace noise.

7 7.1e1.e1 3.7e1 2.e1 3.e 3.e 2.1e 1.3e Iteration Figure 7: Asymptotic behavior on minimum- sensor network localization problem in the presence of uniform noise. We plot the values of (blue) and (red) against the number of iterations, as well as how the sensors have been located at different stages of the process (1, 2, 3,, ) (a) FM, =.39, = (b) TAM, =.3, = (c) LPAM, =.3, = 2... (d) LDHM, =.32, = (e) PDA, =., =. Figure : Performance comparison on 2d data in the presence of uniform noise... (f) PADM, =.31, = (a) FM, =., = (b) TAM, =.7, = (c) LPAM, =.9, = (d) LDHM, =., = (e) PDA, =., = 31. Figure 9: Performance comparison on 3d data in the presence of uniform noise (f) PADM, =.3, = (b) and comparisons on 2d data (c) and comparisons on 3d data (d) and comparisons on 7d data Figure 1: Performance comparison with varying the number of sensor u in the presence of uniform noise (a) and comparisons on 2d data (b) and comparisons on 3d data (c) and comparisons on 7d data Figure 11: Performance comparison with varying the s in the presence of uniform noise (a) and comparisons on 2d data (b) and comparisons on 3d data (c) and comparisons on 7d data Figure : Performance comparison with varying the r in the presence of uniform noise.

Block Coordinate Descent for Regularized Multi-convex Optimization

Block Coordinate Descent for Regularized Multi-convex Optimization Block Coordinate Descent for Regularized Multi-convex Optimization Yangyang Xu and Wotao Yin CAAM Department, Rice University February 15, 2013 Multi-convex optimization Model definition Applications Outline

More information

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16 XVI - 1 Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16 A slightly changed ADMM for convex optimization with three separable operators Bingsheng He Department of

More information

I P IANO : I NERTIAL P ROXIMAL A LGORITHM FOR N ON -C ONVEX O PTIMIZATION

I P IANO : I NERTIAL P ROXIMAL A LGORITHM FOR N ON -C ONVEX O PTIMIZATION I P IANO : I NERTIAL P ROXIMAL A LGORITHM FOR N ON -C ONVEX O PTIMIZATION Peter Ochs University of Freiburg Germany 17.01.2017 joint work with: Thomas Brox and Thomas Pock c 2017 Peter Ochs ipiano c 1

More information

Coordinate Update Algorithm Short Course Operator Splitting

Coordinate Update Algorithm Short Course Operator Splitting Coordinate Update Algorithm Short Course Operator Splitting Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 25 Operator splitting pipeline 1. Formulate a problem as 0 A(x) + B(x) with monotone operators

More information

Convergence of Bregman Alternating Direction Method with Multipliers for Nonconvex Composite Problems

Convergence of Bregman Alternating Direction Method with Multipliers for Nonconvex Composite Problems Convergence of Bregman Alternating Direction Method with Multipliers for Nonconvex Composite Problems Fenghui Wang, Zongben Xu, and Hong-Kun Xu arxiv:40.8625v3 [math.oc] 5 Dec 204 Abstract The alternating

More information

A semi-algebraic look at first-order methods

A semi-algebraic look at first-order methods splitting A semi-algebraic look at first-order Université de Toulouse / TSE Nesterov s 60th birthday, Les Houches, 2016 in large-scale first-order optimization splitting Start with a reasonable FOM (some

More information

Optimization methods

Optimization methods Lecture notes 3 February 8, 016 1 Introduction Optimization methods In these notes we provide an overview of a selection of optimization methods. We focus on methods which rely on first-order information,

More information

Douglas-Rachford splitting for nonconvex feasibility problems

Douglas-Rachford splitting for nonconvex feasibility problems Douglas-Rachford splitting for nonconvex feasibility problems Guoyin Li Ting Kei Pong Jan 3, 015 Abstract We adapt the Douglas-Rachford DR) splitting method to solve nonconvex feasibility problems by studying

More information

Dual and primal-dual methods

Dual and primal-dual methods ELE 538B: Large-Scale Optimization for Data Science Dual and primal-dual methods Yuxin Chen Princeton University, Spring 2018 Outline Dual proximal gradient method Primal-dual proximal gradient method

More information

A user s guide to Lojasiewicz/KL inequalities

A user s guide to Lojasiewicz/KL inequalities Other A user s guide to Lojasiewicz/KL inequalities Toulouse School of Economics, Université Toulouse I SLRA, Grenoble, 2015 Motivations behind KL f : R n R smooth ẋ(t) = f (x(t)) or x k+1 = x k λ k f

More information

Primal-dual Subgradient Method for Convex Problems with Functional Constraints

Primal-dual Subgradient Method for Convex Problems with Functional Constraints Primal-dual Subgradient Method for Convex Problems with Functional Constraints Yurii Nesterov, CORE/INMA (UCL) Workshop on embedded optimization EMBOPT2014 September 9, 2014 (Lucca) Yu. Nesterov Primal-dual

More information

Binary Optimization via Mathematical Programming with Equilibrium Constraints

Binary Optimization via Mathematical Programming with Equilibrium Constraints 1 Binary Optimization via Mathematical Programg with Equilibrium Constraints Ganzhao Yuan, Bernard Ghanem arxiv:108.05v [math.oc] Dec 017 Abstract Binary optimization is a central problem in mathematical

More information

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 9 Alternating Direction Method of Multipliers Shiqian Ma, MAT-258A: Numerical Optimization 2 Separable convex optimization a special case is min f(x)

More information

2 Regularized Image Reconstruction for Compressive Imaging and Beyond

2 Regularized Image Reconstruction for Compressive Imaging and Beyond EE 367 / CS 448I Computational Imaging and Display Notes: Compressive Imaging and Regularized Image Reconstruction (lecture ) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement

More information

LINEARIZED AUGMENTED LAGRANGIAN AND ALTERNATING DIRECTION METHODS FOR NUCLEAR NORM MINIMIZATION

LINEARIZED AUGMENTED LAGRANGIAN AND ALTERNATING DIRECTION METHODS FOR NUCLEAR NORM MINIMIZATION LINEARIZED AUGMENTED LAGRANGIAN AND ALTERNATING DIRECTION METHODS FOR NUCLEAR NORM MINIMIZATION JUNFENG YANG AND XIAOMING YUAN Abstract. The nuclear norm is widely used to induce low-rank solutions for

More information

On the Convergence of Multi-Block Alternating Direction Method of Multipliers and Block Coordinate Descent Method

On the Convergence of Multi-Block Alternating Direction Method of Multipliers and Block Coordinate Descent Method On the Convergence of Multi-Block Alternating Direction Method of Multipliers and Block Coordinate Descent Method Caihua Chen Min Li Xin Liu Yinyu Ye This version: September 9, 2015 Abstract The paper

More information

ON THE GLOBAL AND LINEAR CONVERGENCE OF THE GENERALIZED ALTERNATING DIRECTION METHOD OF MULTIPLIERS

ON THE GLOBAL AND LINEAR CONVERGENCE OF THE GENERALIZED ALTERNATING DIRECTION METHOD OF MULTIPLIERS ON THE GLOBAL AND LINEAR CONVERGENCE OF THE GENERALIZED ALTERNATING DIRECTION METHOD OF MULTIPLIERS WEI DENG AND WOTAO YIN Abstract. The formulation min x,y f(x) + g(y) subject to Ax + By = b arises in

More information

Sparse Optimization Lecture: Dual Methods, Part I

Sparse Optimization Lecture: Dual Methods, Part I Sparse Optimization Lecture: Dual Methods, Part I Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know dual (sub)gradient iteration augmented l 1 iteration

More information

Beyond Heuristics: Applying Alternating Direction Method of Multipliers in Nonconvex Territory

Beyond Heuristics: Applying Alternating Direction Method of Multipliers in Nonconvex Territory Beyond Heuristics: Applying Alternating Direction Method of Multipliers in Nonconvex Territory Xin Liu(4Ð) State Key Laboratory of Scientific and Engineering Computing Institute of Computational Mathematics

More information

Linearized Alternating Direction Method with Adaptive Penalty for Low-Rank Representation

Linearized Alternating Direction Method with Adaptive Penalty for Low-Rank Representation Linearized Alternating Direction Method with Adaptive Penalty for Low-Rank Representation Zhouchen Lin Visual Computing Group Microsoft Research Asia Risheng Liu Zhixun Su School of Mathematical Sciences

More information

Sequential convex programming,: value function and convergence

Sequential convex programming,: value function and convergence Sequential convex programming,: value function and convergence Edouard Pauwels joint work with Jérôme Bolte Journées MODE Toulouse March 23 2016 1 / 16 Introduction Local search methods for finite dimensional

More information

Accelerated primal-dual methods for linearly constrained convex problems

Accelerated primal-dual methods for linearly constrained convex problems Accelerated primal-dual methods for linearly constrained convex problems Yangyang Xu SIAM Conference on Optimization May 24, 2017 1 / 23 Accelerated proximal gradient For convex composite problem: minimize

More information

Active sets, steepest descent, and smooth approximation of functions

Active sets, steepest descent, and smooth approximation of functions Active sets, steepest descent, and smooth approximation of functions Dmitriy Drusvyatskiy School of ORIE, Cornell University Joint work with Alex D. Ioffe (Technion), Martin Larsson (EPFL), and Adrian

More information

arxiv: v1 [math.oc] 23 May 2017

arxiv: v1 [math.oc] 23 May 2017 A DERANDOMIZED ALGORITHM FOR RP-ADMM WITH SYMMETRIC GAUSS-SEIDEL METHOD JINCHAO XU, KAILAI XU, AND YINYU YE arxiv:1705.08389v1 [math.oc] 23 May 2017 Abstract. For multi-block alternating direction method

More information

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization /

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization / Uses of duality Geoff Gordon & Ryan Tibshirani Optimization 10-725 / 36-725 1 Remember conjugate functions Given f : R n R, the function is called its conjugate f (y) = max x R n yt x f(x) Conjugates appear

More information

Solving large Semidefinite Programs - Part 1 and 2

Solving large Semidefinite Programs - Part 1 and 2 Solving large Semidefinite Programs - Part 1 and 2 Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria F. Rendl, Singapore workshop 2006 p.1/34 Overview Limits of Interior

More information

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.18

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.18 XVIII - 1 Contraction Methods for Convex Optimization and Monotone Variational Inequalities No18 Linearized alternating direction method with Gaussian back substitution for separable convex optimization

More information

ALADIN An Algorithm for Distributed Non-Convex Optimization and Control

ALADIN An Algorithm for Distributed Non-Convex Optimization and Control ALADIN An Algorithm for Distributed Non-Convex Optimization and Control Boris Houska, Yuning Jiang, Janick Frasch, Rien Quirynen, Dimitris Kouzoupis, Moritz Diehl ShanghaiTech University, University of

More information

On the iterate convergence of descent methods for convex optimization

On the iterate convergence of descent methods for convex optimization On the iterate convergence of descent methods for convex optimization Clovis C. Gonzaga March 1, 2014 Abstract We study the iterate convergence of strong descent algorithms applied to convex functions.

More information

A memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration

A memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration A memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration E. Chouzenoux, A. Jezierska, J.-C. Pesquet and H. Talbot Université Paris-Est Lab. d Informatique Gaspard

More information

A Customized ADMM for Rank-Constrained Optimization Problems with Approximate Formulations

A Customized ADMM for Rank-Constrained Optimization Problems with Approximate Formulations A Customized ADMM for Rank-Constrained Optimization Problems with Approximate Formulations Chuangchuang Sun and Ran Dai Abstract This paper proposes a customized Alternating Direction Method of Multipliers

More information

ACCELERATED FIRST-ORDER PRIMAL-DUAL PROXIMAL METHODS FOR LINEARLY CONSTRAINED COMPOSITE CONVEX PROGRAMMING

ACCELERATED FIRST-ORDER PRIMAL-DUAL PROXIMAL METHODS FOR LINEARLY CONSTRAINED COMPOSITE CONVEX PROGRAMMING ACCELERATED FIRST-ORDER PRIMAL-DUAL PROXIMAL METHODS FOR LINEARLY CONSTRAINED COMPOSITE CONVEX PROGRAMMING YANGYANG XU Abstract. Motivated by big data applications, first-order methods have been extremely

More information

Coordinate Update Algorithm Short Course Proximal Operators and Algorithms

Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 36 Why proximal? Newton s method: for C 2 -smooth, unconstrained problems allow

More information

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2 EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 2 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory April 5, 2012 Andre Tkacenko

More information

AN AUGMENTED LAGRANGIAN METHOD FOR l 1 -REGULARIZED OPTIMIZATION PROBLEMS WITH ORTHOGONALITY CONSTRAINTS

AN AUGMENTED LAGRANGIAN METHOD FOR l 1 -REGULARIZED OPTIMIZATION PROBLEMS WITH ORTHOGONALITY CONSTRAINTS AN AUGMENTED LAGRANGIAN METHOD FOR l 1 -REGULARIZED OPTIMIZATION PROBLEMS WITH ORTHOGONALITY CONSTRAINTS WEIQIANG CHEN, HUI JI, AND YANFEI YOU Abstract. A class of l 1 -regularized optimization problems

More information

GLOBALLY CONVERGENT ACCELERATED PROXIMAL ALTERNATING MAXIMIZATION METHOD FOR L1-PRINCIPAL COMPONENT ANALYSIS. Peng Wang Huikang Liu Anthony Man-Cho So

GLOBALLY CONVERGENT ACCELERATED PROXIMAL ALTERNATING MAXIMIZATION METHOD FOR L1-PRINCIPAL COMPONENT ANALYSIS. Peng Wang Huikang Liu Anthony Man-Cho So GLOBALLY CONVERGENT ACCELERATED PROXIMAL ALTERNATING MAXIMIZATION METHOD FOR L-PRINCIPAL COMPONENT ANALYSIS Peng Wang Huikang Liu Anthony Man-Cho So Department of Systems Engineering and Engineering Management,

More information

Tight Rates and Equivalence Results of Operator Splitting Schemes

Tight Rates and Equivalence Results of Operator Splitting Schemes Tight Rates and Equivalence Results of Operator Splitting Schemes Wotao Yin (UCLA Math) Workshop on Optimization for Modern Computing Joint w Damek Davis and Ming Yan UCLA CAM 14-51, 14-58, and 14-59 1

More information

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization MATHEMATICS OF OPERATIONS RESEARCH Vol. 29, No. 3, August 2004, pp. 479 491 issn 0364-765X eissn 1526-5471 04 2903 0479 informs doi 10.1287/moor.1040.0103 2004 INFORMS Some Properties of the Augmented

More information

Math 273a: Optimization Overview of First-Order Optimization Algorithms

Math 273a: Optimization Overview of First-Order Optimization Algorithms Math 273a: Optimization Overview of First-Order Optimization Algorithms Wotao Yin Department of Mathematics, UCLA online discussions on piazza.com 1 / 9 Typical flow of numerical optimization Optimization

More information

Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery

Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery Department of Mathematics & Risk Management Institute National University of Singapore (Based on a joint work with Shujun

More information

A Bregman alternating direction method of multipliers for sparse probabilistic Boolean network problem

A Bregman alternating direction method of multipliers for sparse probabilistic Boolean network problem A Bregman alternating direction method of multipliers for sparse probabilistic Boolean network problem Kangkang Deng, Zheng Peng Abstract: The main task of genetic regulatory networks is to construct a

More information

The Multi-Path Utility Maximization Problem

The Multi-Path Utility Maximization Problem The Multi-Path Utility Maximization Problem Xiaojun Lin and Ness B. Shroff School of Electrical and Computer Engineering Purdue University, West Lafayette, IN 47906 {linx,shroff}@ecn.purdue.edu Abstract

More information

1 Computing with constraints

1 Computing with constraints Notes for 2017-04-26 1 Computing with constraints Recall that our basic problem is minimize φ(x) s.t. x Ω where the feasible set Ω is defined by equality and inequality conditions Ω = {x R n : c i (x)

More information

A Majorize-Minimize subspace approach for l 2 -l 0 regularization with applications to image processing

A Majorize-Minimize subspace approach for l 2 -l 0 regularization with applications to image processing A Majorize-Minimize subspace approach for l 2 -l 0 regularization with applications to image processing Emilie Chouzenoux emilie.chouzenoux@univ-mlv.fr Université Paris-Est Lab. d Informatique Gaspard

More information

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem: CDS270 Maryam Fazel Lecture 2 Topics from Optimization and Duality Motivation network utility maximization (NUM) problem: consider a network with S sources (users), each sending one flow at rate x s, through

More information

A multi-stage convex relaxation approach to noisy structured low-rank matrix recovery

A multi-stage convex relaxation approach to noisy structured low-rank matrix recovery A multi-stage convex relaxation approach to noisy structured low-rank matrix recovery Shujun Bi, Shaohua Pan and Defeng Sun March 1, 2017 Abstract This paper concerns with a noisy structured low-rank matrix

More information

Dual methods and ADMM. Barnabas Poczos & Ryan Tibshirani Convex Optimization /36-725

Dual methods and ADMM. Barnabas Poczos & Ryan Tibshirani Convex Optimization /36-725 Dual methods and ADMM Barnabas Poczos & Ryan Tibshirani Convex Optimization 10-725/36-725 1 Given f : R n R, the function is called its conjugate Recall conjugate functions f (y) = max x R n yt x f(x)

More information

GENERALIZED second-order cone complementarity

GENERALIZED second-order cone complementarity Stochastic Generalized Complementarity Problems in Second-Order Cone: Box-Constrained Minimization Reformulation and Solving Methods Mei-Ju Luo and Yan Zhang Abstract In this paper, we reformulate the

More information

HYBRID JACOBIAN AND GAUSS SEIDEL PROXIMAL BLOCK COORDINATE UPDATE METHODS FOR LINEARLY CONSTRAINED CONVEX PROGRAMMING

HYBRID JACOBIAN AND GAUSS SEIDEL PROXIMAL BLOCK COORDINATE UPDATE METHODS FOR LINEARLY CONSTRAINED CONVEX PROGRAMMING SIAM J. OPTIM. Vol. 8, No. 1, pp. 646 670 c 018 Society for Industrial and Applied Mathematics HYBRID JACOBIAN AND GAUSS SEIDEL PROXIMAL BLOCK COORDINATE UPDATE METHODS FOR LINEARLY CONSTRAINED CONVEX

More information

Projection methods to solve SDP

Projection methods to solve SDP Projection methods to solve SDP Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria F. Rendl, Oberwolfach Seminar, May 2010 p.1/32 Overview Augmented Primal-Dual Method

More information

Matrix Rank Minimization with Applications

Matrix Rank Minimization with Applications Matrix Rank Minimization with Applications Maryam Fazel Haitham Hindi Stephen Boyd Information Systems Lab Electrical Engineering Department Stanford University 8/2001 ACC 01 Outline Rank Minimization

More information

From error bounds to the complexity of first-order descent methods for convex functions

From error bounds to the complexity of first-order descent methods for convex functions From error bounds to the complexity of first-order descent methods for convex functions Nguyen Trong Phong-TSE Joint work with Jérôme Bolte, Juan Peypouquet, Bruce Suter. Toulouse, 23-25, March, 2016 Journées

More information

Sparse Approximation via Penalty Decomposition Methods

Sparse Approximation via Penalty Decomposition Methods Sparse Approximation via Penalty Decomposition Methods Zhaosong Lu Yong Zhang February 19, 2012 Abstract In this paper we consider sparse approximation problems, that is, general l 0 minimization problems

More information

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem Michael Patriksson 0-0 The Relaxation Theorem 1 Problem: find f := infimum f(x), x subject to x S, (1a) (1b) where f : R n R

More information

An inertial forward-backward algorithm for the minimization of the sum of two nonconvex functions

An inertial forward-backward algorithm for the minimization of the sum of two nonconvex functions An inertial forward-backward algorithm for the minimization of the sum of two nonconvex functions Radu Ioan Boţ Ernö Robert Csetnek Szilárd Csaba László October, 1 Abstract. We propose a forward-backward

More information

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST) Lagrange Duality Daniel P. Palomar Hong Kong University of Science and Technology (HKUST) ELEC5470 - Convex Optimization Fall 2017-18, HKUST, Hong Kong Outline of Lecture Lagrangian Dual function Dual

More information

Splitting methods for decomposing separable convex programs

Splitting methods for decomposing separable convex programs Splitting methods for decomposing separable convex programs Philippe Mahey LIMOS - ISIMA - Université Blaise Pascal PGMO, ENSTA 2013 October 4, 2013 1 / 30 Plan 1 Max Monotone Operators Proximal techniques

More information

Rank minimization via the γ 2 norm

Rank minimization via the γ 2 norm Rank minimization via the γ 2 norm Troy Lee Columbia University Adi Shraibman Weizmann Institute Rank Minimization Problem Consider the following problem min X rank(x) A i, X b i for i = 1,..., k Arises

More information

Robust Principal Component Analysis Based on Low-Rank and Block-Sparse Matrix Decomposition

Robust Principal Component Analysis Based on Low-Rank and Block-Sparse Matrix Decomposition Robust Principal Component Analysis Based on Low-Rank and Block-Sparse Matrix Decomposition Gongguo Tang and Arye Nehorai Department of Electrical and Systems Engineering Washington University in St Louis

More information

Algorithms for Nonsmooth Optimization

Algorithms for Nonsmooth Optimization Algorithms for Nonsmooth Optimization Frank E. Curtis, Lehigh University presented at Center for Optimization and Statistical Learning, Northwestern University 2 March 2018 Algorithms for Nonsmooth Optimization

More information

Fast Nonnegative Matrix Factorization with Rank-one ADMM

Fast Nonnegative Matrix Factorization with Rank-one ADMM Fast Nonnegative Matrix Factorization with Rank-one Dongjin Song, David A. Meyer, Martin Renqiang Min, Department of ECE, UCSD, La Jolla, CA, 9093-0409 dosong@ucsd.edu Department of Mathematics, UCSD,

More information

Proximal Alternating Linearized Minimization for Nonconvex and Nonsmooth Problems

Proximal Alternating Linearized Minimization for Nonconvex and Nonsmooth Problems Proximal Alternating Linearized Minimization for Nonconvex and Nonsmooth Problems Jérôme Bolte Shoham Sabach Marc Teboulle Abstract We introduce a proximal alternating linearized minimization PALM) algorithm

More information

An Optimization-based Approach to Decentralized Assignability

An Optimization-based Approach to Decentralized Assignability 2016 American Control Conference (ACC) Boston Marriott Copley Place July 6-8, 2016 Boston, MA, USA An Optimization-based Approach to Decentralized Assignability Alborz Alavian and Michael Rotkowitz Abstract

More information

minimize x subject to (x 2)(x 4) u,

minimize x subject to (x 2)(x 4) u, Math 6366/6367: Optimization and Variational Methods Sample Preliminary Exam Questions 1. Suppose that f : [, L] R is a C 2 -function with f () on (, L) and that you have explicit formulae for

More information

TMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM

TMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM TMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM H. E. Krogstad, IMF, Spring 2012 Karush-Kuhn-Tucker (KKT) Theorem is the most central theorem in constrained optimization, and since the proof is scattered

More information

A convergence result for an Outer Approximation Scheme

A convergence result for an Outer Approximation Scheme A convergence result for an Outer Approximation Scheme R. S. Burachik Engenharia de Sistemas e Computação, COPPE-UFRJ, CP 68511, Rio de Janeiro, RJ, CEP 21941-972, Brazil regi@cos.ufrj.br J. O. Lopes Departamento

More information

Relaxations and Randomized Methods for Nonconvex QCQPs

Relaxations and Randomized Methods for Nonconvex QCQPs Relaxations and Randomized Methods for Nonconvex QCQPs Alexandre d Aspremont, Stephen Boyd EE392o, Stanford University Autumn, 2003 Introduction While some special classes of nonconvex problems can be

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Regularization: Ridge Regression and Lasso Week 14, Lecture 2

MA 575 Linear Models: Cedric E. Ginestet, Boston University Regularization: Ridge Regression and Lasso Week 14, Lecture 2 MA 575 Linear Models: Cedric E. Ginestet, Boston University Regularization: Ridge Regression and Lasso Week 14, Lecture 2 1 Ridge Regression Ridge regression and the Lasso are two forms of regularized

More information

Research Article Solving the Matrix Nearness Problem in the Maximum Norm by Applying a Projection and Contraction Method

Research Article Solving the Matrix Nearness Problem in the Maximum Norm by Applying a Projection and Contraction Method Advances in Operations Research Volume 01, Article ID 357954, 15 pages doi:10.1155/01/357954 Research Article Solving the Matrix Nearness Problem in the Maximum Norm by Applying a Projection and Contraction

More information

arxiv: v4 [math.oc] 5 Jan 2016

arxiv: v4 [math.oc] 5 Jan 2016 Restarted SGD: Beating SGD without Smoothness and/or Strong Convexity arxiv:151.03107v4 [math.oc] 5 Jan 016 Tianbao Yang, Qihang Lin Department of Computer Science Department of Management Sciences The

More information

Linearized Alternating Direction Method: Two Blocks and Multiple Blocks. Zhouchen Lin 林宙辰北京大学

Linearized Alternating Direction Method: Two Blocks and Multiple Blocks. Zhouchen Lin 林宙辰北京大学 Linearized Alternating Direction Method: Two Blocks and Multiple Blocks Zhouchen Lin 林宙辰北京大学 Dec. 3, 014 Outline Alternating Direction Method (ADM) Linearized Alternating Direction Method (LADM) Two Blocks

More information

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications Professor M. Chiang Electrical Engineering Department, Princeton University March

More information

A Unified Approach to Proximal Algorithms using Bregman Distance

A Unified Approach to Proximal Algorithms using Bregman Distance A Unified Approach to Proximal Algorithms using Bregman Distance Yi Zhou a,, Yingbin Liang a, Lixin Shen b a Department of Electrical Engineering and Computer Science, Syracuse University b Department

More information

Solving Dual Problems

Solving Dual Problems Lecture 20 Solving Dual Problems We consider a constrained problem where, in addition to the constraint set X, there are also inequality and linear equality constraints. Specifically the minimization problem

More information

Lecture 7: Positive Semidefinite Matrices

Lecture 7: Positive Semidefinite Matrices Lecture 7: Positive Semidefinite Matrices Rajat Mittal IIT Kanpur The main aim of this lecture note is to prepare your background for semidefinite programming. We have already seen some linear algebra.

More information

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.11

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.11 XI - 1 Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.11 Alternating direction methods of multipliers for separable convex programming Bingsheng He Department of Mathematics

More information

arxiv: v5 [cs.na] 22 Mar 2018

arxiv: v5 [cs.na] 22 Mar 2018 Iteratively Linearized Reweighted Alternating Direction Method of Multipliers for a Class of Nonconvex Problems Tao Sun Hao Jiang Lizhi Cheng Wei Zhu arxiv:1709.00483v5 [cs.na] 22 Mar 2018 March 26, 2018

More information

l 0 TV: A Sparse Optimization Method for Impulse Noise Image Restoration

l 0 TV: A Sparse Optimization Method for Impulse Noise Image Restoration IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1 l 0 TV: A Sparse Optimization Method for Impulse Noise Image Restoration Ganzhao Yuan, Bernard Ghanem arxiv:1802.09879v2 [cs.na] 28 Dec

More information

Kaisa Joki Adil M. Bagirov Napsu Karmitsa Marko M. Mäkelä. New Proximal Bundle Method for Nonsmooth DC Optimization

Kaisa Joki Adil M. Bagirov Napsu Karmitsa Marko M. Mäkelä. New Proximal Bundle Method for Nonsmooth DC Optimization Kaisa Joki Adil M. Bagirov Napsu Karmitsa Marko M. Mäkelä New Proximal Bundle Method for Nonsmooth DC Optimization TUCS Technical Report No 1130, February 2015 New Proximal Bundle Method for Nonsmooth

More information

A Continuation Approach Using NCP Function for Solving Max-Cut Problem

A Continuation Approach Using NCP Function for Solving Max-Cut Problem A Continuation Approach Using NCP Function for Solving Max-Cut Problem Xu Fengmin Xu Chengxian Ren Jiuquan Abstract A continuous approach using NCP function for approximating the solution of the max-cut

More information

Robust PCA. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng

Robust PCA. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng Robust PCA CS5240 Theoretical Foundations in Multimedia Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (NUS) Robust PCA 1 / 52 Previously...

More information

Convex Optimization Algorithms for Machine Learning in 10 Slides

Convex Optimization Algorithms for Machine Learning in 10 Slides Convex Optimization Algorithms for Machine Learning in 10 Slides Presenter: Jul. 15. 2015 Outline 1 Quadratic Problem Linear System 2 Smooth Problem Newton-CG 3 Composite Problem Proximal-Newton-CD 4 Non-smooth,

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

Tame variational analysis

Tame variational analysis Tame variational analysis Dmitriy Drusvyatskiy Mathematics, University of Washington Joint work with Daniilidis (Chile), Ioffe (Technion), and Lewis (Cornell) May 19, 2015 Theme: Semi-algebraic geometry

More information

EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6)

EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6) EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement to the material discussed in

More information

Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering

Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering Shuyang Ling Courant Institute of Mathematical Sciences, NYU Aug 13, 2018 Joint

More information

5 Handling Constraints

5 Handling Constraints 5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest

More information

Sequential Quadratic Programming Method for Nonlinear Second-Order Cone Programming Problems. Hirokazu KATO

Sequential Quadratic Programming Method for Nonlinear Second-Order Cone Programming Problems. Hirokazu KATO Sequential Quadratic Programming Method for Nonlinear Second-Order Cone Programming Problems Guidance Professor Masao FUKUSHIMA Hirokazu KATO 2004 Graduate Course in Department of Applied Mathematics and

More information

Key words. first order methods, trust region subproblem, optimality conditions, global optimum

Key words. first order methods, trust region subproblem, optimality conditions, global optimum GLOBALLY SOLVING THE TRUST REGION SUBPROBLEM USING SIMPLE FIRST-ORDER METHODS AMIR BECK AND YAKOV VAISBOURD Abstract. We consider the trust region subproblem which is given by a minimization of a quadratic,

More information

Optimizing Nonconvex Finite Sums by a Proximal Primal-Dual Method

Optimizing Nonconvex Finite Sums by a Proximal Primal-Dual Method Optimizing Nonconvex Finite Sums by a Proximal Primal-Dual Method Davood Hajinezhad Iowa State University Davood Hajinezhad Optimizing Nonconvex Finite Sums by a Proximal Primal-Dual Method 1 / 35 Co-Authors

More information

Numerical optimization. Numerical optimization. Longest Shortest where Maximal Minimal. Fastest. Largest. Optimization problems

Numerical optimization. Numerical optimization. Longest Shortest where Maximal Minimal. Fastest. Largest. Optimization problems 1 Numerical optimization Alexander & Michael Bronstein, 2006-2009 Michael Bronstein, 2010 tosca.cs.technion.ac.il/book Numerical optimization 048921 Advanced topics in vision Processing and Analysis of

More information

14 Singular Value Decomposition

14 Singular Value Decomposition 14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing

More information

A STABILIZED SQP METHOD: GLOBAL CONVERGENCE

A STABILIZED SQP METHOD: GLOBAL CONVERGENCE A STABILIZED SQP METHOD: GLOBAL CONVERGENCE Philip E. Gill Vyacheslav Kungurtsev Daniel P. Robinson UCSD Center for Computational Mathematics Technical Report CCoM-13-4 Revised July 18, 2014, June 23,

More information

A Multilevel Proximal Algorithm for Large Scale Composite Convex Optimization

A Multilevel Proximal Algorithm for Large Scale Composite Convex Optimization A Multilevel Proximal Algorithm for Large Scale Composite Convex Optimization Panos Parpas Department of Computing Imperial College London www.doc.ic.ac.uk/ pp500 p.parpas@imperial.ac.uk jointly with D.V.

More information

We describe the generalization of Hazan s algorithm for symmetric programming

We describe the generalization of Hazan s algorithm for symmetric programming ON HAZAN S ALGORITHM FOR SYMMETRIC PROGRAMMING PROBLEMS L. FAYBUSOVICH Abstract. problems We describe the generalization of Hazan s algorithm for symmetric programming Key words. Symmetric programming,

More information

arxiv: v2 [math.oc] 21 Nov 2017

arxiv: v2 [math.oc] 21 Nov 2017 Unifying abstract inexact convergence theorems and block coordinate variable metric ipiano arxiv:1602.07283v2 [math.oc] 21 Nov 2017 Peter Ochs Mathematical Optimization Group Saarland University Germany

More information

Convergence rates for an inertial algorithm of gradient type associated to a smooth nonconvex minimization

Convergence rates for an inertial algorithm of gradient type associated to a smooth nonconvex minimization Convergence rates for an inertial algorithm of gradient type associated to a smooth nonconvex minimization Szilárd Csaba László November, 08 Abstract. We investigate an inertial algorithm of gradient type

More information

Constrained Optimization and Lagrangian Duality

Constrained Optimization and Lagrangian Duality CIS 520: Machine Learning Oct 02, 2017 Constrained Optimization and Lagrangian Duality Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may

More information

Convergence of Cubic Regularization for Nonconvex Optimization under KŁ Property

Convergence of Cubic Regularization for Nonconvex Optimization under KŁ Property Convergence of Cubic Regularization for Nonconvex Optimization under KŁ Property Yi Zhou Department of ECE The Ohio State University zhou.1172@osu.edu Zhe Wang Department of ECE The Ohio State University

More information

On the convergence of a regularized Jacobi algorithm for convex optimization

On the convergence of a regularized Jacobi algorithm for convex optimization On the convergence of a regularized Jacobi algorithm for convex optimization Goran Banjac, Kostas Margellos, and Paul J. Goulart Abstract In this paper we consider the regularized version of the Jacobi

More information