A Proximal Alternating Direction Method for Semi-Definite Rank Minimization (Supplementary Material)
|
|
- Brent Martin
- 5 years ago
- Views:
Transcription
1 A Proximal Alternating Direction Method for Semi-Definite Rank Minimization (Supplementary Material) Ganzhao Yuan and Bernard Ghanem King Abdullah University of Science and Technology (KAUST), Saudi Arabia The supplementary material is organized as follows. Section 1 presents the details of our proofs. Section 2 presents the convergence analysis of the proposed MPEC-based proximal ADM algorithm. Section 3 presents the classical ADM for solving the subproblem. Finally, Section presents additional experimental results. 1 Proofs Lemma 1. [Semidefinite Embedding Lemma (Fazel, Hindi, and Boyd 23)] Let R R m n be a given matrix. Then (R) r if and only if there exist matrices S = S T R m m and T = T T R n n such that [ ] S R R T, (S) + (T) 2r. (1) T Remark. We remark that the semidefinite optimization problem discussed in this paper is very general. To illustrate this point, we consider the following optimization problem: min g(r) + λ (R) (2) R By the Semidefinite Embedding Lemma, we have the following equivalent semi-definite optimization problem : min g(r) + 1 ([ ]) [ ] S S R R 2, s.t. T R T. T Since [ ] S R R T T is positive semidefinite, by the Schur complement condition 1, it holds that S and T. Then the variable matrix [ ] S T T is also positive semidefinite. Using the MPEC reformulation, we have the following equivalent optimization problem: min g(r) + 1 R,S,T,V 2 tr(v), [ ] [ ] S S R s.t. I V, =, T R T, I V. T Clearly, the equivalent optimization problem in Eq (2) can be solved by our proposed PADM algorithm. Copyright c 2, Association for the Advancement of Artificial Intelligence ( All rights reserved. 1 Schur complement condition implies that: X ( A B B T C ) A, S = C B T A B Here, we prove the variational formulation of the NP-hard function. Lemma 2. For any given positive semidefinite matrix X R n n, it holds that (X) = min tr(i V), s.t. V, X =, V I (3) V and the unique optimal solution of the minimization problem in Eq (3) is given by V = Udiag(sign(σ))U T, where X = Udiag(σ)U T is the eigenvalue decomposition of X, sign is the standard signum function 2. Proof. First of all, we let X and V be arbitrary feasible solutions with their eigenvalue decomposition given by X = Udiag(σ)U T and V = Sdiag(δ)S T, respectively. It always holds that: X, V = Udiag( σ)diag( δ)s T 2 F The second equality is achieved only when V and X are simultaneously unitarily diagonalizable, i.e. both V and X share the same spectral decomposition. Therefore, the feasible set for V in Eq (3) must be contained in the set {V V = Udiag(v)U T, v 1, v R r }. Using the fact that tr(v) = v, 1, tr(x) = σ, 1 and V, X = Udiag(v)U T, Udiag(σ)U T = v, σ, Eq (3) reduces to the vector case: (X) = min v 1, 1 v, s.t. v, σ =, v 1 () Since v and σ, Eq () is equivalent to Eq (): (X) = min v 1, 1 v, s.t. v σ =, v 1 () where denotes the Hadamard product (also known as the entrywise product). Note that for all i [r] when σ i =, v i = 1 will be achieved by minimization, when σ i, v i = will be enforced by the constraint. Since the objective function in Eq () is linear, minimization is always achieved at the boundaries of the feasible solution space. Thus, v = 1 sign(σ). Finally, we have V = Udiag(1 sign(σ))u T. We thus complete the proof of this lemma. 1, x < ; 2 sign(x) =, x = ; 1, x >.
2 The following lemma shows how to compute the generalized Singular Value Thresholding (SVT) operator which is involved in V-subproblem in our PADM Algorithm. Lemma 3. Assume that W has the SVD decomposition that W = Udiag(σ)U T. The optimal solution of the following optimization problem arg min V 1 2 V W 2 F + I (V) () can be computed as Udiag(min(1, max(, σ)))t. Here I is an indicator function{ of the convex set {V, V V I} with I (V)., otherwise. Proof. The proof of this lemma is very natural. For completeness, we present our proof here. For notation convenience, we use σ and z to denote the singular values of W and V, respectively. We naturally derive the following inequalities: 1 2 V W 2 F + I (V) = 1 2 ( z 2 + σ 2 2 V, W ) + I Θ (z) 1 2 ( z 2 + σ 2 2 z, σ ) + I Θ (z) = 1 2 z σ 2 + I Θ (z) From the von Veumann s trace inequality, the solution set of () must be contained in the set {V V = Udiag(v )U T }, where v is given by v 1 = arg min z 2 z σ 2 + I Θ (z) (7) Since the optimization problem in Eq(7) is decomposable, a simple computation yields that the solution can be computed in closed form as: v = min(1, max(, σ)). Therefore, V = Udiag(min(1, max(, σ)))u T. We thus complete the proof of this lemma. 2 Convergence Analysis The global convergence of ADM for convex problems was given by He and Yuan in (He and Yuan 2) under an elegant variational inequality framework. However, since our MPEC optimization problem is non-convex, the convergence analysis for ADM needs additional conditions. In non-convex optimization, convergence to a stationary point (local minimum) is the best convergence property that we can hope for. Under boundedness condition, we show that the sequence generated by the proximal ADM converges to a KKT point. For the ease of discussions, we define: and u {X, V}, s {X, V, π} () Ω {X X κi}, {V V I} (9) First of all, we present the first-order KKT conditions of the MPEC reformulation optimization problem. Based on the augmented Lagrangian function of the MPEC reformulation, we naturally derive the following KKT conditions of the optimization problem for {X, V, π }: I Ω (X ) + A T g(a(x ) b) + πv I (V ) λi + πx = V, X whose existence can be guaranteed by Robinson s constraint qualification (Rockafellar, Wets, and Wets 199). First of all, we prove the subgradient lower bound for the iterates gap by the following lemma. Lemma. Assume that π k is bounded for all k, then there exists a constant ϖ > such that the following inequality holds: L(s k+1 ) ϖ s k+1 s k (1) Proof. By the optimal condition of the X-subproblem and V-subproblem, we have: D(X k+1 X k ) + A T g(a(x k+1 ) b) + π k V k + α V k, X k+1 V k + I Ω (X k+1 ) (11) E(V k+1 V k ) λi + π k X k+1 +α V k+1, X k+1 X k+1 + I (V k+1 ) () By the definition of L we have that L X (s k+1 ) = A T g(ax k+1 b) + π k+1 V k+1 + α V k+1, X k+1 V k+1 + I Ω (X k+1 ) = (π k + α V k, X k+1 )V k + (π k+1 + α V k+1, X k+1 )V k+1 D(X k+1 X k ) = (π k + α V k+1, X k+1 + α V k V k+1, X k+1 )V k +(π k+1 + α V k+1, X k+1 )V k+1 D(X k+1 X k ) = (π k+1 + α V k V k+1, X k+1 )V k + (2π k+1 π k )V k+1 D(X k+1 X k ) = (α V k V k+1, X k+1 )V k + (π k+1 π k )V k+1 +π k+1 (V k+1 V k ) D(X k+1 X k ) The first step uses the definition of L X (s k+1 ), the second step uses Eq (11), the third step uses V k + V k+1 V k+1 = V k, the fourth step uses the multiplier update rule for π. Assume that π k+1 is bounded by a constant ρ that π k+1 ρ. We have: L X (s k+1 ) F α V k V k+1, X k+1 V k + (π k+1 π k )V k+1 + π k+1 (V k+1 V k ) + D(X k+1 X k ) 2nκα V k V k+1 F + 2 π k+1 π k ρ V k V k+1 F + D X k+1 X k F (13)
3 Similarly, we have L V (s k+1 ) = I (V k+1 ) λi + π k+1 X k+1 +α V k+1, X k+1 X k+1 = (π k+1 π k )X k+1 E(V k+1 V k ) L π (s k+1 ) = I V k+1, X k+1 = 1 α (πk+1 π k ) Then we derive the following inequalities: L V (s k+1 ) F κ π k π k+1 + E V k+1 V k F () L π (s k+1 ) 1 α πk+1 π k () Combining Eqs (13-), we conclude that there exists ϖ > such that the following inequality holds L(s) F ϖ s s k. Thus, we complete the proof of this lemma. The following lemma is useful in our convergence analysis. Lemma. Assume that π k is bounded for all k, then we have the following inequality: s k s k+1 2 < + () k= In particular the sequence s k s k+1 is asymptotic regular, namely s k s k+1 as k. Moreover any cluster point of s k is a stationary point of L. Proof. Due to the initialization and the update rule of π, we conclude that π k is nonnegative and monotone non-decreasing. Moreover, as k, we have: X k+1, V k+1 =. This can be proved by contradiction. Suppose that X k+1, V k+1, then π k = + as k. This contradicts our assumption that π k is bounded. Therefore, we conclude that as k + it holds that X k+1, V k+1 =, π i+1 π i < + (17) π i+1 π i 2 < + On the other hand, we naturally derive the following inequalities: L(X k+1, V k+1 ; π k+1 ) = L(X k+1, V k+1 ; π k ) + π k+1 π k, V k+1, X k+1 = L(X k+1, V k+1 ; π k ) + 1 α πk+1 π k 2 L(X k, V k+1 ; π k ) 2 Xk+1 X k α πk+1 π k 2 L(X k, V k ; π k ) 2 Xk+1 X k 2 2 Vk+1 V k α πk+1 π k 2 (1) The first step uses the definition of L; the second step uses update rule of the Lagrangian multiplier π; the third and fourth step use the -strongly convexity of L with respect to X and v, respectively. We define C = L(X, V ; π ) L(X k+1, V k+1 ; π k+1 )+ 1 α k πi+1 π i 2. Clearly, by the boundedness of X k, V k and π k, both L(X k+1, V k+1 ; π k+1 ) and C are bounded. Summing Eq (1) over i = 1, 2..., k, we have: 2 u i+1 u i 2 C (19) Therefore, combining Eq(17) and Eq (19), we have + k=1 sk+1 s k 2 < + ; in particular s k+1 s k. By Eq (1), we have that: L(s k+1 ) ϖ s k+1 s k (2) which implies that any cluster point of s k is a stationary point of L. We complete the proof of this lemma. Remarks: Lemma states that any cluster point is the KKT point. Strictly speaking, this result does not imply the convergence of the algorithm. This is because the boundedness of k= sk s k+1 2 does not imply that the sequence s k is convergent 3. In what follows, we aim to prove stronger result in Theorem 1. Our analysis is mainly based on a recent non-convex analysis tool called Kurdyka-Łojasiewicz inequality (Attouch et al. 21; Bolte, Sabach, and Teboulle 2). One key condition of our proof requires that the Lagrangian function L(s) satisfies the so-call (KL) property in its effective domain. It is so-called the semi-algebraic function satisfy the Kurdyka- Łojasiewicz property. It is not hard to validate that the Lagrangian function L(s) is a semi-algebraic function. This is not surprising since semi-algebraic function is ubiquitous in applications. Interested readers can refer to (Xu and Yin 213) for more details. We now present the following proposition established in (Attouch et al. 21). Proposition 1. For a given semi-algebraic function L(s), for all s doml, there exists θ [, 1), η (, + ] a neighborhood S of s and a concave and continuous function ϕ(t) = c t 1 θ, t [, η) such that for all s S and satisfies L( s) (L(s), L(s) + η), the following inequality holds: (, L( s))ϕ k (L(s) L( s)) 1, s where (, L( s)) = min{ u : u L( s)}. 3 One typical counter-example is s k = k 1. Clearly, i k= sk s k+1 2 = k=1 ( 1 k )2 is bounded by π2 ; however, s k is divergent since s k = ln(k) + C e, where C e is the well-known Euler s constant. Note that semi-algebraic functions include (i) real polynomial functions, (ii) finite sums and products of semi-algebraic functions, and (iii) indicator functions of semi-algebraic sets. Using these definitions repeatedly, the graph of L(s) : {(s, t) t = L(s)} can be proved to be a semi-algebraic set. Therefore, L(s) is a semialgebraic function.
4 The following theorem establishes the convergence properties of the proposed algorithm under a boundedness condition. Theorem 1. Assume that π k is bounded for all k. Then we have the following inequality: + k= s k s k+1 < (21) Moreover, as k, Algorithm 1 converges to the first order KKT point of the MPEC reformulation optimization problem. Proof. For simplicity, we define R k = ϕ(l(s k ) L(s )) ϕ(l(s k+1 ) L(s )). We naturally derive the following inequalities: 2 uk+1 u k 2 1 α πk+1 π k 2 L(s k ) L(s k+1 ) = (L(s k ) L(s )) (L(s k+1 ) L(s )) R k ϕ k (L(s k ) L(s )) R k (, L(s k )) R k ϖ( X k X k 1 + V k V k 1 + π k π k 1 ) R k ϖ( 2 u k u k 1 + π k π k 1 ) = R k ϖ ( uk u k 1 + πk π k 1 ) (Rk ϖ) 2 + ( uk u k 1 + πk π k 1 ) 2 The first step uses Eq (1); the third step uses the concavity of ϕ such that ϕ(a) ϕ(b) ϕ k (a)(a b) for all a, b R; the fourth step uses the KL property such that (, L(s k ))ϕ k (L(s k ) L(s )) 1 as in Proposition 1; the fifth step uses Eq (1); the sixth step uses the inequality that x; y x + y 2 x; y, where ; in [ ] denotes the row-wise partitioning indicator as in Matlab; the last step uses that fact that 2ab a 2 + b 2 for all a, b R. As k, we have: 2 uk+1 u k 2 (Rk κ) 2 + ( ) 2 uk u k 1 Take the squared root of both side and use the inequality that a + b a + b for all a, b R+, we have uk+1 u k Rk κ + 32 uk u k 1 Rk κ + ( uk u k 1 u k+1 u k ) Summing the inequality above over i = 1, 2..., k, we have: u i+1 u i Z + ( u 1 u + u k+1 u k ) where Z = κ k (ϕ(l(s ) L(s )) ϕ(l(s k+1 ) L(s ))) is bounded real number. Therefore, we conclude that as k, we obtain: u i+1 u i < + () Moreover, by Eq (1) in Lemma we have L(s k+1 ) =. In other words, we have the following results: I (X k+1 ) + A T g(a(x k+1 ) b) + π k+1 V k+1 I (V k+1 ) + λi + π k+1 X k+1 = V k+1, X k+1 which imply that {X k+1, V k+1, π k+1 } is a first-order KKT point. Algorithm 2 Classical Alternating Direction Method for Solving the Convex X-Subproblem in Eq (). (S.) Initialize X, Y = R n n, y = R m. z = R m, Z = R n n. Set t =. (S.1) Solve the following (y, X)-subproblem: X t+1 = arg min J (X, y t, Y t ; Z t, z t ) (23) X (S.2) Solve the following Y-subproblem: (y t+1, Y t+1 ) = arg min J (X t+1, y, Y; Z t, z t ) (2) y, Y κi (S.3) Update the Lagrange multiplier via the following formula: Z t+1 = Z t + β(x t+1 Y t+1 ) z t+1 = z t + γ(a(x k+1 ) b y k+1 ) (S.) Set t := t + 1 and then go to Step (S.1). 3 Solving the Convex Subproblem The efficiency of Algorithm proximal ADM in Algorithm 1 relies whether the convex subproblem can be efficiently solved. In this section, we aim to solve the following semidefinite optimization subproblem involved in the proposed PADM algorithm: min g (A(X) b) + α X κi 2 B(X) 2 F + 2 X 2 F + X, C, () Our solution is naturally based on the classical ADM (He and Yuan 2; Lin, Liu, and Su 211). For completeness,
5 we include our algorithm details here. First, we introduce t- wo auxiliary vectors y R m and Y R n n to reformulate Eq () as: min g (y) + α y,x,y 2 B(X) 2 F + 2 X 2 F + X, C, s.t. A(X) b = y, X = Y, Y κi. (2) Let J β,γ : R m R n n R n n R m R n n R be the augmented Lagrangian function in Eq(2) J (y, X, Y; z, Z) = g (y) + α 2 B(X) 2 F + 2 X 2 F + X, C + z, A(X) b y + γ 2 A(X) b y 2 F + Z, X Y + β 2 X Y 2 F, s.t. Y κi z and Z are the Lagrange multipliers associated to the constraints A(Y) b y = and X Y =, respectively, and γ, β > are the penalty parameters. The detailed iteration steps of the classical ADM for Eq (2) are described in Algorithm 2. Next, we focus our attention on the solutions of subproblems (23) and (2) arising in Algorithm 2. (i) X-subproblem. The first-order optimality condition for the variable X is: αb B(X t+1 ) + γa A(X t+1 ) + ( + β)x t+1 = E where E = βy + γa (b + y) C A z Z. Solving this linear system gives: X t+1 = (αb B + γa A + ( + β)i) 1 E (27) When the dimension of the solution is high, solving this linear system may dominant the computation time. However, one can use iterative conjugate gradient to alleviate this computational burden. We remark that it is also possible to utilize linearized ADM to address this issue (He and Yuan 2; Lin, Liu, and Su 211). (ii) (y, Y)-subproblem. Variable y in Eq (2) is updated by solving the following problem: y t+1 = arg min g (y) + γ y R m 2 q y 2, with q = A(X) b+z/γ. It reduces to the Moreau proximal operator prox g ( ) that can be evaluated efficiently by our assumption in the introduction section. Variable Y in Eq (2) is updated by solving the following problem: Y t+1 β = arg min Y κi 2 Y S 2 F with S = X t+1 + Z t /β. Assume that S has the spectral decomposition that S = Vdiag(s)V T. A simple computation yields that the solution Y t+1 can be computed in closed form as: Y t+1 = Vdiag(max(, min(κ1, s)))v T. The exposition above shows that the computation required in each iteration of Algorithm 2 is insignificant. Classical ADM has excellent convergence both in theory and in practice for convex problems. The convergence of Algorithm 2 can be obtained since the Féjer monotonicity of iterative sequences {y t, X t, Y t, z t, Z t } holds due to convexity. For the proof of convergence of Algorithm 2, interested readers can refer to (He and Yuan 2) for more details. Additional Experimental Results In this section, we present some additional experimental results to demonstrate the superiority of our proposed proximal ADM algorithm. Due to page limitations, we were not able to add these results in the submission. We extend our method for sensor network localization problem in the presence of laplace noise and uniform noise. We show our results of laplace noise in Figure 1- and uniform noise in Figure 7-, these experimental results strengthen our conclusions drawn in our submission. References [Attouch et al. 21] Attouch, H.; Bolte, J.; Redont, P.; and Soubeyran, A. 21. Proximal alternating minimization and projection methods for nonconvex problems: An approach based on the kurdyka-lojasiewicz inequality. Mathematics of Operations Research (MOR) 3(2): [Bolte, Sabach, and Teboulle 2] Bolte, J.; Sabach, S.; and Teboulle, M. 2. Proximal alternating linearized minimization for nonconvex and nonsmooth problems. Mathematical Programming (1-2): [Fazel, Hindi, and Boyd 23] Fazel, M.; Hindi, H.; and Boyd, S. P. 23. Log-det heuristic for matrix minimization with applications to hankel and euclidean ance matrices. In Proceedings of the American Control Conference, volume 3, 2. 1 [He and Yuan 2] He, B., and Yuan, X. 2. On the O(1/n) convergence rate of the douglas-rachford alternating direction method. SIAM Journal on Numerical Analysis (SINUM) (2): ,, [Lin, Liu, and Su 211] Lin, Z.; Liu, R.; and Su, Z Linearized alternating direction method with adaptive penalty for low- representation. In NIPS, 2., [Rockafellar, Wets, and Wets 199] Rockafellar, R. T.; Wets, R. J.-B.; and Wets, M Variational analysis. springe, berlin, heidelberg, new york, [Xu and Yin 213] Xu, Y., and Yin, W A block coordinate descent method for regularized multiconvex optimization with applications to nonnegative tensor factorization and completion. SIAM Journal on Imaging Sciences (SIIMS) (3):
6 7.e1.3e1 3.7e1 2.e1 3.9e 3.e 2.1e 1.1e Iteration Figure 1: Asymptotic behavior on minimum- sensor network localization problem in the presence of laplace noise. We plot the values of (blue) and (red) against the number of iterations, as well as how the sensors have been located at different stages of the process (1, 2, 3,, ) (a) FM, =.2, = (b) TAM, =.21, = (c) LPAM, =.19, = (d) LDHM, =.1, = (e) PDA, =.27, = 11. Figure 2: Performance comparison on 2d data in the presence of laplace noise (f) PADM, =., = (a) FM, =.32, = (b) TAM, =.29, = (c) LPAM, =.21, = (d) LDHM, =.32, = (e) PDA, =., = 9. Figure 3: Performance comparison on 3d data in the presence of laplace noise (f) PADM, =.21, = (b) and comparisons on 2d data (c) and comparisons on 3d data (d) and comparisons on 7d data Figure : Performance comparison with varying the number of sensor u in the presence of laplace noise (a) and comparisons on 2d data (b) and comparisons on 3d data (c) and comparisons on 7d data Figure : Performance comparison with varying the s in the presence of laplace noise (a) and comparisons on 2d data (b) and comparisons on 3d data (c) and comparisons on 7d data Figure : Performance comparison with varying the r in the presence of laplace noise.
7 7.1e1.e1 3.7e1 2.e1 3.e 3.e 2.1e 1.3e Iteration Figure 7: Asymptotic behavior on minimum- sensor network localization problem in the presence of uniform noise. We plot the values of (blue) and (red) against the number of iterations, as well as how the sensors have been located at different stages of the process (1, 2, 3,, ) (a) FM, =.39, = (b) TAM, =.3, = (c) LPAM, =.3, = 2... (d) LDHM, =.32, = (e) PDA, =., =. Figure : Performance comparison on 2d data in the presence of uniform noise... (f) PADM, =.31, = (a) FM, =., = (b) TAM, =.7, = (c) LPAM, =.9, = (d) LDHM, =., = (e) PDA, =., = 31. Figure 9: Performance comparison on 3d data in the presence of uniform noise (f) PADM, =.3, = (b) and comparisons on 2d data (c) and comparisons on 3d data (d) and comparisons on 7d data Figure 1: Performance comparison with varying the number of sensor u in the presence of uniform noise (a) and comparisons on 2d data (b) and comparisons on 3d data (c) and comparisons on 7d data Figure 11: Performance comparison with varying the s in the presence of uniform noise (a) and comparisons on 2d data (b) and comparisons on 3d data (c) and comparisons on 7d data Figure : Performance comparison with varying the r in the presence of uniform noise.
Block Coordinate Descent for Regularized Multi-convex Optimization
Block Coordinate Descent for Regularized Multi-convex Optimization Yangyang Xu and Wotao Yin CAAM Department, Rice University February 15, 2013 Multi-convex optimization Model definition Applications Outline
More informationContraction Methods for Convex Optimization and Monotone Variational Inequalities No.16
XVI - 1 Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16 A slightly changed ADMM for convex optimization with three separable operators Bingsheng He Department of
More informationI P IANO : I NERTIAL P ROXIMAL A LGORITHM FOR N ON -C ONVEX O PTIMIZATION
I P IANO : I NERTIAL P ROXIMAL A LGORITHM FOR N ON -C ONVEX O PTIMIZATION Peter Ochs University of Freiburg Germany 17.01.2017 joint work with: Thomas Brox and Thomas Pock c 2017 Peter Ochs ipiano c 1
More informationCoordinate Update Algorithm Short Course Operator Splitting
Coordinate Update Algorithm Short Course Operator Splitting Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 25 Operator splitting pipeline 1. Formulate a problem as 0 A(x) + B(x) with monotone operators
More informationConvergence of Bregman Alternating Direction Method with Multipliers for Nonconvex Composite Problems
Convergence of Bregman Alternating Direction Method with Multipliers for Nonconvex Composite Problems Fenghui Wang, Zongben Xu, and Hong-Kun Xu arxiv:40.8625v3 [math.oc] 5 Dec 204 Abstract The alternating
More informationA semi-algebraic look at first-order methods
splitting A semi-algebraic look at first-order Université de Toulouse / TSE Nesterov s 60th birthday, Les Houches, 2016 in large-scale first-order optimization splitting Start with a reasonable FOM (some
More informationOptimization methods
Lecture notes 3 February 8, 016 1 Introduction Optimization methods In these notes we provide an overview of a selection of optimization methods. We focus on methods which rely on first-order information,
More informationDouglas-Rachford splitting for nonconvex feasibility problems
Douglas-Rachford splitting for nonconvex feasibility problems Guoyin Li Ting Kei Pong Jan 3, 015 Abstract We adapt the Douglas-Rachford DR) splitting method to solve nonconvex feasibility problems by studying
More informationDual and primal-dual methods
ELE 538B: Large-Scale Optimization for Data Science Dual and primal-dual methods Yuxin Chen Princeton University, Spring 2018 Outline Dual proximal gradient method Primal-dual proximal gradient method
More informationA user s guide to Lojasiewicz/KL inequalities
Other A user s guide to Lojasiewicz/KL inequalities Toulouse School of Economics, Université Toulouse I SLRA, Grenoble, 2015 Motivations behind KL f : R n R smooth ẋ(t) = f (x(t)) or x k+1 = x k λ k f
More informationPrimal-dual Subgradient Method for Convex Problems with Functional Constraints
Primal-dual Subgradient Method for Convex Problems with Functional Constraints Yurii Nesterov, CORE/INMA (UCL) Workshop on embedded optimization EMBOPT2014 September 9, 2014 (Lucca) Yu. Nesterov Primal-dual
More informationBinary Optimization via Mathematical Programming with Equilibrium Constraints
1 Binary Optimization via Mathematical Programg with Equilibrium Constraints Ganzhao Yuan, Bernard Ghanem arxiv:108.05v [math.oc] Dec 017 Abstract Binary optimization is a central problem in mathematical
More informationShiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers
Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 9 Alternating Direction Method of Multipliers Shiqian Ma, MAT-258A: Numerical Optimization 2 Separable convex optimization a special case is min f(x)
More information2 Regularized Image Reconstruction for Compressive Imaging and Beyond
EE 367 / CS 448I Computational Imaging and Display Notes: Compressive Imaging and Regularized Image Reconstruction (lecture ) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement
More informationLINEARIZED AUGMENTED LAGRANGIAN AND ALTERNATING DIRECTION METHODS FOR NUCLEAR NORM MINIMIZATION
LINEARIZED AUGMENTED LAGRANGIAN AND ALTERNATING DIRECTION METHODS FOR NUCLEAR NORM MINIMIZATION JUNFENG YANG AND XIAOMING YUAN Abstract. The nuclear norm is widely used to induce low-rank solutions for
More informationOn the Convergence of Multi-Block Alternating Direction Method of Multipliers and Block Coordinate Descent Method
On the Convergence of Multi-Block Alternating Direction Method of Multipliers and Block Coordinate Descent Method Caihua Chen Min Li Xin Liu Yinyu Ye This version: September 9, 2015 Abstract The paper
More informationON THE GLOBAL AND LINEAR CONVERGENCE OF THE GENERALIZED ALTERNATING DIRECTION METHOD OF MULTIPLIERS
ON THE GLOBAL AND LINEAR CONVERGENCE OF THE GENERALIZED ALTERNATING DIRECTION METHOD OF MULTIPLIERS WEI DENG AND WOTAO YIN Abstract. The formulation min x,y f(x) + g(y) subject to Ax + By = b arises in
More informationSparse Optimization Lecture: Dual Methods, Part I
Sparse Optimization Lecture: Dual Methods, Part I Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know dual (sub)gradient iteration augmented l 1 iteration
More informationBeyond Heuristics: Applying Alternating Direction Method of Multipliers in Nonconvex Territory
Beyond Heuristics: Applying Alternating Direction Method of Multipliers in Nonconvex Territory Xin Liu(4Ð) State Key Laboratory of Scientific and Engineering Computing Institute of Computational Mathematics
More informationLinearized Alternating Direction Method with Adaptive Penalty for Low-Rank Representation
Linearized Alternating Direction Method with Adaptive Penalty for Low-Rank Representation Zhouchen Lin Visual Computing Group Microsoft Research Asia Risheng Liu Zhixun Su School of Mathematical Sciences
More informationSequential convex programming,: value function and convergence
Sequential convex programming,: value function and convergence Edouard Pauwels joint work with Jérôme Bolte Journées MODE Toulouse March 23 2016 1 / 16 Introduction Local search methods for finite dimensional
More informationAccelerated primal-dual methods for linearly constrained convex problems
Accelerated primal-dual methods for linearly constrained convex problems Yangyang Xu SIAM Conference on Optimization May 24, 2017 1 / 23 Accelerated proximal gradient For convex composite problem: minimize
More informationActive sets, steepest descent, and smooth approximation of functions
Active sets, steepest descent, and smooth approximation of functions Dmitriy Drusvyatskiy School of ORIE, Cornell University Joint work with Alex D. Ioffe (Technion), Martin Larsson (EPFL), and Adrian
More informationarxiv: v1 [math.oc] 23 May 2017
A DERANDOMIZED ALGORITHM FOR RP-ADMM WITH SYMMETRIC GAUSS-SEIDEL METHOD JINCHAO XU, KAILAI XU, AND YINYU YE arxiv:1705.08389v1 [math.oc] 23 May 2017 Abstract. For multi-block alternating direction method
More informationUses of duality. Geoff Gordon & Ryan Tibshirani Optimization /
Uses of duality Geoff Gordon & Ryan Tibshirani Optimization 10-725 / 36-725 1 Remember conjugate functions Given f : R n R, the function is called its conjugate f (y) = max x R n yt x f(x) Conjugates appear
More informationSolving large Semidefinite Programs - Part 1 and 2
Solving large Semidefinite Programs - Part 1 and 2 Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria F. Rendl, Singapore workshop 2006 p.1/34 Overview Limits of Interior
More informationContraction Methods for Convex Optimization and Monotone Variational Inequalities No.18
XVIII - 1 Contraction Methods for Convex Optimization and Monotone Variational Inequalities No18 Linearized alternating direction method with Gaussian back substitution for separable convex optimization
More informationALADIN An Algorithm for Distributed Non-Convex Optimization and Control
ALADIN An Algorithm for Distributed Non-Convex Optimization and Control Boris Houska, Yuning Jiang, Janick Frasch, Rien Quirynen, Dimitris Kouzoupis, Moritz Diehl ShanghaiTech University, University of
More informationOn the iterate convergence of descent methods for convex optimization
On the iterate convergence of descent methods for convex optimization Clovis C. Gonzaga March 1, 2014 Abstract We study the iterate convergence of strong descent algorithms applied to convex functions.
More informationA memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration
A memory gradient algorithm for l 2 -l 0 regularization with applications to image restoration E. Chouzenoux, A. Jezierska, J.-C. Pesquet and H. Talbot Université Paris-Est Lab. d Informatique Gaspard
More informationA Customized ADMM for Rank-Constrained Optimization Problems with Approximate Formulations
A Customized ADMM for Rank-Constrained Optimization Problems with Approximate Formulations Chuangchuang Sun and Ran Dai Abstract This paper proposes a customized Alternating Direction Method of Multipliers
More informationACCELERATED FIRST-ORDER PRIMAL-DUAL PROXIMAL METHODS FOR LINEARLY CONSTRAINED COMPOSITE CONVEX PROGRAMMING
ACCELERATED FIRST-ORDER PRIMAL-DUAL PROXIMAL METHODS FOR LINEARLY CONSTRAINED COMPOSITE CONVEX PROGRAMMING YANGYANG XU Abstract. Motivated by big data applications, first-order methods have been extremely
More informationCoordinate Update Algorithm Short Course Proximal Operators and Algorithms
Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 36 Why proximal? Newton s method: for C 2 -smooth, unconstrained problems allow
More informationEE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2
EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 2 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory April 5, 2012 Andre Tkacenko
More informationAN AUGMENTED LAGRANGIAN METHOD FOR l 1 -REGULARIZED OPTIMIZATION PROBLEMS WITH ORTHOGONALITY CONSTRAINTS
AN AUGMENTED LAGRANGIAN METHOD FOR l 1 -REGULARIZED OPTIMIZATION PROBLEMS WITH ORTHOGONALITY CONSTRAINTS WEIQIANG CHEN, HUI JI, AND YANFEI YOU Abstract. A class of l 1 -regularized optimization problems
More informationGLOBALLY CONVERGENT ACCELERATED PROXIMAL ALTERNATING MAXIMIZATION METHOD FOR L1-PRINCIPAL COMPONENT ANALYSIS. Peng Wang Huikang Liu Anthony Man-Cho So
GLOBALLY CONVERGENT ACCELERATED PROXIMAL ALTERNATING MAXIMIZATION METHOD FOR L-PRINCIPAL COMPONENT ANALYSIS Peng Wang Huikang Liu Anthony Man-Cho So Department of Systems Engineering and Engineering Management,
More informationTight Rates and Equivalence Results of Operator Splitting Schemes
Tight Rates and Equivalence Results of Operator Splitting Schemes Wotao Yin (UCLA Math) Workshop on Optimization for Modern Computing Joint w Damek Davis and Ming Yan UCLA CAM 14-51, 14-58, and 14-59 1
More informationSome Properties of the Augmented Lagrangian in Cone Constrained Optimization
MATHEMATICS OF OPERATIONS RESEARCH Vol. 29, No. 3, August 2004, pp. 479 491 issn 0364-765X eissn 1526-5471 04 2903 0479 informs doi 10.1287/moor.1040.0103 2004 INFORMS Some Properties of the Augmented
More informationMath 273a: Optimization Overview of First-Order Optimization Algorithms
Math 273a: Optimization Overview of First-Order Optimization Algorithms Wotao Yin Department of Mathematics, UCLA online discussions on piazza.com 1 / 9 Typical flow of numerical optimization Optimization
More informationMulti-stage convex relaxation approach for low-rank structured PSD matrix recovery
Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery Department of Mathematics & Risk Management Institute National University of Singapore (Based on a joint work with Shujun
More informationA Bregman alternating direction method of multipliers for sparse probabilistic Boolean network problem
A Bregman alternating direction method of multipliers for sparse probabilistic Boolean network problem Kangkang Deng, Zheng Peng Abstract: The main task of genetic regulatory networks is to construct a
More informationThe Multi-Path Utility Maximization Problem
The Multi-Path Utility Maximization Problem Xiaojun Lin and Ness B. Shroff School of Electrical and Computer Engineering Purdue University, West Lafayette, IN 47906 {linx,shroff}@ecn.purdue.edu Abstract
More information1 Computing with constraints
Notes for 2017-04-26 1 Computing with constraints Recall that our basic problem is minimize φ(x) s.t. x Ω where the feasible set Ω is defined by equality and inequality conditions Ω = {x R n : c i (x)
More informationA Majorize-Minimize subspace approach for l 2 -l 0 regularization with applications to image processing
A Majorize-Minimize subspace approach for l 2 -l 0 regularization with applications to image processing Emilie Chouzenoux emilie.chouzenoux@univ-mlv.fr Université Paris-Est Lab. d Informatique Gaspard
More informationMotivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:
CDS270 Maryam Fazel Lecture 2 Topics from Optimization and Duality Motivation network utility maximization (NUM) problem: consider a network with S sources (users), each sending one flow at rate x s, through
More informationA multi-stage convex relaxation approach to noisy structured low-rank matrix recovery
A multi-stage convex relaxation approach to noisy structured low-rank matrix recovery Shujun Bi, Shaohua Pan and Defeng Sun March 1, 2017 Abstract This paper concerns with a noisy structured low-rank matrix
More informationDual methods and ADMM. Barnabas Poczos & Ryan Tibshirani Convex Optimization /36-725
Dual methods and ADMM Barnabas Poczos & Ryan Tibshirani Convex Optimization 10-725/36-725 1 Given f : R n R, the function is called its conjugate Recall conjugate functions f (y) = max x R n yt x f(x)
More informationGENERALIZED second-order cone complementarity
Stochastic Generalized Complementarity Problems in Second-Order Cone: Box-Constrained Minimization Reformulation and Solving Methods Mei-Ju Luo and Yan Zhang Abstract In this paper, we reformulate the
More informationHYBRID JACOBIAN AND GAUSS SEIDEL PROXIMAL BLOCK COORDINATE UPDATE METHODS FOR LINEARLY CONSTRAINED CONVEX PROGRAMMING
SIAM J. OPTIM. Vol. 8, No. 1, pp. 646 670 c 018 Society for Industrial and Applied Mathematics HYBRID JACOBIAN AND GAUSS SEIDEL PROXIMAL BLOCK COORDINATE UPDATE METHODS FOR LINEARLY CONSTRAINED CONVEX
More informationProjection methods to solve SDP
Projection methods to solve SDP Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria F. Rendl, Oberwolfach Seminar, May 2010 p.1/32 Overview Augmented Primal-Dual Method
More informationMatrix Rank Minimization with Applications
Matrix Rank Minimization with Applications Maryam Fazel Haitham Hindi Stephen Boyd Information Systems Lab Electrical Engineering Department Stanford University 8/2001 ACC 01 Outline Rank Minimization
More informationFrom error bounds to the complexity of first-order descent methods for convex functions
From error bounds to the complexity of first-order descent methods for convex functions Nguyen Trong Phong-TSE Joint work with Jérôme Bolte, Juan Peypouquet, Bruce Suter. Toulouse, 23-25, March, 2016 Journées
More informationSparse Approximation via Penalty Decomposition Methods
Sparse Approximation via Penalty Decomposition Methods Zhaosong Lu Yong Zhang February 19, 2012 Abstract In this paper we consider sparse approximation problems, that is, general l 0 minimization problems
More informationLecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem
Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem Michael Patriksson 0-0 The Relaxation Theorem 1 Problem: find f := infimum f(x), x subject to x S, (1a) (1b) where f : R n R
More informationAn inertial forward-backward algorithm for the minimization of the sum of two nonconvex functions
An inertial forward-backward algorithm for the minimization of the sum of two nonconvex functions Radu Ioan Boţ Ernö Robert Csetnek Szilárd Csaba László October, 1 Abstract. We propose a forward-backward
More informationLagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)
Lagrange Duality Daniel P. Palomar Hong Kong University of Science and Technology (HKUST) ELEC5470 - Convex Optimization Fall 2017-18, HKUST, Hong Kong Outline of Lecture Lagrangian Dual function Dual
More informationSplitting methods for decomposing separable convex programs
Splitting methods for decomposing separable convex programs Philippe Mahey LIMOS - ISIMA - Université Blaise Pascal PGMO, ENSTA 2013 October 4, 2013 1 / 30 Plan 1 Max Monotone Operators Proximal techniques
More informationRank minimization via the γ 2 norm
Rank minimization via the γ 2 norm Troy Lee Columbia University Adi Shraibman Weizmann Institute Rank Minimization Problem Consider the following problem min X rank(x) A i, X b i for i = 1,..., k Arises
More informationRobust Principal Component Analysis Based on Low-Rank and Block-Sparse Matrix Decomposition
Robust Principal Component Analysis Based on Low-Rank and Block-Sparse Matrix Decomposition Gongguo Tang and Arye Nehorai Department of Electrical and Systems Engineering Washington University in St Louis
More informationAlgorithms for Nonsmooth Optimization
Algorithms for Nonsmooth Optimization Frank E. Curtis, Lehigh University presented at Center for Optimization and Statistical Learning, Northwestern University 2 March 2018 Algorithms for Nonsmooth Optimization
More informationFast Nonnegative Matrix Factorization with Rank-one ADMM
Fast Nonnegative Matrix Factorization with Rank-one Dongjin Song, David A. Meyer, Martin Renqiang Min, Department of ECE, UCSD, La Jolla, CA, 9093-0409 dosong@ucsd.edu Department of Mathematics, UCSD,
More informationProximal Alternating Linearized Minimization for Nonconvex and Nonsmooth Problems
Proximal Alternating Linearized Minimization for Nonconvex and Nonsmooth Problems Jérôme Bolte Shoham Sabach Marc Teboulle Abstract We introduce a proximal alternating linearized minimization PALM) algorithm
More informationAn Optimization-based Approach to Decentralized Assignability
2016 American Control Conference (ACC) Boston Marriott Copley Place July 6-8, 2016 Boston, MA, USA An Optimization-based Approach to Decentralized Assignability Alborz Alavian and Michael Rotkowitz Abstract
More informationminimize x subject to (x 2)(x 4) u,
Math 6366/6367: Optimization and Variational Methods Sample Preliminary Exam Questions 1. Suppose that f : [, L] R is a C 2 -function with f () on (, L) and that you have explicit formulae for
More informationTMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM
TMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM H. E. Krogstad, IMF, Spring 2012 Karush-Kuhn-Tucker (KKT) Theorem is the most central theorem in constrained optimization, and since the proof is scattered
More informationA convergence result for an Outer Approximation Scheme
A convergence result for an Outer Approximation Scheme R. S. Burachik Engenharia de Sistemas e Computação, COPPE-UFRJ, CP 68511, Rio de Janeiro, RJ, CEP 21941-972, Brazil regi@cos.ufrj.br J. O. Lopes Departamento
More informationRelaxations and Randomized Methods for Nonconvex QCQPs
Relaxations and Randomized Methods for Nonconvex QCQPs Alexandre d Aspremont, Stephen Boyd EE392o, Stanford University Autumn, 2003 Introduction While some special classes of nonconvex problems can be
More informationMA 575 Linear Models: Cedric E. Ginestet, Boston University Regularization: Ridge Regression and Lasso Week 14, Lecture 2
MA 575 Linear Models: Cedric E. Ginestet, Boston University Regularization: Ridge Regression and Lasso Week 14, Lecture 2 1 Ridge Regression Ridge regression and the Lasso are two forms of regularized
More informationResearch Article Solving the Matrix Nearness Problem in the Maximum Norm by Applying a Projection and Contraction Method
Advances in Operations Research Volume 01, Article ID 357954, 15 pages doi:10.1155/01/357954 Research Article Solving the Matrix Nearness Problem in the Maximum Norm by Applying a Projection and Contraction
More informationarxiv: v4 [math.oc] 5 Jan 2016
Restarted SGD: Beating SGD without Smoothness and/or Strong Convexity arxiv:151.03107v4 [math.oc] 5 Jan 016 Tianbao Yang, Qihang Lin Department of Computer Science Department of Management Sciences The
More informationLinearized Alternating Direction Method: Two Blocks and Multiple Blocks. Zhouchen Lin 林宙辰北京大学
Linearized Alternating Direction Method: Two Blocks and Multiple Blocks Zhouchen Lin 林宙辰北京大学 Dec. 3, 014 Outline Alternating Direction Method (ADM) Linearized Alternating Direction Method (LADM) Two Blocks
More informationELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications
ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications Professor M. Chiang Electrical Engineering Department, Princeton University March
More informationA Unified Approach to Proximal Algorithms using Bregman Distance
A Unified Approach to Proximal Algorithms using Bregman Distance Yi Zhou a,, Yingbin Liang a, Lixin Shen b a Department of Electrical Engineering and Computer Science, Syracuse University b Department
More informationSolving Dual Problems
Lecture 20 Solving Dual Problems We consider a constrained problem where, in addition to the constraint set X, there are also inequality and linear equality constraints. Specifically the minimization problem
More informationLecture 7: Positive Semidefinite Matrices
Lecture 7: Positive Semidefinite Matrices Rajat Mittal IIT Kanpur The main aim of this lecture note is to prepare your background for semidefinite programming. We have already seen some linear algebra.
More informationContraction Methods for Convex Optimization and Monotone Variational Inequalities No.11
XI - 1 Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.11 Alternating direction methods of multipliers for separable convex programming Bingsheng He Department of Mathematics
More informationarxiv: v5 [cs.na] 22 Mar 2018
Iteratively Linearized Reweighted Alternating Direction Method of Multipliers for a Class of Nonconvex Problems Tao Sun Hao Jiang Lizhi Cheng Wei Zhu arxiv:1709.00483v5 [cs.na] 22 Mar 2018 March 26, 2018
More informationl 0 TV: A Sparse Optimization Method for Impulse Noise Image Restoration
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 1 l 0 TV: A Sparse Optimization Method for Impulse Noise Image Restoration Ganzhao Yuan, Bernard Ghanem arxiv:1802.09879v2 [cs.na] 28 Dec
More informationKaisa Joki Adil M. Bagirov Napsu Karmitsa Marko M. Mäkelä. New Proximal Bundle Method for Nonsmooth DC Optimization
Kaisa Joki Adil M. Bagirov Napsu Karmitsa Marko M. Mäkelä New Proximal Bundle Method for Nonsmooth DC Optimization TUCS Technical Report No 1130, February 2015 New Proximal Bundle Method for Nonsmooth
More informationA Continuation Approach Using NCP Function for Solving Max-Cut Problem
A Continuation Approach Using NCP Function for Solving Max-Cut Problem Xu Fengmin Xu Chengxian Ren Jiuquan Abstract A continuous approach using NCP function for approximating the solution of the max-cut
More informationRobust PCA. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng
Robust PCA CS5240 Theoretical Foundations in Multimedia Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (NUS) Robust PCA 1 / 52 Previously...
More informationConvex Optimization Algorithms for Machine Learning in 10 Slides
Convex Optimization Algorithms for Machine Learning in 10 Slides Presenter: Jul. 15. 2015 Outline 1 Quadratic Problem Linear System 2 Smooth Problem Newton-CG 3 Composite Problem Proximal-Newton-CD 4 Non-smooth,
More informationPreliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012
Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.
More informationTame variational analysis
Tame variational analysis Dmitriy Drusvyatskiy Mathematics, University of Washington Joint work with Daniilidis (Chile), Ioffe (Technion), and Lewis (Cornell) May 19, 2015 Theme: Semi-algebraic geometry
More informationEE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6)
EE 367 / CS 448I Computational Imaging and Display Notes: Image Deconvolution (lecture 6) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement to the material discussed in
More informationCertifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering
Certifying the Global Optimality of Graph Cuts via Semidefinite Programming: A Theoretic Guarantee for Spectral Clustering Shuyang Ling Courant Institute of Mathematical Sciences, NYU Aug 13, 2018 Joint
More information5 Handling Constraints
5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest
More informationSequential Quadratic Programming Method for Nonlinear Second-Order Cone Programming Problems. Hirokazu KATO
Sequential Quadratic Programming Method for Nonlinear Second-Order Cone Programming Problems Guidance Professor Masao FUKUSHIMA Hirokazu KATO 2004 Graduate Course in Department of Applied Mathematics and
More informationKey words. first order methods, trust region subproblem, optimality conditions, global optimum
GLOBALLY SOLVING THE TRUST REGION SUBPROBLEM USING SIMPLE FIRST-ORDER METHODS AMIR BECK AND YAKOV VAISBOURD Abstract. We consider the trust region subproblem which is given by a minimization of a quadratic,
More informationOptimizing Nonconvex Finite Sums by a Proximal Primal-Dual Method
Optimizing Nonconvex Finite Sums by a Proximal Primal-Dual Method Davood Hajinezhad Iowa State University Davood Hajinezhad Optimizing Nonconvex Finite Sums by a Proximal Primal-Dual Method 1 / 35 Co-Authors
More informationNumerical optimization. Numerical optimization. Longest Shortest where Maximal Minimal. Fastest. Largest. Optimization problems
1 Numerical optimization Alexander & Michael Bronstein, 2006-2009 Michael Bronstein, 2010 tosca.cs.technion.ac.il/book Numerical optimization 048921 Advanced topics in vision Processing and Analysis of
More information14 Singular Value Decomposition
14 Singular Value Decomposition For any high-dimensional data analysis, one s first thought should often be: can I use an SVD? The singular value decomposition is an invaluable analysis tool for dealing
More informationA STABILIZED SQP METHOD: GLOBAL CONVERGENCE
A STABILIZED SQP METHOD: GLOBAL CONVERGENCE Philip E. Gill Vyacheslav Kungurtsev Daniel P. Robinson UCSD Center for Computational Mathematics Technical Report CCoM-13-4 Revised July 18, 2014, June 23,
More informationA Multilevel Proximal Algorithm for Large Scale Composite Convex Optimization
A Multilevel Proximal Algorithm for Large Scale Composite Convex Optimization Panos Parpas Department of Computing Imperial College London www.doc.ic.ac.uk/ pp500 p.parpas@imperial.ac.uk jointly with D.V.
More informationWe describe the generalization of Hazan s algorithm for symmetric programming
ON HAZAN S ALGORITHM FOR SYMMETRIC PROGRAMMING PROBLEMS L. FAYBUSOVICH Abstract. problems We describe the generalization of Hazan s algorithm for symmetric programming Key words. Symmetric programming,
More informationarxiv: v2 [math.oc] 21 Nov 2017
Unifying abstract inexact convergence theorems and block coordinate variable metric ipiano arxiv:1602.07283v2 [math.oc] 21 Nov 2017 Peter Ochs Mathematical Optimization Group Saarland University Germany
More informationConvergence rates for an inertial algorithm of gradient type associated to a smooth nonconvex minimization
Convergence rates for an inertial algorithm of gradient type associated to a smooth nonconvex minimization Szilárd Csaba László November, 08 Abstract. We investigate an inertial algorithm of gradient type
More informationConstrained Optimization and Lagrangian Duality
CIS 520: Machine Learning Oct 02, 2017 Constrained Optimization and Lagrangian Duality Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may or may
More informationConvergence of Cubic Regularization for Nonconvex Optimization under KŁ Property
Convergence of Cubic Regularization for Nonconvex Optimization under KŁ Property Yi Zhou Department of ECE The Ohio State University zhou.1172@osu.edu Zhe Wang Department of ECE The Ohio State University
More informationOn the convergence of a regularized Jacobi algorithm for convex optimization
On the convergence of a regularized Jacobi algorithm for convex optimization Goran Banjac, Kostas Margellos, and Paul J. Goulart Abstract In this paper we consider the regularized version of the Jacobi
More information