The Optimization Work of Paul Tseng

Size: px
Start display at page:

Download "The Optimization Work of Paul Tseng"

Transcription

1 The Optimization Work of Paul Tseng Zhi-Quan Luo University of Minnesota May 23, 2010

2 Paul Y. Tseng ( )

3 Overview Paul s thesis work Paul s work on algorithm analysis (first order methods) Paul s work on applications (QCQP/SDP relaxations in wireless communication)

4 Paul s Optimization Work Central issues in optimization: analysis, algorithms, software, applications Paul made fundamental contributions to all four aspects of optimization Paul was a leader in the field of optimization publishing over 120 journal articles Editorial positions in leading optimization/computational Math journals Invited, plenary talks in various conferences the impact of his work still being felt

5 Thesis work Ph.D. Thesis: Relaxation Methods for Monotropic Programming Problems, MIT, Supervisor: Dimitri Bertsekas Dealt with Gauss-Seidel-type relaxation methods for general monotropic programming, problems with convex, separable objective, linear constraints f(x) = i f i(x i ), X = a polytope. Included are algorithms for several important special cases: - network flow problems with linear and convex cost functions; - network flow problems with and without gains; - standard linear programming problems; etc. Wrote very popular codes (with Bertsekas) called RELAX,ERELAX for network flows

6 Thesis work Received 2nd place for George Nicholson Student Paper Competition in 1985 Published 7 journal papers on thesis work Another 9 journal papers on post-thesis work with Bertsekas

7 Coordinate Descent for QP In the next 20+ years, Paul worked on many important theoretical and application problems in optimization First to establish linear convergence of matrix splitting method (a long standing open question, L.-Tseng 91) First to show the convergence of affine scaling algorithm for linear programming (Tseng-L. 92) Analysis of first order methods for smooth/nonsmooth optimization Interior point methods and smoothing methods for LP/QP/LCP/... Applications in image denoising, wireless communication, statistics, compressive sensing, etc

8 Coordinate Descent for QP Consider a convex quadratic minimization problem over R n + minimize 1 2 xt Ax + b T x subject to x X (1) If X = R, we solve (1) by a simple coordinate descent (Gauss-Seidel type) method: write A = A low + D + A upp 0 and iterate (A low + D)x r+1 + A upp x r + b = 0 x r+1 = (A low + D) 1 (A upp x r + b). If X = R n +, the coordinate descent algorithm becomes x r+1 = [ x r+1 (A low + D)x r+1 A upp x r b ] +. where [u] + = max{0,u}; related to SOR, block SOR.

9 Coordinate Descent (CD) for QP Consider a convex quadratic minimization problem over R n + minimize 1 2 xt Ax + b T x subject to x X CD has many applications in image processing, MRI,... Convergence was known for symmetric A 0, with unique optimal solution Convergence for the case symmetric A 0 was unresolved for many years..., unbounded optimal solution set. Studied by Hildreth, Cryer, Mangasarian, Pang,... in various contexts since 1950 s.

10 Matrix Splitting Algorithm Given a symmetric A 0, consider a splitting A = B + C with B C 0. Optimality condition: minimize f(x) = 1 2 xt Ax + b T x subject to x X = R n + x = Proj X [x Ax b], with Proj X [ ] = projection to X Optimal solution set X = {x x = Proj X [x f(x)]}. Matrix splitting algorithm: x r+1 = Proj R n + [ x r+1 Bx r+1 Cx r b ], r = 1,2,... CD is a special case Linear convergence was settled in 1989 (L.-Tseng 91).

11 Convergence Analysis of Matrix Splitting Algorithm Key proof steps Sufficient decrease: for some c 1 0. Cost to go estimate: f(x r ) f(x r+1 ) c 1 x r+1 x r 2, f(x r ) f c 2 φ(x r ) 2, where c 2 > 0 and φ(x) := min x X x x Local error bound: φ(x) c 3 x Proj X [x f(x)], for all x such that x Proj X [x f(x)] c 4 for some c 3, c 4 > 0.

12 Convergence Analysis of Matrix Splitting Algorithm Comments from leading researchers... The notable iterate convergence recently established by Luo and Tseng for exact subproblem solution is extended here to inexact subproblem solution for a symmetric matrix splitting., Olvi Managasarian, SIAM J. Optimization, Vol. 1, pp , Luo and Tseng [13] were the first to provide an affirmative answer to this outstanding question; but their proof was not easy to follow. Shortly afterwards, Mangasarian established a similar result by assuming a symmetric splitting but allowing inexact solutions in the subproblems., Jong-Shi Pang, Math. Prog., Vol. 79, pp , 1997.

13 Many Extensions Let X be a polyhedral set, g( ) be strictly convex. Consider minimize f(x) = g(kx) + c T x subject to x X The three key steps remain valid for a broad class of algorithms x r+1 = Proj X [x r α r f(x r ) + e r ], r = 1,2,... where step size α r > 0 can be Armijo rule, or Goldstein, Levin-Polyak s rule, etc e r satisfies e r c 5 x r+1 x r for some c 5 > 0. Includes many known algorithms: extra-gradient method, proximal point method, coordinate descent method, matrix splitting method Convergence result: the iterates converge linearly to an optimal solution in X. (L.-Tseng 93,94,95)

14 Extensions to Non-monotone Affine Equilibrium Problems Consider symmetric affine variational inequality problem (A = A T ) find a x such that Ax + b,x x 0, for all x X. includes linearly constrained non-convex QP solution set X := {x x = Proj X [x Ax b]} may be unbounded or disconnected Convergence result: the iterates converge linearly to an optimal solution in X. (L.-Tseng 93)

15 Dual Coordinate Ascent Consider a convex minimization problem min f(x), s.t. Ax b Assume feasibility X := {x Ax b}. Its dual is maxf ( A T y) b T y, s.t. y 0 where f is the Legendre transform of f. Apply CD to the dual corresponds to row action algorithm, successive projection to convex sets (POCS), etc Convergence result: If X, then primal/dual iterates converge linearly to an primal/dual optimal solution pair. (L.-Tseng 93)

16 Row Action/Projection to Convex Set (POCS) Method D x 0 x 1 x 2 x 3 x C

17 Incremental Gradient Method If X =, we can still use row action or POCS type algorithms. min f 1 (x) + f 2 (x) + + f m (x) (2) Incremental gradient method: x r+ i m = x r+ i m α r f i (x r+ i m ), i = 1,2,...,m, r = 1,2,... back propagation for neural nets,... reduces to dual coordinate ascent for linear feasibility problem applications in distributed optimization, sensor nets Nedic/Bertsekas, Mangasarian, Solodov, Lewis,... Convergence result: If r αr =, r (αr ) 2 <, then the iterates x r converges to a stationary point of (2). (L.-Tseng 94)

18 Sparse Signal Reconstruction, Compressive Sensing Suppose x is sparse. Given (A,b), find the sparsest point: ˆx = argmin x x 0, subject to Ax = b (3) where x 0 = number of nonzero entries in x From a hard combinatorial problem to convex optimization: ˆx = argmin x x 1, subject to Ax = b (4) Motivation: 1-norm is sparsity promoting Basis pursuit (Donoho et al 98) Many variants; e.g., Ax b 2 σ for noisy b Greedy algorithms (e.g., Tropp-Gilbert 05,...) Key question: when are (3) and (4) equivalent?

19 Compressive Sensing: Exact Recovery via Convex Optimization Assume Ψ = I. Let x be the true sparse solution and ˆx = argmin x x 1, subject to: Ax = A x. Theorem: (Candes-Tao 2005, Candes-Romberg-Tao 2005) If A R m n is iid standard normal and suppose x 0 O(m/[1 + log(n/m)]) < m < n. Then, with high probability (WHP) ˆx = x.

20 Magnetic Resonance Imaging MRI Scan = Fourier Coefficients = Images Is it possible to reduce the time to scan and reconstruct images? Compressive sensing: reconstruct image from subset of the Fourier coefficients

21 Compressive Sensing for Magnetic Resonance Imaging Pick 25% coefficients at random (with bias) Reconstruct image from the 25% coefficients

22 Compressive Sensing for Magnetic Resonance Imaging Use 1/4 of fourier Coefficients. Figure: Original vs. Reconstructed Image (courtesy of Y. Zhang, Rice University)

23 Convex Formulations min Ψx 1, subject to: Ax = b x C n min x 1, subject to: Ax = b; x 0 x R n min Ψx x C n ρ Ax b 2 2 min Ax x R b 2 n 2, subject to: x 1 ν.

24 Convex Formulations Simple and convex problems, but involve Large and dense matrices real-time processing requirement Standard (simplex, interior-point) methods not suitable First-order algorithms can be efficient: involve matrix multiplications using A or A T Fast transforms and structures help Sparsity reduces computational complexity

25 Coordinated Descent for Non-smooth Optimization Smooth case Non-smooth case CD method for non-smooth problems can get stuck at non-interesting points

26 Coordinated Descent for Non-smooth Optimization Consider the nonsmooth formulation: min x R n τ x 1 + Ax b 2 2 Coordinate descent algorithm (CD): Iteratively and cyclically minimize wrt each variable Soft thresholding: let x + = arg min τ y + (y y R x)2, then x + τ, x τ x + = 0, τ x τ x τ x τ Convergence: In general, the CD algorithm does not converge for nonsmooth optimization For nonsmoothness on the variables directly, e.g., x 1, the CD algorithm does converge linearly (Tseng, 2000).

27 Projected Landweber Algorithm Consider the formulation: min Kx x R c 2 n 2, s.t. x 0. Optimality condition: x = Proj R n + [x αk T (Kx c)] where α > 0. Projected Landweber Algorithm: x r+1 = Proj R n + [x r αk T (Kx r c)], r = 0,1,2,... Convergence: If K is full column rank, then x k converges linearly to the unique optimum. (Bertsekas, 1999) If K 0 (entry-wise) and each row is nonzero, then the iterates {x r } converge linearly. (Johansson-Elfving-Kozlovc-Censor-Forssen-Granlund, 2006)

28 Projected Landweber Algorithm Consider the formulation: Optimality condition: min Kx x R c 2 2, subject to: x 1 ν, n x = Proj X [x αk T (Kx c)] where α > 0 and X = {x x 1 ν} Projection to X is simple: iterative thresholding Projected Landweber Algorithm: x r+1 = Proj X [x r α r K T (Kx r c)], r = 0,1,2,... Convergence (Daubechies-Fornasier-Loris, 2008; Bredies-Lorenz, 2008;...): {x k } converges in Hilbert space; if K has full column rank, then linear convergence

29 Recall A General Convergence Result Let X = {x Cx d} be an empty polyhedral set. Consider min f(x) = g(kx) + x X pt x where g is strictly convex, g is Lipschitz continuous. Optimality: x = Proj X [x α f(x)] Gradient projection algorithm: x k+1 = Proj X [x k α k f(x k )] Convergence (L.-Tseng, 1991): For various stepsize rules (including Armijo rule), x k x linearly, for some optimal solution x. If g( ) = c 2, p = 0 and X = {x x 1 ν}, then we recover the Projected Landweber algorithm.

30 Extensions: Proximal Gradient Methods Consider min f 1 (x) + f 2 (x), s.t. x R (5) where f 1 is convex, nonsmooth, while f 2 is convex and smooth. Prox operator: Prox f1 (x) = arg min f 1 (y) + 1 x y 2 2 If f 1 (x) = 1 X (x) (indicator function) for some convex set X, then problem (5) becomes min f 2(x) and X 1 Prox f1 (x) = arg min y X 2 x y 2 = Proj X [x]. Optimality: x = Prox f1 (x α f 2 (x)), α > 0. Proximal gradient (proximal splitting) method: x r+1 = Prox f1 (x r α r f 2 (x r ))

31 Proximal Gradient Method D x 0 y 0 x 1 x 2 x 3 y 1 y 2 y 3 y 4 x 4 C x Paul s formulation: x r+1 = Argmin{ f 2 (x r ),x x r + f 1 (x) + 1 x 2α r x xr 2 } 2 can be replaced by Bregman distance Also related to Mirror-descent, Augmented Lagrangian

32 Analysis of Proximal Gradient Method Consider f 1 (x) = x 1, f 2 (x) = 1 Kx c 2 2 Then Prox τf1 (x) = soft thresholding to [ τ,τ] n In general, consider f 1 (x) = J J x J 2, f 2 (x) = g(kx) where J is a partition of coordinates, g( ) is strongly convex. Then Prox τf1 (x) can still be computed efficiently (e.g., in closed form). When g( ) = b 2, this corresponds to the Group Lasso algorithm in statistics. Also includes the Total Variation (TV) minimization in image processing.

33 Linear Convergence of Proximal Gradient Method Proximal gradient method consists of two steps: gradient descent wrt f 2 (x), then soft thresholding. The three key proof steps: sufficient decrease, cost to go estimate, local error bound, as established by L. and Tseng in early 90 s still hold true for proximal gradient method. Convergence (Tseng, 2009): For any matrix K, linear convergence of iterates to an optimal solution.

34 Application Work: Transmit Beamforming Downlink transmission: basestation has K antennas; m receivers n multicast groups, {G 1,..., G n }, G k G l = k G k = {1,..., m}, G k := G k, n k=1 G k = m. w k : beamforming vector for G k s k : signal sent to group G k transmitted signal: s(t) = n s k (t)w k k=1

35 Transmit Beamforming assume each receiver has one antenna, with channel vector h i For ULA, LoS, far field, ) h i = (1, e jφi,..., e j(k 1)φi where φ i = 2π d λ sinθ i; Vandermonde. signal at receiver i G k : s h i +v i = s k w k h i + l =k s l w l h i + v i SINR for receiver i G k w k h i 2 l =k w l h i 2 + σi 2

36 Problem Description: Transmit Beamforming Transmit beamforming problem: minimize transmit power, subject to QoS constraint for each receiver in each group. minimize subject to minimize subject to n w k 2 k=1 w k h i 2 l =k w l h c i, i G k, k {1,..., n} i 2 + σi 2 n w k 2 k=1 w k h i 2 c i w l h i 2 c i σi 2, i G k, k l =k separable homogeneous QCQP How effective is SDP relaxation for this problem?

37 Empirical Behavior Measured VDSL channel data by France Telecom R&D; 17 measured channel scenarios, 28 loops, 300m, Mhz; Compared SDP solution vs no precoding (ignore crosstalk) SDP yields nearly doubling of minimum received signal power SDP relaxation is tight in over 50% of instances. (SDL 05) SDP solved by the basestation Question: a theoretical performance bound?

38 Problem Description: Separable Homogeneous QCQP Let C i, A ij be Hermitian matrices, and H = R or C. n min wi H K i i=1 w i Ciwi s. t. w 1 A11w1 + w 2 A12w2 + + w na 1nw n b 1 w 1 A21w1 + w 2 A22w2 + + w na 2nw n b 2... NP-hard w 1 Am1w1 + w 2 Am2w2 + + w na mnw n b m n minimize Tr(C iw i) i=1 subject to Tr(A 11W 1 + A 12W A 1nW n) b 1... SDP relaxation Tr(A m1w 1 + A m2w A mnw n) b m W i 0. Question: How good is the SDP relaxation?

39 Special Case: Vandermonde h i For uniform linear array, line of sight and far-field propagation, h i is Vandermonde h i = ( 1, e jφi,..., e j(k 1)φi ) := a(φi ). In this case, the SDP relaxation of the separable homogeneous QCQP always has a rank-1 solution. (m > n allowed) minimize subject to minimize subject to n w k 2 k=1 w k h i 2 c i w l h i 2 c i σi 2, i G k, k {1,..., n} l =k n Tr(W k ) k=1 Tr(W k h ih i ) ci Tr(W l h ih i ) ciσ2 i, i G k, k, W k 0. l k

40 Why? Suppose W i = l w il w il is optimal. Goal: find a rank-1 W i, s.t. Tr(W i h jh j ) = Tr( W i h jh j ) and Tr( W i ) = Tr(W i ). Recall a(φ) = (1, e jφ,..., e j(k 1)φ ). Write Tr(W i h jh j ) = l a(φ j) w il 2 The trig polynomial f i (φ) := l a(φ) w il 2 0, φ. f i (φ) = a(φ) w i 2 for some w i. (Riesz-Fejer) Integrating out φ in a(φ) w i 2 = l a(φ) w il 2 yields Tr( w i ( w i ) ) = w i 2 = l w il 2 = l Tr(w il (w il ) ). Tr(W i h jh j ) = l a(φ j) w il 2 = f(φ j ) = Tr( w i ( w i ) h j h j ) Thus, { W i = w i ( w i ) } n i=1 is feasible and has the same objective value as {Wi }n i=1. a rank-1 solution.

41 Worst-case SDP Approximation Performance In general, SDP relaxation for separable homogeneous QCQP is not tight. Example: For any M > 0, consider υ qp := min x 2 + y 2 s.t. y 2 1, x 2 Mxy 1, x 2 + Mxy 1. Notice that the last two constraints of QCQP imply x 2 M x y + 1, x 2 1, which, in light of the constraints y 2 1, further imply x 2 M + 1. υ qp M + 2. However, for the SDP relaxation υ sdp := min X(11) + X(22) s.t.y (11) 1, X(11) MX(12) 1, X(22) + MX(12) 1,X 0. X = I is clearly a feasible solution, implying υ sdp 2. the performance ratio υ qp /υ sdp is at least (M + 2)/2, which can be arbitrarily large.

42 Formulation is dual to the max norm over ellipsoids problem considered by NRT 99. Special Case: Single Group Multicasting When n = 1 (single group), the transmit beamforming problem becomes υ qp := min w H n w 2 s.t. w H iw := l I i h l w 2 1, i = 1,..., m, h l 0 H n (H = C or R), I 1 I m = {1,..., M} z = x + iy (x, y R n ), z = x T iy T I i : number of antennas at receiver i. w = the least norm vector in the exteriors of ellipsoids.

43 SDP Relaxation Finding a global minimum of QP is NP-hard (reduction from PARTITION). Approximate QP by an easy convex optimization problem, a semidefinite program (SDP) relaxation (Lovász 91, Shor 87). Let H i = l I i h l h l. The SDP relaxation is υ sdp := min Tr(W) s.t. Tr(H i W) 1, i = 1,..., m, W 0. Then 0 υ sdp υ qp? Cυsdp (C 1)

44 Work and Hike

45 Work and Hike

46 Approximation Upper & Lower Bounds Theorem 1 (LSTZ 07): υ qp Cυ sdp where 1 2π 2m2 C 27 π m2 if H = R 1 m C 8m if H = C 2(3.6π) 2

47 Numerical experience Simulation with randomly generated h l (m = 8, n = 4) shows that both the mean and the maximum of the upper bound υ qp υ sdp are lower in the H = C case than the H = R case.

48 Summary: SDP Relaxation for Nonconvex QP A i,āi 0, i = 0,1, 2,..., m B j 0 indefinite, j = 0, 1,2,..., d R, d = 0 R, d = 1 or C, d = 0,1 R or C, d 2 minw H A 0w s.t. w H A iw 1, w H B jw 1 maxw H B 0w s.t. w H A iw 1, w H B jw 1 max min 1 i m s.t. w 2 P w H A iw whāiw + σ2 Θ(m 2 ) Θ(m) Θ(log 1 m) Θ(log 1 m) Θ(m 2 ) Θ(m) N.A. Blue: NRT 99, Red: LSTZ 07, CLC 07, HLNZ 07,SYZ 08

49 The Optimization Work of Paul Tseng The Paul s work is very broad in scope has tremendous depth Only able to give a glimpse of Paul s overall contribution... Still has several unpublished work, unfinished collaborations... Sadly missed by all...

50 Thank You! Questions/Comments?

SDP Relaxation of Quadratic Optimization with Few Homogeneous Quadratic Constraints

SDP Relaxation of Quadratic Optimization with Few Homogeneous Quadratic Constraints SDP Relaxation of Quadratic Optimization with Few Homogeneous Quadratic Constraints Paul Tseng Mathematics, University of Washington Seattle LIDS, MIT March 22, 2006 (Joint work with Z.-Q. Luo, N. D. Sidiropoulos,

More information

Optimization in Wireless Communication

Optimization in Wireless Communication Zhi-Quan (Tom) Luo Department of Electrical and Computer Engineering University of Minnesota 200 Union Street SE Minneapolis, MN 55455 2007 NSF Workshop Challenges Optimization problems from wireless applications

More information

Large-Scale L1-Related Minimization in Compressive Sensing and Beyond

Large-Scale L1-Related Minimization in Compressive Sensing and Beyond Large-Scale L1-Related Minimization in Compressive Sensing and Beyond Yin Zhang Department of Computational and Applied Mathematics Rice University, Houston, Texas, U.S.A. Arizona State University March

More information

Recent Developments of Alternating Direction Method of Multipliers with Multi-Block Variables

Recent Developments of Alternating Direction Method of Multipliers with Multi-Block Variables Recent Developments of Alternating Direction Method of Multipliers with Multi-Block Variables Department of Systems Engineering and Engineering Management The Chinese University of Hong Kong 2014 Workshop

More information

Accelerated Proximal Gradient Methods for Convex Optimization

Accelerated Proximal Gradient Methods for Convex Optimization Accelerated Proximal Gradient Methods for Convex Optimization Paul Tseng Mathematics, University of Washington Seattle MOPTA, University of Guelph August 18, 2008 ACCELERATED PROXIMAL GRADIENT METHODS

More information

Minimizing the Difference of L 1 and L 2 Norms with Applications

Minimizing the Difference of L 1 and L 2 Norms with Applications 1/36 Minimizing the Difference of L 1 and L 2 Norms with Department of Mathematical Sciences University of Texas Dallas May 31, 2017 Partially supported by NSF DMS 1522786 2/36 Outline 1 A nonconvex approach:

More information

SOR- and Jacobi-type Iterative Methods for Solving l 1 -l 2 Problems by Way of Fenchel Duality 1

SOR- and Jacobi-type Iterative Methods for Solving l 1 -l 2 Problems by Way of Fenchel Duality 1 SOR- and Jacobi-type Iterative Methods for Solving l 1 -l 2 Problems by Way of Fenchel Duality 1 Masao Fukushima 2 July 17 2010; revised February 4 2011 Abstract We present an SOR-type algorithm and a

More information

Enhanced Compressive Sensing and More

Enhanced Compressive Sensing and More Enhanced Compressive Sensing and More Yin Zhang Department of Computational and Applied Mathematics Rice University, Houston, Texas, U.S.A. Nonlinear Approximation Techniques Using L1 Texas A & M University

More information

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Low-rank matrix recovery via convex relaxations Yuejie Chi Department of Electrical and Computer Engineering Spring

More information

Generalized Orthogonal Matching Pursuit- A Review and Some

Generalized Orthogonal Matching Pursuit- A Review and Some Generalized Orthogonal Matching Pursuit- A Review and Some New Results Department of Electronics and Electrical Communication Engineering Indian Institute of Technology, Kharagpur, INDIA Table of Contents

More information

Coordinate Update Algorithm Short Course Proximal Operators and Algorithms

Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Coordinate Update Algorithm Short Course Proximal Operators and Algorithms Instructor: Wotao Yin (UCLA Math) Summer 2016 1 / 36 Why proximal? Newton s method: for C 2 -smooth, unconstrained problems allow

More information

Constrained optimization

Constrained optimization Constrained optimization DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Compressed sensing Convex constrained

More information

Optimization methods

Optimization methods Optimization methods Optimization-Based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_spring16 Carlos Fernandez-Granda /8/016 Introduction Aim: Overview of optimization methods that Tend to

More information

Multiuser Downlink Beamforming: Rank-Constrained SDP

Multiuser Downlink Beamforming: Rank-Constrained SDP Multiuser Downlink Beamforming: Rank-Constrained SDP Daniel P. Palomar Hong Kong University of Science and Technology (HKUST) ELEC5470 - Convex Optimization Fall 2018-19, HKUST, Hong Kong Outline of Lecture

More information

Data Sparse Matrix Computation - Lecture 20

Data Sparse Matrix Computation - Lecture 20 Data Sparse Matrix Computation - Lecture 20 Yao Cheng, Dongping Qi, Tianyi Shi November 9, 207 Contents Introduction 2 Theorems on Sparsity 2. Example: A = [Φ Ψ]......................... 2.2 General Matrix

More information

1 Sparsity and l 1 relaxation

1 Sparsity and l 1 relaxation 6.883 Learning with Combinatorial Structure Note for Lecture 2 Author: Chiyuan Zhang Sparsity and l relaxation Last time we talked about sparsity and characterized when an l relaxation could recover the

More information

Optimization for Learning and Big Data

Optimization for Learning and Big Data Optimization for Learning and Big Data Donald Goldfarb Department of IEOR Columbia University Department of Mathematics Distinguished Lecture Series May 17-19, 2016. Lecture 1. First-Order Methods for

More information

STA141C: Big Data & High Performance Statistical Computing

STA141C: Big Data & High Performance Statistical Computing STA141C: Big Data & High Performance Statistical Computing Lecture 8: Optimization Cho-Jui Hsieh UC Davis May 9, 2017 Optimization Numerical Optimization Numerical Optimization: min X f (X ) Can be applied

More information

Solution Recovery via L1 minimization: What are possible and Why?

Solution Recovery via L1 minimization: What are possible and Why? Solution Recovery via L1 minimization: What are possible and Why? Yin Zhang Department of Computational and Applied Mathematics Rice University, Houston, Texas, U.S.A. Eighth US-Mexico Workshop on Optimization

More information

Strengthened Sobolev inequalities for a random subspace of functions

Strengthened Sobolev inequalities for a random subspace of functions Strengthened Sobolev inequalities for a random subspace of functions Rachel Ward University of Texas at Austin April 2013 2 Discrete Sobolev inequalities Proposition (Sobolev inequality for discrete images)

More information

Convex Optimization and l 1 -minimization

Convex Optimization and l 1 -minimization Convex Optimization and l 1 -minimization Sangwoon Yun Computational Sciences Korea Institute for Advanced Study December 11, 2009 2009 NIMS Thematic Winter School Outline I. Convex Optimization II. l

More information

Tutorial on Convex Optimization: Part II

Tutorial on Convex Optimization: Part II Tutorial on Convex Optimization: Part II Dr. Khaled Ardah Communications Research Laboratory TU Ilmenau Dec. 18, 2018 Outline Convex Optimization Review Lagrangian Duality Applications Optimal Power Allocation

More information

Compressive Sensing Theory and L1-Related Optimization Algorithms

Compressive Sensing Theory and L1-Related Optimization Algorithms Compressive Sensing Theory and L1-Related Optimization Algorithms Yin Zhang Department of Computational and Applied Mathematics Rice University, Houston, Texas, USA CAAM Colloquium January 26, 2009 Outline:

More information

ECS289: Scalable Machine Learning

ECS289: Scalable Machine Learning ECS289: Scalable Machine Learning Cho-Jui Hsieh UC Davis Sept 29, 2016 Outline Convex vs Nonconvex Functions Coordinate Descent Gradient Descent Newton s method Stochastic Gradient Descent Numerical Optimization

More information

Algorithms for Nonsmooth Optimization

Algorithms for Nonsmooth Optimization Algorithms for Nonsmooth Optimization Frank E. Curtis, Lehigh University presented at Center for Optimization and Statistical Learning, Northwestern University 2 March 2018 Algorithms for Nonsmooth Optimization

More information

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs 2015 American Control Conference Palmer House Hilton July 1-3, 2015. Chicago, IL, USA Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs Raphael Louca and Eilyan Bitar

More information

HYBRID JACOBIAN AND GAUSS SEIDEL PROXIMAL BLOCK COORDINATE UPDATE METHODS FOR LINEARLY CONSTRAINED CONVEX PROGRAMMING

HYBRID JACOBIAN AND GAUSS SEIDEL PROXIMAL BLOCK COORDINATE UPDATE METHODS FOR LINEARLY CONSTRAINED CONVEX PROGRAMMING SIAM J. OPTIM. Vol. 8, No. 1, pp. 646 670 c 018 Society for Industrial and Applied Mathematics HYBRID JACOBIAN AND GAUSS SEIDEL PROXIMAL BLOCK COORDINATE UPDATE METHODS FOR LINEARLY CONSTRAINED CONVEX

More information

Accelerated primal-dual methods for linearly constrained convex problems

Accelerated primal-dual methods for linearly constrained convex problems Accelerated primal-dual methods for linearly constrained convex problems Yangyang Xu SIAM Conference on Optimization May 24, 2017 1 / 23 Accelerated proximal gradient For convex composite problem: minimize

More information

Math 273a: Optimization Overview of First-Order Optimization Algorithms

Math 273a: Optimization Overview of First-Order Optimization Algorithms Math 273a: Optimization Overview of First-Order Optimization Algorithms Wotao Yin Department of Mathematics, UCLA online discussions on piazza.com 1 / 9 Typical flow of numerical optimization Optimization

More information

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear

More information

A user s guide to Lojasiewicz/KL inequalities

A user s guide to Lojasiewicz/KL inequalities Other A user s guide to Lojasiewicz/KL inequalities Toulouse School of Economics, Université Toulouse I SLRA, Grenoble, 2015 Motivations behind KL f : R n R smooth ẋ(t) = f (x(t)) or x k+1 = x k λ k f

More information

Optimizing Nonconvex Finite Sums by a Proximal Primal-Dual Method

Optimizing Nonconvex Finite Sums by a Proximal Primal-Dual Method Optimizing Nonconvex Finite Sums by a Proximal Primal-Dual Method Davood Hajinezhad Iowa State University Davood Hajinezhad Optimizing Nonconvex Finite Sums by a Proximal Primal-Dual Method 1 / 35 Co-Authors

More information

A Unified Approach to Proximal Algorithms using Bregman Distance

A Unified Approach to Proximal Algorithms using Bregman Distance A Unified Approach to Proximal Algorithms using Bregman Distance Yi Zhou a,, Yingbin Liang a, Lixin Shen b a Department of Electrical Engineering and Computer Science, Syracuse University b Department

More information

Convex Optimization. (EE227A: UC Berkeley) Lecture 15. Suvrit Sra. (Gradient methods III) 12 March, 2013

Convex Optimization. (EE227A: UC Berkeley) Lecture 15. Suvrit Sra. (Gradient methods III) 12 March, 2013 Convex Optimization (EE227A: UC Berkeley) Lecture 15 (Gradient methods III) 12 March, 2013 Suvrit Sra Optimal gradient methods 2 / 27 Optimal gradient methods We saw following efficiency estimates for

More information

Optimization methods

Optimization methods Lecture notes 3 February 8, 016 1 Introduction Optimization methods In these notes we provide an overview of a selection of optimization methods. We focus on methods which rely on first-order information,

More information

Frist order optimization methods for sparse inverse covariance selection

Frist order optimization methods for sparse inverse covariance selection Frist order optimization methods for sparse inverse covariance selection Katya Scheinberg Lehigh University ISE Department (joint work with D. Goldfarb, Sh. Ma, I. Rish) Introduction l l l l l l The field

More information

A NEW ITERATIVE METHOD FOR THE SPLIT COMMON FIXED POINT PROBLEM IN HILBERT SPACES. Fenghui Wang

A NEW ITERATIVE METHOD FOR THE SPLIT COMMON FIXED POINT PROBLEM IN HILBERT SPACES. Fenghui Wang A NEW ITERATIVE METHOD FOR THE SPLIT COMMON FIXED POINT PROBLEM IN HILBERT SPACES Fenghui Wang Department of Mathematics, Luoyang Normal University, Luoyang 470, P.R. China E-mail: wfenghui@63.com ABSTRACT.

More information

Sparse and TV Kaczmarz solvers and the linearized Bregman method

Sparse and TV Kaczmarz solvers and the linearized Bregman method Sparse and TV Kaczmarz solvers and the linearized Bregman method Dirk Lorenz, Frank Schöpfer, Stephan Wenger, Marcus Magnor, March, 2014 Sparse Tomo Days, DTU Motivation Split feasibility problems Sparse

More information

Sparse Optimization Lecture: Dual Methods, Part I

Sparse Optimization Lecture: Dual Methods, Part I Sparse Optimization Lecture: Dual Methods, Part I Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know dual (sub)gradient iteration augmented l 1 iteration

More information

An Introduction to Sparse Approximation

An Introduction to Sparse Approximation An Introduction to Sparse Approximation Anna C. Gilbert Department of Mathematics University of Michigan Basic image/signal/data compression: transform coding Approximate signals sparsely Compress images,

More information

ARock: an algorithmic framework for asynchronous parallel coordinate updates

ARock: an algorithmic framework for asynchronous parallel coordinate updates ARock: an algorithmic framework for asynchronous parallel coordinate updates Zhimin Peng, Yangyang Xu, Ming Yan, Wotao Yin ( UCLA Math, U.Waterloo DCO) UCLA CAM Report 15-37 ShanghaiTech SSDS 15 June 25,

More information

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Sparse Recovery using L1 minimization - algorithms Yuejie Chi Department of Electrical and Computer Engineering Spring

More information

Lecture: Introduction to Compressed Sensing Sparse Recovery Guarantees

Lecture: Introduction to Compressed Sensing Sparse Recovery Guarantees Lecture: Introduction to Compressed Sensing Sparse Recovery Guarantees http://bicmr.pku.edu.cn/~wenzw/bigdata2018.html Acknowledgement: this slides is based on Prof. Emmanuel Candes and Prof. Wotao Yin

More information

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization /

Uses of duality. Geoff Gordon & Ryan Tibshirani Optimization / Uses of duality Geoff Gordon & Ryan Tibshirani Optimization 10-725 / 36-725 1 Remember conjugate functions Given f : R n R, the function is called its conjugate f (y) = max x R n yt x f(x) Conjugates appear

More information

Oslo Class 6 Sparsity based regularization

Oslo Class 6 Sparsity based regularization RegML2017@SIMULA Oslo Class 6 Sparsity based regularization Lorenzo Rosasco UNIGE-MIT-IIT May 4, 2017 Learning from data Possible only under assumptions regularization min Ê(w) + λr(w) w Smoothness Sparsity

More information

About Split Proximal Algorithms for the Q-Lasso

About Split Proximal Algorithms for the Q-Lasso Thai Journal of Mathematics Volume 5 (207) Number : 7 http://thaijmath.in.cmu.ac.th ISSN 686-0209 About Split Proximal Algorithms for the Q-Lasso Abdellatif Moudafi Aix Marseille Université, CNRS-L.S.I.S

More information

ON THE GLOBAL AND LINEAR CONVERGENCE OF THE GENERALIZED ALTERNATING DIRECTION METHOD OF MULTIPLIERS

ON THE GLOBAL AND LINEAR CONVERGENCE OF THE GENERALIZED ALTERNATING DIRECTION METHOD OF MULTIPLIERS ON THE GLOBAL AND LINEAR CONVERGENCE OF THE GENERALIZED ALTERNATING DIRECTION METHOD OF MULTIPLIERS WEI DENG AND WOTAO YIN Abstract. The formulation min x,y f(x) + g(y) subject to Ax + By = b arises in

More information

Sparse Covariance Selection using Semidefinite Programming

Sparse Covariance Selection using Semidefinite Programming Sparse Covariance Selection using Semidefinite Programming A. d Aspremont ORFE, Princeton University Joint work with O. Banerjee, L. El Ghaoui & G. Natsoulis, U.C. Berkeley & Iconix Pharmaceuticals Support

More information

1 Computing with constraints

1 Computing with constraints Notes for 2017-04-26 1 Computing with constraints Recall that our basic problem is minimize φ(x) s.t. x Ω where the feasible set Ω is defined by equality and inequality conditions Ω = {x R n : c i (x)

More information

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications Professor M. Chiang Electrical Engineering Department, Princeton University March

More information

Sparsity Regularization

Sparsity Regularization Sparsity Regularization Bangti Jin Course Inverse Problems & Imaging 1 / 41 Outline 1 Motivation: sparsity? 2 Mathematical preliminaries 3 l 1 solvers 2 / 41 problem setup finite-dimensional formulation

More information

Randomized Coordinate Descent Methods on Optimization Problems with Linearly Coupled Constraints

Randomized Coordinate Descent Methods on Optimization Problems with Linearly Coupled Constraints Randomized Coordinate Descent Methods on Optimization Problems with Linearly Coupled Constraints By I. Necoara, Y. Nesterov, and F. Glineur Lijun Xu Optimization Group Meeting November 27, 2012 Outline

More information

Motivation Sparse Signal Recovery is an interesting area with many potential applications. Methods developed for solving sparse signal recovery proble

Motivation Sparse Signal Recovery is an interesting area with many potential applications. Methods developed for solving sparse signal recovery proble Bayesian Methods for Sparse Signal Recovery Bhaskar D Rao 1 University of California, San Diego 1 Thanks to David Wipf, Zhilin Zhang and Ritwik Giri Motivation Sparse Signal Recovery is an interesting

More information

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16 XVI - 1 Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16 A slightly changed ADMM for convex optimization with three separable operators Bingsheng He Department of

More information

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 9. Alternating Direction Method of Multipliers Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 9 Alternating Direction Method of Multipliers Shiqian Ma, MAT-258A: Numerical Optimization 2 Separable convex optimization a special case is min f(x)

More information

A Note on the Complexity of L p Minimization

A Note on the Complexity of L p Minimization Mathematical Programming manuscript No. (will be inserted by the editor) A Note on the Complexity of L p Minimization Dongdong Ge Xiaoye Jiang Yinyu Ye Abstract We discuss the L p (0 p < 1) minimization

More information

MIT 9.520/6.860, Fall 2018 Statistical Learning Theory and Applications. Class 08: Sparsity Based Regularization. Lorenzo Rosasco

MIT 9.520/6.860, Fall 2018 Statistical Learning Theory and Applications. Class 08: Sparsity Based Regularization. Lorenzo Rosasco MIT 9.520/6.860, Fall 2018 Statistical Learning Theory and Applications Class 08: Sparsity Based Regularization Lorenzo Rosasco Learning algorithms so far ERM + explicit l 2 penalty 1 min w R d n n l(y

More information

A Multilevel Proximal Algorithm for Large Scale Composite Convex Optimization

A Multilevel Proximal Algorithm for Large Scale Composite Convex Optimization A Multilevel Proximal Algorithm for Large Scale Composite Convex Optimization Panos Parpas Department of Computing Imperial College London www.doc.ic.ac.uk/ pp500 p.parpas@imperial.ac.uk jointly with D.V.

More information

Randomness-in-Structured Ensembles for Compressed Sensing of Images

Randomness-in-Structured Ensembles for Compressed Sensing of Images Randomness-in-Structured Ensembles for Compressed Sensing of Images Abdolreza Abdolhosseini Moghadam Dep. of Electrical and Computer Engineering Michigan State University Email: abdolhos@msu.edu Hayder

More information

Compressed sensing. Or: the equation Ax = b, revisited. Terence Tao. Mahler Lecture Series. University of California, Los Angeles

Compressed sensing. Or: the equation Ax = b, revisited. Terence Tao. Mahler Lecture Series. University of California, Los Angeles Or: the equation Ax = b, revisited University of California, Los Angeles Mahler Lecture Series Acquiring signals Many types of real-world signals (e.g. sound, images, video) can be viewed as an n-dimensional

More information

Conditional Gradient (Frank-Wolfe) Method

Conditional Gradient (Frank-Wolfe) Method Conditional Gradient (Frank-Wolfe) Method Lecturer: Aarti Singh Co-instructor: Pradeep Ravikumar Convex Optimization 10-725/36-725 1 Outline Today: Conditional gradient method Convergence analysis Properties

More information

MLCC 2018 Variable Selection and Sparsity. Lorenzo Rosasco UNIGE-MIT-IIT

MLCC 2018 Variable Selection and Sparsity. Lorenzo Rosasco UNIGE-MIT-IIT MLCC 2018 Variable Selection and Sparsity Lorenzo Rosasco UNIGE-MIT-IIT Outline Variable Selection Subset Selection Greedy Methods: (Orthogonal) Matching Pursuit Convex Relaxation: LASSO & Elastic Net

More information

Compressed Sensing: a Subgradient Descent Method for Missing Data Problems

Compressed Sensing: a Subgradient Descent Method for Missing Data Problems Compressed Sensing: a Subgradient Descent Method for Missing Data Problems ANZIAM, Jan 30 Feb 3, 2011 Jonathan M. Borwein Jointly with D. Russell Luke, University of Goettingen FRSC FAAAS FBAS FAA Director,

More information

2 Regularized Image Reconstruction for Compressive Imaging and Beyond

2 Regularized Image Reconstruction for Compressive Imaging and Beyond EE 367 / CS 448I Computational Imaging and Display Notes: Compressive Imaging and Regularized Image Reconstruction (lecture ) Gordon Wetzstein gordon.wetzstein@stanford.edu This document serves as a supplement

More information

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs Raphael Louca & Eilyan Bitar School of Electrical and Computer Engineering American Control Conference (ACC) Chicago,

More information

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition NONLINEAR CLASSIFICATION AND REGRESSION Nonlinear Classification and Regression: Outline 2 Multi-Layer Perceptrons The Back-Propagation Learning Algorithm Generalized Linear Models Radial Basis Function

More information

Randomized Block Coordinate Non-Monotone Gradient Method for a Class of Nonlinear Programming

Randomized Block Coordinate Non-Monotone Gradient Method for a Class of Nonlinear Programming Randomized Block Coordinate Non-Monotone Gradient Method for a Class of Nonlinear Programming Zhaosong Lu Lin Xiao June 25, 2013 Abstract In this paper we propose a randomized block coordinate non-monotone

More information

SOCP Relaxation of Sensor Network Localization

SOCP Relaxation of Sensor Network Localization SOCP Relaxation of Sensor Network Localization Paul Tseng Mathematics, University of Washington Seattle IWCSN, Simon Fraser University August 19, 2006 Abstract This is a talk given at Int. Workshop on

More information

Frank-Wolfe Method. Ryan Tibshirani Convex Optimization

Frank-Wolfe Method. Ryan Tibshirani Convex Optimization Frank-Wolfe Method Ryan Tibshirani Convex Optimization 10-725 Last time: ADMM For the problem min x,z f(x) + g(z) subject to Ax + Bz = c we form augmented Lagrangian (scaled form): L ρ (x, z, w) = f(x)

More information

Majorization Minimization (MM) and Block Coordinate Descent (BCD)

Majorization Minimization (MM) and Block Coordinate Descent (BCD) Majorization Minimization (MM) and Block Coordinate Descent (BCD) Wing-Kin (Ken) Ma Department of Electronic Engineering, The Chinese University Hong Kong, Hong Kong ELEG5481, Lecture 15 Acknowledgment:

More information

Coordinate descent. Geoff Gordon & Ryan Tibshirani Optimization /

Coordinate descent. Geoff Gordon & Ryan Tibshirani Optimization / Coordinate descent Geoff Gordon & Ryan Tibshirani Optimization 10-725 / 36-725 1 Adding to the toolbox, with stats and ML in mind We ve seen several general and useful minimization tools First-order methods

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Instructor: Moritz Hardt Email: hardt+ee227c@berkeley.edu Graduate Instructor: Max Simchowitz Email: msimchow+ee227c@berkeley.edu

More information

Block Coordinate Descent for Regularized Multi-convex Optimization

Block Coordinate Descent for Regularized Multi-convex Optimization Block Coordinate Descent for Regularized Multi-convex Optimization Yangyang Xu and Wotao Yin CAAM Department, Rice University February 15, 2013 Multi-convex optimization Model definition Applications Outline

More information

Semidefinite Programming Basics and Applications

Semidefinite Programming Basics and Applications Semidefinite Programming Basics and Applications Ray Pörn, principal lecturer Åbo Akademi University Novia University of Applied Sciences Content What is semidefinite programming (SDP)? How to represent

More information

Sparse Solutions of an Undetermined Linear System

Sparse Solutions of an Undetermined Linear System 1 Sparse Solutions of an Undetermined Linear System Maddullah Almerdasy New York University Tandon School of Engineering arxiv:1702.07096v1 [math.oc] 23 Feb 2017 Abstract This work proposes a research

More information

Lecture 1 Introduction

Lecture 1 Introduction L. Vandenberghe EE236A (Fall 2013-14) Lecture 1 Introduction course overview linear optimization examples history approximate syllabus basic definitions linear optimization in vector and matrix notation

More information

Conditional Gradient Algorithms for Rank-One Matrix Approximations with a Sparsity Constraint

Conditional Gradient Algorithms for Rank-One Matrix Approximations with a Sparsity Constraint Conditional Gradient Algorithms for Rank-One Matrix Approximations with a Sparsity Constraint Marc Teboulle School of Mathematical Sciences Tel Aviv University Joint work with Ronny Luss Optimization and

More information

Gradient based method for cone programming with application to large-scale compressed sensing

Gradient based method for cone programming with application to large-scale compressed sensing Gradient based method for cone programming with application to large-scale compressed sensing Zhaosong Lu September 3, 2008 (Revised: September 17, 2008) Abstract In this paper, we study a gradient based

More information

Compressive Sensing and Beyond

Compressive Sensing and Beyond Compressive Sensing and Beyond Sohail Bahmani Gerorgia Tech. Signal Processing Compressed Sensing Signal Models Classics: bandlimited The Sampling Theorem Any signal with bandwidth B can be recovered

More information

Lecture 4: Linear and quadratic problems

Lecture 4: Linear and quadratic problems Lecture 4: Linear and quadratic problems linear programming examples and applications linear fractional programming quadratic optimization problems (quadratically constrained) quadratic programming second-order

More information

arxiv: v1 [math.oc] 14 Oct 2014

arxiv: v1 [math.oc] 14 Oct 2014 arxiv:110.3571v1 [math.oc] 1 Oct 01 An Improved Analysis of Semidefinite Approximation Bound for Nonconvex Nonhomogeneous Quadratic Optimization with Ellipsoid Constraints Yong Hsia a, Shu Wang a, Zi Xu

More information

Math 273a: Optimization Subgradient Methods

Math 273a: Optimization Subgradient Methods Math 273a: Optimization Subgradient Methods Instructor: Wotao Yin Department of Mathematics, UCLA Fall 2015 online discussions on piazza.com Nonsmooth convex function Recall: For ˉx R n, f(ˉx) := {g R

More information

New Rank-One Matrix Decomposition Techniques and Applications to Signal Processing

New Rank-One Matrix Decomposition Techniques and Applications to Signal Processing New Rank-One Matrix Decomposition Techniques and Applications to Signal Processing Yongwei Huang Hong Kong Baptist University SPOC 2012 Hefei China July 1, 2012 Outline Trust-region subproblems in nonlinear

More information

An ADMM algorithm for optimal sensor and actuator selection

An ADMM algorithm for optimal sensor and actuator selection An ADMM algorithm for optimal sensor and actuator selection Neil K. Dhingra, Mihailo R. Jovanović, and Zhi-Quan Luo 53rd Conference on Decision and Control, Los Angeles, California, 2014 1 / 25 2 / 25

More information

Convex Optimization M2

Convex Optimization M2 Convex Optimization M2 Lecture 8 A. d Aspremont. Convex Optimization M2. 1/57 Applications A. d Aspremont. Convex Optimization M2. 2/57 Outline Geometrical problems Approximation problems Combinatorial

More information

Bias-free Sparse Regression with Guaranteed Consistency

Bias-free Sparse Regression with Guaranteed Consistency Bias-free Sparse Regression with Guaranteed Consistency Wotao Yin (UCLA Math) joint with: Stanley Osher, Ming Yan (UCLA) Feng Ruan, Jiechao Xiong, Yuan Yao (Peking U) UC Riverside, STATS Department March

More information

Optimization for Compressed Sensing

Optimization for Compressed Sensing Optimization for Compressed Sensing Robert J. Vanderbei 2014 March 21 Dept. of Industrial & Systems Engineering University of Florida http://www.princeton.edu/ rvdb Lasso Regression The problem is to solve

More information

Convex optimization. Javier Peña Carnegie Mellon University. Universidad de los Andes Bogotá, Colombia September 2014

Convex optimization. Javier Peña Carnegie Mellon University. Universidad de los Andes Bogotá, Colombia September 2014 Convex optimization Javier Peña Carnegie Mellon University Universidad de los Andes Bogotá, Colombia September 2014 1 / 41 Convex optimization Problem of the form where Q R n convex set: min x f(x) x Q,

More information

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit New Coherence and RIP Analysis for Wea 1 Orthogonal Matching Pursuit Mingrui Yang, Member, IEEE, and Fran de Hoog arxiv:1405.3354v1 [cs.it] 14 May 2014 Abstract In this paper we define a new coherence

More information

COM Optimization for Communications 8. Semidefinite Programming

COM Optimization for Communications 8. Semidefinite Programming COM524500 Optimization for Communications 8. Semidefinite Programming Institute Comm. Eng. & Dept. Elect. Eng., National Tsing Hua University 1 Semidefinite Programming () Inequality form: min c T x s.t.

More information

An interior-point gradient method for large-scale totally nonnegative least squares problems

An interior-point gradient method for large-scale totally nonnegative least squares problems An interior-point gradient method for large-scale totally nonnegative least squares problems Michael Merritt and Yin Zhang Technical Report TR04-08 Department of Computational and Applied Mathematics Rice

More information

ACCELERATED FIRST-ORDER PRIMAL-DUAL PROXIMAL METHODS FOR LINEARLY CONSTRAINED COMPOSITE CONVEX PROGRAMMING

ACCELERATED FIRST-ORDER PRIMAL-DUAL PROXIMAL METHODS FOR LINEARLY CONSTRAINED COMPOSITE CONVEX PROGRAMMING ACCELERATED FIRST-ORDER PRIMAL-DUAL PROXIMAL METHODS FOR LINEARLY CONSTRAINED COMPOSITE CONVEX PROGRAMMING YANGYANG XU Abstract. Motivated by big data applications, first-order methods have been extremely

More information

Coordinate descent methods

Coordinate descent methods Coordinate descent methods Master Mathematics for data science and big data Olivier Fercoq November 3, 05 Contents Exact coordinate descent Coordinate gradient descent 3 3 Proximal coordinate descent 5

More information

Proximal Gradient Descent and Acceleration. Ryan Tibshirani Convex Optimization /36-725

Proximal Gradient Descent and Acceleration. Ryan Tibshirani Convex Optimization /36-725 Proximal Gradient Descent and Acceleration Ryan Tibshirani Convex Optimization 10-725/36-725 Last time: subgradient method Consider the problem min f(x) with f convex, and dom(f) = R n. Subgradient method:

More information

DOWNLINK transmit beamforming has recently received

DOWNLINK transmit beamforming has recently received 4254 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 58, NO. 8, AUGUST 2010 A Dual Perspective on Separable Semidefinite Programming With Applications to Optimal Downlink Beamforming Yongwei Huang, Member,

More information

Does Alternating Direction Method of Multipliers Converge for Nonconvex Problems?

Does Alternating Direction Method of Multipliers Converge for Nonconvex Problems? Does Alternating Direction Method of Multipliers Converge for Nonconvex Problems? Mingyi Hong IMSE and ECpE Department Iowa State University ICCOPT, Tokyo, August 2016 Mingyi Hong (Iowa State University)

More information

Low-Rank Matrix Recovery

Low-Rank Matrix Recovery ELE 538B: Mathematics of High-Dimensional Data Low-Rank Matrix Recovery Yuxin Chen Princeton University, Fall 2018 Outline Motivation Problem setup Nuclear norm minimization RIP and low-rank matrix recovery

More information

Inverse problems and sparse models (1/2) Rémi Gribonval INRIA Rennes - Bretagne Atlantique, France

Inverse problems and sparse models (1/2) Rémi Gribonval INRIA Rennes - Bretagne Atlantique, France Inverse problems and sparse models (1/2) Rémi Gribonval INRIA Rennes - Bretagne Atlantique, France remi.gribonval@inria.fr Structure of the tutorial Session 1: Introduction to inverse problems & sparse

More information

I P IANO : I NERTIAL P ROXIMAL A LGORITHM FOR N ON -C ONVEX O PTIMIZATION

I P IANO : I NERTIAL P ROXIMAL A LGORITHM FOR N ON -C ONVEX O PTIMIZATION I P IANO : I NERTIAL P ROXIMAL A LGORITHM FOR N ON -C ONVEX O PTIMIZATION Peter Ochs University of Freiburg Germany 17.01.2017 joint work with: Thomas Brox and Thomas Pock c 2017 Peter Ochs ipiano c 1

More information

A Tutorial on Compressive Sensing. Simon Foucart Drexel University / University of Georgia

A Tutorial on Compressive Sensing. Simon Foucart Drexel University / University of Georgia A Tutorial on Compressive Sensing Simon Foucart Drexel University / University of Georgia CIMPA13 New Trends in Applied Harmonic Analysis Mar del Plata, Argentina, 5-16 August 2013 This minicourse acts

More information