Convex Optimization for Signal Processing and Communications: From Fundamentals to Applications

Similar documents
Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

A Brief Review on Convex Optimization

Convex Optimization for Signal Processing and Communications:

CSCI : Optimization and Control of Networks. Review on Convex Optimization

Convex Optimization M2

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

5. Duality. Lagrangian

Lecture: Duality.

Convex Optimization Boyd & Vandenberghe. 5. Duality

Convex Optimization in Communications and Signal Processing

Lecture Note 5: Semidefinite Programming for Stability Analysis

Lecture 6: Conic Optimization September 8

12. Interior-point methods

Convex Optimization for Signal Processing and Communications: From Fundamentals to Applications

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Lecture: Duality of LP, SOCP and SDP

Constrained Optimization and Lagrangian Duality

Convex Optimization & Lagrange Duality

Convex optimization problems. Optimization problem in standard form

4. Convex optimization problems

Interior Point Algorithms for Constrained Convex Optimization

Optimisation convexe: performance, complexité et applications.

Lecture: Convex Optimization Problems

Tutorial on Convex Optimization: Part II

CS-E4830 Kernel Methods in Machine Learning

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Tutorial on Convex Optimization for Engineers Part II

4. Convex optimization problems

EE/AA 578, Univ of Washington, Fall Duality

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

Duality. Lagrange dual problem weak and strong duality optimality conditions perturbation and sensitivity analysis generalized inequalities

LMI MODELLING 4. CONVEX LMI MODELLING. Didier HENRION. LAAS-CNRS Toulouse, FR Czech Tech Univ Prague, CZ. Universidad de Valladolid, SP March 2009

Semidefinite Programming Basics and Applications

Convex Optimization and Modeling

Convex Optimization. Dani Yogatama. School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA. February 12, 2014

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

subject to (x 2)(x 4) u,

Lecture 7: Convex Optimizations

Module 04 Optimization Problems KKT Conditions & Solvers

CS711008Z Algorithm Design and Analysis

What can be expressed via Conic Quadratic and Semidefinite Programming?

Constrained optimization

LP Duality: outline. Duality theory for Linear Programming. alternatives. optimization I Idea: polyhedra

Lagrange duality. The Lagrangian. We consider an optimization program of the form

ICS-E4030 Kernel Methods in Machine Learning

A : k n. Usually k > n otherwise easily the minimum is zero. Analytical solution:

E5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming

Solution to EE 617 Mid-Term Exam, Fall November 2, 2017

15. Conic optimization

A : k n. Usually k > n otherwise easily the minimum is zero. Analytical solution:

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

12. Interior-point methods

3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions

1. Introduction. mathematical optimization. least-squares and linear programming. convex optimization. example. course goals and topics

1. Introduction. mathematical optimization. least-squares and linear programming. convex optimization. example. course goals and topics

Lecture 8. Strong Duality Results. September 22, 2008

Convex Optimization. (EE227A: UC Berkeley) Lecture 6. Suvrit Sra. (Conic optimization) 07 Feb, 2013

minimize x subject to (x 2)(x 4) u,

Lecture 18: Optimization Programming

Linear and non-linear programming

3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Convex Optimization M2

Geometric problems. Chapter Projection on a set. The distance of a point x 0 R n to a closed set C R n, in the norm, is defined as

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

Convex Optimization and l 1 -minimization

Primal/Dual Decomposition Methods

Convex Functions. Wing-Kin (Ken) Ma The Chinese University of Hong Kong (CUHK)

Introduction to Alternating Direction Method of Multipliers

Convex Functions. Pontus Giselsson

COURSE ON LMI PART I.2 GEOMETRY OF LMI SETS. Didier HENRION henrion

4. Convex optimization problems (part 1: general)

Lecture 2: Convex functions

A function(al) f is convex if dom f is a convex set, and. f(θx + (1 θ)y) < θf(x) + (1 θ)f(y) f(x) = x 3

4. Algebra and Duality

8. Geometric problems

Subgradients. subgradients and quasigradients. subgradient calculus. optimality conditions via subgradients. directional derivatives

Math 273a: Optimization Subgradients of convex functions

Introduction to Machine Learning Lecture 7. Mehryar Mohri Courant Institute and Google Research

The Lagrangian L : R d R m R r R is an (easier to optimize) lower bound on the original problem:

Convex Optimization and Modeling

Subgradient. Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes. definition. subgradient calculus

8. Geometric problems

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

l p -Norm Constrained Quadratic Programming: Conic Approximation Methods

ELE539A: Optimization of Communication Systems Lecture 6: Quadratic Programming, Geometric Programming, and Applications

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 18

Conic Linear Programming. Yinyu Ye

Subgradients. subgradients. strong and weak subgradient calculus. optimality conditions via subgradients. directional derivatives

Lecture 1: Background on Convex Analysis

Lagrangian Duality and Convex Optimization

In English, this means that if we travel on a straight line between any two points in C, then we never leave C.

Optimization for Machine Learning

9. Geometric problems

L. Vandenberghe EE236C (Spring 2016) 18. Symmetric cones. definition. spectral decomposition. quadratic representation. log-det barrier 18-1

Semidefinite Programming

Lecture: Examples of LP, SOCP and SDP

Optimality, Duality, Complementarity for Constrained Optimization

CITY UNIVERSITY OF HONG KONG

Transcription:

Convex Optimization for Signal Processing and Communications: From Fundamentals to Applications Chong-Yung Chi Institute of Communications Engineering & Department of Electrical Engineering National Tsing Hua University, Taiwan 30013 E-mail: cychi@ee.nthu.edu.tw Web: http://www.ee.nthu.edu.tw/cychi/ Invited tutorial talk at Aalborg University, Denmark, 2018/9/20 Acknowledgment: My post doctor, Dr. Chia-Hsiang Lin for preparing slides of Part II. My Ph.D. student, Yao-Rong Syu for some slides preparation. 1 / 122

Outline 1 Part I: Fundamentals of Convex Optimization 2 Part II: Application in Hyperspectral Image Analysis: (Big Data Analysis and Machine Learning) 3 Part III: Application in Wireless Communications (5G Systems) Subsection I: Outage Constrained Robust Transmit Optimization for Multiuser MISO Downlinks Subsection II: Outage Constrained Robust Hybrid Coordinated Beamforming for Massive MIMO Enabled Heterogeneous Cellular Networks 2 / 122

Optimization problem Optimization problem: minimize subject to f(x) x C (1) where f(x) is the objective function to be minimized and C is the feasible set from which we try to find an optimal solution. Let x = arg min f(x) (optimal solution or global minimizer) (2) x C Challenges in applications: Local optima; large problem size; decision variable x involving real and/or complex vectors, matrices; feasible set C involving generalized inequalities, etc. Computational complexity: NP-hard; polynomial-time solvable. Performance analysis: Performance insights, properties, perspectives, proofs (e.g., identifiability and convergence), limits and bounds. 3 / 122

Convex sets and convex functions-1 Affine (convex) combination: Provided that C is a nonempty set, x = θ 1x 1 + + θ Kx K, x i C i (3) is called an affine (a convex) combination of x 1,..., x K (K vectors or points of a set) if θi = 1, θi R (θi R+), K Z++. K i=1 Affine (convex) set: C is an affine (a convex) set if C is closed under the operation of affine (convex) combination; an affine set is constructed by lines; a convex set is constructed by line segments. Conic set: If θx C for any θ R + and any x C, then the set C is a cone and it is constructed by rays starting from the origin; the linear combination (3) is called a conic combination if θ i 0 i; 4 / 122

Convex Set Examples Ellipsoid and Euclidean ball Second-order cone centered at x c C = { (x, t) R n+1 x 2 t } Left plot: An ellipsoid with semiaxes λ 1, λ 2, and an Euclidean ball with radius r > max{ λ 1, λ 2} in R 2 ; right plot: Second-order cone in R 3. 5 / 122

Convex sets and convex functions-2 Let A = {a 1,..., a N } R M. Affine hull of A (the smallest affine set containing A) is defined as { N N } aff A x = θ ia i θ i = 1, θ i R i i=1 i=1 = { x = Cα + d α R p} (affine set representation) where C R M p is of full column rank, d aff A, and affdim A = p min{n 1, M}. A is affinely independent (A.I.) with affdim A = N 1 if the set {a a i a A, a a i} is linearly independent for any i; moreover, aff A = {x b T x = h} H(b, h) (when M = N) is a hyperplane, where (b, h) can be determined from A (closed-form expressions available). 6 / 122

Convex sets and convex functions-3 = {x R 3 b T 4 x = h 4} Figure 1: An illustration in R 3, where conv{a 1, a 2, a 3 } is a simplex defined by the shaded triangle, and conv{a 1, a 2, a 3, a 4 } is a simplex (and also a simplest simplex) defined by the tetrahedron with the four extreme points {a 1, a 2, a 3, a 4 }. 7 / 122

Convex sets and convex functions-4 Let A = {a 1,..., a N } R M. Convex hull of A (the smallest convex set containing A) is defined as conv A = { x = N N θ ia i i=1 i=1 } θ i = 1, θ i R + i aff A conv A is called a simplex (a polytope with N vertices) if A is A.I. When A is A.I. and M = N 1, conv A is called a simplest simplex of N vertices (i.e., a 1,..., a N ); aff (A \ {a i}) = H(b i, h i), i I N = {1,..., N} is a hyperplane, where the N boundary hyperplane parameters {(b i, h i), i = 1,..., N} can be uniquely determined from A (closed-form expressions available) and vice versa. 8 / 122

Convex Set Examples conic C { k i=1 θixi xi C, θi R+, k Z++ } = {θx x conv C, θ 0} Left plot: conic C (called the conic hull of C) is a convex cone formed by C = {x 1, x 2} via conic combinations, i.e., the smallest conic set that contains C; right plot: conic C formed by another set C (star). 9 / 122

Convex sets and convex functions-5 y Left plot: 1 +y 2 / B, implying that B is not a convex set; right plot: f(x) 2 is a convex function (by (4)). 10 / 122

Convex sets and convex functions-6 Convex function: f is convex if dom f (the domain of f) is a convex set, and for all x, y dom f, f(θx + (1 θ)y) θf(x) + (1 θ)f(y), 0 θ 1. (4) f is concave if f is convex. Some Examples of Convex Functions An affine function f(x) = a T x + b is both convex and concave on R n. f(x) = x T Px + 2q T x + r, where P S n, q R n and r R is convex if and only if P S n +. Every norm on R n (e.g., p for p 1) is convex. Linear function f(x) = Tr(AX) (where Tr(V) denotes the trace of a square matrix V) is both convex and concave on R n n. f(x) = log det(x) is convex on S n ++. 11 / 122

Ways Proving Convexity of a Function First-order Condition Suppose that f is differentiable. f is a convex function if and only if dom f is a convex set and f(y) f(x) + f(x) T (y x) x, y dom f. (5) Second-order Condition Suppose that f is twice differentiable. f is a convex function if and only if dom f is a convex set and 2 f(x) 0 (positive-semidefinite), x dom f. (6) Epigraph The epigraph of a function f : R n R is defined as epi f = {(x, t) x dom f, f(x) t} R n+1. (7) A function f is convex if and only if epi f is a convex set. 12 / 122

First-order Condition and Epigraph Left plot: first-order condition for a convex function f for the one-dimensional case: f(b) f(a) + f (a)(b a), for all a, b dom f; right plot: the epigraph of a convex function f : R R. 13 / 122

Convex Functions (Cont d) Convexity Preserving Operations Intersection, i Ci of convex sets Ci is a convex set; Nonnegative weighted sum, i θifi (where θi 0) of convex functions f i is a convex function. Image, h(c) = {h(x) x C)}, of a convex set C via affine mapping h(x) Ax + b, is a convex set; Composition f(h(x)) of a convex function f with affine mapping h is a convex function; Image, p(c), of a convex set C via perspective mapping p(x, t) x/t is a convex set; Perspective, g(x, t) = tf(x/t) (where t > 0) of a convex function f is a convex function. 14 / 122

Perspective Mapping & Perspective of a Function Left plot: pinhole camera interpretation of the perspective mapping p(x, t) = x/t, t > 0; right plot: epigraph of the perspective g(x, t) = tf(x/t), t > 0 of f(x) = x 2, where each ray is associated with g(at, t) = a 2 t for a different value of a. 15 / 122

Convex optimization problem Convex problem: (CVXP) p = min x C f(x) (8) is a convex problem if the objective function f( ) is a convex function and C is a convex set (called the feasible set) in standard form as follows: C = {x D f i(x) 0, h j(x) = 0, i = 1,..., m, j = 1,..., p}, where f i(x) is convex for all i and h j(x) is affine for all j and { m } { p } D = dom f dom f i dom h i is called the problem domain. i=1 Advantages: Global optimality: x can be obtained by closed-form solution, analytically (algorithm), or numerically by convex solvers (e.g., CVX and SeDuMi). Computational complexity: Polynomial-time solvable. Performance analysis: KKT conditions are the backbone for analysis. i=1 16 / 122

Global optimality and solution An optimality criterion: Any suboptimal solution to CVXP (8) is globally optimal. Assume that f is differentiable. Then a point x C is optimal if and only if f(x ) T (x x ) 0, x C (9) (where int C is assumed) 17 / 122

Global optimality and solution Besides the optimality criterion (9), a complementary approach for solving CVXP (8) is founded on the duality theory. Dual problem: L(x, λ, ν) f(x) + m i=1 λifi(x) + p i=1 νihi(x) (Lagrangian) g(λ, ν) = inf x D L(x, λ, ν)> (dual function) d = max {g(λ, ν) λ 0, ν R p } (dual problem) p = min {f(x) x C)} (primal problem CVXP (8)) where λ = (λ 1,..., λ m) and ν = (ν 1,..., ν p) are dual variables; stands for an abbreviated generalized inequality defined by the proper cone K = R m +, i.e., the first orthant, (a closed convex solid and pointed cone), i.e., λ K 0 λ K. CVXP (8) and its dual can be solved simultaneously by solving the so-called KKT conditions. (10) 18 / 122

Global optimality and solution KKT conditions: Suppose that f, f 1,..., f m, h 1,..., h p are differentiable and x is primal optimal and (λ, ν ) is dual optimal to CVXP (8). Under strong duality, i.e., p = d = L(x, λ, ν ) (which holds true under Slater s condition: a strictly feasible point exists, i.e., relint C ), the KKT conditions for solving x and (λ, ν ) are as follows: xl(x, λ, ν ) = 0, f i(x ) 0, i = 1,..., m, h i(x ) = 0, i = 1,..., p, λ i 0, i = 1,..., m, λ i f i(x ) = 0, i = 1,..., m. (complementary slackness) The above KKT conditions (11) and the optimality criterion (9) are equivalent under Slater s condition. (11a) (11b) (11c) (11d) (11e) 19 / 122

Strong Duality Lagrangian L(x, λ), dual function g(λ), and primal-dual optimal solution (x, λ ) = (1, 1) of the convex problem min{f 0(x) = x 2 (x 2) 2 1} with strong duality. Note that f 0(x ) = g(λ ) = L(x, λ ) = 1. 20 / 122

Standard Convex Optimization Problems Linear Programming (LP) - Inequality Form min c T x s.t. Gx h, Ax = b, ( stands for componentwise inequality) (12) where c R n, A R p n, b R p, G R m n, h R m, and x R n is the unknown vector variable. LP- Standard Form min c T x (13) s.t. x 0, Ax = b, where c R n, A R p n, b R p, and x R n is the unknown vector variable. 21 / 122

Standard Convex Optimization Problems (Cont d) Quadratic Programming (QP): Convex if and only if P 0 (i.e., P is positive semidefinite) min 1 2 xt Px + q T x + r (14) s.t. Ax = b, Gx h, where P S n, G R m n, and A R p n. Quadratically constrained QP (QCQP): Convex if and only if P i 0, i min 1 2 xt P 0x + q T 0 x + r 0 (15) s.t. 1 2 xt P ix + q T i x + r i 0, i = 1,..., m, Ax = b, where P i S n, i = 0, 1,..., m, and A R p n. 22 / 122

Standard Convex Optimization Problems (Cont d) Second-order cone programming (SOCP) min c T x s.t. A ix + b i 2 f T i x + d i, i = 1,..., m, Fx = g, (16) where A i R n i n, b i R n i, f i R n, d i R, F R p n, g R p, c R n, and x R n is the vector variable. Semidefinite programming (SDP) - Standard Form min Tr(CX) (17) s.t. X 0, Tr(A ix) = b i, i = 1,..., p, with variable X S n, where A i S n, C S n, and b i R. 23 / 122

Alternating direction method of multiplers (ADMM) Consider the following convex optimization problem: min x R n,z R m f1(x) + f2(z) s.t. x S 1, z S 2 z = Ax where f 1 : R n R and f 2 : R m R are convex functions, A is an m n matrix, and S 1 R n and S 2 R m are nonempty convex sets. (18) The considered dual problem of (18) is given by max min {f c } 1(x)+f 2(z)+ Ax z 2 ν R m x S 1,z S 2 2 + 2 νt (Ax z), (19) where c is a penalty parameter, and ν is the dual variable associated with the equality constraint in (18). 24 / 122

ADMM (Cont d) Inner minimization (convex problems): z(q + 1) = arg min {f 2(z) ν(q) T z + c Ax(q) z } 2, (20a) z S 2 2 2 x(q + 1) = arg min {f 1(x) + ν(q) T Ax + c Ax z(q + 1) } 2. (20b) x S 1 2 2 ADMM Algorithm 1: Set q = 0, choose c > 0. 2: Initialize ν(q) and x(q). 3: repeat 4: Solve (20a) and (20b) for z(q + 1) and x(q + 1) by two distributed equipments including the information exchange of z(q + 1) and x(q + 1) between them; 5: ν(q+1) = ν(q) + c (Ax(q + 1) z(q + 1)); 6: q := q + 1; 7: until the predefined stopping criterion is satisfied. When S 1 is bounded or A T A is invertible, ADMM is guaranteed to converge and the obtained {x(q), z(q)} is an optimal solution of problem (18). 25 / 122

Nonconvex problem Reformulation into a convex problem: Equivalent representations (e.g. epigraph representations); function transformation; change of variables, etc. Stationary-point solutions: Provided that C is closed and convex but f is nonconvex, a point x is a stationary point of the nonconvex problem (1) if f (x ; v) lim inf λ 0 f(x + λv) f(x ) λ f(x ) T (x x ) 0 0 x + v C (21) x C (when f is differentiable) where f (x ; v) is the directional derivative of f at a point x in direction v. Block successive upper bound minimization (BSUM) [Razaviyayn 13] guarantees a stationary-point solution under some convergence conditions. KKT points (i.e., solutions of KKT conditions) are also stationary points provided that the Slater s condition is satisfied. [Razaviyayn 13] M. Razaviyayn, M. Hong, and Z.-Q. Luo, A unified convergence analysis of block successive minimization methods for nonsmooth optimization, SIAM J. Optimiz., vol. 23, no. 2, pp. 11261153, 2013. 26 / 122

Stationary points and BSUM An illustration of stationary points of problem (1) for a nonconvex f and convex C; convergence to a stationary point of (1) by BSUM. 27 / 122

Stationary points for nonconvex feasible set An illustration of stationary points of problem (1) when both f and C are nonconvex. If y 1, y 2, y 3 are stationary points of min x C f(x) where C C is convex, then conic (C {y i}) = {θv v C {y i}, θ 0} and C {y i} {v = x {y i} x C} conic (C {y i}), i = 1, 2 = y 1, y 2 are also stationary points of (1). 28 / 122

Nonconvex problem Approximate solutions to problem (1) when f is convex but C is nonconvex: Convex restriction to C: Successive convex approximation (SCA) where C i C is convex for all i. Then x i = arg min x C i f(x) C i+1 (22) f(x i+1) = min x C i+1 f(x) f(x i ) (23) After convergence, an approximate solution x i is obtained. Convex relaxation to C (e.g., semidefinite relaxation (SDR)): C = {X S n + rank(x) = 1} C relaxed to conv C = S n + (SDR); C = { 3, 1, +1, +3} C relaxed to conv C = [ 3, 3] (24) The obtained X or x may not be feasible to problem (1); for SDR, a good approximate solution can be obtained from X via Gaussian randomization. 29 / 122

Successive Convex Approximation (SCA) Illustration of SCA for (1) when f is convex but C is nonconvex, yielding a solution x i (which is a stationary point under some mild condition). 30 / 122

Algorithm development Foundamental theory and tools: Calculus, linear algebra, matrix analysis and computations, convex sets, convex functions, convex problems (e.g., geometric program (GP), LP, QP, SOCP, SDP), duality, interior-point method; CVX and SeDuMi. 31 / 122

A new book Convex Optimization for Signal Processing and Communications: From Fundamentals to Applications Chong-Yung Chi, Wei-Chiang Li, Chia-Hsiang Lin (Publisher: CRC Press, 2017, 432 pages, ISBN: 9781498776455) Motivation: Most of mathematical books are hard to read for engineering students and professionals due to lack of enough fundamental details and tangible linkage between mathematical theory and applications. The book is written in a causally sequential fashion; namely, one can review/peruse the related materials introduced in early chapters/sections again, to overpass hurdles in reading. Covers convex optimization from fundamentals to advanced applications, while holding a strong link from theory to applications. Provides comprehensive proofs and perspective interpretations, many insightful figures, examples and remarks to illuminate core convex optimization theory. 32 / 122

Book features Illustrates, by cutting-edge applications, how to apply the convex optimization theory, like a guided journey/exploration rather than pure mathematics. Has been used for a 2-week short course under the book title at 9 major universities (Shandong Univ, Tsinghua Univ, Tianjin Univ, BJTU, Xiamen Univ., UESTC, SYSU, BUPT, SDNU) in Mainland China more than 17 times since early 2010. Thank you for your attention! Acknowledgment: Financial support by NTHU; my students, visiting students and scholars, short-course participants for almost uncountable questions, interactions and comments over the last decade. 33 / 122