Convex Functions. Wing-Kin (Ken) Ma The Chinese University of Hong Kong (CUHK)

Similar documents
3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions

Lecture 2: Convex functions

3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions

Convex Functions. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

A function(al) f is convex if dom f is a convex set, and. f(θx + (1 θ)y) < θf(x) + (1 θ)f(y) f(x) = x 3

Convex functions. Definition. f : R n! R is convex if dom f is a convex set and. f ( x +(1 )y) < f (x)+(1 )f (y) f ( x +(1 )y) apple f (x)+(1 )f (y)

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

CSCI : Optimization and Control of Networks. Review on Convex Optimization

Convex Functions. Pontus Giselsson

A Brief Review on Convex Optimization

4. Convex optimization problems (part 1: general)

IE 521 Convex Optimization

Conve functions 2{2 Conve functions f : R n! R is conve if dom f is conve and y 2 dom f 2 [0 1] + f ( + (1 ; )y) f () + (1 ; )f (y) (1) f is concave i

Convex Functions and Optimization

Optimisation convexe: performance, complexité et applications.

1. Introduction. mathematical optimization. least-squares and linear programming. convex optimization. example. course goals and topics

1. Introduction. mathematical optimization. least-squares and linear programming. convex optimization. example. course goals and topics

EC9A0: Pre-sessional Advanced Mathematics Course. Lecture Notes: Unconstrained Optimisation By Pablo F. Beker 1

Lecture 3. Optimization Problems and Iterative Algorithms

Lecture 1: January 12

Constrained Optimization and Lagrangian Duality

Convex Optimization. Convex Analysis - Functions

8. Conjugate functions

4. Convex optimization problems

CS675: Convex and Combinatorial Optimization Fall 2016 Convex Optimization Problems. Instructor: Shaddin Dughmi

1. Gradient method. gradient method, first-order methods. quadratic bounds on convex functions. analysis of gradient method

N. L. P. NONLINEAR PROGRAMMING (NLP) deals with optimization models with at least one nonlinear function. NLP. Optimization. Models of following form:

A : k n. Usually k > n otherwise easily the minimum is zero. Analytical solution:

Lecture 4: Convex Functions, Part I February 1

A : k n. Usually k > n otherwise easily the minimum is zero. Analytical solution:

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient

Linear and non-linear programming

Descent methods. min x. f(x)

Convex Optimization in Communications and Signal Processing

CS-E4830 Kernel Methods in Machine Learning

1. Introduction. mathematical optimization. least-squares and linear programming. convex optimization. example. course goals and topics

1. Introduction. mathematical optimization. least-squares and linear programming. convex optimization. example. course goals and topics

1. Introduction. mathematical optimization. least-squares and linear programming. convex optimization. example. course goals and topics

Lecture: Convex Optimization Problems

EE514A Information Theory I Fall 2013

Subgradients. subgradients. strong and weak subgradient calculus. optimality conditions via subgradients. directional derivatives

Convex optimization problems. Optimization problem in standard form

1/37. Convexity theory. Victor Kitov

Existence of minimizers

ELEG5481 SIGNAL PROCESSING OPTIMIZATION TECHNIQUES 6. GEOMETRIC PROGRAM

Chapter 2: Preliminaries and elements of convex analysis

Convex Optimization for Signal Processing and Communications: From Fundamentals to Applications

Subgradients. subgradients and quasigradients. subgradient calculus. optimality conditions via subgradients. directional derivatives

Concave and Convex Functions 1

HW1 solutions. 1. α Ef(x) β, where Ef(x) is the expected value of f(x), i.e., Ef(x) = n. i=1 p if(a i ). (The function f : R R is given.

Introduction to Machine Learning Lecture 7. Mehryar Mohri Courant Institute and Google Research

Handout 2: Elements of Convex Analysis

Nonlinear Programming Models

Lecture 2: Convex Sets and Functions

4. Convex optimization problems

Convex Sets and Functions (part III)

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

BASICS OF CONVEX ANALYSIS

IFT Lecture 2 Basics of convex analysis and gradient descent

Integral Jensen inequality

Static Problem Set 2 Solutions

Convex Optimization and Modeling

Definition of convex function in a vector space

Convex Optimization. (EE227A: UC Berkeley) Lecture 4. Suvrit Sra. (Conjugates, subdifferentials) 31 Jan, 2013

Functions of Several Variables

1 Introduction to Optimization

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Concave and Convex Functions 1

EE Applications of Convex Optimization in Signal Processing and Communications Dr. Andre Tkacenko, JPL Third Term

Some Background Math Notes on Limsups, Sets, and Convexity

Continuous Random Variables

Speech Recognition Lecture 7: Maximum Entropy Models. Mehryar Mohri Courant Institute and Google Research

Subgradient. Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes. definition. subgradient calculus

6.1 Variational representation of f-divergences

Convex Optimization Boyd & Vandenberghe. 5. Duality

The general programming problem is the nonlinear programming problem where a given function is maximized subject to a set of inequality constraints.

E 600 Chapter 3: Multivariate Calculus

13. Convex programming

Some properties of skew-symmetric distributions

Duality. Lagrange dual problem weak and strong duality optimality conditions perturbation and sensitivity analysis generalized inequalities

Some Expectations of a Non-Central Chi-Square Distribution With an Even Number of Degrees of Freedom

5. Duality. Lagrangian

Lecture 5 Channel Coding over Continuous Channels

Convex Optimization and Modeling

14 Lecture 14 Local Extrema of Function

FALL 2018 MATH 4211/6211 Optimization Homework 1

Convex Analysis and Optimization Chapter 2 Solutions

Exam in TMA4180 Optimization Theory

ENGG5781 Matrix Analysis and Computations Lecture 8: QR Decomposition

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version

Chapter 3a Topics in differentiation. Problems in differentiation. Problems in differentiation. LC Abueg: mathematical economics

GEOMETRIC APPROACH TO CONVEX SUBDIFFERENTIAL CALCULUS October 10, Dedicated to Franco Giannessi and Diethard Pallaschke with great respect

Wire Sizing, Repeater Insertion and Dominant Time Constant Optimization for a Bus Line.

Expectation Maximization (EM) Algorithm. Each has it s own probability of seeing H on any one flip. Let. p 1 = P ( H on Coin 1 )

subject to (x 2)(x 4) u,

This test has 30 questions. ANSWER ALL QUESTIONS. All questions carry equal (4) marks.

Computational Optimization. Mathematical Programming Fundamentals 1/25 (revised)

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 17

EE5138R: Problem Set 5 Assigned: 16/02/15 Due: 06/03/15

Lecture 8: Basic convex analysis

Transcription:

Convex Functions Wing-Kin (Ken) Ma The Chinese University of Hong Kong (CUHK) Course on Convex Optimization for Wireless Comm. and Signal Proc. Jointly taught by Daniel P. Palomar and Wing-Kin (Ken) Ma National Chiao Tung Univ., Hsinchu, Taiwan December 19-21, 28

Convex Functions A function f : R n R is said to be convex if i) domf is convex; and ii) for any x, y domf and θ [,1], f(θx + (1 θ)y) θf(x) + (1 θ)f(y) (y, f(y)) (x, f(x)) W.-K. Ma 1

f is strictly convex if f(θx+(1 θ)y) < θf(x)+(1 θ)f(y) for all < θ < 1 and for all x y. f is concave if f is convex. convex strictly convex (and of course convex) non-convex W.-K. Ma 2

1st and 2nd Order Conditions Gradient (for differentiable f) f(x) = [ f(x),..., f(x) ] T R n x 1 x n Hessian (for twice differentiable f): A matrix function 2 f(x) S n in which [ 2 f(x)] ij = 2 f(x) x i x j Taylor series: f(x + ν) = f(x) + f(x) T ν + 1 2 νt 2 f(x)ν +... W.-K. Ma 3

First-order condition: A differentiable f is convex iff given any x domf, f(x) f(x ) + f(x ) T (x x ), x domf f(x) f(x ) + f(x ) T (x x ) x x Second-order condition: A twice differentiable f is convex if and only if 2 f(x), x domf W.-K. Ma 4

Examples on R ax + b is convex. It is also concave. x 2 is convex (on R). x is convex. e αx is convex. log x is concave on R ++. xlog x is convex on R +. log x e t2 /2 dt is concave. W.-K. Ma 5

Affine function is both convex and concave. Examples on R n n f(x) = a T x + b = a i x i + b i=1 f(x) 2 1.5 1.5.5 1 1.5 2 1.5 1 x 2.5 1 1.5 x 1.5 W.-K. Ma 6

Quadratic function f(x) = x T Px + 2q T x + r = n n P ij x i x j + 2 n q i x i + r i=1 j=1 i=1 is convex if and only if P. 2 1 15 f(x) 1 5 f(x) 5 5 1.5 x 2.5 1 1.5 x 1.5 1 1 1.5 x 2.5 1 1.5 x 1.5 1 (a) P. (b) P. W.-K. Ma 7

p-norm is convex for p 1. f(x) = x p = ( n i=1 x i p )1 p 1.8.6.4 p = p = 2 p = 1 1.8.6.4 p =.5 p =.3 p =.1.2.2.2.2.4.4.6.8.6 1.8 1.8.6.4.2.2.4.6.8 1 (a) Region of x p = 1, p 1. 1 1.8.6.4.2.2.4.6.8 1 (b) Region of x p = 1, p 1. W.-K. Ma 8

Geometric mean is concave on R n ++. f(x) = n i=1 x i Log-sum-exp. f(x) = log ( n is convex on R n. (Log-sum-exp. can be used as an approx. to max i=1,...,n x i) i=1 e x i ) W.-K. Ma 9

Affine function: Examples on S n (intuitively less obvious) is convex and concave. f(x) = tr(ax) + b = Logarithmetic determinant function: n i=1 n A ji X ij + b j=1 f(x) = log det(x) is concave on S n ++. Maximum eigenvalue function: is convex on S n. f(x) = λ max (X) = sup y y T Xy y T y W.-K. Ma 1

Convexity Preserving Operations Affine transformation of the domain: f convex = f(ax + b) convex Example: (Least squares function) f(x) = y Ax 2 is convex, since f(x) = 2 is convex. Example: (MIMO capacity) f(x) = log det(hxh T + I) is concave on S n +, since f(x) = log det(x) is concave on S n ++. W.-K. Ma 11

Composition: Let g : R n R and h : R R. g convex, h convex, extended h nondecreasing = f(x) = h(g(x)) convex g concave, h convex, extended h nonincreasing = f(x) = h(g(x)) convex Example: f(x) = y Ax 2 2 is convex by composition, where g(x) = y Ax 2, h(x) = max{, x 2 }. W.-K. Ma 12

Non-negative weighted sum: f 1,...,f m convex w 1,...,w m = m w i f i convex i=1 Example: Regularized least squares function is convex for γ. f(x) = y Ax 2 2 + γ x 2 2 Extension of non-negative weighted sum: f(x, y) convex in x for each y A w(y) for each y A = A w(y)f(x, y)dy convex Example: Ergodic MIMO capacity f(x) = E H {log det(hxh T + I)} ( = ) p(h)log det(hxh T + I)dH is concave on S n +. W.-K. Ma 13

Pointwise maximum: f 1,...,f m convex = f(x) = max{f 1 (x),...,f m (x)} convex Example: Infinity norm f(x) = x = max i=1,...,m x i is convex because f(x) = max{x 1, x 1, x 2, x 2,...,x n, x n }. Pointwise supremum: f(x, y) convex in x for each y A = sup f(x, y) convex y A Example: Worst-case least squares function f(x) = max E E y (A + E)x 2 2 is convex. (E does not even need to be convex!) W.-K. Ma 14

The basic inequality for convex f Jensen s Inequality f(θx + (1 θ)y) θf(x) + (1 θ)f(y) is also called Jensen s inequality. It can be extended to f ( k ) θ k x k i=1 k θ k f(x k ) i=1 where θ 1,...,θ k, and k i=1 θ k = 1; and to f ( S ) p(x)xdx S p(x)f(x)dx where p(x) on S domf, and S p(x)dx = 1. W.-K. Ma 15

Inequalities derived from Jensen s inequality: Arithmetic-geometric inequality: ab (a + b)/2 for a, b Hadamard inequality: detx n i=1 X ii for X S n + Kullback-Leiber divergence: Let p(x),q(x) be PDFs on S, S p(x) log ( ) p(x) dx q(x) Hölder inequality: where 1/p + 1/q = 1, p > 1. x T y x p x q W.-K. Ma 16

Epigraph The epigraph of f is epif = {(x, t) x domf, f(x) t} A powerful property: f convex epif convex e.g., some convexity preserving properties can be proven quite easily by epigraph. t f(x) epif x W.-K. Ma 17

Sublevel Sets The α-sublevel set of f S α = {x domf f(x) α} f convex = S α convex for every α, but S α convex for every α f convex f(x) f(x) α α S α x S α x convex f and convex S α non-convex f but convex S α W.-K. Ma 18

Quasiconvex Functions f : R n R is called quasiconvex (or unimodal) if domf is convex and every S α is convex. If f is convex then f is quasiconvex (by the definition). f is called quasiconcave if f is quasiconvex. f is called quasilinear if f is quasiconvex and quasiconcave. Example: Linear fractional function (useful in modeling SINR) f(x) = at x + b c T x + d is quasiconvex on {x c T x + d > }, because S α = {x a T x + b t(c T x + d)} is a hyperplane. In fact it is also quasiconcave. W.-K. Ma 19