Helly's Theorem and its Equivalences via Convex Analysis

Similar documents
GEOMETRIC APPROACH TO CONVEX SUBDIFFERENTIAL CALCULUS October 10, Dedicated to Franco Giannessi and Diethard Pallaschke with great respect

Math 341: Convex Geometry. Xi Chen

Chapter 2 Convex Analysis

Division of the Humanities and Social Sciences. Supergradients. KC Border Fall 2001 v ::15.45

POLARS AND DUAL CONES

Appendix B Convex analysis

BASICS OF CONVEX ANALYSIS

Optimality Conditions for Nonsmooth Convex Optimization

Dedicated to Michel Théra in honor of his 70th birthday

Introduction to Convex Analysis Microeconomics II - Tutoring Class

Convex and Nonconvex Optimization Techniques for the Constrained Fermat-Torricelli Problem

Best approximations in normed vector spaces

Convex Optimization Notes

Optimization and Optimal Control in Banach Spaces

The Subdifferential of Convex Deviation Measures and Risk Functions

Set, functions and Euclidean space. Seungjin Han

Convex Analysis and Economic Theory AY Elementary properties of convex functions

Helly s Theorem with Applications in Combinatorial Geometry. Andrejs Treibergs. Wednesday, August 31, 2016

Convex Analysis Background

In English, this means that if we travel on a straight line between any two points in C, then we never leave C.

Convex Functions and Optimization

Week 3: Faces of convex sets

Convex Analysis and Optimization Chapter 2 Solutions

ON GENERALIZED-CONVEX CONSTRAINED MULTI-OBJECTIVE OPTIMIZATION

Basic convexity. 1.1 Convex sets and combinations. λ + μ b (λ + μ)a;

Convex Geometry. Carsten Schütt

Preliminary Version Draft 2015

A SET OF LECTURE NOTES ON CONVEX OPTIMIZATION WITH SOME APPLICATIONS TO PROBABILITY THEORY INCOMPLETE DRAFT. MAY 06

Lecture 4: Convex Functions, Part I February 1

The proximal mapping

MA651 Topology. Lecture 10. Metric Spaces.

Lecture 6 - Convex Sets

Mathematics II, course

Lecture Notes in Advanced Calculus 1 (80315) Raz Kupferman Institute of Mathematics The Hebrew University

Maths 212: Homework Solutions

A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions

Chapter 1. Optimality Conditions: Unconstrained Optimization. 1.1 Differentiable Problems

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Generalized Differential Calculus and Applications to Optimization

Abstract. In this paper, the concepts and the properties of conjugate maps and directional derivatives of convex vector functions from a subset of

THE UNIQUE MINIMAL DUAL REPRESENTATION OF A CONVEX FUNCTION

Chapter 2. Convex Sets: basic results

Subdifferential representation of convex functions: refinements and applications

On Semicontinuity of Convex-valued Multifunctions and Cesari s Property (Q)

Convex Analysis and Economic Theory Winter 2018

g 2 (x) (1/3)M 1 = (1/3)(2/3)M.

Handout 2: Elements of Convex Analysis

Metric regularity and systems of generalized equations

Metric Spaces and Topology

Geometry and topology of continuous best and near best approximations

On John type ellipsoids

On the convexity of piecewise-defined functions

Chapter 1 Preliminaries

MATH 202B - Problem Set 5

Local strong convexity and local Lipschitz continuity of the gradient of convex functions

DEFINABLE VERSIONS OF THEOREMS BY KIRSZBRAUN AND HELLY

Subgradients. subgradients and quasigradients. subgradient calculus. optimality conditions via subgradients. directional derivatives

Convex analysis and profit/cost/support functions

Introduction to Real Analysis Alternative Chapter 1

8. Conjugate functions

Exercise Solutions to Functional Analysis

Chapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space.

Introduction to Convex and Quasiconvex Analysis

In Progress: Summary of Notation and Basic Results Convex Analysis C&O 663, Fall 2009

Locally convex spaces, the hyperplane separation theorem, and the Krein-Milman theorem

The extreme points of symmetric norms on R^2

Lecture 1: Convex Sets January 23

Lectures on Geometry

Real Analysis Notes. Thomas Goller

Functional Analysis. Franck Sueur Metric spaces Definitions Completeness Compactness Separability...

Course 212: Academic Year Section 1: Metric Spaces

AN INTRODUCTION TO EXTREME POINTS AND APPLICATIONS IN ISOMETRIC BANACH SPACE THEORY

USING FUNCTIONAL ANALYSIS AND SOBOLEV SPACES TO SOLVE POISSON S EQUATION

REAL VARIABLES: PROBLEM SET 1. = x limsup E k

IE 521 Convex Optimization

Appendix PRELIMINARIES 1. THEOREMS OF ALTERNATIVES FOR SYSTEMS OF LINEAR CONSTRAINTS

MAT 570 REAL ANALYSIS LECTURE NOTES. Contents. 1. Sets Functions Countability Axiom of choice Equivalence relations 9

2. Dual space is essential for the concept of gradient which, in turn, leads to the variational analysis of Lagrange multipliers.

3.1 Convexity notions for functions and basic properties

Optimization Theory. A Concise Introduction. Jiongmin Yong

Translative Sets and Functions and their Applications to Risk Measure Theory and Nonlinear Separation

SOME PROPERTIES ON THE CLOSED SUBSETS IN BANACH SPACES

Some Background Math Notes on Limsups, Sets, and Convexity

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

This chapter reviews some basic geometric facts that we will need during the course.

Convex Sets with Applications to Economics

Inequality Constraints

Near convexity, metric convexity, and convexity

Econ Lecture 3. Outline. 1. Metric Spaces and Normed Spaces 2. Convergence of Sequences in Metric Spaces 3. Sequences in R and R n

1 Radon, Helly and Carathéodory theorems

Convexity and constructive infima

Subgradient. Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes. definition. subgradient calculus

Notions such as convergent sequence and Cauchy sequence make sense for any metric space. Convergent Sequences are Cauchy

Subgradients. subgradients. strong and weak subgradient calculus. optimality conditions via subgradients. directional derivatives

Convex Optimization Conjugate, Subdifferential, Proximation

Continuity of convex functions in normed spaces

Epiconvergence and ε-subgradients of Convex Functions

Topology. Xiaolong Han. Department of Mathematics, California State University, Northridge, CA 91330, USA address:

IE 521 Convex Optimization Homework #1 Solution

Semicontinuous functions and convexity

Transcription:

Portland State University PDXScholar University Honors Theses University Honors College 2014 Helly's Theorem and its Equivalences via Convex Analysis Adam Robinson Portland State University Let us know how access to this document benefits you. Follow this and additional works at: http://pdxscholar.library.pdx.edu/honorstheses Recommended Citation Robinson, Adam, "Helly's Theorem and its Equivalences via Convex Analysis" (2014). University Honors Theses. Paper 67. 10.15760/honors.62 This Thesis is brought to you for free and open access. It has been accepted for inclusion in University Honors Theses by an authorized administrator of PDXScholar. For more information, please contact pdxscholar@pdx.edu.

Helly s Theorem and Its Equivalences via Convex Analysis Adam Robinson Advisor: Dr. Mau Nam Nguyen MTH 401 Honors Project Submitted in Partial Fulfillment of the Requirements for the Degree of Bachelor of Science at Portland State University 2014

Helly s Theorem and Its Equivalences via Convex Analysis Adam Robinson Advisor: Dr. Mau Nam Nguyen Portland State University, 2014 ABSTRACT Helly s theorem is an important result from Convex Geometry. It gives sufficient conditions for a family of convex sets to have a nonempty intersection. A large variety of proofs as well as applications are known. Helly s theorem also has close connections to two other well-known theorems from Convex Geometry: Radon s theorem and Carathéodory s theorem. In this project we study Helly s theorem and its relations to Radon s theorem and Carathéodory s theorem by using tools of Convex Analysis and Optimization. More precisely, we will give a novel proof of Helly s theorem, and in addition we show in a complete way that these three famous theorems are equivalent in the sense that using one of them allows us to derive the others.

Contents Ch. 1. Elements of Convex Analysis and Optimization 3 Ch. 2. The Theorems of Carathédory, Radon, and Helly: Their Statements and Proofs 12 2.1 Carathédory Theorem.............................. 12 2.2 Radon Theorem.................................. 14 2.3 Helly Theorem.................................. 15 Ch. 3. The Theorems of Carathédory, Radon, and Helly: Their Equivalence 17 3.1 Carathéodory s and Helly s Theorem...................... 17 3.2 Carathéodory s and Radon s Theorem..................... 22 iii

Introduction In Convex Geometry, geometric properties of convex sets and functions are investigated. The foundations of this field were developed by many accomplished mathematicians, such as Hermann Brunn, Hermann Minkowski, Werner Fenchel, Constantin Carathéodory, and Eduard Helly. At the beginning of the 1960 s, Convex Analysis was grown out of Convexity, and this new field was systematically developed by the works of R. Tyrrell Rockafellar, Jean-Jacques Moreau, and others. Convex Analysis is more concerned with the generalized differentiation theory of convex functions and sets rather than only with their geometric properties. The presence of the convexity makes it possible to develop calculus rules for a generalized derivative concept called the subdifferential, which can be used to deal with convex functions that are not necessarily differentiable. Convex Analysis then becomes the mathematical foundation for Convex Optimization, a fast growing field with numerous applications to Control Systems, Estimation and Signal Processing, Communications and Networks, Electronic Circuit Design, Data Analysis and Modeling, Statistics, Economics and Finance, etc. In this project we use Convex Analysis and Optimization to study some basic results of Convex Geometry. We mainly focus on a theorem introduced by Eduard Helly in 1913 which gives sufficient conditions for a family of convex sets to have a nonempty intersection. Using modern tools from convex analysis and optimization, we will study Helly s 1

2 theorem from both theoretical and numerical viewpoints. In particular, we will give a novel proof, via Convex Analysis, of Helly s theorem and its connections to Radon s theorem and Carathéodory s theorem, which are also important results from Convex Geometry. It has been mentioned in several references that the above-mentioned theorems of Helly, Radon, and Carathéodory are equivalent in the sense that using one of them allows us to derive the others. However, in [8, p. 47] P. M. Gruber says that we were not able to locate in the literature a complete proof in the context of R n ; see also the excellent surveys [5] and [6], as well as the monograph [3]. We will use the tools of Convex Analysis to provide a complete treatment for the equivalence of these theorems. The analysis involves the use of generalized differentiation properties of the class of distance functions associated with convex sets. This class of functions, which reflects the connection between convex functions and sets, allows us to give a simple, self-contained, complete proof of the described equivalence. Further on, using the concept of distance function we are able to study effective numerical algorithms for finding a point in the intersection of the given family of convex sets, whose existence is guaranteed by Helly s theorem.

Chapter 1 Elements of Convex Analysis and Optimization In this chapter we introduce some important concepts and results of Convex Analysis and Optimization that will be used in the subsequent chapters. The materials presented here can be found in many books on Convex Analysis and Optimization; see, e.g., [10, 11] and the references therein. Throughout the thesis we consider the Euclidean space R n equipped with the Euclidean norm of an element x = (x 1,..., x n ) given by x := x 2 1 +... + x2 n, and the inner product of any two elements x = (x 1,..., x n ) and y = (y 1,..., y n ) given by x, y := x 1 y 1 + + x n y n. It follows from the definition that x 2 = x, x. 3

4 A subset Ω of R n is said to be convex if λx + (1 λ)y Ω whenever x, y Ω and λ (0, 1). Geometrically, a subset Ω is convex if for any x, y Ω, the line segment connecting x and y belongs to the set. Figure 1.0.1: Convex set Ω 1 and nonconvex set Ω 2. From the definition, it is obvious that if {Ω i } i I is a collection of convex sets in R n, then the intersection i I Ω i is also convex. In particular, the intersection of any two convex sets is also a convex set. This property motivates the definition of the convex hull of an arbitrary subset of R n. Given a subset Ω R n, define the convex hull of Ω by co Ω := {C C is convex and Ω C}. Equivalently, the convex hull of a set Ω is the smallest convex set containing Ω. The following important result is a direct consequence of the definition. Proposition 1.0.1. For any convex subset Ω of R n, its convex hull admits the representation { m co Ω = λ i w i m } λ i = 1, λ i 0, w i Ω, m N.

5 A function f : R n R defined on a convex set Ω is called convex if f(λx + (1 λ)y) λf(x) + (1 λ)f(y) for all x, y Ω and λ (0, 1). If this inequality becomes strict for whenever x y, x, y Ω, we say that f is strictly convex on Ω. Figure 1.0.2: A convex function f(x) and its epigraph, the shaded region above f(x). A mapping B : R n R m is called affine if there exist an m n matrix A and an element b R m such that B(x) = Ax + b, for all x R n. Let us present in the proposition below some operations that preserve convexity of functions. Proposition 1.0.2. (i) Let f i : R n R be convex functions for all i = 1,..., m. Then the following functions are convex as well: - The multiplication by scalars λf for any λ > 0. - The sum function m f i. - The maximum function max 1 i m f i. (ii) Let f : R n R be convex and let ϕ : R R be nondecreasing. Then the composition ϕ f is convex. (iii) Let B : R n R p be an affine mapping and let f : R p R be a convex function. Then

6 the composition f B is convex. (iv) Let f i : R n R, for i I, be a collection of convex functions with a nonempty index set I. Then the supremum function f(x) := sup i I f i (x) is convex. The class of distance functions presented in what follows plays a crucial role throughout the thesis. Given a set Ω R n, the distance function associated with Ω is defined by d(x; Ω) := inf { x ω ω Ω }. For each x R n, the Euclidean projection from x to Ω is defined by Π(x; Ω) := { ω Ω x ω = d(x; Ω) }. A function f : R n R is called Lipschitz continuous on R n if there exists a constant l 0 such that f(x) f(y) l x y for all x, y R n. From the definition, it is obvious that any Lipschitz continuous function is uniformly continuous on R n. We will prove in the proposition below that the distance function associated with a nonempty set is Lipschitz continuous on R n. Proposition 1.0.3. Given a nonempty set Ω, the distance function d( ; Ω) is Lipschitz continuous on R n with Lipschitz constant l = 1. Proof. We will show that d(x; Ω) d(y; Ω) x y for all x, y R n.

7 For any ω Ω, one has by the triangle inequality that d(x; Ω) x ω x y + y ω. This implies d(x; Ω) x y + inf{ y ω ω Ω} = x y + d(y; Ω), and hence d(x; Ω) d(y; Ω) x y. Similarly, we have that d(y; Ω) d(x; Ω) x y. Therefore, d(y; Ω) d(x; Ω) y x. Proposition 1.0.4. If Ω is a convex set, then the distance function d( ; Ω) is a convex function. Proof. Take any x,y R n and λ (0, 1). We will show that the distance function given by f(x) := d(x; Ω) for x R n satisfies the convex inequality: f(λx + (1 λ)y) λf(x) + (1 λ)f(y). (1.0.1) Fix any ɛ > 0. By the properties of the infimum, there exists u Ω such that x u < d(x; Ω) + ɛ 2. Similarly, there exists v Ω such that y v < d(y; Ω) + ɛ 2.

8 Since Ω is a convex set, λu + (1 λ)v Ω. Then d(λx + (1 λ)y; Ω) λx + (1 λ)y (λu + (1 λ)v) λ(x u) + (1 λ)(y v) = λ x u + (1 λ) y v [ λ d(x; Ω) + ɛ ] [ + (1 λ) d(y; Ω) + ɛ ] 2 2 = λd(x; Ω) + (1 λ)d(y; Ω) + ɛ. Then we obtain (1.0.1) by letting ɛ 0. Let f : R n R be a convex function and let x R n. A vector v R n is called a subgradient of f at x if v, x x f(x) f( x) for all x R n. The set of all subgradients of f at x is called the subdifferential of the function at this point and is denoted by f( x). Another important concept of Convex Analysis is called the normal cone to a nonempty convex set Ω R n at a point x Ω and defined by N( x; Ω) := {v R n v, x x 0 for all x Ω}. In what follows we study subdifferential formulas for distance functions to convex sets. We first pay attention to the case where the reference point belongs to the set. Proposition 1.0.5. Suppose that x Ω. Then d( x; Ω) = N( x; Ω) B, where B is the closed unit ball of R n.

9 Proof. Fix any v d( x; Ω). Then v, x x d(x; Ω) d( x; Ω) = d(x; Ω) for all x R n. (1.0.2) Since the distance function d( ; Ω) satisfies a Lipschitz condition with the Lipschitz constant l = 1, one has v, x x x x for all x R n. This implies v 1 or v B. It also follows from (1.0.2) that v, x x 0 for all x Ω. Thus v N( x; Ω). Therefore, v N( x; Ω) B. Let us now prove the opposite inclusion. Fix any v N( x; Ω) B. Then v 1 and v, w x 0 for all w Ω. Thus for any x R n and for any w Ω, one has v, x x = v, x w + w x = v, x w + v, w x v, x w v x w x w. This implies v, x x d(x; Ω) = d(x; Ω) d( x; Ω), and hence v d( x; Ω). Now we pay attention to the case where the reference point does not belong to the set.

10 Proposition 1.0.6. Let Ω be a nonempty closed convex set and let x / Ω. Then d( x; Ω) = { x Π( x; Ω) }. d( x; Ω) Proof. Fix z := Π( x; Ω) Ω and fix any v d( x; Ω). By the definition of subdifferential, v, x x d(x; Ω) d( x; Ω) = d(x; Ω) x z x z x z for all x R n. Denoting p(x) := x z, we have v, x x p(x) p( x) for all x R n, { x z } x z which implies that v p( x) =. Let us show that v = x z x z of d( ; Ω) at x. Indeed, for any x R n, denote p x := Π(x; Ω) and get is a subgradient v, x x = v, x z + v, z x = v, x z x z = v, x p x + v, p x z x z. Since z = Π( x; Ω), it follows that x z, p x z 0, and so we have v, p x z 0. Using the fact that v = 1 and the Cauchy-Schwarz inequality gives us v, x x = v, x p x + v, p x z x z v x p x x z = x p x x z = d(x; Ω) d( x; Ω) for all x R n. Thus we arrive at v d( x; Ω). In what follows we present some useful subdifferential rules for convex functions.

11 Proposition 1.0.7. (i) Let f : R n R be a convex function and let x R n. Then the following holds for any α > 0: (αf)( x) = α f( x). (ii) Let f i : R n R be convex functions for i = 1, 2 and let x R n. Then (f 1 + f 2 )( x) = f 1 ( x) + f 2 ( x). (iii) Let B : R n R m be an affine mapping given by B(x) = Ax + b, where A is an m n matrix. Consider a convex function f : R m R. Let x R n and let ȳ = B( x). Then (f B)( x) = A T ( f(ȳ)) = {A T (v) v f(ȳ)}. (1.0.3) Given convex functions f i : R n R for i = 1,..., m, define f(x) := max{f i (x) i = 1,..., m}. The proposition below gives a formula for computing the subdifferential of the maximum function. Theorem 1.0.8. Suppose that f i : R n R for i = 1,..., m are convex functions. Then f( x) = co i I( x) f i ( x),, where I( x) := {i = 1,..., m f i ( x) = f( x)}.

Chapter 2 The Theorems of Carathédory, Radon, and Helly: Their Statements and Proofs In this chapter we give a survey of the statements and proofs of the theorems of Carathédory, Radon, and Helly; see, e.g., [10]. 2.1 Carathédory Theorem Given a set Ω, the convex cone generated by Ω, denoted by K Ω, is the smallest convex cone containing Ω. Lemma 2.1.1. Consider a nonempty set Ω R n. The convex cone generated by Ω has the following representation: { m } K Ω = λ i a i λi 0, a i Ω, m N. 12

13 When Ω is convex, K Ω = R + Ω. Proof. Denote by C the set on the right-hand side of the equality above. Since C is clearly a convex cone containing Ω, C K Ω. It remains to prove the opposite inclusion. Consider an element of C given by x = m λ ia i, where λ i 0, a i Ω, and m N. If λ i = 0 for all i, then x = 0 K Ω. If λ i > 0 for some i, define λ := m λ i > 0. Thus x = λ( λ i λ a i) K Ω. Hence C K Ω. We have proved that C = K Ω. Lemma 2.1.2. Consider a set Ω R n, Ω, and any element x K Ω \{ }. The following holds: x = k λ i a i, λ i > 0, a i Ω and k n. Proof. For an element x K Ω \{ }, we have, by 2.1.1, the representation x = m µ ia i, with µ i > 0, a i Ω and m N. If the elements a 1,..., a m are linearly dependent, then there exist γ i R for i = 1,..., m, not all zeros, such that γ i a i = 0. Define the nonempty set I := {i = 1,..., m γ i > 0}. For any ɛ > 0 we have x = µ i a i = µ i a i ɛ( γ i a i ) = (µ i ɛγ i )a i. Choose ɛ := min{ µ i γ i i I} = µ i 0 γ i0, with i 0 I, and define β i := µ i ɛγ i for i = 1,..., m. We now have the following definition for an element x K Ω \{ }: x = β i a i.

14 Then β i0 = 0, and β i 0, for i = 1,..., m. Continuing this process, we can represent x with a positive linear combination of the linearly independent elements {a j j J}, with J {1,..., m}. Linear independence in an n-dimensional space requires that J n. Theorem 2.1.3. (Carathéodory s Theorem) For Ω R n, Ω, any element x co Ω may be expressed as a combination of, at most, n+1 elements of Ω. Proof. Define B := {1} Ω R n+1. We observe that co B = {1} co Ω, and that co B K B. Consider any x co Ω. Take an element (1, x) co B; there exist λ i 0 and (1, a i ) B, i = 0,..., m with m n so that (1, x) = λ i (1, a i ). i=0 We thus conclude that x = m i=0 λ ia i, m i=0 λ i = 1 with λ i 0 and m n. 2.2 Radon Theorem Next we shall prove a lemma, Radon s and then Helly s theorem. Lemma 2.2.1. For a set of vectors in R n, Υ = {υ 1,..., υ m }, if Υ n + 2, then the vectors are affine dependent. Proof. Consider the set Υ = {υ 2 υ 1,..., υ m υ 1 }. Since Υ n + 1, those vectors are linearly dependent, and it follows that the vectors {υ 1,..., υ m } are affine dependent. Theorem 2.2.2. (Radon s Theorem) Define the set Ω := {ω 1,..., ω m }, with Ω n + 2, in R n. There exist two nonempty, disjoint subsets Ω 1 Ω and Ω 2 Ω so that Ω 1 Ω 2 = Ω and co Ω 1 co Ω 2.

15 Proof. Since we have that Ω = m n + 2, the vectors {ω 1,..., ω m } are affine dependent by 2.2.1. So there exist real numbers λ 1,..., λ m, some of which are positive, so that λ i ω i = 0, λ i = 0. We may define the index sets I 1 := {1,..., m λ i 0} and I 2 := {1,..., m λ i < 0}. Furthermore, I 1, I 2 and i I 1 λ i = i I 2 λ i. Define λ := i I 1 λ i, so we have λ i ω i = λ i ω i and i I 1 i I 2 i I 1 λ i λ ω i = i I 2 λ i λ ω i. Define sets Ω 1 := {ω i i I 1 } and Ω 2 := {ω i i I 2 }. These sets are nonempty and disjoint subsets of Ω with Ω 1 Ω 2 = Ω and co Ω 1 co Ω 2. Thus we have i I 1 λ i λ ω i = i I 2 λ i λ ω i co Ω 1 co Ω 2. The proof is now complete. 2.3 Helly Theorem In this section we present Helly s theorem and its proof. The proof presented below is based on induction and Radon s theorem. Theorem 2.3.1. (Helly s Theorem) Consider O := {Ω 1,..., Ω m }, a collection of convex sets in R n, with O n + 1. If the intersection of any n+1 of these sets is nonempty, then all the sets in the collection O have a nonempty intersection; more formally m Ω i. Proof. We shall prove the theorem by induction on O. For m = n + 1, the result of the theorem is trivial. For the induction step, suppose that the theorem holds for O = m n + 1. Consider a collection of convex sets in R n, {Ω 1,..., Ω m, Ω m+1 }, with the property

16 that any n + 1 of the sets have a nonempty intersection. For i = 1,..., m + 1, define the following: Θ i := m+1 j=1,j i Ω j, which has the property that Θ i Ω j whenever j i and i, j {1,..., m + 1}. By hypothesis, Θ i, so there exists an element ω i Θ i for each i {1,..., m + 1}. Define the set W := {ω 1,..., ω m+1 }. Applying Radon s theorem on the set W, we may find two nonempty, disjoint subsets W 1 := {ω i i I 1 }, and W 2 := {ω i i I 2 }. These subsets have the following properties: W 1 W 2 = W, and co W 1 co W 2. We may select an element w co W 1 co W 2, and show that w m+1 Ω i. In order to do this, we first note that i j for i I 1 and j I 2, which implies that θ i Ω j when i I 1 and j I 2. For a particular i {1,..., m + 1} and i I 1, ω i Θ i Ω j for any j I 2. Because Ω i is convex, ω co W 2 = co {ω j j I 2 } Ω i. Thus we have that ω Ω i for any i I 1. We may, by a similar argument, show that ω Ω i for every i I 2.

Chapter 3 The Theorems of Carathédory, Radon, and Helly: Their Equivalence 3.1 Carathéodory s and Helly s Theorem In this section we study the equivalence of the theorems of Carathéodory and Helly based on subdifferential properties of distance functions. Let us start with two useful lemmas. Lemma 3.1.1. Let Ω i, i = 1,..., m, be nonempty closed, convex sets. If at least one of the sets Ω i is bounded, then the maximum function f(x) := max{d(x; Ω i ) i = 1,..., m}, x R n, (3.1.1) has an absolute minimum on R n. Proof. Let γ := inf{f(x) x R n }. Since f(x) 0 for all x R n, it is obvious that γ is a real number. Let {x k } be a sequence in R n such that lim k f(x k ) = γ. Without loss of 17

18 generality, assume that Ω 1 is bounded. By the definition of limit, one finds k 0 IN such that 0 d(x k ; Ω 1 ) f(x k ) < γ + 1 for all k k 0. Choosing w k Ω 1 with x k w k = d(x k ; Ω), we get x k w k γ + 1, and hence x k w k + 1 for all k k 0. Since Ω 1 is bounded, the sequence {x k } is bounded as well, and so we can assume that it has a subsequence {x kl } that converges to x. By the continuity of f, γ = lim l f(x kl ) = f( x) f(x) for all x R n. Therefore, f has an absolute minimum at x. Given any u R n, define the active index set at u, associated with the function f given in (3.1.1), by I(u) := {i = 1,..., m d(u; Ω i ) = f(u)}. Lemma 3.1.2. Let Ω i, i = 1,..., m, be nonempty closed, convex sets with m Ω i =. Consider the function f defined in (3.1.1). Then f has an absolute minimum at x if and only if x co {w i i I( x)}, where w i := Π( x; Ω i ). Proof. The assumption m Ω i = implies that f(x) = max{d(x; Ω i ) i = 1,..., m} > 0 for all x R n and x / Ω i for every i I(x). It follows from Theorem 1.0.8 that the function f has an absolute minimum at x if and only if { x wi } 0 f( x) = co{ d( x; Ω i ) i I( x)} = co d( x; Ω i ) i I( x). Since f( x) = d( x; Ω i ) > 0 for all i I( x), we can denote this common value by r. By

19 Proposition 1.0.1, there exist λ i 0 for i I( x) with i I( x) λ i = 1 such that 0 = i I( x) x w i λ i, r which is equivalent to x co {w i i I( x)}. Now we are ready to prove the main theorem of this section, namely showing that Carathéodory s theorem and Helly s theorem are equivalent in the described sense. Theorem 3.1.3. Consider the following statements: (i) Carathéodory s theorem: For a nonempty convex set Ω R n with x co A there exist λ i 0 and w i Ω for i = 1,..., n + 1 such that n+1 x = λ i ω i. (ii) Helly s theorem: For any collection of nonempty closed, convex sets {Ω 1,..., Ω m }, m n + 2, in R n with the property that the intersection of any n + 1 sets from this collection is nonempty, one has m Ω i. Then statement (i) implies statement (ii), and vice versa. Proof. Let us first prove (ii) assuming that (i) holds. Consider the case where Ω i, for i = 1,..., m, are nonempty closed, convex sets such that at least one of them is bounded. The assumption implies that the intersection of any collection of k sets from the whole collection, where k n + 1, is nonempty. Suppose that m Ω i =. Consider the function f defined by (3.1.1) and let x be a point in R n at which f has an absolute minimum. Note that this point exists by Lemma 3.1.1. From Lemma 3.1.2 one has x co {w i i I( x)},

20 where w i := Π( x; Ω i ). Applying Carathéodory s theorem, we get J I, J n + 1, such that x = j J λ j w i. Defining the function g(x) := max{d(x; Ω j ) j J}, by the given assumption, j J Ω j, there exists a point u in this intersection with g(u) = 0. Since d( x; Ω j ) = r > 0 for all j J, where r := f( x) = g( x) > 0, the active index set at x associated with the function g is J. By Lemma 3.1.2, one has that g has an absolute minimum at x, and hence 0 < r = g( x) g(u) = 0. This contradiction implies the statement (ii). In the case where we do not assume that at least one of the sets is bounded, we choose an element in each intersection of n + 1 sets. Let t > 0 be sufficiently large such that the closed ball B(0; t) covers all such points. Then we only need to apply the previous case for the collection {Θ i i = 1,..., m}, where Θ i := Ω i B(0; t). Now we show how to derive (i) from (ii). Fix any element ȳ co Ω. If ȳ Ω, then it is obviously a convex combination of itself, so the conclusion is obvious. Suppose that ȳ / Ω. We follow and simplify significantly the proof in [7], pp. 40-41. In particular, we do not assume that Ω R n is closed and bounded as in [7], pp. 40-41. For each point x Ω, denote by L 1 ( x) and L 2 ( x) the closed half-spaces with bounding hyperplane passing through x and being perpendicular to the line connecting x and ȳ, where L 2 ( x) contains ȳ, i.e., L 1 ( x) := {w R n x ȳ, w x 0}, L 2 ( x) := {w R n x ȳ, w x 0}.

21 Let us now show that L 1 ( x) =. x Ω Indeed, suppose that this set is nonempty. Then there exists z x Ω L 1( x), implying that x ȳ, z x 0 for all x Ω. It follows that z ȳ, ȳ x = z x + x ȳ, ȳ x = z x, ȳ x + x ȳ, ȳ x = z x, ȳ x x ȳ 2 < 0 for all x Ω (note that x ȳ 2 > 0, since ȳ / Ω). Since ȳ co Ω, there exist x i Ω and λ i 0 for i = 1,..., m such that ȳ = λ i x i and λ i = 1. Then 0 = z ȳ, ȳ ȳ = z ȳ, ȳ λ i x i = This contradiction shows that L 1 ( x) =. x Ω λ i z ȳ, ȳ x i < 0. Since L 1 (x) is a nonempty closed, convex set for every x, by (ii) there exist x 1,..., x n+1 Ω such that n+1 L 1 (x i ) =. However, this implies ȳ co {x 1,..., x n+1 }. Let us prove this by contradiction. Assuming that this is not the case, there exist a, b R n such that a, ȳ > b and a, x i b for all i = 1,..., n + 1.

22 Define l := {ȳ ta t 0}. Then x i ȳ, ȳ ta x i = x i ȳ, ȳ x i t a, x i ȳ. Thus we can find a value t 0 such that this expression is positive for all t > t 0 and for all i, but this implies n+1 L 1(x i ). This contradiction shows that ȳ co {x 1,..., x n+1 }. 3.2 Carathéodory s and Radon s Theorem In this section we study the equivalence of the theorems of Carathéodory and Radon, thus establishing the equivalence of all three theorems. Theorem 3.2.1. Consider the following statements: (i) Carathéodory s theorem: For a nonempty convex set A R n, with x co A, there exist λ i 0 with n+1 λ i = 1 and w i A for i = 1,..., n + 1 such that n+1 x = λ i w i. (ii) Radon s theorem: Given a set Ω := {ω 1,..., ω m }, m n + 2, in R n, there exist two nonempty, disjoint subsets Ω 1 Ω and Ω 2 Ω such that Ω 1 Ω 2 = Ω and co Ω 1 co Ω 2. The statements (i) and (ii) are equivalent.

23 Proof. (i)= (ii): Consider the set Ω R n defined by Ω = {w 1,..., w m } with m n + 2. Then w 1 + + w m m co {w 1,..., w m }. By Carathéodory s theorem, we have the representation w 1 + + w m m n+1 = λ i w i, with λ i 0 and n+1 λ i = 1. It follows that β i w i = λ i w i, where β i = 1 m, i = 1,..., m, λ i 0, n+1 λ i = 1, and λ i = 0 for all i = n + 2,..., m. So We rewrite the above as (β i λ i )w i = 0. γ i w i = 0, with γ i = β i λ i, m γ i = 0, where not all γ i s are zeros. Then γ i x i + γ j x j = 0, i I j J where I := {i {1,..., m} γ i > 0} and J := {i {1,..., m} γ i 0}. Observe that I and J are both nonempty. Define γ := i I γ i, Ω 1 := {w i i I}, and Ω 2 := {w j j J}. Then γ = j J γ j and 1 γ γ i w i = 1 γ j w j co Ω 1 co Ω 2. γ i I j J

24 At this point, we can see that Ω 1 and Ω 2 satisfy the requirements of Radon s theorem. To prove that Carathéodory s theorem follows from Radon s theorem, we complete the approach to [8, Theorem 3.2]. (ii) = (i): Fix any x co A. Then there exist λ i > 0 for i = 1,..., m, with m λ i = 1 and x = λ i w i, where w i A for all i = 1,..., m and the representation is chosen such that m is minimal. We will show that m n+1. Assume the contrary, i.e., that m > n+1. By Radon s theorem applied to the set Ω := {w 1,..., w m } we can assume that, without loss of generality, k γ i w i = i=k+1 γ i w i, where γ i 0 for i = 1,..., m and k γ i = m i=k+1 γ i = 1, 1 k < m. Then we have the representation β i w i = 0, where β i := γ i for i = 1,..., k, and β i = γ i for i = k+1,..., m. Observe that m β i = 0. Choose an index i 0 {1,..., k} such that ɛ := λ i 0 β i0 = min{ λ i β i i = 1,..., kwith β i > 0}. Define α i := λ i ɛβ i for i = 1,..., m. Then α i 0 for i = 1,..., m, α i0 = 0, m α i = 1, and x = α i w i. This contradicts the minimal property of m. The proof is now complete. Remark 3.2.2. (i) Note that the closedness property assumed in Radon s theorem pre-

25 sented in Theorem 3.1.3(ii) can be relaxed, i.e., the statements in Theorem 3.1.3 and Theorem 3.2.1 are equivalent to the following statement: For any collection of nonempty convex sets {Ω 1,..., Ω m }, m n+2, in R n with the property that the intersection of any n + 1 sets from this collection is nonempty, one has m Ω i. Indeed, this statement obviously implies Theorem 3.1.3(ii), and it follows from Theorem 3.2.1(ii); see, e.g., [10, Theorem 3.13]. (ii) As a next step, it would be natural to extend the present investigations to generalized convexity notions, like for example H-convexity (see, e.g., [1] and [2]) or d-convexity (cf. [3, Chapter II]). Summarizing chapters 3 and 4,we have demonstrated the equivalence of Carathéodory s, Radon s and Helly s theorem in convex geometry by demonstrating the following implications: Figure 3.2.1: C=Carathéodory s theorem R=Radon s Theorem H=Helly s Theorem Diagram summarizing the equavalence of the three main theorems.

Bibliography [1] V. Boltyanski, H. Martini: Carathéodory s theorem and H-convexity. J. Combin. Theory, Ser. A 93 (2001), 292-309. [2] V. Boltyanski, H. Martini: Minkowski addition of H-convex sets and related Hellytype theorems. J. Combin. Theory, Ser. A 103 (2003), 323-336. [3] V. Boltyanski, H. Martini, and P. S. Soltan: Excursions into Combinatorial Geometry. Springer, Berlin and Heidelberg, 1997. [4] S. Boyd, L. Vandenberghe: Convex Optimization. Cambridge University Press, Cambridge, England, 2004. [5] L. Danzer, B. Grünbaum, and V. Klee: Helly s theorem and its relatives. Proceedings of Symposia in Pure Mathematics, Vol. 7 Convexity, American Mathematical Society, 1963. [6] J. Eckhoff: Helly, Radon and Carathéodory type theorems. Handbook of Convex Geometry, Vol. 1, P.M. Gruber and J.M. Wills Editors. North-Holland, 1993, pp. [7] H. G. Eggleston: Convexity. Cambridge Tracts in Mathematics and Mathematical Physics, No. 47, Cambridge University Press, 1958. [8] P. M. Gruber: Convex and Discrete Geometry. Springer-Verlag, 2007. 26

27 [9] H. Martini, N.M. Nam, and A. Robinson: On the equivalence of the theorems of Helly, Radon, and Carathéodory via convex analysis, preprint, 2014. [10] B. Mordukhovich, N. M. Nam: An Easy Path to Convex Analysis and Applications, Morgan and Claypool Publishers, 2014. [11] R. T. Rockafellar: Convex Analysis, Princeton University Press, Princeton, NJ, 1970.