Convex Eval

3 Sublinearity and Support Functions

3.1 Sublinear Functions

3.1.1 Definitions and first properties

Definition 3.1.1
#

A function \(\sigma :\mathbb {R}^n\to \mathbb {R}\cup \{ +\infty \} \) is said to be sublinear if it is convex and positively homogeneous (of degree 1): \(\sigma \in \mathrm{Conv}\, \mathbb {R}^n\) and

\[ \sigma (tx)=t\sigma (x)\qquad \text{for all }x\in \mathbb {R}^n\text{ and }t{\gt}0. \tag {1.1.1} \]
Proposition 3.1.1
#

A function \(\sigma :\mathbb R^n\to \mathbb R\cup \{ +\infty \} \) is sublinear if and only if its epigraph \(\operatorname {epi}\sigma \) is a nonempty convex cone in \(\mathbb R^n\times \mathbb R\).

Proof

We know that \(\sigma \) is a convex function if and only if \(\operatorname {epi}\sigma \) is a nonempty convex set in \(\mathbb R^n\times \mathbb R\) (Proposition B.1.1.6). Therefore, we just have to prove the equivalence between positive homogeneity and \(\operatorname {epi}\sigma \) being a cone.

Let \(\sigma \) be positively homogeneous. For \((x,r)\in \operatorname {epi}\sigma \), the relation \(\sigma (x)\le r\) gives

\[ \sigma (tx)=t\sigma (x)\le tr\qquad \text{for all }t{\gt}0, \]

so \(\operatorname {epi}\sigma \) is a cone. Conversely, if \(\operatorname {epi}\sigma \) is a cone in \(\mathbb R^n\times \mathbb R\), the property \((x,\sigma (x))\in \operatorname {epi}\sigma \) implies \((tx,t\sigma (x))\in \operatorname {epi}\sigma \), i.e.

\[ \sigma (tx)\le t\sigma (x)\qquad \text{for all }t{\gt}0. \]

From Remark 1.1.2, this is just positive homogeneity.

Proposition 3.1.2
#

A function \(\sigma :\mathbb {R}^n\to \mathbb {R}\cup \{ +\infty \} \), not identically equal to \(+\infty \), is sublinear if and only if one of the following two properties holds:

\[ \sigma (t_1x_1+t_2x_2)\le t_1\sigma (x_1)+t_2\sigma (x_2)\qquad \text{for all }x_1,x_2\in \mathbb {R}^n\text{ and }t_1,t_2{\gt}0, \tag {1.1.4} \]

or

\[ \sigma \text{ is positively homogeneous and subadditive}. \tag {1.1.5} \]
sublinearity \(\implies \) (1.1.4)

For \(x_1,x_2\in \mathbb {R}^n\) and \(t_1,t_2{\gt}0\), set \(t:=t_1+t_2{\gt}0\); we have

\[ \begin{aligned} \sigma (t_1x_1+t_2x_2)& =\sigma \bigl(t\bigl[\tfrac {t_1}{t}x_1+\tfrac {t_2}{t}x_2\bigr]\bigr)\\ & =t\sigma \bigl(\tfrac {t_1}{t}x_1+\tfrac {t_2}{t}x_2\bigr)\qquad \text{[positive homogeneity]}\\ & \le t\bigl[\tfrac {t_1}{t}\sigma (x_1)+\tfrac {t_2}{t}\sigma (x_2)\bigr]\qquad \text{[convexity]}, \end{aligned} \]

and (1.1.4) is proved.

[(1.1.4) \(\implies \) (1.1.5)] A function satisfying (1.1.4) is obviously subadditive (take \(t_1=t_2=1\)) and satisfies (take \(x_1=x_2=x\), \(t_1=t_2=1/2t\))

\[ \sigma (tx)=t\sigma (x), \]

i.e. it is positively homogeneous.

[(1.1.5)\(\Rightarrow \text{sublinearity}\)] Take \(t_1,t_2{\gt}0\) with \(t_1+t_2=1\) and apply successively subadditivity and positive homogeneity:

\[ \sigma (t_1x_1+t_2x_2)\le \sigma (t_1x_1)+\sigma (t_2x_2)=t_1\sigma (x_1)+t_2\sigma (x_2), \]

hence \(\sigma \) is convex.

Corollary 3.1.1
#

If \(\sigma \) is sublinear, then

\[ \sigma (x)+\sigma (-x)\ge 0\quad \text{for all }x\in \mathbb {R}^n. \tag {1.1.6} \]
Proof

Take \(x_2=-x_1\) in (1.1.3) and remember that \(\sigma (0)\ge 0\).

Proposition 3.1.3
#

Let \(\sigma \) be sublinear and suppose that there exist \(x_1,\dots ,x_m\) in \(\text{dom }\sigma \) such that

\[ \sigma (x_j)+\sigma (-x_j)=0\quad \text{for } j=1,\dots ,m. \tag {1.1.7} \]

Then \(\sigma \) is linear on the subspace spanned by \(x_1,\dots ,x_m\).

Proof

With \(x_1,\dots ,x_m\) as stated, each \(-x_j\) is in \(\text{dom }\sigma \). Let \(x:=\sum _{j=1}^m t_j x_j\) be an arbitrary linear combination of \(x_1,\dots ,x_m\); we must prove that \(\sigma (x)=\sum _{j=1}^m t_j\sigma (x_j)\). Set

\[ J_1 := \{ j : t_j {\gt} 0\} ,\qquad J_2 := \{ j : t_j {\lt} 0\} , \]

and obtain (as usual, \(\sum _\emptyset =0\)):

\begin{align} \sigma (x) & =\sigma \! \biggl(\sum _{j\in J_1} t_j x_j + \sum _{j\in J_2}(-t_j)(-x_j)\biggr)\\ & \le \sum _{j\in J_1} t_j\sigma (x_j)+\sum _{j\in J_2}(-t_j)\sigma (-x_j) \tag *{[from (1.1.4)]}\\ & = \sum _{j\in J_1} t_j\sigma (x_j)+\sum _{j\in J_2} t_j\sigma (x_j)=\sum _{j=1}^m t_j\sigma (x_j) \tag *{[from (1.1.7)]}\\ & = -\sum _{j\in J_1} t_j\sigma (-x_j)-\sum _{j\in J_2}(-t_j)\sigma (x_j) \tag *{[from (1.1.7)]}\\ & \le -\sigma \! \bigl(-\sum _{j=1}^m t_j x_j\bigr)\tag *{[from (1.1.4)]}\\ & = -\sigma (-x)\le \sigma (x). \tag *{[from (1.1.6)]} \end{align}

In summary, we have proved \(\sigma (x)\le \sum _{j=1}^m t_j\sigma (x_j)\le -\sigma (-x)\le \sigma (x).\)

Proposition 3.1.4
#

Let \(\sigma \) be sublinear. If \(x\in U\), i.e. if

\[ \sigma (x)+\sigma (-x)=0, \]

then there holds

\[ \sigma (x+y)=\sigma (x)+\sigma (y)\qquad \text{for all }y\in \mathbb {R}^n. \]
Proof

In view of subadditivity, we just have to prove “\(\ge \)” in (1.1.10). Start from the identity \(y=x+y-x\); apply successively subadditivity and (1.1.9) to obtain

\[ \sigma (y)\le \sigma (x+y)+\sigma (-x)=\sigma (x+y)-\sigma (x). \]

3.1.2 Some examples

Definition 3.1.2
#

(Gauge) Let \(C\) be a closed convex set containing the origin. The function \(\gamma _C\) defined by

\[ \gamma _C(x) := \inf \{ \lambda {\gt}0 : x\in \lambda C\} \tag {1.2.2} \]

is called the gauge of \(C\). As usual, we set \(\gamma _C(x):=+\infty \) if \(x\notin \lambda C\) for no \(\lambda {\gt}0\).

Theorem 3.1.1
#

Let \(C\) be a closed convex set containing the origin. Then

  1. its gauge \(\gamma _C\) is a nonnegative closed sublinear function;

  2. \(\gamma _C\) is finite everywhere if and only if \(0\) lies in the interior of \(C\);

  3. \(C_\infty \) being the asymptotic cone of \(C\),

    \[ \{ x\in \mathbb {R}^n:\ \gamma _C(x)\le r\} =rC\quad \text{for all }r{\gt}0, \qquad \{ x\in \mathbb {R}^n:\ \gamma _C(x)=0\} =C_\infty . \]
(i) and (iii)

Nonnegativity and positive homogeneity are obvious from the definition of \(\gamma _C\); also, \(\gamma _C(0)=0\) because \(0\in C\). We prove convexity via a geometric interpretation of (1.2.2). Let

\[ K_C:=\operatorname {cone}(C\times \{ 1\} )=\{ (\lambda c,\lambda )\in \mathbb {R}^n\times \mathbb {R}:\ c\in C,\ \lambda \ge 0\} \]

be the convex conical hull of \(C\times \{ 1\} \subset \mathbb {R}^n\times \mathbb {R}\). It is convex (beware that \(K_C\) need not be closed) and \(\gamma _C\) is clearly given by

\[ \gamma _C(x)=\inf \{ \lambda :\ (x,\lambda )\in K_C\} . \]

Thus, \(\gamma _C\) is the lower-bound function of §B.1.3(g), constructed on the convex set \(K_C\); this establishes the convexity of \(\gamma _C\), hence its sublinearity.

Now we prove

\begin{equation} \label{eq:1.2.3} \{ x\in \mathbb {R}^n:\ \gamma _C(x)\le 1\} =C. \end{equation}
7

This will imply the first part in (iii), thanks to positive homogeneity. Then the second part will follow because of (A.2.2.2): \(C_\infty =\cap \{ rC:\ r{\gt}0\} \) and closedness of \(\gamma _C\) will also result from (iii) via Proposition B.1.2.2.

So, to prove (1.2.3), observe first that \(x\in C\) implies from (1.2.2) that certainly \(\gamma _C(x)\le 1\). Conversely, let \(x\) be such that \(\gamma _C(x)\le 1\); we must prove that \(x\in C\). For this we prove that \(x_k:= (1-1/k)x\in C\) for \(k=1,2,\dots \) (and then, the desired property will come from the closedness of \(C\)). By positive homogeneity, \(\gamma _C(x_k)=(1-1/k)\gamma _C(x)\le 1\), so there is \(\lambda _k\in [0,1]\) such that \(x_k=\lambda _k C\), or equivalently \(x_k/\lambda _k\in C\). Because \(C\) is convex and contains the origin, \(\lambda _k (x_k/\lambda _k)+(1-\lambda _k)0 = x_k\) is in \(C\), which is what we want.

[(ii)] Assume \(0\in \operatorname {int}C\). There is \(\varepsilon {\gt}0\) such that for all \(x\neq 0\), \(x_\varepsilon := x/\| x\| \in C\); hence \(\gamma _C(x_\varepsilon )\le 1\) because of (1.2.3). We deduce by positive homogeneity

\[ \gamma _C(x)=\frac{\| x\| }{\varepsilon }\, \gamma _C(x_\varepsilon )\le \frac{\| x\| }{\varepsilon }\, ; \]

this inequality actually holds for all \(x\in \mathbb {R}^n\) (\(\gamma _C(0)=0\)) and \(\gamma _C\) is a finite function.

Conversely, suppose \(\gamma _C\) is finite everywhere. By continuity (Theorem B.3.1.2), \(\gamma _C\) has an upper bound \(L{\gt}0\) on the unit ball:

\[ \| x\| \le 1 \quad \Longrightarrow \quad \gamma _C(x)\le L \quad \Longrightarrow \quad x\in L C, \]

where the last implication comes from (iii). In other words, \(B(0,1/L)\subset C\).

Corollary 3.1.2
#

C is compact if and only if \(\gamma _C(x){\gt}0\) for all \(x\neq 0\).

3.1.3 The convex cone of all closed sublinear functions

Proposition 3.1.5
#

If \(\sigma _1\) and \(\sigma _2\) are [closed] sublinear and \(t_1,t_2\) are positive numbers, then \(\sigma := t_1\sigma _1 + t_2\sigma _2\) is [closed] sublinear, if not identically \(+\infty \).

Proof

Concerning convexity and closedness, everything is known from §B.2. Note in passing that a closed sublinear function is zero (hence finite) at zero. As for positive homogeneity, it is straightforward.

Proposition 3.1.6
#

If \(\{ \sigma _j\} _{j\in J}\) is a family of [closed] sublinear functions, then \(\sigma := \sup _{j\in J}\sigma _j\) is [closed] sublinear, if not identically \(+\infty \).

Proof

Concerning convexity and closedness, everything is known from §B.2. Note in passing that a closed sublinear function is zero (hence finite) at zero. As for positive homogeneity, it is straightforward.

Proposition 3.1.7
#

Let \(\{ \sigma _j\} _{j\in J}\) be a family of sublinear functions all minorized by some linear function. Then \(\sigma := \operatorname {co}\big(\inf _{j\in J}\sigma _j\big)\) is sublinear.

Proof

Once again, the only thing to prove for (i) is positive homogeneity. Actually, it suffices to multiply \(x\) and each \(x_j\) by \(t{\gt}0\) in a formula giving \(\operatorname {co}\big(\inf _j\sigma _j\big)(x)\), say (B.2.5.3).

Proposition 3.1.8
#

Let \(\{ \sigma _j\} _{j\in J}\) be a family of sublinear functions all minorized by some linear function. Then if \(J=\{ 1,\dots ,m\} \) is a finite set, we obtain the infimal convolution

\[ \operatorname {co}\min \{ \sigma _1,\dots ,\sigma _m\} =\sigma _1\mathbin {\square }\cdots \mathbin {\square }\sigma _m. \]
Proof

By definition, computing \(\operatorname {co}\big(\min _j\sigma _j\big)(x)\) amounts to solving the minimization problem in the \(m\) couples of variables \((x_j,\alpha _j)\in \operatorname {dom}\sigma _j\times \mathbb {R}\)

\[ \begin{aligned} & \inf \sum _{j=1}^m\alpha _j\sigma _j(x_j)\qquad \alpha _j\ge 0\\ & \sum _{j=1}^m\alpha _j=1,\ \sum _{j=1}^m\alpha _j x_j=x. \end{aligned} \tag {1.3.1} \]

In view of positive homogeneity, the variables \(\alpha _j\) play no role by themselves: the relevant variables are actually the products \(\alpha _j x_j\) and (1.3.1) can be written – denoting \(\alpha _j x_j\) again by \(x_j\):

\[ \operatorname {co}\big(\min _j\sigma _j\big)(x)=\inf \Big\{ \sum _{j=1}^m\sigma _j(x_j):\sum _{j=1}^m x_j=x\Big\} . \]

We recognize the infimal convolution of the \(\sigma _j\)’s.

Theorem 3.1.2

For \(\sigma _1\) and \(\sigma _2\) in the set \(\Phi \) of sublinear functions that are finite everywhere, define

\[ \Delta (\sigma _1,\sigma _2):=\max _{\| x\| \le 1}|\sigma _1(x)-\sigma _2(x)|. \tag {1.3.2} \]

Then \(\Delta \) is a distance on \(\Phi \).

Proof

Clearly \(\Delta (\sigma _1,\sigma _2){\lt}+\infty \) and \(\Delta (\sigma _1,\sigma _2)=\Delta (\sigma _2,\sigma _1)\). Now positive homogeneity of \(\sigma _1\) and \(\sigma _2\) gives for all \(x\neq 0\)

\[ |\sigma _1(x)-\sigma _2(x)|=\| x\| \Big|\sigma _1\Big(\frac{x}{\| x\| }\Big)-\sigma _2\Big(\frac{x}{\| x\| }\Big)\Big| \le \| x\| \max _{\| u\| =1}|\sigma _1(u)-\sigma _2(u)| \le \| x\| \, \Delta (\sigma _1,\sigma _2). \]

In addition, \(\sigma _1(0)=\sigma _2(0)=0\), so

\[ |\sigma _1(x)-\sigma _2(x)|\le \| x\| \, \Delta (\sigma _1,\sigma _2)\qquad \text{for all }x\in \mathbb R^n \]

and \(\Delta (\sigma _1,\sigma _2)=0\) if and only if \(\sigma _1=\sigma _2\).

As for the triangle inequality, we have for arbitrary \(\sigma _1,\sigma _2,\sigma _3\) in \(\Phi \)

\[ |\sigma _1(x)-\sigma _3(x)|\le |\sigma _1(x)-\sigma _2(x)|+|\sigma _2(x)-\sigma _3(x)|\qquad \text{for all }x\in \mathbb R^n, \]

so there holds

\[ \Delta (\sigma _1,\sigma _3)\le \max _{\| x\| =1}\big[|\sigma _1(x)-\sigma _2(x)|+|\sigma _2(x)-\sigma _3(x)|\big] \le \max _{\| x\| =1}|\sigma _1(x)-\sigma _2(x)|+\max _{\| x\| =1}|\sigma _2(x)-\sigma _3(x)|, \]

which is the required inequality.

Theorem 3.1.3
#

Let \((\sigma _k)\) be a sequence of finite sublinear functions and let \(\sigma \) be a finite function. Then the following are equivalent when \(k\to +\infty \):

  1. \((\sigma _k)\) converges pointwise to \(\sigma \);

  2. \((\sigma _k)\) converges to \(\sigma \) uniformly on each compact set of \(\mathbb {R}^n\);

  3. \(\Delta (\sigma _k,\sigma )\to 0\).

Proof

First, the (finite) function \(\sigma \) is of course sublinear whenever it is the pointwise limit of sublinear functions. The equivalence between (i) and (ii) comes from the general Theorem B.3.1.4 on the convergence of convex functions.

Now, (ii) clearly implies (iii). Conversely \(\Delta (\sigma _k,\sigma )\to 0\) is the uniform convergence on the unit ball, hence on any ball of radius \(L{\gt}0\) (the maxmind in (1.3.2) is positively homogeneous), hence on any compact set.

3.2 The Support Function of a Nonempty Set

3.2.1 Definitions, Interpretations

Definition 3.2.1
#

( Support Function) Let \(S\) be a nonempty set in \(\mathbb {R}^n\). The function

\[ \sigma _S:\mathbb {R}^n\to \mathbb {R}\cup \{ +\infty \} \]

defined by

\[ \mathbb {R}^n\ni x\mapsto \sigma _S(x):=\sup \{ \langle s,x\rangle :\; s\in S\} \tag {2.1.1} \]

is called the support function of \(S\).

Proposition 3.2.1
#

A support function is closed and sublinear.

Proof

This results from Proposition 1.3.1(ii) (a linear form is closed and convex!). Observe in particular that a support function is null (hence \({\lt} +\infty \)) at the origin.

Proposition 3.2.2
#

The support function of \(S\) is finite everywhere if and only if \(S\) is bounded.

Proof

Let \(S\) be bounded, say \(S\subset B(0,L)\) for some \(L{\gt}0\). Then

\[ \langle s,x\rangle \le \| s\| \, \| x\| \le L\| x\| \qquad \text{for all } s\in S, \]

which implies \(\sigma _S(x)\le L\| x\| \) for all \(x\in \mathbb {R}^n\).

Conversely, finiteness of the convex \(\sigma _S\) implies its continuity on the whole space (Theorem B.3.1.2), hence its local boundedness: for some \(L\),

\[ \langle s,x\rangle \le \sigma _S(x)\le L\qquad \text{for all }(s,x)\in S\times B(0,1). \]

If \(s\neq 0\), we can take \(x=s/\| s\| \) in the above relation, which implies \(\| s\| \le L\).

Definition 3.2.2
#

(Breadth of a Set) The breadth of a nonempty set \(S\) along \(x\neq 0\) is

\[ \sigma _S(x)+\sigma _S(-x)=\sup _{s\in S}\langle s,x\rangle -\inf _{s\in S}\langle s,x\rangle , \]

a number in \([0,+\infty ]\). It is \(0\) if and only if \(S\) lies entirely in some affine hyperplane orthogonal to \(x\); such a hyperplane is expressed as

\[ \{ y\in \mathbb {R}^n:\ \langle y,x\rangle =\sigma _S(x)\} , \]

which in particular contains \(S\). The intersection of all these hyperplanes is just the affine hull of \(S\).

3.2.2 Basic properties

Proposition 3.2.3
#

For \(S\subset \mathbb {R}^n\) nonempty, there holds \(\sigma _S=\sigma _{\operatorname {cl} S}=\sigma _{\operatorname {co} S}\); whence

\[ \sigma _S=\sigma _{\overline{\operatorname {co}}\, S}. \tag {2.2.1} \]
Proof

The continuity [resp. linearity, hence convexity] of the function \(\langle s,\cdot \rangle \), which is maximized over \(S\), implies that \(\sigma _S=\sigma _{\operatorname {cl} S}\) [resp. \(\sigma _S=\sigma _{\operatorname {co} S}\)]. Knowing that \(\overline{\operatorname {co}}\, S=\operatorname {cl}\, \operatorname {co}\, S\) (Proposition A.1.4.2), (2.2.1) follows immediately.

Theorem 3.2.1
#

For the nonempty \(S\subset \mathbb {R}^n\) and its support function \(\sigma _S\), there holds

\[ s\in \overline{\operatorname {co}}S\iff \bigl[\langle s,d\rangle \le \sigma _S(d)\quad \text{for all }d\in X\bigr], \tag {2.2.2} \]

where the set \(X\) can be indifferently taken as: the whole of \(\mathbb {R}^n\), the unit ball \(B(0,1)\) or its boundary the unit sphere \(\tilde{B}\), or \(\operatorname {dom}\sigma _S\).

Proof

First, the equivalence between all the choices for \(X\) is clear enough; in particular due to positive homogeneity. Because “\(\Rightarrow \)” is Proposition 2.2.1, we have to prove “\(\Leftarrow \)” only, with \(X=\mathbb {R}^n\) say.

So suppose that \(s\notin \overline{\operatorname {co}}S\). Then \(\{ s\} \) and \(\overline{\operatorname {co}}S\) can be strictly separated (Theorem A.4.1.1): there exists \(d_0\in \mathbb {R}^n\) such that

\[ \langle s,d_0\rangle {\gt}\sup \{ \langle s',d_0\rangle :s'\in \overline{\operatorname {co}}S\} =\sigma _S(d_0), \]

where the last equality is (2.2.1). Our result is proved by contradiction.

Theorem 3.2.2
#

Let \(S\) be a nonempty closed convex set in \(\mathbb {R}^n\). Then

  1. \(s\in \operatorname {aff}S\) if and only if

    \[ \langle s,d\rangle =\sigma _S(d)\quad \text{for all }d\text{ with }\sigma _S(d)+\sigma _S(-d)=0; \tag {2.2.3} \]
  2. \(s\in \operatorname {ri}S\) if and only if

    \[ \langle s,d\rangle {\lt}\sigma _S(d)\quad \text{for all }d\text{ with }\sigma _S(d)+\sigma _S(-d){\gt}0; \tag {2.2.4} \]
  3. in particular, \(s\in \operatorname {int}S\) if and only if

    \[ \langle s,d\rangle {\lt}\sigma _S(d)\quad \text{for all }d\neq 0. \tag {2.2.5} \]
(i)

Let first \(s\in S\). We have already seen in Definition 2.1.4 that

\[ - \sigma _S(-d) \le \langle s,d\rangle \le \sigma _S(d)\qquad \text{for all } d\in \mathbb {R}^n. \]

If the breadth of \(S\) along \(d\) is zero, we obtain a pair of equalities: for such \(d\), there holds

\[ \langle s,d\rangle = \sigma _S(d), \]

an equality which extends by affine combination to any element \(s\in \operatorname {aff}S\).

Conversely, let \(s\) satisfy (2.2.3). A first case is when the only \(d\) described in (2.2.3) is \(d = 0\); as a consequence of our observations in Definition 2.1.4, there is no affine hyperplane containing \(S\), i.e. \(\operatorname {aff}S = \mathbb {R}^n\) and there is nothing to prove. Otherwise, there does exist a hyperplane \(H\) containing \(S\); it is defined by

\[ H := \{ p \in \mathbb {R}^n : \langle p,d_H\rangle = \sigma _S(d_H)\} , \tag {2.2.6} \]

for some \(d_H \neq 0\). We proceed to prove \(\langle s,\cdot \rangle \le \sigma _H\).

In fact, the breadth of \(S\) along \(d_H\) is certainly 0, hence \(\langle s,d_H\rangle = \sigma _S(d_H)\) because of (2.2.3), while (2.2.6) shows that \(\sigma _S(d_H) = \sigma _H(d_H)\). On the other hand, it is obvious that \(\sigma _H(d) = +\infty \) if \(d\) is not collinear to \(d_H\). In summary, we have proved \(\langle s,d\rangle \le \sigma _H(d)\) for all \(d\), i.e. \(s \in H\). We conclude that our \(s\) is in any affine manifold containing \(S\): \(s \in \operatorname {aff}S\).

[(iii)] In view of positive homogeneity, we can normalize \(d\) in (2.2.5). For \(s \in \operatorname {int} S\), there exists \(\varepsilon {\gt} 0\) such that \(s + \varepsilon d \in S\) for all \(d\) in the unit sphere \(\widetilde{B}\). Then, from the very definition (2.1.1),

\[ \sigma _S(d) \ge \langle s + \varepsilon d,d\rangle = \langle s,d\rangle + \varepsilon \quad \text{for all } d \in \widetilde{B}. \]

Conversely, let \(s \in \mathbb {R}^n\) be such that

\[ \sigma _S(d) - \langle s,d\rangle {\gt} 0 \quad \text{for all } d \in \widetilde{B}, \]

which implies, because \(\sigma _S\) is closed and the unit sphere is compact:

\[ 0 {\lt} \varepsilon := \inf \{ \sigma _S(d) - \langle s,d\rangle : d \in \widetilde{B}\} \le +\infty . \]

Thus

\[ \langle s,d\rangle + \varepsilon \le \sigma _S(d)\quad \text{for all } d \in \widetilde{B}. \]

Now take \(u\) with \(\| u\| {\lt} \varepsilon \). From the Cauchy-Schwarz inequality, we have for all \(d \in \widetilde{B}\)

\[ \langle s+u,d\rangle = \langle s,d\rangle + \langle u,d\rangle \le \langle s,d\rangle + \varepsilon \le \sigma _S(d) \]

and this implies \(s + u \in S\) because of Theorem 2.2.2: \(s \in \operatorname {int} S\) and (iii) is proved.

[(ii)] Look at Fig. 2.2.2 again: decompose \(\mathbb {R}^n = V \oplus U\), where \(V\) is the subspace parallel to \(\operatorname {aff}S\) and \(U = V^\perp \). In the decomposition \(d = d_V + d_U\), \(\langle \cdot ,d_U\rangle \) is constant over \(S\), so \(S\) has 0-breadth along \(d_U\) and

\[ \sigma _S(d) = \sup _{s\in S}\langle s,d_V + d_U\rangle = \langle s,d_U\rangle + \sup _{s\in S}\langle s,d_V\rangle \]

for any \(s \in S\). With these notations, a direction described as in (2.2.4) is a \(d\) such that

\[ \sigma _S(d) + \sigma _S(-d) = \sigma _S(d_V) + \sigma _S(-d_V) {\gt} 0. \]

Then, (ii) is just (iii) written in the subspace \(V\).

Proposition 3.2.4
#

Let \(S\) be a nonempty closed convex set in \(\mathbb {R}^n\). Then \(\overline{\operatorname {dom}\sigma _S}\) and the asymptotic cone \(S_\infty \) of \(S\) are mutually polar cones.

Proof

Recall from §A.3.2 that, if \(K_1\) and \(K_2\) are two closed convex cones, then \(K_1\subset K_2\) if and only if \((K_1)^\circ \supset (K_2)^\circ \).

Let \(p\in S_\infty \). Fix \(s_0\) arbitrary in \(S\) and use the fact that \(S_\infty = \bigcap _{t{\gt}0} t(S-s_0)\) ( §A.2.2); for all \(t{\gt}0\), we can find \(s_t\in S\) such that \(p=t(s_t-s_0)\). Now, for \(q\in \operatorname {dom}\sigma _S\), there holds

\[ \langle p,q\rangle = t\langle s_t-s_0,q\rangle \le t\big[\sigma _S(q)-\langle s_0,q\rangle \big]{\lt}+\infty \]

and letting \(t\downarrow 0\) shows that \(\langle p,q\rangle \le 0\). In other words, \(\operatorname {dom}\sigma _S\subset (S_\infty )^\circ \); then \(\overline{\operatorname {dom}\sigma _S}\subset (S_\infty )^\circ \) since the latter is closed.

Conversely, let \(q\in (\operatorname {dom}\sigma _S)^\circ \), which is a cone, hence \(tq\in (\operatorname {dom}\sigma _S)^\circ \) for any \(t{\gt}0\). Thus, given \(s_0\in S\), we have for arbitrary \(p\in \operatorname {dom}\sigma _S\)

\[ \langle s_0+tq,p\rangle = \langle s_0,p\rangle + t\langle q,p\rangle \le \langle s_0,p\rangle \le \sigma _S(p), \]

so \(s_0+tq\in S\) by virtue of Theorem 2.2.2. In other words: \(q\in (S-s_0)/t\) for all \(t{\gt}0\) and \(q\in S_\infty \).

3.2.3 Examples

3.3 The Isomorphism Between Closed Convex Sets and Closed Sublinear Functions

3.3.1 The fundamental correspondence

Theorem 3.3.1
#

Let \(\sigma \) be a closed sublinear function; then there is a linear function minorizing \(\sigma \). In fact, \(\sigma \) is the supremum of the linear functions minorizing it. In other words, \(\sigma \) is the support function of the nonempty closed convex set

\[ S_\sigma := \{ s\in \mathbb {R}^n : \langle s,d\rangle \le \sigma (d)\ \text{for all } d\in \mathbb {R}^n\} . \tag {3.1.1} \]
Proof

Being convex, \(\sigma \) is minorized by some affine function (Proposition B.1.2.1): for some \((s,r)\in \mathbb {R}^n\times \mathbb {R}\),

\[ \langle s,d\rangle - r \le \sigma (d)\qquad \text{for all } d\in \mathbb {R}^n. \tag {3.1.2} \]

Because \(\sigma (0)=0\), the above \(r\) is nonnegative. Also, by positive homogeneity,

\[ \langle s,d\rangle - \tfrac {1}{t}r \le \sigma (d)\qquad \text{for all } d\in \mathbb {R}^n\ \text{and all } t{\gt}0. \]

Letting \(t\to +\infty \), we see that \(\sigma \) is actually minorized by a linear function:

\[ \langle s,d\rangle \le \sigma (d)\qquad \text{for all } d\in \mathbb {R}^n. \tag {3.1.3} \]

Now observe that the minorization (3.1.3) is sharper than (3.1.2): when expressing the closed convex \(\sigma \) as the supremum of all the affine functions minorizing it (Proposition B.1.2.8), we can restrict ourselves to linear functions. In other words

\[ \sigma (d)=\sup \{ \langle s,d\rangle :\ \text{the linear }\langle s,\cdot \rangle \ \text{minorizes }\sigma \} ; \]

in the above index-set, we just recognize \(S_\sigma \).

Corollary 3.3.1
#

For a nonempty closed convex set \(S\) and a closed sublinear function \(\sigma \), the following are equivalent:

  1. \(\sigma \) is the support function of \(S\).

  2. \(S=\{ s:\ \langle s,d\rangle \le \sigma (d)\text{ for all }d\in X\} ,\) where the set \(X\) can be indifferently taken as: the whole of \(\mathbb {R}^n\), the unit ball \(B(0,1)\) or its boundary, or \(\operatorname {dom}\sigma \).

Proof

The case \(X=\mathbb {R}^n\) is just Theorem 3.1.1. The other cases are then clear.

Definition 3.3.1 Direction Exposing a Face
#

Let \(C\) be a nonempty closed convex set, with support function \(\sigma \). For given \(d\neq 0\), the set

\[ F_C(d):=\{ x\in C:\ \langle x,d\rangle =\sigma (d)\} \]

is called the exposed face of \(C\) associated with \(d\), or the face exposed by \(d\).

Proposition 3.3.1
#

For \(x\) in a nonempty closed convex set \(C\), it holds

\[ x\in F_C(d)\iff d\in N_C(x). \]
Proposition 3.3.2

For a nonempty closed convex set \(C\), it holds

\[ \operatorname {bd} C = \bigcup \{ F_C(d):\; d\in X\} \]

where \(X\) can be indifferently taken as: \(\mathbb {R}^n\setminus \{ 0\} \), the unit sphere \(\widetilde{B}\), or \(\operatorname {dom}\sigma _C\setminus \{ 0\} \).

Proof

Observe from Definition 3.1.3 that the face exposed by \(d\neq 0\) does not depend on \(\| d\| \). This establishes the equivalence between the first two choices for \(X\). As for the third choice, it is due to the fact that \(F_C(d)=\varnothing \) if \(d\notin \operatorname {dom}\sigma _C\).

Now, if \(x\) is interior to \(C\) and \(d\neq 0\), then \(x+\varepsilon d\in C\) and \(x\) cannot be a maximizer of \(\langle \cdot ,d\rangle \); \(x\) is not in the face exposed by \(d\). Conversely, take \(x\) on the boundary of \(C\). Then \(N_C(x)\) contains a nonzero vector \(d\); by Proposition 3.1.4, \(x\in F_C(d)\).

3.3.2 Example: Norms and Their Duals, Polarity

Proposition 3.3.3

Let \(B\) and \(B^*\) be defined by (3.2.1) and (3.2.2), where \(\| \cdot \| \) is a norm on \(\mathbb {R}^n\). The support function of \(B\) and the gauge of \(B^*\) are the same function \(\| \cdot \| _* \) defined by

\[ \| s\| _* := \max \{ \langle s,x\rangle : \| x\| \le 1\} . \tag {3.2.3} \]

Furthermore, \(\| \cdot \| _*\) is a norm on \(\mathbb {R}^n\). The support function of its unit ball \(B^*\) and the gauge of its supported set \(B\) are the same function \(\| \cdot \| \): there holds

\[ \| x\| = \max \{ \langle s,x\rangle : \| s\| _* \le 1\} . \tag {3.2.4} \]
Proof

It is a particular case of the results 3.2.4 and 3.2.5 below.

Proposition 3.3.4
#

Let \(C\) be a closed convex set containing the origin. Its gauge \(\gamma _C\) is the support function of a closed convex set containing the origin, namely

\[ C^\circ := \{ s\in \mathbb {R}^n : \langle s,d\rangle \le 1\ \text{for all } d\in C\} , \tag {3.2.8} \]

which defines the polar (set) of \(C\).

Proof

We know that \(\gamma _C\) (which, by Theorem 1.2.5(i), is closed, sublinear and nonnegative) is the support function of some closed convex set containing the origin, say \(D\); from (3.1.1),

\[ D=\{ s\in \mathbb {R}^n:\ \langle s,d\rangle \le r\ \text{ for all }(d,r)\in \operatorname {epi}\gamma _C\} . \]

As seen in (1.2.4), \(\operatorname {epi}\gamma _C\) is the closed convex conical hull of \(C\times \{ 1\} \); we can use positive homogeneity to write

\[ D=\{ s\in \mathbb {R}^n:\ \langle s,d\rangle \le 1\ \text{ for all }d\ \text{such that }\gamma _C(d)\le 1\} . \]

In view of Theorem 1.2.5(iii), the above index-set is just \(C\); in other words, \(D=C^\circ \).

Corollary 3.3.2
#

Let \(C\) be a closed convex set containing the origin. Its support function \(\sigma _C\) is the gauge of \(C^\circ \).

Proposition 3.3.5
#

Let \(C\) be a nonempty compact convex set having \(0\) in its interior, so that \(C^\circ \) enjoys the same properties. Then, for all \(d\) and \(s\) in \(\mathbb {R}^n\), the following statements are equivalent (the notation (3.2.9) is used)

  1. \(H(s)\) is a supporting hyperplane to \(C\) at \(d\);

  2. \(H(d)\) is a supporting hyperplane to \(C^\circ \) at \(s\);

  3. \(d\in \operatorname {bd}C,\ s\in \operatorname {bd}C^\circ \text{ and }\langle s,d\rangle =1\);

  4. \(d\in C,\ s\in C^\circ \text{ and }\langle s,d\rangle =1\).

Proof

Left as an exercise; the assumptions are present to make sure that every nonzero vector in \(\mathbb {R}^n\) does expose a face in each set.

3.3.3 Calculus with support functions

Theorem 3.3.2
#

Let \(S_1\) and \(S_2\) be nonempty closed convex sets; call \(\sigma _1\) and \(\sigma _2\) their support functions. Then

\[ S_1\subset S_2 \iff \sigma _1(d)\le \sigma _2(d)\text{ for all }d\in \mathbb {R}^n. \]
Proof

Apply the equivalence stated in Corollary 3.1.2:

\[ S_1\subset S_2 \iff s\in S_2\text{ for all }s\in S_1 \]
\[ \iff \sigma _2(d)\ge \langle s,d\rangle \text{ for all }s\in S_1\text{ and all }d\in \mathbb {R}^n \]
\[ \iff \sigma _2(d)\ge \sup _{s\in S_1}\langle s,d\rangle \text{ for all }d\in \mathbb {R}^n. \]
Theorem 3.3.3
#

Let \(\sigma _1\) and \(\sigma _2\) be the support functions of the nonempty closed convex sets \(S_1\) and \(S_2\). If \(t_1\) and \(t_2\) are positive, then

\[ t_1\sigma _1+t_2\sigma _2\ \text{is the support function of }\ \overline{(t_1S_1+t_2S_2)}. \]
Proof

Call \(S\) the closed convex set \(\operatorname {cl}(t_1S_1+t_2S_2)\). By definition, its support function is

\[ \sigma _S(d)=\sup \{ \langle t_1s_1+t_2s_2,d\rangle :\; s_1\in S_1,\; s_2\in S_2\} . \]

In the above expression, \(s_1\) and \(s_2\) run independently in their index sets \(S_1\) and \(S_2\), \(t_1\) and \(t_2\) are positive, so

\[ \sigma _S(d)=t_1\sup _{s\in S_1}\langle s,d\rangle +t_2\sup _{s\in S_2}\langle s,d\rangle . \]
Theorem 3.3.4
#

Let \(\{ \sigma _j\} _{j\in J}\) be the support functions of the family of nonempty closed convex sets \(\{ S_j\} _{j\in J}\). Then

\[ \sup _{j\in J}\sigma _j\text{ is the support function of }\overline{\operatorname {co}}\{ \, \bigcup _{j\in J}S_j:\; j\in J\} . \]
Proof

The support function of \(S:=\bigcup _{j\in J}S_j\) is

\[ \sup _{s\in \bigcup _{j\in J}S_j}\langle s,d\rangle =\sup _{j\in J}\sup _{s_j\in S_j}\langle s_j,d\rangle =\sup _{j\in J}\sigma _j(d). \]

This implies (ii) since \(\sigma _S=\sigma _{\overline{\operatorname {co}}\, S}\).

Theorem 3.3.5
#

Let \(\{ \sigma _j\} _{j\in J}\) be the support functions of the family of closed convex sets \(\{ S_j\} _{j\in J}\). If

\[ S:=\bigcap _{j\in J}S_j\neq \varnothing , \]

then

\[ \sigma _S=\overline{\operatorname {co}}\{ \inf \sigma _j:\; j\in J\} . \]
Proof

The set \(S:=\bigcap _j S_j\) being nonempty, it has a support function \(\sigma _S\). Now, from Corollary 3.1.2,

\[ s\in S\iff s\in S_j\text{ for all }j\in J \]
\[ \iff \langle s,\cdot \rangle \le \sigma _j\text{ for all }j\in J \]
\[ \iff \langle s,\cdot \rangle \le \inf _{j\in J}\sigma _j \]
\[ \iff \langle s,\cdot \rangle \le \overline{\operatorname {co}}\bigl(\inf _{j\in J}\sigma _j\bigr), \]

where the last equivalence comes directly from the Definition B.2.5.3 of a closed convex hull. Again Corollary 3.1.2 tells us that the closed sublinear function \(\overline{\operatorname {co}}(\inf \sigma _j)\) is just the support function of \(S\).

Proposition 3.3.6
#

Let \(A:\mathbb {R}^n\to \mathbb {R}^m\) be a linear operator, with adjoint \(A^*\) (for some scalar product \(\langle \cdot ,\cdot \rangle \) in \(\mathbb {R}^m\)). For \(S\subset \mathbb {R}^n\) nonempty, we have

\[ \sigma _{\mathrm{cl}\, A(S)}(y)=\sigma _{S}(A^*y)\qquad \text{for all }y\in \mathbb {R}^m. \]
Proof

Just write the definitions

\[ \sigma _{A(S)}(y)=\sup _{s\in S}\langle As,y\rangle =\sup _{s\in S}\langle s,A^*y\rangle \]

and use Proposition 2.2.1 to obtain the result.

Proposition 3.3.7
#

Let \(A:\mathbb {R}^m\to \mathbb {R}^n\) be a linear operator, with adjoint \(A^*\) (for some scalar product \(\langle \cdot ,\cdot \rangle \) in \(\mathbb {R}^m\)). Let \(\sigma \) be the support function of a nonempty closed convex set \(S\subset \mathbb {R}^m\). If \(\sigma \) is minorized on the inverse image

\[ A^{-1}(d)=\{ p\in \mathbb {R}^m:\; Ap=d\} \tag {3.3.3} \]

of each \(d\in \mathbb {R}^n\), then the support function of the set \((A^{-1})^*(S)\) is the closure of the image-function \(A\sigma \).

Proof

Our assumption is tailored to guarantee \(A\sigma \in \operatorname {Conv}\mathbb {R}^n\) (Theorem B.2.4.2). The positive homogeneity of \(A\sigma \) is clear: for \(d\in \mathbb {R}^n\) and \(t{\gt}0\),

\[ (A\sigma )(td)=\inf _{Ap=td}\sigma (p)=\inf _{A(p/t)=d}t\sigma (p/t)=t\inf _{Aq=d}\sigma (q)=t(A\sigma )(d). \]

Thus, the closed sublinear function \(\operatorname {cl}(A\sigma )\) supports some set \(S'\); by definition, \(s\in S'\) if and only if

\[ \langle s,d\rangle \le \inf \{ \sigma (p):\; Ap=d\} \qquad \text{for all }d\in \mathbb R^n; \]

but this just means

\[ \langle s,Ap\rangle \le \sigma (p)\qquad \text{for all }p\in \mathbb R^m, \]

i.e. \(A^*s\in S\), because \(\langle s,Ap\rangle =\langle A^*s,p\rangle \).

Theorem 3.3.6

Let \(S\) and \(S'\) be two nonempty compact convex sets of \(\mathbb {R}^n\). Then

\[ \Delta (\sigma _S,\sigma _{S'}) := \max _{\| d\| \le 1} |\sigma _S(d)-\sigma _{S'}(d)| = \Delta _H(S,S') . \tag {3.3.5} \]
Proof

As mentioned in §0.5.1, for all \(r\ge 0\), the property

\[ \max \{ d_S(d):\, d\in S'\} \le r \tag {3.3.6} \]

simply means \(S'\subset S+B(0,r)\).

Now, the support function of \(B(0,1)\) is \(\| \cdot \| \) — see (2.3.1). Calculus rules on support functions therefore tell us that (3.3.6) is also equivalent to

\[ \sigma _{S'}(d)\le \sigma _S(d)+r\| d\| \qquad \text{for all }d\in \mathbb {R}^n, \]

which in turn can be written

\[ \max _{\| d\| \le 1}\big[\sigma _{S'}(d)-\sigma _S(d)\big]\le r. \]

In summary, we have proved

\[ \max _{d\in S'} d_S(d) = \max _{\| d\| \le 1}\big[\sigma _{S'}(d)-\sigma _S(d)\big] \]

and symmetrically

\[ \max _{d\in S} d_{S'}(d) = \max _{\| d\| \le 1}\big[\sigma _S(d)-\sigma _{S'}(d)\big]; \]

the result follows.

Proposition 3.3.8

A convex-compact-valued and locally bounded multifunction \(F:\mathbb {R}^n\longrightarrow 2^{\mathbb {R}^n}\) is outer [resp. inner] semi-continuous at \(x_0\in \operatorname {int}\text{dom }F\) if and only if its support function \(x\mapsto \sigma _{F(x)}(d)\) is upper [resp. lower] semi-continuous at \(x_0\) for all \(d\) of norm \(1\).

Proof

Calculus with support functions tells us that our definition (0.5.2) of outer semi-continuity is equivalent to

\[ \forall \varepsilon {\gt}0,\ \exists \delta {\gt}0:\ y\in B(x_0,\delta )\implies \; \sigma _{F(y)}(d)\le \sigma _{F(x_0)}(d)+\varepsilon \lVert d\rVert \quad \text{for all }d\in \mathbb {R}^n \]

and division by \(\lVert d\rVert \) shows that this is exactly upper semi-continuity of the support function for \(\lVert d\rVert =1\). Same proof for inner/lower semi-continuity.

Corollary 3.3.3
#

Let \((S_k)\) be a sequence of nonempty convex compact sets and \(S\) a nonempty convex compact set. When \(k \to +\infty \), the following are equivalent

  1. \(S_k \to S\) in the Hausdorff sense, i.e. \(\Delta _H(S_k,S)\to 0\);

  2. \(\sigma _{S_k}\to \sigma _S\) pointwise;

  3. \(\sigma _{S_k}\to \sigma _S\) uniformly on each compact set of \(\mathbb {R}^n\).

3.3.4 Example: Support functions of closed convex polyhedra