Where here we discover some interesting facts about continuous convex functions.

We know that a function is convex if

for all and .

We see that if is a continuous function, then an equivalent condition for convexity is that either of the following inequalities holds

If is convex it is straightforward to prove the right hand side inequality. Conversely, notice that the right-hand side of that inequality is

So then, we have

which means that is convex. Actually, it means that for almost all , but since is assumed to be continuous, it holds for all .

There is another interesting inequality we may prove: We may use the integral form of Jensen’s inequality according to which if is convex and and are integrable on then

Here, choose ; we have

Then, we have the double-sided inequality

Assuming that is continuous, then is convex if and only if at least one of these inequalities holds.

Interesting identities I wasn’t familiar with! With algorithms like Newton’s method (more specifically the multidimensional case) it’s the goal to find absolute extrema, and to find them a convex function is necessary – otherwise you’re not guaranteed to find that absolute minima or maxima. I’m looking forward to taking analysis the semester after next so I can get into more equations like this and how to formulate them and work with them.

LikeLike

Indeed, there are many interesting inequalities. Maybe a good starting point is the study of those related to the first and second-order approximations of a function which is either assumed to be (i) continuously differentiable with Lipschitz gradient or (ii) twice continuously differentiable with Lipschitz Hessian. These are used in the analysis of first and second-order methods (gradient and Newton methods).

I find Nesterov’s book “Introductory Lectures on Convex Optimization” a very good reference. I uploaded a new post you might find interesting: https://mathematix.wordpress.com/2016/08/03/second-order-approximation-of-twice-differentiable-functions/

LikeLike