In a convex function, every chord (line segment between two points on the surface) lies above the function and does not intersect it.

A surface is guaranteed to be convex if the Hessian matrix is positive definite (has positive eigenvalues) for all possible parameter values.

For example, for a linear regression loss surface characterized by , we have:

(see UDL Chapter 6 Problems).

For any loss function, the eigenvalues of the Hessian matrix at places where the gradient is zero allow us to classify this position as:

  • a minimum (the eigenvalues are all positive)
  • a maximum (the eigenvalues are all negative)
  • (iii) a saddle point (positive eigenvalues are associated with directions in which we are at a minimum and negative ones with directions where we are at a maximum).