ConvexAnalysisMachineLearningStatistics
As the indicator risk functional is non-convex, one defines a convex surrogate loss in order to make the corresponding optimization problem convex. A convex loss is a convex function of the margin that upper bounds the indicator loss,
Indeed, for a convex surrogate margin-based loss, the optimal hypothesis is
One desires Fisher Consistency of
is differentiable at 0 To show that these conditions imply consistency, one wants for any choice of .
: is non-decreasing & minimizes set : is non-decreasing & maximizes set : is convex so & is minimized at is differentiable at 0 so : is convex so minimizes iff zero is in the subdifferential of the expected -loss: and consider the subgradients of i.e. so that summing and subtracting the equations yield As and it must hold that when and when . The second equation implies that when it holds that and . Likewise, when it holds that and .
Therefore, for convex