logo

First-Order Necessary Conditions for Extrema of Multivariable Functions 📂Optimization

First-Order Necessary Conditions for Extrema of Multivariable Functions

Theorem1

Let’s assume the function f:RnRf : \mathbb{R}^{n} \to \mathbb{R} is given. If xx^{\ast} is a local optimizer and fC1f \in C^{1} in the vicinity of xx^{\ast}, then,

f(x)=0 \nabla f(x^{\ast}) = 0

f\nabla f is the gradient of ff. Note here that 00 is not the numeric zero, but a zero vector.

Explanation

The first-order necessary condition tells us about the property of the gradient, which is the first-order derivative of ff, when xx^{\ast} is a local minimizer of ff. Named and extended to multivariable functions, the concept that differentiation at a maximum or minimum yields 00 is something we learn even in high school.

Proof

We prove by contradiction. Assume f(x)0\nabla f (x^{\ast}) \ne 0. And denote as follows:

p=f(x),ptf(x)=f(x)2<0 p = - \nabla f (x^{\ast}),\quad p^{t}\nabla f (x^{\ast}) = - \left\| \nabla f (x^{\ast}) \right\|^{2} \lt 0

Here, ptp^{t} denotes the transpose matrix. Then, because f\nabla f is continuous, there exists s>0s \gt 0 such that the following equation holds:

ptf(x+ξp)<0,ξ[0,s] \begin{equation} p^{t}\nabla f (x^{\ast} + \xi p) \lt 0, \qquad \forall \xi \in [0, s] \end{equation}

Furthermore, by the Taylor expansion formula for multivariable functions,

f(x+ξp)=f(x)+ξptf(x+ξˉp),for some ξˉ(0,ξ) \begin{equation} f(x^{\ast} + \xi p) = f (x^{\ast}) + \xi p^{t} \nabla f(x^{\ast} + \bar{\xi} p),\quad \text{for some } \bar{\xi} \in (0, \xi) \end{equation}

Then, by (1)(1) and (2)(2), we obtain the following:

f(x+ξp)<f(x),ξ[0,s] f(x^{\ast} + \xi p) \lt f (x^{\ast}), \qquad \forall \xi \in [0, s]

This contradicts the fact that xx^{\ast} is a local minimizer. Therefore, the assumption is wrong, and f(x)=0\nabla f (x^{\ast}) = 0 is true.

See Also


  1. J. Nocedal and Stephen J. Wright, Numerical Optimization (2nd), p14-15 ↩︎