First-Order Necessary Conditions for Extrema of Multivariable Functions
Theorem1
Let’s assume the function is given. If is a local optimizer and in the vicinity of , then,
is the gradient of . Note here that is not the numeric zero, but a zero vector.
Explanation
The first-order necessary condition tells us about the property of the gradient, which is the first-order derivative of , when is a local minimizer of . Named and extended to multivariable functions, the concept that differentiation at a maximum or minimum yields is something we learn even in high school.
Proof
We prove by contradiction. Assume . And denote as follows:
Here, denotes the transpose matrix. Then, because is continuous, there exists such that the following equation holds:
Furthermore, by the Taylor expansion formula for multivariable functions,
Then, by and , we obtain the following:
This contradicts the fact that is a local minimizer. Therefore, the assumption is wrong, and is true.
■
See Also
J. Nocedal and Stephen J. Wright, Numerical Optimization (2nd), p14-15 ↩︎