Optimization with Newton Raphson Method

newton raphsonnonlinear optimizationoperations researchoptimization

Actually i don't know how to distinguish maximization and minimization of nonlinear function with Newton Raphson Method.

What i know is, for finding the optimization points, we have to calculate this iteration:

$$x_{i+1}=x_i-[\mathbf Hf(x_i)]^{-1}\nabla f(x_i)$$

Then, what is actually the difference between maximization and minimization using this method?

Best Answer

Newton-Raphson is based on a local quadratic approximation. The iterate moves to the optimum of the quadratic approximation. Whether you minimize or maximize does not depend on the iteration calculation (you cannot modify it to turn minimization into maximization or vice versa) but on the shape of the approximation. The approximation is convex when $Hf(x_i)$ is positive semidefinite (psd), and concave when $Hf(x_i)$ is negative semidefinite (nsd). When $Hf(x_i)$ is psd, you expect to move to a local minimum, whereas when it is nsd,you expect to move to a local maximum.

The easiest way to think about this is for functions $\mathbb{R} \to \mathbb{R}$, so let's take $f(x) = x^3$. At $x=1$ the local quadratic approximation is $g(x) = 1+3(x-1)+3(x-1)^2$ which is convex. So if you perform an iteration of Newton raphson, you move to the minimum of $g$ and you hope to find a minimum of $f$.

On the other hand, if you start at $x=-1$, the local quadratic approximation is $g(x) = -1 + 3(x+1) - 3(x+1)^2$, which is concave. So if you perform an iteration of Newton raphson, you move to the maximum of $g$ and you hope to find a maximum of $f$.

If the definiteness of $Hf$ does not agree with your goal (e.g., $Hf$ is nsd but you want to minimize), then a quadratic approximation is not useful. It might be better to switch to other methods such as gradient descent.

Related Question