Is it fair to call Newton's method for optimization a second-order method?

905 Views Asked by At

In my machine learning class, we've discussed two methods for optimization - gradient descent and Newton's method. I had learned Newton's method before in calculus, but always in the context of finding roots of $f(x)=0$.

In retrospect it is obvious that it can just as easily be applied to $f'(x)=0$, but I guess it had never occurred to me before.

When solving $f(x)=0$, Newton's method requires us to use $f'(x)$, so I was always taught that this makes it a "first-order method".

But of course when solving $f'(x)=0$ using Newton's method (which we do in the context of optimization), we must use $f''(x)$. Does this now make it a "second-order method", while gradient descent would be a first-order method?

Or, should Newton's method applied to $f'(x)$ still be considered a first-order optimization method, and gradient descent a "zeroth-order optimization method"?

1

There are 1 best solutions below

2
On BEST ANSWER

The order of a method refers to the speed with with it approaches the objective. If you have $x_{n + 1}$ computed from $x_n$, converging to $x^*$, the method is called order $\alpha$ if, approximately when near the objective:

$$\lvert x_{n + 1} - x^* \rvert \approx c \lvert x_n - x^* \rvert^\alpha$$

Newton's root finding method (either on the function or on the derivative, for optimization) is of second order unless the zero is multiple, when it is linear.