Im practicing for an exam and this question popped up on an old exam. the following are the subquestions:
a) Show that $s_0=(1,-1)^T$ is a descent direction in $x_0=(0,0)^T.$
b) Determine $x_1$ by exactly solving the linesearching problem with startingpoint $x_0$ and search direction $s_0$.
c) What happens if you solve the optmization problem with Newtons method for multidimensional optimiztion?
Here is what the examinator proposes:
a) We have that
$$\frac{d}{da}f(x_0+as_0) \rvert_{a=0} = \frac{d}{da}\left(\frac{a^2}{2}+\frac{3a^2}{2}-a^2-2a\right)\Bigg\rvert_{a=0}=-2<0.$$
The direction derivative is negative in that direction, thus it is a descent direction.
Question 1: I understand that we have to show that the direction derivative is less than zero in order for it to be a descent, but why are they choosing the stepsize $a=0?$ We haven't been given the optimal stepsize yet, according to the book I'm supposed to find it using the jacobian and hessian.
Why are we even interested in a function in terms of $a$?
b) $f(x_0+as_0)=a^2-2a,$ so the minimal value is attained when $a=1,$ then $x_0+as_0=(1,-1)^T.$
Question 2: I think the question I have here is the result of the lack of understanding of a). I don't understand when I can use the derivative like this to find the stepsize and when I should use the fact that
$$a_k=-\frac{\nabla f(x_k)^Ts_k}{s_k^THs_k},$$
where $H= \ $Hessian.
c) This is an optimization problem with a quadratic objectfunction. Such will Newtons method solve exactly in one step.
Question 3: Why?