On page 487 of Boyd & Vandenberghe's Convex Optimization, the convergenge analysis of Newton's method (Algorithm 9.5) is based on the backtracking line search. My questions are around (9.33) on Page 488. The assumption is $\eta \leq m^2/L$ where $m$ is the constant of strongly convexity, and $L$ is the Lipschitz constant for the Hessian of $f$, i.e., $\nabla^2f(x)$.
In the paragraph under (9.33), it says that if $\left\|\nabla f\left(x^{(k)}\right)\right\|_{2}<\eta$, then $\left\|\nabla f\left(x^{(k+1)}\right)\right\|_{2}<\eta$ because of $\eta\leq m^2/L$. Why is this? How can we get this result for the iteration $k+1$? That is to say, $\left\|\nabla f\left(x^{(l)}\right)\right\|_{2}<\eta$ holds for all $l\geq k$.
Also, in the end of that paragraph, it says that '' Therefore for all $l\geq k$, the algorithm takes a full Newton step $t=1$''. Why is $t^{(k)}=1$ once $\left\|\nabla f\left(x^{(k)}\right)\right\|_{2}<\eta$, not a number less than $1$?
For the first question, let $g(t)=\|\nabla f(x+t\Delta x_{nt})\|_2^2$ with $t\ge 0$, then since $\Delta x_{nt}=-\nabla^2f(x)^{-1}\nabla f(x)$ we have
$$ \begin{aligned} g'(t) &= 2\nabla f(x+t\Delta x_{nt})^T \nabla^2 f(x+t\Delta x_{nt})\Delta x_{nt} \end{aligned} $$ $$ \begin{aligned} g'(0) &= 2\nabla f(x)^T\nabla^2 f(x)\Delta x_{nt}\\ &=2\nabla f(x)^T\nabla^2 f(x)(-\nabla^2f(x)^{-1}\nabla f(x))\\ &=-2\nabla f(x)^T\nabla f(x)\\ &<0 \end{aligned} $$ Since $g'(t)$ is continuous, there exists $t>0$ such that $g(t)<g(0)$, i.e., $\|\nabla f(x^{(k)}+t\Delta x^{(k)}_{nt})\|_2^2=\|\nabla f(x^{(k+1)})\|_2^2<\|\nabla f(x^{(k)})\|_2^2$. But this has nothing to do with $\eta\le m^2/L$(we didn't use this condition in the above derivation) and only answers the first question, why is $t=1$ for all $l\geq k$?