Suppose that I have a (possibly non-convex) function that maps $R^N$ to $R$, and is differentiable at all points except finitely many points. I want to use gradient descent algorithm to show convergence to a local optima, $$ x(k+1) = x(k) + \alpha(k) \nabla f |_{x=x(k)}, $$ where $\nabla f |_{x=x(k)}$ is derivative computed at point $x(k)$, $\alpha(k)$ is step-size at iteration $k$. When derivative is not defined, a random step is taken (say Gaussian) with variance $\alpha(k)$.
Assume that (wherever derivative is defined) the norm of derivative is uniformly bounded by $B$, i.e. $\sup_{x} \|\nabla f \| \le B$.
Is it true that the iterates converge to a local optima?
Edit : This answers the original question. The question has been edited after this answer was made.
No, this does not work :
Consider $f : x \mapsto \frac{1}{x}$, which isn't differentiable at $0$. For any positive value of $\alpha(k)$, if you take $x(k)$ whose absolute value is small enough (eg. smaller than $\alpha(k)$ and close to $0$), $x(k+1)$ will be a number with HUGE absolute value. I suspect that the sequences formed as described in original post will exhibit a chaotic behavior.