Matrix function converges, how about the eigenvalues?

1.3k Views Asked by At

Suppose I have a matrix function $A(t)$ with $$\lVert A(t) - B\rVert \le ct^\alpha$$ in some matrix norm (this will work for any norm, I guess). So, in a sense $A(t)\rightarrow B$ for $t\rightarrow 0$ in $\mathcal{O}(t^\alpha)$. Plus, we have $A(0) = B$.

I happen to know the eigenvalues of $B$, but I don't know a thing about the eigenvalues of $A(t)$. Plus, $A(t)$ does not have any favorable structure, in particular, no symmetry.

So, what can you say about the eigenvalues of $A(t)$? In particular:

  • What about the spectral radius of $\lambda_{A(t)}$? Does it converge to the spectral radius of $B$?
  • Do we have $\lambda_{A(t)}\rightarrow\lambda_B$ for $t\rightarrow 0$ for all eigenvalues $\lambda_{A(t)}$ of $A(t)$?
  • And finally: is the speed of convergence $\mathcal{O}(t^\alpha)$ the same for the eigenvalues/spectral radius as for the matrix function?

The last question is actually the most important one. If the eigenvalues of $B$ are all zero, the eigenvalues as well as the spectral radius of $A(t)$ would go to zero as $\mathcal{O}(t^\alpha)$...

Any help would be appreciated, incl. references to (standard?) textbooks or papers on this matter. Maybe there is a counterexample? So far, in all numerical examples I have seen/done, all properties above do hold.

Edit: To provide a bit more background: The matrices $A(t)$ are iteration matrices which depend on a time-step size $t$. They are not this ugly, but showing convergence of this iteration has proven to be rather difficult. In the simplest case, they look like $$A(t) = (I-tQ_1)^{-1}(t(Q_2-Q_1)+B)$$ with identity matrix $I$ and some matrices $Q_1,Q_2$, which do not have any particular structure we were able to exploit so far. Now, if I can make that conclusion about the spectral radius as described above, I can state that the spectral radius is smaller than 1, i.e. the iteration converges, if the time-step size $t$ is small enough.

Edit: Does this answer help? Also, this question might be related to perturbation theory for eigenvalue problems (with non-symmetric matrices, though, and $B$ is not diagonalizable).

2

There are 2 best solutions below

6
On BEST ANSWER

Parts (1) and (2) -- Yes. The coefficients of the characteristic polynomial are continuous functions of $A(t)$ (they are polynomials in the entries of $A(t)$!) and the roots of a polynomial are continuous functions of the coefficients.

Part (3): If $B$ has non-trivial Jordan blocks, this can fail. For example, $$\begin{pmatrix} 0 & 1 \\ -t & 0 \end{pmatrix} = \begin{pmatrix} 0 & 1 \\ 0 & 0 \end{pmatrix} + O(t)$$ as $t \to 0$, but the eigenvalues are $\pm \sqrt{t}$, which is not $O(t)$.

I believe that this is true when $B$ is diagonalizable. (I know that it is diagonalizable if $B$ has distinct eigenvalues, and I'll write up that argument if this one fails.)

We may as well assume $B$ is diagonalized. Let's concentrate on a particular eigenvalue $\lambda$ of $B$, with multiplicity $m$; we might as well assume $\lambda =0$. So $B = \mathrm{diag}(0,0,\cdots,0, \lambda_2, \lambda_3, ..., \lambda_k)$ with $m$ zeroes.

Suppose that $B-A(t)$ is $O(t^a)$. Explicitly expanding the determinant, the coefficient of $x^k$ in $\det(x \mathrm{Id} - B)$ is $O(t^{a(m-k)})$ for $k < m$, and the coefficient of $x^m$ does not go to $0$ as $t \to 0$. So the Newton polygon of the characteristic polynomial passes through $(m,0)$ and stays above the line from $(m,0)$ to $(0,ma)$. This shows that the bottom $m$ roots of the characteristic polynomial vanish at rate $O(t^a)$ or faster as $t \to 0$ (and the other roots do not vanish.)


It occurs to me that it is worth sketching the argument for the Newton polygon claim directly so you can see how straightforward it is without learning the whole Newton polygon technology. Here is what I am claiming:

Lemma Fix $C>0$. Then there is a $D>0$ such that, if $f(u,z) = \sum_i f_i(u) z^i$ is holomorphic in $z$ on the disc of radius $1$, with $$|f_i(u)/f_m(u)| < \begin{cases} C u^{m-i} & i < m \\ C & i > m \end{cases}$$ then $f(u, \ )$ has $m$ roots in the disc of radius $D u$ for all sufficiently small $u$.

Proof sketch Consider $\oint \tfrac{f'}{f} dz$ where the integral is around the circle of radius $Du$. Write $f = f_m(u) z^m (1+\mbox{other terms})$ and $f' = m f_m(u) z^{m-1} (1+\mbox{other terms})$ so the integral is $\oint m \tfrac{dz}{z} + \mbox{other terms}$. Use the above conditions to bound the other terms.

2
On

This is not answer. Just trying to make observations. Note that when $t$ is very close to zero, you have $$(I-tQ_1)^{-1}=I+tQ_1+t^2Q_1^2+\dots$$ Let $C=Q_2-Q1$. Then $$A(t) =(tI+t^2Q_1+t^3Q_1^2+..)C~+~(I+tQ_1+t^2Q_1^2+\dots)B $$ Now, let's look at the matrix $(I-tQ_1)^{-1}$. Let $Q_1 = T\Lambda T^{-1}$ for some invertible matrix $T$. Thus, if $\lambda_i(Q_1)$ is the $i^{th}$ eigenvalue of $Q_1$. Then we have $$\lambda_i((I-tQ_1)^{-1})=1+\lambda_it+\lambda_i^2t^2+\dots$$ Frankly, I don't know what conclusions you can draw from these. Hope it helps you or anyone else trying to answer these in some way.