Substochastic matrix spectral radius

6.3k Views Asked by At

Let $M$ be a row substochastic matrix, with at least one row having sum less than 1. Also, suppose $M$ is irreducible in the sense of a Markov chain. Is there an easy way to show the largest eigenvalue must be strictly less than 1? I hope that this result is true as stated. I know that Cauchy interlacing theorem gives me $\leq$,

3

There are 3 best solutions below

4
On

You can try completing your matrix to a Markov chain, adding a self loop at the additional state. The new Markov chain is irreducible and aperiodic and so has a unique stationary distribution, which is concentrated on the additional state. It is also the unique eigenvector with eigenvalue at least $1$.

Now take a purported eigenvector for your original matrix with eigenvalue $1$, and add a zero coordinate. The result is an eigenvector for the Markov chain, contradicting the properties we listed above.

In effect, you have a Markov chain with an invisible absorbing state, which is actually reachable from any other state. This ensures that in the long run the state will be reached, and so applying your matrix lots of times on any given vector will yield the zero vector. So all eigenvalues must be less than 1 in magnitude.

0
On

This is essentially Yuval's probabilistic argument with the probability removed. The goal is to show that powers of $M$ converge to zero.

For any state $i$ and integer $n\geq 0$, let $r^n_i=\sum_k M^n_{i k}$ denote the $i$th row sum of $M^n$. For $n=1$, we write $r_i$ rather than $r^1_i$. Since $M$ is substochastic we have $0\leq r^n_i\leq 1$.

Let $k^*$ be an index with $r_{k^*}<1$, and note that for $n\geq 1$
$$r^n_{k^*}=\sum_k M_{k^* k}\ r_k^{n-1}\leq \sum_k M_{k^* k} =r_{k^*}<1.$$

By irreducibility, for any $i$, there is an $m$ with $M_{i k^*}^m>0$. In fact, if $M$ is $N\times N$ matrix, and $i\neq k^*$ then we can assume $m<N$. (Take the shortest path from $i$ to $k^*$ with positive "probability").
Since $M_{i k}^m$ puts positive weight on the index $k=k^*$, we have $$r^N_i=\sum_k M^m_{i k}\ r^{N-m}_k < r^m_i \leq 1.$$

That is, every row sum of $M^N$ is strictly less than one. Now you can show that $M^{jN}\to 0$ as $j\to \infty$ and this shows that $M^N$ (and hence $M$) cannot have any eigenvalue with modulus 1.

0
On

Here is a simple proof. First, for a (sub-)stochastic matrix we have $\rho(\boldsymbol{A})\leq1$, which follows, e.g., from $\rho(\boldsymbol{A})\leq\|\boldsymbol{A}\|_\infty$. Thus, all we want to show is $\rho(\boldsymbol{A})\neq1$. Second, the "substochasticity" can be written as $\boldsymbol{A}\mathsf{1}\leq\mathsf{1}$, $\boldsymbol{A}\mathsf{1}\neq\mathsf{1}$. If $\boldsymbol{A}$ is irreducible, then we apply the following clever proposition with $\alpha=1$ and $\boldsymbol{z}=\mathsf{1}$. This does the job.

Proposition. Let $\boldsymbol{A}\geq0$ be irreducible and $r:=\rho(\boldsymbol{A})$. Then $\boldsymbol{A}\boldsymbol{z}\leq\alpha\boldsymbol{z}$, $\boldsymbol{A}\boldsymbol{z}\neq\alpha\boldsymbol{z}$ together imply $\alpha\neq r$.

Proof. Let $\boldsymbol{c}:=\left(\boldsymbol{A}-\alpha\boldsymbol{I}\right)\boldsymbol{z}$ then $\boldsymbol{A}\boldsymbol{z}\leq\alpha\boldsymbol{z}$ and $\boldsymbol{A}\boldsymbol{z}\neq\alpha\boldsymbol{z}$ become $\boldsymbol{c}\leq0$ and $\boldsymbol{c}\neq0$. Since $\boldsymbol{A}\geq0$ is irreducible, there exists the left Frobenius vector $\boldsymbol{\phi}>0$, such that $\boldsymbol{\phi}^{T}(\boldsymbol{A}-r\boldsymbol{I})=\boldsymbol{0}$ (by the Perron-Frobenius theorem). Note that $\boldsymbol{\phi}>0$ and $\boldsymbol{c}\leq0,\boldsymbol{c}\neq0$ imply $\boldsymbol{\phi}^{T}\boldsymbol{c}\neq0$. If $\alpha=r$ then $\boldsymbol{\phi}^{T}(\boldsymbol{A}-r\boldsymbol{I})\boldsymbol{z}=\boldsymbol{\phi}^{T}\boldsymbol{c}=0$, which contradicts $\boldsymbol{\phi}^{T}\boldsymbol{c}\neq0$. Therefore, $\alpha\neq r$. $\blacksquare$


Background. This proposition is similar to Example 8.3.1 on p.674 of Carl Meyer's "Matrix Analysis and Applied Linear Algebra", where Meyer makes and proves a different but somewhat similar statement, which he later applies to substochastic matrices. I couldn't figure out how Meyer's statement on p.674 was relevant, and I believe he just made a mistake, which I think is corrected by replacing his statement with my proposition.