I was following a proof provided in Gilbert Strang's book "Introduction to Linear Algebra". And I am confused by one step of the proof.
Suppose we have a $n$ by $n$ stochastic matrix $A$, where all elements are not negative and sum of every element in each row is $1$.
There exists a proof that there is only one biggest eigenvalue of $A$ equal to $1$ and other eigenvalues are less then $1$. I found it there: Proof that the largest eigenvalue of a stochastic matrix is 1
Now I want to get a proof that Markov chain has a steady state that is not affected by initial probability distribution:
$$ u_0 = \left( \begin{smallmatrix} u_1 \\ u_2 \\ \ldots \\ u_n \end{smallmatrix} \right) $$
Let's apply diagonalization to our matrix $A$:
$$ A = S \Lambda S^{-1} $$
Where $S$ consists of eigenvectors placed there as columns, $\Lambda$ is a diagonal matrix with corresponding eigenvalues.
Suppose we want to get a presentation of our initial distribution as a linear combination of eigenvectors of $A$ like this:
$$ u_0 = c_1 x_1 + c_2x_2 + \ldots + c_nx_n $$
In a matrix form:
$$u_o = SC$$
We can get $C$ from:
$$C = S^{-1}u_0$$
So, when we apply A to $u_0$ multiple times ($k$ times):
$$ u_k = A \ldots Au_o = S \Lambda S^{-1} \ldots S \Lambda S^{-1} u_o = S \Lambda S^{-1} \ldots S \Lambda S^{-1} SC = S \Lambda^{k} C$$
So, literally we get:
$$ u_k = c_1(\lambda_1)^kx_1 + c_2(\lambda_2)^kx_2 + \ldots + c_n(\lambda_n)^kx_n $$
But then author writes this in the book:
$$ u_k = x_1 + c_2(\lambda_2)^kx_2 + \ldots + c_n(\lambda_n)^kx_n $$
I understand that the author omits $\lambda_1$ because it is equal to 1. Why does the author omit $c_1$?
EDIT: I found out that $c_1$ is equal to 1. But I don't know why. This is why the author omits it.
Later in his proof the author shows that:
$$ \lim_{k\rightarrow \infty } u_k = x_1 $$
So, the author concludes that steady state is equal to the eigenvector with corresponding eigenvalue of 1.
This proof seems a little confusing to me...first of all, consider the stochastic matrix \begin{equation} \begin{bmatrix} 0&1 \\ 1&0 \end{bmatrix},\end{equation} which is diagonalizable, but which has no steady state.
So I think it must be specified that $A$ is regular, ie some power of $A$ has only positive entries.
So now assuming this is true, with regard to your question, if we have the coordinate vector of $u_0$ with respect to the standard basis as $\left( \begin{smallmatrix} u_1 \\ u_2 \\ \ldots \\ u_n \end{smallmatrix} \right)$, and since we know $S$ is the change of basis matrix taking vectors from the standard basis to the basis of eigenvectors, we must have \begin{equation} S\left( \begin{smallmatrix} u_1 \\ u_2 \\ \ldots \\ u_n \end{smallmatrix} \right)=\sum_{i=1}^n u_ix_i,\end{equation} so that in fact we must have $u_i=c_i$, and there is definitely no reason why we must have $u_1=1$.
A much better approach (imho) is to use the property that for a regular stochastic matrix $A$: $\lim_{m \rightarrow \infty}A^m=L$, is a matrix with every column equal to $x_1$, the eigenvector associated with eigenvalue $1$. The existence of $L$ follows in part from the Perron-Frobenius theorem, or equivalently, the link you have posted and some other results - I am not going to prove that here - so if we can accept that $L$ indeed exists, then we need two things - one: \begin{equation} AL=A\lim_{m \rightarrow \infty}A^m=\lim_{m \rightarrow \infty}AA^m=\lim_{m \rightarrow \infty}A^{m+1}=L,\end{equation} and secondly since $AL=L$ every column of $L$ is an eigenvector of $A$ with associated with eigenvalue $1$.
So having established the above, using $x_1$ as in your notation, we have \begin{equation}(\lim_{m \rightarrow \infty}A^m)u_0=Lu_0=u_1x_1+u_2x_2+\cdots+u_nx_1=(u_1+u_2+\cdots+u_n)x_1=x_1,\end{equation} (since $u_0$ is a probability vector its entries sum to 1).