For reference, eigendecomposition of a matrix $A$ $\in R^{n \times n}$ is defined as:
$A = P \Lambda P^{-1}$
where $P$ is a matrix whose columns are the eigenvectors of $A$, and $\Lambda$ is a diagonal matrix whose entries are the corresponding eigenvalues. This is only possible when $P$ is invertible, which is true when the eigenvectors of $A$ form a basis of $R^n$.
What is a complete geometric interpretation of the eigendecomposition of matrix $A$? Many interpretations that I see on the internet and in lectures narrow down on a special case where $A$ is symmetric. I understand that when $A$ is symmetric, its eigenvectors form an orthonormal basis of the space, so $P$ is an orthogonal matrix signifying rotations (maybe flips too). And in this special case, $P \Lambda P^{-1}$ can be interpreted as a transformation where:
- You first rotate the space ($P^{-1}$).
- Then, you scale the rotated space alone the axes ($\Lambda$).
- Finally, you rotate it back ($P$).
But eigendecomposition is defined for any general square matrix $A \in R^{n \times n}$ whose eigenvectors form a basis of $R^n$. The matrix $P$ need not be an orthogonal matrix signifying rotation/flips. What is the geometric interpretation of eigendecomposition for this general case?
In a first step, group the eigenvalues by value.
Any group belongs to a subspace.
The choice of the basis of any of the subspaces is arbitrary, completeness and linear independece implied.
A nonempty kernel, generally speaking, shows, that the linear representation space of the map has been to large.
By an invertible basis transform and index resorting, the matrix can be transformed into an upper triangualar block matrix of a zero block on the diagonal acting on the kernel and a nonsingular matrix on the diagonal acting on the complement.
So concentrate to nonsingular matrices.
Only normal matrices, commuting with their adjoints, yield a basis of eigenvectors.
Commuting with the adjoint means in essence, that within the given scalar product representation space of the linear map A, A and A* have the same eigenbasis.
So they can be diagonalized with the same transformation.
If all eigenvalues are different, the procedure is trivial.
Two eigenvectors of different eigenvalues are orthogonal in the given hermitean scalar product
$$\left< u_\lambda, (\lambda -\mu)\ u_\mu \right> = \left< u_\lambda , \lambda \ u_\mu \right>- \left< u_\lambda , \mu\ u_\mu \right> = \left< \lambda^* \ u_\lambda , u_\mu \right> - \left< u_\lambda , \mu\ u_\mu \right>= \left< A^* \ u_\lambda , u_\mu \right> - \left< u_\lambda , A u_\mu \right> =0$$
$$ (\lambda -\mu) \left< u_\lambda,\ u_\mu \right> = 0$$
It follows, that the matrix of the eigenbasis as columns is an eigenvector of A in matrix form with a digonal matrix as the generalized eigenvalue
$$ A \cdot \left(\begin{array}{c} e_{\lambda_1}^T\\ \vdots \\ e_{\lambda_n}^T \end{array} \right) = \left(\begin{array}{ccc} e_{\lambda_1}, 0, \dots,\dots\\ \vdots,\vdots,\vdots,\vdots \\ 0, \dots, 0, e_{\lambda_n}\end{array} \right) \cdot \left(\begin{array}{c} e_{\lambda_1}^T\\ \vdots \\ e_{\lambda_n}^T \end{array} \right) $$
Since all eigenvectors are orthogonal the matrix of eigenvectors is invertible and diagonalizes $A$ and $A^*$ simultanuously.
If there are multiple eigenvalues, any basis of the corresponding subspaces can be used. That is an arbitrary choice, implying a Gram Schmidt orthogonalization procedure, or, more frequently applied, one choses of a naturally given complete system of pairwise commuting matrices, that split the subspaces of degeneration by their different eigenvalues in that subspaces.