BACKGROUND: I have recently found (probably well known, but I had never seen this before) that a matrix can be written as a linear combination of the outer products of its eigenvectors where the coefficients are the corresponding eigenvalues. Specifically, let $\boldsymbol{u}_k=(u_{1,k},\dots,u_{K,k})^T$ be the $k$-th eigenvector and $\lambda_k$ the $k$-th eigenvalue of a symmetric matrix $M$ with $k\in [1,K]$. Then $M$ can be written as
$$M=\sum_{k=1}^K \lambda_k \boldsymbol{u}_k \boldsymbol{u}_k^T \tag{1} \label{eq1}$$
Which, is just another way of writing the standard eigendecomposition
$$M=U \Lambda U^{-1} \tag{2} \label{eq2}$$
where $U$ is the matrix whose $k$-th column is $\boldsymbol{u}_k$ and $\Lambda$ is the diagonal matrix whose element $\Lambda_{kk}=\lambda_k$. What is interesting to me is that the series expression is reminiscent of Fourier series expansions with $\lambda_k$ being analogous to the Fourier coefficient and the matrix $\boldsymbol{u}_k \boldsymbol{u}_k^T$ being analogous to a basis function.
HERE'S THE QUESTIONS:
(1) Can someone point me to a good reference (or references) about expressing matrices as a series expansion of "basis" matrices?
(2) Can two different matrices, say $A$ and $B$ be expressed as a series expansion using the same set of basis matrices and just different coefficients? (Eq. \ref{eq1} above could be used for both $A$ and $B$, but the set of basis matrices $\boldsymbol{a}_k \boldsymbol{a}_k^T$ and $\boldsymbol{b}_k \boldsymbol{b}_k^T$ would be different, and I would like to instead write them in the same basis). If so, under what conditions?
Your first question: I would advise the different books named "Linear Algebra" with different subtitles of Gilbert Strang (MIT Press).
Your second question; yes, there is such a theorem that you will find in A property of positive definite matrices
Edit: As underlined by @icurays1, it is in the SVD framework
$$M=USV^T=\sum_{k=1}^K \sigma_k \boldsymbol{u}_k \boldsymbol{v}_k^T $$
where $K=min(n,p)$ (if $M \in \mathbb{R}^{n \times p}$), with $\sigma_k \geq 0$ the singular values that this kind of decomposition (as a cumulative sum of the $N$ first "components" closer and closer to the initial matrix $M$ as $N$ increses) is important. Let us take the following notation:
$$M_N:=\sum_{k=1}^N \sigma_k \boldsymbol{u}_k \boldsymbol{v}_k^T $$
(we assume that the singular values are ranked by decreasing amplitude).
with the fundamental property that $M_N$ is, among all rank-$N$ approximations of $M$, the closest to $M$ in the sense of Frobenius norm.
I will take two examples of applications:
Left: The original image; Center: the reconstructed image taking only the $N=10$ first components ; Right: the same as the center image but with $N=30$ first components.
One can see that with 30 components, the reconstruction is almost perfect. On the point of view of the compression ratio, as it is a $629 \times 590$ image, instead of sending $629 \times 590 \approx 371 K$ real numbers (grey values of pixels) for the whole image, we have only to send $(629 + 590) \times 30 \approx 36.6 K$ reals, which results in a 1:10 ratio compression.
$$\begin{bmatrix}0.02&0.08&0.02\\0.06&0.24&0.08\\0.06&0.28&0.04\\ 0.02&0.10&0.00\end{bmatrix}\approx 0.4089\begin{bmatrix}0.2074\\0.6321\\0.7056\\ 0.2442\end{bmatrix}\times\begin{bmatrix}0.2184&0.9545&0.2029\end{bmatrix} \ \ (*)$$
with a very good agreement between the LHS and the RHS because of a quasi independence between the $X$ and $Y$ components; this explains that the joint distribution $f_{X,Y}$ can almost be "reconstructed" as the product $f_X \times f_Y$ of its marginal distributions. An interesting thing is that in (*), column vector $\boldsymbol{u}_1$ and row vector $\boldsymbol{v}_1^T$ contain, (up to a renormalization, because the sum of their components have to be equal to 1) the "closest distributions" $\phi_X$ and $\phi_Y$ such that $f_{X,Y}\approx\phi_X\times \phi_Y$.
Note: the very good agreement between matrix $M$ and its approximation $M_1$ is reflected in the rapid decay of singular values : $\sigma_1=0.4089, \sigma_2=0.0404, \sigma_3=0.0015$.