When inverting a matrix, we can consider the inverse of A to be equivalent to the inverse of the determinant of A multiplied by the adjugate matrix of A. Or rather, that the product of the adjugate matrix of A with A is defined to be equal to det(A) * I.
This is considered a "definition", but as of yet I cannot find any explanation of why this should be the case, why this works, or even who invented the concept of an adjugate matrix.
Clearly this "definition" must be "correct", in the sense that I could "define" the product of the "adjugate" of A with A to be anything (say, ...3), yet this "definition" is clearly "wrong".
Looking in more detail, we see that the adjugate matrix of A is the transpose of the cofactor matrix of A. Why should this be the case? I cannot find any information on why cofactors enable matrix inversion. Again this is given as a "definition":
$\mathbf A^{-1} = \frac{1}{\operatorname{det}(\mathbf A)} \mathbf C^\mathsf{T}.$
And hence apparently just assumed to be true, and I can find no information on why this works.
My core questions are really:
- What actually is an adjugate matrix?
- How does it allow for matrix inversion? There must be something special about it.
- Or did somebody just figure out having a matrix with determinants on the diagonal was useful? If so - how did they figure this out? What is the logical thought-process to reach this conclusion?
- How does the machinery of cofactors play into this? Why is the adjugate the transpose of the matrix cofactors? How does this help?
Or to summarise: why does this all "magically" seem to work, anyway?
One way to think about this is using exterior powers. If $A : V \to V$ and $V$ is $d$-dimensional, then you get an induced map $C = \wedge^{d-1} A : \wedge^{d-1} V \to \wedge^{d-1} V$. If you fix an identification of $\wedge^d V \to k$, the base field, then $\wedge : V \otimes\wedge^{d-1} V \to \wedge^d V \to k$ is a perfect pairing, and if you correspondingly choose dual bases then the matrix of $\wedge^{d-1} A$ is the adjugate of the matrix of $A$. The formula $C^T A = (\det A) I$ is a consequence of the above pairing and the identification $\wedge^d A = \det A$ induced by $\wedge^d V \cong k$.
A bit more explicitly, if $V$ has basis $e_1, \dots, e_d$ then $\wedge^k V$ has basis $e_I = e_{i_1} \wedge \dots \wedge e_{i_k}$ for $I = (i_1, \dots, i_k)$ with $1 \le i_1 < \dots < i_k \le d$. The induced map $\wedge^k A$ takes $e_I$ to $\sum a_{I, J} e_J$ where $a_{I, J}$ is the determinant of the submatrix with columns indexed by $I$ and rows indexed by $J$. In particular, $\wedge^d V$ is $1$-dimensional, generated by $e_1 \wedge \dots \wedge e_d$ and $\wedge^d A$ is given by (multiplication by) $\det A$. To understand the adjugate, note that $\wedge^{d-1} V$ has basis $$e_1 \wedge \dots \wedge \hat{e_j} \wedge \dots \wedge e_d,$$ where $\hat{e_j}$ signifies that that factor is omitted. Up to some signs this is the dual basis of $e_i$ as mentioned above, explicitly $$(e_i) \wedge (e_1 \wedge \dots \wedge \hat{e_j} \wedge \dots \wedge e_d) = (-1)^{i-1} \delta_{i, j} \, (e_1 \wedge \dots \wedge e_d) .$$ Once these signs are incorporated, the general description of the matrix for $\wedge^k A$ specializes to the adjugate matrix when $k = d - 1$.
This is surely not how anyone came up with either the idea or the explicit formula for the adjugate matrix but for me it explains the "magic trick".