If $\mathbf x$ and $\mathbf y$ are matrices such that the function $\mathbf z=\mathbf x^n \mathbf y^m $ is well defined, how can I find the derivatives $$ \frac{\partial \mathbf z}{\partial \mathbf x} \qquad \frac{\partial \mathbf z}{\partial \mathbf y} $$
From some example in Quantum Mechanics I know that, as an example, if $\mathbf z =\mathbf x^2 \mathbf y$ than $$ \frac{\partial \mathbf z}{\partial \mathbf x} =\mathbf{xy+ yx} $$
But where can I find the proof for this result? More general: where can I find some references about this kind of calculus with matrices?
Added after the answers.
My reference is the historical paper of Born and Jordan on foundation of matrix mechanic (it can be found in https://archive.org/details/SourcesOfQuantumMechanics). Here, in the paragraph 2 ( pag.282 of ref.), is defined the ''Symbolic differentiation'' of a matrix function and, as an example, is given: $$ \mathbf{ y=x_1^n x_2^m \qquad \frac{\partial y}{\partial x_1}=x_1^{n-1}x_2^m+x_1^{n-2}x_2^mx_1+\cdots +x_2^mx_1^{n-1} } $$
so it seems totally different from the answers. Is it wrong or it depends only on a different definition of the derivative?
My problem is to justify the result of Born and Jordan that is essential to prove the classical commutation relation of Quantum Mechanic.
If the definition of Born is so different from the ''standard'' mathematical definition, how we can motivate such difference?
The best suggestion is to work with the definition of the directional derivative. If $f(\mathbf x,\mathbf y) = \mathbf x^2\mathbf y$, then if we change $\mathbf x$ in direction $\mathbf v$, we have \begin{align*} D_{(\mathbf v,\mathbf 0)}f(\mathbf x,\mathbf y) &= \lim_{t\to 0} \frac{f(\mathbf x + t\mathbf v,\mathbf y) - f(\mathbf x,\mathbf y)}t = \lim_{t\to 0}\frac{(\mathbf x+t\mathbf v)^2\mathbf y - \mathbf x^2\mathbf y}t \\ &= \lim_{t\to 0}\frac{(\mathbf x^2 + t\mathbf x\mathbf v + t\mathbf v\mathbf x+ \mathbf v^2)\mathbf y - \mathbf x^2\mathbf y}t = (\mathbf x\mathbf v + \mathbf v\mathbf x)\mathbf y, \end{align*} so your formula is totally incorrect (for starters, there's no way $\mathbf y$ can end up on the other side of $\mathbf x$). More importantly, there's no way to write the correct formula as a simple matrix multiplied by the vector $\mathbf v$ because the vector $\mathbf v$ gets intertwined.
Now if you think of $\mathbf x = [x_{ij}]$ and you want to compute $\partial f/\partial x_{ij}$, then you apply what I wrote with $\mathbf v = E_{ij}$, the matrix with a $1$ in the $ij$-position and $0$'s elsewhere.