Consider a column $n\times 1$ vector $\overline {z}$ and an $n\times m$ matrix $W$. What would one call and denote an $n\times m$ matrix of derivatives defined by $M_{ij}=\frac{\partial z_i}{\partial W_{ij}}$?
To provide some context, this matrix appears when deriving the matrix form of backpropagation updates for neural networks, whereby $\overline {z}$ is the vector of inputs to a particular layer (which are linear combinations of actions received from the previous layer) and a particular training instance, and $W$ is the matrix of weights at that layer.
$\def\p#1#2{\frac{\partial #1}{\partial #2}}$The network layer equation in matrix and index notation is $$\eqalign{ z &= Wa + b \\ z_i &= W_{i\ell}a_\ell +b_i \\ }$$ When taking the derivative in index notation, take care not to repeat an index, because that implies summation over that index (like the $\ell$-index in the above equation).
With that in mind, the derivative of $z$ with respect to $W$ should be calculated as $$\eqalign{ \p{z_i}{W_{kj}} &= \left(\p{W_{i\ell}}{W_{kj}}\right)a_\ell \\ &= \left(\delta_{ik}\delta_{j\ell}\right)a_\ell \\ &= \delta_{ik}\,a_j \\ }$$ This third-order tensor is the correct derivative to use for backpropagation problems.
If you set the index $k=i$, that usually implies summation (aka the Einstein convention) $$\eqalign{ \p{z_i}{W_{ij}} &= \delta_{ii}\,a_j \\ &= \sum_{i=1}^N \delta_{ii}\,a_j \\ &= Na_j \\ }$$ If instead, you wanted to equate the indices without summing then you'd obtain $$\eqalign{ M_{ij} = \delta_{ii}\,a_j = {\tt1}_i\,a_j \\ }$$ but this matrix is not the right quantity to use for backprop.