I would like to know how to build intuition for the concept of a tensor using the following reasoning:
If I conceive of a vector as an extension of the scalar concept, i.e. an $N \times 1$ "array of scalars" and similarly a matrix is a "vector of vectors", how can I apply a similar intuition to the idea of a tensor?
Is a scalar a matrix of matrices? A vector of matrices? Or something altogether different?
Formally, a tensor is a multilinear, real-valued function of vectors; this, I find, is the best guide to intuition. For it means that once you can identify some collection of objects as having the properties of a vector space, then, to find a tensor, all you need to do is identify linear operations that can be performed on these vectors, where these operations are real-valued.
For instance, if the set of real numbers is considered as a vector space (you can verify this by checking that the set of real numbers obey the properties of a vector-space with respect to itself as its underlying field), so that each real number is treated as a vector, then the multiplication of real numbers is a billinear, real-valued, operation over the reals: a tensor!
And, if you consider the set of all $N\times1$ column vectors as a vector space over the field of real-numbers, then the multiplication of one column vector by the transpose of another column vector to give a scalar is yet another billinear tensor (this is deeply intertwined with the so-called dot product of vectors).
In fact, when you consider quite natural definitions of common operations, like calculating volumes or lengths, tensors tend to "pop up" all over the place.
Now tensors are interesting, but of little practical use if you cannot compute things with them. It turns out that any tensor you define over a finite dimensional vector-space has a unique representation as an appropriate "matrix". The way to obtain this representation is to allow the tensor operate on basis vectors. This is why it is quite usual in applications to refer to Matrices as tensors.