I am well over my head on this topic, but I'm asking for a couple of conceptual, framing statements to understand the broader picture at an intuitive level.
In Hilbert Space Methods in Probability and Statistical Inference by Christopher G. Small, Don L. McLeish it can be read:
Let $\mathbf G$ be the set of all functions ${\bf x}: \Omega \to > \mathbb R$ such that $({\mathbf x \land n{\bf 1}})\lor(-n{\bf1})\in > {\mathbf H}$ for all natural numbers $n$. The set $\mathbf G$ can be shown to be a vector space under the usual pointwise addition and scalar multiplication.
So what is the connection between the random variables (and specifically their probability density functions) and vector or inner product spaces. Specifically, and if these statements / questions are remotely close to reality, in what way are pdf's linear with scalar multiplication?
In answer to the comment, this is the passage quoted:
And this is the link provided by @symplectomorphic, which really resolves the question:
Many of the concepts in this chapter have elegant interpretations if we think of real-valued random variables as vectors in a vector space. In particular, variance and higher moments are related to the concept of norm and distance, while covariance is related to inner product. These connections can help unify and illuminate some of the ideas in the chapter from a different point of view. Of course, real-valued random variables are simply measurable, real-valued functions defined on the sample space, so much of the discussion in this section is a special case of our discussion of function spaces in the chapter on Distributions, but recast in the notation of probability.
As usual, our starting point is a random experiment modeled by a probability space (Ω,F,P). Thus, Ω is the sample space, F is the σ-algebra of events, and P is the probability measure. Our basic vector space V consists of all real-valued random variables defined on (Ω, F, P). Recall that random variables X1 and X2 are equivalent if P(X1 = X2) = 1, in which case we write X1 ≡ X2. We consider two such random variables as the same vector, so that technically, our vector space consists of equivalence classes under this equivalence relation. The addition operator corresponds to the usual addition of two real-valued random variables, and the operation of scalar multiplication corresponds to the usual multiplication of a real-valued random variable by a real (non-random) number. These operations are compatible with the equivalence relation in the sense that if X1 ≡ X2 and Y1≡Y2 then X1 + Y1 ≡ X2 + Y2 and cX1 ≡ cX2 for c ∈ R. In short, the vector space V is well-defined.

You are mistaking "pointwise scalar multiplication" for "scalar product". For $\alpha \in \Bbb R$ and $\mathbf x \in \mathbf G$ one may define the poinwise scalar multiplication by $(\alpha \mathbf x) (\omega) = \alpha \big( \mathbf x (\omega) \big)$, for $\omega \in \Omega$. There is no mention of scalar product here.
Of course, if you really wanted, you could also define a scalar product by $\langle \mathbf x, \mathbf y \rangle = \int _\Omega \mathbf x (\omega) \mathbf y (\omega) \ \Bbb d \mu$ where $\mu$ is the measure on $\Omega$, but care must be taken in order to assure the convergence of the above integral for all $\mathbf x$ and $\mathbf y$, and in order to make this scalar product non-degenerate (you'll have to work with classes of functions etc.). This could be done, for instance, by requiring $\mu$ to be finite. In any case, this is not what your text tries to convey.