I have a question concerning the covariation of two time-discrete stochastic processes.
Let $(\mathcal{F}_i)_{i\in T}$ be a filtration. We call a time-discrete, real-valued, adapted process $X$ (e.i. $X : \mathbb{N} \to \mathbb{R}^{\Omega}$ and $X_i$ is $\mathcal{F}_i$-measurable) a discrete semi-martingale.
A theorem from Doob states, that every discrete semi-martingale $X$ can be uniquely represented as $X = M + A + X_0$, where $M$ is a Martingale, $A$ is a predictable process and $M_0 = A_0 = 0$. This representation is called the Doob decomposition of $X$. The process $A$ is called the compensator of $X$.
For two discrete semi-martingales $X,Y$ we call $[X,Y]_n := \sum_{i=1}^n \Delta X_i\Delta Y_i$ the variation of $X$ and $Y$. The compensator of the discrete semi-martingale $[X,Y]:=([X,Y]_n)_{n\in \mathbb{N}}$ is called the covariation of $X$ and $Y$ andis denoted by $\langle X,Y \rangle$.
One can show that the variation as well as the covariation are symmetric and bilinear. The variation maps a tuple of discrete semi-martingales onto a discrete semi-martingale. The covariation maps a tuple of discrete semi-martingales onto a predictable process.
Here is my first question: Is there any connection between a scalar product (or even a Hilbertspace) and the variation or covariation?
Here is my second question: Is there any functional analytic interpretation of the Doob decomposition (e.g. in words of projections) ?(Answered, thank you A Blumenthal!)
I'll try to answer your second question.
Here we have a filtration $\mathcal{F}_n$ on a probability space $(\Omega, \mathcal{F},\mathbb{P})$. Letting $H = L^2(\mathcal{F},\mathbb{P})$, we can see the sigma-algebra filtration as giving rise to a filtration of subspaces $H_n = L^2(\Omega,\mathcal{F}_n, \mathbb{P})$ (all this means is that $H_1 \subset H_2 \subset \cdots$ is an increasing sequence of closed subspaces of $H$ ordered by inclusion).
Since we're working with a single measure $\mathbb{P}$, from here I'll write $L^2(\mathcal{F})$ as the space of $L^2$ random variables measurable with respect to $\mathcal{F}$.
It is a fact that the conditional expectation of $L^2$ functions/random variables can be thought of as orthogonal projection onto a subspace. To wit, if $X \in L^2(\mathcal{F})$ and $\mathcal{G} \subset \mathcal{F}$ is a sub-sigma-algebra, then the mapping $X \mapsto \mathbb{E}[X \mid \mathcal{G}]$ is the orthogonal projection of $L^2(\mathcal{F}) \rightarrow L^2(\mathcal{G})$.
We can paint a geometrical picture of martingales by using this fact. Let $\{X_n\}$ be an $\mathcal{F}_n$, square-integrable martingale. Then the martingale property asserts that each martingale difference $X_{n+1}-X_n$ is the orthogonal component to $L^2(\mathcal{F}_n)$ of $X_{n+1}$; thus such a martingale is 'built up' to the present time $n$ from $X_1$ by, at each time step, adding an orthogonal piece to the previous timesteps. This should remind you of $\mathbb{R}^n$, where you can think of a vector $(x_1,\cdots,x_n)$ as being 'built up' by $(x_1,0,\cdots,0)$, then $(0,x_2,0,\cdots,0)$ and so on.
Now let $X_n$ be a 'semimartingale' (I've never heard that terminology; I call this an adapted process). The meaning of the predictable compensator $A_n$ in this context is the component of the difference $X_{n} - X_{n-1}$ lying in $L^2(\mathcal{F}_{n-1})$ (and so it is the 'nonorthogonal obstruction' to the martingale property). It should be no surprise then that it's useful to take $A_n$ to be predictable- it'll always be the part of the difference that 'spills over' into the previous subspace.