Unfortunately I could not make it to the last probability theory lecture and now I am reading through the notes and have some troubles understanding what is going on.
So we defined Brownian motion as a stochastic process on $[0,\infty)$ such that
1.) $t \mapsto X_t(\omega)$ is a.s. continuous 2.) $X_t$ has stationary $X_{t_i}-X_{t_{i-1}} \sim N(0,t_i-t_{i-1})$ and independent increments where $X_t \sim N(0,t).$
Now, we are following the book "Continuous Time Markov Processes" and I refer to the beginning of Chapter 1.7.
It says: It is most convenient in the case of Brownian motion to take the probability space $\Omega$ to be the space $C[0,\infty)$ of all continuous functions $\omega(.)$ on $[0,\infty).$ This choice is natural because Brownian paths are continuous. The process is defined by $X(t,\omega) = \omega(t).$
The $\sigma$- algebra $F$ is taken to be the smallest one for which the projection $\omega \mapsto \omega(t)$ is measurable for each $t$. Rather than having one probability measue on $(\Omega,F)$ we now have a family $(P^x)$ of probability measures indexed by $x \in \mathbb{R}$. The probability measure is the distribution of $x+B(.),$ where $B$ is standard Brownian motion.
This is all a little bit confusing to me. So far I regarded Brownian motion as a map $X: \Omega \rightarrow \big(C[0,\infty), B(C[0,\infty))\big)$ where $B$ is the Borel sigma algebra on $C[0,\infty)$ and $\Omega$ was a space that I did not really care about.
But now they seem to be changing $\Omega$ which is fairly bizarre to me. Does anybody understand what they want to do and from where to where (in particular equipped with which sigma algebra and measures) they want the brownian motion to go? As it could still be that they want to denote something different with omega, I wanted to ask the experts here.
There are at least two concepts that you might want to consider separately, before you look at this definition again: the canonical process, and time continuous Markov process. They are changing $\Omega$ because they want the process to be canonical, and they define a whole bunch of probability measures, because this is necessary to define a Markov process.
Just go step by step through the definition:
The process $(X_t)_t$ with distributions $(P^x)_{x\in\mathbb{R}}$ on the space $(C[0,\infty), \mathcal{F})$ is the Brownian motion explicitly written down as a time-continuous Markov process.
Here is a simpler construction that might help to understand why we can get rid of $\Omega$ and make everything canonical: Suppose you have a probability space $(\Omega, \mathcal{A}, P)$, a measurable space $(\Psi, \mathcal{B})$, and an $\mathcal{A}-\mathcal{B}$-measurable random variable $X$. This variable induces a measure on $(\Psi, \mathcal{B})$ as follows: $B \in\mathcal{B} \mapsto P(X^{-1}(B))$. Thus, $(\Psi, \mathcal{B}, P\circ X^{-1})$ also becomes a probability space. The identity function $Id: \Psi \to \Psi$ is trivially measurable, thus we can interpret the identity function $Id$ as a $\Psi$-valued random variable. Furthermore, notice that $Id$ and $X$ have the same law, because $P[X\in B] = P[X^{-1}(B)] = (P\circ X^{-1})[B] = (P\circ X^{-1})[Id^{-1}(B)] = (P\circ X^{-1})[Id \in B]$. Therefore, we can actually get rid of the original $\Omega$, forget $X$, forget $P$, and instead build a new probability measure on $(\Psi, \mathcal{B})$ and use the identity function on $\Psi$ instead of $X$.