$X=(X_1,X_2,\ldots, X_P)$ is a $p$-dimensional random variable on $(\Omega, S, P) $ iff $X_i$'s are univariate random variables on the same probability space $(\Omega, S, P)$ ."
We all know that. it is written and proved in several books.
But I'm confused with the 'same probability space' thing.
Let me explain my problem with an example : Suppose $X_1$= the height of a boy selected at random from a class of 50 boys. $X_2$ =the weight of the same boy selected at random from a class of 50 boys.
Now in both the cases the sample space $\Omega = (-\infty,+\infty)$ . but the $P$ in the probability space $(\Omega, S, P)$ for both the random variable must be different . let's consider $X_1 \sim \mathrm{N}(5, 1)$ and $X_2\sim \mathrm{N}(45, 4)$. Then $X_1$ and $X_2$ are not defined on same probability space. (or , are they ? I'm kinda confused)
So , $X = ( X_1, X_2)$ , is it not a bi-variate random variable?
[Our professor was talking about constructing multivariate random variable taking different bio-metric measures of an individual . But, if, all the univariate random variables (components of a multivariate random variable) to be defined on the same probability space be a necessary and sufficient condition, then constructing a multivariate random variable taking different bio-metric measures must not be possible.]
This is my first question here, hoping for good responses.Although I don't know whether I'm able to put my query in a compact form. If you find any ambiguity kindly ask.
First, the text below requires that you know definitions of a real random variable (or vector), measurable function, Borel $\sigma $-algebra and cumulative distribution function.
Let $X$ be an $n$-dimensional random vector with the CDF ${F}$ (this is a fundamental assumption which begs the question does the appropriate probability space and the random vector exist, we proceed to show that they do).
Then the relation ${F}\left( x \right) = {\mathbb{P}}\left( {X \leqslant x} \right)$ (here $X \leqslant x \Leftrightarrow {X_i} \leqslant {x_i},\forall i$) uniquely determines a probability $\mathbb{P}$ on ${\mathbb{R}^n}$ (this is non-trivial result and the proof is cumbersome). Let's first construct the underlying probability space.
Let $\Omega = {\mathbb{R}^n}$, $\mathcal{F} = {\mathcal{B}^n}$ the Borel $\sigma $-algebra on ${\mathbb{R}^n}$ and $\mathbb{P}$ as above. Now we define the random vector $X:{\mathbb{R}^n} \to {\mathbb{R}^n},X\left( x \right) = x$. It follows, using the standard notation ${\mathbb{P}_X}\left( B \right) = \mathbb{P}\left( {{X^{ - 1}}\left( B \right)} \right)$ for the probability ${\mathbb{P}_X}$ induced by $X$ on ${\mathbb{R}^n}$, that ${F_X}\left( x \right) = {\mathbb{P}_X}\left( {\left( { - \infty } \right.,\left. x \right]} \right) = \mathbb{P}\left( {{X^{ - 1}}\left( { - \infty } \right.,\left. x \right]} \right) = \mathbb{P}\left( {\left( { - \infty } \right.,\left. x \right]} \right) = F\left( x \right)$, so $\left( {\Omega ,\mathcal{F},\mathbb{P}} \right)$ and $X$ are indeed such a probability space and a random variable that the given $F$ is the corresponding cumulative distribution function.
Now on to your specific example: ${X_i}$ is a random variable on $\left( {\Omega ,\mathcal{F},\mathbb{P}} \right)$ because $X_i^{ - 1}\left( {{B_i}} \right) = {\left( {{\pi _i}\left( X \right)} \right)^{ - 1}}\left( {{B_i}} \right) = {X^{ - 1}}\left( {\pi _i^{ - 1}\left( {{B_i}} \right)} \right)$, where ${{\pi _i}}$ is $i$-th coordinate projection and ${B_i} \in \mathcal{B}\left( \mathbb{R} \right)$. Since projections are continuous, they are measurable, so $\pi _i^{ - 1}\left( {{B_i}} \right) \in {\mathcal{B}^n}$ (it's helpful to imagine this Borel $\sigma $ algebra as a $\sigma $ algebra in the codomain of $X$) which implies that ${X^{ - 1}}\left( {\pi _i^{ - 1}\left( {{B_i}} \right)} \right) \in \mathcal{F}$, since $X$ is a random vector.
So, if ${X_1}$ is the height and ${X_2}$ the weight of an individual, then $\Omega = {\mathbb{R}^2}$ (and not $\Omega = \mathbb{R}$ as you assumed) and $X_1^{ - 1}\left( {\left[ {a,b} \right]} \right) = {X^{ - 1}}\left( {\left[ {a,b} \right] \times \mathbb{R}} \right) = \left[ {a,b} \right] \times \mathbb{R} \in {\mathcal{B}^2} = \mathcal{F}$. Also, ${X_1},{X_2}:\Omega \to \mathbb{R}$ and $X = \left( {{X_1},{X_2}} \right):\Omega \to {\mathbb{R}^2}$, so ${X_1},{X_2},X$ have the same domain.
The reason why you CAN forget about the underlying probability space is exactly the construction given above which guarantees that such a space always exists. However, I've found that it's sometimes helpful to be aware of it.