I'm currently study conditional probaility form the book Probability and measure Book by Patrick Billingsley and I stuck to understanding the following thing.
We know that $\mathbb{P}(A||\mathscr{G})$ is a random variable with two properties:
(1) $\mathbb{P}(A||\mathscr{G})$ is measurable $\mathscr{G}$ and integrable.
(2) $\mathbb{P}(A||\mathscr{G})$ satisfies the functional equation $$\int_G \mathbb{P}(A||\mathscr{G})dP=\mathbb{P}(A\cap G), \ \ \ G\in \mathscr{G}.$$
Condition (i) in the definition above in effect requires that the values of $\mathbb{P}(A||\mathscr{G})$ depends only on the sets in $\mathscr{G}$. An observer who knows the outcome of $\mathscr{G}$ viewed as an experiment knows for each $G$ in $\mathscr{G}$ whether it contains $\omega$ or not; for each $x$ he knows this in particular for the set $[\omega' :\mathbb{P}(A||\mathscr{G})_{\omega^{'}}=x]$, and hence he in principle knows the functional value $\mathbb{P}(A||\mathscr{G})_{\omega}$ even if does know $\omega$ itself.
I want a explanation and a simple example of this bold font statement.
Thanks in advance
The conditional probability $\mathbb{P}(A\mid \mid \mathscr{G})$ can be interpreted as the belief about the occurence of the event $A$ is one knows which events in the sigma-field $\mathscr{G}$ occur and which ones do not occur. The simplest and most meaningful scenario is when $\mathscr{G}$ is generated by a real-valued random variable $X$, that is the sigma-field made by all the events which can be represented as $\{\omega: X(\omega)\in B\}$ for some $B$ in the Borel sigma-field. Knowing which events in such sigma-field occur is equivalent to know the value of $X$. In this scenario, what is said in the bold statement is this: if one knows the value of $X$ then he/she also knows the conditional probability $\mathbb{P}(A\mid \mid \mathscr{G})$ even if he/she does not know $\omega$.