Why not defining a measure as a function on functions?

1.3k Views Asked by At

A measure $\mu$ is a function to $\left[0,\infty\right]$ on the sets belonging to a $\sigma$-algebra.

Then for integrable functions $f$ the integral $\int fd\mu$ comes in, having nice properties like: $$\int f+gd\mu=\int fd\mu+\int gd\mu$$ and $$\int cfd\mu=c\int fd\mu$$ Often I wonder: is there anything that keeps us from defining measure $\mu$ as function on integrable functions (or eventually nonnegative integrable functions) instead of sets?

And in the same line: why not writing $\mu\left(f\right)$ or $\mu f$ instead of $\int fd\mu$? The original value on set $A$ can easily be found back as $\mu\left(1_{A}\right)$.

Are there good reasons not to do this? And if so can you give me some?

1

There are 1 best solutions below

0
On BEST ANSWER

First of all, if one wants a "measure defined on a set $\mathcal{F}$ of functions" and if

1) $\mathcal{F}$ only contains functions measurable w.r.t. some $\sigma$-algebra $\mathcal{A}$,

2) $\mathcal{F}$ contains all indicator functions $1_A$ for $A \in \mathcal{A}$,

then it is just more economical to only require a measure to be defined on the sets $A \in \mathcal{A}$ and then use the usual measure-theoretic arguments/theorems to extend the integral.


But of course, there are cases where the two conditions above are not fulfilled, for example one starts from a functional $\mu : C(K) \to \Bbb{R}$, where (e.g.) $C(K)$ are the continuous functions on some compact Hausdorff space $K$. If $\mu(f) \geq 0$ in case of $f \geq 0$, then the Riesz representation theorem shows that there actually is a measure $\nu$ such that $$ \mu(f) = \int f \, d \nu \qquad \forall f \in C(K). $$ The same holds if the space $C(K)$ is replaced by the space $C_c (\Omega)$ of all compactly supported functions on a locally compact Hausdorff space $\Omega$.


There is even a more general construction, the Daniell integral, which is described well in the Wikipedia article and also in Section 4.5 of the (excellent) book "Real Analysis and Probability" by Richard M. Dudley (who calls it the Daniell-Stone integral).

I will follow here the path of Dudley. We start with a so-called vector lattice $F$ of functions $f : X \to \Bbb{R}$ for some base-set $X$. This means that $F$ is a vector space and that if $f,g \in F$, then also $\max\{f,g\} \in F$.

Then, suppose that we are given a linear functional $\mu : F \to \Bbb{R}$ such that

1) If $f \in F$ satisfies $f \geq 0$, then $\mu(f) \geq 0$,

2) If the sequence $(f_n)_n$ in $F$ converges pointwise nonincreasing to $0$, then $\mu(f_n) \to 0$.

Finally, if $F$ is a Stone vector lattice, i.e., if $\min\{f,1\} \in F$ for all $f \in F$, then there is a measure $\nu : \mathcal{A} \to [0,\infty]$, defined on some $\sigma$-algebra $\mathcal{A}$, such that all $f \in F$ are $\mathcal{A}$-measurable and $$ \mu(f) = \int_X f \, d\nu \qquad \forall f \in F. $$

Hence, in most reasonable cases, there is not much difference between the two approaches.

Furthermore, the measure $\nu$ is uniquely determined by the properties above on the smallest $\sigma$-ring for which all $f \in F$ are measurable.


Finally, I have seen the notation $\mu(f)$ some times, in particular (IIRC) in the book "Foundations of modern probability" by Olaf Kallenberg. Thus, it seems to be more or less common in the probabilistic community.