When introducing Kolmogorov's axiomatic approach probability, it is often claimed that this is are a way out of the problems associated with the following two interpretations of probability:
• frequentistic approach: probability is the limit of relative frequence of an event occuring in long strings of repetitions of an experiment
• subjective approach: the (hypothetical) monetary value that I would put on a bet that an event occurs
My question is: What exactly are the problems, associated to these interpretations?
[Careful! Long sentence ahead!] Are they merely philosophical (e.g. in case of the frequentistic approach: we can't associated a probability to all events, that we would like to associated probabilities to, since not all events can formulated within repeatable experiments; e.g. to estimate probability of a politician being elected, we can hold 1000 times an election, to approximate, as the fraction of those 1000 times in which he was elected, the probability of election) or are there ``hard'' mathematical obstacles that arise when one formalizes these other approaches and states them axiomatically (e.g. in case of the frequentistic approach: we assume that $(\Omega,\mathcal{F})$ is a measurable space, let $A\in\mathcal{F}$ be arbitrary and consider for each $k\in\mathbb{N}$ a finite sequence $(B_{i}^{k})_{i\leq k}\in\mathcal{F}^{k}$ such that $\lim_{k\rightarrow\infty}\frac{N_{k}((B_{i}^{k})_{i\leq k})}{k}$ exist (which then necessarily lies within $[0,1]$), where $N_{k}$$((B_{i}^{k})_{i\leq k}):=|\{i:B_{i}^{k}=A\}|$; the value of this limit is then called probability of $A$. For the subjective approach I don't know how to formalize this)?
You can easily dispose of random variables for elementary probability (on finite sample spaces), just look at Feller’s volume 1 which does not formally introduce them until over halfway through the book. However, in the theory of stochastic processes, especially in continuous time, the concept of random variable is indispensable.
In short the first half of this post addresses the fact that the elementary approach fails (somewhat dramatically!) when combinatorics no longer applies. The second half recounts a passage of Cramer essentially commenting that while other axiomatic approaches to probability theory are certainly valuable, none seem to be as amenable to actually doing mathematics as Kolmogorov's.
Allow me to try to digest the introduction I referenced as well as provide some quoted passages from H. Cramer's book on distributions:
First on the introductory chapter of Probability with Martingales. As often in probability models, we have some event $A$ we are interested in [in the chapter its the event of extinction of the Branching process], and a sequence of events $A_n$ that are associated with $A$ somehow. When, how, and why can we say $\mathbb{P}(A)=\lim_n \mathbb{P}(A_n)$? In the specific case $A_n=\{Z_n =0\}$, the event of extinction in the $n$-th generation and $A=\{Z_m = 0 \text{ for some } m\}$ the event of extinction, ever, and $\pi_n := \mathbb{P}(Z_n=0)$ and $\pi:=\mathbb{P}(Z_m = 0 \text{ for some } m)$. It is intuitive to posit that this limit holds in this case, but how do we prove it? And even if we do prove it, what if I have a different model now, or some slight variation?
In the elementary theory we can identify the sample space with the set of all outcomes without issue and combinatorics greatly aids in our work. Now recall for this branching process, for $n\in \mathbb{Z^+}$ and $r\in \mathbb{N}$, the RV $X_r^{(n+1)}$ represent the number of children in the $(n+1)$-st generation of the $r$-th animal from the $n$-th generation (if there is one). Then the total number in the $(n+1)$-th generation is given by $$Z_{n+1}=X_1^{(n+1)}+\dotsc +X_{Z_n}^{(n+1)}$$
The $X$ are non-negative integer valued RVs and the doubly-infinite sequence $X_r^{(s)}$ for $r,s\in \mathbb{N}$ is assumed to be IID. For a single copy of $X$ we might just as well say $\Omega = \{0,1,2,\dotsc\}$. But for this experiment, as long as we want to compute $\pi :=\mathbb{P}(Z_m = 0 \text{ for some } m)$, we need to know the results of the entire sequence $(X_s^{(r)}: r,s \in \mathbb{N})$ as a single outcome of our experiment. We aren't just interested in one generation or one animal, so our sample space needs to reflect that. Thus following the author now,
Kolmogorov's measure-theoretic treatement resolves this (or settles on a compromise, depending on your preferences). We know exactly which sets can be measured (those in $\Sigma$ or $\mathcal{F}$), and we have useful limiting properties like left-continuity of measures and $\sigma$-additivity for computations. Further this answer discusses why finite additivity is not enough, even for countable sample spaces. Finally, you might be interested in example 3.7 Coin Tossing on page 32, Chapter 3 of the same book. It seems similar/related to the example at the end of the OP.
Let me also add that it is hard to formulate and state results about existence of continuous modifications of stochastic processes, let alone prove them (or any other sample-path property for that matter) without random variables, for example Kolmogorov’s continuous modification theorem: en.m.wikipedia.org/wiki/Kolmogorov_continuity_theorem.
Note also the followingm with regards to the book Rogers and Williams Volume 1, p123: The canonical sample space is nice (then the sample-path is the outcome of the experiment, as it is in elementary theory), but "probability theory gets most of its depth from being able to construct (certainly non-canonical!) processes from other processes by time transformations, or as solutions of SDEs."
(This book, volume 1, opens up with a survey of results on Brownian motion, exemplifying nearly every property/concept in the general theory (martingale, gaussian process, markov process, infinitesimal generator etc). The next chapter reviews measure theoretic probability, stochastic processes (Daniel-Kolmogorov theorem), and discrete time and continuous time martingales. After is the general theory of Markov processes.)
Now I leave you with a moderately lengthy passage from Cramer's little book on the difficulties with other axiomatic approaches proposed then (first printed 1937, 2nd edition 1962):
from Random Variables and Probability Distributions 2nd edition by H. Cramer, p3-5.