In introductory undergraduate probability courses, even those with a focus on set theory, I've often seen the definition of a sigma field entirely skipped over. Indeed, I've often seen the definition of a probability space include the claim "where $\mathcal{F}$ is the set of all events". Now, I've got no doubt that sigma fields are necessary in probability theory, so I'm forced to ask. How is it that these undergraduate courses can manage to teach probability theory without making any mention of sigma fields? What do they lose by skipping over this definition? Or rather, what do they have to manipulate in order to avoid it?
At the very least, I can recall from my own time as a first year undergraduate that I would occasionally get confused over what exactly an "event" is and at least in principle, I can see where "the set of all events" part of the definition could be confusing. For example, rolling a six-sided dice would give you a perfectly valid sigma field of { {6}, {not 6}, $\emptyset $, $\Omega$ } which wouldn't seem to fit the earlier definition (e.g. where's the "not 5" event?), unless you cheat the sample space in some way that I would be unsure of the validity of.
Admittedly, there's an awfully cynical part of me that thinks that the answer may be "by hoping that nobody thinks too hard about it" and indeed, inspecting my notes from such courses suggests that may genuinely be the answer. But I'm hoping that the good people at Stack Exchange can show otherwise.
It is crucial for students of probability to learn that an event is simply a subset of the sample space. This is not trivial. It is deep. What we can describe in words about “what happened” in the probability experiment can be captured precisely by a subset. Arguably, a heavy emphasis on sigma fields early in a probability course can distract attention away from this important concept.
If you have ever taught an introductory probability course, you will know that students struggle with very basic things. Students forget what the equal sign means. They write “equations” where the left-hand-side is not the same type of object as the right hand side, such as having a random variable on the left and a PDF on the right. Consider the following tragic equations of this flavor:
\begin{align} &X \sim U(0,1) \implies \frac{X}{2} = \int_0^1 x dx\\ &P[X] = f(x)dx\\ &P[A] = P[A | B_i]P[B_i] \quad \forall i \end{align}
A detailed discussion of sigma fields, with repeated pointers to its heavy notation and scary terminology, will not help. Arguably, even the most advanced students in the class must first master basic probability calculations before they can appreciate the foundational issues regarding non-measurability that arise (only) when we deal with uncountably infinite sample spaces. [Recall that if the sample space $S$ is finite or countably infinite, we can without loss of rigor define the sigma field to be the set of all subsets of $S$. More restrictive sigma fields are only required for uncountably infinite sample spaces.]*
Rhetorical question: Would students ever learn arithmetic if, first, they were forced to read the foundational work Principia Mathematica by Russell and Whitehead?
*Edit [A footnote to my bracketed sentence above]: Let $S$ be a finite or countably infinite sample space. Define $F$ as the set of all subsets of $S$. Then for any collection of values $(p(\omega))_{\omega \in S}$ that satisfy $p(\omega) \geq 0$ for all $\omega \in S$ and $\sum_{\omega \in S} p(\omega) = 1$, we can define the following function $P:F\rightarrow\mathbb{R}$: $$ \boxed{P[A] = \sum_{\omega \in A} p(\omega) \quad \forall A \subseteq S \quad (\mbox{Equation *})}$$ It is easy to see that this function $P$ satisfies the three axioms of probability and hence it is a valid probability measure:
$P[A] \geq 0$ for all $A \subseteq S$.
$P[S] = 1$.
If $\{A_1, A_2, A_3, ...\}$ is a sequence of disjoint subsets of $S$ then $P[\cup_{i=1}^{\infty} A_i] = \sum_{i=1}^{\infty} P[A_i]$.
Conversely, any function $P:F\rightarrow\mathbb{R}$ that satisfies the above three axioms must have the form of (Equation *) with $p(\omega) = P[\{\omega\}]$ for all $\omega \in S$. The values $p(\omega)$ form the probability masses for each outcome of the finite or countably infinite sample space.
Thus, for finite or countably infinite sample spaces, it makes sense to define the sigma algebra $F$ (which contains all events) as the set of all possible subsets of the sample space.