Consider the following characterization of the Bayes' theorem:
Bayes' Theorem
Given some observed data $x$, the posterior probability that the paramater $\Theta$ has the value $\theta$ is $p(\theta \mid x) = p(x \mid \theta) p (\theta) / p(x)$, where $p(x \mid \theta)$ is the likelihood, $p(\theta)$ is the prior probability of the value $\theta$, and $p(x)$ is the marginal probability of the value $x$.
Is there any special reason why we call $p(x)$ the "marginal probability"? What is "marginal" about it?
The explanation I was given when I was taught conditional probabilities is that if you draw up a table of the probabilities $p(x,y)$, then the row/column sums $$ p(x) = \sum_{y} p(x,y) $$ (by the law of total probability) are written in the margins of the table.