The definition of Conditional Probability for events $A$ and $B$ in sample space $S$ is $$\mathbb{P}(A|B)=\frac{\mathbb{P}(A\cap B)}{\mathbb{P}(B)}.$$
Sometimes, we use a rearranged version of this formula to calculate the probability of the intersection of events - called the multiplicative law of probability:
$$\mathbb{P}(A\cap B)=\mathbb{P}(A|B)\times \mathbb{P}(B)$$
When using this formula, how does one calculate $\mathbb{P}(A|B)?$ Since by definition the intersection is required to find the conditional probability? Is there an alternative definition/way to compute the conditional probability when you don't know the intersection?
I have calculated the conditional probability through intuition many times (E.g. picking Red/Blue marbles out of a bag, without replacement), but I was wondering if there was some sort of standard convention on how calculate the conditional probability when you don't know the intersection?
Example.
Say we have three people (Alex, Bob, Carol) with their three hats. Say I take all their hats, mix them up, and then return one to each person. What is the probability that person A and B get exactly their own hat back?
"Solution": The way I would think of it is: Let $E_A$ and $E_B$ be the events that Alex and Bob get their hats back respectively. Then, $$\mathbb{P}(E_A\cap E_B)= \mathbb{P}(E_B)\times \mathbb{P}(E_A|E_B)$$
The probability of $E_B$ would be $\frac{1}{3}$. Now, the way I would calculate $\mathbb{P}(E_A|E_B) $ intuitively, even though I don't know what the intersection is (because that's what I'm trying to find), is "Since Bob has his hat, I have two hats left, which gives a probability of $\frac{1}{2}$ for Alex to get his hat back."
This intuitive logic of getting to the conditional probability directly, when I didn't use/bypassed the definition, is what I would like to clarify/formalise.
This doesn't answer the question exactly, but is a bit much for a comment:
Sometimes the conditional probability is already known. For example, let $\{X_n:n=0,1,2,\ldots\}$ be a Markov chain on the nonnegative integers with initial distribution $\alpha$ and transition matrix $P$, that is, for each nonnegative integer $i$ we have $\mathbb P(X_0=i)=\alpha_i$ and for each pair of nonnegative integers $i,j$ we have $$ \mathbb P(X_{n+1} = j\mid X_n = i) = P_{ij}, $$ (the $(i,j)$-entry of $P$). Then, the distribution of $X_1$ would be given by $$ \mathbb P(X_1 = j) = \sum_{i=0}^\infty \mathbb P(X_1 = j\mid X_0=i)\mathbb P(X_0=i) = \sum_{i=0}^\infty P_{ij}\alpha_i. $$