A coin is thrown until it comes up tails. What is the probability of making at most $3$ throws?
The answer is :$\frac{7}{8}$.
I don't understand why. What are the possible cases and what are the favorable ones?
Let 1=Tail 0=Head then an event is for example: 001 or 00001 or 01 or 1 but this events are not equiprobable we have more chances to appear 1 than to appear 00000000000000000001.
In these conditions we can say :$\color{blue}1)$the probability to appear 1 is 1/2 (because we 0 and 1 possible cases) $\color{blue}2)$the probability to appear 01 is 1/4 (because we have 4 possible cases 11, 00, 10, 01) and $\color{blue}3)$ the probability to appear 001 is 1/8 (from the same reason)
Then the answer is 1/2+1/4+1/8=7/8. what is unclear for me is why actually in $\color{blue}1),\color{blue}2),\color{blue}3)$ we have prob. 1/2, 1/4, resp. 1/8 I think the explanation I gave in brackets is not rigorous.
$\color{red}{------------------------------}$
I rephrased the question.
Let $\mathcal{E}$ be an experiment. Let $Ω$ be the sample space on $\mathcal E$, and let $Σ$ be the event space of $\mathcal E$.
We know that: $\Omega=\{\underbrace1_{\omega_1},\underbrace {01}_{\omega_2},\underbrace{001}_{\omega_3},...\}$
Probability: $P:Σ=\mathscr{P}(\Omega)\rightarrow[0,\infty)$ respect Kolmogorov Axioms. Now how to prove $P(\omega_1)=1/2,P(\omega_2)=1/4,P(\omega_3)=1/8$ using just axioms.
$\def\Om{\Omega}$ $\def\om{\omega}$ $\def\si{\sigma}$ $\def\cF{\mathcal{F}}$ $\def\bN{\mathbb{N}}$
The axioms of probability are analogous to the rules of inference in formal logic. The rules of inference cannot prove anything by themselves. They must be applied to a set of premises. Without premises, there can be no proof. Similarly, the axioms of probability must be applied to premises, or assumptions.
The assumptions in your probability model are that
Unfortunately, these assumptions cannot be formulated in the probability space you wrote down in your reformulation of the question. For example, in your space, there is no event that corresponds to "The second coin flip is tails." (You have an event corresponding to "The second coin flip is tails and the first is heads.", but not "The second coin flip is tails and the first is tails." Without the latter event, there is no way to construct the event which denotes "The second coin flip is tails.")
To properly formulate the assumptions of this model, we need a more robust probability space, such as $$ \Om = \{\om = (\om_1,\om_2,\ldots):\om_j\in\{0,1\}\}, $$ with $\cF$ being the smallest $\si$-algebra containing sets of the form $$ A_{a_1,\ldots,a_n} = \{(a_1, \ldots, a_n, \om_{n+1}, \om_{n+2}, \ldots): \om_j \in \{0,1\} \text{ for $j>n$}\}, $$ where $(a_1,\ldots,a_n)\in\{0,1\}^n$. With the measurable space $(\Om,\cF)$ in place, we next need to identify the events we will need. We need events related to our assumptions as well as the conclusions we are aiming for. Informally, we need
Formally, these translate to
The assumptions can now be formulated mathematically as follows:
Only now can we begin the process of applying the axioms of probability to these assumptions in order to derive $P(F_j) = 2^{-j}$.
The first thing we need to do is prove that there exists a probability measure $P$ on $(\Om,\cF)$ that satisfies our assumptions. This can be done using the Kolmogorov extension theorem. Once we know that such a $P$ exists, the proof is very easy and follows immediately from our assumptions and the fact that $P(A^c)=1-P(A)$ (which is easily derived from the axioms of probability): \begin{align} P(F_j) &= P\bigg(T_j \cap \bigg(\bigcap_{k=1}^{j-1} T_j^c\bigg)\bigg)\\ &= P(T_1^c \cap \cdots \cap T_{j-1}^c \cap T_j)\\ &= P(T_1^c)\cdots P(T_{j-1}^c)P(T_j)\\ &= \left({\frac12}\right)\cdots\left({\frac12}\right)\quad\text{($j$ factors)}\\ &= \left({\frac12}\right)^j = 2^{-j}. \end{align}
EDIT:
I do not know what the level of your probability knowledge is. Based on the original formulation of your question, I would have guessed that your experience is at the level of a basic introductory course in non-rigorous probability, like the kind we teach to undergraduates here in the US. But your reformulation of the question mentioned Kolmogorov axioms and asked for proof. This is why I posted this answer.
It may be surprising that such a seemingly simple situation (a sequence of coin flips) requires such elaborate mathematics, involving concepts usually taught only in math graduate school. But I assure you, this is what is required, almost from the very beginning, if you want to study the rigorous proofs of probability theory.
I am sorry if this post is too complicated to be helpful, but perhaps someday you will want to return to it.