I need some help understanding the fallacy in the following reasoning (thank you in advance!). It is essentially implying that for a single sample from a population, you can know a population parameter (like the EV) precisely with a few seemingly reasonable assumptions. I’ll frame it as an example.
Let’s say that we are trying to infer something about the return distribution (assume normal) of a financial instrument (population distribution). We make an assumption about the variance of this distribution. We then create samples and calculate their averages, and each sample consists of 100 draws from the population distribution (which again we don’t know).
Using our assumption about the variance of the population distribution, we can calculate the variance of the sampling distribution, which is the distribution which results after sampling (again, each sample being of 100 draws from the population distribution and the resultant average) infinitely many times. We can compute a 95% confidence or Bayesian credible interval using the variance of the sampling distribution, which implies that if we take 1 incremental sample and calculate its average, the probability that the population mean falls within said interval around the sample average is 95% (Yes I realize frequentists will challenge this for confidence intervals, but the credible interval unequivocally states this). So, if we take an incremental sample, it would follow that the probability that the population mean falls within the band described by the confidence interval around the estimated mean of the sample is 95%.
We’ve now basically created a probability distribution for the population mean itself, and the midpoint of the distribution is the estimated sample mean. If we assume that this distribution is normally distributed or even just that the likelihood that the population mean/EV falls within the 95% confidence band around the estimated sample mean or the in the 2.5% tails on either side of the band is uniform, then that would imply that in expectation, the expected value of the population mean/EV is the estimated sample mean.
It would seem obvious that that cannot be true given it would imply that that you could then know your true EV for any population simply by referencing a sample of any size. You can’t say Steph Curry’s free throw make probability is 50% after watching him shoot 2 free throws and miss 1.
Where exactly does this go wrong?
Suppose we have a random variable $X$ with finite mean. That random variable has some distribution.
For a fixed $n$ and iid samples $X_1, \ldots, X_n$ from the distribution of $X$, we can construct the random variable $\frac{1}{n}\sum_{i=1}^n X_i$, which has its own distribution (possibly) distinct from the distribution of $X$. When you talk about the distribution we created from the sample, it doesn't sound like you're talking about this distribution.
If we observe actual values $x_1, \ldots, x_n$, we can construct the empirical distribution, which has cdf $F(x)=\frac{1}{n}\sum_{i=1}^n 1(x\leqslant x_i)$, where $1(x_i\leqslant x)=1$ if $x\leqslant x_i$ and $1(x_i\leqslant x)=0$ otherwise. This is also (possibly) distinct from the two previous distributions. When you talk about the distribution we created from the sample, it doesn't sound like you're talking about this distribution.
If we finally take the observed sample mean $\overline{x}=\frac{1}{n}\sum_{i=1}^n x_i$ and select some distribution which is symmetric about $\overline{x}$ (note that there are many such distributions), we get another distribution which is generally not equal to any of the previous three. When you talk about the distribution that we created from the sample, based on your comments, it sounds like you're talking about one of these distributions. But knowing this distribution doesn't tell us the population distribution (or the other two preceding distributions).
The fact that the confidence interval is symmetric about the sample mean doesn't mean that the population distribution (or the distribution of $\frac{1}{n}\sum_{i=1}^n X_i$ or the empirical distribution) is symmetric about the sample mean. For example, if we have a normal population with mean $0$ and we pick a sample which has sample mean $1$, the population distribution is still symmetric around $0$, not $1$. The confidence interval we construct is symmetric about $1$.