I was reading this text on probability and I have confusion interpreting a statement. The text reads as follows:
Sampling is a very common technique for estimating the fraction of elements in a set that have a certain property. For example, suppose that you would like to know how many Americans plan to vote for the Republican candidate in the next presidential election. It is infeasible to ask every American how they intend to vote, so pollsters will typically contact n Americans selected at random and then compute the fraction of those Americans that will vote republican. This value is then used as the estimate of the number of all Americans that will vote Republican. For example, if 45% of the n contacted voters report that they will vote Republican, the pollster reports that 45% of all Americans will vote Republican. In addition, the pollster will usually also provide some sort of qualifying statement such as
"There is a 95% probability that the poll is accurate to within (+/-) 4 percentage points."
Many people interpret the qualifying statement to mean that there is a 95% chance that between 41% and 49% of Americans intend to vote Republican. But this is wrong!
Later they do a bit of math and conclude with the actual meaning of that qualifying statement and thus say that:
There is a 95% chance that the sample group will produce an estimate that is within (+/-) 4 percentage points of the correct value for the overall population. So either we were “unlucky” in selecting the people to poll or the results of the poll will be correct to within (+/-) 4 points.
I don't see any difference between the two interpretations. Can anyone please help?
What I understand from their last statement is that there is a 95% chance that the result they obtained from that particular sample group (i.e., 45% of Americans will vote Republican) is within (+/-) 4 % of the correct fraction (Let it be p). So in other words there is 95% chance that 'p' is between 0.41 and 0.49, which is precisely what the first interpretation (which they consider wrong) meant.
The only actual philosophical aspect to this question has to do whether we are talking about parameters in the Bayesian sense, or the frequentist sense. The quoted text implies that the discussion falls within the latter context.
So, in the frequentist view of statistical inference, one must be clear about what is random and what is fixed; what is observed or observable, and what is unknown or unknowable.
For instance, suppose we are interested in the mean age (measured as the number of whole years lived) of the human population of Earth as of January 1, 2018. This is represented by a single number whose value theoretically could be calculated, but is so impractical to do that it is effectively impossible. Yet we know it is definite and fixed. Such a quantity is a parameter. It is not subject to randomness. It is a fixed but unknown quantity that is a property of the distribution of ages of all people on Earth at that given moment.
Now, if we take a simple random sample of people and calculate the mean age of the sample, we intuitively understand that this may give some idea of the value of this parameter, but each time we take such a random sample, the sample mean is merely an estimate whose value is not fixed, but may change from one sample to another. The process of selecting people for each sample is where the randomness comes from. The outcome of each sample is a realization of this underlying random process, and it is called a statistic.
So, to recap, the parameter is unknowable but fixed; the statistic (and the random variable(s) from which it is calculated) is observable but random--its value changes each time it is realized.
Instead of merely providing the sample mean each time you take a sample, and using this as your estimate of the true population mean, you could incorporate some measure of uncertainty about your estimate, and this generally depends on the size of your sample. Intuition would suggest that the larger your sample, the more information you have about the population, and the more precise you are able to say the resulting estimate might be. So a confidence interval captures this idea by giving an estimate that is an interval rather than a single number. But because a confidence interval is also an estimate that is derived from the sample, it too is random.
To emphasize, the parameter is fixed but unknown. Each time a sample is taken, the calculated confidence interval varies. By random chance, some confidence intervals might not contain the value of the parameter (but you don't and can't know when this happens for a given sample). However, by adjusting the width of the confidence interval, you can state that the interval has a certain coverage probability that represents the probability that a random sample is taken that results in a confidence interval containing the value of the parameter. The larger you want this coverage probability or confidence level, the wider the resulting interval.
Back to the original question, we see that it makes no sense to state that there is "95% chance that between 41% and 49% of Americans intend to vote Republican," because this amounts to saying that the parameter (the true proportion of eligible Americans intending to vote Republican) is random from sample to sample, and that the confidence limits of 41% and 49% are fixed; when in fact, it is the 41% and 49% that are random, having been calculated from the sample, and the parameter, reflecting the true state of reality, is fixed.
This is a very confusing concept to understand, because when a statistic is realized, it is easy to forget that this was just one possible outcome. We are tempted to think, "well, we know what we saw, so it is fixed." But no! The observation is what it is, of course, but it is just one possible outcome of many. We would not find it reasonable to suppose that a fair six-sided die, rolled once, would then continue to give us the same outcome henceforth.
To be absolutely clear about your example, the incorrect interpretation is to say that there is some "chance" that the parameter is going to fall between two fixed values. The correct interpretation is to say that there is some chance that a random sample will result in an interval estimate that contains the true value of the parameter.
A Bayesian thinks all of the above is misguided, and instead regards the data as fixed, and the parameter as random. And this leads to some very interesting ways of conducting statistical inference, but this discussion is not within the scope of your question.