How can we calculate the standard error from a single sample?

4.5k Views Asked by At

Say we sample 19 dolphins for mercury levels. Say the summary of the sample is here:

summary stats of dolphin mercury levels

So the mean is 4.4 and the sample's standard deviation is 2.3

From this, why do we say the standard error (which is the standard deviation of the sampling distribution of means... which doesn't exist yet) is $\frac{2.3}{\sqrt{4.4}}$

What's the intuition behind this? How can be infer the SE from one sample?

1

There are 1 best solutions below

3
On

The standard error should be $$\frac{s}{\sqrt{n}} = \frac{2.3}{\sqrt{19}},$$ not $2.3/\sqrt{4.4}$ as you wrote.

To understand why, suppose we observe the sample $(X_1, X_2, \ldots, X_n)$, and furthermore, assume that these observations are independent and identically distributed from a common parametric model with finite mean $\mu$ and variance $\sigma^2$. Then the sample mean is $$\bar X = \frac{1}{n} \sum_{i=1}^n X_i,$$ and the variance of the sample mean is $$\operatorname{Var}[\bar X] = \operatorname{Var}\left[\frac{1}{n}\sum_{i=1}^n X_i\right] \overset{\text{ind}}{=} \frac{1}{n^2} \sum_{i=1}^n \operatorname{Var}[X_i] = \frac{1}{n^2} (n \sigma^2) = \frac{\sigma^2}{n}.$$ Since $\sigma^2$ is an unknown parameter, the standard error of the sample mean is computed by replacing the population variance with the sample variance; i.e., $$\operatorname{SE}[\bar X] = \sqrt{\frac{s^2}{n}} = \frac{s}{\sqrt{n}}.$$ This is where the formula comes from.