My robot has a laser time-of-flight distance sensor. Holding the distance and target constant: When I take single set of readings, three standard deviations usually is 3% to 6% of the mean.
If I take multiple sets of readings, each set will show three standard deviations to be 3-6% of the mean, BUT, averaging the averages from each set will show much lower standard deviation, such that three standard deviations will be between 0.5 to 2% of the mean.
Am I getting a more accurate reading by the average of averages of subsets, than the average of a whole set?
(Or am I just seeing the benefit of an average over individual readings?)
17:30:52
Readings : 20
Reading Delay : 0.010 s
Average Reading: 1112 mm
Minimum Reading: 1075 mm
Maximum Reading: 1156 mm
Std Dev Reading: 19 mm
Three SD readings vs ave reading: 5.1 %
Adjusted For Error Average Distance: 1099 mm
17:30:55
Readings : 20
Reading Delay : 0.010 s
Average Reading: 1110 mm
Minimum Reading: 1078 mm
Maximum Reading: 1140 mm
Std Dev Reading: 17 mm
Three SD readings vs ave reading: 4.6 %
Adjusted For Error Average Distance: 1097 mm
17:30:59
Readings : 20
Reading Delay : 0.010 s
Average Reading: 1112 mm
Minimum Reading: 1057 mm
Maximum Reading: 1157 mm
Std Dev Reading: 20 mm
Three SD readings vs ave reading: 5.5 %
Adjusted For Error Average Distance: 1099 mm
17:31:03
Readings : 20
Reading Delay : 0.010 s
Average Reading: 1115 mm
Minimum Reading: 1090 mm
Maximum Reading: 1146 mm
Std Dev Reading: 14 mm
Three SD readings vs ave reading: 3.8 %
Adjusted For Error Average Distance: 1101 mm
17:31:07
Readings : 20
Reading Delay : 0.010 s
Average Reading: 1118 mm
Minimum Reading: 1060 mm
Maximum Reading: 1157 mm
Std Dev Reading: 21 mm
Three SD readings vs ave reading: 5.6 %
Adjusted For Error Average Distance: 1104 mm
Average Average: 1113 mm
Minimum Average: 1110 mm
Maximum Average: 1118 mm
Std Dev Average: 3 mm
Three SD averages vs ave reading: 0.7 %
Ave all Readings: 1112 mm
SDev all Reading: 16 mm
Three SD all vs ave all readings: 5.0 %
There is natural fluctuation in the measurements, and the standard deviation $(\sigma)$ measures the size of those fluctuations.
The averages fluctuate less, by $\sigma/\sqrt n$. That is the point of taking a mean. So you can divide those $\sigma$, which are about 16, by $\sqrt{20}$, to say how precise each mean is. The precision is 3 or 4.
In the final answer, you have 100 measurements, so your final average is accurate to $16/\sqrt{100}\approx1.6$. On the other hand, you have five measured averages, each accurate to 3mm, so the overall precision is $3/\sqrt5\approx1.34$. I think the difference between 1.6 and 1.34 is roundoff error.
Three standard deviations in each of the five estimates would be $3×16/\sqrt{20}\approx11mm$. Three standard deviations in the overall average would be$3×1.6\approx5mm$.