Why normal approximation to binomial distribution uses np> 5 as a condition

30.6k Views Asked by At

I was reading about normal approximation to binomial distribution and I dunno how it works for cases when you say for example p is equal to 0.3 where p is probability of success.

On most websites it is written that normal approximation to binomial distribution works well if average is greater than 5. I.e. np> 5 But I am unable to find where did this empirical formula came from?

If n is quite large and probability of success is equal to .5 then i agree that normal approximation to binomial distribution is going to be quite accurate. But what about other cases? How can one say np> 5 is the condition for doing normal approximation?

4

There are 4 best solutions below

4
On

The condition $np > 5$ is not the condition, merely a rough estimate of what should be true in order for the normal distribution approximation to be "good enough".

From Wikipedia:

One rule is that both $x=np$ and $n(1 − p)$ must be greater than 5. However, the specific number varies from source to source, and depends on how good an approximation one wants.

There you can also find a list of other "rules".

5
On

So I did some experiments. I think np>5 condition is not correct at all. It depends on Excess Kurtosis value for a given binomial distribution. If it is Mesokurtic then approximation will give accurate results.

Check following table enter image description here

for n=11 and p=0.5 kurtosis will be around 0.18. That is platykurtic and so I don't think approximation will give accurate results, even though n*p=5.5 > 5. The table shows results which manifests what I am trying to say.

0
On

Here's how I'm thinking of these conditions.

Note that if a random variable is truncated near its mean (i.e. the absolute value of the z score of the truncated value isn't too large) then the random variable's distribution will be skewed away from the truncated value and toward its mean.

That being said, observe that a binomial random variable X~B(n,p) is truncated at 0 and n. The condition np>10 pushes the distribution away from the truncation at 0, while n(1-p)>10 pushes the distribution away from the truncation at n. This will assure us that the distribution of X won't be undesirably skewed in any direction.

Think of np and n(1-p) as the expected number of success and failures in a series of n trials, respectively.

Hope this helps.

0
On

For $np$ and $nq$ to increase $n$ must increase. $n$ is the number of independent trials, so it should be clear that the more independent trials made, the more accurate your approximation is. The probability histogram approximates a normal curve pretty accurately when $np$ and $nq$ are greater(or equal to) $5$. However bigger is better! If $np$ and $nq$ were greater than $10$ the probability histogram would approximate the normal curve even more.