I have $N$ (about a 30) data values of which $n$ (about 4) are missing. Domain knowledge tells me that the overall distribution of my data should be roughly normal. Because of the missing values, the distribution is currently not normal.
How should I go about imputing the missing $n$ values such that my final data is as close to normal as possible? Or, is this an altogether foolish way of dealing with missing data?
Comment continued:
The scenario you propose, that normality is spoiled by a few missing observations (deleted without regard to data values), seems unlikely. So I will check it out (using R statistical software).
I sample 30 observations from a normal population. I test them for 'normality' (using a Shapiro-Wilk test). In advance, I decide to delete observations 11 through 15 (in order of collection). I test the remaining ones for normality.
Data are consistent with normal before and after deletion. (Both P-values are substantially greater than 5%.)
Another way to check for normality is to see if a normal probability plot is (roughly) linear. Plots for both datasets seem consistent with normality by this criterion also. [Slight 'wobbles' are OK, especially with such small sample sizes.]
Note: What you say can happen if I delete the largest five observations. After a dozen tries I encountered a dataset that barely passed the Shapiro-Wilk test (P-value about 0.06), but failed after deletion of the five largest observations (P-value about 0.005). Also, the normal probability plot of the data after deletion seems distinctly non-linear.