I have a fundamental question on statistical tests, particularly tests for normal distribution. As I understand statistical tests in general, they have a null hypothesis $H_0$ (e.g. the samples were drawn from a normal distribution) and an alternative hypothesis $H_1$ (e.g. the samples were not drawn from a normal distribution). If the test is significant ($p < p_\alpha$) one can reject $H_0$ and assume that $H_1$ is true. However, if the test is not significant, one can not automatically assume that $H_0$ is true.
Now, all tests for normal distribution that I read about have a $H_0$ that the samples were drawn from a normal distribution. Hence, the only thing you can do with these tests is to assume that the samples were not drawn from a normal distribution if the test is significant. You can't assume that the samples are drawn from a normal distribution if the test is not significant. But that's what everybody seems to be doing.
Is there anything fundamentally wrong with my understanding of statistical tests? How can I "prove" that a given sample was drawn from a normal distribution?
When the test statistic falls in the significant region, we reject the null hypothesis that the samples are drawn from a Gaussian distribution.
When the test statistic does not fall in the significant region, we fail to reject the null hypothesis that samples are drawn from a Gaussian distribution - we cannot say anything more.
Even this 'failing to reject' has a lot of use cases.
For e.g. in the output of OLS regression where $\beta_i$'s (the coefficients) are estimated, we see the p-values reported. These p-values are nothing but they are rejecting the null hypothesis that $\beta_i = 0$. Essentially, they are reporting a significant linear relationship.
With regards to your question, your understanding is correct.
One can never be 100% sure that a sample comes from a Gaussian distribution. We can only give a confidence level. Typically, more the sample size, better the confidence level.
However, there are variations across each of these statistical tests readily available in statistical software. Certain tests are 'good' for detecting certain aspects based on their statistical power.