Update:
This question arises when I read a neuroscience paper. In the natural science community, people are generally less careful about the correctness of statistical language. I am clearly aware of the definitions of Type I error and Type II error. In statistics classes, we were taught that we should say either "reject the null" or "fail to reject the null" at a given significance level regarding Type I error rate. But we were rarely taught about the language used to describe Type II error.
The author hypothesized that two samples are from the same distribution. The statistical test was performed using the following hypotheses.
The null hypothesis: the two samples are from the same distribution.
The alternative hypothesis: the two samples are from different distributions.
The author got a p-value=0.83. Here the p-value is the same as the type I error rate. At significance level=0.05, we fail to reject the null hypothesis. That is, there is 83% chance that they are from different distributions when we say they are not.
When the neuroscientists see something like this, they are more or less convinced that the two samples are from the same distribution. Of course, as a mathematician, this is far from rigorous.
I am wondering if calculating the Type II error will strengthen the argument. If my type II error rate is very small, say 0.03. What would be the correct statistical language to draw conclusion?
Basically, I want to know the implication of the null/alternative given a high type I error and a low type II error and the rigorous statistical language to describe it.
We could obtain a high p-value of $0.83$ if $F=G$ or close to equal or if the sample size was small for this type of test. By determining the $\beta$ or the power $1-\beta$ we can gain some information. Unfortunately that is not always easy.
The Type II error is, in general, a function not just a value. As drphil pointed out, you choose a Type I value (along with the specific hypothesis test and sample size) and then Type II value(s) are determined. We can use that information (if easily computable) to choose between test statistics. e.g., chi-square or Kolmogorov-Smirnov test. That is the quest for uniformly most powerful tests. Or we can use the information to determine a sample size to use to give chosen $\alpha$ and $\beta$ values. Here you would have to specify a particular alternative hypothesis for the $\beta$ value.
In the example you cite, if the alternative is true then $F\ne G$ but that is not enough to compute $\beta.$ In this problem, suppose we want to detect a difference between F=exponential mean 1 and G=exponential with mean 5. With these $F,G$ as the population cdfs then we need to find the probability of rejecting $F=G$ using our particular test. That is the power $1-\beta.$ If that is an important difference to detect if it is true in this example then we want $1-\beta = 0.95$ or higher. Select the sample size to do that. (Might have to be done by simulation.) This will then fix the $\beta$ values for all other types of alternative hypotheses.
So in summary the power function of the hypothesis test is what you are looking for. There is no further "reject\do not reject" decision.
P.S. The interpretation of $0.83$ is not correct:
"83% chance that they are from the same distribution when we say they are not." That is (almost) the statement of a Type II error probability. The correct interpretation is: if we had used $\alpha \gt 0.83$ then we would have Rejected the null hypothesis. Or it is the probability of getting a sample result more extreme (more favorable to the alternative) than the result we actually obtained.