let' say I have two different phenomena classes, and I extract two different kinds of values for each of them. For example, comparing two different leaves, I extract length and weight of several hundreds of instances.
From this experimental observation values, I calculate the mean and standard derivation, and assume they follow a normal distribution, so:
$L_{1}\sim\mathcal N(\mu_{11},\sigma_{11})$ Length distribution for first class of leaf
$W_{1}\sim\mathcal N(\mu_{21},\sigma_{22})$ Weight distribution for first class of leaf
$L_{2}\sim\mathcal N(\mu_{31},\sigma_{32})$ Length distribution for second class of leaf
$W_{2}\sim\mathcal N(\mu_{41},\sigma_{42})$ Weight distribution for second class of leaf
I want to take the characteristic that better distinguishes between both classes, so I need some kind of measurement of distance between $L_{1},L_{2}$ and $W_{1},W_{2}$, to take the one with longest distance. Which mathematical notion helps me here?


This could be very late a solution but it may benefit future reads.
In general in pattern-recognition, when the two distributions have equal variance we apply mahalanobis distance. But your features have different variance and the mahalanobis distance would tend to zero (for details, read on wikipedia ).
For your case, Bhattacharyya bound would work. This is used in general to compare Gaussian distributions with different variance. (It can also be used with other distributions).
For your example, distance between $L_1$ and $L_2$ can be computed by following equation:
\begin{equation} D_{L_1L_2} = \frac{1}{8} (\mu_{11}-\mu_{31})^T \sigma^{-1}(\mu_{11}-\mu_{31}) + \frac{1}{2} \ln (\frac{\sigma}{\sqrt{\sigma_{11}\sigma_{32}}}) \end{equation}
where, $\sigma=\frac{\sigma_{11}+\sigma_{32}}{2}$.
Similarly, you can compute distance between $W_1$ and $W_2$.