Let we have a competitive survival game in which a player has choice between different resources to earn. The question here is which resource should he prefer to maximize the chance of survival. I tried to find out if there already established measures of such sort, but so far without success.
Some thoughts led me to the following measure of the resource' value:
$$V=\int_{t_0}^{t_1} e^{\int_0^t \log (p_1(u))du}\log \frac{p_1(t)}{p_0(t)} dt$$
where $p_0(t)$ is the probability density of survival without the resource, $p_1(t)$ is the probability density of survival with the resource and $(t_0,t_1)$ is the period of time through which the resource affects the probability of survival (the time for which the value is evaluated considered t=0)
I would like to know if there any similar measures already proposed and about the possible drawbacks of this proposed value. I also wonder to which extent this can be applied to real economics.
I'm not sure about your formula. Nor about your goal. For example, assume your char survives 80 years by default. But if your strategy is to search a dangerous jungle for the Fountain of YouthTM, you die at the age of 78 from exhaustion with probability 0.99, but you live 1000000 years if you succeed in finding the fountain (with probability 0.01).