For $r\ge -1$, the exponential of the negative Renyi entropy is defined as $$M(p):=\Big(\sum_i p_i^{1+r}\Big)^{\frac1r},$$ for a probability measure as tuples $p:=(p_i)_i$ I would like to prove the convexity of $M(\cdot)$, or $$M(ap+bq)\le aM(p)+bM(q),$$ $\forall\,a+b=1 \wedge a,b\ge0$, and two probability measures $p$ and $q$ with the same cardinalities.
For $r>0$, I can show the convexity via the Minkowski inequality for $\big(\sum_i x_i^{1+r}\big)^{\frac1{1+r}}$ then the convexity of $f(x):=x^{1+\frac1r}$.
But how would one show the convexity for $-1<r<0$? The above technique does not work since the inequality signs from the two steps point in the opposite directions.
Let $h(t) = t^{1/r}$ and $g(x) = \sum_{i=1}^n x_i^{1+r}$, $0< r < -1$.
We have that $h'(t) = \frac{t^{1/r-1}}{r} \leq 0$ for all $0 \leq t$, so $h$ is nonincreasing on $\mathbb{R}_+$, and $h''(t) = (1/r-1)\frac{t^{1/r-2}}{r} \geq 0$ for all $0\leq t$, so $h$ is convex.
Similarly, $u(t) = t^{1+r}$ is concave on $0 \leq t$, since $u''(t) = r(1+r)t^{r-1} \leq 0$, which implies that $g(x)$ is concave (sum of component-wise concave functions).
Now, $g(\theta x + (1-\theta)y)\geq \theta g(x) + (1-\theta)g(y),$ so $h(g(\theta x + (1-\theta)y)) \leq h(\theta g(x) + (1-\theta)g(y)) \leq \theta h(g(x)) + (1-\theta)h(g(y))$.
Therefore $f=h\circ g$ is convex, as desired.