What is the maximum entropy of a softmax output layer of a neural network?

100 Views Asked by At

Neural networks oftentimes apply the softmax operation to their output neurons which is defined by

$$s(x)_i = \frac{e^{x_i}}{\sum_{j=1}^N e^{x_j}}$$

Here, $i$ is the index of an output neuron, and $N$ represents the total number of output neurons. That means that the sum of all output activations sum up to 1 and that the value of each output neuron lies between 0 and 1:

$$\sum_{i=1}^N s(x)_i = 1$$ $$0 \leq s(x)_i \leq 1$$

I would like to know if it is possible to calculate an upper bound of the output layer's entropy? Or do I first have to discretize the network's output in order to derive an upper bound for the entropy?