I'm reading about variational inference and expectation maximization, I often see this assertion without proof or explanation:
We have no problem with maximizing sum of logs as in
$\max_\theta \sum{\log{p(x_i, z_i)}}$
But we do have trouble with maximizing log of sum, as when $p(x_i,z_i)$ is marginalized to $p(x_i)$ using law of total probability, and as such you get a sum inside the log. $\max_\theta \sum{\log{\sum_z{p(x_i, z)}}}$
Why is that?
A sum of logs is easily simplified as the log of the product, i.e.
$$\sum_{i=1}^n \log a_n = \log\left(\prod_{i=1}^n a_n\right).$$
There is no such conversion for the log of a sum.