I know $\text{posterior} = \frac{\text{proir} \cdot \text{likelyhood}}{\text{evidence}}$, but how it get $\alpha +\sum _{i=1}^{n}x_{i},\,\beta +n-\sum _{i=1}^{n}x_{i} $ from that formula and what does sum of $x_{i}$ means? Given Bernoulli is :
$$ p(x,\mu) = \mu^x(1-\mu)^{1-x} x \in \{0,1\} $$
Can some one give me some intuition about the posterior hyperparameters? e.g. explain how those term could help me get something from $\text{posterior} = \frac{\text{proir} \cdot \text{likelyhood}}{\text{evidence}}$.
As the link explains and I quote:
One important assumption left out is that these $x_{i}$ are independent observations, although this is as far as I have read the norm anyways.
To be explicit, each $x_{i}$ is an observation of a Bernoulli experiment with parameter $\mu$; $$p(x_{i} \mid \mu) = \mu^{x_{i}}(1-\mu)^{1-x_{i}}.$$
Let me denote $\mathcal{D} = \{x_{1}, \dots, x_{n}\}$. To get the hyperparameters for the posterior it is easiest to use the unnormalized Bayes theorem:
$$p(\mu\mid \mathcal{D})\propto p(\mathcal{D}|\mu)p(\mu).$$ Since the data in $\mathcal{D}$ is (implicitly) assumed to be independent: $$p(\mathcal{D}\mid \mu) = \prod_{i=1}^{n}p(x_{i}\mid \mu) = \prod_{i=1}^{n} \mu^{x_{i}}(1-\mu)^{1-x_{i}} = \mu^{\sum_{i=1}^{n}x_{i}}(1-\mu)^{n - \sum_{i=1}^{n}x_{i}}.$$ Which therefore implies, since $\mu\sim\text{Beta}(\alpha,\beta)$ and hence $p(\mu) \propto \mu^{\alpha-1}(1-\mu)^{\beta-1}$, that the posterior is:
$$p(\mu\mid \mathcal{D})\propto \mu^{\sum_{i=1}^{n}x_{i} + \alpha - 1}(1-\mu)^{n - \sum_{i=1}^{n}x_{i} + \beta - 1}.$$
To find the normalizing constant, you could either integrate, but more simply note that by comparing coefficients that $p(\mu\mid \mathcal{D})$ is proportional to the density of a $\text{Beta}(\alpha + \sum_{i=1}^{n} x_{i}, n + \beta - \sum_{i=1}^{n} x_{i})$ and hence the constant is the one from this Beta distribution.