Sufficient statistics for $\lambda$ poisson distribution.

8.1k Views Asked by At

I have the following question out of a book. I even have the solution from solutions manual that I cannot really follow either, so I thought I would ask here to see if someone could dumb it down for me.

Let $Y_!, Y_2, ..., Y_n$ denote a random sample from a Poisson distribution with parameter $\lambda$. Show by conditioning that $\sum_{i=1}^{n}Y_i$ is sufficient for $\lambda$

So that is the question. From the solution I can see that they are using the conditional probability, but from the chapter that this question is from you are taught that you show sufficiency by using the likelihood. So what does it mean by "show by conditioning"? What does the act of conditioning do in this instance? Why would we use conditioning instead of likelihood? Also, just how would you do this question?

2

There are 2 best solutions below

2
On BEST ANSWER

We have that \begin{equation} f(y_k,\lambda) = \frac{\lambda^{y_k}e^{-\lambda}}{y_k!} \end{equation} So the joint distribution of all your samples is \begin{equation} f(y_1 \ldots y_n,\lambda) = \prod_{i=1}^n \frac{\lambda^{y_k}e^{-\lambda}}{y_k!} \end{equation} which is, by collecting similar terms, \begin{equation} f(y_1 \ldots y_n,\lambda) = \frac{\lambda^{\sum_{k=1}^n y_k}e^{-n\lambda}}{y_1!y_2! \ldots y_n!} \end{equation} Now conditioning on the statistics $T = \sum Y_i$, we have \begin{equation} f(y_1 \ldots y_n,\lambda,T=t) = \frac{\lambda^t e^{-n\lambda}}{y_1!y_2! \ldots y_n!} \end{equation} Furthermore, we know that the sum of Poisson distributions with parameter $\lambda$ , is a Poisson distribution with parameter $n\lambda$, i.e. \begin{equation} g(t,\lambda) = \frac{(n\lambda)^t}{t!}e^{-n\lambda} \end{equation} We can say that the conditional distribution of the sample is \begin{equation} f(y_1 \ldots y_n \mid T=t, \lambda) = \frac{f(y_1 \ldots y_n,\lambda,T=t)}{g(t,\lambda)} = \frac{ \frac{\lambda^t e^{-n\lambda}}{y_1!y_2! \ldots y_n!}}{\frac{(n\lambda)^t}{t!}e^{-n\lambda}} = \frac{t!}{n^t y_1! \ldots y_n!} \end{equation} which is independent of $\lambda$. So, we say $T$ is a sufficient statistic.

2
On

"Why would we use conditioning instead of likelihood?"

The reason to use conditioning is that the definition of sufficiency uses conditioning. To say that $Y_1+\cdots+Y_n$ is sufficient for $\lambda$ means that the conditional distribution of $(Y_1,\ldots,Y_n)$ given $Y_1+\cdots+Y_n$ does not depend on $\lambda.$ Therefore, in order to show that $Y_1+\cdots+Y_n$ is sufficient for $\lambda,$ one finds the conditional distribution of $(Y_1,\ldots,Y_n)$ given $Y_1+\cdots+Y_n.$ \begin{align} & \Pr(Y_1=y_n\ \&\ \cdots\ \&\ Y_n=y_n \mid Y_1+\cdots+Y_n=y) \\[10pt] = {} & \frac{\Pr(Y_1=y_1\ \&\ \cdots\ \&\ Y_n=y_n\ \&\ Y_1+\cdots+Y_n=y)}{\Pr(Y_1+\cdots+Y_n=y)}. \tag 0 \end{align}

To evaluate this, first note that \begin{align} & \text{the event } \big[Y_1=y_1\ \&\ \cdots\ \&\ Y_n=y_n\ \&\ Y_1+\cdots+Y_n=y\big] \\[10pt] & \text{is the same as the event } \big[Y_1=y_1\ \&\ \cdots\ \&\ Y_n=y_n\big] \tag 1 \end{align} and $$ Y_1+\cdots+Y_n \sim \operatorname{Poisson}(n\lambda). \tag 2 $$

Use $(1)$ to simplify the numerator in $(0).$

Use $(2)$ to evaluate the denominator in $(0).$

Then observe that when you're done simplifying, no $\text{“}\lambda\text{''}$ appears in the bottom line.