I was confronted with the next problem and a little help would be welcomed. This is what I have so far.
Problem
Let $X_1, X_2,...,X_n$ be a random sample from
$f(x;\theta)=e^{-(x-\theta)}I_{[\theta,\infty)}(x)$, for $-\infty<\theta<\infty$.
(a) Find a sufficient statistic.
(b) Find a maximum-likelihood estimator of $\theta$.
(c) Find a method-of-moments estimator of $\theta$.
(d) Is there a complete sufficient statistic? If so, find it.
(e) Find the UMVUE of $\theta$ if one exists.
Attempt of solution
Let $I=\{1,2,...,n\}$
(a)
Let $Y=min\{X_i:i \in I\}$.
Clearly, if $Y \geq \theta$, $f(x;y)=\prod_{i \in I} f(x_i;y)$ is independent of $\theta$. Therefore, $Y=min\{X_i:i \in I\}$ must be a sufficient statistic.
(b)
Analogously, the likelihood function is maximized when $min\{X_i:i \in I\}=Y \geq \theta$ (It is $0$ otherwise), so the MLE of $\theta$ must be Y, as above defined.
I need help with (c)
(d)
Since this density function is not in the exponential family given the Indicator function, it's needed a different approach.
Integrating over $[\theta,x)$ the density function $f(x;\theta)$, the distribution function $F(x;\theta)=(\frac{-e^\theta}{e^x}-1)I_{[\theta,\infty)}(x)$ is obtained. And since the event $[Y<y]$ is the intersection of $[X_i \geq y]$ for $i \in I$ events (meaning $1-P[X_i < y]$), the distribution function of Y is given by $F(y;\theta)=(\frac{e^\theta}{e^x})^nI_{[\theta,\infty)}(x)$ and $Y$ has density
$f(y;\theta)=n(\frac{e^\theta}{e^y})^{n-1}I_{[\theta,\infty)}(x)$
Now, let g be a measurable function such that $E[g(Y)]=0, \enspace \forall \theta \in \mathbb{R}-\{0\}$. ($\theta=0$ trivializes the exercise)
Observe that,
if $\forall \theta \in \mathbb{R}-\{0\}$,
$0=E[g(Y)]=n\int_{\theta}^{\infty}g(y)[\frac{e^\theta}{e^y}]^{n-1}dy$. Then, since $n \in \mathbb{N}$, and $\frac{e^\theta}{e^y} \neq 0, \enspace \forall y \in \mathbb{R}$.
Then $g(Y) \equiv 0$.
Therefore, $Y=min\{X_i:i \in I\}$ is complete, and given (a), it is sufficient as well.
(e)
In this step we should search for $T=t(Y)$, an unbiased estimator of $\theta$.
Let's then consider that $E[Y]=\frac{n\theta}{n-1}+\frac{n}{(n-1)^2}$.
I am not sure about this, but does the first member of the result imply that such UMVUE doesn't exist? i.e., is possible to make $\frac{n\theta}{n-1}$ "become" $\theta$?
Note:
I am aware of the length of this exercise, and aware -as well- of the possible existence of mistakes here. Corrections are welcomed in such case.
Thanks in advance.
Your solutions to parts a), b) and d) are correct.
For c), use the first moment based estimator. For the given distribution, $$\mathbb{E}(X_1)=\int_\theta^\infty xe^{\theta-x}dx=\int_\theta^\infty (x-\theta+\theta)e^{\theta-x}dx=\int_0^\infty (t+\theta)e^{-t}dt=\Gamma(2)+\theta=\theta+1,$$
where we use the substitution $t=x-\theta$, and the fact that $\Gamma(1)=\Gamma(2)=1.$ This means
$$\mathbb{E}(\bar{X})=\theta+1.$$
Naturally the method of moments estimator will be $\bar{X}-1.$
e) We need $t(Y)$ to be unbiased for $\theta.$ It can be checked that $\mathbb{E}(Y)=\theta+\dfrac{1}{n}.$
We can then take $t(Y)=Y-\dfrac{1}{n}.$ This is of course unbiased and is a function of the complete sufficient statistic $Y,$ and hence must be the UMVUE.