Let $X$ be a random variable defined on a compact set $K\subset \mathbb{R}$. The moment generating function (MGF) of $X$, denoted as $M_X(t), t\in \mathbb{R}$, is defined as $$M_X(t) = \mathrm{E} [e^{tX}] = \int_K e^{tX}d\mathbb{F}(x).$$
From the Wikipedia page, one can compute the expansion of
$\begin{align*} M_X(t) = \mathrm{E} [e^{tX}] &= 1 + tE(X) + \frac{t^2 E(X^2)}{2!} + \cdots \\ & = 1 + tm_1 + \frac{t^2 m_2}{2!} + \cdots, \end{align*}$
where $m_k = E(X^k)$ is the $k$-th moment of $X$. However, I have a few questions regarding MGFs and using MGF to determine $X$.
What is the region of convergence (ROC) of the above Taylor expansion? I suppose that this is related to $E(X^k), k = 1, 2, \ldots$. Are there sufficient and necessary conditions on $\{E(X^k)\}_{k = 1}^\infty$ so that the ROC of $M_X(t)$ has a positive radius?
I remember a statement that `an MGF uniquely determines a random variable'. My question is that, when we say two MGFs equal, do we automatically imply that the ROCs of these two MGFs are the same? Is it possible to have two MGFs agree on an interval, but with different ROC?
Let us consider another random variable $Y$ as a function of $X$. Then the expectation of $Y^k$ is computed as $\mathrm{E}(Y^k) = \int_K Y^k(X) d\mathbb{F}(X)$. The MGF of $Y$, denoted as $M_Y(t)$, can be computed by $$\begin{align*} M_Y(t) = \mathrm{E} [e^{tY}] &= 1 + tE(Y) + \frac{t^2 E(Y^2)}{2!} + \cdots. \end{align*}$$ From my understanding, the '$k$-th moment' of $Y$ should be $Y^k$ integrated with respect to the distribution of $Y$. So I don't see any reason for $\mathrm{E}(Y^k)$ to be called as the `$k$-th moment' of $Y$ in this case. Then what is $\mathrm{E}(Y^k)$ called? Is $M_Y(t)$ still called the 'moment generating function'? Does $M_Y(t)$ still uniquely determine $Y$?
As a preamble to the following answers, I note that if the moment generating function of a random variable has a positive radius of convergence, it uniquely detetmines the distribution of that random variable, but not the random variable itself, because there are always many different random variables with any given distribution.