From Cox, Little and O'Shea's book Ideals, Varieties and Algorithms. I really don't understand their proof on the following lemma about monomial ideals.
Let $I=\langle x^{\alpha}|\alpha \in A\rangle$ be a monomial ideal. Then a monomial $x^{\beta}$ lies in $I$ iff $x^{\beta}$ is divisible by $x^{\alpha}$ for some $\alpha \in A$.
Here's the proof.
If $x^{\beta}$ is a multiple of $x^{\alpha}$ for some $\alpha \in A$ then $x^{\beta} \in I$ by definition of an ideal. Conversely, $x^{\beta} \in I$ then $x^{\beta} =\sum_{i=1}^s h_ix^{\alpha (i)}$ where $h_i \in k[x_1,...,x_n]$ and $\alpha (i) \in A$. If we expand each $h_i$ as a linear combination of monomials we see that every term on the right side of the equation is divisible by some $x^{\alpha (i)}$. Hence the left side must have the same property.
Firstly, $x^{\beta}$ is a "monomial" right? If I understood monomials correctly, it should be $x_1^{\alpha _1}x_2^{\alpha _2}...$, so basically, it's a multiple of exponents of variables. However, $x^{\beta} =\sum_{i=1}^s h_ix^{\alpha (i)}$ seems to imply, since it's a "sum" notation, that the monomial $x^{\beta}$ is somehow a "linear combination" of monomials (or polynomials, but in the end, I guess it is essentially the same as linear combinations of monomials).
How? I can only think of expressing a monomial as a linear combination of monomials only with every single coefficient being $0$ except for the term where the exponent $n$-tuple ($\beta$ in this case) coincide. Then, why bother with the "sigma" notation? Isn't it redundant?
Second, what does it mean by "expand $h_i$ as a linear combination of monomials"? Does it literally want me to express each coefficient $h_i$ in terms of a liner combination of monomials? Even if so, how does this immediately tell me that "every term on the right side of the equation is divisible by some $x^{\alpha (i)}$?
I mean, so this proof is saying $x^{\beta}= h_1x^{\alpha (1)}+h_2x^{\alpha (2)}+...+h_sx^{\alpha (s)}$ and further, if I expand each $h_i$, namely, $$x^{\beta}= (h_{1,0}x_1^{\gamma_{1,0}}+h_{0,1}x_2^{\gamma_{1,1}}+...+h_{1,n}x_n^{\gamma {0,n}})x^{\alpha (1)}+(h_{2,0}x_1^{\gamma_{2,0}}+h_{2,1}x_2^{\gamma_{2,1}}+...+h_{2,n}x_n^{\gamma {2,n}})x^{\alpha (2)}+...+(h_{s,0}x_1^{\gamma_{s,0}}+h_{s,1}x_2^{\gamma_{s,1}}+...+h_{s,n}x_n^{\gamma {s,n}})x^{\alpha (s)}$$ where $h_i=h_{i,0}x_1^{\gamma_{i,0}}+h_{i,1}x_2^{\gamma_{i,1}}+...+h_{i,n}x_n^{\gamma {i,n}} \in k[x_1,...,x_n]$. I mean, this expanding business, to me, seems like it's just made it super complicated. How is this supposed to make me "see" that it is divisible by some $x^{\alpha (i)}$? I really can't see anything even close to it.
What's going on with this proof? Can somebody please please help me? Thanks...
(A bit late, anyway ...) To prove the claim, the implication needs to be shown both ways:
If $x^\beta$ is a multiple of $x^\alpha$ for some $\alpha \in A$, then $x^\beta \in I$, and,
If $x^\beta \in I$, then it is a multiple of $x^\alpha$ for some $\alpha \in A$.
I assume that you understand how to prove the first part. Then, for the second one, since $x^\beta \in I$, $x$ is a linear combination of multiples of the generators of $I$, hence the notation,
$$ x^\beta = \sum_i h_i x^{\alpha(i)} $$
The left hand side is a monomial, but we can not assume that all but one of the $h_i$ must be zero on the right hand side. (If we assumed that, then we are already assuming that $x^\beta$ is a multiple of some $x^\alpha$, the statement we need to prove!) For example, if $n=2$ and $x^2y$, $x^1y^2$ are three of the many generators of $I$, and $\beta = (4,2)$, then the following is a valid way of expanding $x^\beta$.
$$ x^\beta = (x^2y+xy)x^2y + (-x^2)x^1y^2 $$
This is why they use that notation.