I am reviewing the method of Lagrange multiplier and this time it strikes me as to why don't we just eliminate the multiplier $\lambda$ once and for all and just work with the remaining equations - since we are (mostly) only interested in locating the points at which extrema occur. I believe that, for most purposes, it is safe to assume that $\lambda$ does not vanish (see this for example), or even if we want to be safe, we only need to check that particular occurrence. So, instead of solving for $\lambda$ and plugging it back into the equations to compute the $x, y, z$, we could have eliminated the need to go through $\lambda$ to get $x, y, z$.
So, for instance, for a two-variable situation, we might want to recast the equations as
$\frac{f_x(x, y)}{f_y(x, y)}=\frac{g_x(x, y)}{g_y(x, y)}$.
I figure that there might be some difficulties with this approach since many sample solutions I see involve the computation of $\lambda$, but what are they?
Lagrange Multipliers say that, given that $f$ is a function you are trying to minimize/maximize, and $g$ is a constraint, you can use the clever $\nabla g=\lambda\nabla f$ to create a system of equations for $x$, $y$, $z$, and $\lambda$. With as many equations as variables, one can solve for $x$, $y$, and $z$, with the quickest ways to do so sometimes requiring solving for $\lambda$ first.
If I am understanding your question correctly, you are asking why we cannot just set $\nabla g=\nabla f$ (functionally having $\lambda=1$) and thus ignore the $\lambda$. The issue is that it misunderstands the motivation for the method of Lagrange Multipliers; the method does not work without the nonzero $\lambda$ scaling factor.
The method guarantees that $f$ is maximized/minimized where $\nabla f$ is parallel to $\nabla g$, since the smallest/largest level curve/surface of $f$ should just barely touch $g$, and so their gradients should be parallel. It does not guarantee that these gradients should be equal, and so we cannot assume the gradients have equal magnitude or direction. The $\lambda$ is necessary in that sense.