If we model $Y = \beta X$, we can estimate $\beta$ to minimize
$$\sum (Y_i - \beta X_i)^2$$
Taking derivatives and solving for 0, we get $\sum 2\beta X_i^2 - 2Y_1X_i = 0 \implies \beta = \frac{\sum Y_i X_i}{\sum X_i^2}.$
Why does our best fit for $Y = \beta X$ not even satisfy that it has the same variance?

Depending on the measured data -- for example it forms a "dot cloud" with a distinct distance to the origin -- it might be reasonable to set as balance line the line from the origin through the dot cloud's barycentre $(\bar{x}, \bar{y})$ instead of a "best fit" according the LS method using vertical distances of the observed points to the resulting line as target figure.
The slope of this line through the origin is $\displaystyle\beta=\frac {\sum_k y_k}{\sum_k x_k}=\frac{\bar{y}}{\bar{x}}$.
(This conforms to what I remember from studies decades ago.)