Where do we place the decimal point due to a binary multiplication of two decimal binary numbers?

5.4k Views Asked by At

An example can be $$ 0010000.010 * 0001000.010 $$ which will give $$1000110.000100$$

But how was the operation done? More precisely, how was the decimal point placed there? Is the rule for placing the decimal point in the previous example exclusive for that example because the two numbers have equal number decimal places (3) ? or there is a general rule for placing the decimal point whether or not the number of decimal places are the same? (like will the rule you'll provide me still apply when multiplying 1011 * 0.010 ? Kindly place all zeros even the useless ones (after the last non-zero digit in the decimal part).

enter image description here

2

There are 2 best solutions below

2
On

Let's do decimal first. Say $1.5\times 0.24$. But I'm going to write it like this:

$$(1\times 10^0 + 5\times 10^{-1})\times (2\times 10^{-1}+4\times 10^{-2})$$ Those negative powers of $10$ are very annoying. So let's write it as $$10^{-1}(1\times 10^1+5\times 10^0)\times 10^{-2}(2\times 10^{1}+4\times 10^0)$$ But this is just $$10^{-3}(15\times 24)$$ If you know how to multiply integers, you're home free.

Nothing here depended on the base being $10$. For your problem, just replace the $10$ with a $2$.

0
On

Here's a simple explanation using a concrete example. If there are 3 decimal places in the first number, you are dealing with thousandths. If there are 7 decimal places in the second number, you are dealing with ten millionths. The product of thousandths and ten millionths gives ten billionths, which has 10 decimal places.

5.021 = $\frac {5021} {1,000}$, 3.3942037 = $\frac {33,942,037} {10,000,000}$

$5.021*3.3942037=\frac {5021} {1,000} * \frac {33,942,037} {10,000,000}$=$\frac {5,021*33,942,037} {10,000,000,000}$