As I was reading through Justin Abrahms' "Understanding the formal definition of Big-O" article to better understand Big-O notation I came across something at the end that caught my attention:
You might notice that this definition means that the $O(45n)$ function is also $O(n^3)$ because it will also never cross. This is true, but not particularly useful.
(emphasis mine)
Which fits the definition given in the Rosen's "Discrete Mathematics and its applications" book.
So then, at which point does Big-O becomes useful (or not)? It has do something to do with the witnesses you choose, right?
At the risk of being too vague...
In practice you often have a large number of error terms, like
$$ f(x) = g(x) + \epsilon_1(x) + \epsilon_2(x) + \cdots + \epsilon_n(x), $$
where $g(x)$ is a useful approximation to $f(x)$ and each $\epsilon_k(x)$ is small by comparison. Say you're interested in how good the approximation $f(x) \approx g(x)$ is.
Rather than spending lots of effort estimating each $\epsilon_k(x)$ precisely, you can quickly estimate them each roughly, and you'll often end up with different big-O bounds on each of them.
Note that it might be EXTREMELY difficult to estimate each of them precisely. Further, it might be the case that finding reasonable big-O bounds is orders of magnitude easier than finding their respective big-$\Theta$ classes. Another possibility is that making a distinction between $O(\cdots)$ and $\Theta(\cdots)$ just doesn't make a difference for your particular purpose.
Then you use the "not particularly useful" fact you're asking about to combine all of these big-O bounds into one big-O, the biggest one, and you end up with a simple expression like
$$f(x) = g(x) + O(h(x)).$$