So now that I have moved on to differentiation in multiple dimensions and a more general treatment, I recognize a derivative (let us treat the 1-d case for now, it is easy to generalize in more than 1 with norms and such) as an operator $D$ that takes $f \to f'$and $f'$ although not (necessarily) a linear function itself, takes points in the domain of $f$ and sends them to scalars $f'(x)$ which we say is the derivative. The mapping $g(h) = f'(x)\cdot h$ then satisfies:
$$f(x+h) = f(x) + f'(x) \cdot h + \epsilon_x(h)$$ where the error term $\epsilon_x(h)$ satisfies: $$ \frac{\epsilon_x(h)}{h} \to 0 $$ as $h \to 0$. Now, it is easy to see why the error term behaves this way by the limit definition of the derivative. My question here however, is why (intuitionally) does the quotient of the error term (rather than the error term itself) need to approach $0$? The more I think about it, the more I realize that the error term approaching $0$ itself is a trivial consequence of continuity, i.e the fact that $\lim_{h \to 0} f(x + h) = f(x) = \lim_{h\to 0}(f'(x)\cdot h + f(x))$. I guess my question is, why does the error (quotient) approaching $0$ give us a "good enough" or satisfactory linear approximation of $f$?