4

From the definition of a derivative, we have that $$f'(a) = \lim\limits_{x\to a}\frac{f(x)-f(a)}{x-a}$$

or $$\lim\limits_{x\to a}f'(x) = \lim\limits_{x\to a}\frac{f(x)-f(a)}{x-a}$$

This leads me to believe we can write $$\lim\limits_{x\to a}(f(a)+(x-a)f'(x))=\lim\limits_{x\to a}f(x)$$

Which to me, implies that derivative allow us to obtain a good approximation for a function near $a$, and that this approximation will become increasingly accurate as $x\to a$.

However, couldn't we just as easily write

$$\lim\limits_{x\to a}(f(a)+2(x-a)f'(x))=\lim\limits_{x\to a}f(x)$$

since $$\lim\limits_{x\to a}(f(a)+(x-a)f'(x))=f(a)+\lim\limits_{x\to a}((x-a)f'(x))$$

and $$\lim\limits_{x\to a}((x-a)f'(a)) = 0 = \lim\limits_{x\to a}(2(x-a)f'(a))$$ provided that $\lim\limits_{x\to a}f'(a)$ exists.

Wouldn't this imply instead that $\lim\limits_{x\to a}(f(a)+2(x-a)f'(x))$ can be used to provide good approximations for the function near $a$? In fact, couldn't we replace $2$ with any other constant or function which has a limit at $a$?

Using such approximations, however, would obviously produce incorrect results in proofs such as those for the chain and product rules. So how can this contradiction be dealt with? Why is using $f'(x)$ simply more correct than using $2f'(x)$, even when the math doesn't necessarily seem to be demonstrating this?

Golden_Ratio
  • 12,591
Ark1409
  • 123
  • From the definition of a derivative, we have that $$f'(a) = \lim\limits_{x\to a}\frac{f(x)-f(a)}{x-a}$$

    or $$\lim\limits_{x\to a}f'(x) = \lim\limits_{x\to a}\frac{f(x)-f(a)}{x-a}$$ I am unsure whether this is right. Let $~g(x)~$ denote $~f'(x).~$ I see 2 problems with your assertion. 1st, although I could be mistaken, just because $~g(x)~$ is well defined at $~(x = a),~$ does not mean that there is some small neighborhood around $~(x=a)~$ such that $~g(x)~$ is well defined there. $\color{red}{\text{I could be mistaken about this first objection}}$ ...see next comment

    – user2661923 Mar 18 '23 at 06:36
  • My second objection is similar, and (again), I could be mistaken. Even assuming that $~g(x) = f'(x)~$ is defined everywhere, that does not necessarily imply that $~g(x)~$ is $\color{red}{\text{continuous}}$ at $~(x = a)~$ which seems to be what you are assuming. – user2661923 Mar 18 '23 at 06:38
  • 1
    (1) Your second formula, $\lim\limits_{x\to a}f'(x) = \lim\limits_{x\to a}\frac{f(x)-f(a)}{x-a}$, is true if only if $f'$ is continuous at $a$. (2) Your formula for a linear approximation is incorrect: it should be $f(x) \approx f(a) + (x-a) f'(a)$, not $f(x) \approx f(a) + (x-a) f'(x)$. Note that if you use $f'(x)$ instead of $f'(a)$, then the approximation is not a priori a linear function of $x$. Not to mention, $f'$ might not be defined at points other than $a$. – Stef Mar 18 '23 at 14:12

3 Answers3

6

The point of the derivative as a linear aproximation of a function $f$ at a point $a$ is that the derivative is not just a linear aproximation, is the best linear aproximation, but, given a function $f : \mathbb{R}\rightarrow \mathbb{R}$ and $a\in \mathbb{R}$, intuitively, what means that a liner aproximation $L(x) = m(x-a) + f(a)$ of $f$ at $a$ is the best? Notice that is not enought if we say that $L$ is the best linear aproximation if: $\lim_{x\to a}L(x) = f(a)$, since every function of the form $m(x-a) + f(a)$ satisfies this, so what conditions are reasonable to say that $L$ is the best linear aproximation? That $L$ is the best linear aproximation means that you don't need to stay very close to $a$ to aproximate very accurately $f$, i.e. given a $\epsilon > 0$ (very very small), you can find a $\delta > 0$ (not too small) such that for $x\in \mathbb{R}$ with $|x-a|<\delta$ we get that $|L(x) - f(x)|<\epsilon$, in other words, $L(x)-f(x)$ goes to $0$ faster than $x-a$ does. This reasoning lead us to say that $L$ is the best linear aproximation if

$$\lim_{x\to a}\frac{f(x)-L(x)}{x-a} = 0$$

and this reasoning gives us the correct answer since if the derivative exist at a point, you can rewrite the limit

$$f'(a) = \lim_{x\to a} \frac{f(x)-f(a)}{x-a} $$

as

$$\lim_{x\to a}\frac{f(x)-[f(a) + (x-a)f'(a)]}{x-a} = 0$$

and is easy to prove the following:

Let $f : A\subseteq\mathbb{R}\rightarrow\mathbb{R}$ be a function and $a\in A'$. $f$ is differentiable at $a$ if and only if there exists $m\in\mathbb{R}$ such that $$\lim_{x\to a}\frac{f(x)-[f(a) + m(x-a)]}{x-a} = 0$$ in any case, we get that $f'(a) = m$.

Jorge S.
  • 782
  • "given a $\epsilon>0$ (very very small), you can find a $\delta>0$ (not too small) such that for $x\in\mathbb{R}$ with $|x−a|<\delta$ we get that $|L(x)−f(x)|<\epsilon$."

    I don't see why this property only holds for the derivative. If we choose $m=2f'(x)$, shouldn't it also be possible to show that we can get as close to $(a,f(a))$ as we want?

    – Ark1409 Mar 18 '23 at 04:54
  • Yes, we can, but given $L(x) = m(x-a) + f(a)$, the important point is not that for all $\epsilon >0$ we can find $\delta > 0$ such that for all $x$ with $|x-a| <\delta$ we get $|L(x) - f(x)| < \delta$, the important point is how much small is $\epsilon$ in comparation with $\delta$, as I said, when $\epsilon$ is very very small but $\delta$ is not too small this means that $L(x)-f(x)$ goes faster to $0$ than $x-a$ does, and the last theorem ensures that if $m\in\mathbb{R}$ satisfies this property, then necesarily $m$ is the derivative of $f$. – Jorge S. Mar 18 '23 at 05:51
  • @Ark1409 In other words, is not enought if we just say that $L$ is the best linear aproximation if we can get as close to $(a, f(a))$ as we want, since as I mentioned, every function of the form $m(x-a) + f(a)$ satisfies this, what makes a linear aproximation the best (intuitively) is that we don't need to be too close to $a$ to get very very close to $f(x)$ around $a$ and the last theorem says that only the derivative satisfies this poperty. – Jorge S. Mar 18 '23 at 06:02
4

Note that when $f$ is continuously differentiable at $a$, then $$\lim\limits_{x\to a}(f(a)+(x-a)f'(x))=\lim\limits_{x\to a}f(x)$$ is certainly true, but it is also true that

$$\lim\limits_{x\to a}(f(a)+g(x))=\lim\limits_{x\to a}f(x)$$

for any function $g$ such that $\lim_{x\to a}g(x)=0.$ Clearly, using the above limit equations as a basis for approximation is not going to get you anywhere meaningful.


Instead, without any limits, we can simply see from the definition of the derivative that when $f$ is differentiable at $a$, we can approximate the value of $f$ at $x$ as

$$f(x)\approx f(a)+(x-a)f'(a).\quad (1)$$

In fact, the expression in $(1)$ is the first order Taylor polynomial of $f$ around $a$. This may not be a "good" approximation; how "good" it is depends on your error tolerance, and there are known results on the bound of the error of such polynomials. See here.

Golden_Ratio
  • 12,591
  • "we can simply see from the definition of the derivative", wouldn't this sort of reasoning come from the limit definition though? How can we show that this approximation works well and that $cf'(x)$ will not? – Ark1409 Mar 18 '23 at 04:10
1

$$\lim\limits_{x\to a}f'(x) = \lim\limits_{x\to a}\frac{f(x)-f(a)}{x-a}$$

This claim assumes that $f'$ is necessarily continuous (what we might write as $f \in \mathcal{C}^1(D)$ for the appropriate domain $D$). This is not always the case; some examples are here, such as the function

$$f(x) := \begin{cases} x^2 \sin(1/x), & x \ne 0 \\ 0, & x = 0 \end{cases}$$


$$\lim\limits_{x\to a}(f(a)+(x-a)f'(x))=\lim\limits_{x\to a}f(x)$$

Claiming this, further, makes certain assumptions about limit laws that need not hold. In particular, more or less what you're doing is claiming

$$\lim_{x \to a} \frac{f(x)-f(a)}{x-a} = \frac{\displaystyle \lim_{x \to a} f(x)-f(a)}{\displaystyle \lim_{x \to a}x-a} \tag{1}$$

and then multiplying by the denominator to get

$$\left( \lim_{x \to a} f'(x) \right) \left( \lim_{x \to a} (x-a) \right) = \lim_{x \to a} \Big( f(x) - f(a) \Big) \tag{2}$$

and then splitting the latter limit to get

$$\left( \lim_{x \to a} f'(x) \right) \left( \lim_{x \to a} (x-a) \right) = - f(a) + \lim_{x \to a} f(x) \tag{3}$$

and then adding $f(a)$ to both sides,

$$f(a)+ \left( \lim_{x \to a} f'(x) \right) \left( \lim_{x \to a} (x-a) \right) = \lim_{x \to a} f(x) \tag{4}$$

and then recombining the left hand side all under the same limit:

$$ \lim_{x \to a} \Big( f(a) + f'(x) (x-a) \Big) = \lim_{x \to a} f(x) \tag{5}$$

Quite a few rules for limits were broken along the way.

  • In $(1)$, we can only claim that $$\lim_{x \to a} \frac{f(x)}{g(x)} = \frac{\displaystyle \lim_{x\to a} f(x)}{\displaystyle \lim_{x\to a} g(x)}$$ when all three limits involved exist (as finite numbers, to be clear: $\pm \infty$ is not a case where we say a limit exists here), and the bottom limit of the fraction is nonzero. A problem arises: the bottom limit goes to zero in $(1)$.

  • In $(2)$, you just multiplied both sides by $0$ as a result; $(2)$ is identical to saying $0=0$, so no actual information of note may be derived.

  • In $(3)$, I would be hesitant to say you can split the limit up like that. Like the above rule, for any $\alpha,\beta \in \mathbb{R}$, we can only say $$ \lim_{x \to a} \Big( \alpha f(x) + \beta g(x) \Big) = \alpha \lim_{x \to a} f(x) + \beta \lim_{x \to a} g(x)$$ when all three limits exist. Does the limit as $x \to a$ of $f(x)$ necessarily exist? One should do more work to justify this claim.

  • $(5)$ has many of the same concerns as all of the previous rules mentioned (aside from a product equivalent to the quotient law), just applied in reverse.


Granted, for a restricted class of functions, your equality does hold, and we have the more general observation mentioned in Golden_Ratio's answer that

$$\lim_{x \to a} \Big( f(a) + g(x) \Big) = \lim_{x \to a} f(x)$$

when $g(x) \xrightarrow{x\to a} 0$, so one begs the question: why is the formula

$$f'(x) \approx f(a) + f'(a) (x-a)$$

considered a good approximation?

One can certainly handwave some details. For $x$ very much near $a$, if

$$f'(a) = \lim_{x \to a} \frac{f(x) - f(a)}{x-a}$$

is expected to hold, then for $x$ very near $a$ (say, $x = a +\varepsilon$ for some $\varepsilon$ very very small with $|\varepsilon|>0$), the limit and derivative should be about the same. That is,

$$f'(a) \approx \frac{f(a+\varepsilon) - f(a)}{a +\varepsilon - a} = \frac{f(a+\varepsilon)-f(a)}{\varepsilon}$$

But then rearranging,

$$f(a+\varepsilon) \approx \varepsilon f'(a) + f(a)$$

and since $\varepsilon = x-a$, the desired approximation arises. Of course, when concerned with rigor, one should avoid such hand-waving.


A past MSE post somewhat handles why the tangent-line approximation is the "best" linear approximation (since your scalings of $f'(x)$ still render the approximation linear) -- it is the only linear function (in the sense of a form $y=mx+b$) whose error tends to $0$ faster than $x-a$ tends to $0$.

Also of interest may be Taylor's theorem, particularly results tied to the remainder term and bounds on it.

PrincessEev
  • 43,815
  • The lines tagged $(1)$ to $(2)$ are actually multiplication:

    $$\begin{align} \lim_{x\to a}(f(x)-f(a)) &= \lim_{x\to a}\left(\frac{f(x)-f(a)}{x-a}\cdot(x-a)\right)\ &= \left(\lim_{x\to a}\frac{f(x)-f(a)}{x-a}\right)\left(\lim_{x\to a}(x-a)\right)\ &= \left(\lim_{x\to a}\frac{f(x)-f(a)}{x-a}\right)\cdot 0\ \end{align}$$

    (for $\lim_{x\to a}\frac{f(x)-f(a)}{x-a}$ that exists). The end result at line $(2)$ is still $0=0$, but there's no invalid $\frac 00$ involved here.

    – peterwhy Mar 18 '23 at 04:41
  • Similarly, lines $(2)$ to $(4)$ are from adding $\lim_{x\to a} f(a)$ to both sides:

    $$\begin{align} \lim_{x\to a} f(x) &= \lim_{x\to a}\left((f(x)-f(a)) + f(a)\right)\ &= \lim_{x\to a}\left(f(x)-f(a)\right) + \lim_{x\to a} f(a)\ &= \lim_{x\to a}\left(f(x)-f(a)\right) + f(a)\ \end{align}$$

    Because both $\lim_{x\to a}\left(f(x)-f(a)\right)$ ($=0$) and $\lim_{x\to a} f(a)$ exist.

    – peterwhy Mar 18 '23 at 04:48
  • How can we show that the existence of the function $\epsilon(h)$ (representing the $\epsilon$ used above as error) is unique to $f'(x)$ (i.e. there exists no such function for $cf'(x)$)? – Ark1409 Mar 18 '23 at 05:10