4

if $y$ is a function of $x$, why is $\frac {dy}{dx} dx = dy$? I have not learnt real analysis, but have done a bunch of math method courses at university and this has been bugging me. Why is it you can treat that as a fraction?

I would like the traditional calculus view first, then maybe differential forms view later. $dx$ to me is $\lim_{\Delta x \to 0} \Delta x$ and $\frac {dy}{dx} = \lim_{h\to 0} \frac {y(x+h)-y(x)}{h}$

I think $dy$ would be $\lim_{h\to 0} {y(x+h)-y(x)}$ but I'm not sure.

I don't see how you can just play around with these limits as if they were fractions. What proof is there that you can do so?

  • Well, think about what $\int y'(x) \mathrm{d}x$ is equal to. It's really just formal a formal expression at the intro calc level though. – Tyler Nov 22 '14 at 19:24
  • What does $dx$ even mean to you? Once you get to more formal definitions, it is basically the definition. – Thomas Andrews Nov 22 '14 at 19:25
  • That equation can be thought of as defining dy as a function of x and dx (so dx is an independent variable). For example, if $y = x^2$, then $dy = 2x dx$. – BigL Nov 22 '14 at 19:26
  • @BigL That's actually silly, since $x$ is the indepdent variable, not $dx$. What do you think $dy$ means? – Thomas Andrews Nov 22 '14 at 19:26
  • The answer will depend on your definition of $dy$ and $dx$. For the differential forms interpretation, see this post from the Math.SE blog. –  Nov 22 '14 at 19:27
  • @Thomas It is a quite standard definition. See for example http://en.wikipedia.org/wiki/Differential_of_a_function – BigL Nov 22 '14 at 19:31
  • 3
    Your view of $dx$ can't be useful; all you've said is that, to you, $dx$ is $0$. (actually, you probably have the wrong idea about limits, if you're imagining you said something different than $0$) –  Nov 22 '14 at 19:39
  • well that's how Feynman and Newton taught it. What's the correct view of dx? – minusatwelfth Nov 22 '14 at 19:41
  • 1
    @user1564795 dx is not 0, but it goes to 0: $dx\to 0$ It represents a very, very small change of x. – callculus42 Nov 22 '14 at 19:44
  • that doesn't tell me what $dx$ is though. How is it defined? – minusatwelfth Nov 22 '14 at 19:47
  • @user1564795 I honestly never understood what my calculus professors were talking about when they introduced differentials. I learned the formula for total differential and how to use it, but not really what it meant. That is until I started studying differential geometry. The differential forms view makes so much sense. I might not recommend learning it to someone who had just started learning calculus, but since you knew about path integrals months ago, you're probably fine to start learning differential forms. As a first, brief source, see that blog post that I linked to. –  Nov 22 '14 at 19:48
  • I'll look into that post, but it looks slightly above my level. I would still like the infinitesimal explanation though as it's still taught – minusatwelfth Nov 22 '14 at 19:58
  • I want the explanation that doesn't involve differential forms, whatever that may be. Isn't that what real analysis is about? – minusatwelfth Nov 22 '14 at 20:46
  • @Bye_World interesting. Arturo Magidin says that it's an 'abuse of notation', but it can be shown that this works most of the time under 'mild assumptions'. This is sort of what I was expecting, but I would like to see the argument that proves this. – minusatwelfth Nov 22 '14 at 21:01

2 Answers2

6

The way that differentials $dx, dy, $ etc are usually defined in elementary calculus is as follows:

Consider a $C_1$ function $f: \Bbb R \to \Bbb R$.

We define the tangent to the graph of $f(x)$ at the point $(c,f(c))$ to be $y=f'(c)(x-c) + f(c)$.

Here the value $x-c$ is called the increment of $x$, and is denoted $\Delta x$. So our tangent function can be written as $y=f'(c)\Delta x + f(c)$. Note that $\Delta x$ is actually evaluated at a specific value $c$ and thus should possibly be more correctly written as $\Delta x(c)$ or $\Delta x|_c$, but that'll get tedious so I'll leave off the explicit $c$ dependence.

Because the increment of $x$ can be just as easily defined on the graph of $f$ as on its tangent line $y$, we can also consider $\Delta x$ to be the differential of $x$, denoted $dx$. That is, we'll define $dx := \Delta x$. You can see that we could define all of the independent variables of a function in the same way for a function of multiple variables -- so for $g(x,y)$, we'd have $dx=\Delta x = x-c_1$ and $dy=\Delta y = y-c_2$.

The increment of the function $f$ is defined to be $\Delta f := f(c+\Delta x) - f(c)$. This is a measure of how much $f$ has changed given a change in $x$.

One of the fundamental ideas of calculus is that the increment $\Delta f$ should be approximately the same as the increment of its tangent $\Delta y$ for sufficiently small changes of $x$. From the above definition of the increment of a function, we can see that $\Delta y = y(c+\Delta x) - y(c) = [f'(c)((c+\Delta x) -c) + f(c)] - [f'(c)(c-c)+f(c)]$ $=f'(c)\Delta x + f(c) -0 -f(c) = f'(c)\Delta x$. Thus $\Delta y = f'(c)\Delta x$.

We define the differential of the function $f$ at $c$ to be equal to the increment of its tangent function. So $df|_c := \Delta y|_c$, or, dropping the $c$ dependence from our notation, $df:=\Delta y=f'(c)\Delta x$.

Notice that with these definitions, there is a fundamental difference between the differentials of independent variables -- like $dx$ -- and functions of those variables -- like $df$. Differentials of an independent variable is just the increment of that variable whereas the differential of a function is defined to be the increment of the tangent to that function, i.e. $df=\Delta y=f'(c)\Delta x=f'(c)dx$.

Also you should notice, that we define differentials (at least the differentials of functions) in terms of the derivative, not the other way around like early developers of calculus did. So we assume that a definition of the derivative has already been devised and agreed upon and we can then define our differentials in terms of that derivative.

NOTE: It is important that you realize that $\frac {df}{dx}$ is just a suggestive notation for $f'(x)$. It is NOT a fraction of differentials. In fact the only reason we use it is because it helps students learn things like the chain rule. But the derivative is actually defined by a limit, NOT by a ratio of infinitesimals (unless you're learning non-standard analysis).

So in response to your question: using the very first definitions given to students in elementary calculus. $df$ is literally defined by $df=f'(x)dx = \frac {df}{dx}dx$.

Now what are differentials used for? Differentials are an approximation of the change of a function for a very small change in the independent variable(/s). That is, $df \approx \Delta f$ when $\Delta x$ is sufficiently small.

In fact, the smaller that we make $\Delta x$, the smaller the difference between $df$ and $\Delta f$ becomes. The mathematical statement of this is that $\lim_{\Delta x \to 0} \frac {df}{\Delta f} = 1$. So $df$ and $\Delta f$ are "the same" for very, very small values of $\Delta x$.

Because OP seems to think that calculus is based on infinitesimals, I will try to elucidate this as fallacy. Notice that I did not use infinitesimals anywhere in this. $dx=x-c$ is finite. $df=f'(c)dx=f'(c)(x-c)$ is finite. There are no infinitesimals in this answer. And more generally in calculus, we don't use infinitesimals. Your first course in calculus you may have talked about them, but they are just meant to be a heuristic. No definitions actually require them.

One more thing to note: if you're familiar with Taylor's theorem, then let's look at the Taylor expansion of a $C_k$ function $f$ near the point $(c,f(c))$. We have $f(x) = f(c) + f'(c)(x-c) + \frac {f''(c)}{2!}(x-c)^2 + \cdots + \frac {f^{(k)}(c)}{k!}(x-c)^k + R$, where the remainder, $R$, is very small. Then you can see that the tangent function to the graph of $f$ is just the $1$st order Taylor expansion of $f$. And the differential $df$ is just the second term of Taylor expansion.

NOTE: IMO, a better, more useful definition of differentials is as differential forms. For a good reference, pick up the book Advanced Calculus: A Differential Forms Approach by Harold M. Edwards.

  • You have shown that $\Delta f = \Delta y = f'(c)(x-c)=f'(c)\Delta x$ However $\Delta f$ could be large. For example $f'(x-c)$ was incredibly large, enough to to overpower $\Delta x$. So I think $df =/= \Delta f$ as $df$ should also be infinitesimal – minusatwelfth Nov 22 '14 at 21:40
  • Let me correct my above sentence.I meant: "For example if $f′(c)$ was incredibly large, enough to overpower Δx." – minusatwelfth Nov 22 '14 at 21:49
  • The condition that $\Delta x$ is very small isn't enough to ensure $df$ is very small. It's a hole in your argument that the definition of a derivative in my OP didn't have. You're basically saying $df=\Delta y = any size$. – minusatwelfth Nov 22 '14 at 21:59
  • Ok after some thought I agree with your proof. However one thing still bugs me a little. Doesn't this mean dx has to be infinitely small to get perfect accuracy? In which case the independent variables are always infinitesimals and the dependent is not. Right? – minusatwelfth Nov 22 '14 at 22:18
0

If you pick two points on the function, let's say $x_1$ and $x_2$, the change in between is $\Delta x = x_2-x_1$. If you insert $x_1$ into the function $f(x)$, you will get $f(x_1)=y_1$. If you insert $f(x_2)$, you will get $y_2$. The change in between is $y_2-y_1=\Delta y$. The slope can be calculated with the formula: $\frac{\mathrm{change\:in\:}y}{\mathrm{change\:in\:}x}$, which is the same as $\frac{\Delta y}{\Delta x}$. Let's say that in the figure below $\frac{\Delta y}{\Delta x}=2$. If you go $1$ to the right, the $\Delta x=1$. Plugging that into the formula gives you $\frac{\Delta y}{1}=2 \rightarrow \Delta y=2$. So every step you go to the right, you will go two steps up.

enter image description here

To get the slope as accurate as possible, we need to make the change of $x$ as small as possible. Let's say that $x_2=x_1+h$. To make the change as small as possible, we need to make $h$ as small as possible.

The change of $x$ is $\Delta x$, which is $h$. So $\Delta x = h$. Now we need to get $\Delta y$, which is $y_2-y_1$. We know that $y_2=f(x_2)$ and $x_2=x+h$. So $y_2=f(x+h)$. We also know that $y_1 = f(x)$. $\Delta y = y_2-y_1 = f(x+h)-f(x)$. Plugging these in the formula $\frac{\Delta y}{\Delta x}$ makes:

$$\frac{\Delta y}{\Delta x}=\frac{f(x+h)-f(x)}{h}$$

Now the problem is that $h\ne0$, because division by $0$ is undefined. So we need to make $h$ as small as possible, approaching $0$. So:

$$\frac{\Delta y}{\Delta x}=\lim_{h\to0}\frac{f(x+h)-f(x)}{h}$$

Adnan
  • 1,021
  • You mean dy/dx equals that limit, not change in y/ change in x. Also, this doesn't answer my question at all. Thanks for the effort. – minusatwelfth Nov 22 '14 at 20:35