2

I am interested in the continued fraction

$$1 + \dfrac{1}{1 + \dfrac{1}{1 + ...}}$$

You can solve this by letting $$y = 1 + \dfrac{1}{1 + \dfrac{1}{1 + ...}}$$

Then since it is infinite

$$y = 1 + \dfrac{1}{y}$$

As such

$$y^2 - y - 1 = 0 \\ y = \dfrac{1 \pm \sqrt{5}}{2} $$

So I was initially inclined to discard the negative solution since $\dfrac{1 - \sqrt{5}}{2}$ is a negative number, and it wouldn't make sense for a bunch of additions and divisions to result in a negative number. However, then I realised that we are dealing with infinity here and so I am now intrigued by the negative solution and what it might mean.

Could anyone help clear up the significance of this and whether or not it should be discarded, and why?

Thanks.

J P
  • 121
  • 5
    Notice that we first need to establish convergence of your sequence of fractions (which is not too hard). After we have done that, we notice that the limit of those fractions must satisfy $y=1+\frac1y$. In other words, we have $$y=1 + \dfrac{1}{1 + \dfrac{1}{1 + ...}}\implies y=1+\frac1y,$$ but the other direction of the implication is not true! – Maximilian Janisch Feb 22 '20 at 00:03

1 Answers1

4

Introduction:

So this ties a lot into how we define many "infinite processes." Some of the more familiar ones are these:

  • Infinite sums, such as

$$\sum_{k=0}^\infty x^k = \frac{1}{1-x} \; \text{whenever} \; |x| <1$$

  • Infinite products, such as

$$\prod_{p \; prime} \frac{1}{1-p^{-2}} = \frac{\pi^2}{6}$$

  • Infinitely-nested radicals, such as

$$\sqrt{1+2\sqrt{1+3\sqrt{1+4\sqrt{1+\cdots}}}}=3$$

  • And, of course, continued (infinite) fractions as well...

Defining Infinite Processes: Convergents:

Basically, we find a "convergent" of sorts - a "partial" and finite version of the infinite thing - and find the limit as we take infinitely many. What might these convergents look like? Usually, it depends on the context. For instance, for sums and products, we just take finitely many terms. For radicals and fractions, we truncate the expression and find the limit as we truncate further down the line.

Thus, for instance,

$$\sum_{k=0}^\infty x^k = \lim_{n \to \infty} \sum_{k=0}^n x^k \;\;\;\;\; \text{and} \;\;\;\;\; \prod_{p \; prime} \frac{1}{1-p^{-2}} = \lim_{n \to \infty} \prod_{\text{the first n primes}} \frac{1}{1-p^{-2}}$$

For radicals and fractions, it's easier to think of them in terms of a sequence, in which we continue adding one more term at each step. For instance, for the radical expression above,

$$\sqrt 1 \;\;\; , \;\;\; \sqrt{1 + \sqrt 2} \;\;\; , \;\;\; \sqrt{1 + \sqrt{2 + \sqrt 3}} \;\;\; , \;\;\; \sqrt{1 + \sqrt{2 + \sqrt{3 + \sqrt 4}}} \;\;\; , \;\;\; \cdots$$

...in which the obvious pattern holds to get the infinite radical from before, is a sequence which approaches $3$.

Similarly, for an infinite nested fraction, we can truncate before each plus or minus sign:

$$1 \;\;\; , \;\;\; 1 + \frac 1 1 \;\;\; , \;\;\; 1 + \frac{1}{1 + \frac 1 1} \;\;\; , \;\;\; 1 + \frac{1}{1 + \frac{1}{1 + \frac 1 1}} \;\;\; , \;\;\; \cdots$$

will be the sequence of convergents for your infinitely nested fraction. In light of this, it is clear that your sequence is always positive. You have logically assigned a value for your sequence as well (albeit on the premise that it converges in the first place).

In that light, there is only one value which this expression could reasonably have: $\varphi$, or $(1+ \sqrt 5)/2$. The negative solution won't work!


Summary & Connections:

In short, for infinite processes such as these, we like to define a series of convergents by truncating the infinite operation or expression or whatever at finite places, each further down the line. Then we consider the limit of these truncated expressions as we go further and further down the line. This is much akin to how, in calculus, we define an infinite sum as the limit of the partial sums, but more generally!

This method, provided the limit exists of course, allows us to properly assign a value to a given infinite expression. Sometimes other methods like yours will result in that value being vague - you get a multiplicity of values, but you can't be sure which value is "right," much akin to how certain methods of equation solving introduce extraneous solutions. However, this method allows us to, if nothing else, verify which of that plurality is correct!


A Note On Stability:

There is one more topic I'd like to touch on: stability of solutions to an iterative process. As you might be able to guess from the sequence of convergents for your infinite fraction, we can devise a recurrence relation that describes how to get the next convergent from the previous:

$$a_{n+1} = 1 + \frac{1}{a_{n}}$$

Dynamical systems and a lot of application-related things also have stability as a core concept. So it is true here.

Of course, as you might imagine, we want to consider the limit $\lim_{n \to \infty} a_n$ - that would be your infinite fraction again, no? But there's an issue: what is our first term of the sequence? Our initial condition of the recurrence?

Here's where things get really interesting, and why your negative solution is not something to be thrown away carelessly...

Start by defining, say, $a_1 = 1$. Playing around in Excel, then, we get this:

enter image description here

Let's play around with other initial values. Here's a few samples: $a_1=2,\pi,1000$.

enter image description here

Notice something? Each sequence converges (quite quickly) to $1.618$ or so. That is, it converges to the golden ratio, $(1+\sqrt 5)/2$.

In fact, let's even get close to the negative solution. We call this the conjugate of the golden ratio, $\bar \varphi = (1-\sqrt 5)/2 \approx -0.618.$ Playing around in Excel some more, let's set $a_1=-0.9,-0.618,$ and $-0.618034$, increasingly close approximations to $\bar \varphi$.

enter image description here

But I wonder... What if $a_1 = \bar \varphi$? Let's try that!

enter image description here

What we notice is quite incredible: even after comparatively many iterations, $a_n \approx \bar \varphi$ even still! In fact, the divergence itself is only a result of computer error - truncated decimals, rounding and approximation errors, and so on! If you do the math by hand, you'll find that if $a_1 = \bar \varphi$, so does $a_2$. And $a_3.$ And $a_4$. And so on.

This touches on the notion of stability. Let's summarize our observations.

  • Whenever $a_1 \ne \bar \varphi$, $a_n$ approaches $\varphi$. It might take a while, but no matter how large or small or negative or close to $\bar \varphi$, it eventually approaches it. (Unless we run into a divide-by-zero error along the way but that's just something to ignore.)

  • Whenever and only whenever $a_1 = \bar \varphi$, $a_n$ does not approach $\varphi$ but rather it stays at $\bar \varphi$.

In such a case, we call $\varphi$ a "stable" solution because (within some amount of reason, since some systems will have multiple stable/unstable solutions) assigning $a_1$ a value near $\varphi$ (or anywhere aside the unstable solution in this case) will ensure $a_n \to \varphi$.

On the other hand, $\bar \varphi$ is an "unstable" solution, because even if you start near it, $a_n$ eventually goes away from it, towards $\varphi$ in this case. However, if $a_1 = \bar \varphi$, then $a_n \to \bar \varphi$.

So while your fraction might only have $\varphi$ as a reasonable way of assigning its value, $\bar \varphi$ has an interesting role to play with respect to stability.

PrincessEev
  • 43,815