1

It is possible to represent the sum of two geometric Brownian paths as another geometric Brownian motion?

Intro_________________

I am trying to understand how it would behave a weighted composition of two arbitrarily chosen geometric Brownian paths, let say: $$S_1(t) = S_{01}e^{(\mu_1-\sigma_1^2/2)t+\sigma_1 W_{1t}}$$ $$S_2(t) = S_{02}e^{(\mu_2-\sigma_2^2/2)t+\sigma_2 W_{2t}}$$ where $W_{1t}$ and $W_{2t}$ are standard Brownian motions $\sim N(0,\ t)$, not necessarily independent, so $\rho_{12} = \frac{\text{COV}[\sigma_1 W_{1t},\ \sigma_2 W_{2t}]}{\sigma_1 \sigma_2}$. Now, considering weights $0 \leq a_i \leq 1$ such as $\sum_i a_i = 1$, I want to know if is possible to approximate the process: $$S_3(t) = a_1 S_1(t)+a_2 S_2(t)$$ as it were built as: $$S_3(t) = S_{03}e^{(\mu_3-\sigma_3^2/2)t+\sigma_3 W_{3t}}$$ which looks like it is a classic approximation for the sum of Log-Normal variables.

For trying to find the parameters, I am trying to find the expected value and the variance of $S_3(t)$ since I could estimate from there the approximate parameters $\mu_3$, $\sigma_3$ and some random process $W_{3t}$, but I got stuck on splitting the logarithm sum. I am aiming to find these parameters such later I can add other terms by extension as $\ln(S_5) = \ln(S_4+S_3)$ an so on.


What I have done so far_________________

I started by trying the following: since I could represent $$S_3(t) = S_{03}e^{(\mu_3-\sigma_3^2/2)t+\sigma_3 W_{3t}} = e^{\ln(S_{03})+(\mu_3-\sigma_3^2/2)t+\sigma_3 W_{3t}} = e^{Z}$$ $$a_1 S_1(t) = a_1 S_{01}e^{(\mu_1-\sigma_1^2/2)t+\sigma_1 W_{1t}} = e^{\ln(a_1 S_{01})+(\mu_1-\sigma_1^2/2)t+\sigma_1 W_{1t}} = e^{X}$$ $$a_2 S_2(t) = a_2 S_{02}e^{(\mu_2-\sigma_2^2/2)t+\sigma_2 W_{2t}} = e^{\ln(a_2 S_{02})+(\mu_2-\sigma_2^2/2)t+\sigma_2 W_{2t}} = e^{Y}$$ and I am trying to find $E[\ln(e^Z)] = E[Z]$ and $\text{Var}[\ln(e^Z)] = \text{Var}[Z]$, I need to figure out a way to split: $$Z = \ln(e^X+e^Y)$$ since I don't know how to apply this operators to the logarithm of a sum. At first I tried to split the logarithm of the sum by factorization: $$Z=\ln(e^X+e^Y) = \ln(e^X(1+e^{Y-X}))=\ln(e^X)+\ln(1+e^{Y-X})= X+\ln(1+e^{Y-X})$$ so I tried to study how $\ln(1+e^N)$ behaves, and as could be seen in plots in Wolfram-Alpha the function fastly behaves as $\ln(1+e^N)\approx N$, son for large $N$, or equivalently $Y \gg X$, I will have that: $$Z = X+\ln(1+e^{Y-X}) \approx X+Y-X = Y$$ which tells me that the term who dominates the sum will saturate the logarithm drowning within him the other variable as it were a slightly noise in the main trend (which is an interesting insight thinking both are terms that drift from each other "exponentially fast" as time passes).

But for gaining insight of how represent the combination of both terms at the beginning is not much helpful at all. But the plot from $\ln(1+e^N)$ resembles the comments of this another question, so using something similar to a Swish function or to integrating a Sigmoid function, I found that this function could be approximated quite close by: $$\ln(1+e^N) \approx \frac{N}{1-e^{-\frac{N}{\ln(2)}}}$$

I don't know How good it actually is, I am asking that in other question, but I think is good enough from the plot: approximation used

Later I found on Wikipedia that the function is called Softplus function, and is somehow related to the LogSumExp function so also to the Partition function.

Using this approximation, I will have that I can split: $$Z = \ln(e^X+e^Y) \approx f(X,Y) = \begin{cases} X+\frac{Y-X}{1-e^{-\frac{(Y-X)}{\ln(2)}}},\quad X \neq Y \\ X+\ln(2), \quad X\equiv Y \end{cases}$$

Ignoring the case $X\equiv Y$ I could represent also: $$f(X,Y) = \frac{Xe^{\frac{X}{\ln(2)}}-Ye^{\frac{Y}{\ln(2)}}}{e^{\frac{X}{\ln(2)}}-e^{\frac{Y}{\ln(2)}}} = X \left[\frac{e^{\frac{X}{\ln(2)}}}{e^{\frac{X}{\ln(2)}}-e^{\frac{Y}{\ln(2)}}}\right]+Y\left[\frac{e^{\frac{Y}{\ln(2)}}}{e^{\frac{Y}{\ln(2)}}-e^{\frac{X}{\ln(2)}}}\right]$$

so for values were $X$ have the same order of $Y$ the combination is driven by highly non-linear coefficients, and here is were I got stuck.

Does somebody have an idea of how to get a decent approximation of $E[f(X,Y)]$ and $\text{Var}[f(X,Y)]$???


Added later:

I realized later, that since if one random variable take advantage the sum got saturated $\ln(e^X+e^Y)\overset{Y\gg X}{\approx} Y$, I will only need to focus when they are comparable in size, so I tried to use the Taylor expansion: $$\ln(1+e^x) \approx \ln(2)+\frac{x}{2}+\frac{x^2}{8}+\mathbb{O}(x^4)$$

which leads to splitting the sum approximately as: $$\begin{array}{r c l} Z = X+\ln(1+e^{Y-X}) & \approx & X+\ln(2)+\frac{Y-X}{2}+\frac{(Y-X)^2}{8} \\ & = & \underbrace{\ln(2)+\frac{Y+X}{2}}_{\text{1st order approx.}}+\frac{(Y-X)^2}{8} \end{array}$$

So far I was been able to perform a really interesting insight from the $1st$ order approximation, but I think the $2nd$ order term is required for getting something useful, and I got stuck splitting the $(Y-X)^2$ term so any help is welcome.

2nd Added later

Maybe it could be more useful to approximate the function $\ln(1+e^x)$ through a piecewise function:

$$h(x)=\begin{cases} 0,\quad x\leq -\ln(16)\\ \frac{(x+\ln(16))^2}{4\ln(16)} ,\quad -\ln(16)<x<\ln(16)\\ x,\quad x\geq \ln(16)\end{cases}$$

which plot could be seen in Desmos: piecewise approx

with this, I would have a similar approximation for: $$Z \approx \ln(2)+\frac{Y+X}{2}+\frac{(Y-X)^2}{4\ln(16)}$$

So at least for both approximation the $1st$ order expansion is the same.

Unfortunately it leads to the same issue it got me stuck: how to evaluate the variance of something of the form: $$\text{Var}[aX+bY+cXY+dX^2+eY^2]$$

Last update

I made the question for the variance for the second order expansion here, and it looks that in the best of the scenarios is a nightmare with $+20$ terms.

Since the first order approximation was fruitful for my understanding, and at the end, if one variable grow this mess-of-terms will behave at last just like the variance of the bigger variable, I think it lost the point of what I could gain as increasing the complexity of the analysis. With this I close the question.

RobPratt
  • 45,619
Joako
  • 1,380
  • 1
    "To find the expected value and the variance of S3(t)," we do have $ {\rm E}[S_3(t)] = a_1{\rm E}[S_1(t)] + a_2{\rm E}[S_2(t)] $ and $ {\rm Var}[S_3(t)] = a_1^2{\rm Var}[S_1(t)] + a_2^2{\rm Var}[S_2(t)] +2a_1a_2{\rm Cov}[S_1(t),S_2(t)] $. – ir7 Dec 28 '23 at 21:17
  • @ir7 Thanks for commenting. I am trying to understand the Geometric Brownian Motion, and I am founding that the traditional expected value $E[e^Z]$ and variance $\text{Var}[e^Z]$ are not really representative (just notice that $E[e^Z]-\sqrt{\text{Var}[e^Z]}$ will grow well below zero when $e^Z$ is always non-negative), and instead I am looking for better figures like $e^{E[Z]}$ which correspond to the Median Value of $e^Z$. This is why $E[S_3]$ is not what I am looking for, but instead $q = E[\ln(S_3)]$ such I could make later $e^q$ as a representative path (same happen for the variance). – Joako Dec 29 '23 at 02:16
  • We can assume S3(t) to be lognormal and match its first two moments (already computed by me), or to be shifted lognormal and match the first three moments. Another approach is to estimate its pdf from cumulants using the Edgeworth series (expansion around the normal pdf). – ir7 Dec 30 '23 at 20:05
  • (Also, a mixture of lognormals can be used, if higher moment matching is needed.) – ir7 Dec 30 '23 at 20:48

0 Answers0