It is possible to represent the sum of two geometric Brownian paths as another geometric Brownian motion?
Intro_________________
I am trying to understand how it would behave a weighted composition of two arbitrarily chosen geometric Brownian paths, let say: $$S_1(t) = S_{01}e^{(\mu_1-\sigma_1^2/2)t+\sigma_1 W_{1t}}$$ $$S_2(t) = S_{02}e^{(\mu_2-\sigma_2^2/2)t+\sigma_2 W_{2t}}$$ where $W_{1t}$ and $W_{2t}$ are standard Brownian motions $\sim N(0,\ t)$, not necessarily independent, so $\rho_{12} = \frac{\text{COV}[\sigma_1 W_{1t},\ \sigma_2 W_{2t}]}{\sigma_1 \sigma_2}$. Now, considering weights $0 \leq a_i \leq 1$ such as $\sum_i a_i = 1$, I want to know if is possible to approximate the process: $$S_3(t) = a_1 S_1(t)+a_2 S_2(t)$$ as it were built as: $$S_3(t) = S_{03}e^{(\mu_3-\sigma_3^2/2)t+\sigma_3 W_{3t}}$$ which looks like it is a classic approximation for the sum of Log-Normal variables.
For trying to find the parameters, I am trying to find the expected value and the variance of $S_3(t)$ since I could estimate from there the approximate parameters $\mu_3$, $\sigma_3$ and some random process $W_{3t}$, but I got stuck on splitting the logarithm sum. I am aiming to find these parameters such later I can add other terms by extension as $\ln(S_5) = \ln(S_4+S_3)$ an so on.
What I have done so far_________________
I started by trying the following: since I could represent $$S_3(t) = S_{03}e^{(\mu_3-\sigma_3^2/2)t+\sigma_3 W_{3t}} = e^{\ln(S_{03})+(\mu_3-\sigma_3^2/2)t+\sigma_3 W_{3t}} = e^{Z}$$ $$a_1 S_1(t) = a_1 S_{01}e^{(\mu_1-\sigma_1^2/2)t+\sigma_1 W_{1t}} = e^{\ln(a_1 S_{01})+(\mu_1-\sigma_1^2/2)t+\sigma_1 W_{1t}} = e^{X}$$ $$a_2 S_2(t) = a_2 S_{02}e^{(\mu_2-\sigma_2^2/2)t+\sigma_2 W_{2t}} = e^{\ln(a_2 S_{02})+(\mu_2-\sigma_2^2/2)t+\sigma_2 W_{2t}} = e^{Y}$$ and I am trying to find $E[\ln(e^Z)] = E[Z]$ and $\text{Var}[\ln(e^Z)] = \text{Var}[Z]$, I need to figure out a way to split: $$Z = \ln(e^X+e^Y)$$ since I don't know how to apply this operators to the logarithm of a sum. At first I tried to split the logarithm of the sum by factorization: $$Z=\ln(e^X+e^Y) = \ln(e^X(1+e^{Y-X}))=\ln(e^X)+\ln(1+e^{Y-X})= X+\ln(1+e^{Y-X})$$ so I tried to study how $\ln(1+e^N)$ behaves, and as could be seen in plots in Wolfram-Alpha the function fastly behaves as $\ln(1+e^N)\approx N$, son for large $N$, or equivalently $Y \gg X$, I will have that: $$Z = X+\ln(1+e^{Y-X}) \approx X+Y-X = Y$$ which tells me that the term who dominates the sum will saturate the logarithm drowning within him the other variable as it were a slightly noise in the main trend (which is an interesting insight thinking both are terms that drift from each other "exponentially fast" as time passes).
But for gaining insight of how represent the combination of both terms at the beginning is not much helpful at all. But the plot from $\ln(1+e^N)$ resembles the comments of this another question, so using something similar to a Swish function or to integrating a Sigmoid function, I found that this function could be approximated quite close by: $$\ln(1+e^N) \approx \frac{N}{1-e^{-\frac{N}{\ln(2)}}}$$
I don't know How good it actually is, I am asking that in other question, but I think is good enough from the plot:
Later I found on Wikipedia that the function is called Softplus function, and is somehow related to the LogSumExp function so also to the Partition function.
Using this approximation, I will have that I can split: $$Z = \ln(e^X+e^Y) \approx f(X,Y) = \begin{cases} X+\frac{Y-X}{1-e^{-\frac{(Y-X)}{\ln(2)}}},\quad X \neq Y \\ X+\ln(2), \quad X\equiv Y \end{cases}$$
Ignoring the case $X\equiv Y$ I could represent also: $$f(X,Y) = \frac{Xe^{\frac{X}{\ln(2)}}-Ye^{\frac{Y}{\ln(2)}}}{e^{\frac{X}{\ln(2)}}-e^{\frac{Y}{\ln(2)}}} = X \left[\frac{e^{\frac{X}{\ln(2)}}}{e^{\frac{X}{\ln(2)}}-e^{\frac{Y}{\ln(2)}}}\right]+Y\left[\frac{e^{\frac{Y}{\ln(2)}}}{e^{\frac{Y}{\ln(2)}}-e^{\frac{X}{\ln(2)}}}\right]$$
so for values were $X$ have the same order of $Y$ the combination is driven by highly non-linear coefficients, and here is were I got stuck.
Does somebody have an idea of how to get a decent approximation of $E[f(X,Y)]$ and $\text{Var}[f(X,Y)]$???
Added later:
I realized later, that since if one random variable take advantage the sum got saturated $\ln(e^X+e^Y)\overset{Y\gg X}{\approx} Y$, I will only need to focus when they are comparable in size, so I tried to use the Taylor expansion: $$\ln(1+e^x) \approx \ln(2)+\frac{x}{2}+\frac{x^2}{8}+\mathbb{O}(x^4)$$
which leads to splitting the sum approximately as: $$\begin{array}{r c l} Z = X+\ln(1+e^{Y-X}) & \approx & X+\ln(2)+\frac{Y-X}{2}+\frac{(Y-X)^2}{8} \\ & = & \underbrace{\ln(2)+\frac{Y+X}{2}}_{\text{1st order approx.}}+\frac{(Y-X)^2}{8} \end{array}$$
So far I was been able to perform a really interesting insight from the $1st$ order approximation, but I think the $2nd$ order term is required for getting something useful, and I got stuck splitting the $(Y-X)^2$ term so any help is welcome.
2nd Added later
Maybe it could be more useful to approximate the function $\ln(1+e^x)$ through a piecewise function:
$$h(x)=\begin{cases} 0,\quad x\leq -\ln(16)\\ \frac{(x+\ln(16))^2}{4\ln(16)} ,\quad -\ln(16)<x<\ln(16)\\ x,\quad x\geq \ln(16)\end{cases}$$
which plot could be seen in Desmos:
with this, I would have a similar approximation for: $$Z \approx \ln(2)+\frac{Y+X}{2}+\frac{(Y-X)^2}{4\ln(16)}$$
So at least for both approximation the $1st$ order expansion is the same.
Unfortunately it leads to the same issue it got me stuck: how to evaluate the variance of something of the form: $$\text{Var}[aX+bY+cXY+dX^2+eY^2]$$
Last update
I made the question for the variance for the second order expansion here, and it looks that in the best of the scenarios is a nightmare with $+20$ terms.
Since the first order approximation was fruitful for my understanding, and at the end, if one variable grow this mess-of-terms will behave at last just like the variance of the bigger variable, I think it lost the point of what I could gain as increasing the complexity of the analysis. With this I close the question.