Suppose $f\in C([a,b])$ is twice continuously differentiable and $f''(x)>0$ on the interval. Show that the best linear approximation $p$ to $f$ has the slope $p'(x)=(f(b)-f(a))/(b-a)$.
To my understanding, due to the second derivative of $f$ being greater than 0 on the entire interval, then we know for sure $f$ itself has to be concave up on the entire interval. This and the fact that $f$ is concave up tells us that the maximum of $f$ must be in the interval, more specifically, at one of the endpoints. Also, there is a minimum somewhere else in the interval.
Now, by Taylor's theorem, we know there is a Taylor polynomial $p(x)=f(c)+f'(c)(x-c)$ at $c\in [a,b]$. Due to the linearity of $f$, we know it is a polynomial of degree $1$ and so $f$ must resemble $f(x)=mx+b$ where $m,b\in\mathbb{R}$.
But now I am stuck. Looking at $p'(x)=(f(b)-f(a))/(b-a)$, the right side seems to resemble the definition of the derivative of $f$ and intuitively, it makes sense to me that $p$ approximating $f$ will have this matching slope but I cannot seem to prove it. How to proceed?