I am trying to derive Newton step in an iterative optimization. I know the step is:
$$\Delta x=-H^{-1}g$$
where H is Hessian and $g$ is gradient of a vector function $f(x)$ at $x$.
I also know the step is derived from second order Taylor expansion.
For example, let's have a vector fuction $f(x)$ which is $\mathbb{R}^{3}\to\mathbb{R}$. Its second-order Taylor expansion is:
$$f(x+\Delta x)\approx f(x)+g^{T}\Delta x+ \frac{1}{2}(\Delta x)^{T}H(\Delta x)$$
where $g$ is gradient and $H$ is Hessian of $f$ at $x$.
Taking partial derivative with respect to $\Delta x$ should be:
$$\frac{\partial f(x+\Delta x)}{\partial \Delta x}=g^{T}+H(\Delta x)$$
The first term is clear since $g^{T}$ is multiplier of $\Delta x$. But why
$$\frac{\partial}{\partial\Delta x} \frac{1}{2}(\Delta x)^{T}H(\Delta x)=H(\Delta x)$$ ?
Using product rule leads me to a different result (ommiting the $\frac{1}{2}$ for more clarity):
$$\frac{\partial}{\partial\Delta x} (\Delta x)^{T}H(\Delta x)= \\ =\frac{\partial(\Delta x)^{T}}{\partial\Delta x}H(\Delta x)+(\Delta x)^{T}\frac{\partial H}{\partial\Delta x}(\Delta x)+(\Delta x)^{T}H\frac{\partial\Delta x}{\Delta x}= \\ =1\cdot H\cdot\Delta x + 0 + (\Delta x)^{T}H\cdot 1= \\ =H\cdot\Delta x + (\Delta x)^{T}H$$