I was reading this paper where they want to find the saddle point of this equation:
$$\text{min }\text{max }\{x^TAy + \frac{\lambda}{2} ||By−z||^2\}$$
Where $x,y,z$ are vectors and $A,B$ are matrices. $\lambda$ is a real number.
On section 2.1 they reformulate this problem as a Variational Inequality. I read here (page 2332) that is easier to consider variational inequalities of the function's differentials, so the authors of the first paper (I'll call it "the paper of doom") make:
$$\Delta_x (x^TAy + \frac{\lambda}{2} ||By−z||^2) = -A^Ty \tag{1}$$ $$\Delta_y (x^TAy + \frac{\lambda}{2} ||By−z||^2) = Ax + \lambda B^T(By - z) \tag{2}$$
I assume that $\Delta_x (||By−z||^2) = 0$.
I'm calculating $\Delta_y(||By−z||^2)$ as follows:
$$\Delta_y(||By−z||^2) = \Delta_y \sum_i^L \sum_j^N (b_{ij}y_j - z_j)^2$$ $$ = \Delta_y \sum_i^L \sum_j^N |(b_{ij}y_j)^2 - 2b_{ij}y_jz_j + z_j^2|$$ $$ = \sum_i^L \sum_j^N \Delta_y |(b_{ij}y_j)^2 - 2b_{ij}y_jz_j + z_j^2|$$ $$ = \sum_i^L \sum_j^N |2b_{ij}y_j - 2b_{ij}z_j|$$ $$ = 2\sum_i^L \sum_j^N |b_{ij}y_j - b_{ij}z_j|$$
After reading this post and this post I'm calculating the derivatives of $y^TAx$ as follows: (typo here, see the edit)
Edit:
Thanks Greg for your comment, you helped me to find a typo in my question and other mistakes from the authors of the paper of doom, here's the correction:
$A \in \mathbb{R}^{LxN}$, $x \in \mathbb{R}^{L}$, $y \in \mathbb{R}^{N}$.
I'm calculating the derivatives of $x^TAy$ as follows:
$$\Delta_x x^TAy = Ay$$ $$\Delta_y x^TAy = x^TA$$
Thanks to this wonderful paper, I knew that my calculation of $\Delta_x x^TAy$ and $\Delta_y x^TAy$ were right, in the case of $\Delta_x x^TAy$ I was missing a last step: $$\Delta_y x^TAy = x^TA = a \in \mathbb{R}$$ $$a = a^T = (x^TA)^T = A^Tx$$
For the euclidean norm, in the paper of doom authors make (or at least seem to make, I can not trust them anymore):
$$\frac{\lambda}{2} ||By−z||^2 = \lambda B^T(By−z)$$
Where $z \in \mathbb{R}^L, B \in \mathbb{R}^{LxN}$
My mistake was trying to force my results to look like the ones in the paper of doom. How to calculate the derivative of the euclidean norm is well explained here and more closely to my case in here. $(By)_i$ is the $i$ element of the vector $By$, and $y_i$ is the $i$ element of the vector $y$:
$$\frac{\partial}{\partial y_i}||By - z||^2 = 2(\sum_i^L ((By_i) - z_i)) \frac{\partial}{\partial y_i}((By)_i - z_i) $$
So I'm getting:
$$\Delta_x (x^TAy + \frac{\lambda}{2} ||By−z||^2) = Ay \tag{3}$$ $$\Delta_y (x^TAy + \frac{\lambda}{2} ||By−z||^2) = A^Tx + 2(\sum_i^L ((By_i) - z_i)) \frac{\partial}{\partial y_i}((By)_i - z_i) \tag{4}$$
At this point, I'm unsure how to calculate $\frac{\partial}{\partial y_i}((By)_i - z_i)$ Does anyone know how to do it?