I'm trying to understand a conclusion in [1], [2].
There they state a definition and a theorem:
Definition(5.2 in [1], 2.2 in [2]): For a general linear System $Ax=y$, we say that $\hat{x}$ is a least-squares solution, if $||A\hat{x}-y|| = \min_x ||Ax-y||$. We say that $\hat{x}$ is a minimum least-squares solution, if $\hat{x}$ has the smallest norm among all least-square solutions.
Theorem(5.1 in [1], 2.1 in [2]): Let there exist a matrix $G$ such that $Gy$ is a minimum least-squares solution of a linear system $Ax=y$. Then it is necessarcy and sufficent that $G = A^+$, the Moore-Penrose generalized inverse of matrix A.
Now, both papers want to find a minimum least-squares solution of the linear system $H\beta = T$, that is $$||H\hat{\beta} - T|| = \min_\beta ||H\beta - T||, \qquad (1)$$ where $H \in \mathbb{R}^{N \times \tilde{N}}$, $\beta \in \mathbb{R}^{\tilde{N} \times m}$ and $T \in \mathbb{R}^{N \times m}$.
They argue as follows: "According to the theorem, the minimum least-squares solution of above linear system is $\hat{\beta} = H^+T$, where $H^+$ is the Moore-Penrose generalized inverse of matrix $H$.".
Let be $\hat{\beta}$ the solution found in $(1)$. To apply the theorem, there must exist a matrix $G$ so that $\hat{\beta} = GT$. But can we make sure there always exist such a matrix $G$? From my point of view, I can't see that this always holds, because of the dimensions of $\hat{\beta}$ and $T$.