0

When multiplying 2 square matrices the resulting first row ends up being a combination of the first column of a matrix and the first row of one another. Aren't we geometrically missing information?

Edit: What I mean is that to compute each composition of transformations there is a certain algorithm which leads to the correct result: 1. I find it hard to ignore the way it is built up; why is it that to find the new image of i-hat (as shown in the picture) we multiply its first component by the x components (first column) of the second transform, and its second component by the y components? I at first thought we were just "saving" the motion/displacement exherted upon the first transform's "i-hat" components and that it didn't really matter in which component we saved it cause, well, it will end up being a single vector -combination of all the components. But I then discovered that the order followed was the unique path to get the correct composition. All the answers I've seen regarding this question state that this arises the "natural properties" of composition, but the matter is, why? There must be a geometrical explanation as to why we map the given vectors into the images of certain components and not otherwise.

I am deeply sorry if I am being unclear or if this is is just a stupid question that arises from a sequence of misconceptions about matrix multiplication, but it has been stoping me from going any further in linear algebra so I'd appreciate if someone told me what is really going on.

  • Matrices themselves represent linear transformations and essentially tell one how a coordinate system changes. Matrix multiplication represents a composition of linear transformations (they are functions after all) and hence the product of two matrixes represents two consecutive coordinate changes done at once. Here is a link to neat video that might help with a geometric viewpoint to matrix multiplication: https://youtu.be/XkY2DOUCWMU – user328442 Dec 24 '18 at 00:19
  • Welcome to stackexchange. I can't tell from your question what you really want to know. Please [edit] the question. Perhaps show us the product of two square matrices and tell us just what "geometrically missing information" you have in mind. Do that with an [edit] to the question, not in a comment, and use mathjax: https://math.meta.stackexchange.com/questions/5020/mathjax-basic-tutorial-and-quick-reference – Ethan Bolker Dec 24 '18 at 00:40
  • 1
    Missing something in terms of what? What would you like to know that you feel you don't have enough information about? – Joel Pereira Dec 24 '18 at 00:50

2 Answers2

1

1.Let $E=\{e_1,...,e_n\}$ be any vector-space basis for $\Bbb R^n.$ For every $v\in\Bbb R^n$ there is a unique sequence $v_1,...,v_n$ of numbers such that $v=v_1e_1+...+v_ne_n.$ And $v_j$ is called "the $j$-th co-ordinate of $v$ with respect to the basis $E$".

Note that the $j$-th co-ordinate of a sum of some $k$ vectors is the sum of the individual vectors' $j$-th co-ordinates. That is, $(\,v(1)+...+v(k)\,)_j=v(1)_j \,+...+\,v(k)_j\;.$ And of course $(rv)_j=r\cdot v_j$ for any $r\in \Bbb R.$

  1. A linear map $f:\Bbb R^n\to \Bbb R^n$ is uniquely and completely determined by $f(e_1),...,f(e_n).$ Because the co-ordinates $v_1,...,v_n$ are uniquely determined by a given $v,$ and $$f(v)=f(v_1e_1+..+v_ne_n)=f(v_1e_1)+...+f(v_ne_n)=v_1f(e_1)+...+v_n f(e_n).$$

  2. Let $f_{j,i}$ be the $j$-th co-ordinate (with respect to $E$) of $f(e_i)$. For any $v\in \Bbb R^n,$ and any $j$, the $j$-th co-ordinate of $f(v)$ is $$ (f(v))_j=(\,v_1f(e_1)\,+...+\,v_n f(e_n)\,)_j =$$ $$=(v_1f(e_1))_j\,+...+\,(v_nf(e_n))_j=$$ $$=v_1f_{j,1}\,+...+\,v_nf_{j,n}\;.$$

The array $(f_{j,i})_{1\leq i\leq n, 1\leq j\leq n}$ is called "the matrix representation of $f$ with respect to $E$".

Remark: Any $v\in \Bbb R^n$ is a sequence $(r_1,...,r_n)$ of numbers, and one choice for a basis $E$ is, for each $i$, to let $e_i$ be the sequence $(e_{i,1}\,,...,e_{i,n})$ in which $e_{i,i}=1$ and $e_{i,j}=0$ when $j\ne i$. With respect to this basis, if $v=(r_1,...,r_n)$ then $v_j=r_j.$

0

Matrix multiplication is feeding one linear change of variables into another. See my answer at Why, historically, do we multiply matrices as we do?.

KCd
  • 46,062