31

Can someone please explain how polynomials are vector spaces? I don't understand this at all. Vectors are straight, so how could a polynomial be a vector. Is it just that the coefficients are vector spaces? I'm really not understanding how polynomials tie in to linear algebra at all? I can do all the problems with them, because they are done the same way as any other vector problem, but I can't understand intuitively what I'm doing at all.

Matt Samuel
  • 58,164
tokola
  • 697
  • 14
    A vector space is any set with an addition and scalar multiplication defined that satisfies the axioms. The idea is that we are motivated by our thoughts of vectors in Euclidean n-space, but we then abstract the key features so that we can apply our geometric insight in far more general settings. We call such abstract vectors being on the down-low. No just kidding that was just one of the many bad jokes I promised not to make. – user4894 Nov 01 '14 at 20:56
  • 3
    Why polynomials form a vector space? Because the set of polynomial with 2 operations ("+", "x") do satisfy the axioms of vector space. – SamC Nov 01 '14 at 20:59
  • 2
    A mathematical vector space is defined abstractly - I suggest you look up vector spaces online or in a text book, and check the definition. One motivating factor for looking at vector spaces is that Euclidean space fits the definition, and you can do geometry using vectors. But having abstracted the key idea from the Euclidean case, we find that there are lots of other objects which fit the definition - that's one of the reasons it is so useful. The ring of polynomials in one variable over a field is an infinite dimensional vector space. – Mark Bennet Nov 01 '14 at 21:04
  • 1
    Perhaps your own comment "because they are done the same way as any other vector problem" is the key thing here. Since there is basically no difference, why make a difference? We can expand the meaning of the word "vector", and include any type of objects that we can form linear combinations of. – Hans Lundmark Nov 01 '14 at 21:40
  • 3
    To get the best answers, I think you should write down in your question what your idea of a vector space is and what you are expecting it to do for you. This way, the replies can be tailored to proceed from your current understanding to a more comprehensive one. As it stands, people are giving definitions and axioms disregarding your own idas about what a vector space is (and piling up notions like ring and field that I think will only serve to confuse you further). – guest Nov 02 '14 at 04:22

7 Answers7

29

A (real) vector space means, by definition any set $V$ together with operations ${+_V}: V\times V\to V$ and ${\times_V}: \mathbb R\times V \to V$, such that

  1. $+_V$ is associative and commutative and has a neutral element, and has inverses for every element of $V$.
  2. $\times_V$ associates with real multiplication: $a\times_V(b\times_V v)=ab\times_V v$ for all $a,b\in \mathbb R$ and $v\in V$.
  3. $\times_V$ distributes over $+_V$ in the sense that $a\times_V(v+_Vw)=(a\times_V v)+_V(a\times_V w)$ and $(a+b)\times_V v=(a\times_V v)+_V(b\times_V v)$for all $a,b\in \mathbb R$ and $v,w\in V$.

It happens that if you take the set of all polynomials together with addition of polynomials and multiplication of a polynomial with a number, the resulting structure satisfies these conditions. Therefore it is a vector space -- that is all there is to it.

It can be useful intuitively to visualize vectors as little arrows (or however you're used to thinking about geometric vectors), but this is just intuition which may or may not be helpful -- it is deliberately not part of the linear-algebra concept of a vector space.

  • 3
    I also think that it's important to note that the notion of vector doesn't necessarily implies the existence of a coordinate system, a metric system; it's also a bit counterintuitive because a vector space requires a field, and usually a field is like the $x$ and $y$ in a cartesian system, but a vector space, at least from the mathematical point of view, doesn't really has a mandatory coordinate system . – user2485710 Nov 01 '14 at 22:35
  • 5
    As a non-mathematician, I always saw Vectors as 'lists of numbers', or arrays if you will. Indeed, any polynomial can be expressed as a list of the coefficients for each power. – Mark Jeronimus Nov 02 '14 at 01:02
  • 3
    @Mark: Yes, that meaning is common in (among other places?) programming languages -- but that is a different generalization of geometric vectors than the one used in linear algebra. For example the set of "all functions $\mathbb R\to\mathbb R$, whether continuous or not" is a vector space too, and such a function cannot be represented simply as a list of numbers. – hmakholm left over Monica Nov 02 '14 at 10:57
  • 1
    $1 \times_V v = v, \forall v \in V$ is missing! I don't think it can be omitted – Geralt of Rivia Oct 07 '17 at 12:32
  • @hmakholmleftoverMonica it can if you use computable functions over the reals and Godel numbering. Then addition becomes uncomputable though. – James Gilles Dec 30 '22 at 05:17
9

When we imagine vectors, we all see arrows from our undergraduate physics lessons. And it's fine. However, what is a vector space ? Well it's any set where you can add elements between them with proportionality factors. And you can definitely add polynomials and multiply them with reals.

I think you just need to accept that the graphical representation you like is only a way to remember the only properties that define a vector : you can add them together, and you can "extend" them by multiplication with a real (or a complex number of course...).

I've always thought that "vector space" is too complex a word to simply represent the possibility of adding together and "extending" the elements of a set.

  • 13
    'Vector space' is actually a criminally misleading term. Why aren't we all using the more neutral term 'linear space' in our undergraduate courses and note that this is the same as 'vector space' but for psychological reasons we will reserve the latter term for tuples of real or complex numbers (behind the scenes: basis-dependent realizations of linear spaces)? I spend copious amounts of time scooping these preconceived notions out of my students' brains, something that could be avoided to a large extend by a simple switch in terminology. Many authors underestimate the power of terminology! – guest Nov 02 '14 at 04:09
6

A vector space has two kinds of things: vectors and scalars. It must be possible to add two vectors, and it must be possible to multiply a vector by a scalar. There are also some laws that the addition and the multiplication must obey, such as $s\cdot(\mathbf{a} + \mathbf{b}) = s\cdot\mathbf{a} + s\cdot\mathbf{b}$.

One example of this is to take the vectors to be $4$-tuples of real numbers, say $\langle a,b,c,d\rangle$, and the scalars to be single real numbers. We can add two vectors (by adding them component-wise) and we can multiply a vector by a scalar, multiplying each of the components of the vector by the scalar.

Another example of this is to take the vectors to be $2\times 2$ matrices of real numbers $$\begin{pmatrix}a&b\\c&d\end{pmatrix}$$ and the scalars to be single real numbers. We can add two of these vectors using ordinary matrix addition, and we can multiply a matrix by a scalar, multiplying each of its entries by the single number.

Of course this example is exactly the same as the one in the previous paragraph. The vectors behave exactly the same way whether we write them as $\begin{pmatrix}a&b\\c&d\end{pmatrix}$ or as $\langle a,b,c,d\rangle$. It doesn't matter whether the brackes are curvy or angled, or whether we write the numbers piled up or not.

Here is another example: take the vectors to be third-degree polynomials, say $ax^3 +bx^2+cx+d$, and the scalars to be real numbers. We can add two of these vectors using ordinary polynomial addition, and we can multiply a vector by a scalar by multiplying the coefficients of the polynomial by the single number.

This example is just like the previous two. We could abbreviate $ax^3+bx^2+cx+d$ as $\langle a,b,c,d\rangle$ (since the parts involving $x$ are aways the same, we could just agree not to write them down) or as $\begin{pmatrix}a&b\\c&d\end{pmatrix}$. It doesn't matter if we write the four numbers in a line, or in a pile, or with $x^3$es and plus signs in between; the vectors are still behaving the same way.

So why bother to do it? This is one of the great strengths of mathematics, to see when two different-seeming kinds of objects are actually the same, and to develop an abstract theory that applies in many different situations. Once we develop a theory of vector spaces, we can apply that theory to all sorts of things that behave like vectors—such as polynomials—even if we don't normally think of them as vectors. In short, by understanding polynomials as vectors, we can use the theory of vector spaces to help us solve problems that are about polynomials.

For example, in Constructing a degree 4 rational polynomial satisfying $f(\sqrt{2}+\sqrt{3}) = 0$ I used the theory of vector spaces to show that the required polynomial actually exists, and my method for calculating that polynomial draws heavily on the theory of vector spaces; it boils down to the problem of finding the coordinates of a certain vector in a different basis.

MJD
  • 65,394
  • 39
  • 298
  • 580
2

The ring of polynomials with coefficients in a field is a vector space with basis $1, x, x^2,x^3,\ldots$. Every polynomial is a finite linear combination of the powers of $x$ and if a linear combination of powers of $x$ is 0 then all coefficients are zero (assuming $x$ is an indeterminate, not a number).

Matt Samuel
  • 58,164
2

You are thinking of a vector as being a geometric object. They are straight lines when the space is $\mathbb{R}^n$. However vector spaces are more general than this and intuitive visualizations fall apart.

Vector spaces have is any space which has certain properties, such as closure under addition and properties of scalar multiplication.

http://en.wikipedia.org/wiki/Vector_space#Definition

To show that something is a vector space, you simply need to show that all the properties of vector spaces hold.

1

Perhaps it is best to define what a vector space is. First, they must be over a field; in this case, probably $\mathbb{R}$ or $\mathbb{C}$. For polynomials it is best to think of certain nice properties it satisfies. For example, $a(P(x) + Q(x)) = aP(x) + aQ(x)$ or $(a+b)P(x) = aP(x) + bP(x)$ would be two distributive properties it must satisfy. Look up what a vector space is for more of the axioms.

When you think of taking $w = cu+v$ with $u$ and $v$ going in the same direction, you find that $w$ goes in the same direction (up to orientation). Here you are thinking of the closed subspace of vectors on a given line. For polynomials, your "direction" might be that $P(0) = 0$. Then the direction isn't changed by adding polynomials or scaling by a real or complex number. But here your notion of "direction" that has changed. A vector space abstracts a geometric construction and so it can be impossible to explain these notions geometrically.

-1

Polynomial is just a function. It can have integer, decimal, vector, or matrix inputs. let $P(x)$ bedefined by,

$P(x) = 1 + 3x^2 = \sum{}_{i=0}^{k=2}\alpha_i x^i$

where $\alpha = [1,0,3]$. Given a matrix $A$,

$P(A) = 1 + 3A^2 = \sum{}_{i=0}^{2}\alpha_i A^i$

As it turns out, any vector $v$ with the same dimension as some square matrix* $A$ can be expressed as some polynomial P(A) so long as we have freedom to choose our basis. To see this requires some tricky albeit basic math.

Let $u$ be some basis vector we don't yet know the value of. I claim the following is valid,

$T^jv = \sum{}_{i =j}^{2}\alpha_{i-j} T^iu$

The above equation forms a triangular system of equations that can be solved for beginning with the highest index $j=2$. By 'triangular' I mean at the highest index $j=2$ you get only one term in the summation on the right side. We already know everything on the left (because we know T and v are arbitrary*).

What we are solving for are the values $T^iu $ on the right side. For $j=2$ we can solve for $T^2u$. Then we use this in turn to move down to $j=1$ and solve $Tu$. We now know the values of $T^2u$ and $Tu$ which allows us to go down (or up depending on how you see it) to $j=0$ and solve for $u$ itself.

We then have $u_0=u, u_1= T^u, u_2 = T^2u$ as our basis vectors which we have represented v in.

But did we represent v in terms of these vectors? Yes. If you perform the above steps the equation you get for $j=0$,

$v = \sum{}_{i =0}^{2}\alpha_{i} ui \ = \sum{}_{i =0}^{2}\alpha_{i} T^iu \ = P(A)u $

*These statements are not intended to be rigorous. More exact stipulations are warranted and discussion is welcome.

wayne
  • 11