At first, one might define a vector as a directed line-segment; in other words, an arrow of a given length pointing in some direction. We say that two vectors are equal if they have the same length and the same direction (though not necessarily the same position in space). In these notes we will usually denote vectors by small letters in boldface or with small arrows over them, such as . When handwriting, it is also customary to underline vectors: .
Usually, we will think of vectors as being elements of ; in other words, an ordered list of real numbers, , where . Here are called the coordinates of (see also 2.E). Instead of picturing this as a point in space, I recommend visualizing this element as a directed line-segment (i.e. an arrow) starting at the origin , and ending at the point . If and , then the following statements should be familiar to you:
has a binary operation called vector addition: If , then .
Addition is commutative: If , then .
Addition is associative: If , then .
There is an additive identity: , called a zero vector, such that we have .
There are additive inverses: If , then such that , where is a zero vector from V4.
is a field, and there is a scalar multiplication operation: If and then .
The multiplicative identity in operates as follows: If then .
Scalar multiplication is compatible with multiplication in : If then .
Distributivity of vector addition: If , then .
Distributivity of field addition: If , then .
If is a set and is a field, and there is an addition (as in V1), and a scalar multiplication (as in V6), which obey all of the above axioms, then the quadruple is called a vector space. In that case, elements of the set are called vectors, and is called its field of scalars; we will also say that is a vector space over the field . The main example of a vector space that we will consider in this module is , but there are other important ones as well. This is another abstract definition (see the discussion in Section 1.A). The above rules are called the vector space axioms.
The set for any positive integer is a vector space over . If , , and , then the addition and scalar multiplication operations are defined as follows:
The set of square complex matrices is a vector space over , with the usual matrix addition and scalar multiplication.
As a generalization of the previous example, matrices with rows and columns and whose coefficients are in , with usual addition and scalar multiplication, is a vector space over the field .
The set of polynomials of degree less than or equal to , with coefficients in the field . In other words,
You add polynomials by adding their coefficients. The zero vector in is . In fact, all of the vector space axioms are satisfied.
The set of all functions from to is a vector space over . Addition and scalar multiplication are defined as follows:
where and .
Let be the set of real matrices, with the usual matrix addition and scalar multiplication.
is a vector space over . Why is not a vector space over ?
[End of Exercise]
A key property of vector spaces is that they contain all linear combinations of all of their vectors.
A linear combination of the vectors is an vector of the following form:
For some scalars .
The above sum is written without brackets, which is only possible without risk of ambiguity due to the axiom V3.
At first, it is easiest to think about linear combinations within the vector space . But you will also need to think about linear combinations when is a vector space of matrices (so is matrix addition), or when is a vector space of functions (so refers to the addition of functions).
In the vector space , determine whether or not is a linear combination of the two vectors and .
[Hint: You may need to solve a system of linear equations.]
[End of Exercise]
Here are some other elementary properties that follow from the axioms. These might seem “obvious” to you, but once you attempt to deduce them using only the axioms, you will realise how tricky and unintuitive the proofs can be.
Let be a vector space over a field .
There is a unique zero vector; in other words, any two zero vectors are equal to each other, so we may call it “the zero vector”.
For each , there is a unique additive inverse; in other words, there is only one vector that obeys . We write that vector “”.
, for any .
, for any and (notation as in part (ii)).
For any ,
(i) Assume are two zero vectors; that means we have the following two equations for any :
,
.
That is how a zero vector was defined, in axiom V4. Since these equations are true for any , they must be true for , in which case the second equation becomes . But they must also be true for , in which case the first equation becomes . Finally, by V2, we have the following equalities:
So there can be only one zero vector.
(iii) We carefully proceed as follows. By F7 we have that in , and so
Now by V10 the above is equal to
Since , by V5, there is an additive inverse which is also in . By V1, we can add this element to both sides of , which we have already established, to obtain
Now by V3 the right hand side is equal to
By V5 the above is equal to
By V4, the above is eqaul to . But by V5, the left hand side of the above equation equals . Therefore .
[End of Exercise]