Skip to main content

Section 2.3 Vector Spaces

We now move to our next task of abstraction. We have generalized the real numbers and introduced the idea of a field, and we will now generalize the set and structure of the vectors \(\rr^n\) (see SectionΒ 1.4) to a vector space.

Definition 2.3.1.

A vector space over a field \(\ff\) is a set \(V\) on which are defined the operations of addition and scalar multiplication such that all of the following properties hold.
  1. For all \(\bfu, \bfv \in V\text{,}\) \(\bfu+\bfv \in V\text{.}\) (The sum of two vectors is a vector.)
  2. For all \(c \in \ff\) and all \(\bfv \in V\text{,}\) \(c\bfv \in V\text{.}\) (The scalar multiple of a vector is a vector.)
  3. For all \(\bfu, \bfv \in V\text{,}\) \(\bfu+\bfv=\bfv+\bfu\text{.}\) (Vector addition is commutative.)
  4. For all \(\bfu, \bfv, \bfw \in V\text{,}\) \((\bfu+\bfv)+\bfw = \bfu+(\bfv+\bfw)\text{.}\) (Vector addition is associative.)
  5. There is a vector \(\mathbf{0} \in V\) such that \(\bfv + \mathbf{0} = \bfv\) for all \(\bfv \in V\text{.}\) (There is a vector which is the identity for vector addition. We call this the zero vector.)
  6. For each \(\bfu \in V\) there exists a vector \(\bfv \in V\) such that \(\bfu + \bfv = \mathbf{0}\text{.}\) (Each vector has an additive inverse.)
  7. For each \(\bfv \in V\text{,}\) \(1\bfv = \bfv\text{.}\) (Scalar multiplication by \(1\) is an identity.)
  8. For each \(\bfv \in V\) and all \(c, d \in \ff\text{,}\) \(c(d\bfv)=(cd)\bfv\text{.}\) (Scalar multiplication of a vector is associative.)
  9. For all \(\bfu, \bfv \in V\) and each \(c \in \ff\text{,}\) \(c(\bfu+\bfv) = c\bfu + c\bfv\text{.}\) (Scalar multiplication distributes over the sum of vectors.)
  10. For each \(\bfv \in V\) and all \(c, d \in \ff\text{,}\) \((c+d)\bfv = c\bfv + d\bfv\text{.}\) (Scalar multiplication distributes over the sum of field elements.)
The elements of \(V\) are called vectors and the elements of \(\ff\) are called scalars.

Note 2.3.2.

The reader will want to note when the multiplication of field elements is in view and when scalar multiplication (of a scalar times a vector) is in view. The context should help, as should our practice of using bold notation for vectors. The product of vectors (in the way that we take the product of field elements) is not a defined construction in a general vector space.
Faced with this rather abstract definition, some examples are in order.

Example 2.3.3.

The set \(\rr^n\text{,}\) with the operations of scalar multiplication and vector addition defined in DefinitionΒ 1.4.3, is a vector space over \(\rr\text{.}\) As with our definition of a field, if \(\rr^n\) is not a vector space then we have carried out the enterprise of abstraction rather poorly. (See FactΒ 1.4.10 where we stated most of the properties of a vector space for \(\rr^n\text{,}\) though we did not use that language.)

Example 2.3.4.

Let \(\ff\) be a field and let \(\ff^n\) be defined in the following way:
\begin{equation*} \ff^n = \left\{ \left[\begin{array}{@{}c@{}} a_1 \\ a_2 \\ \vdots \\ a_n \end{array}\right] \Bigm\vert a_i \in \ff \right\}\text{.} \end{equation*}
The operations of scalar multiplication and vector addition are defined for \(\ff^n\) over \(\ff\) in the same way that the operations of scalar multiplication and vector addition are defined for \(\rr^n\) over \(\rr\text{.}\) Then \(\ff^n\) is a vector space.
This should be relatively easy for the reader to believe, but checking the details could be a helpful exercise in the definitions of both fields and vector spaces.

Example 2.3.5.

Let \(A\) be a subset of the real numbers. Define the set \(V\) as the set of all functions \(A \to \rr\text{.}\) Given appropriate operations, this is a vector space.
If \(f\) and \(g\) are elements of \(V\text{,}\) then we define the sum as
\begin{equation*} (f+g)(t) = f(t)+g(t) \end{equation*}
for all \(t \in A\text{.}\) (The vector \(f+g\) must be a function \(A \to \rr\text{,}\) so we here define how to calculate the value of this function for every \(t \in A\text{.}\)) Additionally, if \(c \in \rr\) and \(f \in V\text{,}\) then we define scalar multiplication as
\begin{equation*} (cf)(t) = cf(t) \end{equation*}
for all \(t \in A\text{.}\) (Again, we are defining the value of the function \(cf\) at each element of \(A\text{.}\)) We will now argue that \(V\) is a vector space.
Looking back at DefinitionΒ 2.3.1, the first two axioms hold by the way we defined the operations. Axioms three and four hold because addition in \(\rr\) is both commutative and associative. (Since these vectors in \(V\) are functions which take values in \(\rr\text{,}\) the properties of \(\rr\) are once again crucial.)
The zero vector in \(V\) is the function \(f(t)=0\text{,}\) the function which is uniformly zero for each value \(t \in A\text{.}\) This function has the properties of the zero vector mandated by axiom five.
The additive inverse of a function \(g \in V\) is the function \((-1)g=-g\) since
\begin{equation*} (g+(-g))(t) = g(t)+(-g)(t)=g(t)-g(t)=0 \end{equation*}
for all \(t \in A\text{.}\) This shows that axiom six holds.
The rest of the axioms hold because of the definitions of vector addition and scalar multiplication and the properties of the real numbers.

Note 2.3.6.

The previous example can be difficult to digest, because we are considering functions to be vectors. It may take some adjustment to think of functionsβ€”rather than single real numbers (or even ordered lists of real numbers)β€”as the objects of study .

Example 2.3.7.

Let \(P_n\) denote the set of all real-valued polynomials of degree less than or equal to \(n\) with real coefficients. (Recall that the degree of a polynomial is the largest exponent of the variable that has a nonzero coefficient.) This means that the polynomial \(p(x) = 7x^{10}-\tfrac{1}{2}x^6 + 9\) has degree \(10\) and is an element of \(P_n\) for all \(n \ge 10\text{.}\)
We consider vector addition to be the usual algebraic sum of polynomials and scalar multiplication to be the usual product of a polynomial by a constant. Then \(P_n\) is a vector space over \(\rr\text{.}\)

Example 2.3.8.

Consider the set \(V=\rr^2\) but with non-standard operations. (In this example and the next, we will use horizontal instead of vertical notation for \(\rr^2\text{.}\) This is purely for ease of notation in these limited instances.) We define the sum of two vectors \(\bfu=(u_1,u_2)\) and \(\bfv=(v_1,v_2)\) to be
\begin{equation*} \bfu\oplus \bfv = (u_1v_1,1)\text{,} \end{equation*}
and we define scalar multiplication of a vector \(\bfu=(u_1,u_2)\) by a real number \(c\) to be
\begin{equation*} c\odot \bfu = (cu_1,1)\text{.} \end{equation*}
It is a good exercise to grapple with the axioms and determine whether or not \(V\) with these operations is a vector space.
(We are using the symbols \(\oplus\) and \(\odot\) because these operations we are defining are not familiar ones, and most readers likely have no previous associations with these symbols. For our purposes, these symbols have no overarching meaning; they will be redefined for the purposes of specific examples and exercises.)
With some work, the reader will find that this is not a vector space. All we need to do in order to show this is not a vector space is to find one axiom that does not hold. We will show that the fifth axiom (regarding the zero vector) does not hold.
We must be careful here, because the terminology can become confusing. By β€œzero vector” we do not always mean in \(\rr^2\) the ordered pair \((0,0)\text{.}\) Depending on the operations that are in view, the zero vector may exist in a form other than \((0,0)\text{.}\) To show this axiom does not hold, we must argue that no element of \(\rr^2\) can have the properties of a zero vector given these operations.
We will argue by contradiction. Suppose there is a vector \(\bfu = (u_1,u_2)\) which has the properties of a zero vector. This means that, for any \(\bfv \in V\text{,}\) we have \(\bfu \oplus \bfv = \bfv \oplus \bfu = \bfv\text{.}\) We consider the vector \(\bfv = (2,2)\text{.}\) If \(\bfu\) is the zero vector, then we must have
\begin{equation*} (u_1,u_2) \oplus (2,2) = (2,2)\text{,} \end{equation*}
but the addition on this set means
\begin{equation*} (u_1,u_2) \oplus (2,2) = (2u_1,1)\text{.} \end{equation*}
Since it is impossible to have \((2,2)=(2u_1,1)\) as the second coordinates are not equal as elements of \(\rr\text{,}\) we have a contradiction, and therefore there can be no zero vector in \(V\) with these operations.

Example 2.3.9.

We again consider the set \(V=\rr^2\text{.}\) Vector addition will be the usual sum in \(\rr^2\text{,}\) but scalar multiplication will be different. For any real number \(c\) and any \(\bfu = (u_1,u_2) \in \rr^2\text{,}\) we define \(c\odot \bfu\) by
\begin{equation*} c\odot \bfu = (cu_1,\tfrac{1}{2}cu_2)\text{.} \end{equation*}
Once again we pose the question: Is this a vector space?
While it is good practice to check all of the axioms, again we only need to find one axiom which does not hold in order to show this is not a vector space. (If it were a vector space, we would need to prove that all of the axioms hold.) We will show that scalar multiplication by \(1\) is not an identity (axiom 7).
We will examine the scalar product \(1\odot (1,1)\text{:}\)
\begin{equation*} 1\odot (1,1) = (1,\tfrac{1}{2})\text{.} \end{equation*}
If axiom 7 were to hold, we would have \(1\odot (1,1)=(1,1)\text{.}\) Since \((1,1)\neq (1,\tfrac{1}{2})\text{,}\) we have shown that axiom 7 does not hold. Therefore, this is not a vector space.

Example 2.3.10.

If we fix positive integers \(m\) and \(n\) and a field \(\ff\text{,}\) then \(M_{m,n}(\ff)\) is the set of all \(m\times n\) matrices with entries in \(\ff\text{.}\) When \(m=n\text{,}\) we use the notation \(M_n(\ff)\text{.}\)
With the standard addition and scalar multiplication of matrices (see NoteΒ 1.4.8 for a description), the set \(M_{m,n}(\ff)\) is a vector space over \(\ff\) for a fixed \(m\) and \(n\text{.}\) The quick explanation here is that because of the way scalar multiplication and addition of matrices are defined, once a reader believes that \(\ff^n\) is a vector space (see ExampleΒ 2.3.4), believing that \(M_{m,n}(\ff)\) is a vector space just amounts to considering several columns at the same time. But since these columns have no interaction with each other as far as these operations are concerned, none of the vector space axioms will be violated. The zero vector for this vector space is the zero matrix.
As evidence of the fact that \(\rr^n\) is the model for our general definition of a vector space, we now repeat the definitions of linear combination and span. These definitions in the general setting will be necessaary for the following section. (See the original definitions in DefinitionΒ 1.4.6.)

Definition 2.3.11.

Let \(\ff\) be a field and let \(c_1, c_2, \ldots, c_m \in \ff\text{.}\) Let \(V\) be a vector space over \(\ff\) and let \(\bfv_1, \bfv_2, \ldots, \bfv_m \in V\text{.}\) Then the linear combination of the vectors \(\bfv_1, \bfv_2, \ldots, \bfv_m \) with weights \(c_1, c_2, \ldots, c_m\) is
\begin{equation*} c_1\bfv_1 + c_2\bfv_2 + \cdots + c_m\bfv_m\text{.} \end{equation*}
The span of the vectors \(\bfv_1, \bfv_2, \ldots, \bfv_m\) is the set of all possible linear combinations of \(\bfv_1, \bfv_2, \ldots, \bfv_m\) and is written \(\spn\{\bfv_1, \ldots, \bfv_m\}\text{.}\) In other words, the span is defined to be
\begin{equation*} \spn\{\bfv_1,\ldots,\bfv_m\} = \left\{ \sum_{i=1}^m c_i\bfv_i \mid c_1,\ldots,c_m \in \ff \right\}\text{.} \end{equation*}
The final result in this section is similar to TheoremΒ 2.1.9 in which we summarize some of the β€œobvious” facts which are true in any vector space. We will prove a few of these facts here and leave some others as exercises.

Proof.

We will first prove property 2. Suppose that \(\bfu \in V\) and that both \(\bfv, \bfw \in V\) have the properties of being additive inverses of \(\bfu\text{.}\) We will show that \(\bfv=\bfw\text{.}\)
Since \(\bfu+\bfv = \mathbf{0}\) and \(\bfu+\bfw=\mathbf{0}\text{,}\) we have
\begin{equation*} \bfu+\bfv = \bfu+\bfw\text{.} \end{equation*}
Adding \(\bfv\) to both sides (and using both the associative law and the properties of the zero vector) we have
\begin{align*} (\bfv+\bfu)+\bfv \amp = (\bfv+\bfu)+\bfw \\ \mathbf{0}+\bfv \amp =\mathbf{0}+\bfw \\ \bfv \amp = \bfw \text{.} \end{align*}
We will now show that property 4 holds. Since \(0=0+0\) in \(\ff\text{,}\) we can use the distributive property (axiom 10 in DefinitionΒ 2.3.1) to find the following:
\begin{equation*} 0\bfv = (0+0)\bfv = 0\bfv + 0\bfv\text{.} \end{equation*}
We have \(0\bfv=0\bfv+0\bfv\text{,}\) and subtracting \(0\bfv\) from both sides gives us \(\mathbf{0}=0\bfv\text{.}\) (The reader should recall that by β€œsubtracting” we mean adding the additive inverse, which every vector in a vector space possesses.)
To prove property 6, we must show that \((-1)\bfv\) has the properties of the additive inverse of \(\bfv\text{.}\) (Then, by property 2, we can conclude what we want.) We will use the distributive law (axiom 10 again) as well as the just-proved property 4:
\begin{align*} \bfv + (-1)\bfv \amp =1\bfv + (-1)\bfv\\ \amp =(1-1)\bfv\\ \amp =0\bfv = \mathbf{0}\text{.} \end{align*}

Reading Questions Reading Questions

1.

In this question we will explore \(\ff_5^3\text{.}\) (See ExampleΒ 2.3.4. The vector space \(\ff_5^3\) should be thought of as being the same as \(\rr^3\) except that the entries and the operations are coming from \(\ff_5\text{.}\))
  1. Write down two nonzero elements of \(\ff_5^3\) and compute their sum.
  2. Write down one nonzero element of \(\ff_5^3\) and compute the scalar multiple of this vector by the element \(3\) of \(\ff_5\text{.}\)

3.

Consider the vector \(\bfu = (1,3)\) in \(\ff_5^2\text{.}\) (Again, we are writing the vector horizontally for convenience.) How many elements does \(\mathrm{Span}\{\bfu\}\) contain? Write all of them down.

Exercises Exercises

1.

For each of the following, determine whether the given subset \(V\subseteq \rr^2\) is a real vector space, using the usual operations of vector addition and scalar multiplication in \(\rr^2\text{.}\) (By a β€œreal vector space” we simply mean a vector space over the field \(\rr\text{.}\))
  1. \(\displaystyle V = \{ (x,y) \mid x,y \ge 0 \}\)
  2. \(\displaystyle V = \{ (x,y) \mid xy \ge 0 \}\)
  3. \(\displaystyle V = \{ (x,y) \mid y = x \}\)
Answer.
  1. No, this is not a vector space. For example, it is not closed under scalar multiplication. We can see that \(-2(1,1) = (-2,-2) \not \in V\text{.}\)
  2. No, this is not a vector space. For example, it is not closed under addition. We can see that \((1,4) + (-2,-1)= (-1,3) \not \in V\text{.}\)
  3. Yes, this is a vector space. All of the vector space axioms hold.

2.

Consider the set \(V = \ff_5^2\) with the usual scalar multiplication but with vector addition defined this way:
\begin{equation*} (u_1,v_1) \oplus (u_2,v_2) = (u_1 + u_2, v_1+v_2+1)\text{.} \end{equation*}
This is not a vector space over \(\ff_5\text{.}\) Determine which of the vector space axioms fail, and give an explanation for each such axiom.

3.

Consider the set \(V = \qq^2\) with the usual vector addition but with scalar multiplication defined this way:
\begin{equation*} c \odot (x,y) = (0, cy)\text{.} \end{equation*}
This is not a vector space over \(\qq\text{.}\) Determine which of the vector space axioms fail, and give an explanation for each such axiom.
Answer.
The only axiom that fails is axiom 7. We can see, for example, that \(1 \cdot (1,1) = (0,1) \neq (1,1)\text{,}\) so this axiom does not hold.

4.

Consider the following three elements of \(P_3\text{:}\)
\begin{align*} p_1 \amp = 2t^2 - t + 6\\ p_2 \amp = -t^3 - t^2 + 4t\\ p_3 \amp = 4t^3 + \tfrac{1}{2}t^2 + 2t + 1\text{.} \end{align*}
Calculate the linear combination of \(p_1\text{,}\) \(p_2\text{,}\) and \(p_3\) with weights \(2\text{,}\) \(-3\text{,}\) and \(-2\text{,}\) respectively.
Answer.
We calculate that
\begin{equation*} 2p_1 -3p_2 -2p_3 = -5t^3 + 6t^2 -18t +10\text{.} \end{equation*}

5.

Consider the following three elements of \(\ff_3^2\text{:}\)
\begin{equation*} \bfu = \begin{bmatrix} 1 \\ 2 \end{bmatrix}, \hspace{6pt} \bfu = \begin{bmatrix} 2 \\ 0 \end{bmatrix}, \hspace{6pt} \bfw = \begin{bmatrix} 1 \\ 1 \end{bmatrix}\text{.} \end{equation*}
Calculate the linear combination of \(\bfu\text{,}\) \(\bfv\text{,}\) and \(\bfw\) with weights \(1\text{,}\) \(2\text{,}\) and \(2\text{,}\) respectively.

6.

Consider the vector space \(\ff_2^3\text{.}\)
  1. How many distinct vectors are in \(\ff_2^3\text{?}\)
  2. If \(\bfu\) and \(\bfv\) are the following vectors,
    \begin{equation*} \bfu = \begin{bmatrix} 1 \\ 0 \\ 1 \end{bmatrix}, \hspace{12pt} \bfv = \begin{bmatrix} 0 \\ 1 \\ 1 \end{bmatrix}\text{,} \end{equation*}
    how many distinct vectors are in \(\mathrm{Span}\{\bfu, \bfv\}\text{?}\)
Answer.
  1. This vector space contains \(2^3 = 8\) distinct vectors.
  2. The set \(\spn\{\bfu, \bfv\}\) contains four distinct vectors: \(\bfo\text{,}\) \(\bfu\text{,}\) \(\bfv\text{,}\) and \(\bfu + \bfv\text{.}\) These are all distinct.

7.

Consider the vector space \(V\) of all functions \(\rr \to \rr\text{.}\) (This general setting was introduced in ExampleΒ 2.3.5.) Let \(f\) and \(g\) be the following vectors in \(V\text{:}\)
\begin{equation*} f(t) = \sin^2(t), \hspace{12pt} g(t) = \cos^2(t)\text{.} \end{equation*}
Is the constant function \(h(t) = 4\) an element of \(\mathrm{Span}\{f, g\}\text{?}\) Explain your answer.

8.

Consider the following vectors in \(\ff_5^3\text{:}\)
\begin{equation*} \bfu = \begin{bmatrix} 2 \\ 4 \\ 0 \end{bmatrix}, \hspace{6pt} \bfv = \begin{bmatrix} 3 \\ 0 \\ 2 \end{bmatrix}, \hspace{6pt} \bfw = \begin{bmatrix} 1 \\ 1 \\ 2 \end{bmatrix}\text{.} \end{equation*}
Is \(\bfw \in \mathrm{Span}\{\bfu, \bfv\}\text{?}\) Explain your answer.

9.

Consider the following vectors in \(\cc^3\text{:}\)
\begin{equation*} \bfu = \begin{bmatrix} 2i \\ 1-i \\ 2-i \end{bmatrix}, \hspace{6pt} \bfv = \begin{bmatrix} 1+2i \\ 0 \\ 4+i \end{bmatrix}, \hspace{6pt} \bfw = \begin{bmatrix} 2+2i \\ 2 - i \\ 3i \end{bmatrix}\text{.} \end{equation*}
Is \(\bfw \in \mathrm{Span}\{\bfu, \bfv\}\text{?}\) Explain your answer.
Answer.
No, \(\bfw\) is not in \(\spn\{\bfu,\bfv\}\text{.}\) When we form the matrix with \(\bfu\text{,}\) \(\bfv\text{,}\) and \(\bfw\) as its columns, it reduces to the following:
\begin{equation*} \begin{bmatrix} 1 \amp 0 \amp 0 \\ 0 \amp 1 \amp 0 \\ 0 \amp 0 \amp 1 \end{bmatrix}\text{.} \end{equation*}
Since there is a pivot in the final column, the corresponding vector equation \(x_1\bfu + x_2\bfv = \bfw\) has no solutions.

Writing Exercises

10.
Consider the following subset of \(\rr^2\text{:}\)
\begin{equation*} V = \{ (x,y) \mid y = 4x \}\text{.} \end{equation*}
Prove that \(V\text{,}\) with the usual operations of vector addition and scalar multiplication in \(\rr^2\text{,}\) is a vector space over \(\rr\text{.}\) (Hint: Since \(V\) is a subset of \(\rr^2\) with the same operations as \(\rr^2\text{,}\) some of the axioms may not need to be proved from scratch.)
11.
Is \(\rr\) a vector space over \(\qq\text{?}\) (Addition and scalar multiplication should be understood as the obvious operations in \(\rr\text{.}\)) Explain your answer.
12.
Is \(\qq\) a vector space over \(\rr\text{?}\) (Addition and scalar multiplication should be understood as the obvious operations in \(\qq\text{.}\)) Explain your answer.
Answer.
No, \(\qq\) is not a vector space over \(\rr\text{.}\) It is not closed under scalar multiplication. For example, we see that \(\pi (1,1) = (\pi,\pi) \not \in \qq^2\text{.}\)
13.
Prove that \(\cc\) is a vector space over \(\rr\text{.}\) (Addition and scalar multiplication should be understood as the obvious operations in \(\cc\text{.}\))
14.
Prove that every vector space over \(\cc\) is a vector space over \(\rr\text{.}\)
15.
Solution.
Suppose that \(\bfo\) is a zero vector in \(V\) and that \(\mathbf{z}\) also has the properties of a zero vector. Since \(\bfo\) is a zero vector, we have \(\bfo + \mathbf{z} = \mathbf{z}\text{.}\) Since \(\mathbf{z}\) has the properties of a zero vector, we have \(\bfo + \mathbf{z} = \bfo\text{.}\) Therefore, we have
\begin{equation*} \mathbf{z} = \bfo + \mathbf{z} = \bfo\text{,} \end{equation*}
so \(\mathbf{z} = \bfo\text{.}\) This proves that the zero vector is unique.