Let
\(\ff\) be a field and let
\(A\) be an
\(m\times n\) matrix with entries from
\(\ff\text{.}\) (We will refer to this in what follows as βa matrix over
\(\ff\text{.}\)β) Then multiplication by
\(A\) is a linear transformation from
\(\ff^n\) to
\(\ff^m\text{.}\) (We will denote the function which is multiplication by
\(A\) by
\(T_A:\ff^n \to \ff^m\text{.}\))
To justify this claim we must first explain what we mean by βmultiplication by \(A\text{.}\)β We will let \(\bfv \in \ff^n\) and denote entry \((i,j)\) in \(A\) by \(a_{ij}\text{.}\) We will further denote the entries of \(\bfv\) by
\begin{equation*}
\bfv = \left[\begin{array}{@{}c@{}}
v_1 \\ \vdots \\ v_n
\end{array}\right]\text{.}
\end{equation*}
Then the matrix-vector product \(A\bfv\) is defined to be the following vector in \(\ff^m\text{:}\)
\begin{equation}
A\bfv = \left[\begin{array}{@{}c@{}}
a_{11}v_1 + \cdots + a_{1n}v_n \\ \vdots \\
a_{m1}v_1 + \cdots + a_{mn}v_n
\end{array}\right]\text{.}\tag{3.1}
\end{equation}
One way to state this is that entry \(j\) in \(A\bfv\) is the sum of the entry-wise product of row \(j\) in \(A\) with \(\bfv\text{.}\) Since \(A\bfv\) is an element of \(\ff^m\text{,}\) the domain and codomain of \(T_A\) are correct.
What we have defined is the product of a matrix and a vector. However, an alternate description of this product will be more useful in proving that
\(T_A\) is a linear transformation.
If the columns of \(A\) are thought of as vectors \(\mathbf{a}_1, \ldots, \mathbf{a}_n\text{,}\) then the product \(A\bfv\) is also
\begin{equation}
A\bfv = v_1\mathbf{a}_1 + \cdots + v_n\mathbf{a}_n = \sum_{i=1}^n v_i\mathbf{a}_i\text{.}\tag{3.2}
\end{equation}
In words,
\(A\bfv\) is a linear combination of the columns of
\(A\) with weights coming from the entries of
\(\bfv\text{.}\) (We have reserved proving the equivalence of these two formulations to
ExerciseΒ 3.1.7.16.)
With this equivalent definition, proving that \(T_A\) is a linear transformation is a snap. Let \(\bfu\) and \(\bfv\) be vectors in \(\ff^n\) and let \(c \in \ff\text{.}\) We will further denote the entries of \(\bfu\) and \(\bfv\) by
\begin{equation*}
\bfu = \left[\begin{array}{@{}c@{}}
u_1 \\ \vdots \\ u_n
\end{array}\right] \hspace{.3in} \text{and} \hspace{.3in} \bfv = \left[\begin{array}{@{}c@{}}
v_1 \\ \vdots \\ v_n
\end{array}\right]\text{.}
\end{equation*}
Then we have the following:
\begin{align*}
T_A(\bfu + \bfv) \amp = A(\bfu+\bfv) = \sum_{i=1}^n(u_i+v_i)\mathbf{a}_i\\
T_A(\bfu) + T_A(\bfv) \amp = A\bfu + A\bfv = \sum_{i=1}^n u_i\mathbf{a}_i + \sum_{i=1}^n v_i\mathbf{a}_i\text{.}
\end{align*}
These two expressions are equal due to the fact that \(\ff^m\) is a vector space.
We have one final calculation to prove that \(T_A\) is a linear transformation. Let \(\bfv\) be a vector in \(\ff^n\) and let \(c\) be in \(\ff\text{.}\) Then we have
\begin{align*}
T_A(c\bfv) \amp = A(c\bfv) = \sum_{i=1}^n (cv_i)\mathbf{a}_i\\
cT_A(\bfv) \amp = c A(\bfv) = c \sum_{i=1}^n v_i\mathbf{a}_i.
\end{align*}
Once again, thsse expressions are equal because \(\ff^m\) is a vector space.
These calculations prove that
\(T_A\) is a linear transformation.