2.2: Matrix multiplication and linear combinations
\(\newcommand{\zerovec}{\mathbf 0}\) \(\newcommand{\twovec}[2]{\begin{pmatrix} #1 \\ #2 \end{pmatrix} } \) \(\newcommand{\threevec}[3]{\begin{pmatrix} #1 \\ #2 \\ #3 \end{pmatrix} } \) \(\newcommand{\fourvec}[4]{\begin{pmatrix} #1 \\ #2 \\ #3 \\ #4 \end{pmatrix} } \) \(\newcommand{\fivevec}[5]{\begin{pmatrix} #1 \\ #2 \\ #3 \\ #4 \\ #5 \end{pmatrix} } \)The previous section introduced vectors and linear combinations and demonstrated how they provide a means of thinking about linear systems geometrically. In particular, we saw that the vector \(\mathbf b\) is a linear combination of the vectors \(\mathbf v_1,\mathbf v_2,\ldots,\mathbf v_n\) if the linear system corresponding to the augmented matrix
is consistent.
Our goal in this section is to introduction matrix multiplication, another algebraic operation that connects linear systems and linear combinations.
2.2.1 Matrices
We first thought of a matrix as a rectangular array of numbers. When the number of rows is \(m\) and columns is \(n\text{,}\) we say that the dimensions of the matrix are \(m\times n\text{.}\) For instance, the matrix below is a \(3\times4\) matrix:
We may also think of the columns of a matrix as a collection of vectors. For instance, the matrix above may be represented as
where
In this way, we see that our \(3\times 4\) matrix is the same as a collection of 4 vectors in \(\mathbb R^3\text{.}\)
This means that we may define scalar multiplication and matrix addition operations using the corresponding vector operations.
Preview Activity 2.2.1. Matrix operations.
-
Compute the scalar multiple
\begin{equation*} -3\left[ \begin{array}{rrr} 3 & 1 & 0 \\ -4 & 3 & -1 \\ \end{array} \right]\text{.} \end{equation*}
- Suppose that \(A\) and \(B\) are two matrices. What do we need to know about their dimensions before we can form the sum \(A+B\text{?}\)
-
Find the sum
\begin{equation*} \left[ \begin{array}{rr} 0 & -3 \\ 1 & -2 \\ 3 & 4 \\ \end{array} \right] + \left[ \begin{array}{rrr} 4 & -1 \\ -2 & 2 \\ 1 & 1 \\ \end{array} \right]\text{.} \end{equation*}
-
The matrix \(I_n\text{,}\) which we call the
identity
matrix is the \(n\times n\) matrix whose entries are zero except for the diagonal entries, which are 1. For instance,
\begin{equation*} I_3 = \left[ \begin{array}{rrr} 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 1 \\ \end{array} \right]\text{.} \end{equation*}
If we can form the sum \(A+I_n\text{,}\) what must be true about the matrix \(A\text{?}\)
-
Find the matrix \(A - 2I_3\) where
\begin{equation*} A = \left[ \begin{array}{rrr} 1 & 2 & -2 \\ 2 & -3 & 3 \\ -2 & 3 & 4 \\ \end{array} \right]\text{.} \end{equation*}
As this preview activity shows, both of these operations are relatively straightforward. Some care, however, is required when adding matrices. Since we need the same number of vectors to add and since the vectors must be of the same dimension, two matrices must have the same dimensions as well if we wish to form their sum.
The identity matrix will play an important role at various points in our explorations. It is important to note that it is a square matrix, meaning it has an equal number of rows and columns, so any matrix added to it must be square as well. Though we wrote it as \(I_n\) in the activity, we will often just write \(I\) when the dimensions are clear.
2.2.2 Matrix-vector multiplication and linear combinations
A more important operation will be matrix multiplication as it allows us to compactly express linear systems. For now, we will work with the product of a matrix and vector, which we illustrate with an example.
Suppose we have the matrix \(A\) and vector \(\mathbf x\) as given below.
Their product will be defined to be the linear combination of the columns of \(A\) using the components of \(\mathbf x\) as weights. This means that
Let's take note of the dimensions of the matrix and vectors. The two components of the vector \(\mathbf x\) are weights used to form a linear combination of the columns of \(A\text{.}\) Since \(\mathbf x\) has two components, \(A\) must have two columns. In other words, the number of columns of \(A\) must equal the dimension of the vector \(\mathbf x\text{.}\)
In the same way, the columns of \(A\) are 3-dimensional so any linear combination of them is 3-dimensional as well. Therefore, \(A\mathbf x\) will be 3-dimensional.
We then see that if \(A\) is a \(3\times2\) matrix, \(\mathbf x\) must be a 2-dimensional vector and \(A\mathbf x\) will be 3-dimensional.
More generally, we have the following definition.
The product of a matrix \(A\) by a vector \(\mathbf x\) will be the linear combination of the columns of \(A\) using the components of \(\mathbf x\) as weights.
If \(A\) is an \(m\times n\) matrix, then \(\mathbf x\) must be an \(n\)-dimensional vector, and the product \(A\mathbf x\) will be an \(m\)-dimensional vector. If
then
The next activity introduces some properties of matrix multiplication.
Activity 2.2.2. Matrix-vector multiplication.
-
Find the matrix product
\begin{equation*} \left[ \begin{array}{rrrr} 1 & 2 & 0 & -1 \\ 2 & 4 & -3 & -2 \\ -1 & -2 & 6 & 1 \\ \end{array} \right] \left[ \begin{array}{r} 3 \\ 1 \\ -1 \\ 1 \\ \end{array} \right]\text{.} \end{equation*}
-
Suppose that \(A\) is the matrix
\begin{equation*} \left[ \begin{array}{rrr} 3 & -1 & 0 \\ 0 & -2 & 4 \\ 2 & 1 & 5 \\ 1 & 0 & 3 \\ \end{array} \right]\text{.} \end{equation*}
If \(A\mathbf x\) is defined, what is the dimension of the vector \(\mathbf x\) and what is the dimension of \(A\mathbf x\text{?}\)
- A vector whose entries are all zero is denoted by \(\zerovec\text{.}\) If \(A\) is a matrix, what is the product \(A\zerovec\text{?}\)
- Suppose that \(I = \left[\begin{array}{rrr} 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 1 \\ \end{array}\right]\) is the identity matrix and \(\mathbf x=\threevec{x_1}{x_2}{x_3}\text{.}\) Find the product \(I\mathbf x\) and explain why \(I\) is called the identity matrix.
-
Suppose we write the matrix \(A\) in terms of its columns as
\begin{equation*} A = \left[ \begin{array}{rrrr} \mathbf v_1 & \mathbf v_2 & \ldots & \mathbf v_n \\ \end{array} \right]\text{.} \end{equation*}
If the vector \(\mathbf e_1 = \left[\begin{array}{r} 1 \\ 0 \\ \vdots \\ 0 \end{array}\right]\text{,}\) what is the product \(A\mathbf e_1\text{?}\)
-
Suppose that
\begin{equation*} A = \left[ \begin{array}{rrrr} 1 & 2 \\ -1 & 1 \\ \end{array} \right], \mathbf b = \left[ \begin{array}{r} 6 \\ 0 \end{array} \right]\text{.} \end{equation*}
Is there a vector \(\mathbf x\) such that \(A\mathbf x = \mathbf b\text{?}\)
Multiplication of a matrix \(A\) and a vector is defined as a linear combination of the columns of \(A\text{.}\) However, there is a shortcut for computing such a product. Let's look at our previous example and focus on the first row of the product.
To find the first component of the product, we consider the first row of the matrix. We then multiply the first entry in that row by the first component of the vector, the second entry by the second component of the vector, and so on, and add the results. In this way, we see that the third component of the product would be obtained from the third row of the matrix by computing \(2(3) + 3(1) = 9\text{.}\)
You are encouraged to evaluate Item a using this shortcut and compare the result to what you found while completing the previous activity.
Activity 2.2.3.
In addition, Sage can find the product of a matrix and vector using the
*
operator. For example,
A = matrix(2,2,[1,2,2,1]) v = vector([3,-1]) A*v
-
Use Sage to evaluate the product Item a yet again.
-
In Sage, define the matrix and vectors
\begin{equation*} A = \left[ \begin{array}{rrr} -2 & 0 \\ 3 & 1 \\ 4 & 2 \\ \end{array} \right], \zerovec = \left[ \begin{array}{r} 0 \\ 0 \end{array} \right], \mathbf v = \left[ \begin{array}{r} -2 \\ 3 \end{array} \right], \mathbf w = \left[ \begin{array}{r} 1 \\ 2 \end{array} \right]\text{.} \end{equation*}
- What do you find when you evaluate \(A\zerovec\text{?}\)
- What do you find when you evaluate \(A(3\mathbf v)\) and \(3(A\mathbf v)\) and compare your results?
- What do you find when you evaluate \(A(\mathbf v+\mathbf w)\) and \(A\mathbf v + A\mathbf w\) and compare your results?
- If \(I=\left[\begin{array}{rrr} 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 1 \\ \end{array}\right]\) is the \(3\times3\) identity matrix, what is the product \(IA\text{?}\)
This activity demonstrates several general properties satisfied by matrix multiplication that we record here.
If \(A\) is a matrix, \(\mathbf v\) and \(\mathbf w\) vectors, and \(c\) a scalar, then
- \(A\zerovec = \zerovec\text{.}\)
- \(A(c\mathbf v) = cA\mathbf v\text{.}\)
- \(A(\mathbf v+\mathbf w) = A\mathbf v + A\mathbf w\text{.}\)
2.2.3 Matrix-vector multiplication and linear systems
So far, we have begun with a matrix \(A\) and a vector \(\mathbf x\) and formed their product \(A\mathbf x = \mathbf b\text{.}\) We would now like to turn this around: beginning with a matrix \(A\) and a vector \(\mathbf b\text{,}\) we will ask if we can find a vector \(\mathbf x\) such that \(A\mathbf x = \mathbf b\text{.}\) This will naturally lead back to linear systems.
To see the connection between the matrix equation \(A\mathbf x = \mathbf b\) and linear systems, let's write the matrix \(A\) in terms of its columns \(\mathbf v_i\) and \(\mathbf x\) in terms of its components.
We know that the matrix product \(A\mathbf x\) forms a linear combination of the columns of \(A\text{.}\) Therefore, the equation \(A\mathbf x = \mathbf b\) is merely a compact way of writing the equation for the weights \(c_i\text{:}\)
We have seen this equation before: Remember that Proposition 2.1.7 says that the solutions of this equation are the same as the solutions to the linear system whose augmented matrix is
This gives us three different ways of looking at the same solution space.
If \(A=\left[\begin{array}{rrrr} \mathbf v_1& \mathbf v_2& \ldots\mathbf v_n \end{array}\right]\) and \(\mathbf x=\left[ \begin{array}{r} x_1 \\ x_2 \\ \vdots \\ x_n \\ \end{array}\right] \text{,}\) then the following are equivalent.
- The vector \(\mathbf x\) satisfies \(A\mathbf x = \mathbf b \text{.}\)
-
The vector \(\mathbf b\) is a linear combination of the columns of \(A\) with weights \(x_j\text{:}\)
\begin{equation*} x_1\mathbf v_1 + x_2\mathbf v_2 + \ldots + x_n\mathbf v_n = \mathbf b\text{.} \end{equation*}
-
The components of \(\mathbf x\) form a solution to the linear system corresponding to the augmented matrix
\begin{equation*} \left[\begin{array}{rrrr|r} \mathbf v_1 & \mathbf v_2 & \ldots & \mathbf v_n & \mathbf b \end{array}\right]\text{.} \end{equation*}
When the matrix \(A = \left[\begin{array}{rrrr} \mathbf v_1& \mathbf v_2& \ldots& \mathbf v_n\end{array}\right]\text{,}\) we will frequently write
and say that we augment the matrix \(A\) by the vector \(\mathbf b\text{.}\)
We may think of \(A\mathbf x = \mathbf b\) as merely giving a notationally compact way of writing a linear system. This form of the equation, however, will allow us to focus on important features of the system that determine its solution space.
Describe the solution space of the equation
By Proposition 2.2.4, the solution space to this equation is the same as the equation
which is the same as the linear system corresponding to
We will study the solutions to this linear system by finding the reduced row echelon form of the augmented matrix:
This gives us the system of equations
The variable \(x_3\) is free so we may write the solution space parametrically as
Since we originally asked to describe the solutions to the equation \(A\mathbf x = \mathbf b\text{,}\) we will express the solution in terms of the vector \(\mathbf x\text{:}\)
This shows that the solutions \(\mathbf x\) may be written in the form \(\mathbf v + x_3\mathbf w\text{,}\) for appropriate vectors \(\mathbf v\) and \(\mathbf w\text{.}\) Geometrically, the solution space is a line in \(\mathbb R^3\) through \(\mathbf v\) moving parallel to \(\mathbf w\text{.}\)
Activity 2.2.4. The equation \(A\mathbf x = \mathbf b\).
-
Consider the linear system
\begin{equation*} \begin{alignedat}{4} 2x & {}+{} & y & {}-{} & 3z & {}={} & 4 \\ -x & {}+{} & 2y & {}+{} & z & {}={} & 3 \\ 3x & {}-{} & y & & & {}={} & -4 \\ \end{alignedat}\text{.} \end{equation*}
Identify the matrix \(A\) and vector \(\mathbf b\) to express this system in the form \(A\mathbf x = \mathbf b\text{.}\)
-
If \(A\) and \(\mathbf b\) are as below, write the linear system corresponding to the equation \(A\mathbf x=\mathbf b\text{.}\)
\begin{equation*} A = \left[\begin{array}{rrr} 3 & -1 & 0 \\ -2 & 0 & 6 \end{array} \right], \mathbf b = \left[\begin{array}{r} -6 \\ 2 \end{array} \right] \end{equation*}
and describe the solution space.
-
Describe the solution space of the equation
\begin{equation*} \left[ \begin{array}{rrrr} 1 & 2 & 0 & -1 \\ 2 & 4 & -3 & -2 \\ -1 & -2 & 6 & 1 \\ \end{array} \right] \mathbf x = \left[\begin{array}{r} -1 \\ 1 \\ 5 \end{array} \right]\text{.} \end{equation*}
- Suppose \(A\) is an \(m\times n\) matrix. What can you guarantee about the solution space of the equation \(A\mathbf x = \zerovec\text{?}\)
2.2.4 Matrix products
In this section, we have developed some algebraic operations on matrices with the aim of simplifying our description of linear systems. We will now introduce a final operation, the product of two matrices, that will become important when we study linear transformations in Section 2.5 .
Given matrices \(A\) and \(B\text{,}\) we will form their product \(AB\) by first writing \(B\) in terms of its columns:
We then define
Given the matrices
we have
It is important to note that we can only multiply matrices if the dimensions of the matrices are compatible. More specifically, when constructing the product \(AB\text{,}\) the matrix \(A\) multiplies the columns of \(B\text{.}\) Therefore, the number of columns of \(A\) must equal the number of rows of \(B\text{.}\) When this condition is met, the number of rows of \(AB\) is the number of rows of \(A\text{,}\) and the number of columns of \(AB\) is the number of columns of \(B\text{.}\)
Activity 2.2.5.
Consider the matrices
- Suppose we want to form the product \(AB\text{.}\) Before computing, first explain how you know this product exists and then explain what the dimensions of the resulting matrix will be.
- Compute the product \(AB\text{.}\)
-
Sage can multiply matrices using the
*
operator. Define the matrices \(A\) and \(B\) in the Sage cell below and check your work by computing \(AB\text{.}\) - Are you able to form the matrix product \(BA\text{?}\) If so, use the Sage cell above to find \(BA\text{.}\) Is it generally true that \(AB = BA\text{?}\)
-
Suppose we form the three matrices.
\begin{equation*} A = \left[\begin{array}{rr} 1 & 2 \\ 3 & -2 \\ \end{array}\right], B = \left[\begin{array}{rr} 0 & 4 \\ 2 & -1 \\ \end{array}\right], C = \left[\begin{array}{rr} -1 & 3 \\ 4 & 3 \\ \end{array}\right]\text{.} \end{equation*}
Compare what happens when you compute \(A(B+C)\) and \(AB + AC\text{.}\) State your finding as a general principle.
- Compare the results of evaluating \(A(BC)\) and \((AB)C\) and state your finding as a general principle.
-
When we are dealing with real numbers, we know if \(a\neq 0\) and \(ab = ac\text{,}\) then \(b=c\text{.}\) Define matrices
\begin{equation*} A = \left[\begin{array}{rr} 1 & 2 \\ -2 & -4 \\ \end{array}\right], B = \left[\begin{array}{rr} 3 & 0 \\ 1 & 3 \\ \end{array}\right], C = \left[\begin{array}{rr} 1 & 2 \\ 2 & 2 \\ \end{array}\right] \end{equation*}
and compute \(AB\) and \(AC\text{.}\)
If \(AB = AC\text{,}\) is it necessarily true that \(B = C\text{?}\) -
Again, with real numbers, we know that if \(ab = 0\text{,}\) then either \(a = 0\) or \(b=0\text{.}\) Define
\begin{equation*} A = \left[\begin{array}{rr} 1 & 2 \\ -2 & -4 \\ \end{array}\right], B = \left[\begin{array}{rr} 2 & -4 \\ -1 & 2 \\ \end{array}\right] \end{equation*}
and compute \(AB\text{.}\)
If \(AB = 0\text{,}\) is it necessarily true that either \(A=0\) or \(B=0\text{?}\)
This activity demonstrated some general properties about products of matrices, which mirror some properties about operations with real numbers.
Properties of Matrix-matrix Multiplication.
If \(A\text{,}\) \(B\text{,}\) and \(C\) are matrices such that the following operations are defined, it follows that
- Associativity:
-
\(A(BC) = (AB)C\text{.}\)
- Distributivity:
-
\(A(B+C) = AB+AC\text{.}\)
\((A+B)C = AC+BC\text{.}\)
At the same time, there are a few properties that hold for real numbers that do not hold for matrices.
Things to be careful of.
The following properties hold for real numbers but not for matrices.
- Commutativity:
-
It is not generally true that \(AB = BA\text{.}\)
- Cancellation:
-
It is not generally true that \(AB = AC\) implies that \(B = C\text{.}\)
- Zero divisors:
-
It is not generally true that \(AB = 0\) implies that either \(A=0\) or \(B=0\text{.}\)
Summary
In this section, we have found an especially simple way to express linear systems using matrix multiplication.
- If \(A\) is an \(m\times n\) matrix and \(\mathbf x\) an \(n\)-dimensional vector, then \(A\mathbf x\) is the linear combination of the columns of \(A\) using the components of \(\mathbf x\) as weights. The vector \(A\mathbf x\) is \(m\)-dimensional.
- The solution space to the equation \(A\mathbf x = \mathbf b\) is the same as the solution space to the linear system corresponding to the augmented matrix \(\left[ \begin{array}{r|r} A & \mathbf b \end{array}\right]\text{.}\)
- If \(A\) is an \(m\times n\) matrix and \(B\) is an \(n\times p\) matrix, we can form the product \(AB\text{,}\) which is an \(m\times p\) matrix whose columns are the products of \(A\) and the columns of \(B\text{.}\)
Exercises 2.2.6Exercises
Consider the system of linear equations
- Find the matrix \(A\) and vector \(\mathbf b\) that expresses this linear system in the form \(A\mathbf x=\mathbf b\text{.}\)
-
Give a description of the solution space to the equation \(A\mathbf x = \mathbf b\text{.}\)
Suppose that \(A\) is a \(135\times2201\) matrix. If \(A\mathbf x\) is defined, what is the dimension of \(\mathbf x\text{?}\) What is the dimension of \(A\mathbf x\text{?}\)
Suppose that \(A \) is a \(3\times2\) matrix whose columns are \(\mathbf v_1\) and \(\mathbf v_2\text{;}\) that is,
- What is the dimension of the vectors \(\mathbf v_1\) and \(\mathbf v_2\text{?}\)
- What is the product \(A\twovec{1}{0}\) in terms of \(\mathbf v_1\) and \(\mathbf v_2\text{?}\) What is the product \(A\twovec{0}{1}\text{?}\) What is the product \(A\twovec{2}{3}\text{?}\)
-
Suppose that
\begin{equation*} A\twovec{1}{0} = \threevec{3}{-2}{1}, A\twovec{0}{1} = \threevec{0}{3}{2}\text{.} \end{equation*}
What is the matrix \(A\text{?}\)
Shown below are vectors \(\mathbf v_1\) and \(\mathbf v_2\text{.}\) Suppose that the matrix \(A\) is
- What are the dimensions of the matrix \(A\text{?}\)
-
On the plot above, indicate the vectors
\begin{equation*} A\twovec{1}{0}, A\twovec{2}{3}, A\twovec{0}{-3}\text{.} \end{equation*}
- Find all vectors \(\mathbf x\) such that \(A\mathbf x=\mathbf b\text{.}\)
- Find all vectors \(\mathbf x\) such that \(A\mathbf x = \zerovec\text{.}\)
Suppose that
- Describe the solution space to the equation \(A\mathbf x = \zerovec\text{.}\)
- Find a \(3\times2\) matrix \(B\) with no zero entries such that \(AB = 0\text{.}\)
Consider the matrix
-
Find the product \(A\mathbf x\) where
\begin{equation*} \mathbf x = \fourvec{1}{-2}{0}{2}\text{.} \end{equation*}
-
Give a description of the vectors \(\mathbf x\) such that
\begin{equation*} A\mathbf x = \threevec{-1}{15}{17}\text{.} \end{equation*}
- Find the reduced row echelon form of \(A\) and identify the pivot positions.
- Can you find a vector \(\mathbf b\) such that \(A\mathbf x=\mathbf b\) is inconsistent?
- For a general 3-dimensional vector \(\mathbf b\text{,}\) what can you say about the solution space of the equation \(A\mathbf x = \mathbf b\text{?}\)
The operations that we perform in Gaussian elimination can be accomplished using matrix multiplication. This observation is the basis of an important technique that we will investigate in a subsequent chapter.
Let's consider the matrix
-
Suppose that
\begin{equation*} S = \left[\begin{array}{rrr} 1 & 0 & 0 \\ 0 & 7 & 0 \\ 0 & 0 & 1 \\ \end{array}\right]\text{.} \end{equation*}
Verify that \(SA\) is the matrix that results when the second row of \(A\) is scaled by a factor of 7. What matrix \(S\) would scale the third row by -3?
-
Suppose that
\begin{equation*} P = \left[\begin{array}{rrr} 0 & 1 & 0 \\ 1 & 0 & 0 \\ 0 & 0 & 1 \\ \end{array}\right]\text{.} \end{equation*}
Verify that \(PA\) is the matrix that results from interchanging the first and second rows. What matrix \(P\) would interchange the first and third rows?
-
Suppose that
\begin{equation*} L_1 = \left[\begin{array}{rrr} 1 & 0 & 0 \\ -2 & 1 & 0 \\ 0 & 0 & 1 \\ \end{array}\right]\text{.} \end{equation*}
Verify that \(L_1A\) is the matrix that results from multiplying the first row of \(A\) by \(-2\) and adding it to the second row. What matrix \(L_2\) would multiply the first row by 3 and add it to the third row?
-
When we performed Gaussian elimination, our first goal was to perform row operations that brought the matrix into a triangular form. For our matrix \(A\text{,}\) find the row operations needed to find a row equivalent matrix \(U\) in triangular form. By expressing these row operations in terms of matrix multiplication, find a matrix \(L\) such that \(LA = U\text{.}\)
In this exercise, you will construct the inverse of a matrix, a subject that we will investigate more fully in the next chapter. Suppose that \(A\) is the \(2\times2\) matrix:
-
Find the vectors \(\mathbf b_1\) and \(\mathbf b_2\) such that the matrix \(B=\left[\begin{array}{rr} \mathbf b_1 & \mathbf b_2 \end{array}\right]\) satisfies
\begin{equation*} AB = I = \left[\begin{array}{rr} 1 & 0 \\ 0 & 1 \\ \end{array}\right]\text{.} \end{equation*}
-
In general, it is not true that \(AB = BA\text{.}\) Check that it is true, however, for the specific \(A\) and \(B\) that appear in this problem.
- Suppose that \(\mathbf x = \twovec{x_1}{x_2}\text{.}\) What do you find when you evaluate \(I\mathbf x\text{?}\)
-
Suppose that we want to solve the equation \(A\mathbf x = \mathbf b\text{.}\) We know how to do this using Gaussian elimination; let's use our matrix \(B\) to find a different way:
\begin{equation*} \begin{aligned} A\mathbf x & {}={} \mathbf b \\ B(A\mathbf x) & {}={} B\mathbf b \\ (BA)\mathbf x & {}={} B\mathbf b \\ I\mathbf x & {}={} B\mathbf b \\ \mathbf x & {}={} B\mathbf b \\ \end{aligned}\text{.} \end{equation*}
In other words, the solution to the equation \(A\mathbf x=\mathbf b\) is \(\mathbf x = B\mathbf b\text{.}\)
Consider the equation \(A\mathbf x = \twovec{5}{-2}\text{.}\) Find the solution in two different ways, first using Gaussian elimination and then as \(\mathbf x = B\mathbf b\text{,}\) and verify that you have found the same result.
Determine whether the following statements are true or false and provide a justification for your response.
- If \(A\mathbf x\) is defined, then the number of components of \(\mathbf x\) equals the number of rows of \(A\text{.}\)
- The solution space to the equation \(A\mathbf x = \mathbf b\) is equivalent to the solution space to the linear system whose augmented matrix is \(\left[\begin{array}{r|r} A & \mathbf b \end{array}\right]\text{.}\)
- If a linear system of equations has 8 equations and 5 unknowns, then the dimensions of the matrix \(A\) in the corresponding equation \(A\mathbf x = \mathbf b\) is \(5\times8\text{.}\)
- If \(A\) has a pivot in every row, then every equation \(A\mathbf x = \mathbf b\) is consistent.
- If \(A\) is a \(9\times5\) matrix, then \(A\mathbf x=\mathbf b\) is inconsistent for some vector \(\mathbf b\text{.}\)
Suppose that \(A\) is an \(4\times4\) matrix and that the equation \(A\mathbf x = \mathbf b\) has a unique solution for some vector \(\mathbf b\text{.}\)
- What does this say about the pivots of the matrix \(A\text{?}\) Write the reduced row echelon form of \(A\text{.}\)
- Can you find another vector \(\mathbf c\) such that \(A\mathbf x = \mathbf c\) is inconsistent?
- What can you say about the solution space to the equation \(A\mathbf x = \zerovec\text{?}\)
- Suppose \(A=\left[\begin{array}{rrrr} \mathbf v_1 & \mathbf v_2 & \mathbf v_3 & \mathbf v_4 \end{array}\right]\text{.}\) Explain why every four-dimensional vector can be written as a linear combination of the vectors \(\mathbf v_1\text{,}\) \(\mathbf v_2\text{,}\) \(\mathbf v_3\text{,}\) and \(\mathbf v_4\) in exactly one way.
Define the matrix
-
Describe the solution space to the homogeneous equation \(A\mathbf x = \zerovec\text{.}\) What does this solution space represent geometrically?
-
Describe the solution space to the equation \(A\mathbf x=\mathbf b\) where \(\mathbf b = \threevec{-3}{-4}{1}\text{.}\) What does this solution space represent geometrically and how does it compare to the previous solution space?
-
We will now explain the relationship between the previous two solution spaces. Suppose that \(\mathbf x_h\) is a solution to the homogeneous equation; that is \(A\mathbf x_h=\zerovec\text{.}\) We will also suppose that \(\mathbf x_p\) is a solution to the equation \(A\mathbf x = \mathbf b\text{;}\) that is, \(A\mathbf x_p=\mathbf b\text{.}\)
Use the Linearity Principle expressed in Proposition 2.2.3 to explain why \(\mathbf x_h+\mathbf x_p\) is a solution to the equation \(A\mathbf x = \mathbf b\text{.}\) You may do this by evaluating \(A(\mathbf x_h+\mathbf x_p)\text{.}\)
That is, if we find one solution \(\mathbf x_p\) to an equation \(A\mathbf x = \mathbf b\text{,}\) we may add any solution to the homogeneous equation to \(\mathbf x_p\) and still have a solution to the equation \(A\mathbf x = \mathbf b\text{.}\) In other words, the solution space to the equation \(A\mathbf x = \mathbf b\) is given by translating the solution space to the homogeneous equation by the vector \(\mathbf x_p\text{.}\)
Suppose that a city is starting a bicycle sharing program with bicycles at locations \(B\) and \(C\text{.}\) Bicycles that are rented at one location may be returned to either location at the end of the day. Over time, the city finds that 80% of bicycles rented at location \(B\) are returned to \(B\) with the other 20% returned to \(C\text{.}\) Similarly, 50% of bicycles rented at location \(C\) are returned to \(B\) and 50% to \(C\text{.}\)
To keep track of the bicycles, we form a vector
where \(B_k\) is the number of bicycles at location \(B\) at the beginning of day \(k\) and \(C_k\) is the number of bicycles at \(C\text{.}\) The information above tells us
where
-
Let's check that this makes sense.
- Suppose that there are 1000 bicycles at location \(B\) and none at \(C\) on day 1. This means we have \(\mathbf x_1 = \twovec{1000}{0}\text{.}\) Find the number of bicycles at both locations on day 2 by evaluating \(\mathbf x_2 = A\mathbf x_1\text{.}\)
- Suppose that there are 1000 bicycles at location \(C\) and none at \(B\) on day 1. Form the vector \(\mathbf x_1\) and determine the number of bicycles at the two locations the next day by finding \(\mathbf x_2 = A\mathbf x_1\text{.}\)
- Suppose that one day there are 1050 bicycles at location \(B\) and 450 at location \(C\text{.}\) How many bicycles were there at each location the previous day?
- Suppose that there are 500 bicycles at location \(B\) and 500 at location \(C\) on Monday. How many bicycles are there at the two locations on Tuesday? on Wednesday? on Thursday?
This problem is a continuation of the previous problem.
-
Let us define vectors
\begin{equation*} \mathbf v_1 = \twovec{5}{2}, \mathbf v_2 = \twovec{-1}{1}\text{.} \end{equation*}
Show that
\begin{equation*} A\mathbf v_1 = \mathbf v_1, A\mathbf v_2 = 0.3\mathbf v_2\text{.} \end{equation*} -
Suppose that \(\mathbf x_1 = c_1 \mathbf v_1 + c_2 \mathbf v_2\) where \(c_2\) and \(c_2\) are scalars. Use the Linearity Principle expressed in Proposition 2.2.3 to explain why
\begin{equation*} \mathbf x_{2} = A\mathbf x_1 = c_1\mathbf v_1 + 0.3c_2\mathbf v_2\text{.} \end{equation*}
-
Continuing in this way, explain why
\begin{equation*} \begin{aligned} \mathbf x_{3} = A\mathbf x_2 & {}={} c_1\mathbf v_1 +0.3^2c_2\mathbf v_2 \\ \mathbf x_{4} = A\mathbf x_3 & {}={} c_1\mathbf v_1 +0.3^3c_2\mathbf v_2 \\ \mathbf x_{5} = A\mathbf x_4 & {}={} c_1\mathbf v_1 +0.3^4c_2\mathbf v_2 \\ \end{aligned}\text{.} \end{equation*}
- Suppose that there are initially 500 bicycles at location \(B\) and 500 at location \(C\text{.}\) Write the vector \(\mathbf x_1\) and find the scalars \(c_1\) and \(c_2\) such that \(\mathbf x_1=c_1\mathbf v_1 + c_2\mathbf v_2\text{.}\)
- Use the previous part of this problem to determine \(\mathbf x_2\text{,}\) \(\mathbf x_3\) and \(\mathbf x_4\text{.}\)
- After a very long time, how are all the bicycles distributed?