9.1: Algebraic Considerations
- Page ID
- 14552
\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)
\( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)
( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)
\( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)
\( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)
\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)
\( \newcommand{\Span}{\mathrm{span}}\)
\( \newcommand{\id}{\mathrm{id}}\)
\( \newcommand{\Span}{\mathrm{span}}\)
\( \newcommand{\kernel}{\mathrm{null}\,}\)
\( \newcommand{\range}{\mathrm{range}\,}\)
\( \newcommand{\RealPart}{\mathrm{Re}}\)
\( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)
\( \newcommand{\Argument}{\mathrm{Arg}}\)
\( \newcommand{\norm}[1]{\| #1 \|}\)
\( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)
\( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)
\( \newcommand{\vectorA}[1]{\vec{#1}} % arrow\)
\( \newcommand{\vectorAt}[1]{\vec{\text{#1}}} % arrow\)
\( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vectorC}[1]{\textbf{#1}} \)
\( \newcommand{\vectorD}[1]{\overrightarrow{#1}} \)
\( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}} \)
\( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)
\( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)
\( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)
\(\newcommand{\avec}{\mathbf a}\) \(\newcommand{\bvec}{\mathbf b}\) \(\newcommand{\cvec}{\mathbf c}\) \(\newcommand{\dvec}{\mathbf d}\) \(\newcommand{\dtil}{\widetilde{\mathbf d}}\) \(\newcommand{\evec}{\mathbf e}\) \(\newcommand{\fvec}{\mathbf f}\) \(\newcommand{\nvec}{\mathbf n}\) \(\newcommand{\pvec}{\mathbf p}\) \(\newcommand{\qvec}{\mathbf q}\) \(\newcommand{\svec}{\mathbf s}\) \(\newcommand{\tvec}{\mathbf t}\) \(\newcommand{\uvec}{\mathbf u}\) \(\newcommand{\vvec}{\mathbf v}\) \(\newcommand{\wvec}{\mathbf w}\) \(\newcommand{\xvec}{\mathbf x}\) \(\newcommand{\yvec}{\mathbf y}\) \(\newcommand{\zvec}{\mathbf z}\) \(\newcommand{\rvec}{\mathbf r}\) \(\newcommand{\mvec}{\mathbf m}\) \(\newcommand{\zerovec}{\mathbf 0}\) \(\newcommand{\onevec}{\mathbf 1}\) \(\newcommand{\real}{\mathbb R}\) \(\newcommand{\twovec}[2]{\left[\begin{array}{r}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\ctwovec}[2]{\left[\begin{array}{c}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\threevec}[3]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\cthreevec}[3]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\fourvec}[4]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\cfourvec}[4]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\fivevec}[5]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\cfivevec}[5]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\mattwo}[4]{\left[\begin{array}{rr}#1 \amp #2 \\ #3 \amp #4 \\ \end{array}\right]}\) \(\newcommand{\laspan}[1]{\text{Span}\{#1\}}\) \(\newcommand{\bcal}{\cal B}\) \(\newcommand{\ccal}{\cal C}\) \(\newcommand{\scal}{\cal S}\) \(\newcommand{\wcal}{\cal W}\) \(\newcommand{\ecal}{\cal E}\) \(\newcommand{\coords}[2]{\left\{#1\right\}_{#2}}\) \(\newcommand{\gray}[1]{\color{gray}{#1}}\) \(\newcommand{\lgray}[1]{\color{lightgray}{#1}}\) \(\newcommand{\rank}{\operatorname{rank}}\) \(\newcommand{\row}{\text{Row}}\) \(\newcommand{\col}{\text{Col}}\) \(\renewcommand{\row}{\text{Row}}\) \(\newcommand{\nul}{\text{Nul}}\) \(\newcommand{\var}{\text{Var}}\) \(\newcommand{\corr}{\text{corr}}\) \(\newcommand{\len}[1]{\left|#1\right|}\) \(\newcommand{\bbar}{\overline{\bvec}}\) \(\newcommand{\bhat}{\widehat{\bvec}}\) \(\newcommand{\bperp}{\bvec^\perp}\) \(\newcommand{\xhat}{\widehat{\xvec}}\) \(\newcommand{\vhat}{\widehat{\vvec}}\) \(\newcommand{\uhat}{\widehat{\uvec}}\) \(\newcommand{\what}{\widehat{\wvec}}\) \(\newcommand{\Sighat}{\widehat{\Sigma}}\) \(\newcommand{\lt}{<}\) \(\newcommand{\gt}{>}\) \(\newcommand{\amp}{&}\) \(\definecolor{fillinmathshade}{gray}{0.9}\)- Develop the abstract concept of a vector space through axioms.
- Deduce basic properties of vector spaces.
- Use the vector space axioms to determine if a set and its operations constitute a vector space.
In this section we consider the idea of an abstract vector space. A vector space is something which has two operations satisfying the following vector space axioms.
A vector space \(V\) is a set of vectors with two operations defined, addition and scalar multiplication, which satisfy the axioms of addition and scalar multiplication.
In the following definition we define two operations; vector addition, denoted by \(+\) and scalar multiplication denoted by placing the scalar next to the vector. A vector space need not have usual operations, and for this reason the operations will always be given in the definition of the vector space. The below axioms for addition (written +) and scalar multiplication must hold for however addition and scalar multiplication are defined for the vector space.
It is important to note that we have seen much of this content before, in terms of \(\mathbb{R}^n\). We will prove in this section that \(\mathbb{R}^n\) is an example of a vector space and therefore all discussions in this chapter will pertain to \(\mathbb{R}^n\). While it may be useful to consider all concepts of this chapter in terms of \(\mathbb{R}^n\), it is also important to understand that these concepts apply to all vector spaces.
In the following definition, we will choose scalars \(a,b\) to be real numbers and are thus dealing with real vector spaces. However, we could also choose scalars which are complex numbers. In this case, we would call the vector space \(V\) complex.
Let \(\vec{v}, \vec{w}, \vec{z}\) be vectors in a vector space \(V\). Then they satisfy the following axioms of addition:
- Closed under Addition \[\mbox{If}\; \vec{v}, \vec{w} \;\mbox{are in}\; V, \;\mbox{then}\; \vec{v}+\vec{w} \;\mbox{is also in}\; V.\nonumber \]
- The Commutative Law of Addition \[\vec{v}+\vec{w}=\vec{w}+\vec{v}\nonumber \]
- The Associative Law of Addition \[\left( \vec{v}+\vec{w}\right) +\vec{z}=\vec{v}+\left( \vec{w}+ \vec{z}\right)\nonumber \]
- The Existence of an Additive Identity \[\vec{v}+\vec{0}=\vec{v}\nonumber \]
- The Existence of an Additive Inverse \[\vec{v}+\left( -\vec{v}\right) =\vec{0}\nonumber \]
vectorspaceaxiomsaddition
Let \(a, b \in \mathbb{R}\) and let \(\vec{v}, \vec{w}, \vec{z}\) be vectors in a vector space \(V\). Then they satisfy the following axioms of scalar multiplication:
- Closed under Scalar Multiplication \[\mbox{If}\; a \;\mbox{is a real number, and}\; \vec{v} \;\mbox{is in}\; V, \;\mbox{then}\; a\vec{v} \;\mbox{is in}\; V.\nonumber\]
- \[a \left( \vec{v}+\vec{w}\right) = a \vec{v}+ a \vec{w} \nonumber \]
- \[\left( a + b \right) \vec{v}= a \vec{v}+ b \vec{v} \nonumber \]
- \[a \left( b \vec{v}\right) = (a b) \vec{v} \nonumber \]
- \[1\vec{v}=\vec{v} \nonumber \]
Consider the following example, in which we prove that \(\mathbb{R}^n\) is in fact a vector space.
\(\mathbb{R}^n\), under the usual operations of vector addition and scalar multiplication, is a vector space.
Solution
To show that \(\mathbb{R}^n\) is a vector space, we need to show that the above axioms hold. Let \(\vec{x}, \vec{y}, \vec{z}\) be vectors in \(\mathbb{R}^n\). We first prove the axioms for vector addition.
- To show that \(\mathbb{R}^n\) is closed under addition, we must show that for two vectors in \(\mathbb{R}^n\) their sum is also in \(\mathbb{R}^n\). The sum \(\vec{x} + \vec{y}\) is given by: \[\left[ \begin{array}{c} x_1 \\ x_2 \\ \vdots \\ x_n \end{array}\right] + \left[ \begin{array}{c} y_1 \\ y_2 \\ \vdots \\ y_n \end{array}\right] = \left[ \begin{array}{c} x_1 + y_1 \\ x_2 + y_2 \\ \vdots \\ x_n + y_n \end{array}\right]\nonumber \] The sum is a vector with \(n\) entries, showing that it is in \(\mathbb{R}^n\). Hence \(\mathbb{R}^n\) is closed under vector addition.
- To show that addition is commutative, consider the following: \[\begin{aligned} \vec{x} + \vec{y} &= \left[ \begin{array}{c} x_1 \\ x_2 \\ \vdots \\ x_n \end{array} \right] + \left[ \begin{array}{c} y_1 \\ y_2 \\ \vdots \\ y_n \end{array} \right] \\ &= \left[ \begin{array}{c} x_1 + y_1 \\ x_2 + y_2 \\ \vdots \\ x_n + y_n \end{array} \right] \\ &= \left[ \begin{array}{c} y_1 + x_1 \\ y_2 + x_2 \\ \vdots \\ y_n + x_n \end{array} \right] \\ &= \left[ \begin{array}{c} y_1 \\ y_2 \\ \vdots \\ y_n \end{array} \right] + \left[ \begin{array}{c} x_1 \\ x_2 \\ \vdots \\ x_n \end{array} \right] \\ &= \vec{y} + \vec{x}\end{aligned}\] Hence addition of vectors in \(\mathbb{R}^n\) is commutative.
- We will show that addition of vectors in \(\mathbb{R}^n\) is associative in a similar way. \[\begin{aligned} (\vec{x} + \vec{y}) + \vec{z} &= \left( \left[ \begin{array}{c} x_1 \\ x_2 \\ \vdots \\ x_n \end{array} \right] + \left[ \begin{array}{c} y_1 \\ y_2 \\ \vdots \\ y_n \end{array} \right] \right) + \left[ \begin{array}{c} z_1 \\ z_2 \\ \vdots \\ z_n \end{array} \right] \\ &= \left[ \begin{array}{c} x_1 + y_1 \\ x_2 + y_2 \\ \vdots \\ x_n + y_n \end{array} \right] + \left[ \begin{array}{c} z_1 \\ z_2 \\ \vdots \\ z_n \end{array} \right] \\ &= \left[ \begin{array}{c} (x_1 + y_1) + z_1 \\ (x_2 + y_2) + z_2\\ \vdots \\ (x_n + y_n) + z_n \end{array} \right] \\ &= \left[ \begin{array}{c} x_1 + (y_1 + z_1) \\ x_2 + (y_2 + z_2)\\ \vdots \\ x_n + (y_n + z_n) \end{array} \right ) \\ &= \left ( \begin{array}{c} x_1 \\ x_2 \\ \vdots \\ x_n \end{array} \right] + \left[ \begin{array}{c} y_1 + z_1 \\ y_2 + z_2\\ \vdots \\ y_n + z_n \end{array} \right] \\ &= \left[ \begin{array}{c} x_1 \\ x_2 \\ \vdots \\ x_n \end{array} \right] + \left( \left[ \begin{array}{c} y_1 \\ y_2 \\ \vdots \\ y_n \end{array} \right] + \left[ \begin{array}{c} z_1 \\ z_2 \\ \vdots \\ z_n \end{array} \right] \right) \\ &= \vec{x} + \left( \vec{y} + \vec{z} \right)\end{aligned}\]Hence addition of vectors is associative.
- Next, we show the existence of an additive identity. Let \(\vec{0} = \left[ \begin{array}{c} 0 \\ 0 \\ \vdots \\ 0 \end{array} \right].\) \[\begin{aligned}\vec{x}+\vec{0}&=\left[\begin{array}{c}x_1\\x_2\\ \vdots \\x_n\end{array}\right]+\left[\begin{array}{c}0\\0\\ \vdots \\0\end{array}\right] \\ &=\left[\begin{array}{c}x_1+0 \\ x_2+0 \\ \vdots \\ x_n+0\end{array}\right] \\ &=\left[\begin{array}{c}x_1\\x_2\\ \vdots \\ x_n\end{array}\right] \\ &=\vec{x}\end{aligned}\] Hence the zero vector \(\vec{0}\) is an additive identity.
- Next, we prove the existence of an additive inverse. Let \(-\vec{x} = \left ( \begin{array}{c} -x_1 \\ -x_2 \\ \vdots \\ -x_n \end{array}\right )\). \[\begin{aligned}\vec{x}+(-\vec{x})&=\left[\begin{array}{c}x_1\\x_2\\ \vdots\\x_n\end{array}\right]+\left[\begin{array}{c}-x_1 \\ -x_2\\ \vdots \\ -x_n\end{array}\right] \\ &=\left[\begin{array}{c}x_1-x_1 \\ x_2-x_2 \\ \vdots \\ x_n-x_n\end{array}\right] \\ &=\left[\begin{array}{c}0\\0\\ \vdots\\0\end{array}\right] \\ &=\vec{0}\end{aligned}\] Hence \(-\vec{x}\) is an additive inverse.
We now need to prove the axioms related to scalar multiplication. Let \(a,b\) be real numbers and let \(\vec{x}, \vec{y}\) be vectors in \(\mathbb{R}^n\).
- We first show that \(\mathbb{R}^n\) is closed under scalar multiplication. To do so, we show that \(a\vec{x}\) is also a vector with \(n\) entries. \[a\vec{x} = a\left[ \begin{array}{c} x_1 \\ x_2 \\ \vdots \\ x_n \end{array} \right] = \left[ \begin{array}{c} ax_1 \\ ax_2 \\ \vdots \\ ax_n \end{array} \right]\nonumber \] The vector \(a\vec{x}\) is again a vector with \(n\) entries, showing that \(\mathbb{R}^n\) is closed under scalar multiplication.
- We wish to show that \(a (\vec{x} + \vec{y}) = a\vec{x} + a\vec{y}\). \[\begin{aligned}a(\vec{x}+\vec{y})&=a\left(\left[\begin{array}{c}x_1\\x_2\\ \vdots\\x_n\end{array}\right]+\left[\begin{array}{c}x_1\\x_2\\ \vdots x_n\end{array}\right]\right)\\ &=a\left[\begin{array}{c}x_1+y_1 \\ x_2+y_2 \\ \vdots \\ x_n+y_n\end{array}\right] \\ &=\left[\begin{array}{c}a(x_1+y_1) \\ a(x_2+y_2) \\ \vdots \\ a(x_n+y_n)\end{array}\right] \\ &=\left[\begin{array}{c}ax_1+ay_1 \\ ax_2+ay_2 \\ \vdots \\ ax_n+ay_n\end{array}\right] \\ &=\left[\begin{array}{c}ax_1 \\ ax_2 \\ \vdots \\ ax_n\end{array}\right] +\left[\begin{array}{c}ay_1\\ay_2\\ \vdots\\ay_n\end{array}\right] \\ &=a\vec{x}+a\vec{y}\end{aligned}\]
- Next, we wish to show that \((a+b)\vec{x} = a\vec{x} + b\vec{x}\). \[\begin{aligned}(a+b)\vec{x}&=(a+b)\left[\begin{array}{c}x_1\\x_2\\ \vdots\\x_n\end{array}\right] \\ &=\left[\begin{array}{c}(a+b)x_1 \\ (a+b)x_2 \\ \vdots \\ (a+b)x_n\end{array}\right] \\ &=\left[\begin{array}{c}ax_1+bx_1 \\ ax_2+bx_2\\ \vdots \\ ax_n+bx_n\end{array}\right] \\ &=\left[\begin{array}{c}ax_1 \\ ax_2 \\ \vdots \\ ax_n\end{array}\right]+\left[\begin{array}{x}bx_1 \\ bx_2\\ \vdots \\ bx_n\end{array}\right] \\ &=a\vec{x}+b\vec{x}\end{aligned}\]
- We wish to show that \(a(b\vec{x}) = (ab) \vec{x}\). \[\begin{aligned} a(b\vec{x}) &= a\left(b \left[ \begin{array}{c} x_1 \\ x_2 \\ \vdots \\ x_n \end{array} \right] \right) \\ &= a\left( \left[ \begin{array}{c} bx_1 \\ bx_2 \\ \vdots \\ bx_n \end{array} \right] \right) \\ &= \left[ \begin{array}{c} a(bx_1) \\ a(bx_2) \\ \vdots \\ a(bx_n) \end{array} \right] \\ &= \left[ \begin{array}{c} (ab)x_1 \\ (ab)x_2 \\ \vdots \\ (ab)x_n \end{array} \right] \\ &= (ab) \left[ \begin{array}{c} x_1 \\ x_2 \\ \vdots \\ x_n \end{array} \right] \\ &= (ab)\vec{x}\end{aligned}\]
- Finally, we need to show that \(1\vec{x} = \vec{x}\). \[\begin{aligned} 1\vec{x} &= 1 \left[ \begin{array}{c} x_1 \\ x_2 \\ \vdots \\ x_n \end{array} \right] \\ &= \left[ \begin{array}{c} 1x_1 \\ 1x_2 \\ \vdots \\ 1x_n \end{array} \right] \\ &= \left[ \begin{array}{c} x_1 \\ x_2 \\ \vdots \\ x_n \end{array} \right] \\ &= \vec{x}\end{aligned}\]
By the above proofs, it is clear that \(\mathbb{R}^n\) satisfies the vector space axioms. Hence, \(\mathbb{R}^n\) is a vector space under the usual operations of vector addition and scalar multiplication.
We now consider some examples of vector spaces.
Let \(\mathbb{P}_2\) be the set of all polynomials of at most degree \(2\) as well as the zero polynomial. Define addition to be the standard addition of polynomials, and scalar multiplication the usual multiplication of a polynomial by a number. Then \(\mathbb{P}_2\) is a vector space.
Solution
We can write \(\mathbb{P}_2\) explicitly as \[\mathbb{P}_2 = \left\{ a_2x^2 + a_1x + a_0 | a_i \in \mathbb{R} \; \mbox{for all} \; i \right\}\nonumber \] To show that \(\mathbb{P}_2\) is a vector space, we verify the axioms. Let \(p(x), q(x), r(x)\) be polynomials in \(\mathbb{P}_2\) and let \(a,b,c\) be real numbers. Write \(p(x)=p_2x^2 + p_1x + p_0\), \(q(x)=q_2x^2 + q_1x + q_0\), and \(r(x)=r_2x^2 + r_1x + r_0\).
- We first prove that addition of polynomials in \(\mathbb{P}_2\) is closed. For two polynomials in \(\mathbb{P}_2\) we need to show that their sum is also a polynomial in \(\mathbb{P}_2\). From the definition of \(\mathbb{P}_2\), a polynomial is contained in \(\mathbb{P}_2\) if it is of degree at most \(2\) or the zero polynomial. \[\begin{aligned} p(x) + q(x) &= p_2x^2 + p_1x + p_0 + q_2x^2+ q_1x + q_0 \\ &= (p_2+q_2)x^2 + (p_1+q_1)x + (p_0+q_0) \end{aligned}\] The sum is a polynomial of degree \(2\) and therefore is in \(\mathbb{P}_2\). It follows that \(\mathbb{P}_2\) is closed under addition.
- We need to show that addition is commutative, that is \(p(x)+q(x) = q(x) + p(x)\). \[\begin{aligned} p(x) + q(x) &= p_2x^2 + p_1x + p_0 + q_2x^2 + q_1x + q_0\\ &= (p_2+q_2)x^2 + (p_1+q_1)x + (p_0+q_0) \\ &= (q_2+p_2)x^2 + (q_1+p_1)x + (q_0+p_0) \\ &= q_2x^2 + q_1x + q_0 + p_2x^2 + p_1x + p_0\\ &= q(x) + p(x)\end{aligned}\]
- Next, we need to show that addition is associative. That is, that \((p(x) + q(x)) + r(x) = p(x) + (q(x)+r(x))\). \[\begin{aligned} (p(x) + q(x)) + r(x) &= \left( p_2x^2 +p_1x + p_0 + q_2x^2 + q_1x + q_0 \right) + r_2x^2 +r_1x + r_0 \\ &= (p_2+q_2)x^2 + (p_1+q_1)x + (p_0 +q_0) + r_2x^2 + r_1x + r_0\\ &= (p_2+q_2+r_2)x^2 + (p_1+q_1+r_1)x + (p_0+q_0+r_0) \\ &= p_2x^2 + p_1x + p_0 + (q_2+r_2)x^2 + (q_1+r_1)x + (q_0+r_0) \\ &= p_2x^2 + p_1x + p_0 + \left( q_2x^2 +q_1x + q_0 + r_2x^2 + r_1x + r_0 \right)\\ &= p(x) + \left( q(x) + r(x) \right)\end{aligned}\]
- Next, we must prove that there exists an additive identity. Let \(0(x)=0x^2+0x+0\). \[\begin{aligned} p(x) + 0(x) &= p_2x^2 + p_1x + p_0 + 0x^2 + 0x + 0 \\ &= (p_2 + 0)x^2 + (p_1 + 0)x + (p_0 + 0)\\ &= p_2x^2 + p_1x + p_0 \\ &= p(x)\end{aligned}\] Hence an additive identity exists, specifically the zero polynomial.
- Next we must prove that there exists an additive inverse. Let \(-p(x) = -p_2x^2 - p_1x - p_0\) and consider the following: \[\begin{aligned} p(x) + (-p(x)) &= p_2x^2 + p_1x + p_0 + \left( - p_2x^2 - p_1x - p_0\right) \\ &= (p_2 - p_2)x^2 + (p_1 - p_1)x + (p_0 - p_0) \\ &= 0x^2 + 0x + 0 \\ &= 0(x)\end{aligned}\] Hence an additive inverse \(-p(x)\) exists such that \(p(x) + (-p(x)) = 0(x)\).
We now need to verify the axioms related to scalar multiplication.
- First we prove that \(\mathbb{P}_2\) is closed under scalar multiplication. That is, we show that \(ap(x)\) is also a polynomial of degree at most \(2\). \[ap(x) = a\left( p_2x^2 + p_1x + p_0 \right) = ap_2x^2 +ap_1x+ ap_0\nonumber \] Therefore \(\mathbb{P}_2\) is closed under scalar multiplication.
- We need to show that \(a(p(x) + q(x)) = ap(x) + aq(x)\). \[\begin{aligned} a(p(x) + q(x)) &= a \left( p_2x^2 + p_1x + p_0 + q_2x^2 + q_1x + q_0 \right)\\ &= a \left( (p_2+q_2)x^2 + (p_1+q_1)x + (p_0+q_0) \right)\\ &= a(p_2+q_2)x^2 + a(p_1+q_1)x + a(p_0 + q_0) \\ &= (ap_2 + aq_2)x^2 + (ap_1+aq_1)x + (ap_0 + aq_0) \\ &= ap_2x^2 + ap_1x + ap_0 + aq_2x^2 +aq_1x + aq_0\\ &= ap(x) + aq(x) \end{aligned}\]
- Next we show that \((a+b) p(x) = ap(x) + bp(x)\). \[\begin{aligned} (a+b) p(x) &= (a+b) ( p_2x^2 + p_1x + p_0)\\ &= (a+b)p_2x^2 + (a+b)p_1x + (a+b)p_0 \\ &= ap_2x^2 + ap_1x + ap_0 + bp_2x^2 +bp_1x + bp_0\\ &= ap(x) + bp(x)\end{aligned}\]
- The next axiom which needs to be verified is \(a(bp(x)) = (ab)p(x)\). \[\begin{aligned} a(bp(x)) &= a \left( b \left(p_2x^2 + p_1x +p_0\right)\right) \\ &= a \left( bp_2x^2 +bp_1x + bp_0 \right)\\ &=abp_2x^2+abp_1x+abp_0 \\ &=(ab)(p_2x^2+p_1x+p_0) \\ &=(ab)p(x)\end{aligned}\]
- Finally, we show that \(1p(x) = p(x)\). \[\begin{aligned} 1p(x) &= 1 \left( p_2x^2 + p_1x + p_0\right)\\ &= 1p_2x^2 + 1p_1x + 1p_0\\ &= p_2x^2 + p_1x + p_0\\ &= p(x)\end{aligned}\] Since the above axioms hold, we know that \(\mathbb{P}_2\) as described above is a vector space.
Another important example of a vector space is the set of all matrices of the same size.
Let \(\mathbb{M}_{2,3}\) be the set of all \(2 \times 3\) matrices. Using the usual operations of matrix addition and scalar multiplication, show that \(\mathbb{M}_{2,3}\) is a vector space.
Solution
Let \(A, B\) be \(2 \times 3\) matrices in \(\mathbb{M}_{2,3}\). We first prove the axioms for addition.
- In order to prove that \(\mathbb{M}_{2,3}\) is closed under matrix addition, we show that the sum \(A+B\) is in \(\mathbb{M}_{2,3}\). This means showing that \(A+B\) is a \(2 \times 3\) matrix. \[\begin{aligned} A+B &= \left[ \begin{array}{rrr} a_{11} & a_{12} & a_{13}\\ a_{21} & a_{22} & a_{23} \end{array} \right] + \left[ \begin{array}{rrr} b_{11} & b_{12} & b_{13}\\ b_{21} & b_{22} & b_{23} \end{array} \right] \\ &= \left[ \begin{array}{rrr} a_{11} + b_{11} & a_{12}+b_{12} & a_{13}+b_{13}\\ a_{21} +b_{21}& a_{22}+b_{22} & a_{23}+b_{23} \end{array} \right]\end{aligned}\] You can see that the sum is a \(2\times 3\) matrix, so it is in \(\mathbb{M}_{2,3}\). It follows that \(\mathbb{M}_{2,3}\) is closed under matrix addition.
- The remaining axioms regarding matrix addition follow from properties of matrix addition. Therefore \(\mathbb{M}_{2,3}\) satisfies the axioms of matrix addition.
We now turn our attention to the axioms regarding scalar multiplication. Let \(A, B\) be matrices in \(\mathbb{M}_{2,3}\) and let \(c\) be a real number.
- We first show that \(\mathbb{M}_{2,3}\) is closed under scalar multiplication. That is, we show that \(cA\) a \(2 \times 3\) matrix. \[\begin{aligned} cA &= c\left[ \begin{array}{rrr} a_{11} & a_{12} & a_{13}\\ a_{21} & a_{22} & a_{23} \end{array} \right] \\ &= \left[ \begin{array}{rrr} ca_{11} & ca_{12} & ca_{13}\\ ca_{21} & ca_{22} & ca_{23} \end{array} \right]\end{aligned}\] This is a \(2 \times 3\) matrix in \(\mathbb{M}_{2,3}\) which proves that the set is closed under scalar multiplication.
- The remaining axioms of scalar multiplication follow from properties of scalar multiplication of matrices. Therefore \(\mathbb{M}_{2,3}\) satisfies the axioms of scalar multiplication.
In conclusion, \(\mathbb{M}_{2,3}\) satisfies the required axioms and is a vector space.
While here we proved that the set of all \(2 \times 3\) matrices is a vector space, there is nothing special about this choice of matrix size. In fact if we instead consider \(\mathbb{M}_{m,n}\), the set of all \(m \times n\) matrices, then \(\mathbb{M}_{m,n}\) is a vector space under the operations of matrix addition and scalar multiplication.
We now examine an example of a set that does not satisfy all of the above axioms, and is therefore not a vector space.
Let \(V\) denote the set of \(2 \times 3\) matrices. Let addition in \(V\) be defined by \(A + B = A\) for matrices \(A,B\) in \(V\). Let scalar multiplication in \(V\) be the usual scalar multiplication of matrices. Show that \(V\) is not a vector space.
Solution
In order to show that \(V\) is not a vector space, it suffices to find only one axiom which is not satisfied. We will begin by examining the axioms for addition until one is found which does not hold. Let \(A,B\) be matrices in \(V\).
- We first want to check if addition is closed. Consider \(A+B\). By the definition of addition in the example, we have that \(A+B = A\). Since \(A\) is a \(2 \times 3\) matrix, it follows that the sum \(A+B\) is in \(V\), and \(V\) is closed under addition.
- We now wish to check if addition is commutative. That is, we want to check if \(A + B = B + A\) for all choices of \(A\) and \(B\) in \(V\). From the definition of addition, we have that \(A + B = A\) and \(B + A = B\). Therefore, we can find \(A\), \(B\) in \(V\) such that these sums are not equal. One example is \[A = \left[ \begin{array}{rrr} 1 & 0 & 0 \\ 0 & 0 & 0 \end{array}\right], B = \left[ \begin{array}{rrr} 0 & 0 & 0 \\ 1 & 0 & 0 \end{array}\right]\nonumber \] Using the operation defined by \(A+B=A\), we have \[\begin{aligned}A+B&=A \\ &=\left[\begin{array}{ccc}1&0&0\\0&0&0\end{array}\right] \\ B+A&=B \\ &=\left[\begin{array}{ccc}0&0&0\\1&0&0\end{array}\right]\end{aligned}\] It follows that \(A+B\neq B+A\). Therefore addition as defined for \(V\) is not commutative and \(V\) fails this axiom. Hence \(V\) is not a vector space.
Consider another example of a vector space.
Let \(S\) be a nonempty set and define \(\mathbb{F}_S\) to be the set of real functions defined on \(S\). In other words, we write \(\mathbb{F}_S: S \mapsto \mathbb{R}\). Letting \(a,b,c\) be scalars and \(f,g,h\) functions, the vector operations are defined as \[\begin{aligned} \left( f+g\right) \left( x\right) &=f\left( x\right) +g\left( x\right) \\ \left( af\right) \left( x\right) &=a\left( f\left( x\right) \right)\end{aligned}\] Show that \(\mathbb{F}_S\) is a vector space.
Solution
To verify that \(\mathbb{F}_S\) is a vector space, we must prove the axioms beginning with those for addition. Let \(f, g, h\) be functions in \(\mathbb{F}_S\).
- First we check that addition is closed. For functions \(f, g\) defined on the set \(S\), their sum given by \[(f+g)(x) = f(x)+g(x)\nonumber \] is again a function defined on \(S\). Hence this sum is in \(\mathbb{F}_S\) and \(\mathbb{F}_S\) is closed under addition.
- Secondly, we check the commutative law of addition: \[\left( f+g\right) \left( x\right) =f\left( x\right) +g\left( x\right) =g\left( x\right) +f\left( x\right) =\left( g+f\right) \left( x\right)\nonumber \] Since \(x\) is arbitrary, \(f+g=g+f\).
- Next we check the associative law of addition: \[\left( \left( f+g\right) +h\right) \left( x\right) = \left( f+g\right) \left( x\right) +h\left( x\right) =\left( f\left( x\right) +g\left( x\right) \right) +h\left( x\right)\nonumber \] \[=f\left( x\right) +\left( g\left( x\right) +h\left( x\right) \right) =\left( f\left( x\right) +\left( g+h\right) \left( x\right) \right) =\left( f+\left( g+h\right) \right) \left( x\right)\nonumber \] and so \(\left( f+g\right) +h=f+\left( g+h\right) .\)
- Next we check for an additive identity. Let \(0\) denote the function which is given by \(0\left( x\right) =0.\) Then this is an additive identity because \[\left( f+0\right) \left( x\right) =f\left( x\right) +0\left( x\right) =f\left( x\right)\nonumber \] and so \(f+0=f\).
- Finally, check for an additive inverse. Let \(-f\) be the function which satisfies \(\left( -f\right) \left( x\right) = -f\left( x\right) .\) Then \[\left( f+\left( -f\right) \right) \left( x\right) = f\left( x\right) +\left( -f\right) \left( x\right) = f\left( x\right) +-f\left( x\right) =0\nonumber \] Hence \(f+\left( -f\right) =0\).
Now, check the axioms for scalar multiplication.
- We first need to check that \(\mathbb{F}_S\) is closed under scalar multiplication. For a function \(f(x)\) in \(\mathbb{F}_S\) and real number \(a\), the function \((af)(x) = a(f(x))\) is again a function defined on the set \(S\). Hence \(a(f(x))\) is in \(\mathbb{F}_S\) and \(\mathbb{F}_S\) is closed under scalar multiplication.
- \[\left( \left( a+b\right) f\right) \left( x\right) = \left( a+b\right) f\left( x\right) =af\left( x\right) +bf\left( x\right) = \left( af+bf\right) \left( x\right)\nonumber \] and so \(\left( a+b\right) f=af+bf\).
- \[\left( a\left( f+g\right) \right) \left( x\right) = a\left( f+g\right) \left( x\right) = a\left( f\left( x\right) +g\left( x\right) \right)\nonumber \] \[=af\left( x\right) +bg\left( x\right) = \left( af+bg\right) \left( x\right)\nonumber \] and so \(a\left( f+g\right) =af+bg\).
- \[\left( \left( ab\right) f\right) \left( x\right) = \left( ab\right) f\left( x\right) =a\left( bf\left( x\right) \right) = \left( a\left( bf\right) \right) \left( x\right)\nonumber \] so \(\left( abf\right) =a\left( bf\right)\).
- Finally \(\left( 1f\right) \left( x\right) = 1f\left( x\right) =f\left( x\right)\) so \(1f=f\).
It follows that \(V\) satisfies all the required axioms and is a vector space.
Consider the following important theorem.
In any vector space, the following are true:
- \(\vec{0}\), the additive identity, is unique
- \(-\vec{x}\), the additive inverse, is unique
- \(0\vec{x}=\vec{0}\) for all vectors \(\vec{x}\)
- \(\left( -1\right) \vec{x}=-\vec{x}\) for all vectors \(\vec{x}\)
- Proof
-
- When we say that the additive identity, \(\vec{0}\), is unique, we mean that if a vector acts like the additive identity, then it is the additive identity. To prove this uniqueness, we want to show that another vector which acts like the additive identity is actually equal to \(\vec{0}\). Suppose \(\vec{0}^{\prime }\) is also an additive identity. Then, \[\vec{0} + \vec{0}^{\prime} = \vec{0}\nonumber \] Now, for \(\vec{0}\) the additive identity given above in the axioms, we have that \[\vec{0}^{\prime} + \vec{0} = \vec{0}^{\prime}\nonumber \] So by the commutative property: \[0 = 0 + 0^{\prime} = 0^{\prime} + 0 = 0^{\prime}\nonumber \] This says that if a vector acts like an additive identity (such as \(\vec{0}^{\prime}\)), it in fact equals \(\vec{0}\). This proves the uniqueness of \(\vec{0}\).
- When we say that the additive inverse, \(-\vec{x}\), is unique, we mean that if a vector acts like the additive inverse, then it is the additive inverse. Suppose that \(\vec{y}\) acts like an additive inverse: \[\vec{x}+\vec{y}=\vec{0}\nonumber \] Then the following holds: \[\vec{y} = \vec{0} + \vec{y} = (-\vec{x} + \vec{x}) + \vec{y} = -\vec{x} + (\vec{x} + \vec{y}) = -\vec{x} + \vec{0} = -\vec{x}\nonumber \] Thus if \(\vec{y}\) acts like the additive inverse, it is equal to the additive inverse \(-\vec{x}\). This proves the uniqueness of \(-\vec{x}\).
- This statement claims that for all vectors \(\vec{x}\), scalar multiplication by \(0\) equals the zero vector \(\vec{0}\). Consider the following, using the fact that we can write \(0=0+0\): \[0\vec{x}=\left( 0+0\right) \vec{x}=0\vec{x}+0\vec{x}\nonumber \] We use a small trick here: add \(-0\vec{x}\) to both sides. This gives \[\begin{aligned} 0\vec{x} + (-0\vec{x})&=0\vec{x}+0\vec{x}+(-\vec{x})\\ \vec{0} + 0 &=0\vec{x} + 0 \\ \vec{0} &= 0\vec{x}\end{aligned}\] This proves that scalar multiplication of any vector by \(0\) results in the zero vector \(\vec{0}\).
- Finally, we wish to show that scalar multiplication of \(-1\) and any vector \(\vec{x}\) results in the additive inverse of that vector, \(-\vec{x}\). Recall from \(2.\) above that the additive inverse is unique. Consider the following: \[\begin{aligned} \left( -1\right) \vec{x}+\vec{x} & =\left( -1\right) \vec{x}+1\vec{x}\\ &=\left(-1+1\right) \vec{x} \\ &=0\vec{x} \\ &=\vec{0}\end{aligned}\] By the uniqueness of the additive inverse shown earlier, any vector which acts like the additive inverse must be equal to the additive inverse. It follows that \(\left( -1\right) \vec{x}=-\vec{x}\).
An important use of the additive inverse is the following theorem.
Let \(V\) be a vector space. Then \(\vec{v} + \vec{w} = \vec{v} + \vec{z}\) implies that \(\vec{w} = \vec{z}\) for all \(\vec{v}, \vec{w}, \vec{z} \in V\)
- Proof
-
The proof follows from the vector space axioms, in particular the existence of an additive inverse (\(-\vec{u}\)). The proof is left as an exercise to the reader.