Skip to main content
Mathematics LibreTexts

20.3: Linear Independence

  • Page ID
    81201
  • \( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \) \( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)\(\newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\) \( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\) \( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\) \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\) \( \newcommand{\Span}{\mathrm{span}}\) \(\newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\) \( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\) \( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\) \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\) \( \newcommand{\Span}{\mathrm{span}}\)\(\newcommand{\AA}{\unicode[.8,0]{x212B}}\)

    Let \(S = \{v_1, v_2, \ldots, v_n\}\) be a set of vectors in a vector space \(V\text{.}\) If there exist scalars \(\alpha_1, \alpha_2 \ldots \alpha_n \in F\) such that not all of the \(\alpha_i\)'s are zero and

    \[ \alpha_1 v_1 + \alpha_2 v_2 + \cdots + \alpha_n v_n = {\mathbf 0 }\text{,} \nonumber \]

    then \(S\) is said to be linearly dependent. If the set \(S\) is not linearly dependent, then it is said to be linearly independent. More specifically, \(S\) is a linearly independent set if

    \[ \alpha_1 v_1 + \alpha_2 v_2 + \cdots + \alpha_n v_n = {\mathbf 0 } \nonumber \]

    implies that

    \[ \alpha_1 = \alpha_2 = \cdots = \alpha_n = 0 \nonumber \]

    for any set of scalars \(\{ \alpha_1, \alpha_2 \ldots \alpha_n \}\text{.}\)

    Proposition \(20.9\)

    Let \(\{ v_1, v_2, \ldots, v_n \}\) be a set of linearly independent vectors in a vector space. Suppose that

    \[ v = \alpha_1 v_1 + \alpha_2 v_2 + \cdots + \alpha_n v_n = \beta_1 v_1 + \beta_2 v_2 + \cdots + \beta_n v_n\text{.} \nonumber \]

    Then \(\alpha_1 = \beta_1, \alpha_2 = \beta_2, \ldots, \alpha_n = \beta_n\text{.}\)

    Proof

    If

    \[ v = \alpha_1 v_1 + \alpha_2 v_2 + \cdots + \alpha_n v_n = \beta_1 v_1 + \beta_2 v_2 + \cdots + \beta_n v_n\text{,} \nonumber \]

    then

    \[ (\alpha_1 - \beta_1) v_1 + (\alpha_2 - \beta_2) v_2 + \cdots + (\alpha_n - \beta_n) v_n = {\mathbf 0}\text{.} \nonumber \]

    Since \(v_1, \ldots, v_n\) are linearly independent, \(\alpha_i - \beta_i = 0\) for \(i = 1, \ldots, n\text{.}\)

    The definition of linear dependence makes more sense if we consider the following proposition.

    Proposition \(20.10\)

    A set \(\{ v_1, v_2, \dots, v_n \}\) of vectors in a vector space \(V\) is linearly dependent if and only if one of the \(v_i\)'s is a linear combination of the rest.

    Proof

    Suppose that \(\{ v_1, v_2, \dots, v_n \}\) is a set of linearly dependent vectors. Then there exist scalars \(\alpha_1, \ldots, \alpha_n\) such that

    \[ \alpha_1 v_1 + \alpha_2 v_2 + \cdots + \alpha_n v_n = {\mathbf 0 }\text{,} \nonumber \]

    with at least one of the \(\alpha_i\)'s not equal to zero. Suppose that \(\alpha_k \neq 0\text{.}\) Then

    \[ v_k = - \frac{\alpha_1}{\alpha_k} v_1 - \cdots - \frac{\alpha_{k - 1}}{\alpha_k} v_{k-1} - \frac{\alpha_{k + 1}}{\alpha_k} v_{k + 1} - \cdots - \frac{\alpha_n}{\alpha_k} v_n\text{.} \nonumber \]

    Conversely, suppose that

    \[ v_k = \beta_1 v_1 + \cdots + \beta_{k - 1} v_{k - 1} + \beta_{k + 1} v_{k + 1} + \cdots + \beta_n v_n\text{.} \nonumber \]

    Then

    \[ \beta_1 v_1 + \cdots + \beta_{k - 1} v_{k - 1} - v_k + \beta_{k + 1} v_{k + 1} + \cdots + \beta_n v_n = {\mathbf 0}\text{.} \nonumber \]

    The following proposition is a consequence of the fact that any system of homogeneous linear equations with more unknowns than equations will have a nontrivial solution. We leave the details of the proof for the end-of-chapter exercises.

    Proposition \(20.11\)

    Suppose that a vector space \(V\) is spanned by \(n\) vectors. If \(m \gt n\text{,}\) then any set of \(m\) vectors in \(V\) must be linearly dependent

    A set \(\{ e_1, e_2, \ldots, e_n \}\) of vectors in a vector space \(V\) is called a basis for \(V\) if \(\{ e_1, e_2, \ldots, e_n \}\) is a linearly independent set that spans \(V\text{.}\)

    Example \(20.12\)

    The vectors \(e_1 = (1, 0, 0)\text{,}\) \(e_2 = (0, 1, 0)\text{,}\) and \(e_3 =(0, 0, 1)\) form a basis for \({\mathbb R}^3\text{.}\) The set certainly spans \({\mathbb R}^3\text{,}\) since any arbitrary vector \((x_1, x_2, x_3)\) in \({\mathbb R}^3\) can be written as \(x_1 e_1 + x_2 e_2 + x_3 e_3\text{.}\)

    Solution

    Also, none of the vectors \(e_1, e_2, e_3\) can be written as a linear combination of the other two; hence, they are linearly independent. The vectors \(e_1, e_2, e_3\) are not the only basis of \({\mathbb R}^3\text{:}\) the set \(\{ (3, 2, 1), (3, 2, 0), (1, 1, 1) \}\) is also a basis for \({\mathbb R}^3\text{.}\)

    Example \(20.13\)

    Let \({\mathbb Q}( \sqrt{2}\, ) = \{ a + b \sqrt{2} : a, b \in {\mathbb Q} \}\text{.}\) The sets

    Solution

    \(\{1, \sqrt{2}\, \}\) and \(\{1 + \sqrt{2}, 1 - \sqrt{2}\, \}\) are both bases of \({\mathbb Q}( \sqrt{2}\, )\text{.}\)

    From the last two examples it should be clear that a given vector space has several bases. In fact, there are an infinite number of bases for both of these examples. In general, there is no unique basis for a vector space. However, every basis of \({\mathbb R}^3\) consists of exactly three vectors, and every basis of \({\mathbb Q}(\sqrt{2}\, )\) consists of exactly two vectors. This is a consequence of the next proposition.

    Proposition \(20.14\)

    Let \(\{ e_1, e_2, \ldots, e_m \}\) and \(\{ f_1, f_2, \ldots, f_n \}\) be two bases for a vector space \(V\text{.}\) Then \(m = n\text{.}\)

    Proof

    Since \(\{ e_1, e_2, \ldots, e_m \}\) is a basis, it is a linearly independent set. By Proposition 20.11, \(n \leq m\text{.}\) Similarly, \(\{ f_1, f_2, \ldots, f_n \}\) is a linearly independent set, and the last proposition implies that \(m \leq n\text{.}\) Consequently, \(m = n\text{.}\)

    If \(\{ e_1, e_2, \ldots, e_n \}\) is a basis for a vector space \(V\text{,}\) then we say that the dimension of \(V\) is \(n\) and we write \(\dim V =n\text{.}\) We will leave the proof of the following theorem as an exercise.

    Theorem \(20.15\)

    Let \(V\) be a vector space of dimension \(n\text{.}\)

    1. If \(S = \{v_1, \ldots, v_n \}\) is a set of linearly independent vectors for \(V\text{,}\) then \(S\) is a basis for \(V\text{.}\)
    2. If \(S = \{v_1, \ldots, v_n \}\) spans \(V\text{,}\) then \(S\) is a basis for \(V\text{.}\)
    3. If \(S = \{v_1, \ldots, v_k \}\) is a set of linearly independent vectors for \(V\) with \(k \lt n\text{,}\) then there exist vectors \(v_{k + 1}, \ldots, v_n\) such that

      \[ \{v_1, \ldots, v_k, v_{k + 1}, \ldots, v_n \} \nonumber \]

      is a basis for \(V\text{.}\)


    This page titled 20.3: Linear Independence is shared under a GNU Free Documentation License 1.3 license and was authored, remixed, and/or curated by Thomas W. Judson (Abstract Algebra: Theory and Applications) via source content that was edited to the style and standards of the LibreTexts platform; a detailed edit history is available upon request.

    • Was this article helpful?