Skip to main content
Mathematics LibreTexts

10.4: Isometries

  • Page ID
    58893
  • \( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \) \( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)\(\newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\) \( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\) \( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\) \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\) \( \newcommand{\Span}{\mathrm{span}}\) \(\newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\) \( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\) \( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\) \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\) \( \newcommand{\Span}{\mathrm{span}}\)\(\newcommand{\AA}{\unicode[.8,0]{x212B}}\)

    We saw in Section [sec:2_6] that rotations about the origin and reflections in a line through the origin are linear operators on \(\mathbb{R}^2\). Similar geometric arguments (in Section [sec:4_4]) establish that, in \(\mathbb{R}^3\), rotations about a line through the origin and reflections in a plane through the origin are linear. We are going to give an algebraic proof of these results that is valid in any inner product space. The key observation is that reflections and rotations are distance preserving in the following sense. If \(V\) is an inner product space, a transformation \(S : V \to V\) (not necessarily linear) is said to be distance preserving if the distance between \(S(\mathbf{v})\) and \(S(\mathbf{w})\) is the same as the distance between \(\mathbf{v}\) and \(\mathbf{w}\) for all vectors \(\mathbf{v}\) and \(\mathbf{w}\); more formally, if

    \[\label{eq:distance_preserving} \left\| S(\mathbf{v}) - S(\mathbf{w}) \right\| =\left\| \mathbf{v} - \mathbf{w} \right\| \quad \mbox{for all } \mathbf{v} \mbox{ and } \mathbf{w} \mbox{ in } V \]

    Distance-preserving maps need not be linear. For example, if \(\mathbf{u}\) is any vector in \(V\), the transformation \(S_{\mathbf{u}} : V \to V\) defined by \(S_{\mathbf{u}}(\mathbf{v}) = \mathbf{v} + \mathbf{u}\) for all \(\mathbf{v}\) in \(V\) is called translation by \(\mathbf{u}\), and it is routine to verify that \(S_{\mathbf{u}}\) is distance preserving for any \(\mathbf{u}\). However, \(S_{\mathbf{u}}\) is linear only if \(\mathbf{u} = \mathbf{0}\) (since then \(S_{\mathbf{u}}(\mathbf{0}) = \mathbf{0}\)). Remarkably, distance-preserving operators that do fix the origin are necessarily linear.

    Lemma \(\PageIndex{1}\)

    Let \(V\) be an inner product space of dimension \(n\), and consider a distance-preserving transformation \(S : V \to V\). If \(S(\mathbf{0}) = \mathbf{0}\), then \(S\) is linear.

    We have \(\left\| S(\mathbf{v}) - S(\mathbf{w})\right\|^{2} = \left\| \mathbf{v} - \mathbf{w}\right\|^{2}\) for all \(\mathbf{v}\) and \(\mathbf{w}\) in \(V\) by ([eq:distance_preserving]), which gives

    \[\label{eq:lemma1_proof} \langle S(\mathbf{v}), S(\mathbf{w}) \rangle = \langle \mathbf{v}, \mathbf{w} \rangle \quad \mbox{ for all } \mathbf{v} \mbox{ and } \mathbf{w} \mbox{ in } V \]

    Now let \(\{\mathbf{f}_{1}, \mathbf{f}_{2}, \dots, \mathbf{f}_{n}\}\) be an orthonormal basis of \(V\). Then \(\{S(\mathbf{f}_{1}), S(\mathbf{f}_{2}), \dots, S(\mathbf{f}_{n})\}\) is orthonormal by ([eq:lemma1_proof]) and so is a basis because \( dim \textbf {V} = n\). Now compute:

    \[\begin{aligned} \langle S(\mathbf{v} + \mathbf{w}) - S(\mathbf{v}) - S(\mathbf{w}), S(\mathbf{f}_i) \rangle &= \langle S(\mathbf{v} + \mathbf{w}), S(\mathbf{f}_i) \rangle - \langle S(\mathbf{v}), S(\mathbf{f}_i) \rangle - \langle S(\mathbf{w}), S(\mathbf{f}_i) \rangle \\ &= \langle \mathbf{v} + \mathbf{w}, \mathbf{f}_i \rangle - \langle \mathbf{v}, \mathbf{f}_i \rangle - \langle \mathbf{w}, \mathbf{f}_i \rangle \\ &= 0\end{aligned} \nonumber \]

    for each \(i\). It follows from the expansion theorem (Theorem [thm:030904]) that \(S(\mathbf{v} + \mathbf{w}) - S(\mathbf{v}) - S(\mathbf{w}) = 0\); that is, \(S(\mathbf{v} + \mathbf{w}) = S(\mathbf{v}) + S(\mathbf{w})\). A similar argument shows that \(S(a\mathbf{v}) = aS(\mathbf{v})\) holds for all \(a\) in \(\mathbb{R}\) and \(\mathbf{v}\) in \(V\), so \(S\) is linear after all.

    Definition: Isometries

    Distance-preserving linear operators are called isometries.

    It is routine to verify that the composite of two distance-preserving transformations is again distance preserving. In particular the composite of a translation and an isometry is distance preserving. Surprisingly, the converse is true.

    Theorem \(\PageIndex{1}\)

    If \(V\) is a finite dimensional inner product space, then every distance-preserving transformation \(S : V \to V\) is the composite of a translation and an isometry.

    Proof. If \(S : V \to V\) is distance preserving, write \(S(\mathbf{0}) = \mathbf{u}\) and define \(T : V \to V\) by \(T(\mathbf{v}) = S(\mathbf{v}) - \mathbf{u}\) for all \(\mathbf{v}\) in \(V\). Then \(\left\| T(\mathbf{v}) - T(\mathbf{w})\right\| = \left\| \mathbf{v} - \mathbf{w}\right\|\) for all vectors \(\mathbf{v}\) and \(\mathbf{w}\) in \(V\) as the reader can verify; that is, \(T\) is distance preserving. Clearly, \(T(\mathbf{0}) = \mathbf{0}\), so it is an isometry by Lemma [lem:032019]. Since

    \[S(\mathbf{v}) = \mathbf{u} + T(\mathbf{v}) = (S_{\mathbf{u}} \circ T)(\mathbf{v}) \quad \mbox{for all } \mathbf{v} \mbox{ in } V \nonumber \]

    we have \(S = S_{\mathbf{u}} \circ T\), and the theorem is proved.

    In Theorem [thm:032040], \(S = S_{\mathbf{u}} \circ T\) factors as the composite of an isometry \(T\) followed by a translation \(S_{\mathbf{u}}\). More is true: this factorization is unique in that \(\mathbf{u}\) and \(T\) are uniquely determined by \(S\); and \(\mathbf{w} \in V\) exists such that \(S = T \circ S_{\mathbf{w}}\) is uniquely the composite of translation by \(\mathbf{w}\) followed by the same isometry \(T\) (Exercise [ex:10_4_12]).

    Theorem [thm:032040] focuses our attention on the isometries, and the next theorem shows that, while they preserve distance, they are characterized as those operators that preserve other properties.

    Theorem \(\PageIndex{2}\)

    Let \(T : V \to V\) be a linear operator on a finite dimensional inner product space \(V\). The following conditions are equivalent:

    =1.2mm

    1. & \(T\) is an isometry. & (\(T\) preserves distance)
    2. & \(\left\| T(\mathbf{v})\right\| = \left\| \mathbf{v}\right\|\) for all \(\mathbf{v}\) in \(V\). & (\(T\) preserves norms)
    3. & \(\langle T(\mathbf{v}), T(\mathbf{w}) \rangle = \langle\mathbf{v}, \mathbf{w} \rangle\) for all \(\mathbf{v}\) and \(\mathbf{w}\) in \(V\). & (\(T\) preserves inner products)
    4. & If \(\{\mathbf{f}_{1}, \mathbf{f}_{2}, \dots, \mathbf{f}_{n}\}\) is an orthonormal basis of \(V\), & & then \(\{T(\mathbf{f}_1), T(\mathbf{f}_2), \dots, T(\mathbf{f}_n) \}\) is also an orthonormal basis. & (\(T\) preserves orthonormal bases)
    5. &T carries some orthonormal basis to an orthonormal basis.

    Proof.

    1. \(\Rightarrow\) (2). Take \(\mathbf{w} = \mathbf{0}\) in ([eq:distance_preserving]).
    2. \(\Rightarrow\) (3). Since \(T\) is linear, (2) gives \(\left\| T(\mathbf{v}) - T(\mathbf{w})\right\|^{2} = \left\| T(\mathbf{v} - \mathbf{w})\right\|^{2} = \\left\| \mathbf{v} - \mathbf{w}\right\|^{2}\). Now (3) follows.
    3. \(\Rightarrow\) (4). By (3), \(\{T(\mathbf{f}_{1}), T(\mathbf{f}_{2}), \dots, T(\mathbf{f}_{n})\}\) is orthogonal and \(\left\| T(\mathbf{f}_{i})\right\|^{2} = \left\|\mathbf{f}_{i}\right\|^{2} = 1\). Hence it is a basis because \(dim \textbf{V} = n\).
    4. \(\Rightarrow\) (5). This needs no proof.
    5. \(\Rightarrow\) (1). By (5), let \(\{\mathbf{f}_{1}, \dots, \mathbf{f}_{n}\}\) be an orthonormal basis of \(V\) such that\(\{T(\mathbf{f}_{1}), \dots, T(\mathbf{f}_{n})\}\) is also orthonormal. Given \(\mathbf{v} = v_{1}\mathbf{f}_{1} + \dots + v_{n}\mathbf{f}_{n}\) in \(V\), we have \(T(\mathbf{v}) = v_{1}T(\mathbf{f}_{1}) + \dots + v_{n}T(\mathbf{f}_{n})\) so Pythagoras’ theorem gives

    \[\left\| T(\mathbf{v}) \right\| ^2 = v_1^2 + \dots + v_n^2 = \left\| \mathbf{v} \right\| ^2 \nonumber \]

    Hence \(\left\| T(\mathbf{v})\right\| = \left\|\mathbf{v}\right\|\) for all \(\mathbf{v}\), and (1) follows by replacing \(\mathbf{v}\) by \(\mathbf{v} - \mathbf{w}\).

    Before giving examples, we note some consequences of Theorem [thm:032053].

    Corollary \(\PageIndex{1}\)

    Let \(V\) be a finite dimensional inner product space.

    1. Every isometry of \(V\) is an isomorphism.
      1. \(1_{V} : V \to V\) is an isometry.
      2. The composite of two isometries of \(V\) is an isometry.
      3. The inverse of an isometry of \(V\) is an isometry.

    Proof. (1) is by (4) of Theorem \(\PageIndex{2}\) and Theorem 10.3.1. (2a) is clear, and (2b) is left to the reader. If \(T : V \to V\) is an isometry and \(\{\mathbf{f}_{1}, \dots, \mathbf{f}_{n}\}\) is an orthonormal basis of \(V\), then (2c) follows because \(T^{-1}\) carries the orthonormal basis \(\{T(\mathbf{f}_{1}), \dots, T(\mathbf{f}_{n})\}\) back to \(\{\mathbf{f}_{1}, \dots, \mathbf{f}_{n}\}\).

    The conditions in part (2) of the corollary assert that the set of isometries of a finite dimensional inner product space forms an algebraic system called a group. The theory of groups is well developed, and groups of operators are important in geometry. In fact, geometry itself can be fruitfully viewed as the study of those properties of a vector space that are preserved by a group of invertible linear operators.

    Example \(\PageIndex{1}\)

    Rotations of \(\mathbb{R}^2\) about the origin are isometries, as are reflections in lines through the origin: They clearly preserve distance and so are linear by Lemma \(\PageIndex{1}\). Similarly, rotations about lines through the origin and reflections in planes through the origin are isometries of \(\mathbb{R}^3\).

    Example \(\PageIndex{2}\)

    Let \(T : \mathbf{M}_{nn} \to \mathbf{M}_{nn}\) be the transposition operator: \(T(A) = A^{T}\). Then \(T\) is an isometry if the inner product is \(\langle A, B \rangle = tr \textbf{(AB^T)} = \displaystyle \sum_{i, j} a_{ij}b_{ij}\). In fact, \(T\) permutes the basis consisting of all matrices with one entry \(1\) and the other entries \(0\).

    Let \(T : \mathbf{M}_{nn} \to \mathbf{M}_{nn}\) be the transposition operator: \(T(A) = A^{T}\). Then \(T\) is an isometry if the inner product is \(\langle A, B \rangle = tr \mathbf{(AB^T)} = \displaystyle \sum_{i, j} a_{ij}b_{ij}\). In fact, \(T\) permutes the basis consisting of all matrices with one entry \(1\) and the other entries \(0\).

    The proof of the next result requires the fact (see Theorem [thm:032053]) that, if \(B\) is an orthonormal basis, then \(\langle\mathbf{v}, \mathbf{w} \rangle = C_{B}(\mathbf{v}) \bullet C_{B}(\mathbf{w})\) for all vectors \(\mathbf{v}\) and \(\mathbf{w}\).

    Theorem \(\PageIndex{3}\)

    Let \(T : V \to V\) be an operator where \(V\) is a finite dimensional inner product space. The following conditions are equivalent.

    1. \(T\) is an isometry.
    2. \(M_{B}(T)\) is an orthogonal matrix for every orthonormal basis \(B\).
    3. \(M_{B}(T)\) is an orthogonal matrix for some orthonormal basis \(B\).

    Proof.

    (1) \(\Rightarrow\) (2). Let \(B = \{\mathbf{e}_{1}, \dots, \mathbf{e}_{n}\}\) be an orthonormal basis. Then the \(j\)th column of \(M_{B}(T)\) is \(C_{B}[T(\mathbf{e}_{j})]\), and we have

    \[C_B[T(\mathbf{e}_j)] \bullet C_B[T(\mathbf{e}_k)] = \langle T(\mathbf{e}_j), T(\mathbf{e}_k) \rangle = \langle \mathbf{e}_j, \mathbf{e}_k \rangle \nonumber \]

    using (1). Hence the columns of \(M_{B}(T)\) are orthonormal in \(\mathbb{R}^n\), which proves (2).

    (2) \(\Rightarrow\) (3). This is clear.

    (3) \(\Rightarrow\) (1). Let \(B = \{\mathbf{e}_{1}, \dots, \mathbf{e}_{n}\}\) be as in (3). Then, as before,

    \[\langle T(\mathbf{e}_j), T(\mathbf{e}_k) \rangle = C_B[T(\mathbf{e}_j)] \bullet C_B[T(\mathbf{e}_k)] \nonumber \]

    so \(\{T(\mathbf{e}_{1}), \dots, T(\mathbf{e}_{n})\}\) is orthonormal by (3). Hence Theorem [thm:032053] gives (1).

    It is important that \(B\) is orthonormal in Theorem [thm:032147]. For example, \(T : V \to V\) given by \(T(\mathbf{v}) = 2\mathbf{v}\) preserves orthogonal sets but is not an isometry, as is easily checked.

    If \(P\) is an orthogonal square matrix, then \(P^{-1} = P^{T}\). Taking determinants yields \((\det P)^{2} = 1\), so \(\det P = \pm 1\). Hence:

    Corollary \(\PageIndex{2}\)

    If \(T : V \to V\) is an isometry where \(V\) is a finite dimensional inner product space, then \(\det T = \pm 1\).

    Example \(\PageIndex{3}\)

    If \(A\) is any \(n \times n\) matrix, the matrix operator \(T_{A}: \mathbb{R}^n \to \mathbb{R}^n\) is an isometry if and only if \(A\) is orthogonal using the dot product in \(\mathbb{R}^n\). Indeed, if \(E\) is the standard basis of \(\mathbb{R}^n\), then \(M_{E}(T_{A}) = A\) by Theorem 9.2.4.

    Rotations and reflections that fix the origin are isometries in \(\mathbb{R}^2\) and \(\mathbb{R}^3\) (Example [exa:032132]); we are going to show that these isometries (and compositions of them in \(\mathbb{R}^3\)) are the only possibilities. In fact, this will follow from a general structure theorem for isometries. Surprisingly enough, much of the work involves the two–dimensional case.

    Theorem \(\PageIndex{4}\)

    Let \(T : V \to V\) be an isometry on the two-dimensional inner product space \(V\). Then there are two possibilities.

    Either There is an orthonormal basis \(B\) of \(V\) such that

    \[M_B(T) = \left[ \begin{array}{rr} \cos \theta & - \sin \theta \\ \sin \theta & \cos \theta \end{array} \right], \ 0 \leq \theta < 2\pi \nonumber \]

    or There is an orthonormal basis \(B\) of \(V\) such that

    \[M_B(T) = \left[ \begin{array}{rr} 1 & 0 \\ 0 & -1 \end{array} \right] \nonumber \]

    Furthermore, type (1) occurs if and only if \(\det T = 1\), and type (2) occurs if and only if \(\det T = -1\).

    Proof. The final statement follows from the rest because \(\det T = \det [M_{B}(T)]\) for any basis \(B\). Let \(B_{0} = \{\mathbf{e}_{1}, \mathbf{e}_{2}\}\) be any ordered orthonormal basis of \(V\) and write

    \[A = M_{B_0}(T) = \left[ \begin{array}{rr} a & b \\ c & d \end{array} \right]; \mbox{ that is, } \begin{array}{l} T(\mathbf{e}_1) = a \mathbf{e}_1 + c \mathbf{e}_2 \\ T(\mathbf{e}_2) = b \mathbf{e}_1 + d \mathbf{e}_2 \\ \end{array} \nonumber \]

    Then \(A\) is orthogonal by Theorem [thm:032147], so its columns (and rows) are orthonormal. Hence

    \[a^{2} + c^{2} = 1 = b^{2} + d^{2} \nonumber \]

    so \((a, c)\) and \((d, b)\) lie on the unit circle. Thus angles \(\theta\) and \(\varphi\) exist such that

    \[\begin{array}{lll} a = \cos \theta, & c = \sin \theta & 0 \leq \theta < 2 \pi \\ d = \cos \varphi, & b = \sin \varphi & 0 \leq \varphi < 2 \pi \end{array} \nonumber \]

    Then \(\sin(\theta + \varphi) = cd + ab = 0\) because the columns of \(A\) are orthogonal, so \(\theta + \varphi = k\pi\) for some integer \(k\). This gives \(d = \cos(k\pi - \theta) = (-1)^{k} \cos \theta\) and \(b = \sin(k\pi - \theta) = (-1)^{k+1} \sin \theta\). Finally

    \[A = \left[ \begin{array}{cc} \cos \theta & (-1)^{k + 1} \sin \theta \\ \sin \theta & (-1)^k \cos \theta \end{array} \right] \nonumber \]

    If \(k\) is even we are in type (1) with \(B = B_{0}\), so assume \(k\) is odd. Then \(A = \left[ \begin{array}{rr} a & c \\ c & -a \end{array} \right]\). If \(a = -1\) and \(c = 0\), we are in type (1) with \(B = \{\mathbf{e}_{2}, \mathbf{e}_{2}\}\). Otherwise \(A\) has eigenvalues \(\lambda_{1} = 1\) and \(\lambda_{2} = -1\) with corresponding eigenvectors \(\mathbf{x}_1 = \left[ \begin{array}{c} 1 + a \\ c \end{array} \right]\) and \(\mathbf{x}_2 = \left[ \begin{array}{c} -c \\ 1 + a \end{array} \right]\) as the reader can verify. Write

    \[\mathbf{f}_1 = (1 + a)\mathbf{e}_1 + c\mathbf{e}_2 \quad \mbox{ and } \quad \mathbf{f}_2 = -c\mathbf{e}_2 + (1 + a)\mathbf{e}_2 \nonumber \]

    Then \(\mathbf{f}_{1}\) and \(\mathbf{f}_{2}\) are orthogonal (verify) and \(C_{B_0}(\mathbf{f}_i) = C_{B_0}(\lambda_i \mathbf{f}_i) = \mathbf{x}_i\) for each \(i\). Moreover

    \[C_{B_0} [T(\mathbf{f}_i)] = AC_{B_0}(\mathbf{f}_i) = A \mathbf{x}_i = \lambda_i \mathbf{x}_i = \lambda_i C_{B_0}(\mathbf{f}_i) = C_{B_0}(\lambda_i \mathbf{f}_i) \nonumber \]

    so \(T(\mathbf{f}_{i}) = \lambda_{i}\mathbf{f}_{i}\) for each \(i\). Hence \(M_B(T) = \left[ \begin{array}{cc} \lambda_1 & 0 \\ 0 & \lambda_2 \end{array} \right] = \left[ \begin{array}{rr} 1 & 0 \\ 0 & -1 \end{array} \right]\) and we are in type (2) with \(B = \left\{\frac{1}{\left\| \mathbf{f}_1 \right\|} \mathbf{f}_1, \frac{1}{\left\| \mathbf{f}_2 \right\|} \mathbf{f}_2 \right\}\).

    Corollary \(\PageIndex{3}\)

    An operator \(T : \mathbb{R}^2 \to \mathbb{R}^2\) is an isometry if and only if \(T\) is a rotation or a reflection.

    In fact, if \(E\) is the standard basis of \(\mathbb{R}^2\), then the clockwise rotation \(R_{\theta}\) about the origin through an angle \(\theta\) has matrix

    \[M_E(R_\theta) = \left[ \begin{array}{rr} \cos \theta & - \sin \theta \\ \sin \theta & \cos \theta \end{array} \right] \nonumber \]

    (see Theorem [thm:006021]). On the other hand, if \(S : \mathbb{R}^2 \to \mathbb{R}^2\) is the reflection in a line through the origin (called the fixed line of the reflection), let \(\mathbf{f}_{1}\) be a unit vector pointing along the fixed line and let \(\mathbf{f}_{2}\) be a unit vector perpendicular to the fixed line. Then \(B = \{\mathbf{f}_{1}, \mathbf{f}_{2}\}\) is an orthonormal basis, \(S(\mathbf{f}_{1}) = \mathbf{f}_{1}\) and \(S(\mathbf{f}_{2}) = -\mathbf{f}_{2}\), so

    \[M_B(S) = \left[ \begin{array}{rr} 1 & 0 \\ 0 & -1 \end{array} \right] \nonumber \]

    Thus \(S\) is of type 2. Note that, in this case, \(1\) is an eigenvalue of \(S\), and any eigenvector corresponding to \(1\) is a direction vector for the fixed line.

    Example \(\PageIndex{4}\)

    In each case, determine whether \(T_{A} : \mathbb{R}^2 \to \mathbb{R}^2\) is a rotation or a reflection, and then find the angle or fixed line:

    \[\begin{array}{lcl} \mbox{(a) } A = \frac{1}{2} \left[ \begin{array}{rr} 1 & \sqrt{3} \\ -\sqrt{3} & 1 \end{array} \right] & \quad & \mbox{(b) } A = \frac{1}{5} \left[ \begin{array}{rr} -3 & 4 \\ 4 & 3 \end{array} \right] \end{array} \nonumber \]

    Solution

    Both matrices are orthogonal, so (because \(M_{E}(T_{A}) = A\), where \(E\) is the standard basis) \(T_{A}\) is an isometry in both cases. In the first case, \(\det A = 1\), so \(T_{A}\) is a counterclockwise rotation through \(\theta\), where \(\cos \theta = \frac{1}{2}\) and \(\sin \theta = - \frac{\sqrt{3}}{2}\). Thus \(\theta = - \frac{\pi}{3}\). In (b), \(\det A = -1\), so \(T_{A}\) is a reflection in this case. We verify that \(\mathbf{d} = \left[ \begin{array}{r} 1 \\ 2 \end{array} \right]\) is an eigenvector corresponding to the eigenvalue \(1\). Hence the fixed line \(\mathbb{R}\mathbf{d}\) has equation \(y = 2x\).

    We now give a structure theorem for isometries. The proof requires three preliminary results, each of interest in its own right.

    Lemma \(\PageIndex{2}\)

    Let \(T : V \to V\) be an isometry of a finite dimensional inner product space \(V\). If \(U\) is a \(T\)-invariant subspace of \(V\), then \(U^{\perp}\) is also \(T\)-invariant.

    Proof. Let \(\mathbf{w}\) lie in \(U^{\perp}\). We are to prove that \(T(\mathbf{w})\) is also in \(U^{\perp}\); that is, \(\langle T(\mathbf{w}), \mathbf{u} \rangle = 0\) for all \(\mathbf{u}\) in \(U\). At this point, observe that the restriction of \(T\) to \(U\) is an isometry \(U \to U\) and so is an isomorphism by the corollary to Theorem [thm:032053]. In particular, each \(\mathbf{u}\) in \(U\) can be written in the form \(\mathbf{u} = T(\mathbf{u}_{1})\) for some \(\mathbf{u}_{1}\) in \(U\), so

    \[\langle T(\mathbf{w}), \mathbf{u} \rangle = \langle T(\mathbf{w}), T(\mathbf{u}_1) \rangle = \langle \mathbf{w}, \mathbf{u}_1 \rangle = 0 \nonumber \]

    because \(\mathbf{w}\) is in \(U^{\perp}\). This is what we wanted.

    To employ Lemma [lem:032292] above to analyze an isometry \(T : V \to V\) when \(dim \mathbf{V} = n\), it is necessary to show that a \(T\)-invariant subspace \(U\) exists such that \(U \neq 0\) and \(U \neq V\). We will show, in fact, that such a subspace \(U\) can always be found of dimension \(1\) or \(2\). If \(T\) has a real eigenvalue \(\lambda\) then \(\mathbb{R}\mathbf{u}\) is \(T\)-invariant where \(\mathbf{u}\) is any \(\lambda\)-eigenvector. But, in case (1) of Theorem [thm:032199], the eigenvalues of \(T\) are \(e^{i\theta}\) and \(e^{-i\theta}\) (the reader should check this), and these are nonreal if \(\theta \neq 0\) and \(\theta \neq \pi\). It turns out that every complex eigenvalue \(\lambda\) of \(T\) has absolute value \(1\) (Lemma [lem:032309] below); and that \(U\) has a \(T\)-invariant subspace of dimension \(2\) if \(\lambda\) is not real (Lemma [lem:032323]).

    Lemma \(\PageIndex{3}\)

    Let \(T : V \to V\) be an isometry of the finite dimensional inner product space \(V\). If \(\lambda\) is a complex eigenvalue of \(T\), then \(|\lambda| = 1\).

    Proof. Choose an orthonormal basis \(B\) of \(V\), and let \(A = M_{B}(T)\). Then \(A\) is a real orthogonal matrix so, using the standard inner product \(\langle \mathbf{x}, \mathbf{y} \rangle = \mathbf{x}^T \overline{\mathbf{y}}\) in \(\mathbb{C}\), we get

    \[\left\| A\mathbf{x} \right\| ^2 = (A\mathbf{x})^T(\overline{A\mathbf{x}}) = \mathbf{x}^T A^T \overline{A\mathbf{x}} = \mathbf{x}^TI\mathbf{x} = \left\| \mathbf{x} \right\| ^2 \nonumber \]

    for all \(\mathbf{x}\) in \(\mathbb{C}^n\). But \(A\mathbf{x} = \lambda\mathbf{x}\) for some \(\mathbf{x} \neq \mathbf{0}\), whence \(\left\| \mathbf{x}\right\|^{2} = \left\| \lambda\mathbf{x}\right\|^{2} = |\lambda|^{2}\left\|\mathbf{x}\right\|^{2}\). This gives \(|\lambda| = 1\), as required.

    Lemma \(\PageIndex{4}\)

    Let \(T : V \to V\) be an isometry of the \(n\)-dimensional inner product space \(V\). If \(T\) has a nonreal eigenvalue, then \(V\) has a two-dimensional \(T\)-invariant subspace.

    Proof. Let \(B\) be an orthonormal basis of \(V\), let \(A = M_{B}(T)\), and (using Lemma [lem:032309]) let \(\lambda = e^{i\alpha}\) be a nonreal eigenvalue of \(A\), say \(A\mathbf{x} = \lambda\mathbf{x}\) where \(\mathbf{x} \neq \mathbf{0}\) in \(\mathbb{C}^n\). Because \(A\) is real, complex conjugation gives \(A\overline{\mathbf{x}} = \overline{\lambda} \overline{\mathbf{x}}\), so \(\overline{\lambda}\) is also an eigenvalue. Moreover \(\lambda \neq \overline{\lambda}\) (\(\lambda\) is nonreal), so \(\{\mathbf{x}, \overline{\mathbf{x}} \}\) is linearly independent in \(\mathbb{C}^n\) (the argument in the proof of Theorem [thm:016090] works). Now define

    \[\mathbf{z}_1 = \mathbf{x} + \overline{\mathbf{x}} \quad \mbox{ and } \quad \mathbf{z}_2 = i(\mathbf{x} - \overline{\mathbf{x}}) \nonumber \]

    Then \(\mathbf{z}_{1}\) and \(\mathbf{z}_{2}\) lie in \(\mathbb{R}^n\), and \(\{\mathbf{z}_{1}, \mathbf{z}_{2}\}\) is linearly independent over \(\mathbb{R}\) because \(\{\mathbf{x}, \overline{\mathbf{x}} \}\) is linearly independent over \(\mathbb{C}\). Moreover

    \[\mathbf{x} = \frac{1}{2} (\mathbf{z}_1 - i \mathbf{z}_2) \quad \mbox{ and } \quad \overline{\mathbf{x}} = \frac{1}{2} (\mathbf{z}_1 + i\mathbf{z}_2) \nonumber \]

    Now \(\lambda + \overline{\lambda} = 2 \cos \alpha\) and \(\lambda - \overline{\lambda} = 2i \sin \alpha\), and a routine computation gives

    \[\begin{aligned} A \mathbf{z}_1 &= \mathbf{z}_1 \cos \alpha + \mathbf{z}_2 \sin \alpha \\ A \mathbf{z}_2 &= -\mathbf{z}_1 \sin \alpha + \mathbf{z}_2 \cos \alpha\end{aligned} \nonumber \]

    Finally, let \(\mathbf{e}_{1}\) and \(\mathbf{e}_{2}\) in \(V\) be such that \(\mathbf{z}_{1} = C_{B}(\mathbf{e}_{1})\) and \(\mathbf{z}_{2} = C_{B}(\mathbf{e}_{2})\). Then

    \[C_B[T(\mathbf{e}_1)] = AC_B(\mathbf{e}_1) = A\mathbf{z}_1 = C_B(\mathbf{e}_1 \cos \alpha + \mathbf{e}_2 \sin \alpha) \nonumber \]

    using Theorem [thm:027955]. Because \(C_{B}\) is one-to-one, this gives the first of the following equations (the other is similar):

    \[\begin{aligned} T(\mathbf{e}_1) &= \mathbf{e}_1 \cos \alpha + \mathbf{e}_2 \sin \alpha \\ T(\mathbf{e}_2) &= -\mathbf{e}_1 \sin \alpha + \mathbf{e}_2 \cos \alpha\end{aligned} \nonumber \]

    Thus \(U = span\{\mathbf{e}_{1}, \mathbf{e}_{2}\}\) is \(T\)-invariant and two-dimensional.

    We can now prove the structure theorem for isometries.

    Theorem \(\PageIndex{5}\)

    Let \(T : V \to V\) be an isometry of the \(n\)-dimensional inner product space \(V\). Given an angle \(\theta\), write \(R(\theta) = \left[ \begin{array}{rr} \cos \theta & - \sin \theta \\ \sin \theta & \cos \theta \end{array} \right]\). Then there exists an orthonormal basis \(B\) of \(V\) such that \(M_{B}(T)\) has one of the following block diagonal forms, classified for convenience by whether \(n\) is odd or even:

    \[n=2 k+1\left[\begin{array}{cccc}1 & 0 & \cdots & 0 \\ 0 & R\left(\theta_1\right) & \cdots & 0 \\ \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & \cdots & R\left(\theta_k\right)\end{array}\right] \quad \text{or} \quad\left[\begin{array}{cccc}-1 & 0 & \cdots & 0 \\ 0 & R\left(\theta_1\right) & \cdots & 0 \\ \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & \cdots & R\left(\theta_k\right)\end{array}\right]\]

    \[n=2 k\left[\begin{array}{cccc}R\left(\theta_1\right) & 0 & \cdots & 0 \\ 0 & R\left(\theta_2\right) & \cdots & 0 \\ \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & \cdots & R\left(\theta_k\right)\end{array}\right] \quad\ \text{or} \quad\left[\begin{array}{ccccc}-1 & 0 & 0 & \cdots & 0 \\ 0 & 1 & 0 & \cdots & 0 \\ 0 & 0 & R\left(\theta_1\right) & \cdots & 0 \\ \vdots & \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & 0 & \cdots & R\left(\theta_{k-1}\right)\end{array}\right]\]

    Proof. We show first, by induction on \(n\), that an orthonormal basis \(B\) of \(V\) can be found such that \(M_{B}(T)\) is a block diagonal matrix of the following form:

    \[M_B(T) = \left[ \begin{array}{ccccc} I_r & 0 & 0 & \cdots & 0 \\ 0 & -I_s & 0 & \cdots & 0 \\ 0 & 0 & R(\theta_1) & \cdots & 0 \\ \vdots & \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & 0 & \cdots & R(\theta_t) \end{array} \right] \nonumber \]

    where the identity matrix \(I_{r}\), the matrix \(-I_{s}\), or the matrices \(R(\theta_{i})\) may be missing. If \(n = 1\) and \(V = \mathbb{R}\mathbf{v}\), this holds because \(T(\mathbf{v}) = \lambda\mathbf{v}\) and \(\lambda = \pm 1\) by Lemma [lem:032309]. If \(n = 2\), this follows from Theorem [thm:032199]. If \(n \geq 3\), either \(T\) has a real eigenvalue and therefore has a one-dimensional \(T\)-invariant subspace \(U = \mathbb{R}\mathbf{u}\) for any eigenvector \(\mathbf{u}\), or \(T\) has no real eigenvalue and therefore has a two-dimensional \(T\)-invariant subspace \(U\) by Lemma [lem:032323]. In either case \(U^{\perp}\) is \(T\)-invariant (Lemma [lem:032292]) and \(dim \mathbf {U} ^{\perp} = n - dim \mathbf{U} < n\). Hence, by induction, let \(B_{1}\) and \(B_{2}\) be orthonormal bases of \(U\) and \(U^{\perp}\) such that \(M_{B_1}(T)\) and \(M_{B_2}(T)\) have the form given. Then \(B = B_{1} \cup B_{2}\) is an orthonormal basis of \(V\), and \(M_{B}(T)\) has the desired form with a suitable ordering of the vectors in \(B\).

    Now observe that \(R(0) = \left[ \begin{array}{rr} 1 & 0 \\ 0 & 1 \end{array} \right]\) and \(R(\pi) = \left[ \begin{array}{rr} -1 & 0 \\ 0 & -1 \end{array} \right]\). It follows that an even number of \(1\)s or \(-1\)s can be written as \(R(\theta_{1})\)-blocks. Hence, with a suitable reordering of the basis \(B\), the theorem follows.

    As in the dimension \(2\) situation, these possibilities can be given a geometric interpretation when \(V = \mathbb{R}^3\) is taken as euclidean space. As before, this entails looking carefully at reflections and rotations in \(\mathbb{R}^3\). If \(Q : \mathbb{R}^3 \to \mathbb{R}^3\) is any reflection in a plane through the origin (called the fixed plane of the reflection), take \(\{\mathbf{f}_{2}, \mathbf{f}_{3}\}\) to be any orthonormal basis of the fixed plane and take \(\mathbf{f}_{1}\) to be a unit vector perpendicular to the fixed plane. Then \(Q(\mathbf{f}_{1}) = -\mathbf{f}_{1}\), whereas \(Q(\mathbf{f}_{2}) = \mathbf{f}_{2}\) and \(Q(\mathbf{f}_{3}) = \mathbf{f}_{3}\). Hence \(B = \{\mathbf{f}_{1}, \mathbf{f}_{2}, \mathbf{f}_{3}\}\) is an orthonormal basis such that

    \[M_B(Q) = \left[ \begin{array}{rrr} -1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 1 \end{array} \right] \nonumber \]

    Similarly, suppose that \(R : \mathbb{R}^3 \to \mathbb{R}^3\) is any rotation about a line through the origin (called the axis of the rotation), and let \(\mathbf{f}_{1}\) be a unit vector pointing along the axis, so \(R(\mathbf{f}_{1}) = \mathbf{f}_{1}\). Now the plane through the origin perpendicular to the axis is an \(R\)-invariant subspace of \(\mathbb{R}^2\) of dimension \(2\), and the restriction of \(R\) to this plane is a rotation. Hence, by Theorem [thm:032199], there is an orthonormal basis \(B_{1} = \{\mathbf{f}_{2}, \mathbf{f}_{3}\}\) of this plane such that \(M_{B_1}(R) = \left[ \begin{array}{rr} \cos \theta & - \sin \theta \\ \sin \theta & \cos \theta \end{array} \right]\). But then \(B = \{\mathbf{f}_{1}, \mathbf{f}_{2}, \mathbf{f}_{3}\}\) is an orthonormal basis of \(\mathbb{R}^3\) such that the matrix of \(R\) is

    \[M_B(R) = \left[ \begin{array}{ccc} 1 & 0 & 0 \\ 0 & \cos \theta & - \sin \theta \\ 0 & \sin \theta & \cos \theta \end{array} \right] \nonumber \]

    However, Theorem [thm:032367] shows that there are isometries \(T\) in \(\mathbb{R}^3\) of a third type: those with a matrix of the form

    \[M_B(T) = \left[ \begin{array}{ccc} -1 & 0 & 0 \\ 0 & \cos \theta & - \sin \theta \\ 0 & \sin \theta & \cos \theta \end{array} \right] \nonumber \]

    If \(B = \{\mathbf{f}_{1}, \mathbf{f}_{2}, \mathbf{f}_{3}\}\), let \(Q\) be the reflection in the plane spanned by \(\mathbf{f}_{2}\) and \(\mathbf{f}_{3}\), and let \(R\) be the rotation corresponding to \(\theta\) about the line spanned by \(\mathbf{f}_{1}\). Then \(M_{B}(Q)\) and \(M_{B}(R)\) are as above, and \(M_{B}(Q) M_{B}(R) = M_{B}(T)\) as the reader can verify. This means that \(M_{B}(QR) = M_{B}(T)\) by Theorem [thm:028640], and this in turn implies that \(QR = T\) because \(M_{B}\) is one-to-one (see Exercise [ex:9_1_26]). A similar argument shows that \(RQ = T\), and we have Theorem [thm:032447].

    Theorem \(\PageIndex{6}\)

    If \(T : \mathbb{R}^3 \to \mathbb{R}^3\) is an isometry, there are three possibilities.

    1. \(T\) is a rotation, and \(M_B(T) = \left[ \begin{array}{ccc} 1 & 0 & 0 \\ 0 & \cos \theta & - \sin \theta \\ 0 & \sin \theta & \cos \theta \end{array} \right]\) for some orthonormal basis \(B\).
    2. \(T\) is a reflection, and \(M_B(T) = \left[ \begin{array}{rrr} -1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 1 \end{array} \right]\) for some orthonormal basis \(B\).
    3. \(T = QR = RQ\) where \(Q\) is a reflection, \(R\) is a rotation about an axis perpendicular to the fixed plane of \(Q\) and \(M_B(T) = \left[ \begin{array}{ccc} -1 & 0 & 0 \\ 0 & \cos \theta & - \sin \theta \\ 0 & \sin \theta & \cos \theta \end{array} \right]\) for some orthonormal basis \(B\).

    Hence \(T\) is a rotation if and only if \(\det T = 1\).

    Proof. It remains only to verify the final observation that \(T\) is a rotation if and only if \(\det T = 1\). But clearly \(\det T = -1\) in parts (b) and (c).

    A useful way of analyzing a given isometry \(T : \mathbb{R}^3 \to \mathbb{R}^3\) comes from computing the eigenvalues of \(T\). Because the characteristic polynomial of \(T\) has degree \(3\), it must have a real root. Hence, there must be at least one real eigenvalue, and the only possible real eigenvalues are \(\pm 1\) by Lemma \(\PageIndex{3}\). Thus Table (\PageIndex{1}\)includes all possibilities.

    Eigenvalues of T

    Action of T

    (1) 1 , no other real eigenvalues

    Rotation about the line \(\mathbb{R} \mathbf{f}\) where \(\mathbf{f}\) is an eigenvector corresponding to 1. [Case (a) of Theorem (\PageIndex{6}\).]

    (2) -1 , no other real eigenvalues

    Rotation about the line \(\mathbb{R} \mathbf{f}\) followed by reflection in the plane \((\mathbb{R})^{\perp}\) where \(\mathbf{f}\) is an eigenvector corresponding to -1 . [Case (c) of Theorem (\PageIndex{6}\).]

    (3) -1,1,1

    Reflection in the plane \((\mathbb{R} \mathbf{f})^{\perp}\) where \(\mathbf{f}\) is an eigenvector corresponding to -1 . [Case (b) of Theorem (\PageIndex{6}\).]

    (4) 1,-1,-1

    This is as in (1) with a rotation of \(\pi\).

    (5) -1,-1,-1

    Here T(x) = -x for all x. This is (2) with a rotation of \(\pi\).

    (6) 1,1,1

    Here T is the identity isometry.

    Example \(\PageIndex{5}\)

    Analyze the isometry \(T : \mathbb{R}^3 \to \mathbb{R}^3\) given by \(T \left[ \begin{array}{c} x \\ y \\ z \end{array} \right] = \left[ \begin{array}{c} y \\ z \\ -x \end{array} \right]\).

    Solution

    If \(B_{0}\) is the standard basis of \(\mathbb{R}^3\), then \(M_{B_0}(T) = \left[ \begin{array}{rrr} 0 & 1 & 0 \\ 0 & 0 & 1 \\ -1 & 0 & 0 \end{array} \right]\), so \(c_{T}(x) = x^{3} + 1 = (x + 1)(x^{2} - x + 1)\). This is (2) in Table [tab:10_4_1]. Write:

    \[\mathbf{f}_1 = \frac{1}{\sqrt{3}} \left[ \begin{array}{r} 1 \\ -1 \\ 1 \end{array} \right] \quad \mathbf{f}_2 = \frac{1}{\sqrt{6}} \left[ \begin{array}{rr} 1 \\ 2 \\ 1 \end{array} \right] \quad \mathbf{f}_3 = \frac{1}{\sqrt{2}} \left[ \begin{array}{rr} 1 \\ 0 \\ -1 \end{array} \right] \nonumber \]

    Here \(\mathbf{f}_{1}\) is a unit eigenvector corresponding to \(\lambda_{1} = -1\), so \(T\) is a rotation (through an angle \(\theta\)) about the line \(L = \mathbb{R}\mathbf{f}_{1}\), followed by reflection in the plane \(U\) through the origin perpendicular to \(\mathbf{f}_{1}\) (with equation \(x - y + z = 0\)). Then, \(\{\mathbf{f}_{1}, \mathbf{f}_{2}\}\) is chosen as an orthonormal basis of \(U\), so \(B = \{\mathbf{f}_{1}, \mathbf{f}_{2}, \mathbf{f}_{3}\}\) is an orthonormal basis of \(\mathbb{R}^3\) and

    \[M_B(T)=\left[\begin{array}{rrr}-1 & 0 & 0 \\ 0 & \frac{1}{2} & -\frac{\sqrt{3}}{2} \\ 0 & \frac{\sqrt{3}}{2} & \frac{1}{2}\end{array}\right] \nonumber\]

    Hence \(\theta\) is given by \(\cos \theta = \frac{1}{2}, \sin \theta = \frac{\sqrt{3}}{2}\), so \(\theta = \frac{\pi}{3}\).

    Let \(V\) be an \(n\)-dimensional inner product space. A subspace of \(V\) of dimension \(n - 1\) is called a hyperplane in \(V\). Thus the hyperplanes in \(\mathbb{R}^3\) and \(\mathbb{R}^2\) are, respectively, the planes and lines through the origin. Let \(Q : V \to V\) be an isometry with matrix

    \[M_B(Q) = \left[ \begin{array}{rr} -1 & 0 \\ 0 & I_{n - 1} \end{array} \right] \nonumber \]

    for some orthonormal basis \(B = \{\mathbf{f}_{1}, \mathbf{f}_{2}, \dots, \mathbf{f}_{n}\}\). Then \(Q(\mathbf{f}_{1}) = -\mathbf{f}_{1}\) whereas \(Q(\mathbf{u}) = \mathbf{u}\) for each \(\mathbf{u}\) in \(U = span \mathbf{ \{\mathbf{f}_{2}, \dots, \mathbf{f}_{n}\}}\). Hence \(U\) is called the fixed hyperplane of \(Q\), and \(Q\) is called reflection in \(U\). Note that each hyperplane in \(V\) is the fixed hyperplane of a (unique) reflection of \(V\). Clearly, reflections in \(\mathbb{R}^2\) and \(\mathbb{R}^3\) are reflections in this more general sense.

    Continuing the analogy with \(\mathbb{R}^2\) and \(\mathbb{R}^3\), an isometry \(T : V \to V\) is called a rotation if there exists an orthonormal basis \(\{\mathbf{f}_{1}, \dots, \mathbf{f}_{n}\}\) such that

    \[M_B(T) = \left[ \begin{array}{ccc} I_r & 0 & 0 \\ 0 & R(\theta) & 0 \\ 0 & 0 & I_s \end{array} \right] \nonumber \]

    in block form, where \(R(\theta) = \left[ \begin{array}{rr} \cos \theta & - \sin \theta \\ \sin \theta & \cos \theta \end{array} \right]\), and where either \(I_{r}\) or \(I_{s}\) (or both) may be missing. If \(R(\theta)\) occupies columns \(i\) and \(i + 1\) of \(M_{B}(T)\), and if \(W = span\mathbf {\{\mathbf{f}_{i}, \mathbf{f}_{i+1}\}}\), then \(W\) is \(T\)-invariant and the matrix of \(T : W \to W\) with respect to \(\{\mathbf{f}_{i}, \mathbf{f}_{i+1}\}\) is \(R(\theta)\). Clearly, if \(W\) is viewed as a copy of \(\mathbb{R}^2\), then \(T\) is a rotation in \(W\). Moreover, \(T(\mathbf{u}) = \mathbf{u}\) holds for all vectors \(\mathbf{u}\) in the \((n - 2)\)-dimensional subspace \(U = span \mathbf{\{\mathbf{f}_{1}, \dots, \mathbf{f}_{i-1}, \mathbf{f}_{i+1}, \dots, \mathbf{f}_{n}\}} \), and \(U\) is called the fixed axis of the rotation \(T\). In \(\mathbb{R}^3\), the axis of any rotation is a line (one-dimensional), whereas in \(\mathbb{R}^2\) the axis is \(U = \{\mathbf{0}\}\).

    With these definitions, the following theorem is an immediate consequence of Theorem [thm:032367] (the details are left to the reader).

    Theorem \(\PageIndex{7}\)

    Let \(T : V \to V\) be an isometry of a finite dimensional inner product space \(V\). Then there exist isometries \(T_{1}, \dots, T\) such that

    \[T = T_k T_{k - 1} \cdots T_2 T_1 \nonumber \]

    where each \(T_{i}\) is either a rotation or a reflection, at most one is a reflection, and \(T_{i}T_{j} = T_{j}T_{i}\) holds for all \(i\) and \(j\). Furthermore, \(T\) is a composite of rotations if and only if \(\det T = 1\).


    This page titled 10.4: Isometries is shared under a CC BY-NC-SA 4.0 license and was authored, remixed, and/or curated by W. Keith Nicholson (Lyryx Learning Inc.) via source content that was edited to the style and standards of the LibreTexts platform; a detailed edit history is available upon request.