Skip to main content
Mathematics LibreTexts

Symmetric Matrices

  • Page ID
    218320
  • \( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

    \( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)

    \( \newcommand{\dsum}{\displaystyle\sum\limits} \)

    \( \newcommand{\dint}{\displaystyle\int\limits} \)

    \( \newcommand{\dlim}{\displaystyle\lim\limits} \)

    \( \newcommand{\id}{\mathrm{id}}\) \( \newcommand{\Span}{\mathrm{span}}\)

    ( \newcommand{\kernel}{\mathrm{null}\,}\) \( \newcommand{\range}{\mathrm{range}\,}\)

    \( \newcommand{\RealPart}{\mathrm{Re}}\) \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

    \( \newcommand{\Argument}{\mathrm{Arg}}\) \( \newcommand{\norm}[1]{\| #1 \|}\)

    \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

    \( \newcommand{\Span}{\mathrm{span}}\)

    \( \newcommand{\id}{\mathrm{id}}\)

    \( \newcommand{\Span}{\mathrm{span}}\)

    \( \newcommand{\kernel}{\mathrm{null}\,}\)

    \( \newcommand{\range}{\mathrm{range}\,}\)

    \( \newcommand{\RealPart}{\mathrm{Re}}\)

    \( \newcommand{\ImaginaryPart}{\mathrm{Im}}\)

    \( \newcommand{\Argument}{\mathrm{Arg}}\)

    \( \newcommand{\norm}[1]{\| #1 \|}\)

    \( \newcommand{\inner}[2]{\langle #1, #2 \rangle}\)

    \( \newcommand{\Span}{\mathrm{span}}\) \( \newcommand{\AA}{\unicode[.8,0]{x212B}}\)

    \( \newcommand{\vectorA}[1]{\vec{#1}}      % arrow\)

    \( \newcommand{\vectorAt}[1]{\vec{\text{#1}}}      % arrow\)

    \( \newcommand{\vectorB}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

    \( \newcommand{\vectorC}[1]{\textbf{#1}} \)

    \( \newcommand{\vectorD}[1]{\overrightarrow{#1}} \)

    \( \newcommand{\vectorDt}[1]{\overrightarrow{\text{#1}}} \)

    \( \newcommand{\vectE}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash{\mathbf {#1}}}} \)

    \( \newcommand{\vecs}[1]{\overset { \scriptstyle \rightharpoonup} {\mathbf{#1}} } \)

    \(\newcommand{\longvect}{\overrightarrow}\)

    \( \newcommand{\vecd}[1]{\overset{-\!-\!\rightharpoonup}{\vphantom{a}\smash {#1}}} \)

    \(\newcommand{\avec}{\mathbf a}\) \(\newcommand{\bvec}{\mathbf b}\) \(\newcommand{\cvec}{\mathbf c}\) \(\newcommand{\dvec}{\mathbf d}\) \(\newcommand{\dtil}{\widetilde{\mathbf d}}\) \(\newcommand{\evec}{\mathbf e}\) \(\newcommand{\fvec}{\mathbf f}\) \(\newcommand{\nvec}{\mathbf n}\) \(\newcommand{\pvec}{\mathbf p}\) \(\newcommand{\qvec}{\mathbf q}\) \(\newcommand{\svec}{\mathbf s}\) \(\newcommand{\tvec}{\mathbf t}\) \(\newcommand{\uvec}{\mathbf u}\) \(\newcommand{\vvec}{\mathbf v}\) \(\newcommand{\wvec}{\mathbf w}\) \(\newcommand{\xvec}{\mathbf x}\) \(\newcommand{\yvec}{\mathbf y}\) \(\newcommand{\zvec}{\mathbf z}\) \(\newcommand{\rvec}{\mathbf r}\) \(\newcommand{\mvec}{\mathbf m}\) \(\newcommand{\zerovec}{\mathbf 0}\) \(\newcommand{\onevec}{\mathbf 1}\) \(\newcommand{\real}{\mathbb R}\) \(\newcommand{\twovec}[2]{\left[\begin{array}{r}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\ctwovec}[2]{\left[\begin{array}{c}#1 \\ #2 \end{array}\right]}\) \(\newcommand{\threevec}[3]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\cthreevec}[3]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \end{array}\right]}\) \(\newcommand{\fourvec}[4]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\cfourvec}[4]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \end{array}\right]}\) \(\newcommand{\fivevec}[5]{\left[\begin{array}{r}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\cfivevec}[5]{\left[\begin{array}{c}#1 \\ #2 \\ #3 \\ #4 \\ #5 \\ \end{array}\right]}\) \(\newcommand{\mattwo}[4]{\left[\begin{array}{rr}#1 \amp #2 \\ #3 \amp #4 \\ \end{array}\right]}\) \(\newcommand{\laspan}[1]{\text{Span}\{#1\}}\) \(\newcommand{\bcal}{\cal B}\) \(\newcommand{\ccal}{\cal C}\) \(\newcommand{\scal}{\cal S}\) \(\newcommand{\wcal}{\cal W}\) \(\newcommand{\ecal}{\cal E}\) \(\newcommand{\coords}[2]{\left\{#1\right\}_{#2}}\) \(\newcommand{\gray}[1]{\color{gray}{#1}}\) \(\newcommand{\lgray}[1]{\color{lightgray}{#1}}\) \(\newcommand{\rank}{\operatorname{rank}}\) \(\newcommand{\row}{\text{Row}}\) \(\newcommand{\col}{\text{Col}}\) \(\renewcommand{\row}{\text{Row}}\) \(\newcommand{\nul}{\text{Nul}}\) \(\newcommand{\var}{\text{Var}}\) \(\newcommand{\corr}{\text{corr}}\) \(\newcommand{\len}[1]{\left|#1\right|}\) \(\newcommand{\bbar}{\overline{\bvec}}\) \(\newcommand{\bhat}{\widehat{\bvec}}\) \(\newcommand{\bperp}{\bvec^\perp}\) \(\newcommand{\xhat}{\widehat{\xvec}}\) \(\newcommand{\vhat}{\widehat{\vvec}}\) \(\newcommand{\uhat}{\widehat{\uvec}}\) \(\newcommand{\what}{\widehat{\wvec}}\) \(\newcommand{\Sighat}{\widehat{\Sigma}}\) \(\newcommand{\lt}{<}\) \(\newcommand{\gt}{>}\) \(\newcommand{\amp}{&}\) \(\definecolor{fillinmathshade}{gray}{0.9}\)

    Symmetric Matrices

    In this discussion, we will look at symmetric matrices and see that diagonalizing is a pleasure.  Recall that a matrix is symmetric if 

            A  =  AT 

    In other words the columns and rows of A are interchangeable.  The next theorem we state without proof.

    Theorem

    Let A be a symmetric matrix and p(x) be the characteristic polynomial.  Then all the roots of the characteristic polynomial p(x) are real.  In particular the eigenvalues of A are real and there are n linearly independent eigenvectors and A is diagonalizable.


    This says that a symmetric matrix with n linearly independent eigenvalues is always similar to a diagonal matrix.  As good as this may sound, even better is true.  First a definition. 

     

    Definition

    A matrix P is called orthogonal if its columns form an orthonormal set and call a matrix A orthogonally diagonalizable if it can be diagonalized by D  =  P-1AP with P an orthogonal matrix.


    Theorem

    If A is an n x n symmetric matrix, then any two eigenvectors that come from distinct eigenvalues are orthogonal.

     

    If we take each of the eigenvalues to be unit vectors, then the we have the following corollary. 

    Corollary

    Symmetric matrices with n distinct eigenvalues are orthogonally diagonalizable. 

     

    Proof of the Theorem

    We need to show that if v and w are eigenvectors corresponding to distinct real eigenvalues a and b, then v . w  =  0.  We have

            a(v . w)  =  (av) . w  =  (Av) . w

            =  v . (ATw)  =  v . (Aw) =  v . (bw)

            =  b(v . w)

    Hence

           (a - b)(v . w)  =  0

    since a and b are distinct, we can conclude that v and w are orthogonal.

    We have used that

            (Av) . w  =  v . (ATw)

    a fact that is left for you as an exercise.


    There is special property that holds for orthogonal matrices that is worth noting.

     

    Theorem

    Let P be an orthogonal matrix.  Then 

            P-1  =  PT

     

    Proof

    We need to show that if P is orthogonal, then

            PTP  =  I

    This follows immediately from the definition of orthogonal and matrix multiplication.  If vj is the jth column of P, then

            [PTP]ij  =  vi . vj 

    But since {v1, ..., vn} is an orthonormal set of vectors, we have

             vi . vj  = dij

     

    The above theorem is especially useful, since computing a transpose is much easier than computing an inverse.


    Example

    Orthogonally diagonalize 

           \( A = \begin{pmatrix} 7 & -2 & 0 \\ -2 & 6 & -2 \\ 0 & -2 & 5 \end{pmatrix} \)

    Solution

    We find that the eigenvalues of A are 3, 6, and 9.  To find the eigenvectors we find the null spaces.

           \( rref(3I - A) = rref\begin{pmatrix} -4 & 2 & 0 \\ 2 & -3 & 2 \\ 0 & 2 & -2 \end{pmatrix} = \begin{pmatrix} 1 & 0 & -\frac{1}{2} \\ 0 & 1 & -1 \\ 0 & 0 & 0 \end{pmatrix} \)

    An eigenvector is (1/2, 1,1).  We need to normalize this eigenvalue since it has magnitude 3/2.  Dividing by the magnitude gives the unit vector (1/3, 2/3, 2/3).  Next we have 

     

           \( rref(6I - A) = rref\begin{pmatrix} -1 & 2 & 0 \\ 2 & 0 & 2 \\ 0 & 2 & 1 \end{pmatrix} = \begin{pmatrix} 1 & 0 & 1 \\ 0 & 1 & \frac{1}{2} \\ 0 & 0 & 0 \end{pmatrix} \)

    An eigenvector is (-1, -1/2,1).  Again, we need to normalize this eigenvalue since it has magnitude 3/2.  Dividing by the magnitude gives the unit vector (-2/3, -1/3, 2/3).  Next we have 

           \( rref(9I - A) = rref\begin{pmatrix} 2 & 2 & 0 \\ 2 & 3 & 2 \\ 0 & 2 & 4 \end{pmatrix} = \begin{pmatrix} 1 & 0 & -2 \\ 0 & 1 & 2 \\ 0 & 0 & 0 \end{pmatrix} \)

    An eigenvector is (2, -2,1).  Again, we need to normalize this eigenvalue since it has magnitude 3.  Dividing by the magnitude gives the unit vector (2/3, -2/3, 1/3)

    We now have 

           \( P= \begin{pmatrix} -\frac{1}{3} & -\frac{2}{3} & \frac{2}{3} \\ \frac{2}{3} & -\frac{1}{3} & -\frac{2}{3} \\ \frac{2}{3} & \frac{2}{3} & \frac{1}{3} \end{pmatrix} \)     \( D = \begin{pmatrix} 3 & 0 & 0 \\ 0 & 6 & 0 \\ 0 & 9 & 0 \end{pmatrix} \)

    And we can write

           \( A = PDP^{-1} = PDP^{T} = \begin{pmatrix} -\frac{1}{3} & -\frac{2}{3} & \frac{2}{3} \\ \frac{2}{3} & -\frac{1}{3} & -\frac{2}{3} \\ \frac{2}{3} & \frac{2}{3} & \frac{1}{3} \end{pmatrix}  \begin{pmatrix} 3 & 0 & 0 \\ 0 & 6 & 0 \\ 0 & 9 & 0 \end{pmatrix} \begin{pmatrix} -\frac{1}{3} & \frac{2}{3} & \frac{2}{3} \\ -\frac{2}{3} & -\frac{1}{3} & \frac{2}{3} \\ \frac{2}{3} & -\frac{2}{3} & \frac{1}{3} \end{pmatrix}\)



    Back to the Matrices and Vectors Home Page

     

     

    Symmetric Matrices is shared under a CC BY license and was authored, remixed, and/or curated by LibreTexts.

    • Was this article helpful?