Loading [MathJax]/jax/output/HTML-CSS/jax.js
Skip to main content
Library homepage
 

Text Color

Text Size

 

Margin Size

 

Font Type

Enable Dyslexic Font
Mathematics LibreTexts

2.5: Matrix Inverses

( \newcommand{\kernel}{\mathrm{null}\,}\)

Three basic operations on matrices, addition, multiplication, and subtraction, are analogs for matrices of the same operations for numbers. In this section we introduce the matrix analog of numerical division.

To begin, consider how a numerical equation ax=b is solved when a and b are known numbers. If a=0, there is no solution (unless b=0). But if a0, we can multiply both sides by the inverse a1=1a to obtain the solution x=a1b. Of course multiplying by a1 is just dividing by a, and the property of a1 that makes this work is that a1a=1. Moreover, we saw in Section [sec:2_2] that the role that 1 plays in arithmetic is played in matrix algebra by the identity matrix I. This suggests the following definition.

Matrix Inverses004202 If A is a square matrix, a matrix B is called an inverse of A if and only if

AB=I and BA=I

A matrix A that has an inverse is called an invertible matrix.

004207 Show that B=[1110] is an inverse of A=[0111].

Compute AB and BA.

AB=[0111][1110]=[1001]BA=[1110][0111]=[1001]

Hence AB=I=BA, so B is indeed an inverse of A.

004217 Show that A=[0013] has no inverse.

Let B=[abcd] denote an arbitrary 2×2 matrix. Then

AB=[0013][abcd]=[00a+3cb+3d]

so AB has a row of zeros. Hence AB cannot equal I for any B.

The argument in Example [exa:004217] shows that no zero matrix has an inverse. But Example [exa:004217] also shows that, unlike arithmetic, it is possible for a nonzero matrix to have no inverse. However, if a matrix does have an inverse, it has only one.

004227 If B and C are both inverses of A, then B=C.

Since B and C are both inverses of A, we have CA=I=AB. Hence

B=IB=(CA)B=C(AB)=CI=C

If A is an invertible matrix, the (unique) inverse of A is denoted A1. Hence A1 (when it exists) is a square matrix of the same size as A with the property that

AA1=I and A1A=I

These equations characterize A1 in the following sense:

Inverse Criterion: If somehow a matrix B can be found such that AB=I and BA=I, then A is invertible and B is the inverse of A; in symbols, B=A1.

This is a way to verify that the inverse of a matrix exists. Example [exa:004241] and Example [exa:004261] offer illustrations.

004241 If A=[0111], show that A3=I and so find A1.

We have A2=[0111][0111]=[1110], and so

A3=A2A=[1110][0111]=[1001]=I

Hence A3=I, as asserted. This can be written as A2A=I=AA2, so it shows that A2 is the inverse of A. That is, A1=A2=[1110].

The next example presents a useful formula for the inverse of a 2×2 matrix A=[abcd] when it exists. To state it, we define the determinant detA and the adjugate adjA of the matrix A as follows:

det[abcd]=adbc, and adj[abcd]=[dbca]

004261 If A=[abcd], show that A has an inverse if and only if detA0, and in this case

A1=1detAadjA

For convenience, write e=detA=adbc and B=adjA=[dbca]. Then AB=eI=BA as the reader can verify. So if e0, scalar multiplication by 1e gives

A(1eB)=I=(1eB)A

Hence A is invertible and A1=1eB. Thus it remains only to show that if A1 exists, then e0.

We prove this by showing that assuming e=0 leads to a contradiction. In fact, if e=0, then AB=eI=0, so left multiplication by A1 gives A1AB=A10; that is, IB=0, so B=0. But this implies that a, b, c, and d are all zero, so A=0, contrary to the assumption that A1 exists.

As an illustration, if A=[2438] then detA=284(3)=280. Hence A is invertible and A1=1detAadjA=128[8432], as the reader is invited to verify.

The determinant and adjugate will be defined in Chapter [chap:3] for any square matrix, and the conclusions in Example [exa:004261] will be proved in full generality.

Inverses and Linear Systems

Matrix inverses can be used to solve certain systems of linear equations. Recall that a system of linear equations can be written as a single matrix equation

Ax=b

where A and b are known and x is to be determined. If A is invertible, we multiply each side of the equation on the left by A1 to get

A1Ax=A1bIx=A1bx=A1b

This gives the solution to the system of equations (the reader should verify that x=A1b really does satisfy Ax=b). Furthermore, the argument shows that if x is any solution, then necessarily x=A1b, so the solution is unique. Of course the technique works only when the coefficient matrix A has an inverse. This proves Theorem [thm:004292].

004292 Suppose a system of n equations in n variables is written in matrix form as

Ax=b

If the n×n coefficient matrix A is invertible, the system has the unique solution

x=A1b

004298 Use Example [exa:004261] to solve the system {5x13x2=47x1+4x2=8.

In matrix form this is Ax=b where A=[5374], x=[x1x2], and b=[48]. Then detA=54(3)7=41, so A is invertible and A1=141[4375] by Example [exa:004261]. Thus Theorem [thm:004292] gives

x=A1b=141[4375][48]=141[868]

so the solution is x1=841 and x2=6841.

An Inversion Method

If a matrix A is n×n and invertible, it is desirable to have an efficient technique for finding the inverse. The following procedure will be justified in Section [sec:2_5].

Matrix Inversion Algorithm004348 If A is an invertible (square) matrix, there exists a sequence of elementary row operations that carry A to the identity matrix I of the same size, written AI. This same series of row operations carries I to A1; that is, IA1. The algorithm can be summarized as follows:

[AI][IA1]

where the row operations on A and I are carried out simultaneously.

004354 Use the inversion algorithm to find the inverse of the matrix

A=[271141130]

Apply elementary row operations to the double matrix

[AI]=[271100141010130001]

so as to carry A to I. First interchange rows 1 and 2.

[141010271100130001]

Next subtract 2 times row 1 from row 2, and subtract row 1 from row 3.

[141010013120011011]

Continue to reduced row-echelon form.

[1011470013120002111]

[1003232112010121232001121212]

Hence A1=12[3311113111], as is readily verified.

Given any n×n matrix A, Theorem [thm:001017] shows that A can be carried by elementary row operations to a matrix R in reduced row-echelon form. If R=I, the matrix A is invertible (this will be proved in the next section), so the algorithm produces A1. If RI, then R has a row of zeros (it is square), so no system of linear equations Ax=b can have a unique solution. But then A is not invertible by Theorem [thm:004292]. Hence, the algorithm is effective in the sense conveyed in Theorem [thm:004371].

004371 If A is an n×n matrix, either A can be reduced to I by elementary row operations or it cannot. In the first case, the algorithm produces A1; in the second case, A1 does not exist.

Properties of Inverses

The following properties of an invertible matrix are used everywhere.

Cancellation Laws004379 Let A be an invertible matrix. Show that:

  1. If AB=AC, then B=C.
  2. If BA=CA, then B=C.

Given the equation AB=AC, left multiply both sides by A1 to obtain A1AB=A1AC. Thus IB=IC, that is B=C. This proves (1) and the proof of (2) is left to the reader.

Properties (1) and (2) in Example [exa:004379] are described by saying that an invertible matrix can be “left cancelled” and “right cancelled”, respectively. Note however that “mixed” cancellation does not hold in general: If A is invertible and AB=CA, then B and C may not be equal, even if both are 2×2. Here is a specific example:

A=[1101], B=[0012],C=[1111]

Sometimes the inverse of a matrix is given by a formula. Example [exa:004261] is one illustration; Example [exa:004397] and Example [exa:004423] provide two more. The idea is the Inverse Criterion: If a matrix B can be found such that AB=I=BA, then A is invertible and A1=B.

004397 If A is an invertible matrix, show that the transpose AT is also invertible. Show further that the inverse of AT is just the transpose of A1; in symbols, (AT)1=(A1)T.

A1 exists (by assumption). Its transpose (A1)T is the candidate proposed for the inverse of AT. Using the inverse criterion, we test it as follows:

AT(A1)T=(A1A)T=IT=I(A1)TAT=(AA1)T=IT=I

Hence (A1)T is indeed the inverse of AT; that is, (AT)1=(A1)T.

004423 If A and B are invertible n×n matrices, show that their product AB is also invertible and (AB)1=B1A1.

We are given a candidate for the inverse of AB, namely B1A1. We test it as follows:

(B1A1)(AB)=B1(A1A)B=B1IB=B1B=I(AB)(B1A1)=A(BB1)A1=AIA1=AA1=I

Hence B1A1 is the inverse of AB; in symbols, (AB)1=B1A1.

We now collect several basic properties of matrix inverses for reference.

004442 All the following matrices are square matrices of the same size.

  1. I is invertible and I1=I.
  2. If A is invertible, so is A1, and (A1)1=A.
  3. If A and B are invertible, so is AB, and (AB)1=B1A1.
  4. If A1,A2,,Ak are all invertible, so is their product A1A2Ak, and

    (A1A2Ak)1=A1kA12A11.

  5. If A is invertible, so is Ak for any k1, and (Ak)1=(A1)k.
  6. If A is invertible and a0 is a number, then aA is invertible and (aA)1=1aA1.
  7. If A is invertible, so is its transpose AT, and (AT)1=(A1)T.
  1. This is an immediate consequence of the fact that I2=I.
  2. The equations AA1=I=A1A show that A is the inverse of A1; in symbols, (A1)1=A.
  3. This is Example [exa:004423].
  4. Use induction on k. If k=1, there is nothing to prove, and if k=2, the result is property 3. If k>2, assume inductively that (A1A2Ak1)1=A1k1A12A11. We apply this fact together with property 3 as follows:

    [A1A2Ak1Ak]1=[(A1A2Ak1)Ak]1=A1k(A1A2Ak1)1=A1k(A1k1A12A11)

  5. This is property 4 with A1=A2==Ak=A.
  6. This is left as Exercise [ex:ex2_4_29].
  7. This is Example [exa:004397].

The reversal of the order of the inverses in properties 3 and 4 of Theorem [thm:004442] is a consequence of the fact that matrix multiplication is not commutative. Another manifestation of this comes when matrix equations are dealt with. If a matrix equation B=C is given, it can be left-multiplied by a matrix A to yield AB=AC. Similarly, right-multiplication gives BA=CA. However, we cannot mix the two: If B=C, it need not be the case that AB=CA even if A is invertible, for example, A=[1101], B=[0010]=C.

Part 7 of Theorem [thm:004442] together with the fact that (AT)T=A gives

004537 A square matrix A is invertible if and only if AT is invertible.

004541 Find A if (AT2I)1=[2110].

By Theorem [thm:004442](2) and Example [exa:004261], we have

(AT2I)=[(AT2I)1]1=[2110]1=[0112]

Hence AT=2I+[0112]=[2114], so A=[2114] by Theorem [thm:004442](7).

The following important theorem collects a number of conditions all equivalent1 to invertibility. It will be referred to frequently below.

Inverse Theorem004553 The following conditions are equivalent for an n×n matrix A:

  1. A is invertible.
  2. The homogeneous system Ax=0 has only the trivial solution x=0.
  3. A can be carried to the identity matrix In by elementary row operations.
  4. The system Ax=b has at least one solution x for every choice of column b.
  5. There exists an n×n matrix C such that AC=In.

We show that each of these conditions implies the next, and that (5) implies (1).

(1) (2). If A1 exists, then Ax=0 gives x=Inx=A1Ax=A10=0.

(2) (3). Assume that (2) is true. Certainly AR by row operations where R is a reduced, row-echelon matrix. It suffices to show that R=In. Suppose that this is not the case. Then R has a row of zeros (being square). Now consider the augmented matrix [A0] of the system Ax=0. Then [A0][R0] is the reduced form, and [R0] also has a row of zeros. Since R is square there must be at least one nonleading variable, and hence at least one parameter. Hence the system Ax=0 has infinitely many solutions, contrary to (2). So R=In after all.

(3) (4). Consider the augmented matrix [Ab] of the system Ax=b. Using (3), let AIn by a sequence of row operations. Then these same operations carry [Ab][Inc] for some column c. Hence the system Ax=b has a solution (in fact unique) by gaussian elimination. This proves (4).

(4) (5). Write In=[e1e2en] where e1,e2,,en are the columns of In. For each j=1,2,,n, the system Ax=ej has a solution cj by (4), so Acj=ej. Now let C=[c1c2cn] be the n×n matrix with these matrices cj as its columns. Then Definition [def:003447] gives (5):

AC=A[c1c2cn]=[Ac1Ac2Acn]=[e1e2en]=In

(5) (1). Assume that (5) is true so that AC=In for some matrix C. Then Cx=0 implies x=0 (because x=Inx=ACx=A0=0). Thus condition (2) holds for the matrix C rather than A. Hence the argument above that (2) (3) (4) (5) (with A replaced by C) shows that a matrix C exists such that CC=In. But then

A=AIn=A(CC)=(AC)C=InC=C

Thus CA=CC=In which, together with AC=In, shows that C is the inverse of A. This proves (1).

The proof of (5) (1) in Theorem [thm:004553] shows that if AC=I for square matrices, then necessarily CA=I, and hence that C and A are inverses of each other. We record this important fact for reference.

004612 If A and C are square matrices such that AC=I, then also CA=I. In particular, both A and C are invertible, C=A1, and A=C1.

Here is a quick way to remember Corollary [cor:004612]. If A is a square matrix, then

  1. If AC=I then C=A1.
  2. If CA=I then C=A1.

Observe that Corollary [cor:004612] is false if A and C are not square matrices. For example, we have

[121111][111101]=I2 but [111101][121111]I3

In fact, it is verified in the footnote on page that if AB=Im and BA=In, where A is m×n and B is n×m, then m=n and A and B are (square) inverses of each other.

An n×n matrix A has rankn if and only if (3) of Theorem [thm:004553] holds. Hence

004623 An n×n matrix A is invertible if and only if rankA=n.

Here is a useful fact about inverses of block matrices.

004627 Let P=[AX0B] and Q=[A0YB] be block matrices where A is m×m and B is n×n (possibly mn).

  1. Show that P is invertible if and only if A and B are both invertible. In this case, show that

    P1=[A1A1XB10B1]

  2. Show that Q is invertible if and only if A and B are both invertible. In this case, show that

    Q1=[A10B1YA1B1]

We do (a.) and leave (b.) for the reader.

  1. If A1 and B1 both exist, write R=[A1A1XB10B1]. Using block multiplication, one verifies that PR=Im+n=RP, so P is invertible, and P1=R. Conversely, suppose that P is invertible, and write P1=[CVWD] in block form, where C is m×m and D is n×n.

    Then the equation PP1=In+m becomes

    [AX0B][CVWD]=[AC+XWAV+XDBWBD]=Im+n=[Im00In]

    using block notation. Equating corresponding blocks, we find

    AC+XW=Im,BW=0, and BD=In

    Hence B is invertible because BD=In (by Corollary [cor:004537]), then W=0 because BW=0, and finally, AC=Im (so A is invertible, again by Corollary [cor:004537]).

Inverses of Matrix Transformations

Let T=TA:RnRn denote the matrix transformation induced by the n×n matrix A. Since A is square, it may very well be invertible, and this leads to the question:

What does it mean geometrically for T that A is invertible?

To answer this, let T=TA1:RnRn denote the transformation induced by A1. Then

T[T(x)]=A1[Ax]=Ix=xfor all x in RnT[T(x)]=A[A1x]=Ix=x

The first of these equations asserts that, if T carries x to a vector T(x), then T carries T(x) right back to x; that is T “reverses” the action of T. Similarly T “reverses” the action of T. Conditions ([eq:inverse1]) can be stated compactly in terms of composition:

TT=1Rn and TT=1Rn

When these conditions hold, we say that the matrix transformation T is an inverse of T, and we have shown that if the matrix A of T is invertible, then T has an inverse (induced by A1).

The converse is also true: If T has an inverse, then its matrix A must be invertible. Indeed, suppose S:RnRn is any inverse of T, so that ST=1Rn and TS=1Rn. It can be shown that S is also a matrix transformation. If B is the matrix of S, we have

BAx=S[T(x)]=(ST)(x)=1Rn(x)=x=Inx for all x in Rn

It follows by Theorem [thm:002985] that BA=In, and a similar argument shows that AB=In. Hence A is invertible with A1=B. Furthermore, the inverse transformation S has matrix A1, so S=T using the earlier notation. This proves the following important theorem.

004693 Let T:RnRn denote the matrix transformation induced by an n×n matrix A. Then

A is invertible if and only if T has an inverse.

In this case, T has exactly one inverse (which we denote as T1), and T1:RnRn is the transformation induced by the matrix A1. In other words

(TA)1=TA1

The geometrical relationship between T and T1 is embodied in equations ([eq:inverse1]) above:

T1[T(x)]=x and T[T1(x)]=x for all x in Rn

These equations are called the fundamental identities relating T and T1. Loosely speaking, they assert that each of T and T1 “reverses” or “undoes” the action of the other.

This geometric view of the inverse of a linear transformation provides a new way to find the inverse of a matrix A. More precisely, if A is an invertible matrix, we proceed as follows:

  1. Let T be the linear transformation induced by A.
  2. Obtain the linear transformation T1 which “reverses” the action of T.
  3. Then A1 is the matrix of T1.

Here is an example.

004725

Find the inverse of A=[0110] by viewing it as a linear transformation R2R2.

If x=[xy] the vector Ax=[0110][xy]=[yx] is the result of reflecting x in the line y=x (see the diagram). Hence, if Q1:R2R2 denotes reflection in the line y=x, then A is the matrix of Q1. Now observe that Q1 reverses itself because reflecting a vector x twice results in x. Consequently Q11=Q1. Since A1 is the matrix of Q11 and A is the matrix of Q, it follows that A1=A. Of course this conclusion is clear by simply observing directly that A2=I, but the geometric method can often work where these other methods may be less straightforward.


  1. If p and q are statements, we say that p implies q (written pq) if q is true whenever p is true. The statements are called equivalent if both pq and qp (written pq, spoken “p if and only if q”). See Appendix [chap:appbproofs].↩

This page titled 2.5: Matrix Inverses is shared under a CC BY-NC-SA 4.0 license and was authored, remixed, and/or curated by W. Keith Nicholson (Lyryx Learning Inc.) via source content that was edited to the style and standards of the LibreTexts platform.

Support Center

How can we help?