Evanalysis
5.1Embedded interactionEstimated reading time: 15 min

5.1 Invertible matrices

Study invertible matrices through definitions, row reduction, equivalent formulations, and the algebra of inverses.

Course contents

MATH1030: Linear algebra I

Rigorous linear algebra notes on systems, matrices, structure, and proof, with interaction used only where it clarifies the mathematics.

Chapter 1Systems of equations1 sections
Chapter 4Solution structure1 sections
Chapter 5Invertibility1 sections

Invertibility is one of the first places where linear algebra becomes more than a procedure for solving one system. A square matrix is invertible exactly when it can be undone by another square matrix, and that idea turns out to be equivalent to many other statements: row reduction to the identity, consistency of every system Ax=bAx = b, linear independence of the columns, and the ability to write every vector as a linear combination of those columns.

This note develops those equivalences carefully. The goal is not only to know what the symbol A1A^{-1} means, but also to recognize when it exists and how to use it without guessing.

Left and right inverses

Before the square case, it is useful to separate two one-sided notions.

Definition

Left inverse and right inverse

Let AA be a p×qp \times q matrix.

  • A q×pq \times p matrix HH is a left inverse of AA if HA=IqHA = I_q.
  • A q×pq \times p matrix GG is a right inverse of AA if AG=IpAG = I_p.

These definitions matter because matrix multiplication is not commutative. For rectangular matrices, a left inverse and a right inverse need not both exist. The square case is special.

Definition

Invertible matrix

Let AA be a p×pp \times p square matrix. We say that AA is invertible if there exists a p×pp \times p matrix BB such that

BA=AB=Ip.BA = AB = I_p.

The matrix BB is called the inverse of AA, and we write B=A1B = A^{-1}.

Theorem

The inverse is unique

If BB is a left inverse of AA and CC is a right inverse of AA, then B=CB = C. So an invertible matrix has exactly one inverse.

Proof

Why the inverse is unique

What invertibility means

Invertibility is a reversibility statement. Applying AA changes a vector, but if AA is invertible then A1A^{-1} undoes that change exactly.

That is why the identity matrix appears in the definition. The identity matrix does nothing:

Ipx=xI_p x = x

for every compatible vector x. An inverse is precisely a matrix that brings you back to that unchanged state.

Worked example

A diagonal matrix is easy to invert

Let

D=diag(2,1,3).D = \operatorname{diag}(2, -1, 3).

Then

D1=diag(12,1,13).D^{-1} = \operatorname{diag}\left(\tfrac{1}{2}, -1, \tfrac{1}{3}\right).

This works because each diagonal entry is replaced by its reciprocal, and the off-diagonal zeros stay zero. Multiplying DD by D1D^{-1} gives I3I_3.

Row reduction and the inverse

The most practical way to test invertibility is to row-reduce. The key point has two parts:

  1. row-operation matrices are invertible, with inverse given by the reverse row operation;
  2. a square matrix is invertible exactly when it can be row-reduced to IpI_p.

Theorem

Row-operation matrices are invertible

If ρ\rho is a row operation on matrices with p rows, and ρˉ\bar{\rho} is the reverse row operation, then the corresponding row-operation matrices M[ρ]M[\rho] and M[ρˉ]M[\bar{\rho}] satisfy

M[ρ]1=M[ρˉ],M[ρˉ]1=M[ρ].M[\rho]^{-1} = M[\bar{\rho}], \qquad M[\bar{\rho}]^{-1} = M[\rho].

This gives a clean interpretation of row reduction: every row operation is actually multiplication on the left by an invertible matrix.

Theorem

Invertibility and row reduction

For a square matrix AA, the following are equivalent:

  • AA is invertible.
  • AA is row-equivalent to IpI_p.
  • AA is a product of p×pp \times p row-operation matrices.
  • AA is nonsingular.

The practical consequence is very concrete: if row operations transform AA to IpI_p, then those same operations, applied to [AIp][A | I_p], transform it to [IpA1][I_p | A^{-1}].

Read and try

Follow one inverse-by-row-reduction example

The live demo lets you step through [A | I] until the left block becomes I.

Start from [A | I]. If A is invertible, row reduction will turn the left block into I.

121100
011010
234001

The live demo above is the shortest way to see the logic. It is not the definition. It is the computational method that matches the definition.

Read and try

Trace one full row-reduction path

The live stepper walks through one complete elimination path, showing the row operation, the pivot you are focusing on, and the matrix produced at each step.

1224
1335
2656

Row operation

Choose the first pivot in column 1.

What to notice

Column 1 already has a convenient pivot 1 in the first row, so we do not need a row swap.

Start with the augmented matrix. The first pivot should help us clear the entries underneath it.

The second widget shows the shape of a full elimination path. In an invertible case, the left block eventually becomes IpI_p, and that is the moment when the right block becomes the inverse.

Equivalent formulations

Invertibility is useful because it has a dictionary of equivalent conditions. This is the main bridge between algebra, row reduction, and systems of linear equations.

Theorem

Equivalent ways to recognize invertibility

Let AA be a p×pp \times p matrix. Then the following statements are equivalent:

  1. AA is invertible.
  2. AA is row-equivalent to IpI_p.
  3. AA is a product of row-operation matrices.
  4. AA has a left inverse.
  5. AA has a right inverse.
  6. AA is nonsingular.
  7. For every column vector b with p entries, the system Ax=bAx = b is consistent.
  8. For every column vector b with p entries, the system Ax=bAx = b has the unique solution x=A1bx = A^{-1}b.

Two of these statements are especially important in practice.

  • Statement 7 says that the columns of AA span RpR^p.
  • Statement 8 says that invertibility gives you a complete solution formula, not just existence.

That is why invertibility is the exact algebraic condition behind solving a linear system by a matrix inverse.

Row-equivalence through invertible matrices

We can push the row-operation viewpoint one step further. Instead of thinking about row-equivalence as a long list of elementary moves, package the whole list into one invertible matrix on the left.

Theorem

Row-equivalence is left multiplication by an invertible matrix

Suppose AA and BB are matrices with p rows. Then the following are equivalent:

  • AA and BB are row-equivalent.
  • There exists an invertible p×pp \times p matrix GG such that
B=GA.B = GA.

Moreover, once B=GAB = GA, we also have

A=G1B.A = G^{-1}B.

This theorem is not a new computational trick. It is a cleaner language for the same phenomenon. A sequence of row operations can always be compressed into one invertible matrix GG, and the reverse row operations are encoded by G1G^{-1}.

Worked example

Reading a row-equivalence as one matrix equality

Let

A=[101123011],G=[100110001].A = \begin{bmatrix} 1 & 0 & 1 \\ 1 & 2 & 3 \\ 0 & 1 & 1 \end{bmatrix}, \qquad G = \begin{bmatrix} 1 & 0 & 0 \\ -1 & 1 & 0 \\ 0 & 0 & 1 \end{bmatrix}.

The matrix GG is the row-operation matrix for the move

R2R2R1.R_2 \leftarrow R_2 - R_1.

So

GA=[101022011].GA = \begin{bmatrix} 1 & 0 & 1 \\ 0 & 2 & 2 \\ 0 & 1 & 1 \end{bmatrix}.

If we call this new matrix BB, then B=GAB = GA. That single equation records the entire row operation. Since GG is invertible, AA and BB are row-equivalent.

The gain is conceptual. Once you know that row-equivalence means multiplication by an invertible matrix on the left, you can explain many invariants in one line instead of by repeating row-operation arguments.

Theorem

Row operations preserve linear relations among corresponding columns

Let AA and BB be row-equivalent p×qp \times q matrices, and write their columns as

A=[a1  a2    aq],B=[b1  b2    bq].A = [a_1 \; a_2 \; \cdots \; a_q], \qquad B = [b_1 \; b_2 \; \cdots \; b_q].

If

aj=α1ak1+α2ak2++αnakn,a_j = \alpha_1 a_{k_1} + \alpha_2 a_{k_2} + \cdots + \alpha_n a_{k_n},

then

bj=α1bk1+α2bk2++αnbkn.b_j = \alpha_1 b_{k_1} + \alpha_2 b_{k_2} + \cdots + \alpha_n b_{k_n}.

In particular, linear dependence and linear independence among corresponding columns are preserved by row-equivalence.

The proof is short once B=GAB = GA is known. Multiply the relation for the columns of AA by GG. Because matrix multiplication is linear,

Gaj=α1Gak1+α2Gak2++αnGakn,Ga_j = \alpha_1 G a_{k_1} + \alpha_2 G a_{k_2} + \cdots + \alpha_n G a_{k_n},

which is exactly the corresponding relation amongst the columns of BB.

This is the bridge from row reduction to column language. Row operations change the actual columns, but they do not change which columns are redundant or which column relations are forced by the others.

Why the reduced row-echelon form is unique

The reduced row-echelon form in a row-equivalence class is unique. That fact is easy to overlook, but it is what makes later definitions mathematically legitimate.

Theorem

A row-equivalence class has exactly one reduced row-echelon form

Suppose AA is a matrix, and suppose BB and CC are both reduced row-echelon forms. If BB is row-equivalent to AA and CC is row-equivalent to AA, then

B=C.B = C.

A standard proof uses induction on the rank. The basic strategy is:

  1. compare the pivot columns from left to right,
  2. use preserved linear relations to force the same pivot positions, and then
  3. show that every free column must have the same coefficients in terms of the pivot columns.

So reduced row-echelon form is not merely a convenient final answer. It is the final answer inside a row-equivalence class.

Definition

Rank

The rank of a matrix is the number of pivots in its reduced row-echelon form.

This definition works only because the reduced row-echelon form is unique. If different reduction paths could produce different reduced forms with different numbers of pivots, then rank would depend on the calculation. The uniqueness theorem rules that out.

Column independence and linear combinations

The same source also recasts invertibility in terms of columns.

Theorem

Invertibility and the columns of a square matrix

For a p×pp \times p matrix AA, the following are equivalent:

  • AA is invertible.
  • The columns of AA are linearly independent.
  • Every column vector in RpR^p is a linear combination of the columns of AA.

These are not separate facts. They are three ways of reading the same structural statement.

If the columns are linearly independent, then no column is redundant. If they span RpR^p, then every target vector can be built from them. For a square matrix, those two conditions coincide exactly when the matrix is invertible.

Why the transpose also matters

Invertibility behaves well under transpose.

Theorem

Transpose and powers

If AA is invertible, then:

  • AtA^t is invertible, and (At)1=(A1)t(A^t)^{-1} = (A^{-1})^t.
  • AnA^n is invertible for every integer n, and (An)1=(A1)n(A^n)^{-1} = (A^{-1})^n.

The transpose result is useful when you want to turn a statement about columns into a statement about rows. The power rule is useful when a repeated transformation appears in a calculation.

Worked example

Worked example

Find an inverse by row reduction

Let

A=[1235].A = \begin{bmatrix} 1 & 2 \\ 3 & 5 \end{bmatrix}.

Start from [AI2][A | I_2]:

[12103501].\left[\begin{array}{cc|cc} 1 & 2 & 1 & 0 \\ 3 & 5 & 0 & 1 \end{array}\right].

Eliminate the entry below the first pivot:

[12100131].\left[\begin{array}{cc|cc} 1 & 2 & 1 & 0 \\ 0 & -1 & -3 & 1 \end{array}\right].

Now scale the second row and clear the entry above the second pivot:

[10520131].\left[\begin{array}{cc|cc} 1 & 0 & -5 & 2 \\ 0 & 1 & 3 & -1 \end{array}\right].

So

A1=[5231].A^{-1} = \begin{bmatrix} -5 & 2 \\ 3 & -1 \end{bmatrix}.

The computation is not the point by itself. The point is that the right block of the augmented matrix records the inverse because the left block has been driven to the identity.

Common mistakes

Common mistake

Do not confuse one-sided inverses in the rectangular case

For a non-square matrix, having a left inverse does not automatically mean it has a right inverse. The square case is special: once an inverse exists, it is both a left inverse and a right inverse, and it is unique.

Common mistake

Do not guess invertibility from appearance

A matrix can look simple and still fail to be invertible. The correct test is to row-reduce it, or to use one of the equivalent conditions above.

Quick checks

Quick check

If AA is invertible, what is A1AA^{-1}A?

Use the defining property of an inverse.

Solution

Answer

Quick check

If AA is invertible, can the homogeneous system Ax=0Ax = 0 have a nonzero solution?

Use the unique-solution statement.

Solution

Answer

Quick check

If AA is invertible, is AtA^t invertible?

Use the transpose rule above.

Solution

Answer

Quick check

If B=GAB = GA with GG invertible and the columns of AA satisfy a3=2a1a2a_3 = 2a_1 - a_2, what relation must hold among the columns of BB?

Keep the same coefficients and use bj=Gajb_j = Ga_j.

Solution

Answer

Quick check

Why does uniqueness of RREF matter when defining rank?

Answer in one sentence using the phrase “well defined.”

Solution

Answer

Exercise

Quick check

Suppose AA is invertible and AB=IpAB = I_p. Prove that B=A1B = A^{-1}.

Use the fact that the inverse of AA is unique.

Solution

Guided solution

Read this first

This page depends especially on 2.3 Gaussian elimination and RREF, 3.1 Matrix multiplication and identity matrices, and 3.2 Transpose and special matrices.

Section mastery checkpoint

Answer each question correctly to complete this section checkpoint. Correct progress: 0%.

Skills: invertible-matrix, rank, equivalent-statements

If A is invertible n×n, which statement must hold?

Attempts used: 0

Attempts remaining: Unlimited attempts

Preview does not consume an attempt.

Submit records a graded attempt.

  • Preview should confirm which option you selected before the attempt is consumed.

Key terms in this unit