Folioby Interconnected
Log InSign Up

Diagonalization: Simplifying Matrices by Choice of Basis

A matrix is diagonalizable if and only if the sum of its geometric multiplicities equals the matrix size. We prove this criterion, give a step-by-step diagonalization procedure with applications to computing A^n, and prove Schur's theorem that every complex square matrix is unitarily triangularizable.

FO
Folio Official
March 1, 2026

1. Definition of Diagonalizability

Definition 1 (Diagonalizable matrix).
An n×n matrix A is called diagonalizable if there exists an invertible matrix P such that
P−1AP=​λ1​O​⋱​Oλn​​​
is a diagonal matrix.
Remark 2.
Diagonalizability is equivalent to the statement that, in a suitably chosen basis, A is represented by a diagonal matrix. The columns of P are eigenvectors, and the diagonal entries are the corresponding eigenvalues.

2. Conditions for Diagonalizability

Theorem 3 (Necessary and sufficient conditions for diagonalizability).
The following conditions on an n×n matrix A are equivalent:
  1. A is diagonalizable.

  2. A possesses n linearly independent eigenvectors.

  3. For every eigenvalue λi​, the geometric multiplicity equals the algebraic multiplicity.

  4. The characteristic polynomial pA​(λ) splits completely over K, and for each eigenvalue the geometric multiplicity equals the algebraic multiplicity.

Proof.
(1⇔2): The equation P−1AP=D is equivalent to AP=PD. Writing P=(v1​⋯vn​) and D=diag(λ1​,…,λn​), we see that Avi​=λi​vi​ for each i. The matrix P is invertible precisely when v1​,…,vn​ are linearly independent.

(2⇔3): The eigenspaces for distinct eigenvalues are always in direct sum, so the total count of linearly independent eigenvectors is the sum of the geometric multiplicities. This sum equals n if and only if each geometric multiplicity equals the corresponding algebraic multiplicity. □
Example 4 (A diagonalizable matrix).
Let A=(41​−21​). Its characteristic polynomial is pA​(λ)=λ2−5λ+6=(λ−2)(λ−3). Since A has two distinct eigenvalues, it is diagonalizable.

For λ1​=2: v1​=(1,1)T. For λ2​=3: v2​=(2,1)T. Setting P=(11​21​), we obtain
P−1AP=(20​03​).
Example 5 (A non-diagonalizable matrix).
Let A=(20​12​). The characteristic polynomial is pA​(λ)=(λ−2)2, so λ=2 has algebraic multiplicity 2. However, dimV2​=1, so the geometric multiplicity is strictly less than the algebraic multiplicity, and A is not diagonalizable.

3. The Diagonalization Procedure

To diagonalize an n×n matrix A, proceed as follows:

  1. Compute the characteristic polynomial pA​(λ)=det(A−λI).

  2. Solve pA​(λ)=0 to find the eigenvalues λ1​,…,λk​ and their algebraic multiplicities.

  3. For each eigenvalue λi​, compute a basis for the eigenspace Vλi​​=ker(A−λi​I).

  4. Check that the geometric multiplicity equals the algebraic multiplicity for every eigenvalue. (If not, A is not diagonalizable.)

  5. Form P by arranging the eigenvectors as columns; then D=P−1AP is diagonal.

4. Computing Powers of a Matrix

When A is diagonalizable, writing A=PDP−1 gives

An=PDnP−1=P​λ1n​O​⋱​Oλkn​​​P−1.

Example 6.
For A=(41​−21​) with P=(11​21​) and P−1=(−11​2−1​), we obtain
An=(11​21​)(2n0​03n​)(−11​2−1​)=(−2n+2⋅3n−2n+3n​2n+1−2⋅3n2n+1−3n​).

5. Triangularization

Theorem 7 (Schur's triangularization theorem).
Over K=C, every n×n matrix A is unitarily triangularizable: there exists a unitary matrix U such that U∗AU is upper triangular.
Remark 8.
Even matrices that cannot be diagonalized can always be triangularized (provided the base field is algebraically closed). The diagonal entries of the resulting upper triangular matrix are the eigenvalues of A, which confirms the identities trA=∑λi​ and detA=∏λi​.
Theorem 9.
If A is triangularized as T=P−1AP with T upper triangular, then Tn is again upper triangular with diagonal entries λin​. The factorization An=PTnP−1 therefore provides a method for computing powers even for non-diagonalizable matrices.
Linear AlgebraAlgebraTextbookDiagonalizationTriangularizationMatrix Powers
FO
Folio Official

Mathematics "between the lines" — exploring the intuition textbooks leave out, written in LaTeX on Folio.

1 followers·107 articles
Linear Algebra TextbookPart 9 of 13
Previous
Eigenvalues and Eigenvectors: Invariant Directions of Linear Maps
Next
Inner Product Spaces and Orthonormal Bases

Share your expertise with the world

Write articles with LaTeX support, build your audience, and earn from your knowledge.

Start Writing — It's Free

More from Folio Official

Folio Official·March 1, 2026

Matrices and Representation of Linear Maps

Every linear map between finite-dimensional spaces is uniquely represented by a matrix once bases are chosen, and composition of maps corresponds to matrix multiplication. We derive the change-of-basis formula, characterize invertible matrices, and show that matrix rank equals the rank of the corresponding linear map.

Linear AlgebraAlgebraTextbook
Folio Official·March 1, 2026

The Determinant: A Scalar Invariant of Square Matrices

The determinant of a square matrix is the unique scalar-valued function characterized by alternation and multilinearity. We construct it via the Leibniz formula, prove the product formula det(AB) = det(A)det(B), and derive Cramer's rule for solving linear systems.

Linear AlgebraAlgebraTextbook
Folio Official·March 1, 2026

Jordan Normal Form: Beyond Diagonalization

Every complex square matrix is similar to a Jordan normal form — a block-diagonal matrix of Jordan blocks that is unique up to block ordering. We prove existence and uniqueness, connect the Jordan form to the minimal polynomial, and apply it to compute the matrix exponential e^{tA}.

Linear AlgebraAlgebraTextbook
Folio Official·March 24, 2026

The Singular Value Decomposition: Structure of Arbitrary Matrices

Every real m x n matrix factors as A = U Sigma V^T, where U and V are orthogonal and Sigma is diagonal — this is the singular value decomposition. We prove existence, show how the SVD yields optimal low-rank approximations (Eckart–Young theorem), and construct the Moore–Penrose pseudoinverse for least-squares solutions.

Linear AlgebraAlgebraTextbook