Diagonalization
Simplify matrices by changing to the eigenvector basis
You have learned that eigenvalues and eigenvectors reveal the special directions where a matrix acts simply, just by scaling. Now comes the payoff: if a matrix has enough eigenvectors, you can completely simplify how you work with it. Instead of dealing with a complicated transformation, you can change your perspective to one where the matrix becomes diagonal, with all its action happening along the coordinate axes. This process is called diagonalization, and it transforms difficult matrix computations into trivially easy ones.
Think about driving through a city on a grid of streets versus driving through a maze of winding roads. The grid makes navigation simple because everything aligns with your natural sense of “forward” and “sideways.” Diagonalization does something similar: it finds a coordinate system where the matrix’s action is as simple as possible, stretching or compressing along perpendicular directions with no mixing between them.
In this lesson, you will learn when a matrix can be diagonalized, how to find the change-of-basis matrix that accomplishes this, and why this is so powerful for computing matrix powers and solving applied problems.
Core Concepts
What Is Diagonalization?
A square matrix $A$ is diagonalizable if there exists an invertible matrix $P$ and a diagonal matrix $D$ such that
$$A = PDP^{-1}$$
Equivalently, you can rearrange this as $P^{-1}AP = D$, which shows that $D$ is what $A$ “looks like” when you change to a new coordinate system described by $P$.
The matrices in this equation have specific meanings:
- $D$ is a diagonal matrix with the eigenvalues of $A$ on its diagonal
- $P$ is formed by placing the corresponding eigenvectors of $A$ as its columns
- The columns of $P$ must be in the same order as the eigenvalues in $D$
For example, if $A$ has eigenvalues $\lambda_1 = 3$ and $\lambda_2 = 5$ with corresponding eigenvectors $\vec{v}_1$ and $\vec{v}_2$, then:
$$D = \begin{pmatrix} 3 & 0 \ 0 & 5 \end{pmatrix}, \quad P = \begin{pmatrix} | & | \ \vec{v}_1 & \vec{v}_2 \ | & | \end{pmatrix}$$
The equation $A = PDP^{-1}$ tells you that applying $A$ is equivalent to: first changing coordinates (multiply by $P^{-1}$), then scaling along the axes by the eigenvalues (multiply by $D$), and finally changing back to the original coordinates (multiply by $P$).
When Is a Matrix Diagonalizable?
Not every matrix can be diagonalized. The key requirement is having enough linearly independent eigenvectors.
Diagonalizability Criterion: An $n \times n$ matrix $A$ is diagonalizable if and only if $A$ has $n$ linearly independent eigenvectors.
This happens in several important cases:
-
Distinct eigenvalues: If $A$ has $n$ distinct eigenvalues, then $A$ is definitely diagonalizable. Eigenvectors corresponding to different eigenvalues are always linearly independent.
-
Repeated eigenvalues with full eigenspaces: If an eigenvalue $\lambda$ has algebraic multiplicity $k$ (it appears $k$ times as a root of the characteristic polynomial), then $A$ is diagonalizable only if the geometric multiplicity (dimension of the eigenspace) also equals $k$. You need $k$ linearly independent eigenvectors for that eigenvalue.
-
Symmetric matrices: Real symmetric matrices are always diagonalizable. Even better, they can be diagonalized using an orthonormal basis of eigenvectors.
When a matrix fails to be diagonalizable, it is called defective. This happens when there are not enough eigenvectors to form a basis. The classic example is the matrix $\begin{pmatrix} 1 & 1 \ 0 & 1 \end{pmatrix}$, which has only one eigenvalue ($\lambda = 1$) with algebraic multiplicity 2, but only one linearly independent eigenvector.
Similar Matrices
Two matrices $A$ and $B$ are called similar if there exists an invertible matrix $P$ such that
$$B = P^{-1}AP$$
Similar matrices represent the same linear transformation, just described in different coordinate systems. They share many important properties:
- Same eigenvalues (with the same multiplicities)
- Same determinant
- Same trace
- Same rank
When we diagonalize $A$, we are finding a diagonal matrix $D$ that is similar to $A$. The diagonal matrix is the simplest representative of all matrices similar to $A$.
The Diagonalization Process
Here is how to diagonalize a matrix $A$:
Step 1: Find the eigenvalues.
Solve the characteristic equation $\det(A - \lambda I) = 0$.
Step 2: Find the eigenvectors.
For each eigenvalue $\lambda_i$, solve $(A - \lambda_i I)\vec{v} = \vec{0}$ to find the eigenvectors.
Step 3: Check if diagonalization is possible.
Verify that you have $n$ linearly independent eigenvectors total. If not, $A$ is not diagonalizable.
Step 4: Form $P$ and $D$.
- Place the eigenvectors as columns of $P$
- Place the corresponding eigenvalues on the diagonal of $D$ (in the same order)
Step 5: Verify (optional but recommended).
Check that $AP = PD$, which is equivalent to $A = PDP^{-1}$.
Powers of Diagonal Matrices
One of the main reasons diagonalization is so useful is that powers of diagonal matrices are trivially easy to compute. If
$$D = \begin{pmatrix} \lambda_1 & 0 & \cdots & 0 \ 0 & \lambda_2 & \cdots & 0 \ \vdots & \vdots & \ddots & \vdots \ 0 & 0 & \cdots & \lambda_n \end{pmatrix}$$
then
$$D^k = \begin{pmatrix} \lambda_1^k & 0 & \cdots & 0 \ 0 & \lambda_2^k & \cdots & 0 \ \vdots & \vdots & \ddots & \vdots \ 0 & 0 & \cdots & \lambda_n^k \end{pmatrix}$$
You simply raise each diagonal entry to the power $k$. There is no mixing between entries because the off-diagonal entries are all zero.
Computing $A^n$ Using Diagonalization
Here is where diagonalization really pays off. If $A = PDP^{-1}$, then:
$$A^2 = (PDP^{-1})(PDP^{-1}) = PD(P^{-1}P)DP^{-1} = PD \cdot I \cdot DP^{-1} = PD^2P^{-1}$$
The middle terms $P^{-1}P$ collapse to the identity matrix. Continuing this pattern:
$$A^n = PD^nP^{-1}$$
This is enormously powerful. Instead of multiplying $A$ by itself $n$ times (which would require $n-1$ matrix multiplications), you:
- Compute $D^n$ by raising each diagonal entry to the $n$-th power
- Multiply $PD^nP^{-1}$ (just two matrix multiplications, regardless of $n$)
For large $n$, this is vastly more efficient. Computing $A^{100}$ directly would require 99 matrix multiplications. Using diagonalization, it requires the same amount of work as computing $A^2$.
Symmetric Matrices and Orthogonal Diagonalization
Real symmetric matrices (where $A = A^T$) have two remarkable properties:
-
All eigenvalues are real. Even though the characteristic polynomial might seem like it could have complex roots, symmetry guarantees they are all real numbers.
-
Eigenvectors for different eigenvalues are orthogonal. If $\vec{v}_1$ and $\vec{v}_2$ are eigenvectors for different eigenvalues, then $\vec{v}_1 \cdot \vec{v}_2 = 0$.
This means symmetric matrices can always be diagonalized, and moreover, you can choose the eigenvectors to form an orthonormal basis. When you do this, the matrix $P$ becomes an orthogonal matrix, meaning $P^{-1} = P^T$.
This is called orthogonal diagonalization:
$$A = PDP^T$$
where $P$ is orthogonal (its columns are orthonormal eigenvectors) and $D$ is diagonal.
Orthogonal diagonalization is particularly nice because computing $P^{-1}$ is as easy as taking the transpose, and orthogonal matrices preserve lengths and angles.
Notation and Terminology
| Term | Meaning | Example |
|---|---|---|
| Diagonalizable | Can be written as $PDP^{-1}$ where $D$ is diagonal | A matrix with $n$ linearly independent eigenvectors |
| Similar matrices | $B = P^{-1}AP$ for some invertible $P$ | Similar matrices have the same eigenvalues |
| $P$ | Matrix of eigenvectors (as columns) | Columns are eigenvectors in order matching $D$ |
| $D$ | Diagonal matrix of eigenvalues | Eigenvalues on diagonal, zeros elsewhere |
| Defective matrix | Not diagonalizable | Has fewer than $n$ independent eigenvectors |
| Orthogonal diagonalization | $A = PDP^T$ with orthogonal $P$ | Possible for symmetric matrices |
Examples
Compute $D^{10}$ where $D = \begin{pmatrix} 2 & 0 \ 0 & 3 \end{pmatrix}$.
Solution:
For a diagonal matrix, you simply raise each diagonal entry to the power:
$$D^{10} = \begin{pmatrix} 2^{10} & 0 \ 0 & 3^{10} \end{pmatrix} = \begin{pmatrix} 1024 & 0 \ 0 & 59049 \end{pmatrix}$$
Why this works: When you multiply $D$ by itself, the diagonal entries multiply and the off-diagonal entries stay zero. After $n$ multiplications, each diagonal entry has been multiplied by itself $n$ times.
Answer: $D^{10} = \begin{pmatrix} 1024 & 0 \ 0 & 59049 \end{pmatrix}$
If $A = PDP^{-1}$, express $A^3$ in terms of $P$ and $D$.
Solution:
We compute $A^3$ by expanding:
$$A^3 = A \cdot A \cdot A = (PDP^{-1})(PDP^{-1})(PDP^{-1})$$
Watch what happens in the middle:
$$A^3 = PD\underbrace{(P^{-1}P)}{= I}D\underbrace{(P^{-1}P)}{= I}DP^{-1}$$
$$A^3 = PD \cdot I \cdot D \cdot I \cdot DP^{-1} = PD^3P^{-1}$$
General pattern: For any positive integer $n$:
$$A^n = PD^nP^{-1}$$
The key insight is that $P^{-1}P = I$ cancels in the middle, leaving all the $D$s next to each other.
Answer: $A^3 = PD^3P^{-1}$
Diagonalize $A = \begin{pmatrix} 5 & 4 \ 1 & 2 \end{pmatrix}$.
Solution:
Step 1: Find the eigenvalues.
The characteristic equation is $\det(A - \lambda I) = 0$:
$$\det\begin{pmatrix} 5 - \lambda & 4 \ 1 & 2 - \lambda \end{pmatrix} = (5 - \lambda)(2 - \lambda) - 4 \cdot 1$$
$$= 10 - 5\lambda - 2\lambda + \lambda^2 - 4 = \lambda^2 - 7\lambda + 6$$
Factoring: $\lambda^2 - 7\lambda + 6 = (\lambda - 6)(\lambda - 1) = 0$
The eigenvalues are $\lambda_1 = 6$ and $\lambda_2 = 1$.
Step 2: Find the eigenvectors.
For $\lambda_1 = 6$:
Solve $(A - 6I)\vec{v} = \vec{0}$:
$$\begin{pmatrix} 5 - 6 & 4 \ 1 & 2 - 6 \end{pmatrix}\vec{v} = \begin{pmatrix} -1 & 4 \ 1 & -4 \end{pmatrix}\vec{v} = \vec{0}$$
From the first row: $-x + 4y = 0$, so $x = 4y$.
Let $y = 1$, then $x = 4$.
Eigenvector: $\vec{v}_1 = \begin{pmatrix} 4 \ 1 \end{pmatrix}$
For $\lambda_2 = 1$:
Solve $(A - I)\vec{v} = \vec{0}$:
$$\begin{pmatrix} 5 - 1 & 4 \ 1 & 2 - 1 \end{pmatrix}\vec{v} = \begin{pmatrix} 4 & 4 \ 1 & 1 \end{pmatrix}\vec{v} = \vec{0}$$
From the first row: $4x + 4y = 0$, so $x = -y$.
Let $y = 1$, then $x = -1$.
Eigenvector: $\vec{v}_2 = \begin{pmatrix} -1 \ 1 \end{pmatrix}$
Step 3: Form $P$ and $D$.
$$P = \begin{pmatrix} 4 & -1 \ 1 & 1 \end{pmatrix}, \quad D = \begin{pmatrix} 6 & 0 \ 0 & 1 \end{pmatrix}$$
Step 4: Verify.
We can check that $AP = PD$:
$$AP = \begin{pmatrix} 5 & 4 \ 1 & 2 \end{pmatrix}\begin{pmatrix} 4 & -1 \ 1 & 1 \end{pmatrix} = \begin{pmatrix} 24 & -1 \ 6 & 1 \end{pmatrix}$$
$$PD = \begin{pmatrix} 4 & -1 \ 1 & 1 \end{pmatrix}\begin{pmatrix} 6 & 0 \ 0 & 1 \end{pmatrix} = \begin{pmatrix} 24 & -1 \ 6 & 1 \end{pmatrix}$$
Yes, $AP = PD$, confirming our diagonalization is correct.
Answer: $A = PDP^{-1}$ where $P = \begin{pmatrix} 4 & -1 \ 1 & 1 \end{pmatrix}$ and $D = \begin{pmatrix} 6 & 0 \ 0 & 1 \end{pmatrix}$
Use the diagonalization from Example 3 to compute $A^{10}$ for $A = \begin{pmatrix} 5 & 4 \ 1 & 2 \end{pmatrix}$.
Solution:
From Example 3, we have $A = PDP^{-1}$ where:
$$P = \begin{pmatrix} 4 & -1 \ 1 & 1 \end{pmatrix}, \quad D = \begin{pmatrix} 6 & 0 \ 0 & 1 \end{pmatrix}$$
Step 1: Compute $D^{10}$.
$$D^{10} = \begin{pmatrix} 6^{10} & 0 \ 0 & 1^{10} \end{pmatrix} = \begin{pmatrix} 60466176 & 0 \ 0 & 1 \end{pmatrix}$$
Step 2: Compute $P^{-1}$.
For a $2 \times 2$ matrix $\begin{pmatrix} a & b \ c & d \end{pmatrix}$, the inverse is $\frac{1}{ad - bc}\begin{pmatrix} d & -b \ -c & a \end{pmatrix}$.
$$\det(P) = (4)(1) - (-1)(1) = 4 + 1 = 5$$
$$P^{-1} = \frac{1}{5}\begin{pmatrix} 1 & 1 \ -1 & 4 \end{pmatrix} = \begin{pmatrix} 1/5 & 1/5 \ -1/5 & 4/5 \end{pmatrix}$$
Step 3: Compute $A^{10} = PD^{10}P^{-1}$.
First, compute $PD^{10}$:
$$PD^{10} = \begin{pmatrix} 4 & -1 \ 1 & 1 \end{pmatrix}\begin{pmatrix} 60466176 & 0 \ 0 & 1 \end{pmatrix} = \begin{pmatrix} 4 \cdot 60466176 & -1 \ 60466176 & 1 \end{pmatrix} = \begin{pmatrix} 241864704 & -1 \ 60466176 & 1 \end{pmatrix}$$
Then, compute $(PD^{10})P^{-1}$:
$$A^{10} = \begin{pmatrix} 241864704 & -1 \ 60466176 & 1 \end{pmatrix}\begin{pmatrix} 1/5 & 1/5 \ -1/5 & 4/5 \end{pmatrix}$$
For entry $(1,1)$: $241864704 \cdot \frac{1}{5} + (-1) \cdot (-\frac{1}{5}) = \frac{241864704 + 1}{5} = \frac{241864705}{5} = 48372941$
For entry $(1,2)$: $241864704 \cdot \frac{1}{5} + (-1) \cdot \frac{4}{5} = \frac{241864704 - 4}{5} = \frac{241864700}{5} = 48372940$
For entry $(2,1)$: $60466176 \cdot \frac{1}{5} + 1 \cdot (-\frac{1}{5}) = \frac{60466176 - 1}{5} = \frac{60466175}{5} = 12093235$
For entry $(2,2)$: $60466176 \cdot \frac{1}{5} + 1 \cdot \frac{4}{5} = \frac{60466176 + 4}{5} = \frac{60466180}{5} = 12093236$
Answer:
$$A^{10} = \begin{pmatrix} 48372941 & 48372940 \ 12093235 & 12093236 \end{pmatrix}$$
Note: Computing this directly by multiplying $A$ by itself 9 times would be tedious and error-prone. Diagonalization makes it straightforward, even for very large powers.
Determine whether $A = \begin{pmatrix} 1 & 1 \ 0 & 1 \end{pmatrix}$ is diagonalizable.
Solution:
Step 1: Find the eigenvalues.
$$\det(A - \lambda I) = \det\begin{pmatrix} 1 - \lambda & 1 \ 0 & 1 - \lambda \end{pmatrix} = (1 - \lambda)(1 - \lambda) - 0 = (1 - \lambda)^2$$
Setting $(1 - \lambda)^2 = 0$ gives $\lambda = 1$ (with algebraic multiplicity 2).
Step 2: Find the eigenvectors for $\lambda = 1$.
Solve $(A - I)\vec{v} = \vec{0}$:
$$\begin{pmatrix} 1 - 1 & 1 \ 0 & 1 - 1 \end{pmatrix}\vec{v} = \begin{pmatrix} 0 & 1 \ 0 & 0 \end{pmatrix}\vec{v} = \vec{0}$$
This gives the system:
- $0 \cdot x + 1 \cdot y = 0$, which means $y = 0$
- $0 \cdot x + 0 \cdot y = 0$ (always satisfied)
The general solution is $\vec{v} = \begin{pmatrix} x \ 0 \end{pmatrix} = x\begin{pmatrix} 1 \ 0 \end{pmatrix}$.
The eigenspace is one-dimensional, spanned by $\begin{pmatrix} 1 \ 0 \end{pmatrix}$.
Step 3: Check diagonalizability.
For a $2 \times 2$ matrix to be diagonalizable, we need 2 linearly independent eigenvectors. Here, $\lambda = 1$ has:
- Algebraic multiplicity: 2
- Geometric multiplicity: 1 (only one independent eigenvector)
Since $1 < 2$, there are not enough eigenvectors.
Conclusion: The matrix $A = \begin{pmatrix} 1 & 1 \ 0 & 1 \end{pmatrix}$ is not diagonalizable.
This matrix is called defective. It represents a shear transformation: it moves points horizontally by an amount proportional to their height. There is no basis in which this transformation acts purely by scaling.
Geometric insight: The only eigenvector direction is along the $x$-axis (horizontal). Every other direction gets “sheared” and does not simply scale. Since you cannot find two independent directions that just get scaled, you cannot diagonalize.
Answer: Not diagonalizable (defective matrix).
Key Properties and Rules
Summary of Diagonalization Properties
| Property | Statement |
|---|---|
| Diagonalizability criterion | $A$ is diagonalizable if and only if it has $n$ linearly independent eigenvectors |
| Distinct eigenvalues | If all $n$ eigenvalues are distinct, then $A$ is diagonalizable |
| Powers formula | If $A = PDP^{-1}$, then $A^n = PD^nP^{-1}$ |
| Similar matrices | $A$ and $D$ are similar: they have the same eigenvalues, determinant, and trace |
| Symmetric matrices | Real symmetric matrices are always diagonalizable |
| Orthogonal diagonalization | For symmetric $A$: $A = PDP^T$ where $P$ is orthogonal |
The Diagonalization Algorithm
- Compute eigenvalues: Solve $\det(A - \lambda I) = 0$
- Compute eigenvectors: For each $\lambda$, solve $(A - \lambda I)\vec{v} = \vec{0}$
- Check: Count total independent eigenvectors. Need exactly $n$ for an $n \times n$ matrix
- Form matrices: $P$ = [eigenvectors as columns], $D$ = diagonal matrix of eigenvalues
- Result: $A = PDP^{-1}$
When Diagonalization Fails
A matrix is NOT diagonalizable when:
- An eigenvalue has geometric multiplicity strictly less than its algebraic multiplicity
- Equivalently: the total count of linearly independent eigenvectors is less than $n$
Real-World Applications
Long-Term Behavior of Markov Chains
Markov chains model systems that transition between states with fixed probabilities, like weather patterns (sunny vs. rainy), customer behavior (loyal vs. churned), or web browsing (which page you visit next). The transition probabilities form a matrix, and the long-term distribution of states is determined by the eigenvalues.
If you want to know where the system will be after many steps, you need to compute a high power of the transition matrix. Diagonalization makes this tractable: instead of multiplying the matrix by itself hundreds of times, you raise the eigenvalues to that power. The eigenvalue equal to 1 (which always exists for valid Markov chains) determines the steady-state distribution, while eigenvalues less than 1 in magnitude represent transient effects that die out over time.
Solving Systems of Differential Equations
Many physical systems are described by systems of linear differential equations of the form $\vec{x}’(t) = A\vec{x}(t)$. When $A$ is diagonalizable, you can decouple these equations by changing to the eigenvector basis.
In the new coordinates, the system becomes a set of independent equations, each of the form $y’(t) = \lambda y(t)$. These have the simple solution $y(t) = y(0)e^{\lambda t}$. Transforming back gives the complete solution to the original system. This technique is fundamental in physics and engineering for analyzing oscillating systems, electrical circuits, and mechanical vibrations.
Population Dynamics (Leslie Matrices)
Ecologists use Leslie matrices to model how populations with different age groups evolve over time. The matrix entries represent birth rates and survival probabilities for each age class. To predict the population structure many generations in the future, you need to compute high powers of the Leslie matrix.
Diagonalization reveals the long-term growth rate (the largest eigenvalue) and the stable age distribution (the corresponding eigenvector). This helps ecologists understand whether a population will grow, shrink, or stabilize, and what the eventual age structure will look like.
Efficient Computation of Matrix Powers
Beyond the specific applications above, diagonalization is a general-purpose tool for computing matrix powers efficiently. Any problem that involves repeated application of a linear transformation, from image processing to economics to quantum mechanics, benefits from diagonalization.
The computational savings are dramatic. For an $n \times n$ matrix, computing $A^k$ directly requires about $k \cdot n^3$ operations. Using diagonalization, you need about $n^3$ operations to find $P$ and $P^{-1}$, then just $n$ operations to compute $D^k$ (raising each diagonal entry to the $k$-th power), and finally about $n^3$ operations to multiply $PD^kP^{-1}$. This is essentially independent of $k$, making it vastly more efficient for large powers.
Self-Test Problems
Problem 1: Compute $D^5$ where $D = \begin{pmatrix} 3 & 0 \ 0 & -2 \end{pmatrix}$.
Show Answer
For a diagonal matrix, raise each diagonal entry to the power:
$$D^5 = \begin{pmatrix} 3^5 & 0 \ 0 & (-2)^5 \end{pmatrix} = \begin{pmatrix} 243 & 0 \ 0 & -32 \end{pmatrix}$$
Answer: $D^5 = \begin{pmatrix} 243 & 0 \ 0 & -32 \end{pmatrix}$
Problem 2: If $A = PDP^{-1}$, express $A^{-1}$ in terms of $P$ and $D$ (assuming $A$ is invertible).
Show Answer
Starting from $A = PDP^{-1}$, take the inverse of both sides:
$$A^{-1} = (PDP^{-1})^{-1}$$
Using the rule $(ABC)^{-1} = C^{-1}B^{-1}A^{-1}$:
$$A^{-1} = (P^{-1})^{-1}D^{-1}P^{-1} = PD^{-1}P^{-1}$$
Note that $D^{-1}$ is easy to compute: just take the reciprocal of each diagonal entry.
$$D^{-1} = \begin{pmatrix} 1/\lambda_1 & 0 & \cdots \ 0 & 1/\lambda_2 & \cdots \ \vdots & \vdots & \ddots \end{pmatrix}$$
Answer: $A^{-1} = PD^{-1}P^{-1}$
Problem 3: Diagonalize $A = \begin{pmatrix} 4 & 1 \ 0 & 3 \end{pmatrix}$.
Show Answer
Step 1: Find eigenvalues.
Since $A$ is upper triangular, the eigenvalues are the diagonal entries: $\lambda_1 = 4$ and $\lambda_2 = 3$.
Step 2: Find eigenvectors.
For $\lambda_1 = 4$:
$(A - 4I)\vec{v} = \begin{pmatrix} 0 & 1 \ 0 & -1 \end{pmatrix}\vec{v} = \vec{0}$
This gives $y = 0$, so $\vec{v}_1 = \begin{pmatrix} 1 \ 0 \end{pmatrix}$.
For $\lambda_2 = 3$:
$(A - 3I)\vec{v} = \begin{pmatrix} 1 & 1 \ 0 & 0 \end{pmatrix}\vec{v} = \vec{0}$
This gives $x + y = 0$, so $x = -y$. Let $y = 1$: $\vec{v}_2 = \begin{pmatrix} -1 \ 1 \end{pmatrix}$.
Step 3: Form $P$ and $D$.
$$P = \begin{pmatrix} 1 & -1 \ 0 & 1 \end{pmatrix}, \quad D = \begin{pmatrix} 4 & 0 \ 0 & 3 \end{pmatrix}$$
Answer: $A = PDP^{-1}$ where $P = \begin{pmatrix} 1 & -1 \ 0 & 1 \end{pmatrix}$ and $D = \begin{pmatrix} 4 & 0 \ 0 & 3 \end{pmatrix}$
Problem 4: Is the matrix $A = \begin{pmatrix} 2 & 0 \ 0 & 2 \end{pmatrix}$ diagonalizable? If so, diagonalize it.
Show Answer
The matrix is already diagonal. Its only eigenvalue is $\lambda = 2$ (with algebraic multiplicity 2).
For eigenvalue $\lambda = 2$:
$(A - 2I)\vec{v} = \begin{pmatrix} 0 & 0 \ 0 & 0 \end{pmatrix}\vec{v} = \vec{0}$
Every non-zero vector is an eigenvector. The eigenspace is all of $\mathbb{R}^2$, which is 2-dimensional.
Since the geometric multiplicity (2) equals the algebraic multiplicity (2), the matrix is diagonalizable.
In fact, $A$ is already diagonal, so $A = IAI^{-1}$ where $P = I = \begin{pmatrix} 1 & 0 \ 0 & 1 \end{pmatrix}$ and $D = A = \begin{pmatrix} 2 & 0 \ 0 & 2 \end{pmatrix}$.
Answer: Yes, $A$ is diagonalizable. In fact, $D = A$ and $P$ can be any invertible matrix (commonly the identity).
Problem 5: Determine if $A = \begin{pmatrix} 0 & 1 \ 0 & 0 \end{pmatrix}$ is diagonalizable.
Show Answer
Step 1: Find eigenvalues.
$$\det(A - \lambda I) = \det\begin{pmatrix} -\lambda & 1 \ 0 & -\lambda \end{pmatrix} = \lambda^2$$
The only eigenvalue is $\lambda = 0$ with algebraic multiplicity 2.
Step 2: Find eigenvectors for $\lambda = 0$.
$(A - 0I)\vec{v} = A\vec{v} = \begin{pmatrix} 0 & 1 \ 0 & 0 \end{pmatrix}\vec{v} = \vec{0}$
This gives $y = 0$, so eigenvectors are of the form $\begin{pmatrix} x \ 0 \end{pmatrix} = x\begin{pmatrix} 1 \ 0 \end{pmatrix}$.
The eigenspace is 1-dimensional.
Step 3: Check diagonalizability.
Algebraic multiplicity of $\lambda = 0$: 2
Geometric multiplicity of $\lambda = 0$: 1
Since $1 < 2$, there are not enough eigenvectors.
Answer: No, $A$ is not diagonalizable.
Problem 6: A matrix $A$ has eigenvectors $\vec{v}_1 = \begin{pmatrix} 1 \ 0 \end{pmatrix}$ and $\vec{v}_2 = \begin{pmatrix} 1 \ 1 \end{pmatrix}$ with eigenvalues $\lambda_1 = 2$ and $\lambda_2 = 5$. Write $A$ in the form $PDP^{-1}$ and then compute $A$ explicitly.
Show Answer
Step 1: Form $P$ and $D$.
$$P = \begin{pmatrix} 1 & 1 \ 0 & 1 \end{pmatrix}, \quad D = \begin{pmatrix} 2 & 0 \ 0 & 5 \end{pmatrix}$$
Step 2: Compute $P^{-1}$.
$$\det(P) = (1)(1) - (1)(0) = 1$$
$$P^{-1} = \begin{pmatrix} 1 & -1 \ 0 & 1 \end{pmatrix}$$
Step 3: Compute $A = PDP^{-1}$.
First, $PD = \begin{pmatrix} 1 & 1 \ 0 & 1 \end{pmatrix}\begin{pmatrix} 2 & 0 \ 0 & 5 \end{pmatrix} = \begin{pmatrix} 2 & 5 \ 0 & 5 \end{pmatrix}$
Then, $A = (PD)P^{-1} = \begin{pmatrix} 2 & 5 \ 0 & 5 \end{pmatrix}\begin{pmatrix} 1 & -1 \ 0 & 1 \end{pmatrix} = \begin{pmatrix} 2 & 3 \ 0 & 5 \end{pmatrix}$
Verification: Check that $A\vec{v}_1 = 2\vec{v}_1$ and $A\vec{v}_2 = 5\vec{v}_2$.
$A\vec{v}_1 = \begin{pmatrix} 2 & 3 \ 0 & 5 \end{pmatrix}\begin{pmatrix} 1 \ 0 \end{pmatrix} = \begin{pmatrix} 2 \ 0 \end{pmatrix} = 2\begin{pmatrix} 1 \ 0 \end{pmatrix}$ (correct)
$A\vec{v}_2 = \begin{pmatrix} 2 & 3 \ 0 & 5 \end{pmatrix}\begin{pmatrix} 1 \ 1 \end{pmatrix} = \begin{pmatrix} 5 \ 5 \end{pmatrix} = 5\begin{pmatrix} 1 \ 1 \end{pmatrix}$ (correct)
Answer: $A = \begin{pmatrix} 2 & 3 \ 0 & 5 \end{pmatrix}$
Problem 7: Explain why the $3 \times 3$ identity matrix $I$ is diagonalizable. What are its eigenvalues and eigenvectors?
Show Answer
The identity matrix $I = \begin{pmatrix} 1 & 0 & 0 \ 0 & 1 & 0 \ 0 & 0 & 1 \end{pmatrix}$ is already diagonal, so it is trivially diagonalizable.
Eigenvalues: All eigenvalues are $\lambda = 1$ (with algebraic multiplicity 3).
Eigenvectors: Since $I\vec{v} = \vec{v} = 1 \cdot \vec{v}$ for every vector $\vec{v}$, every non-zero vector is an eigenvector.
Why is it diagonalizable? The eigenvalue $\lambda = 1$ has:
- Algebraic multiplicity: 3
- Geometric multiplicity: 3 (the eigenspace is all of $\mathbb{R}^3$)
Since these are equal, the matrix is diagonalizable. We can choose $P$ to be any invertible matrix (commonly the identity itself), and $D = I$.
Answer: Eigenvalue $\lambda = 1$ with multiplicity 3. Every non-zero vector is an eigenvector. The identity is already diagonal.
Summary
-
A matrix $A$ is diagonalizable if it can be written as $A = PDP^{-1}$ where $D$ is a diagonal matrix of eigenvalues and $P$ is a matrix whose columns are the corresponding eigenvectors.
-
When is a matrix diagonalizable? An $n \times n$ matrix is diagonalizable if and only if it has $n$ linearly independent eigenvectors. This is guaranteed when all eigenvalues are distinct. When eigenvalues repeat, you need the geometric multiplicity to equal the algebraic multiplicity.
-
The diagonalization process:
- Find all eigenvalues by solving $\det(A - \lambda I) = 0$
- Find eigenvectors for each eigenvalue by solving $(A - \lambda I)\vec{v} = \vec{0}$
- Form $P$ (eigenvectors as columns) and $D$ (eigenvalues on diagonal, in matching order)
-
Powers of diagonal matrices are easy: $D^n$ is found by raising each diagonal entry to the $n$-th power.
-
Computing $A^n$: If $A = PDP^{-1}$, then $A^n = PD^nP^{-1}$. This is vastly more efficient than multiplying $A$ by itself $n$ times.
-
Symmetric matrices are always diagonalizable. Their eigenvalues are real, and eigenvectors for different eigenvalues are orthogonal. They can be orthogonally diagonalized: $A = PDP^T$ where $P$ is orthogonal.
-
Applications include computing long-term behavior of Markov chains, solving systems of differential equations, modeling population dynamics with Leslie matrices, and any situation requiring efficient computation of matrix powers.
-
A matrix that is not diagonalizable is called defective. This happens when an eigenvalue does not have enough linearly independent eigenvectors.