Section 4.2 Finding eigenvalues and eigenvectors
¶The last section introduced eigenvalues and eigenvectors, presented the underlying geometric intuition behind their definition, and demonstrated their use in understanding the longterm behavior of certain systems. We will now develop a more algebraic understanding of eigenvalues and eigenvectors. In particular, we will find an algebraic method for determining the eigenvalues and eigenvectors of a square matrix.
Preview Activity 4.2.1
Let's begin by reviewing some important ideas that we have seen previously.
Suppose that \(A\) is a square matrix and that the nonzero vector \(\xvec\) is a solution to the homogeneous equation \(A\xvec = \zerovec\text{.}\) What can we conclude about the invertibility of \(A\text{?}\)
How does the determinant \(\det A\) tell us if there is a nonzero solution to the homogeneous equation \(A\xvec = \zerovec\text{?}\)

Suppose that
\begin{equation*}
A = \left[\begin{array}{rrr}
3 \amp 1 \amp 1 \\
0 \amp 2 \amp 4 \\
1 \amp 1 \amp 3 \\
\end{array}\right]\text{.}
\end{equation*}
Find the determinant \(\det A\text{.}\) What does this tell us about the solution space to the homogeneous equation \(A\xvec
= \zerovec\text{?}\)
FInd a basis for \(\nul(A)\text{.}\)
What is the relationship between the rank of a matrix and the dimension of its null space?
Subsection 4.2.1 The characteristic polynomial
We will first see that the eigenvalues of a square matrix appear as the roots of a particular polynomial. To begin, notice that we originally defined an eigenvector as a nonzero vector \(\vvec\) that satisfied the equation \(A\vvec =
\lambda\vvec\text{.}\) We will rewrite this as
\begin{equation*}
\begin{aligned}
A\vvec \amp {}={} \lambda\vvec \\
A\vvec  \lambda\vvec \amp {}={} \zerovec \\
A\vvec  \lambda I\vvec \amp {}={} \zerovec \\
(A\lambda I)\vvec \amp {}={} \zerovec\text{.} \\
\end{aligned}
\end{equation*}
In other words, an eigenvector \(\vvec\) is a solution of the homogeneous equation \((A\lambda I)\vvec=\zerovec\text{.}\) This puts us in familiar territory, which we will explore in the next activity.
Activity 4.2.2
The eigenvalues of a square matrix are defined by the condition that there be a nonzero solution to the homogeneous equation \((A\lambda I)\vvec=\zerovec\text{.}\)
If there is a nonzero solution to the homogeneous equation \((A\lambda I)\vvec = \zerovec\text{,}\) what can we conclude about the invertibility of the matrix \(A\lambda
I\text{?}\)
If there is a nonzero solution to the homogeneous equation \((A\lambda I)\vvec = \zerovec\text{,}\) what can we conclude about the determinant \(\det(A\lambda I)\text{?}\)

Let's consider the matrix
\begin{equation*}
A = \left[\begin{array}{rr}
1 \amp 2 \\
2 \amp 1 \\
\end{array}\right]
\end{equation*}
from which we construct
\begin{equation*}
A\lambda I =
\left[\begin{array}{rr}
1 \amp 2 \\
2 \amp 1 \\
\end{array}\right]
 \lambda
\left[\begin{array}{rr}
1 \amp 0 \\
0 \amp 1 \\
\end{array}\right]
=
\left[\begin{array}{rr}
1\lambda \amp 2 \\
2 \amp 1\lambda \\
\end{array}\right]\text{.}
\end{equation*}
Find the determinant \(\det(A\lambda I)\text{.}\) What kind of equation do you obtain when we set this determinant to zero to obtain \(\det(A\lambda I) = 0\text{?}\)
Use the determinant you found in the previous part to find the eigenvalues \(\lambda\) by solving \(\det(A\lambda I) = 0\text{.}\) We considered this matrix in the previous section so we should find the same eigenvalues for \(A\) that we found by reasoning geometrically there.
Consider the matrix \(A = \left[\begin{array}{rr}
2 \amp 1 \\
0 \amp 2 \\
\end{array}\right]\) and find its eigenvalues by solving the equation \(\det(A\lambda I) = 0\text{.}\)
Consider the matrix \(A = \left[\begin{array}{rr}
0 \amp 1 \\
1 \amp 0 \\
\end{array}\right]\) and find its eigenvalues by solving the equation \(\det(A\lambda I) = 0\text{.}\)
Find the eigenvalues of the triangular matrix \(\left[\begin{array}{rrr}
3 \amp 1 \amp 4 \\
0 \amp 2 \amp 3 \\
0 \amp 0 \amp 1 \\
\end{array}\right]
\text{.}\) What is generally true about the eigenvalues of a triangular matrix?
This activity demonstrates a technique that enables us to find the eigenvalues of a square matrix \(A\text{.}\) Since an eigenvalue \(\lambda\) is a scalar for which the equation \((A\lambda
I)\vvec = \zerovec\) has a nonzero solution, it must be the case that \(A\lambda I\) is not invertible. Therefore, its determinant is zero. This gives us the equation
\begin{equation*}
\det(A\lambda I) = 0
\end{equation*}
whose solutions are the eigenvalues of \(A\text{.}\) This equation is called the characteristic equation of \(A\text{.}\)
If we write the characteristic equation for the matrix \(A = \left[\begin{array}{rr}
1 \amp 2 \\
2 \amp 1 \\
\end{array}\right]
\text{,}\) we see that
\begin{equation*}
\begin{aligned}
\det(A\lambda I) \amp {}={} 0 \\
\\
\det \left[\begin{array}{rr}
1  \lambda \amp 2 \\
2 \amp 1  \lambda \\
\end{array}\right] \amp {}={} 0 \\
\\
(1\lambda)^2  4 \amp {}={} 0 \\
3  2\lambda + \lambda^2 {}={} 0 \\
(3\lambda)(1\lambda) {}={} 0\text{.} \\
\end{aligned}
\end{equation*}
This shows us that the eigenvalues are \(\lambda = 3\) and \(\lambda=1\text{,}\) the same eigenvalues we found by reasoning geometrically in the previous section.
In general, the expression \(\det(A\lambda I)\) is a polynomial in \(\lambda\text{,}\) which is called the characteristic polynomial of \(A\text{.}\) If \(A\) is an \(n\times n\) matrix, the degree of the characteristic polynomial is \(n\text{.}\) For instance, if \(A\) is a \(2\times2\) matrix, then \(\det(A\lambda I)\) is a quadratic polynomial; if \(A\) is a \(3\times3\) matrix, then \(\det(A\lambda I)\) is a cubic polynomial.
The other examples that appear in this activity demonstrate some issues we will need to deal with later. For instance, the matrix \(A = \left[\begin{array}{rr}
2 \amp 1 \\
0 \amp 2 \\
\end{array}\right]\) leads to the characteristic equation, \((2\lambda)^2 = 0\text{.}\) In this case, there is only one eigenvalue \(\lambda = 2\) that appears as a repeated root. For now, we simply note that our work in the previous section showed that it was not possible to form a basis of \(\real^2\) consisting of eigenvectors of \(A\text{.}\)
Finally, when \(A = \left[\begin{array}{rr}
0 \amp 1 \\
1 \amp 0 \\
\end{array}\right]
\text{,}\) we find the characteristic equation \(\lambda^2 + 1 =
0\text{.}\) While this equation has no real solutions, it does have complex solutions \(\lambda = \pm i\text{,}\) and it will be useful for us to work with these complex eigenvalues in the future. In the meantime, remember that this matrix defines a \(90^\circ\) rotation so we do not expect any solutions to the equation \(A\vvec = \lambda \vvec\) for real eigenvalues \(\lambda\) since a vector \(\vvec\) and \(A\vvec\) can never lie on the same line.
Finally, the eigenvalues of a triangular matrix are easily determined because the determinant of a triangular matrix is the product of the entries on the diagonal. Therefore, the characteristic equation is
\begin{equation*}
\begin{aligned}
\det\left(
\left[\begin{array}{rrr}
3 \amp 1 \amp 4 \\
0 \amp 2 \amp 3 \\
0 \amp 0 \amp 1 \\
\end{array}\right]
\lambda I\right) \amp {}={}
\det
\left[\begin{array}{rrr}
3\lambda \amp 1 \amp 4 \\
0 \amp 2\lambda \amp 3 \\
0 \amp 0 \amp 1\lambda \\
\end{array}\right] \\
\\
\amp {}={}(3\lambda)(2\lambda)(1\lambda) = 0\text{,}
\end{aligned}
\end{equation*}
showing that the eigenvalues are the diagonal entries \(\lambda =
3,2,1\text{.}\)
We have how seen now the characteristic equation can be used to determine the eigenvalues of a matrix. Remember, however, that finding the determinant of a matrix using a cofactor expansion is not computationally feasible when the size of the matrix is relatively large. Finding the eigenvalues of a matrix by factoring its characteristic polynomial is therefore a technique limited to relatively small matrices; we will introduce a new technique for finding eigenvalues of larger matrices in the next chapter.
Subsection 4.2.2 Finding eigenvectors
Now that we can find the eigenvalues of a square matrix \(A\) by solving the characteristic equation \(\det(A\lambda
I) = 0\text{,}\) we will turn to the question of finding the eigenvectors associated to an eigenvalue \(\lambda\text{.}\) Once again, the key is to note that an eigenvector is a nonzero solution to the homogeneous equation \((A\lambda I)\vvec =
\zerovec\text{.}\) In other words, the eigenvectors associated to an eigenvalue \(\lambda\) form the null space \(\nul(A\lambda
I)\text{.}\)
This shows that the eigenvectors associated to an eigenvalue form a subspace of \(\real^n\text{.}\) We will use \(E_\lambda\) to denote the subspace of eigenvectors of a matrix \(A\) associated to the eigenvalue \(\lambda\) and note that
\begin{equation*}
E_\lambda = \nul(A\lambda I)\text{.}
\end{equation*}
We say that \(E_\lambda\) is the eigenspace of \(A\) associated to the eigenvalue \(\lambda\text{.}\)
Activity 4.2.3
In this activity, we will find the eigenvectors of a matrix as the null space of the matrix \(A\lambda I\text{.}\)
Let's begin with the matrix \(A = \left[\begin{array}{rr}
1 \amp 2 \\
2 \amp 1 \\
\end{array}\right]
\text{.}\) We have seen that \(\lambda = 3\) is an eigenvalue. Form the matrix \(A3I\) and find a basis for the eigenspace \(E_3 = \nul(A3I)\text{.}\) What is the dimension of this eigenspace? For each of the basis vectors \(\vvec\text{,}\) verify that \(A\vvec = 3\vvec\text{.}\)
We also saw that \(\lambda = 1\) is an eigenvalue. Form the matrix \(A(1)I\) and find a basis for the eigenspace \(E_{1}\text{.}\) What is the dimension of this eigenspace? For each of the basis vectors \(\vvec\text{,}\) verify that \(A\vvec = \vvec\text{.}\)
Is it possible to form a basis of \(\real^2\) consisting of eigenvectors of \(A\text{?}\)
Now consider the matrix \(A = \left[\begin{array}{rr}
3 \amp 0 \\
0 \amp 3 \\
\end{array}\right]
\text{.}\) Write the characteristic equation for \(A\) and use it to find the eigenvalues of \(A\text{.}\) For each eigenvalue, find a basis for its eigenspace \(E_\lambda\text{.}\) Is it possible to form a basis of \(\real^2\) consisting of eigenvectors of \(A\text{?}\)
Next, consider the matrix \(A = \left[\begin{array}{rr}
2 \amp 1 \\
0 \amp 2 \\
\end{array}\right]
\text{.}\) Write the characteristic equation for \(A\) and use it to find the eigenvalues of \(A\text{.}\) For each eigenvalue, find a basis for its eigenspace \(E_\lambda\text{.}\) Is it possible to form a basis of \(\real^2\) consisting of eigenvectors of \(A\text{?}\)
Finally, find the eigenvalues and eigenvectors of the diagonal matrix \(A = \left[\begin{array}{rr}
4 \amp 0 \\
0 \amp 1 \\
\end{array}\right]
\text{.}\) Explain your result by considering the geometric effect of the matrix transformation defined by \(A\text{.}\)
Once we find the eigenvalues of a matrix \(A\text{,}\) describing the eigenspace \(E_\lambda\) amounts to the familiar task of describing the null space \(\nul(A\lambda I)\text{.}\) For instance, we know that \(\lambda = 3\) is an eigenvalue of \(A = \left[\begin{array}{rr}
1 \amp 2 \\
2 \amp 1 \\
\end{array}\right]
\text{.}\) Then, \(E_3 = \nul(A3I)\text{,}\) and we have
\begin{equation*}
A  3I =
\left[\begin{array}{rr}
2 \amp 2 \\
2 \amp 2 \\
\end{array}\right]
\sim
\left[\begin{array}{rr}
1 \amp 1 \\
0 \amp 0 \\
\end{array}\right]\text{.}
\end{equation*}
From the reduced row echelon form, we see that the eigenvectors \(\vvec = \twovec{v_1}{v_2}\) are determined by the single equation \(v_1v_2 = 0\) or \(v_1 = v_2\text{.}\) Therefore the eigenvectors in \(E_3\) have the form
\begin{equation*}
\vvec=\twovec{v_1}{v_2} = \twovec{v_2}{v_2} = v_2\twovec{1}{1}\text{.}
\end{equation*}
In other words, \(E_3\) is a onedimensional subspace of \(\real^2\) with basis \(\twovec{1}{1}\text{.}\) Once again, this agrees with the eigenvectors that we found geometrically in the previous section.
The same reasoning applies to show that the eigenvectors associated to \(\lambda = 1\) have the form
\begin{equation*}
\vvec = v_2\twovec{1}{1}\text{,}
\end{equation*}
which shows that the eigenspace \(E_{1}\) is a onedimensional subspace of \(\real^2\) having basis \(\twovec{1}{1}\text{.}\)
Two more examples from the activity are important. The characteristic equation for the matrix \(A=
\left[\begin{array}{rr}
3 \amp 0 \\
0 \amp 3 \\
\end{array}\right]\) is \(\det(A\lambda I) = (3\lambda)^2 = 0\text{.}\) This shows that there is a single eigenvalue \(\lambda = 3\text{.}\) If we find the eigenspace \(E_3=\nul(A3I)\text{,}\) we have
\begin{equation*}
A 3I =
\left[\begin{array}{rr}
3 \amp 0 \\
0 \amp 3 \\
\end{array}\right]
 3I
=
\left[\begin{array}{rr}
0 \amp 0 \\
0 \amp 0 \\
\end{array}\right]\text{.}
\end{equation*}
This shows that every vector is in \(E_3\) so that \(E_3=\real^2\text{.}\) In this case, there is a basis of \(\real^2\) consisting of eigenvectors of \(A\text{.}\) This aligns with our geometric understanding: this matrix has the effect of scaling vectors by a factor of \(3\) in every direction. Therefore, every vector is an eigenvector with eigenvalue \(\lambda = 3\text{.}\)
However, if we consider the matrix \(A=
\left[\begin{array}{rr}
2 \amp 1 \\
0 \amp 2 \\
\end{array}\right]
\text{,}\) we find the characteristic equation \((2\lambda)^2 = 0\text{,}\) which shows that there is again a single eigenvalue \(\lambda =
2\text{.}\) In this case,
\begin{equation*}
A 2I =
\left[\begin{array}{rr}
2 \amp 1 \\
0 \amp 2 \\
\end{array}\right]
 2I
=
\left[\begin{array}{rr}
0 \amp 1 \\
0 \amp 0 \\
\end{array}\right]\text{,}
\end{equation*}
which shows that \(E_2\) is a onedimensional subspace of \(\real^2\) with basis \(\twovec{1}{0}\text{.}\) Since there are no other eigenvalues, it is not possible to find a basis for \(\real^2\) consisting of eigenvectors of \(A\text{.}\)
Once again, we can understand this result geometrically. The matrix transformation corresponding to the matrix \(A\) is a shear that slides vectors horizontally. This transformation therefore only scales vectors that lie on the horizontal axis.
These last two examples illustrate two types of behavior when there is a single eigenvalue. In one case, we are able to construct a basis of \(\real^2\) using eigenvectors; in the other, we are not. We will explore this behavior more in the next subsection.
A check on our work
When finding eigenvalues and their associated eigenvectors in this way, we first find eigenvalues \(\lambda\) by solving the characteristic equation. If \(\lambda\) is a solution to the characteristic equation, then \(A\lambda I\) is not invertible and, consequently, \(A\lambda I\) must contain a row without a pivot position.
This serves as a check on our work. If we row reduce \(A\lambda I\) and find the identity matrix, then we have made an error either in solving the characteristic equation or in finding \(\nul(A\lambda I)\text{.}\)
Subsection 4.2.3 The characteristic polynomial and the dimension of eigenspaces
Given a square \(n\times n\) matrix \(A\text{,}\) we saw in the previous section the value of being able to express any vector in \(\real^n\) as a linear combination of eigenvectors of \(A\text{.}\) For this reason, we asked Question 4.1.7 to determine when we can construct a basis of \(\real^n\) consisting of eigenvectors. We will explore this question more fully now.
As we saw above, the eigenvalues of \(A\) are the solutions of the characteristic equation \(\det(A\lambda I) = 0\text{.}\) Two examples of characteristic equations we have seen above are
\begin{equation*}
\begin{aligned}
(3\lambda)(2\lambda)(1\lambda) \amp {}={} 0 \\
\text{and}\qquad(2\lambda)^2 \amp {}={} 0 \\
\end{aligned}\text{.}
\end{equation*}
Generally speaking, the characteristic polynomial can always be factored into terms having the form \((\lambda_j\lambda)\) where \(\lambda_j\) is an eigenvalue of \(A\text{.}\) In doing so, we must allow ourselves to consider complex eigenvalues, which we will study in more detail in the next section. This means, however, that we can always write the characteristic equation in the form
\begin{equation*}
(\lambda_1\lambda)^{m_1}
(\lambda_2\lambda)^{m_2}
\ldots
(\lambda_p\lambda)^{m_p}
= 0\text{.}
\end{equation*}
The solutions to the characteristic equation are the eigenvalues \(\lambda_j\text{,}\) and \(m_j\text{,}\) the number of times that \(\lambda_j  \lambda\) appears as a factor in the characteristic polynomial, is called the multiplicity of the eigenvalue \(\lambda_j\text{.}\)
Example 4.2.1
We have seen that the matrix \(A = \left[\begin{array}{rr}
2 \amp 1 \\
0 \amp 2 \\
\end{array}\right]\) has the characteristic equation \((2\lambda)^2 =
0\text{.}\) This matrix \(A\) has a single eigenvalue \(\lambda = 2\text{,}\) which has multiplicity \(2\text{.}\)
Example 4.2.2
If a matrix has the characteristic equation
\begin{equation*}
(4\lambda)^2(5\lambda)(1\lambda)^7(3\lambda)^2 = 0\text{,}
\end{equation*}
then that matrix has four eigenvalues: \(\lambda=4\) having multiplicity 2; \(\lambda=5\) having multiplicity 1; \(\lambda=1\) having multiplicty 7; and \(\lambda=3\) having multiplicty 2. The degree of the characteristic polynomial is the sum of the multiplicities \(2+1+7+2 =
12\) so this matrix must be a \(12\times12\) matrix.
The multiplicities of the eigenvalues are important because they influence the dimension of the eigenspaces. We know that the dimension of an eigenspace must be at least one; the following proposition also tells us the dimension of an eigenspace can be no larger than the multiplicity of its associated eigenvalue.
Proposition 4.2.3
If \(\lambda\) is a real eigenvalue of the matrix \(A\) with multiplicity \(m\text{,}\) then
\begin{equation*}
1 \leq \dim E_\lambda \leq m\text{.}
\end{equation*}
Example 4.2.4
The diagonal matrix \(\left[\begin{array}{rr}
3 \amp 0 \\
0 \amp 3 \\
\end{array}\right]\) has the characteristic equation \((3\lambda)^2 =
0\text{.}\) There is a single eigenvalue \(\lambda = 3\) having multiplicity \(m = 2\text{,}\) and we saw earlier that \(\dim E_3 = 2 \leq m = 2\text{.}\)
Example 4.2.5
The matrix \(\left[\begin{array}{rr}
2 \amp 1 \\
0 \amp 2 \\
\end{array}\right]\) has the characteristic equation \((2\lambda)^2 =
0\text{.}\) Once again, there is a single eigenvalue \(\lambda =
2\) having multiplicity \(m = 2\text{.}\) In contrast with the previous example, we saw that \(\dim E_2 = 1 \leq m = 2\text{.}\)
Example 4.2.6
We saw earlier that the matrix \(\left[\begin{array}{rrr}
3 \amp 1 \amp 4 \\
0 \amp 2 \amp 3 \\
0 \amp 0 \amp 1 \\
\end{array}\right]\) has the characteristic equation
\begin{equation*}
(3\lambda)(2\lambda)(1\lambda)=0\text{.}
\end{equation*}
There are three eigenvalues \(\lambda=3,2,1\) each having multiplicity \(1\text{.}\) By the proposition, we are guaranteed that the dimension of each eigenspace is \(1\text{;}\) that is,
\begin{equation*}
\dim E_3 = \dim E_{2} = \dim E_1 = 1\text{.}
\end{equation*}
It turns out that this is enough to guarantee that there is a basis of \(\real^3\) consisting of eigenvectors.
Example 4.2.7
If a \(12\times12\) matrix has the characteristic equation
\begin{equation*}
(4\lambda)^2(5\lambda)(1\lambda)^7(3\lambda)^2 = 0\text{,}
\end{equation*}
we know there are four eigenvalues \(\lambda=4,5,1,3\text{.}\) Without more information, all we can say about the dimensions of the eigenspaces is
\begin{equation*}
\begin{aligned}
1 \leq \dim E_4 \amp {}\leq{} 2 \\
1 \leq \dim E_{5} \amp {}\leq{} 1 \\
1 \leq \dim E_1 \amp {}\leq{} 7 \\
1 \leq \dim E_3 \amp {}\leq{} 2\text{.} \\
\end{aligned}
\end{equation*}
We can guarantee that \(\dim E_{5} = 1\text{,}\) but we cannot be more specific about the dimensions of the other eigenspaces.
Fortunately, if we have an \(n\times n\) matrix, it most commonly happens that the characteristic equation has the form
\begin{equation*}
(\lambda_1\lambda)
(\lambda_2\lambda)
\ldots
(\lambda_n\lambda) = 0
\end{equation*}
where there are \(n\) distinct eigenvalues, each of which has multiplicity \(1\text{.}\) In this case, the dimension of each of the eigenspaces \(\dim E_{\lambda_j} = 1\text{.}\) With a little work, it can be seen that choosing a basis vector \(\vvec_j\) for each of the eigenspaces produces a basis for \(\real^n\text{.}\) We therefore have the following proposition.
Proposition 4.2.8
If \(A\) is an \(n\times n\) matrix having \(n\) distinct real eigenvalues, then there is a basis of \(\real^n\) consisting of eigenvectors of \(A\text{.}\)
This proposition provides one answer to our Question 4.1.7. The next activity explores this question further.
Activity 4.2.4
Identify the eigenvalues, and their multiplicities, of an \(n\times n\) matrix whose characteristic polynomial is \((2\lambda)^3(3\lambda)^{10}(5\lambda)\text{.}\) What can you conclude about the dimensions of the eigenspaces? What is the dimension of the matrix? Do you have enough information to guarantee that there is a basis of \(\real^n\) consisting of eigenvectors?
Find the eigenvalues of \(\left[\begin{array}{rr}
0 \amp 1 \\
4 \amp 4 \\
\end{array}\right]\) and state their multiplicities. Can you find a basis of \(\real^2\) consisting of eigenvectors of this matrix?

Consider the matrix \(A =
\left[\begin{array}{rrr}
1 \amp 0 \amp 2 \\
2 \amp 2 \amp 4 \\
0 \amp 0 \amp 2 \\
\end{array}\right]\) whose characteristic equation is
\begin{equation*}
(2\lambda)^2(1\lambda) = 0\text{.}
\end{equation*}
Identify the eigenvalues and their multiplicities.
For each eigenvalue \(\lambda\text{,}\) find a basis of the eigenspace \(E_\lambda\) and state its dimension.
Is there a basis of \(\real^3\) consisting of eigenvectors of \(A\text{?}\)

Now consider the matrix \(A =
\left[\begin{array}{rrr}
5 \amp 2 \amp 6 \\
2 \amp 2 \amp 4 \\
2 \amp 1 \amp 2 \\
\end{array}\right]\) whose characteristic equation is also
\begin{equation*}
(2\lambda)^2(1\lambda) = 0\text{.}
\end{equation*}
Identify the eigenvalues and their multiplicities.
For each eigenvalue \(\lambda\text{,}\) find a basis of the eigenspace \(E_\lambda\) and state its dimension.
Is there a basis of \(\real^3\) consisting of eigenvectors of \(A\text{?}\)

Consider the matrix \(A =
\left[\begin{array}{rrr}
5 \amp 2 \amp 6 \\
4 \amp 1 \amp 8 \\
2 \amp 1 \amp 2 \\
\end{array}\right]\) whose characteristic equation is
\begin{equation*}
(2\lambda)(1\lambda)(1\lambda) = 0\text{.}
\end{equation*}
Identify the eigenvalues and their multiplicities.
For each eigenvalue \(\lambda\text{,}\) find a basis of the eigenspace \(E_\lambda\) and state its dimension.
Is there a basis of \(\real^3\) consisting of eigenvectors of \(A\text{?}\)
Subsection 4.2.4 Computational issues in finding eigenvalues and eigenvectors
We can use Sage to find the characteristic polynomial, eigenvalues, and eigenvectors of a matrix. As we will see, however, some care is required when dealing with matrices whose entries are not rational numbers. The next activity demonstrates how Sage can be used in this way and some of the complications that arise. We will revisit this issue in subsequent sections.
Activity 4.2.5
We will use Sage to find the eigenvalues and eigenvectors of a matrix. Let's begin with the matrix \(A = \left[\begin{array}{rr}
1 \amp 2 \\
2 \amp 1 \\
\end{array}\right]
\text{.}\)

We can find the characteristic polynomial of a matrix \(A\) by writing A.charpoly('lam')
. Notice that we have to give Sage a variable in which to write the polynomial; here, we use lam
though you could just as well use x
.
The factored form of the characteristic polynomial may be more useful since it will tell us the eigenvalues and their multiplicities. The factor chacteristic polynomial is found with A.fcp('lam')
.

If we only want the eigenvalues, we can use A.eigenvalues()
.
Notice that the multiplicity of an eigenvalue is the number of times it is repeated in the list of eigenvalues.

Finally, we can find eigenvectors by A.eigenvectors_right()
. (We are looking for right eigenvalues since the vector \(\vvec\) appears to the right of \(A\) in the definition \(A\vvec=\lambda \vvec\text{.}\))
At first glance, the result of this command can be a little confusing to interpret. What we see is a list with one entry for each eigenvalue. For each eigenvalue, there is a triple consisting of (i) the eigenvalue \(\lambda\text{,}\) (ii) a basis for \(E_\lambda\text{,}\) and (iii) the multiplicity of \(\lambda\text{.}\)

When working with decimal entries, which are called floating point numbers in computer science, we must remember that computers perform only approximate arithmetic. This is a problem when we wish to find the eigenvectors of such a matrix. To illustrate, consider the matrix \(A=\left[\begin{array}{rr}
0.4 \amp 0.3 \\
0.6 \amp 0.7 \\
\end{array}\right]
\text{.}\)
Without using Sage, find the eigenvalues of this matrix.
What do you find for the reduced row echelon form of \(AI\text{?}\)

Let's now use Sage to determine the reduced row echelon form of \(AI\text{:}\)
What result does Sage report for the reduced row echelon form? Why is this result not correct?

Because the arithmetic Sage performs with floating point entries is only approximate, we are not able to find the eigenspace \(E_1\text{.}\) In this next chapter, we will learn how to address this issue. In the meantime, we can get around this problem by writing the entries in the matrix as rational numbers:
Subsection 4.2.5 Summary
In this section, we developed a technique for finding the eigenvalues and eigenvectors of an \(n\times n\) matrix \(A\text{.}\)
The expression \(\det(A\lambda I)\) is a degree \(n\) polynomial, known as the characteristic polynomial. The eigenvalues are the roots of the characteristic polynomial \(\det(A\lambda I) = 0\text{.}\)
The set of eigenvectors associated to the eigenvalue \(\lambda\) forms the eigenspace \(E_\lambda =
\nul(A\lambda I)\text{.}\)

If the factor \((\lambda_j  \lambda)\) appears \(m_j\) times in the characteristic polynomial, we say that the eigenvalue \(\lambda_j\) has multiplicity \(m_j\) and note that
\begin{equation*}
1 \leq \dim E_{\lambda_j} \leq m_j\text{.}
\end{equation*}
If each of the eigenvalues is real and has multiplicity \(1\text{,}\) then we can form a basis for \(\real^n\) consisting of eigenvectors of \(A\text{.}\)
We can use Sage to find the eigenvalues and eigenvalues of matrices. However, we need to be careful working with floating point numbers since floating point arithmetic is only an approximation.
Subsection 4.2.6 Exercises
1
For each of the following matrices, find its characteristic polynomial, its eigenvalues, and the multiplicity of each eigenvalue.
\(A=\left[\begin{array}{rr}
4 \amp 1 \\
4 \amp 0 \\
\end{array}\right]
\text{.}\)
\(A=\left[\begin{array}{rrr}
3 \amp 1 \amp 3 \\
0 \amp 4 \amp 0 \\
0 \amp 0 \amp 6
\end{array}\right]
\text{.}\)
\(A = \left[\begin{array}{rr}
2 \amp 0 \\
0 \amp 2 \\
\end{array}\right]
\text{.}\)
\(A=\left[\begin{array}{rr}
1 \amp 2 \\
2 \amp 2 \\
\end{array}\right]
\text{.}\)
2
Given an \(n\times n\) matrix \(A\text{,}\) an important question Question 4.1.7 asks whether we can find a basis of \(\real^n\) consisting of eigenvectors of \(A\text{.}\) For each of the matrices in the previous exericse, find a basis of \(\real^n\) consisting of eigenvectors or state why such a basis does not exist.
3
Determine whether the following statements are true or false and provide a justification for your response.
The eigenvalues of a matrix \(A\) are the entries on the diagonal of \(A\text{.}\)
If \(\lambda\) is an eigenvalue of multiplicity \(1\text{,}\) then \(E_\lambda\) is onedimensional.
If a matrix \(A\) is invertible, then \(\lambda=0\) cannot be an eigenvalue.
If \(A\) is a \(13\times 13\) matrix, the charasteristic polynomial has degree less than \(13\text{.}\)
The eigenspace \(E_\lambda\) of \(A\) is the same as the null space \(\nul(A\lambda I)\text{.}\)
4
Provide a justification for your response to the following questions.
Suppose that \(A\) is a \(3\times 3\) matrix having eigenvalues \(\lambda = 3,3,5\text{.}\) What are the eigenvalues of \(2A\text{?}\)
Suppose that \(D\) is a diagonal \(3\times 3\) matrix. Why can you guarantee that there is a basis of \(\real^3\) consisting of eigenvectors of \(D\text{?}\)
If \(A\) is a \(3\times 3\) matrix whose eigenvalues are \(\lambda = 1,3,5\text{,}\) can you guarantee that there is a basis of \(\real^3\) consisting of eigenvectors of \(A\text{?}\)

Suppose that the characteristic polynomial of a matrix \(A\) is
\begin{equation*}
\det(A\lambda I) = \lambda^3 + 4\lambda\text{.}
\end{equation*}
What are the eigenvalues of \(A\text{?}\) Is \(A\) invertible? Is there a basis of \(\real^n\) consisting of eigenvectors of \(A\text{?}\)

If the characteristic polynomial of \(A\) is
\begin{equation*}
\det(A\lambda I) = (4 \lambda)(2\lambda)(1\lambda)\text{,}
\end{equation*}
what is the characteristic polynomial of \(A^2\text{?}\) what is the characteristic polynomial of \(A^{1}\text{?}\)
5
For each of the following matrices, use Sage to determine its eigenvalues, their multiplicities, and a basis for each eigenspace. For which matrices is it possible to construct a basis for \(\real^3\) consisting of eigenvectors?
\(A = \left[\begin{array}{rrr}
4 \amp 12 \amp 6 \\
4 \amp 5 \amp 4 \\
11 \amp 20 \amp 13 \\
\end{array}\right]\)
\(A = \left[\begin{array}{rrr}
1 \amp 3 \amp 1 \\
4 \amp 8 \amp 5 \\
8 \amp 17 \amp 10 \\
\end{array}\right]\)
\(A = \left[\begin{array}{rrr}
3 \amp 8 \amp 4 \\
2 \amp 3 \amp 2 \\
6 \amp 12 \amp 7 \\
\end{array}\right]\)
6
There is a relationship between the determinant of a matrix and the product of its eigenvalues.
We have seen that the eigenvalues of the matrix \(A = \left[\begin{array}{rr}
1 \amp 2 \\
2 \amp 1 \\
\end{array}\right]\) are \(\lambda = 3,1\text{.}\) What is \(\det A\text{?}\) What is the product of the eigenvalues of \(A\text{?}\)
Consider the triangular matrix \(A = \left[\begin{array}{rrr}
2 \amp 0 \amp 0 \\
1 \amp 3 \amp 0 \\
3 \amp 1 \amp 2 \\
\end{array}\right]
\text{.}\) What are the eigenvalues of \(A\text{?}\) What is \(\det
A\text{?}\) What is the product of the eigenvalues of \(A\text{?}\)
Based on these examples, what do you think is the relationship between the determinant of a matrix and the product of its eigenvalues?

Suppose the characteristic polynomial is written as
\begin{equation*}
\det(A\lambda I) = (\lambda_1\lambda)(\lambda_2\lambda)
\ldots (\lambda_n\lambda)\text{.}
\end{equation*}
By substituting \(\lambda = 0\) into this equation, explain why the determinant of a matrix equals the product of its eigenvalues.
7
Consider the matrix \(A=\left[\begin{array}{rr}
0.5 \amp 0.6 \\
0.3 \amp 1.4 \\
\end{array}\right]
\text{.}\)
Find the eigenvalues of \(A\) and a basis for their associated eigenspaces.
Suppose that \(\xvec_0=\twovec{11}{6}\text{.}\) Express \(\xvec_0\) as a linear combination of eigenvectors of \(A\text{.}\)

Define the vectors
\begin{equation*}
\begin{aligned}
\xvec_1 \amp {}={} A\xvec_0 \\
\xvec_2 \amp {}={} A\xvec_1 = A^2\xvec_0 \\
\xvec_3 \amp {}={} A\xvec_2 = A^3\xvec_0 \\
\vdots \amp {}={} \vdots
\end{aligned}\text{.}
\end{equation*}
Write \(\xvec_1\text{,}\) \(\xvec_2\text{,}\) and \(\xvec_3\) as a linear combination of eigenvectors of \(A\text{.}\)
What happens to \(\xvec_k\) as \(k\) grows larger and larger?
8
Consider the matrix \(A=\left[\begin{array}{rr}
0.4 \amp 0.3 \\
0.6 \amp 0.7 \\
\end{array}\right]\)
Find the eigenvalues of \(A\) and a basis for their associated eigenspaces.
Suppose that \(\xvec_0=\twovec{0}{1}\text{.}\) Express \(\xvec_0\) as a linear combination of eigenvectors of \(A\text{.}\)

Define the vectors
\begin{equation*}
\begin{aligned}
\xvec_1 \amp {}={} A\xvec_0 \\
\xvec_2 \amp {}={} A\xvec_1 = A^2\xvec_0 \\
\xvec_3 \amp {}={} A\xvec_2 = A^3\xvec_0 \\
\vdots \amp {}={} \vdots
\end{aligned}\text{.}
\end{equation*}
Write \(\xvec_1\text{,}\) \(\xvec_2\text{,}\) and \(\xvec_3\) as a linear combination of eigenvectors of \(A\text{.}\)
What happens to \(\xvec_k\) as \(k\) grows larger and larger?