Eigenvalues (translated from German, this means proper values) are a special set of scalars
associated with every square matrix that are sometimes also known as characteristic roots,
characteristic values, or proper values. Each eigenvalue is paired with a corresponding set of
so-called eigenvectors. The determination of the eigenvalues and eigenvectors of a system is
extremely important in physics and engineering, where it arises in such common applications as
stability analysis, the physics of rotating bodies, and small oscillations of vibrating systems, to name
only a few.
If A is a square \( n \times n \) matrix with real entries and v is an \( n \times 1 \)
column vector, then the product w = Av is defined and is another \( n \times 1 \)
column vector. It does not matter whether v is real vector v ∈ ℝn or complex v ∈ ℂn.
Therefore, any square matrix with real entries (we deal only with real matrices) can be considered as a linear operator A : v ↦ w = Av, acting either in ℝn or ℂn. Of course, one can use any Euclidean space not necessarily ℝn or ℂn.
Although a transformation v ↦ Av may move vectors in a variety of directions, it often happen that we are looking for such vectors on which action of A is just multiplication by a constant. Such a linear transformation is usually referred to as the spectral representation of the operator A.
It is important in many applications to determine whether there exist nonzero column vectors v such
that the product vector \( {\bf A}\,{\bf v} \) is a constant multiple (which we denote as λ) of v.
Example 1A:
Let us consider the following matrix and two vectors:
As shown by the plot, v is an eigenvector, but vector u is not an eigenvector of matrix A given that no such constant
exists to map u ↦ A u.
Example 1B:
As another example, we can consider a stochastic matrix that describes the transitions of a Markov chain. Each of its entries is a nonnegative real number representing a probability (in every column, the sum of entries is 1). It is also called a probability matrix, transition matrix, substitution matrix, or Markov matrix.
with respect to the standard basis of ℝ². Here v = [x, y]T is an arbitrary column-vector from ℝ².
In particular, for the vector uk whose coordinates are two consecutive Fibonacci numbers [Fk, Fk-1]T, we have that
Thus, we can produce a vector whose coordinates are two consecutive Fibonacci numbers by applying L many times to the vector u1 with coordinates [F1, F0]T = [1, 0]T:
has a nontrivial solution v (meaning it is not identically zero), then the vector v is called an eigenvector, corresponding to λ,
which is called the associated eigenvalue. The set of all eigenvalues is called the spectrum of matrix A.
Theorem 1:
If A is an n×n triangular matrix (upper triangular, lower triangular, or diagonal). then the eigenvalues of A are the entries on the main diagonal of A.
Example 3:
We consider the following 5×5 lower triangular matrix
For a matrix of this size, it is appropriate to use a software. So we ask Python for help:
We know that the entries on the main diagonal are eigenvalues, but we check the Python output
The corresponding eigenvectors are
The trace of matrix A is zero as well as its determinant. So its characteristic polynomial is
We can calculate it by evaluating the determinant:
■
A number λ can be eigenvalue for a square matrix A only when the determinant of the matrix corresponding to the system \( \lambda \,{\bf v} - {\bf A}\, {\bf v} = {\bf 0} \) vanishes,
namely, \( \det \left( \lambda\, {\bf I} - {\bf A} \right) =0 ,\) where I is the identity matrix.
If A is a square matrix, the determinant \( \chi (\lambda ) = \det \left( \lambda\, {\bf I} - {\bf A} \right) . \) is called the characteristic polynomial and we denote it by χ(λ).
Samuelson's formula allows the characteristic polynomial to be computed recursively without divisions.
Theorem 2:
A scalar (real or complex) is an eigenvalue of a square matrix A if and only if it is a root of characteristic polynomial:
This determinant
is called the characteristic polynomial and we denote it by \( \chi (\lambda ) = \det \left( \lambda\, {\bf I} - {\bf A} \right) . \) Every square matrix has an eigenvalue and corresponding eigenvectors.
Therefore, eigenvalues are the nulls of the characteristic polynomial and they
are the roots of the equation
\( \chi (\lambda ) = 0. \)
The characteristic polynomial is always a polynomial of degree n, where
n is the dimension of the square matrix A. Its
coefficients can be expressed through eigenvalues:
where \( \mbox{tr} {\bf A} = a_{11} + a_{22} + \cdots + a_{nn} = \lambda_1 + \lambda_2 + \cdots + \lambda_n \) is the trace of the matrix A, that is, the sum of its diagonal elements,
which is equal to the sum of all eigenvalues (including their
multiplicities). This is true for arbitrary matrices.
The set of all eigenvalues is called
the spectrum
of the matrix A.
Theorem 3:
A square matrix A is invertible if and only if λ = 0 is not an eigenvalue of A.
Theorem 4:
For any polynomial p(s), if λ is an eigenvalue of a matrix A, and v is a corresponding eigenvector, then p(λ) is an eigenvalue of p(A) and v is a corresponding eigenvector.
Theorem 5:
If λ1, λ2, … , λr are distinct eigenvalues of a square matrix A, and is v1, v2, … , vr are corresponding eigenvectors, then { v1, v2, … , vr } is a linearly independent set.
A defective matrix is a square matrix that does not have a complete basis of eigenvectors.
Mathematica has some special commands (Eigensystem,
Eigenvalues, Eigenvectors, and CharacteristicPolynomial)
to deal with eigenvalues and eigenvectors for square matrices. We show how to
use them in a sequence of examples.
The main command that deals with an eigenvalues problem is
B.eigenvects, which gives a list {values, interger, vectors} of the
eigenvalues, multiplicities, and eigenvectors of the square matrix B.
Its characteristic polynomial
Since this quadratic equation \( \det\left( \lambda{\bf I} - {\bf A} \right) = \lambda \left( \lambda -5 \right) = 0 \) has two real roots λ = 0 and λ = 5, we know thgat these numbers are eigenvalues of matrix A.
Now we find eigenvectors corresponding to these eigenvalues. For λ = 0, we need to solve the vector equation
Therefore, we know that the matrix A has one double
eigenvalue λ = 1, and one simple eigenvalue
λ = 0 (which indicates that
matrix A is singular). Next, we find the corresponding
eigenvectors
So Python provides us only one eigenvector \( \xi = \left[ 1,0,0 \right] \)
corresponding to the eigenvalue λ = 1 (therefore, A is defective)
and one eigenvector v = <-1,1,0> corresponding
eigenvalue λ = 0. To check this, we introduce the matrix
B1:
which means that B1 has one simple eigenvalue \( \lambda = 1 \) and one
double eigenvalue \( \lambda =0. \) Then we check that \( \xi \) is an eigenvector
of the matrix A:
Then we check that v is the eigenvector corresponding to \( \lambda = 0: \)
To find the generalized eigenvector corresponding to \( \lambda = 1, \) we
use the following Python command
This gives us another generalized eigenvector \( \xi_2 = \left[ 0,-1,-1 \right] \)
corresponding to the eigenvalue \( \lambda = 1 \) (which you can multiply by
any constant). To check it, we calculate:
but the first power of B1 does not annihilate it:
■
The characteristic polynomial can be found either with
Mathematica's command
CharacteristicPolynomial or multiplying (λ - λk)m for each eigenvalue λk of multiplicity m, when eigenvalues are available.Remember that for odd dimensions, Mathematica's command
CharacteristicPolynomial provides negative value because it is based on the formula det(A - λI) rather than Eq.\eqref{EqEigen.2}.
This means that two equations do not define any constraint on the coordinates of the eigenvector. In other words, all three equations are equivalent and you can take any of them to define a relation between its coordinates.
Therefore, the coordinates of the eigenvectors corresponding to the double eigenvalue λ = 1 are related via the identity