6.1: Basic Facts about Eigenvalue Problems
( \newcommand{\kernel}{\mathrm{null}\,}\)
Even if a matrix A is real, its eigenvectors and eigenvalues can be complex. For example,
[11−11][1i]=(1+i)[1i].
Eigenvectors are not uniquely defined. Given an eigenvector →x, any nonzero complex multiple of that vector is also an eigenvector of the same matrix, with the same eigenvalue. We can reduce this ambiguity by normalizing eigenvectors to a fixed unit length:
N−1∑n=0|xn|2=1.
Note, however, that even after normalization, there is still an inherent ambiguity in the overall complex phase. Multiplying a normalized eigenvector by any phase factor eiϕ gives another normalized eigenvector with the same eigenvalue.
6.1.1 Matrix Diagonalization
Most matrices are diagonalizable, meaning that their eigenvectors span the N-dimensional complex space (where N is the matrix size). Matrices which are not diagonalizable are called defective. Many classes of matrices that are relevant to physics (such as Hermitian matrices) are always diagonalizable; i.e., never defective.
The reason for the term "diagonalizable" is as follows. A diagonalizable N×N matrix A has eigenvectors that span the N-dimensional space, meaning that we can choose N linearly independent eigenvectors, {→x0,→x1,⋯→xN−1}, with eigenvalues {λ0,λ1,⋯λN−1}. We refer to such a set of N eigenvalues as the "eigenvalues of A". If we group the eigenvectors into an N×N matrix
Q=[→x0,→x1,⋯→xN−1],
then, since the eigenvectors are linearly independent, Q is guaranteed to be invertible. Using the eigenvalue equation, we can then show that
Q−1AQ=[λ00⋯00λ1⋯0⋮⋮⋱⋮00⋯λN−1].
In other words, there exists a similarity transformation which converts A into a diagonal matrix. The N numbers along the diagonal are precisely the eigenvalues of A.
6.1.2 The Characteristic Polynomial
One of the most important consequences of diagonalizability is that the determinant of a diagonalizable matrix A is the product of its eigenvalues:
det(A)=N−1∏n=0λn
This can be proven by taking the determinant of the similarity transformation equation, and using (i) the property of the determinant that det(UV)=det(U)det(V), and (ii) the fact that the determinant of a diagonal matrix is the product of the elements along the diagonal.
In particular, the determinant of A is zero if one of its eigenvalues is zero. This fact can be further applied to the following re-arrangement of the eigenvalue equation:
(A−λI)→x=0,
where I is the N×N identity matrix. This says that the matrix A−λI has an eigenvalue of zero, meaning that for any eigenvalue λ,
det(A−λI)=0.
The left-hand side of the above equation is a polynomial in the variable λ, of degree N. This is called the characteristic polynomial of the matrix A. Its roots are eigenvalues of A, and vice versa.
For 2×2 matrices, the standard way of calculating the eigenvalues is to find the roots of the characteristic polynomial. However, this is not a reliable method for finding the eigenvalues of larger matrices. There is a well-known and important result in mathematics, known as Abel's impossibility theorem, which states that polynomials of degree 5 and higher have no general algebraic solution. (By comparison, degree-2 polynomials have a general algebraic solution, which is the familiar quadratic formula, and similar formulas exist for degree-3 and degree-4 polynomials.) A matrix of size N≥5 has a characteristic polynomial of degree N≥5, and Abel's impossibility theorem tells us that we can't calculate the roots of that characteristic polynomial by ordinary arithmetic.
In fact, Abel's impossibility theorem leads to an even stronger conclusion: there is no general algebraic method for finding the eigenvalues of a matrix of size N≥5, whether using the characteristic polynomial or any other method. For suppose we had such a method for finding the eigenvalues of a matrix. Then, for any polynomial equation of degree N≥5, of the form
a0+a1λ+⋯+aN−1λN−1+λN=0,
we can construct an N×N "companion matrix" of the form
A=[010⋯0001⋯0⋮⋮⋱⋱⋮000⋱1−a0−a1−a2⋯−aN−1].
As you can check for yourself, each root λ of the polynomial is also an eigenvalue of the companion matrix, with corresponding eigenvector
→x=[1λ⋮λN−1].
Hence, if there exists a general algebraic method for finding the eigenvalues of a large matrix, that would allow us to find solve polynomial equations of high degree. Abel's impossibility theorem tells us that no such solution method can exist.
This might seem like a terrible problem, but in fact there's a way around it, as we'll shortly see.
6.1.3 Hermitian Matrices
A Hermitian matrix H is a matrix which has the property
H†=H,
where H† denotes the "Hermitian conjugate", which is matrix transposition accompanied by complex conjugation:
H†≡(HT)∗,i.e.(H†)ij=H∗ji.
Hermitian matrices have the nice property that all their eigenvalues are real. This can be easily proven using index notation:
∑jHijxj=λxi⇒∑jx∗jHji=λ∗x∗i⇒∑ijx∗iHijxj=λ∑i|xi|2=λ∗∑j|xj|2⇒λ=λ∗.
In quantum mechanics, Hermitian matrices play a special role: they represent measurement operators, and their eigenvalues (which are restricted to the real numbers) are the set of possible measurement outcomes.