Symmetric matrices naturally occur in applications. For example the covariance matrix in statistics, and the adjacency matrix in graph theory, are both symmetric. In both of those situations it is desirable to find the eigenvalues of the matrix, because those eigenvalues have certain meaningful interpretations.
But you might ask: “What if there are non-real eigenvalues?” This is a great question, since in general real matrices might have non-real eigenvalues (See Exercise 5.6). Fortunately, we have the following theorem:
Let be a symmetric matrix. Then every eigenvalue of is a real number.
For the proof, see Exercise 5.31.
Choose your own real symmetric matrix which is not diagonal, and find its eigenvalues (they should be real!).
Find a real matrix which has at least one non-real eigenvalue.
[End of Exercise]
The following theorem decomposes into simpler, easier to work with components: and . Another way of finding the matrices and is to use the computer program R, with the command eigen.
Let .
If is a matrix whose columns form an orthonormal basis of eigenvectors of , and is the diagonal matrix of eigenvalues (in the same order), then
has an orthonormal basis of eigenvectors if and only if is symmetric.
This is also called the spectral theorem. The name comes from applications (in particular by physicists) where the set of eigenvalues of a matrix is called its “spectrum”.
To prove (i), assume is an orthonormal basis of eigenvectors of , and and are as in the Theorem. Then is the change of basis matrix from to the standard basis of . So by Theorem 4.54(i), is a diagonal matrix. Rearranging this equation we get . But the columns of form an orthonormal basis, so by Theorem 5.1 we have (i.e. is orthogonal). Therefore .
To prove (ii), firstly notice that if an orthonormal basis of eigenvectors exists, by part (i) we can write . Since , and diagonal matrices are symmetric, we see that , which shows is symmetric.
The difficult part of (ii) is the other direction. We omit the rest of this proof from this module, but include it here for completeness. Assume is symmetric. To construct an orthonormal basis, we proceed by induction on the size of . The base case, , follows because any unit vector is an eigenvector and also an orthonormal basis. So let , and assume for every square matrix of size , the statement of the theorem is true. By Lemma 5.5, we can choose an eigenvalue of , and let be a corresponding eigenvector with norm 1.
By Corollary 3.39, we can extend to an orthonormal basis of , which we can write . Notice that
, for any . So by Thereom 3.33, the vectors have coordinate equal to zero, in the basis . Therefore, if is the change of basis matrix from to the standard basis, then by Theorem 4.52
Since is orthogonal (Theorem 5.1), we know , so the matrix is symmetric; and therefore so is . Since is a real symmetric matrix of dimension , by our induction assumption, there is an orthonormal basis of eigenvectors of (in ), and therefore an orthogonal matrix such that , where is diagonal. We create our final matrix as a matrix product:
because then
So by Theorem 4.54(ii), the columns of form a basis of eigenvectors. Since is the product of orthogonal matrices, it is itself an orthogonal matrix, which means this basis is in fact orthonormal. Therefore the result holds for all by induction. ∎
Assume is symmetric with exactly one eigenvalue, . Prove that . [ Hint: Use the spectral decomposition of .]
[End of Exercise]
Find a basis of orthonormal eigenvectors for the following matrix, and hence find its spectral decomposition:
Solution: First, we find the eigenvalues, by finding the roots of the characteristic polynomial:
Cubic polynomials are, in general, hard to solve. If there is an integer solution (which in general, there is not, but one can hope!), then it must divide the constant term, which is 4. If we try , we see that . Therefore 1 is a root, so we can factor
Hence, the eigenvalues are and . Next, we compute each of the eigenspaces. Omitting details, we find: and .
Eigenvectors for two different eigenvalues are always orthogonal to each other (see Exercise 5.26). But if your eigenspace has dimension two or larger, then the basis you write down for it is not necessarily orthogonal.
In this example, both vectors in are orthogonal to , but they are not orthogonal to each other. How do we produce eigenvectors in which are orthogonal to each other? The answer is to use the Gram-Schmidt process. Let and . Then
.
By Theorem 3.36, both and are eigenvectors in , and are orthogonal to each other. Next, we scale so that they have length 1 (of course, scaling doesn’t change the fact that they are eigenvectors). So we obtain an orthonormal basis of eigenvectors:
Finally, we write down the corresponding change of basis matrix, and spectral decomposition:
As a final check, one could perform the matrix multiplication , and confirm that the resulting matrix is indeed ; one could also check that .
Summary of the method used in the above example:
For each eigenvalue , find a basis for the eigenspace ,
For each such that , find an orthogonal basis (use Gram-Schmidt),
Scale each vector to obtain an orthonormal basis of ,
is the matrix whose columns are the vectors in , and is the diagonal matrix whose entries are the eigenvalues of (in the same order as ).
If done correctly, you should be able to check .
Find a basis of orthonormal eigenvectors for the following matrix, and hence obtain its spectral decomposition.
If , where is orthogonal and diagonal, prove that the columns of are all eigenvectors of .
[End of Exercise]
A matrix formed by deleting a collection of rows and/or columns of a bigger matrix is known as a submatrix. Given a square matrix , the leading principal minor of size is the determinant of the submatrix consisting of the entries in the upper-left corner of , for any . In other words, the determinant of the matrix formed by deleting the right-most columns and the bottom rows.
Find the leading principal minors of .
Solution: The determinant of the upper-left submatrix is 1. The determinant of the upper-left submatrix is . The upper-left submatrix is all of , which has determinant . So the leading principal minors are and .
Find a matrix such that the coefficients of are all non-zero and the leading principal minors of are all positive numbers.
[End of Exercise]
We saw in Theorem 3.11 that a bilinear form is symmetric exactly when its associated matrix is symmetric. Similarly, we will call a matrix associated to a positive definite form a positive definite matrix; in other words:
for all non-zero .
Given a symmetric matrix, there are a few convenient tests for positive definiteness:
Let be real symmetric. The following are equivalent:
is positive definite,
All of the eigenvalues of are positive (i.e. ),
(Sylvester’s criterion) The leading principal minors are positive (i.e. ).
The criterion (iii) is named after English mathematician J.J. Sylvester (1814 - 1897) who discovered many fundamental results in matrix theory.
Assume (ii). By the spectral theorem, we can write , where invertible and with for all . Now take . Since is invertible, . Therefore we have that
For the opposite direction, assume (i). Let be an eigenvalue. By Theorem 5.5, , so we can find an eigenvector . In other words, there is a vector such that and . Since is positive definite by assumption, . But we also know that if then (since the standard scalar product is positive definite bilinear form; see 3.B). Therefore, . So we have shown (i)(ii).
We omit the proof that these are equivalent to (iii). ∎
Let .
Find the eigenvalues of .
Calculate the leading principal minors of .
Define 2 vectors of your choice, and check that for each of them .
In your opinion, which of these methods is the best to show positive definiteness?
Let . Prove that the leading principal minors are all positive, and also prove that is not positive definite. Why doesn’t this contradict Theorem 5.15?
[End of Exercise]