Lecture from 11.12.2024 | Video: Videos ETHZ
Eigenvalues and Eigenvectors: A Summary
Let’s recap some key properties of eigenvalues and eigenvectors for a matrix :
- Complex Conjugate Eigenvalue: If is an eigenvalue-eigenvector pair of , then is also an eigenvalue-eigenvector pair, where and denote the complex conjugates.
- Eigenvalues of and : and share the same eigenvalues, but their eigenvectors are generally different.
- Eigenvalues of : If is invertible and is an eigenvalue-eigenvector pair, then is an eigenvalue-eigenvector pair for . Note that since is invertible.
- Linearity of Eigenvalues: Eigenvalues do not behave linearly under matrix addition or multiplication. In general, the eigenvalues of are not the sum of the eigenvalues of and , and the eigenvalues of are not the product of the eigenvalues of and .
- Gaussian Elimination and Eigenvalues: Gaussian elimination does not preserve eigenvalues.
- Eigenvalues of Orthogonal Matrices: If is an orthogonal matrix () and is an eigenvalue of , then .
Theorem: Basis of Eigenvectors for Distinct Eigenvalues
If has distinct real eigenvalues, then there exists a basis for consisting of the eigenvectors of . This basis enables diagonalization, simplifying various computations involving . However, what happens when eigenvalues are repeated?
Repeated Eigenvalues: Challenges in Building a Basis
When eigenvalues are repeated, constructing a basis of eigenvectors becomes more challenging.
Definition: The algebraic multiplicity of an eigenvalue is the number of times it appears as a root of the characteristic polynomial.
Example
-
The characteristic polynomial is . The eigenvalue 0 has algebraic multiplicity 2. The eigenspace corresponding to is , which is spanned by the vector . Its dimension is 1. Thus, there is only one linearly independent eigenvector.
-
The characteristic polynomial is . is the only eigenvalue, again with algebraic multiplicity 2. However, the eigenspace is all of (any vector is an eigenvector), which has dimension 2. We have two linearly independent eigenvectors and .
These examples illustrate that a repeated eigenvalue might not yield enough linearly independent eigenvectors to form a basis. This leads to the following definition:
Complete Set of Real Eigenvectors
Definition: A matrix has a complete set of real eigenvectors if there exists a basis for consisting of eigenvectors of . So in our example the first matrix doesn’t have such a complete set of eigenvectors, the second matrix however does have such a set.
When Do We Have a Complete Set of Real Eigenvectors?
A crucial question in linear algebra is whether a given matrix possesses a complete set of eigenvectors that form a basis for the vector space. This property, known as diagonalizability, significantly simplifies matrix computations and analysis.
Definition: Complete Set of Eigenvectors: A matrix has a complete set of real eigenvectors if there exists a basis for consisting entirely of eigenvectors of .
Let’s examine specific cases:
1. Distinct Eigenvalues
Proposition: If a matrix has distinct real eigenvalues, then it has a complete set of linearly independent eigenvectors, hence forming a basis for .
Proof: We have already proved (in the last lecture note [[24 Explicit Fibonacci Formula, Eigenvalue and Eigenvector properties, Trace#Key Properties and Results#Linear Independence of Eigenvectors]]) that eigenvectors corresponding to distinct eigenvalues are linearly independent. If a matrix has distinct eigenvalues, it consequently has linearly independent eigenvectors. Since these eigenvectors reside in , linearly independent vectors form a basis. Therefore, the matrix has a complete set of eigenvectors.
2. Diagonal Matrices
Proposition: A diagonal matrix always has a complete set of linearly independent eigenvectors.
Proof: Let be a diagonal matrix. The characteristic polynomial is given by:
Setting the characteristic polynomial to zero, we find that the eigenvalues are the diagonal entries: for Now let’s consider the standard basis vectors (a vector with 1 in the -th position and 0 elsewhere). Observe:
Thus, each standard basis vector is an eigenvector corresponding to the eigenvalue . Since the standard basis vectors are linearly independent and span , the diagonal matrix has a complete set of linearly independent eigenvectors. Note that if there are repeated entries on the diagonal, we still have a complete set of linearly independent eigenvectors; it’s just that multiple eigenvectors will correspond to the same eigenvalue.
3. Repeated Eigenvalues: Algebraic and Geometric Multiplicity
When an eigenvalue is repeated (i.e., its algebraic multiplicity is greater than 1), the situation is more nuanced.
Definition: Algebraic Multiplicity
The algebraic multiplicity of an eigenvalue is the number of times it appears as a root of the characteristic polynomial.
Definition: Geometric Multiplicity
The geometric multiplicity of an eigenvalue is the dimension of its eigenspace, - the number of linearly independent eigenvectors associated with that eigenvalue.
Explanation:
-
Algebraic multiplicity counts how many times an eigenvalue appears as a root of the characteristic polynomial. This tells us how many eigenvectors we expect for that eigenvalue.
-
Geometric multiplicity counts the maximum number of linearly independent eigenvectors we can actually find for an eigenvalue. This is the dimension of the eigenspace.
To form a basis, we need linearly independent eigenvectors for an matrix.
If the geometric multiplicity of an eigenvalue is less than its algebraic multiplicity, we have “missing” eigenvectors—we don’t have enough linearly independent vectors for that eigenvalue to contribute its “full share” to a basis.
Only when the geometric multiplicity equals the algebraic multiplicity for all eigenvalues do we have enough linearly independent eigenvectors to span the entire space and form a complete set, thus making the matrix diagonalizable.
Crucial Condition for a Complete Set
A matrix has a complete set of eigenvectors if and only if for every eigenvalue, its geometric multiplicity is equal to its algebraic multiplicity.
Example (Repeated Eigenvalue, Insufficient Eigenvectors)
The characteristic polynomial is , so has algebraic multiplicity 2. However, the eigenspace has dimension 1 (spanned by ), so the geometric multiplicity is 1. Since the geometric multiplicity is less than the algebraic multiplicity, this matrix does not have a complete set of eigenvectors and is not diagonalizable.
Projection Matrices and Complete Sets of Eigenvectors
Projection matrices, representing projections onto subspaces, possess a special eigenstructure that always guarantees a complete set of eigenvectors.
Proposition: Let be the orthogonal projection matrix onto a subspace . Then has two eigenvalues: 0 and 1, and it always has a complete set of real eigenvectors that form a basis for
Proof
Eigenvalues 0 and 1
Let . We can uniquely decompose as where and (the orthogonal complement of ). By definition of orthogonal projection:
- Eigenvalue 1: If , then . Thus, any non-zero vector in is an eigenvector of with eigenvalue 1.
- Eigenvalue 0: If , then . Thus, any non-zero vector in is an eigenvector of with eigenvalue 0.
Complete Set of Eigenvectors
Let . Let be an orthonormal basis for , and let be an orthonormal basis for .
-
Eigenvectors for eigenvalue 1: The vectors are eigenvectors of with eigenvalue 1 (as shown above).
-
Eigenvectors for eigenvalue 0: The vectors are eigenvectors of with eigenvalue 0 (as shown above).
-
Combined Set: The set forms an orthonormal basis for consisting entirely of eigenvectors of
- Linear Independence: Since it is a basis for the set of vectors are linearly independent.
- Spanning Property: The set spans all of as it is a basis.
Explicit Form of Eigenvectors: The eigenvectors of the projection matrix are any set of orthonormal vectors that form bases for the subspace and its orthogonal complement .
Diagonalization: Leveraging a Complete Set of Eigenvectors
Theorem: Let have a complete set of eigenvectors associated with eigenvalues , respectively. Let be the matrix whose columns are these eigenvectors. Then:
where is a diagonal matrix with the eigenvalues on the diagonal. This process is called diagonalization.
Diagonalization and Powers of A... (MIT)
Proof
-
Invertibility of : Since has a complete set of eigenvectors, these eigenvectors form a basis for which mean they are linearly independent. Consequently, the matrix , having these linearly independent vectors as its columns, is invertible.
-
Eigenvalue Equation in Matrix Form: The individual eigenvalue equations for , can be expressed concisely in matrix form:
-
Introducing : Notice that the rightmost expression in the previous step can be written as the product of and the diagonal matrix : Let’s see why:
-
Diagonalization: Combining the steps, we have . Since is invertible, we can multiply both sides on the left by :
This equation demonstrates that is similar to the diagonal matrix
-
Rewriting the Equation: Finally, to obtain the diagonalization of , we can multiply both sides of on the left by and on the right by :
Similar Matrices
Definition: Two matrices and are similar if there exists an invertible matrix such that
Proposition: Similar matrices have the same eigenvalues.
Proof
If , and , then:
Thus, if is an eigenvalue of with eigenvector , then is an eigenvalue of with eigenvector
Diagonalization, when possible, finds a similar matrix that is diagonal, greatly simplifying computations and revealing the core structure of the linear transformation.
Why Diagonalization Matters
Diagonalization simplifies many matrix computations. For instance, calculating powers of becomes much easier:
Since is diagonal, is simply computed by raising the diagonal entries (eigenvalues) to the power . This simplification extends to other matrix functions, making diagonalization a powerful technique in various applications.
Change of Basis and Diagonalization
Diagonalization, when possible, provides a way to simplify the representation of a linear transformation by choosing a basis of eigenvectors. In this basis, the transformation acts as a simple scaling along each eigenvector direction. Let’s explore this concept in detail, starting with the general case of a linear transformation and then specializing to diagonalizable matrices.
General Case: Change of Basis
Consider a linear transformation and two sets of bases:
- for (the “old” basis).
- for (the “new” basis).
Let and be the matrices whose columns are the respective basis vectors. These matrices are invertible since their columns are linearly independent. Any vector can be expressed in the old basis as . Let . Then .
The transformation results in a vector in , which can be expressed in the new basis as . Similarly, let Then
We want to find a matrix that represents the transformation with respect to these new bases, such that . To do this, let be the matrix representing the linear transformation in the standard bases. Then .
Using the change of basis matrices: To find , we need to express in terms of : Thus, the matrix representing the transformation in the new bases is:
Diagonalization: A Special Case of Change of Basis
When is diagonalizable (i.e., it has a complete set of linearly independent eigenvectors), we can choose the eigenvectors as the new basis for both the domain and codomain ().
Let be the matrix whose columns are the eigenvectors of . Then, in this eigenvector basis, the transformation is represented by the diagonal matrix where has the eigenvalues of on its diagonal.
In this case, the change of basis simplifies the transformation to a scaling along each eigenvector direction, as the matrix representing the transformation becomes diagonal.
Continue here: 26 Symmetric Matrices, Spectral Theorem, Rayleigh Quotient