Last modified: February 15, 2022

This article is written in: 🇺🇸

Eigenvalue Decomposition (EVD)

Eigenvalue Decomposition (EVD), also known as Eigendecomposition, is a fundamental operation in linear algebra that breaks down a square matrix into a simpler form defined by its eigenvalues and eigenvectors. This decomposition provides deep insights into the properties and structure of a matrix, enabling simplifications in numerous computations such as matrix powers, transformations, and the solution of systems of linear equations. In many applications — from vibration analysis in engineering to principal component analysis in data science — EVD plays a critical role.

The key idea is that certain square matrices can be represented as a product of a matrix of their eigenvectors and a diagonal matrix of their eigenvalues. This diagonalization separates the scaling factors (eigenvalues) from the directions of transformations (eigenvectors) encoded by the original matrix.

output(19)

Mathematical Formulation

Consider an n×n matrix A. The Eigenvalue Decomposition of A is given by:

A=PDP1, where:

I. Eigenvalues (λi):

The eigenvalues of A are the roots of the characteristic polynomial:

det(AλI)=0.

If we solve this polynomial equation and find n eigenvalues (not necessarily distinct), we denote them as λ1,λ2,,λn.

II. Eigenvectors (vi):

For each eigenvalue λi, we find the corresponding eigenvector vi by solving:

(AλiI)vi=0.

Each eigenvector vi is a nonzero vector associated with the eigenvalue λi.

III. Eigenvector Matrix (P) and Eigenvalue Matrix (D):

Once we have the set of eigenvalues and eigenvectors:

Construct D as a diagonal matrix whose diagonal entries are the eigenvalues:

D=[λ1000λ2000λn]

Construct P such that its columns are the eigenvectors v1,v2,,vn:

P=[|||v1v2vn|||]

If A is diagonalizable, and if the vi are chosen to be linearly independent, then P is invertible and A=PDP1.

Derivation

The derivation of EVD follows naturally from the definition of eigenvalues and eigenvectors. For each eigenvalue λi, we have:

Avi=λivi.

If we gather all eigenvectors into the matrix P and consider how A acts on the columns of P:

A[v1v2vn]=[Av1Av2Avn]=[λ1v1λ2v2λnvn]=PD.

If P is invertible, we can write:

A=PDP1.

Not every matrix is guaranteed to have a full set of n linearly independent eigenvectors. If it does, the matrix is said to be diagonalizable, and the above decomposition is possible. If not, the matrix cannot be decomposed purely into this form.

Algorithm Steps

I. Find Eigenvalues:

II. Find Eigenvectors:

III. Form P and D:

IV. Verify Invertibility of P:

Example

Let:

A=[4123]

I. Find Eigenvalues:

The characteristic polynomial:

det(AλI)=det[4λ1 23λ]=(4λ)(3λ)2=0.

Expanding:

(4λ)(3λ)2=(127λ+λ2)2=λ27λ+10=0.

Solve λ27λ+10=0:

The roots are λ1=5 and λ2=2.

II. Find Eigenvectors:

For λ1=5:

(A5I)=[11 22].

Solve (A5I)v=0 leads to v1=[1,1]T.

For λ2=2:

(A2I)=[21 21]

Solve (A2I)v=0 leads to v2=[1,2]T.

III. Form P and D:

P=[11 12],D=[50 02]

Thus:

A=PDP1.

Advantages

I. Simplification of Computations:

Once in the form A=PDP1, computing powers of A or applying certain transformations becomes much simpler. For example, Ak=PDkP1, and since D is diagonal, raising it to a power is straightforward.

II. Insights into Matrix Structure:

The eigendecomposition reveals the intrinsic "modes" of the linear transformation represented by A. Eigenvalues show how the transformation scales each eigen-direction, and eigenvectors show the directions of these fundamental modes.

III. Numerical Stability in Some Computations:

Working with D instead of A can improve numerical stability and make some algorithms more efficient, particularly in areas like principal component analysis, spectral clustering, and other advanced data analysis tasks.

Limitations

I. Not All Matrices Are Diagonalizable:

Some matrices cannot be broken down into a pure eigen decomposition if they do not have enough linearly independent eigenvectors. For such matrices, more generalized decompositions like the Jordan normal form are required.

II. Computational Cost for Large Matrices:

Finding eigenvalues and eigenvectors for large matrices can be computationally expensive. Efficient numerical algorithms and approximations exist, but they may still be costly for very large systems.

III. Complex Eigenvalues:

For real matrices, eigenvalues can be complex. While this is not a fundamental limitation, it means we must consider complex arithmetic when performing the decomposition, which may not be desired in some real-world applications.

Table of Contents

    Eigenvalue Decomposition (EVD)
    1. Mathematical Formulation
    2. Derivation
    3. Algorithm Steps
    4. Example
    5. Advantages
    6. Limitations