Matrix Decompositions
Diagonalizable Matrices
is diagonalizable if and only if it has linearly independent eigenvectors, or equivalently, if the geometric multiplicity and the algebraic multiplicity of all the eigenvalues agree. A special case of this is if has distinct eigenvalues. Suppose we have eigenvalues and corresponding eigenvectors . Then
Intuitively, this says that we can find a basis consisting of the eigenvectors of . Useful for computing large powers of , as . An important example is being real and symmetric implies is diagonalizable.
Singular Value Decomposition
SVD is powerful in low-rank approximations of matrices. Unlike eigenvalue decomposition, SVD uses two unique bases (left/right singular vectors). For orthogonal matrices and diagonal matrix with nonnegative diagonal entries in nonincreasing order, we can write any matrix as:
Intuitively, this says that we can express as a diagonal matrix with suitable choices of (orthogonal) bases.
QR Decomposition
For nonsingular , we can write , where is orthogonal and is an upper triangular matrix with positive diagonal elements. QR decomposition assists in increasing the efficiency of solving for nonsingular :
QR decomposition is very useful in efficiently solving large numerical systems and inversion of matrices. Furthermore, it is also used in least-squares when our data is not full rank.
LU and Cholesky Decompositions
For nonsingular , we can write , where is a lower and is an upper triangular matrix. This decomposition assists in solving as well as computing the determinant:
If is symmetric positive definite, then can be expressed as via Cholesky decomposition, where is an upper triangular matrix with positive diagonal entries. Cholesky decomposition is essentially LU decomposition with . These decompositions are both useful for solving large linear systems.
Projections
Fix a vector . The projection of onto is given by
More generally, if has orthogonal basis , then the projection of onto is given by
The main property is that and is orthogonal to any . Linear Regression can be viewed as a projection of our observed data onto the subspace formed by the span of the collected data.