Linear Algebra - Special Matrices, SVD, Least Squares
Wednesday, September 7, 2022
A square matrix \(Q\) is called orthogonal if its columns are an orthonormal set, i.e. \(Q^T Q = I\).
A square matrix \(S\) is called symmetric if it is equal to its transpose, i.e. \(S^T = S\).
Any matrix \(A \in \bbR^{m \times n}\) may be described by two orthogonal matrices \(U \in \bbR^{m \times m}, V \in \bbR^{n \times n}\) and a diagonal (up to the matrix border) non-negative matrix \(\Sigma\) as \[ A = U \Sigma V^T \]
Intuition: The SVD is a powerful tool to describe the action of any matrix in terms of an orthonormal change of basis (\(V^T\)), non-negative component-wise scaling (\(\Sigma\)), and an orthonormal basis reconstruction (\(U\)).
Drill: Let \(A\) have SVD \(A = U \Sigma V^T\). Prove that \(A^T A\) and \(A A^T\) share all of their nonzero eigenvalues, and further show that these eigenvalues are exactly the squares of the nonzero diagonal elements of \(\Sigma\).
Given a matrix \(A \in \bbR^{m \times n}\) and a vector \(b \in \bbR^m\), \[ \underset{x \in \bbR^n}{\text{minimize}} || A x - b ||_2^2 \]
Given a matrix \(A \in \bbR^{m \times n}\) and a vector \(b \in \bbR^m\), \[ \underset{x \in \bbR^n}{\text{minimize}} || A x - b ||_2^2 \]
The pseudoinverse of \(A\) is defined using the compact SVD as \(A^+ = V \Sigma^{-1} U^T\).