Key concepts: special/structured matrices,
SVD, least squares
Key results:
structured matrices often have desirable
properties
the SVD is a powerful tool for analysis of
matrices
the ubiquitous least squares problem has an
intimate relationship to the SVD.
Next workshop: Friday September 09,
2022
Special/Structured Matrices
A square matrix is called
orthogonal if its columns are an orthonormal set, i.e. .
Drill: Prove that every orthogonal matrix
also has orthogonal rows.
Hint: what is one way to interpret the statement
?
Drill: Prove that every orthogonal matrix
has determinant .
Hint: Recall two properties of the determinant are
and .
A square matrix is called
symmetric if it is equal to its transpose, i.e. .
Properties:
The set of symmetric matrices is a vector subspace
of the set of all real square matrices.
is
symmetric if and only if is
symmetric.
All symmetric matrices are diagonalizable. Further,
for any symmetric matrix there
exists an orthonormal set of eigenvectors such that has eigendecomposition , where is an orthogonal matrix; all
eigenvalues are guaranteed to be real, hence is a real diagonal matrix.
(These are results of the spectral theorem, which
really deserves its own lecture…)
Drill: Prove that for symmetric matrix
and any positive integer the matrix is symmetric.
Singular Value Decomposition (SVD)
Any matrix may be described by two orthogonal matrices and a diagonal (up to the matrix border) non-negative matrix
as
Visualization of the SVD for a 2x2 matrix (reproduced with
permission from Wikipedia)
Intuition: The SVD is a powerful tool
to describe the action of any matrix in terms of an
orthonormal change of basis (),
non-negative component-wise scaling (), and an orthonormal basis
reconstruction ().
Drill: Let have SVD . Prove that and share all of
their nonzero eigenvalues, and further show that these eigenvalues are
exactly the squares of the nonzero diagonal elements of .
Least Squares
Given a matrix and a vector ,
Practical Examples:
To find the line of best fit over a series of
points ,
Least Squares
Practical Examples (cont.):
To find the th order polynomial of best fit over
a series of points ,
To find the best fit fourier series over a series
of points ,
where
Least Squares
Given a matrix and a vector ,
Drill: What is the minimum value can take over all
possible ?
For the following drills, assume
Drill: If , how many solutions to the least squares problem are there?
What is a closed form expression for these solutions?
Drill: If , how many solutions to the least squares problem are
there? What is a closed form expression for these solutions?
Hint: Recall that every set of vectors has a
linearly independent subset which have the same span.
Hint: Recall that every linearly independent set of
vectors in which span the
entirety of contains exactly
vectors.
Drill: If , how many solutions to the least squares problem are
there? What is a closed form expression for these solutions?
Least Squares
Luckily, the least squares problem is convex, so we
can solve by setting the derivative equal to 0:
Issue when , is rank-deficient
and can’t be inverted…
The pseudoinverse of is
defined using the compact SVD as .
Drill: Prove directly by substitution that
is a projection matrix onto
the column space of .
Problems
Work together to solve these problems!
Prove for any orthogonal matrix and arbitrary vector . In other words, multiplying
by an orthogonal matrix does not change the norm of a vector.
Prove that the transpose of an orthogonal matrix is still an
orthogonal matrix.
Let . Prove that the action of any
arbitrary orthogonal matrix is a bijective endomorphism, i.e. .
Let be a non-negative diagonal matrix with entries in descending
order, i.e. . Prove that .
Hint: assume that .
In this case, there exists some index such that . Another feasible
point in is , where all the entries are the
same as those in save that
and . Which has
a larger objective value?
The operator norm of a matrix is defined as Using the previous results, solve for the operator norm in
terms of the singular values of .