Mathematical Foundations of Machine Learning
Prof. Matthieu Bloch
Wednesday, November 6, 2024
Last time
- Last class: Monday November 04, 2024
- Today: We will use the SVD to better understand
regression problems
- To be effectively prepared for today's class, you should
have:
- Gone over slides
and read associated lecture notes here
and there
- Planned to submit Homework 6 (due Thursday November 07, 2024)
- Logistics:
- Jack Hill office hours: Wednesday 11:30am-12:30pm in TSRB
and hybrid
- Anuvab Sen office hours: Thursday 12pm-1pm in TSRB and
hybrid
- Dr. Bloch office hours: Friday November 08, 2024 6pm-7pm
online
- Homework 6: due Thursday November 7, 2024
What's next for this semester
Lecture 21 - Monday November 4, 2024: SVD and least
squares
- Lecture 22 - Wednesday November 6, 2024: Gradient descent
- Homework 6 due on Thursday November 7, 2024
- Lecture 23 - Monday November 11, 2024: Estimation
- Lecture 24 - Wednesday November 13, 2024: Estimation
- Homework 7 due on Friday November 15. 2024
- Lecture 25 - Monday November 18, 2024: Classification and
Regression
- Lecture 26 - Wednesday November 20, 2024: Classification and
Regression
- Lecture 27 - Monday November 25, 2024: Principal Component Analysis
- Lecture 28 - Monday December 2, 2024: Principal Component
Analysis
Singular value decomposition
- What happens for non-square matrices?
Let
with . Then
where
- such
that
(orthonormal columns)
- such
that
(orthonormal columns)
- is diagonal with positive entries
and .
The are called the
singular values
SVD and least-squares
- When we cannot solve , we solve instead
- This allows us to pick the minimum norm solution among potentially
infinitely many solutions of the normal equations.
- Recall: when is of rank , then
The solution of is where is
the SVD of .
Pseudo inverse
- is called the
pseudo-inverse of
- Also called Lanczos inverse, or Moore-Penrose inverse
- If is square invertible
then
- If (tall and thin
matrix) of rank then
- If (short and large
matrix) of rank then
- Note
is as "close" to an inverse of as possible (see notes and
Homework)
Stability of least squares
- What if we observe and we apply the pseudo inverse?
We can separate the error analysis into two components
We will express the error in terms of the SVD
With
-
orthobasis of ,
augmented by
to form an orthobasis of
-
orthobasis of ,
augmented by
to form an orthobasis of
The null space error is given by
The noise error is given by
Stable reconstruction by truncation
Stable reconstruction by regularization
Next time
- Next class: Monday November 11, 2024
- To be effectively prepared for next class, you
should:
- Go over today's slides
and read associated lecture notes here
and there
- Work on Homework 7
- Optional
- Export slides for next lecture as PDF (be on the lookout for an
announcement when they're ready)


1/1
Mathematical Foundations of Machine Learning
Prof. Matthieu Bloch
Wednesday, November 6, 2024