Mathematical Foundations of Machine Learning
Prof. Matthieu Bloch
Monday September 16, 2024 (v1.1)
Last time
- Last class: Wednesday, September 11, 2024
- We wrapped up our discussion of infinite dimensional spaces
- Take away: separable Hilbert spaces are like
- To be effectively prepared for today's class, you should
have:
- Come to class last week
- Gone over slides
and read associated lecture
notes
- Have submitted Homework 2
- Logistics
- Office hours
- Jack Hill: 11:30am-12:30pm on Wednesdays in TSRB and hybrid
- Anuvab: 12pm-1pm on Thursdays in TSRB and hybrid
- Gradescope
- Make sure you assign each solution to each question
- Make sure you do some quality control on your submission
Regression
A fundamental problem in supervised machine learning can be cast
as follows
- Often , but
sometimes is a weirder
object (think tRNA string)
- if
with , the
problem is called classification
- if , the
problem is called regression
We need to introduce several ingredients to make the question
well defined
- We need a class to which
should belong
- We need a loss function to measure
the quality of our approximation
We can then formulate the question as
We will focus quite a bit on the square loss
, called
least-square regression
Least square linear regression
Least square affine regression
Canonical form I
- Stack as row vectors
into a matrix , stack as elements
of column vector
Canonical form II
Nonlinear regression using a basis
Solving the least-squares problem
Any solution to the
problem must satisfy This system is called normal
equations
Facts: for any matrix
We can say a lot more about the normal equations
- There is always a solution
- If , there
is a unique solution:
- if
there are infinitely many non-trivial solution
- if , there
exists a solution for
which
In machine learning, there are often infinitely many
solutions
Minimum norm 2 solutions
Reasonable approach: choose a solution among
infinitely many with the minimum energy principle
- We will see the solution is always unique using the SVD
For now, assume that , so that the
problem becomes
Regularization
- Recall the problem
- There are infinitely many solution if is non trivial
- The space of solution is unbounded!
- Even if , the
system can be poorly conditioned
- Regularization with consists in solving
- This problem always has a unique solution
- Note that is in the
row space of
Next time
- Next class: Wednesday September 18, 2024
- To be effectively prepared for next class, you
should:
- Go over today's slides
and read associated lecture
notes
- Work on Homework 3 (due date: Wednesday September 25,
2024)
- Optional
- Export slides for next lecture as PDF (be on the lookout for an
announcement when they're ready)


1/1
Mathematical Foundations of Machine Learning
Prof. Matthieu Bloch
Monday September 16, 2024 (v1.1)