Dr. Matthieu R Bloch
Wednesday, December 1, 2021
General announcements
Assignment 6 posted (last assignment)
Due December 7, 2021 for bonus, deadline December 10, 2021
2 lectures left
Let me know what’s missing
Assignment 5 grades posted
Reviewing Midterm2 grades one last time
The learning problem and why we need probabilities.
Lecture notes 17 and 23
Flip a biased coin, lands on head with unknown probability
Say we flip the coin
Can we relate
It is possible that
An unknown function
A dataset
A set of hypotheses
An algorithm
An unknown conditional distribution
A dataset
A set of hypotheses
An algorithm
A dataset
An unknown conditional distribution
A set of hypotheses
A loss function
An algorithm
Learning is not memorizing
Consider hypothesis
What we really care about is the true risk (a.k.a. out-sample error)
Question #1: Can we generalize?
Question #2: Can we learn well?
Consider a special case of the general supervised learning problem
Dataset
Unknown
Finite set of hypotheses
Binary loss function
In this very specific case, the true risk simplifies
The empirical risk becomes
Our objective is to find a hypothesis
For a fixed
Observe that for
The empirical risk is a sum of iid random variables
We’re in luck! Such bounds, a.k.a, known as concentration inequalities, are a well studied subject
Let
Let
Let
By the law of large number, we know that
Given enough data, we can generalize
How much data?
That’s not quite enough! We care about
If we choose
We can obtain much better bounds than with Chebyshev
Let
In our learning problem
We can now choose
How about learning
If
How do we make
The function
We have effectively already proved the following result
A finite hypothesis set
Ideally we want
In general this is not possible
Remember, we usually have to learn
Questions