Dr. Matthieu R Bloch
Monday, December 6, 2021
General announcements
Assignment 6 due December 7, 2021 for bonus, deadline December 10, 2021
Last lecture!
Let me know what’s missing
Expect an email from me tonight
Midterm 2 statistics
Hilbert spaces
Spaces of functions can be manipulated almost just as easily
Finite dimensional is fairly natural
Infinite dimensional can be manipulated just as well using orthobases
With orthobases, vectors in infinite dimensional separates Hilbert spaces are like square summable sequences
Regression
Who knew solving
SVD provides lots of insights
Regression in Hilbert spaces
More on learning and Bayes classifiers
Lecture notes 17 and 23
Consider a special case of the general supervised learning problem
Dataset
Unknown
Finite set of hypotheses
Binary loss function
In this very specific case, the true risk simplifies
The empirical risk becomes
Our objective is to find a hypothesis
For a fixed
Observe that for
The empirical risk is a sum of iid random variables
We’re in luck! Such bounds, a.k.a, known as concentration inequalities, are a well studied subject
Let
Let
Let
By the law of large number, we know that
Given enough data, we can generalize
How much data?
That’s not quite enough! We care about
If we choose
We can obtain much better bounds than with Chebyshev
Let
In our learning problem
We can now choose
How about learning
If
How do we make
The function
We have effectively already proved the following result
A finite hypothesis set
Ideally we want
In general this is not possible
Remember, we usually have to learn
Questions
We revisit the supervised learning setup (slight change in notation)
Dataset
Unknown
Binary loss function
The risk of a classifier
We will not directly worry about
The classifier
For
If all classes are equally likely
Assume
In practice we do not know
Back to our training dataset
The nearest-neighbor (NN) classifier is
Risk of NN classifier conditioned on
Let
Let
Let
If