Matthieu Bloch
April 9, 2020
Method | Parameter |
---|---|
Polynomial regression | polynomial degree \(d\) |
Ridge regression | \(\lambda\) |
SVMs | margin violation constraint \(C\) |
Kernel methods | kernel choice |
\(K\) nearest neighbor | \(K\) |
In addition to training data \(\calD\), suppose we have access to another validation set \(\calV\eqdef\{(\bfx_i,y_i\}_{i=1}^K\)
Assume \(h\) selected from training set \(\calD\) and use the validation to form an estimate \[\widehat{R}_{\textsf{val}}(h)\eqdef\frac{1}{K}\sum_{i=1}^K\ell(h(\bfx_i),y_i)\]
Using validation for model selection
Effect of bias
Dilemma: we would like \(R(h)\approx R(h^{-})\approx\widehat{R}_{\textsf{val}}(h^{-})\)