Class 20: Introduction to Smoothing Splines
- We reviewed the various methods for smoothing, working from the
class handout.
Smoothing Splines
- Lisp script for today's class
- Class 21
- Handout
- This is a draft. We'll add details in class for some of the proofs, etc.
- Loss Function
- The loss function which leads to splines is a penalized sum of squares.
Among all functions with the same fidelity to the data, cubic splines
minimize this integrated second derivative penalty, and are thus the
smoother of choice for this criterion.
- Computing a Spline
- Forget smoothing for a moment. How does on compute a spline. We have lots
of choices, most of which resemble a regression function of sorts, with various
choices for the regression design matrix. But how does one use these things
for smoothing. You have to add in the loss function. One formulation, the one which
uses B-splines, gives a useful form for the loss function.
An important alternative is to remove knots. This different strategy leads
to the "turbo smoother" of Friedman and Silverman.
- Bayesian Spin
- Though it may be surprising, other arguments lead to the same time of estimating
equations. These penalized least squares estimators can also be motivated from
a "shrinkage perspective" (as in ridge regression) or by posing a prior distribution
located at zero.
Next time
We'll do more calculations, using (if I can make it work on the Sun by then) some fast code
for smoothing splines that is written in C. Also, we'll look more closely at some of the
details of the B-splines that will not be done today.