Localized Regression
#1
Posted 2011-January-13, 08:10
Thought folks might find it amusing to see the fun part of my work...
http://blogs.mathwor...driven-fitting/
(I'm quite serious about the "fun" part)
#2
Posted 2011-January-13, 09:05
The infliction of cruelty with a good conscience is a delight to moralists — that is why they invented hell. — Bertrand Russell
#3
Posted 2011-January-13, 10:00
#4
Posted 2011-January-13, 10:13
On a more on-topic note, the subject is interesting. I learned some curve-fitting when I was undergraduate, but, since it was physics, we always had some underlying theory for the main effect.
#5
Posted 2011-January-13, 10:12
eg Santa Catalina island rattlesnake which has lost its rattle because there are no large mammals to warn off.
Interesting article.
#6
Posted 2011-January-13, 18:17
#7
Posted 2011-January-13, 18:25
In theory, you should be able to extend the technique to cover N independent variables.
In practice, LOWESS requires a KNN search operation and KNN search slows down a lot with large numbers of dimensions.
Most people seem to switch over to boosted or bagged decision trees for large numbers of N.
#9
Posted 2011-January-14, 15:18
#10
Posted 2011-January-14, 16:16
I wouldn't characterize anything in the blog post as novel. I was simply trying to provide a useful illustration of localized regression and the bootstrap.
It's pretty easy to find references about the application of this techniques within Economics for low dimensional spaces.
(Once you hit higher numbers of dimensions boosted and bagged regression trees are preferred).
There's also a lot of debate about the relative merits of localized regression versus smoothing splines in low dimensional spaces.
For example, the following piece preferred localized regression
http://www.cs.pitt.e...er/sisref/3.pdf
while
http://www.polisci.o.../lkeele/TOC.pdf
argued that smoothing splines are preferrable
#11
Posted 2011-January-14, 16:18
Lo(w)ess is linear (or quadratic) regression where one model is fitted for each value of the independent variable, in that the points used for regression are weighted on the basis of their distance to the point of interest, using a dynamic-bandwidth kernel. It is some weird kernel the motivation of which I don't quite understand but I suppose the inventors had their reasons

#12
Posted 2011-January-17, 05:15
Last week we had a guest speaker from New Zealand who had developped kernel smoothing specifically for estimation of local odds, i.e. you have two processes X and Y which you could fit by Lowess or such. Rather than fitting the two independently and then compute odds, he fitted the local log-odds directly. I found that interesting although of course a very specific thing.
#13
Posted 2011-January-18, 10:06
The link to Efron:
http://www.nytimes.c...r=1&ref=science
His letter links to an article in the Times:
http://www.nytimes.c...ence/11esp.html
In the "small world" department, I wrote a term paper on esp when I was in high school (when Rhine at Duke was a big cheese), and I sort of knew Brad Efron then. He went to St. Paul Central, I went to St. Paul Monroe. My friend Joe Auslander pointed me to Efron's letter because he knew of this (weak) high school link. Richard's blog mentions resampling, and that relates back to Efron.
I haven't the time right now, but I am interested in Richard's blog and intend to get to it.