Honestly hurts my feelings a little that I didnโt even make this list ๐ฅฒ๐ฅฒ
@aliciacurth
Machine Learner by day, ๐ฆฎ Statistician at โค๏ธ In search of statistical intuition for modern ML & simple explanations for complex things๐ Interested in the mysteries of modern ML, causality & all of stats. Opinions my own. https://aliciacurth.github.io
Honestly hurts my feelings a little that I didnโt even make this list ๐ฅฒ๐ฅฒ
This is what I came to this app for ๐ฆฎ
Thank you for sharing!! Sounds super interesting, so will definitely check it out :)
Exactly this!! thank you ๐ค
Oh exciting! On which one? :)
To be fair, itโs actually a really really good TLDR!! Iโm honestly just a little scared this will end up on the wrong side of twitter now ๐ณ
Now might be the worst possible point in time to admit that I donโt own a physical copy of the book myself (yet!! Iโm actually building up a textbook bookshelf for myself) BUT because Hastie, Tibshirani & Friedman are the GOATs that they are, they made the pdf free: hastie.su.domains/ElemStatLearn/
Now might be the worst possible point in time to admit that I donโt own a physical copy of the book myself (yet!! Iโm actually building up a textbook bookshelf for myself) BUT because Hastie, Tibshirani & Friedman are the GOATs that they are, they made the pdf free: hastie.su.domains/ElemStatLearn/
Oh friends who are complaining about not enough Real Math^tm in their feed, I am here to help. Well, Alicia is here to help, at least!
To emphasise just how accurately that reflects Alanโs approach to research (which I 100% subscribe to btw), I feel compelled to share that this is the actual slide I use whenever I present the U-turn paper in Alanโs absence ๐ (not a joke)
Now continued below with case study 2: understanding performance differences of neural networks and gradient boosted trees on irregular tabular data!!
btw this is why friends dont let friends skip the โboring classical MLโ chapters in Elements of Statistical Learningโผ๏ธ
(True story: the origin of this case study is that @alanjeffares.bsky.social[big EoSL nerd] looked at the neural net eq&said โkinda looks like GBTs in EoSL Ch10โ&we went from there)
Thereโs one more case study & thoughts on the effect of design choices on function updates leftโ Iโll cover that in a final thread! (next week, giving us all a break๐
)
Until then, find the paper here arxiv.org/abs/2411.00247
and/or recap part 1 of this thread below! ๐ค 14/14
Thus in conclusion this 2nd case study showed that the telescoping approximation of a trained neural network can be a useful lens to investigate performance diffs with other methods!
Here we used it to show how some perf diffs are predicted by specific model diffs(ie diffs in implied kernels)๐ก13/n
Importantly, this growth in performance gap is tracked by the behaviour of the modelsโ kernels:
while there is no difference in kernel weights for GBTs across different input irregularity levels, the neural netโs kernel weights for the most irregular ex grow more extreme! 12/n
We test this hypothesis by varying the proportion of irregular inputs in the testset for fixed trained models.
We find that GBTs outperform NNs already in the absence of irregular ex; this speaks to diff in baseline suitability
The performance gap then indeed grows as we increase irregularity!11/n
This highlights a potential explanation why GBTs outperform neural nets on tabular data in the presence of input irregularities:
The kernels implied by the neural network might behave much much more unpredictably for test inputs different to inputs observed at train time! ๐ก๐ค10/n
Trees issue preds that are proper averages: all kernel weights are between 0 & 1. That is: trees never โextrapolateโ from the convex hull of training observations ๐ก
Neural net tangent kernels OTOH are generally unbounded and could take on very different vals for unseen test inputs!๐ฐ 9/n
One diff is obvious and purely architectural: either kernel might be able to better fit a particular underlying outcome generating process!
A second diff is a lot more subtle and relates to how regular (or: predictable) the two will likely behave on new data: โฆ 8/n
but WAIT A MINUTE โ isnโt that literally the same formula as the kernel representation of the telescoping model of a trained neural network I showed you before?? Just with a different kernel??
Surely this diff in kernel must account for at least some of the observed performance differencesโฆ ๐ค7/n
Gradient boosted trees (aka OG gradient boosting) simply implement this process using trees!
From our previous work on random forests(arxiv.org/abs/2402.01502) we know we can interpret trees as adaptive kernel smoothers, so we can rewrite the GBT preds as weighted avgs over training loss grads!6/n
Quick refresher: what is gradient boosting?
Not to be confused with other forms of boosting (eg Adaboost), *Gradient* boosting fits a sequence of weak learners that execute steepest descent in function space directly by learning to predict the loss gradients of training examples! 5/n
In arxiv.org/abs/2411.00247 we ask: why? What distinguishes gradient boosted trees from deep learning that would explain this?
A first reaction might be โthey are SO different idk where to start ๐ญโ โ BUT we show that through the telescoping lens (see part 1 of this๐งตโฌ๏ธ) things become more clear..4/n
And you know who continues to rule the tabular benchmarks? Gradient boosted trees (GBTs)!!(or their descendants)
While the severity of the perf gap over neural nets is disputed, arxiv.org/abs/2305.02997 still found as recently as last year that GBTs esp outperform when data is irregular! 3/n
First things first, why do we care about tabular?
Deep learning sometimes seems to forget we used to do data formats that werenโt text or image (๐) BUT in data science applications โ from medicine to marketing and econ โ tabular data still rules big parts of the world!!
2/n
Part 2: Why do boosted trees outperform deep learning on tabular data??
@alanjeffares.bsky.social & I suspected that answers to this are obfuscated by the 2 being considered very different algs๐ค
Instead we show they are more similar than youโd think โ making their diffs smaller but predictive!๐งต1/n
No need to leap at all, my original description even had the word "delight" in it!!
Wow, I love that! ๐
Thank you!! I donโt think I know empirical fisher actually โ do you have a ref?
It was hard to fit, but I gave it my best shot! โจgolden retriever energy in stats โจ might be historically underrepresented but I say that can and should change with us