For some reasons I couldn’t foresee, there’s been no blog post here on november 13 and november 20. So, here is the post about LSBoost announced here a few weeks ago.
First things first, what is LSBoost? Gradient boosted nonlinear penalized least squares. More precisely in LSBoost, the ensembles’ base learners are penalized, randomized neural networks.
These previous posts, with several Python and R examples, constitute a good introduction to LSBoost:
More recently, I’ve also written a more formal, short introduction to LSBoost:
The paper’s code – and more insights on LSBoost – can be found in the following Jupyter notebook:
Comments, suggestions are welcome as usual.