Boosting nonlinear penalized least squares

[This article was first published on T. Moudiki's Webpage - Python, and kindly contributed to python-bloggers]. (You can report issue about the content on this page here)
Want to share your content on python-bloggers? click here.

For some reasons I couldn’t foresee, there’s been no blog post here on november 13
and november 20. So, here is the post about LSBoost announced here a few weeks ago.

First things first, what is LSBoost? Gradient boosted nonlinear penalized least squares. More precisely in LSBoost, the ensembles’ base learners are penalized, randomized neural networks.

These previous posts, with several Python and R examples, constitute a good introduction to LSBoost:

More recently, I’ve also written a more formal, short introduction to LSBoost:

The paper’s code – and more insights on LSBoost – can be found in the following Jupyter notebook:

Comments, suggestions are welcome as usual.


To leave a comment for the author, please follow the link and comment on their blog: T. Moudiki's Webpage - Python.

Want to share your content on python-bloggers? click here.