Gradient Boosting Machines: New Insights, Algorithms, and Improved Complexity

Speaker Haihao Lu, MIT
Date: 2/15/2019
Time: 10 a.m.

103 Transportation Building


ISE Grad Seminar

Event Type: Seminar/Symposium


The gradient boosting machine (GBM) is one of the most successful supervised learning algorithms, and it has been the dominant method in many data science competitions, including Kaggle and KDDCup. In spite of its practical success, there has been a huge gap between practice and theoretical understanding.  In this line of research, we show that GBM can be interpreted as a greedy coordinate descent method in the coefficient space and/or a mirror descent method in the “pseudo-residual” space.  Armed with this structural insight, we develop two new algorithms for classification in the context of GBM:  (i) the Random-then-Greedy Gradient Boosting Machine (RtGBM), which lowers the cost per iteration and achieves improved performance in theory as well as practice; and (ii) the Accelerated Gradient Boosting Machine (AGBM), which achieves the computational efficiency of acceleration schemes in general, again both in theory and in practice.  These two algorithms are currently being incorporated by Google into their TensorFlow Boosted Trees software.  The talk is based on joint works with Rahul Mazumder (MIT), Sai Praneeth Karimireddy (EPFL), Natalia Ponomareva (Google) and Vahab Mirrokni (Google).

To request disability-related accommodations for this event, please contact the person listed above, or the unit hosting the event.