August 2024 A large-sample theory for infinitesimal gradient boosting
Clément Dombry, Jean-Jil Duchamps
Author Affiliations +
Bernoulli 30(3): 1894-1920 (August 2024). DOI: 10.3150/23-BEJ1657

Abstract

Infinitesimal gradient boosting (Dombry and Duchamps (2021)) is defined as the vanishing-learning-rate limit of the popular tree-based gradient boosting algorithm from machine learning. It is characterized as the solution of a nonlinear ordinary differential equation in an infinite-dimensional function space where the infinitesimal boosting operator driving the dynamics depends on the training sample. We consider the asymptotic behavior of the model in the large sample limit and prove its convergence to a deterministic process. This population limit is again characterized by a differential equation that depends on the population distribution. We explore some properties of this population limit: we prove that the dynamics makes the test error decrease and we consider its long time behavior.

Citation

Download Citation

Clément Dombry. Jean-Jil Duchamps. "A large-sample theory for infinitesimal gradient boosting." Bernoulli 30 (3) 1894 - 1920, August 2024. https://doi.org/10.3150/23-BEJ1657

Information

Received: 1 September 2022; Published: August 2024
First available in Project Euclid: 14 May 2024

Digital Object Identifier: 10.3150/23-BEJ1657

Keywords: gradient boosting , large sample theory , softmax gradient tree

JOURNAL ARTICLE
27 PAGES

This article is only available to subscribers.
It is not available for individual sale.
+ SAVE TO MY LIBRARY

Vol.30 • No. 3 • August 2024
Back to Top