Measuring the Algorithmic Convergence of Randomized Ensembles: The Regression Setting

4Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.
Get full text

Abstract

When randomized ensemble methods such as bagging and random forests are implemented, a basic question arises: Is the ensemble large enough? In particular, the practitioner desires a rigorous guarantee that a given ensemble will perform nearly as well as an ideal infinite ensemble (trained on the same data). The purpose of the current paper is to develop a bootstrap method for solving this problem in the context of regression—which complements our companion paper in the context of classification [Lopes, Ann. Statist., 47 (2019), 1088–1112]. In contrast to the classification setting, the current paper shows that theoretical guarantees for the proposed bootstrap can be established under much weaker assumptions. In addition, we illustrate the flexibility of the method by showing how it can be adapted to measure algorithmic convergence for variable selection. Lastly, we provide numerical results demonstrating that the method works well in a range of situations.

Cite

CITATION STYLE

APA

Lopes, M. E., Wu, S., & Lee, T. C. M. (2020). Measuring the Algorithmic Convergence of Randomized Ensembles: The Regression Setting. SIAM Journal on Mathematics of Data Science, 2(4), 921–943. https://doi.org/10.1137/20M1343300

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free