Accelerating the XGBoost algorithm using GPU computing

207Citations
Citations of this article
236Readers
Mendeley users who have this article in their library.

Abstract

We present a CUDA-based implementation of a decision tree construction algorithm within the gradient boosting library XGBoost. The tree construction algorithm is executed entirely on the graphics processing unit (GPU) and shows high performance with a variety of datasets and settings, including sparse input matrices. Individual boosting iterations are parallelised, combining two approaches. An interleaved approach is used for shallow trees, switching to a more conventional radix sort-based approach for larger depths. We show speedups of between 3 × and 6 × using a Titan X compared to a 4 core i7 CPU, and 1.2 × using a Titan X compared to 2 × Xeon CPUs (24 cores).We show that it is possible to process the Higgs dataset (10 million instances, 28 features) entirely within GPU memory. The algorithm is made available as a plug-in within the XGBoost library and fully supports all XGBoost features including classification, regression and ranking tasks.

Cite

CITATION STYLE

APA

Mitchell, R., & Frank, E. (2017). Accelerating the XGBoost algorithm using GPU computing. PeerJ Computer Science, 2017(7). https://doi.org/10.7717/peerj-cs.127

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free