ThunderGBM: Fast GBDTs and Random Forests on GPUs
School of Computing, National University of Singapore, 117418, Singapore
National University of Singapore, 2019
@article{wen2019thundergbm,
title={ThunderGBM: Fast GBDTs and Random Forests on GPUs},
author={Wen, Zeyi and Shi, Jiashuai and He, Bingsheng and Li, Qinbin and Chen, Jian},
year={2019}
}
Gradient Boosting Decision Trees (GBDTs) and Random Forests (RFs) have been used in many real-world applications. They are often a standard recipe for building state-of-the-art solutions to machine learning and data mining problems. However, training and prediction are very expensive computationally for large and high dimensional problems. This article presents an efficient and open source software toolkit called ThunderGBM which exploits the high-performance Graphics Processing Units (GPUs) for GBDTs and RFs. ThunderGBM supports classification, regression and ranking. It uses identical command line options and configuration files as XGBoost-one of the most popular GBDT and RF libraries. ThunderGBM can be used through multiple language interfaces including C/C++ and Python, and can run on single or multiple GPUs of a machine. Our experimental results show that ThunderGBM outperforms the existing libraries while producing similar models, and can handle high dimensional problems which existing GPU based libraries fail. Documentation, examples, and more details about ThunderGBM are available at https://github.com/xtra-computing/thundergbm.
February 3, 2019 by hgpu