The Comparison of LightGBM and XGBoost Coupling Factor Analysis and Prediagnosis of Acute Liver Failure

120Citations
Citations of this article
188Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

This paper focuses on the comparison of dimensionality reduction effect between LightGBM and XGBoost-FA. With respect to XGBoost, LightGBM can be built in the effect of dimensionality reduction via both Gradient-based One-Side Sampling(GOSS) and Exclusive Feature Bundling(EFB) algorithms, while XGBoost coupling with traditional dimensionality reduction tool Factor Analysis (XGBoost-FA) may also have dimensionality reduction effect. To present the empirical comparison, the prediagnosis dataset for the 2018 Kaggle competition Acute Liver Failure has been chosen as the research object. And pairwise comparison has been conducted among XGBoost, LightGBM, XGBoost-FA and LightGBM-FA. Concerning the test set, the vector (accuracy, log loss function, training time) of the above first four prediagnostic models are (0.75014, 0.569707, 10.5s), (0.75811, 0.576059,15.1s), (0.67786,0.663924,5.7s) and (0.67274,0.676019, 4.1s) respectively. It's been found that the training time of XGBoost-FA (external dimensionality reduction) is shorter than that of LightGBM (build-in dimensionality reduction). Considering (accuracy, training time) being (0.82, 3.1s) published on Kaggle, the algorithm (logogram as K2a) is better than the four XGBoost-FA and LightGBM in both training time and accuracy. However, K2a removes more than 50% samples with missing values and only performs binary classification. For multi-class classification or data with a large number of missing values, XGBoost-FA is more suggested if higher operational time is required, while LightGBM is preferred if higher predictive accuracy is required. With XGBoost-FA or LightGBM being employed in AI medical services, doctors are more productive in diagnosis and treatment due to much more data support and less workload. Both complement each other.

References Powered by Scopus

XGBoost: A scalable tree boosting system

33197Citations
N/AReaders
Get full text

Toward safer highways, application of XGBoost and SHAP for real-time accident detection and feature analysis

662Citations
N/AReaders
Get full text

Automated Detection of Parkinson's Disease Based on Multiple Types of Sustained Phonations Using Linear Discriminant Analysis and Genetically Optimized Neural Network

161Citations
N/AReaders
Get full text

Cited by Powered by Scopus

hyOPTXg: OPTUNA hyper-parameter optimization framework for predicting cardiovascular disease using XGBoost

119Citations
N/AReaders
Get full text

Machine Learning Techniques for Chronic Kidney Disease Risk Prediction

77Citations
N/AReaders
Get full text

Estimation of the soil arsenic concentration using a geographically weighted XGBoost model based on hyperspectral data

64Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Zhang, D., & Gong, Y. (2020). The Comparison of LightGBM and XGBoost Coupling Factor Analysis and Prediagnosis of Acute Liver Failure. IEEE Access, 8, 220990–221003. https://doi.org/10.1109/ACCESS.2020.3042848

Readers over time

‘20‘21‘22‘23‘24‘250255075100

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 22

59%

Lecturer / Post doc 8

22%

Professor / Associate Prof. 4

11%

Researcher 3

8%

Readers' Discipline

Tooltip

Computer Science 21

54%

Engineering 10

26%

Economics, Econometrics and Finance 4

10%

Mathematics 4

10%

Save time finding and organizing research with Mendeley

Sign up for free
0