The Best Techniques to Deal with Unbalanced Sequential Text Data in Deep Learning

0Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.

Abstract

Datasets with a balanced distribution of data are often difficult to find in real life. Although various methods have been developed and proven successful using shallow learning algorithms, handling unbalanced classes using a deep learning approach is still limited. Most of these studies only focus on image data using the Convolution Neural Network (CNN) architecture. In this study, we tried to apply several class handling techniques to three datasets of unbalanced text data. Both use a data-level approach with resampling techniques on word vectors and algorithm-level using Weighted Cross-Entropy Loss (WCEL) to handle cases of imbalanced text classification. With Bidirectional Long-Short Term Memory (BiLSTM) architecture. We tested each method using three datasets with different characteristics and levels of imbalance. Based on the experiments that have been carried out, each technique applied has a different performance on each dataset.

Cite

CITATION STYLE

APA

Adi, S., Hikmah, A., Sari, B. W., Sunyoto, A., Yaqin, A., & Hayaty, M. (2022). The Best Techniques to Deal with Unbalanced Sequential Text Data in Deep Learning. International Journal of Advanced Computer Science and Applications, 13(11), 664–669. https://doi.org/10.14569/IJACSA.2022.0131177

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free