Arabic language is rich and complex in terms of word morphology compared to other Latin languages. Recently, natural language processing (NLP) field emerges with many researches targeting Arabic language understanding (ALU). In this context, this work presents our developed approach based on the Arabic bidirectional encoder representations from transformers (AraBERT) model where the main required steps are presented in detail. We started by the input text pre-processing, which is, then, segmented using the Farasa segmentation technique. In the next step, the AraBERT model is implemented with the pertinent parameters. The performance of our approach has been evaluated using the ARev dataset which contains more than 40,000 comments-remarks records relate to the tourism sector such as hotel reviews, restaurant reviews and others. Moreover, the obtained results are deeply compared with other relevant states of the art methods, and it shows the competitiveness of our approach that gives important results that can serve as a guide for further improvements in this field.
CITATION STYLE
El Moubtahij, H., Abdelali, H., & Tazi, E. B. (2022). AraBERT transformer model for Arabic comments and reviews analysis. IAES International Journal of Artificial Intelligence, 11(1), 379–387. https://doi.org/10.11591/ijai.v11.i1.pp379-387
Mendeley helps you to discover research relevant for your work.