JASMINE: Arabic GPT Models for Few-Shot Learning

9Citations
Citations of this article
26Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Scholarship on generative pretraining (GPT) remains acutely Anglocentric, leaving serious gaps in our understanding of the whole class of autoregressive models. For example, we have little knowledge about the potential of these models and their societal impacts in diverse linguistic and cultural settings. We alleviate this issue for Arabic, a wide collection of languages and dialectal varieties with ∼ 450 million population, by introducing JASMINE. JASMINE is a suite of powerful Arabic autoregressive Transformer language models ranging in size between 300 million-6.7 billion parameters pretrained on a large and diverse dataset (∼ 235GB of text). We also carefully design and release a comprehensive benchmark for both automated and human evaluation of Arabic autoregressive models, with coverage of potential social biases, harms, and toxicity. Using our novel benchmark, we evaluate JASMINE extensively showing powerful performance intrinsically as well as in few-shot learning on a wide range of NLP tasks. We aim to responsibly release our models and evaluation benchmark with interested researchers, along with code for experimenting with them.

Cite

CITATION STYLE

APA

Nagoudi, E. M. B., Abdul-Mageed, M., Elmadany, A. R., Inciarte, A. A., & Khondaker, T. I. (2023). JASMINE: Arabic GPT Models for Few-Shot Learning. In EMNLP 2023 - 2023 Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 16721–16744). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.emnlp-main.1040

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free