Abstract
Critical studies found NLP systems to bias based on gender and racial identities. However, few studies focused on identities defined by cultural factors like religion and nationality. Compared to English, such research efforts are even further limited in major languages like Bengali due to the unavailability of labeled datasets. This paper describes a process for developing a bias evaluation dataset highlighting cultural influences on identity. We also provide a Bengali dataset as an artifact outcome that can contribute to future critical research.
Cite
CITATION STYLE
Das, D., Guha, S., & Semaan, B. (2023). Toward Cultural Bias Evaluation Datasets: The Case of Bengali Gender, Religious, and National Identity. In Cross-Cultural Considerations in NLP at EACL, C3NLP 2023 - Proceedings of the Workshop (pp. 68–83). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.c3nlp-1.8
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.