KoSBi: A Dataset for Mitigating Social Bias Risks Towards Safer Large Language Model Applications

11Citations
Citations of this article
32Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Large language models (LLMs) learn not only natural text generation abilities but also social biases against different demographic groups from real-world data. This poses a critical risk when deploying LLM-based applications. Existing research and resources are not readily applicable in South Korea due to the differences in language and culture, both of which signifcantly affect the biases and targeted demographic groups. This limitation requires localized social bias datasets to ensure the safe and effective deployment of LLMs. To this end, we present KoSBi, a new social bias dataset of 34k pairs of contexts and sentences in Korean covering 72 demographic groups in 15 categories. We fnd that through fltering-based moderation, social biases in generated content can be reduced by 16.47%p on average for HyperCLOVA (30B and 82B), and GPT-3.

Cite

CITATION STYLE

APA

Lee, H., Hong, S., Park, J., Kim, T., Kim, G., & Ha, J. W. (2023). KoSBi: A Dataset for Mitigating Social Bias Risks Towards Safer Large Language Model Applications. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 5, pp. 208–224). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.acl-industry.21

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free