Can we trust the evaluation on ChatGPT?

33Citations
Citations of this article
83Readers
Mendeley users who have this article in their library.
Get full text

Abstract

ChatGPT, the first large language model with mass adoption, has demonstrated remarkable performance in numerous natural language tasks. Despite its evident usefulness, evaluating ChatGPT's performance in diverse problem domains remains challenging due to the closed nature of the model and its continuous updates via Reinforcement Learning from Human Feedback (RLHF). We highlight the issue of data contamination in ChatGPT evaluations, with a case study in stance detection. We discuss the challenge of preventing data contamination and ensuring fair model evaluation in the age of closed and continuously trained models.

Cite

CITATION STYLE

APA

Aiyappa, R., An, J., Kwak, H., & Ahn, Y. Y. (2023). Can we trust the evaluation on ChatGPT? In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 47–54). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.trustnlp-1.5

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free