Recent progress in Large Language Models (LLMs) has produced models that exhibit remarkable performance across a variety of NLP tasks. However, it remains unclear whether the existing focus of NLP research accurately captures the genuine requirements of human users. This paper provides a comprehensive analysis of the divergence between current NLP research and the needs of real-world NLP applications via a large-scale collection of user-GPT conversations. We analyze a large-scale collection of real user queries to GPT. We compare these queries against existing NLP benchmark tasks and identify a significant gap between the tasks that users frequently request from LLMs and the tasks that are commonly studied in academic research. For example, we find that tasks such as “design” and “planning” are prevalent in user interactions but are largely neglected or different from traditional NLP benchmarks. We investigate these overlooked tasks, dissect the practical challenges they pose, and provide insights toward a roadmap to make LLMs better aligned with user needs.
CITATION STYLE
Ouyang, S., Wang, S., Liu, Y., Zhong, M., Jiao, Y., Iter, D., … Han, J. (2023). The Shifted and The Overlooked: A Task-oriented Investigation of User-GPT Interactions. In EMNLP 2023 - 2023 Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 2375–2393). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.emnlp-main.146
Mendeley helps you to discover research relevant for your work.