With the ubiquitous use of online meeting platforms and robust automatic speech recognition systems, meeting transcripts have emerged as a promising domain for natural language tasks. Most recent works on meeting transcripts primarily focus on summarization and extraction of action items. However, meeting discussions also have a useful question-answering (QA) component, crucial to understanding the discourse or meeting content, and can be used to build interactive interfaces on top of long transcripts. Hence, in this work, we leverage this inherent QA component of meeting discussions and introduce MEETINGQA, an extractive QA dataset comprising of questions asked by meeting participants and corresponding responses. As a result, questions can be open-ended and actively seek discussions, while the answers can be multi-span and distributed across multiple speakers. Our comprehensive empirical study of several robust baselines including long-context language models and recent instruction-tuned models reveals that models perform poorly on this task (F1 = 57.3) and severely lag behind human performance (F1 = 84.6), thus presenting a challenging new task for the community to improve upon.
CITATION STYLE
Prasad, A., Bui, T., Yoon, S., Deilamsalehy, H., Dernoncourt, F., & Bansal, M. (2023). MEETINGQA: Extractive Question-Answering on Meeting Transcripts. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 15000–15025). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.acl-long.837
Mendeley helps you to discover research relevant for your work.