MEETINGQA: Extractive Question-Answering on Meeting Transcripts

9Citations
Citations of this article
17Readers
Mendeley users who have this article in their library.

Abstract

With the ubiquitous use of online meeting platforms and robust automatic speech recognition systems, meeting transcripts have emerged as a promising domain for natural language tasks. Most recent works on meeting transcripts primarily focus on summarization and extraction of action items. However, meeting discussions also have a useful question-answering (QA) component, crucial to understanding the discourse or meeting content, and can be used to build interactive interfaces on top of long transcripts. Hence, in this work, we leverage this inherent QA component of meeting discussions and introduce MEETINGQA, an extractive QA dataset comprising of questions asked by meeting participants and corresponding responses. As a result, questions can be open-ended and actively seek discussions, while the answers can be multi-span and distributed across multiple speakers. Our comprehensive empirical study of several robust baselines including long-context language models and recent instruction-tuned models reveals that models perform poorly on this task (F1 = 57.3) and severely lag behind human performance (F1 = 84.6), thus presenting a challenging new task for the community to improve upon.

Cite

CITATION STYLE

APA

Prasad, A., Bui, T., Yoon, S., Deilamsalehy, H., Dernoncourt, F., & Bansal, M. (2023). MEETINGQA: Extractive Question-Answering on Meeting Transcripts. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 15000–15025). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.acl-long.837

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free