Retrieving relevant answers from heterogeneous data formats, for given for questions, is a challenging problem. The process of pinpointing relevant information suitable to answer a question is further compounded in large document collections containing documents of substantial length. This paper presents the models designed as part of our submission to the DialDoc21 Shared Task (Document-grounded Dialogue and Conversational Question Answering) for span prediction in question answering. The proposed models leverage the superior predictive power of pretrained transformer models like RoBERTa, ALBERT and ELECTRA, to identify the most relevant information in an associated passage for the next agent turn. To further enhance the performance, the models were fine-tuned on different span selection based question answering datasets like SQuAD2.0 and Natural Questions (NQ) corpus. We also explored ensemble techniques for combining multiple models to achieve enhanced performance for the task. Our team SB NITK ranked 6th on the leaderboard for the Knowledge Identification task, and our best ensemble model achieved an Exact score of 58.58 and an F1 score of 73.39.
CITATION STYLE
Bachina, S., Balumuri, S., & Sowmya Kamath, S. (2021). Ensemble ALBERT and RoBERTa for Span Prediction in Question Answering. In DialDoc 2021 - 1st Workshop on Document-Grounded Dialogue and Conversational Question Answering, Proceedings (pp. 63–68). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.dialdoc-1.9
Mendeley helps you to discover research relevant for your work.