Unsupervised FAQ retrieval with question generation and BERT

43Citations
Citations of this article
167Readers
Mendeley users who have this article in their library.

Abstract

We focus on the task of Frequently Asked Questions (FAQ) retrieval. A given user query can be matched against the questions and/or the answers in the FAQ. We present a fully unsupervised method that exploits the FAQ pairs to train two BERT models. The two models match user queries to FAQ answers and questions, respectively. We alleviate the missing labeled data of the latter by automatically generating high-quality question paraphrases. We show that our model is on par and even outperforms supervised models on existing datasets.

Cite

CITATION STYLE

APA

Mass, Y., Carmeli, B., Roitman, H., & Konopnicki, D. (2020). Unsupervised FAQ retrieval with question generation and BERT. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 807–812). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.acl-main.74

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free