Domain Adaptation for Conversational Query Production with the RAG Model Feedback

3Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.

Abstract

Conversational query production is an emerging fundamental task for the dialogue system, where search queries are generated to explore the vast and continually updating knowledge from a search engine. To accelerate this line of research, previous studies have released several datasets with human-annotated search queries. However, the limited annotations still can not cover conversations of various domains. To solve this challenge, we propose a novel domain adaptation framework. It is inspired by a weakly supervised learning algorithm from previous work (Wang et al., 2023b) that guides a model using reinforcement learning with BM25 scores as feedback. Though effective, it is fragile facing noisy content on webpages from a commercial search engine and variance in conversations because of ignoring deep semantic information of dialogue contexts. Thus, we improve the algorithm by taking the advance of retrieval-augmented generation (RAG) and exploring several practical techniques such as knowledge distillation for stable training. We conduct experiments in multiple settings across different languages. Guided by the RAG model feedback, our model is more robust and performs significantly better especially in a more challenging setting over strong baselines.

Cite

CITATION STYLE

APA

Wang, A., Song, L., Xu, G., & Su, J. (2023). Domain Adaptation for Conversational Query Production with the RAG Model Feedback. In Findings of the Association for Computational Linguistics: EMNLP 2023 (pp. 9129–9141). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-emnlp.612

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free