Conversational query production is an emerging fundamental task for the dialogue system, where search queries are generated to explore the vast and continually updating knowledge from a search engine. To accelerate this line of research, previous studies have released several datasets with human-annotated search queries. However, the limited annotations still can not cover conversations of various domains. To solve this challenge, we propose a novel domain adaptation framework. It is inspired by a weakly supervised learning algorithm from previous work (Wang et al., 2023b) that guides a model using reinforcement learning with BM25 scores as feedback. Though effective, it is fragile facing noisy content on webpages from a commercial search engine and variance in conversations because of ignoring deep semantic information of dialogue contexts. Thus, we improve the algorithm by taking the advance of retrieval-augmented generation (RAG) and exploring several practical techniques such as knowledge distillation for stable training. We conduct experiments in multiple settings across different languages. Guided by the RAG model feedback, our model is more robust and performs significantly better especially in a more challenging setting over strong baselines.
CITATION STYLE
Wang, A., Song, L., Xu, G., & Su, J. (2023). Domain Adaptation for Conversational Query Production with the RAG Model Feedback. In Findings of the Association for Computational Linguistics: EMNLP 2023 (pp. 9129–9141). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-emnlp.612
Mendeley helps you to discover research relevant for your work.