Agents relying on large collections of interactions face the challenge of choosing an appropriate answer from such collections. Several works address this challenge by using offline learning approaches, which do not take advantage of how user-agent conversations unfold. In this work, we propose an alternative approach: incorporating user feedback at each interaction with the agent, in order to enhance its ability to choose an answer. We focus on the case of adjusting the weights of the features used by the agent to choose an answer, using an online learning algorithm (the Exponentially Weighted Average Forecaster) for that purpose. We validate our hypothesis with an experiment featuring a specific agent and simulating user feedback using a reference corpus. The results of our experiment suggest that the adjustment of the agent’s feature weights can improve its answers, provided that an appropriate reward function is designed, as this aspect is critical in the agent’s performance.
CITATION STYLE
Mendonça, V., Melo, F. S., Coheur, L., & Sardinha, A. (2017). Online learning for conversational agents. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 10423 LNAI, pp. 739–750). Springer Verlag. https://doi.org/10.1007/978-3-319-65340-2_60
Mendeley helps you to discover research relevant for your work.