Common Sense Knowledge in Large Scale Neural Conversational Models

1Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.
Get full text

Abstract

It was recently shown, that neural language models, trained on large scale conversational corpus such as OpenSubtitles have recently demonstrated ability to simulate conversation and answer questions, that require common-sense knowledge, suggesting the possibility that such networks actually learn a way to represent and use common-sense knowledge, extracted from dialog corpus. If this is really true, the possibility exists of using large scale conversational models for use in information retrieval (IR) tasks, including question answering, document retrieval and other problems that require measuring of semantic similarity. In this work we analyze behavior of a number of neural network architectures, trained on Russian conversations corpus, containing 20 million dialog turns. We found that small to medium neural networks do not really learn any noticeable common-sense knowledge, operating pure on the level of syntactic features, while large very deep networks shows do posses some common-sense knowledge.

Cite

CITATION STYLE

APA

Tarasov, D. S., & Izotova, E. D. (2018). Common Sense Knowledge in Large Scale Neural Conversational Models. In Studies in Computational Intelligence (Vol. 736, pp. 39–44). Springer Verlag. https://doi.org/10.1007/978-3-319-66604-4_6

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free