Continuous learning in a hierarchical multiscale neural network

3Citations
Citations of this article
98Readers
Mendeley users who have this article in their library.

Abstract

We reformulate the problem of encoding a multi-scale representation of a sequence in a language model by casting it in a continuous learning framework. We propose a hierarchical multi-scale language model in which short time-scale dependencies are encoded in the hidden state of a lower-level recurrent neural network while longer time-scale dependencies are encoded in the dynamic of the lower-level network by having a meta-learner update the weights of the lower-level neural network in an online meta-learning fashion. We use elastic weights consolidation as a higher-level to prevent catastrophic forgetting in our continuous learning framework.

Cite

CITATION STYLE

APA

Wolf, T., Chaumond, J., & Delangue, C. (2018). Continuous learning in a hierarchical multiscale neural network. In ACL 2018 - 56th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers) (Vol. 2, pp. 1–7). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/p18-2001

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free