We reformulate the problem of encoding a multi-scale representation of a sequence in a language model by casting it in a continuous learning framework. We propose a hierarchical multi-scale language model in which short time-scale dependencies are encoded in the hidden state of a lower-level recurrent neural network while longer time-scale dependencies are encoded in the dynamic of the lower-level network by having a meta-learner update the weights of the lower-level neural network in an online meta-learning fashion. We use elastic weights consolidation as a higher-level to prevent catastrophic forgetting in our continuous learning framework.
CITATION STYLE
Wolf, T., Chaumond, J., & Delangue, C. (2018). Continuous learning in a hierarchical multiscale neural network. In ACL 2018 - 56th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers) (Vol. 2, pp. 1–7). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/p18-2001
Mendeley helps you to discover research relevant for your work.