State Gradients for RNN Memory Analysis

2Citations
Citations of this article
88Readers
Mendeley users who have this article in their library.

Abstract

We present a framework for analyzing what the state in RNNs remembers from its input embeddings. We compute the gradients of the states with respect to the input embeddings and decompose the gradient matrix with Singular Value Decomposition to analyze which directions in the embedding space are best transferred to the hidden state space, characterized by the largest singular values. We apply our approach to LSTM language models and investigate to what extent and for how long certain classes of words are remembered on average for a certain corpus. Additionally, the extent to which a specific property or relationship is remembered by the RNN can be tracked by comparing a vector characterizing that property with the direction(s) in embedding space that are best preserved in hidden state space.

Cite

CITATION STYLE

APA

Verwimp, L., van Hamme, H., Renkens, V., & Wambacq, P. (2018). State Gradients for RNN Memory Analysis. In EMNLP 2018 - 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, Proceedings of the 1st Workshop (pp. 344–346). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/w18-5443

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free