Do RNN States Encode Abstract Phonological Alternations?

5Citations
Citations of this article
52Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Sequence-to-sequence models have delivered impressive results in word formation tasks such as morphological inflection, often learning to model subtle morphophonological details with limited training data. Despite the performance, the opacity of neural models makes it difficult to determine whether complex generalizations are learned, or whether a kind of separate rote memorization of each morphophonological process takes place. To investigate whether complex alternations are simply memorized or whether there is some level of generalization across related sound changes in a sequence-to-sequence model, we perform several experiments on Finnish consonant gradation—a complex set of sound changes triggered in some words by certain suffixes. We find that our models often—though not always—encode 17 different consonant gradation processes in a handful of dimensions in the RNN. We also show that by scaling the activations in these dimensions we can control whether consonant gradation occurs and the direction of the gradation.

Cite

CITATION STYLE

APA

Silfverberg, M., Tyers, F., Nicolai, G., & Hulden, M. (2021). Do RNN States Encode Abstract Phonological Alternations? In NAACL-HLT 2021 - 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 5501–5513). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.naacl-main.435

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free