Gradient exceptionality in maximum entropy grammar with lexically specific constraints

22Citations
Citations of this article
22Readers
Mendeley users who have this article in their library.

Abstract

The number of exceptions to a phonological generalization appears to gradiently affect its productivity. Generalizations with relatively few exceptions are relatively productive, as measured in tendencies to regularization, as well as in nonce word productions and other psycholinguistic tasks. Gradient productivity has been previously modeled with probabilistic grammars, including Maximum Entropy Grammar, but they often fail to capture the fixed pronunciations of the existing words in a language, as opposed to nonce words. Lexically specific constraints allow existing words to be produced faithfully, while permitting variation in novel words that are not subject to those constraints. When each word has its own lexically specific version of a constraint, an inverse correlation between the number of exceptions and the degree of productivity is straightforwardly predicted.

Cite

CITATION STYLE

APA

Moore-Cantwell, C., & Pater, J. (2016). Gradient exceptionality in maximum entropy grammar with lexically specific constraints. Catalan Journal of Linguistics, 15, 53–66. https://doi.org/10.5565/rev/catjl.183

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free