Gradient Exceptionality in Maximum Entropy Grammar with Lexically Specific Constraints

The number of exceptions to a phonological generalization appears to gradiently affect its productivity. Generalizations with relatively few exceptions are relatively productive, as measured in tendencies to regularization, as well as in nonce word productions and other psycholinguistic tasks. Gradi...

Descripción completa

Guardado en:
Detalles Bibliográficos
Autores principales: Claire Moore-Cantwell, Joe Pater
Formato: article
Lenguaje:CA
EN
Publicado: Universitat Autònoma de Barcelona 2016
Materias:
Acceso en línea:https://doaj.org/article/574cc5847bd44ae0be3ea4dd4c98b52f
Etiquetas: Agregar Etiqueta
Sin Etiquetas, Sea el primero en etiquetar este registro!
Descripción
Sumario:The number of exceptions to a phonological generalization appears to gradiently affect its productivity. Generalizations with relatively few exceptions are relatively productive, as measured in tendencies to regularization, as well as in nonce word productions and other psycholinguistic tasks. Gradient productivity has been previously modeled with probabilistic grammars, including Maximum Entropy Grammar, but they often fail to capture the fixed pronunciations of the existing words in a language, as opposed to nonce words. Lexically specific constraints allow existing words to be produced faithfully, while permitting variation in novel words that are not subject to those constraints. When each word has its own lexically specific version of a constraint, an inverse correlation between the number of exceptions and the degree of productivity is straightforwardly predicted.