Gradient Exceptionality in Maximum Entropy Grammar with Lexically Specific Constraints
The number of exceptions to a phonological generalization appears to gradiently affect its productivity. Generalizations with relatively few exceptions are relatively productive, as measured in tendencies to regularization, as well as in nonce word productions and other psycholinguistic tasks. Gradi...
Guardado en:
Autores principales: | , |
---|---|
Formato: | article |
Lenguaje: | CA EN |
Publicado: |
Universitat Autònoma de Barcelona
2016
|
Materias: | |
Acceso en línea: | https://doaj.org/article/574cc5847bd44ae0be3ea4dd4c98b52f |
Etiquetas: |
Agregar Etiqueta
Sin Etiquetas, Sea el primero en etiquetar este registro!
|
Sumario: | The number of exceptions to a phonological generalization appears to gradiently affect its productivity. Generalizations with relatively few exceptions are relatively productive, as measured in tendencies to regularization, as well as in nonce word productions and other psycholinguistic tasks. Gradient productivity has been previously modeled with probabilistic grammars, including Maximum Entropy Grammar, but they often fail to capture the fixed pronunciations of the existing words in a language, as opposed to nonce words. Lexically specific constraints allow existing words to be produced faithfully, while permitting variation in novel words that are not subject to those constraints. When each word has its own lexically specific version of a constraint, an inverse correlation between the number of exceptions and the degree of productivity is straightforwardly predicted. |
---|