Optimizing Small BERTs Trained for German NER

Currently, the most widespread neural network architecture for training language models is the so-called BERT, which led to improvements in various Natural Language Processing (NLP) tasks. In general, the larger the number of parameters in a BERT model, the better the results obtained in these NLP t...

Descripción completa

Guardado en:
Detalles Bibliográficos
Autores principales: Jochen Zöllner, Konrad Sperfeld, Christoph Wick, Roger Labahn
Formato: article
Lenguaje:EN
Publicado: MDPI AG 2021
Materias:
Acceso en línea:https://doaj.org/article/c269d264e4014373a18286098ba93af0
Etiquetas: Agregar Etiqueta
Sin Etiquetas, Sea el primero en etiquetar este registro!