Optimizing Small BERTs Trained for German NER
Currently, the most widespread neural network architecture for training language models is the so-called BERT, which led to improvements in various Natural Language Processing (NLP) tasks. In general, the larger the number of parameters in a BERT model, the better the results obtained in these NLP t...
Saved in:
| Main Authors: | , , , |
|---|---|
| Format: | article |
| Language: | EN |
| Published: |
MDPI AG
2021
|
| Subjects: | |
| Online Access: | https://doaj.org/article/c269d264e4014373a18286098ba93af0 |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|