language: no | |
license: cc-by-4.0 | |
tags: | |
- norwegian | |
- bert | |
thumbnail: nblogo_3.png | |
pipeline_tag: fill-mask | |
widget: | |
- text: På biblioteket kan du låne en [MASK]. | |
- **Release 1.0beta** (April 29, 2021) | |
# NB-BERT-large (beta) | |
## Description | |
NB-BERT-large is a general BERT-large model built on the large digital collection at the National Library of Norway. | |
This model is trained from scratch on a wide variety of Norwegian text (both bokmål and nynorsk) from the last 200 years using a monolingual Norwegian vocabulary. | |
## Intended use & limitations | |
The 1.0 version of the model is general, and should be fine-tuned for any particular use. Some fine-tuning sets may be found on Github, see | |
* https://github.com/NBAiLab/notram | |
## Training data | |
The model is trained on a wide variety of text. The training set is described on | |
* https://github.com/NBAiLab/notram | |
## More information | |
For more information on the model, see | |
https://github.com/NBAiLab/notram |