NB-BERT
Collection
Models based on BERT from Google, and trained on data from various sources, including the digital collection at the National Library of Norway.
•
4 items
•
Updated
NB-BERT-large is a general BERT-large model built on the large digital collection at the National Library of Norway.
This model is trained from scratch on a wide variety of Norwegian text (both bokmål and nynorsk) from the last 200 years using a monolingual Norwegian vocabulary.
The 1.0 version of the model is general, and should be fine-tuned for any particular use. Some fine-tuning sets may be found on Github, see
The model is trained on a wide variety of text. The training set is described on
For more information on the model, see