Commit
·
762a413
1
Parent(s):
45aff1f
Update README.md
Browse files
README.md
CHANGED
@@ -15,14 +15,14 @@ model-index:
|
|
15 |
|
16 |
# Danish Legal LM
|
17 |
|
18 |
-
This model is pre-training on a combination of the Danish part of the MultiEURLEX (Chalkidis et al., 2021) dataset comprising EU legislation and two subsets (`retsinformationdk`, `retspraksis`) of the Danish Gigaword Corpus (Derczynski et al., 2021)
|
19 |
It achieves the following results on the evaluation set:
|
20 |
- Loss: 0.7302 (up to 128 tokens)
|
21 |
- Loss: 0.7847 (up to 512 tokens)
|
22 |
|
23 |
## Model description
|
24 |
|
25 |
-
This is a RoBERTa (Liu et al., 2019) model pre-
|
26 |
|
27 |
## Intended uses & limitations
|
28 |
|
|
|
15 |
|
16 |
# Danish Legal LM
|
17 |
|
18 |
+
This model is pre-training on a combination of the Danish part of the MultiEURLEX (Chalkidis et al., 2021) dataset comprising EU legislation and two subsets (`retsinformationdk`, `retspraksis`) of the Danish Gigaword Corpus (Derczynski et al., 2021) comprising legal proceedings.
|
19 |
It achieves the following results on the evaluation set:
|
20 |
- Loss: 0.7302 (up to 128 tokens)
|
21 |
- Loss: 0.7847 (up to 512 tokens)
|
22 |
|
23 |
## Model description
|
24 |
|
25 |
+
This is a RoBERTa (Liu et al., 2019) model pre-trained on Danish legal corpora. It follows a base configuration with 12 Transformer layers, each one with 768 hidden units and 12 attention heads.
|
26 |
|
27 |
## Intended uses & limitations
|
28 |
|