Update README.md
Browse files
README.md
CHANGED
@@ -54,7 +54,9 @@ This model is case-sensitive: it makes a difference between English and English.
|
|
54 |
The model has 6 layers, 768 dimension and 12 heads, totalizing 82M parameters (compared to 125M parameters for RoBERTa-base).
|
55 |
On average DistilRoBERTa is twice as fast as Roberta-base.
|
56 |
|
57 |
-
## Training
|
|
|
|
|
58 |
|
59 |
## Training procedure
|
60 |
|
|
|
54 |
The model has 6 layers, 768 dimension and 12 heads, totalizing 82M parameters (compared to 125M parameters for RoBERTa-base).
|
55 |
On average DistilRoBERTa is twice as fast as Roberta-base.
|
56 |
|
57 |
+
## Training Data
|
58 |
+
|
59 |
+
Polar sentiment dataset of sentences from financial news. The dataset consists of 4840 sentences from English language financial news categorised by sentiment. The dataset is divided by agreement rate of 5-8 annotators.
|
60 |
|
61 |
## Training procedure
|
62 |
|