lighteternal
commited on
Commit
•
9fb63b5
1
Parent(s):
19cc467
Update README.md
Browse files
README.md
CHANGED
@@ -62,7 +62,7 @@ print("\n".join([x.get("generated_text") for x in generator(
|
|
62 |
|
63 |
## Training data
|
64 |
|
65 |
-
We used a 23.
|
66 |
This is a better version of our GPT-2 small model (https://huggingface.co/lighteternal/gpt2-finetuned-greek-small)
|
67 |
|
68 |
|
|
|
62 |
|
63 |
## Training data
|
64 |
|
65 |
+
We used a 23.4GB sample from a consolidated Greek corpus from CC100, Wikimatrix, Tatoeba, Books, SETIMES and GlobalVoices containing long senquences.
|
66 |
This is a better version of our GPT-2 small model (https://huggingface.co/lighteternal/gpt2-finetuned-greek-small)
|
67 |
|
68 |
|