HamzaNaser
commited on
Commit
•
e1ef5d7
1
Parent(s):
6474150
Update README.md
Browse files
README.md
CHANGED
@@ -22,7 +22,7 @@ model-index:
|
|
22 |
|
23 |
|
24 |
# Dialects-to-MSA-Transformer overview
|
25 |
-
This Model is optimized to convert written text in various non Standard Classical Arabic into Classic Arabic, the model was Fine-Tuned on 0.8M pairs of sentence generated by OpenAI API gpt-4o-mini Text Generation Model, beside being able to convert Dialects into Classical Arabic, the model can also be used in other NLP tasks such as Text Correction, Diacretization and
|
26 |
|
27 |
|
28 |
|
@@ -85,7 +85,7 @@ Inspecting large paris of texts might be tedious, thus we have taken a sample of
|
|
85 |
| Data Set Size | GPU Device | Epochs | Training Time | Blue Score |
|
86 |
|:-------------:|:----------:|:----------:|:---------------:|:------------:|
|
87 |
| 0.8M | A100 | 3 | 7.7Hrs | 46.9 |
|
88 |
-
|
|
89 |
|
90 |
## Costs and Resources
|
91 |
There are two main computing resources when Dialects to MSA Transformer were built, one is the generation of MSA sequences using GPT model, the second resource is the GPU used to train and adjust the parameters of the pretrained Model.
|
|
|
22 |
|
23 |
|
24 |
# Dialects-to-MSA-Transformer overview
|
25 |
+
This Model is optimized to convert written text in various non Standard Classical Arabic into Classic Arabic, the model was Fine-Tuned on 0.8M pairs of sentence generated by OpenAI API gpt-4o-mini Text Generation Model, beside being able to convert Dialects into Classical Arabic, the model can also be used in other NLP tasks such as Text Correction, Diacretization, Sentence Punctuation and Machine Translation.
|
26 |
|
27 |
|
28 |
|
|
|
85 |
| Data Set Size | GPU Device | Epochs | Training Time | Blue Score |
|
86 |
|:-------------:|:----------:|:----------:|:---------------:|:------------:|
|
87 |
| 0.8M | A100 | 3 | 7.7Hrs | 46.9 |
|
88 |
+
| 2.6M | A100 | 1 | ??? | ??? |
|
89 |
|
90 |
## Costs and Resources
|
91 |
There are two main computing resources when Dialects to MSA Transformer were built, one is the generation of MSA sequences using GPT model, the second resource is the GPU used to train and adjust the parameters of the pretrained Model.
|