HamzaNaser commited on
Commit
e1ef5d7
1 Parent(s): 6474150

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -22,7 +22,7 @@ model-index:
22
 
23
 
24
  # Dialects-to-MSA-Transformer overview
25
- This Model is optimized to convert written text in various non Standard Classical Arabic into Classic Arabic, the model was Fine-Tuned on 0.8M pairs of sentence generated by OpenAI API gpt-4o-mini Text Generation Model, beside being able to convert Dialects into Classical Arabic, the model can also be used in other NLP tasks such as Text Correction, Diacretization and Sentence Punctuation.
26
 
27
 
28
 
@@ -85,7 +85,7 @@ Inspecting large paris of texts might be tedious, thus we have taken a sample of
85
  | Data Set Size | GPU Device | Epochs | Training Time | Blue Score |
86
  |:-------------:|:----------:|:----------:|:---------------:|:------------:|
87
  | 0.8M | A100 | 3 | 7.7Hrs | 46.9 |
88
- | 3.0M | A100 | 1 | ??? | ??? |
89
 
90
  ## Costs and Resources
91
  There are two main computing resources when Dialects to MSA Transformer were built, one is the generation of MSA sequences using GPT model, the second resource is the GPU used to train and adjust the parameters of the pretrained Model.
 
22
 
23
 
24
  # Dialects-to-MSA-Transformer overview
25
+ This Model is optimized to convert written text in various non Standard Classical Arabic into Classic Arabic, the model was Fine-Tuned on 0.8M pairs of sentence generated by OpenAI API gpt-4o-mini Text Generation Model, beside being able to convert Dialects into Classical Arabic, the model can also be used in other NLP tasks such as Text Correction, Diacretization, Sentence Punctuation and Machine Translation.
26
 
27
 
28
 
 
85
  | Data Set Size | GPU Device | Epochs | Training Time | Blue Score |
86
  |:-------------:|:----------:|:----------:|:---------------:|:------------:|
87
  | 0.8M | A100 | 3 | 7.7Hrs | 46.9 |
88
+ | 2.6M | A100 | 1 | ??? | ??? |
89
 
90
  ## Costs and Resources
91
  There are two main computing resources when Dialects to MSA Transformer were built, one is the generation of MSA sequences using GPT model, the second resource is the GPU used to train and adjust the parameters of the pretrained Model.