carlosep93 commited on
Commit
d76d2e0
·
verified ·
1 Parent(s): 86b7da7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -1
README.md CHANGED
@@ -54,7 +54,15 @@ SalamandraTA-2B is a machine translation model that has been continually pre-tra
54
 
55
  ### Description
56
 
57
- Continual pre-trained model from Salamandra-2B on 70 billion tokens of highly curated parallel data.
 
 
 
 
 
 
 
 
58
 
59
  ### Hyperparameters
60
 
 
54
 
55
  ### Description
56
 
57
+ This machine translation model is built upon the foundation of Salamandra 2B. By leveraging the knowledge of the base Salamandra 2B model, this model is able to perform high quality translations between **almost 900 translation directions**.
58
+
59
+ Key Features:
60
+
61
+ * **Continual Pretraining:** The model is trained on 70 Billion tokens of parallel data. All data employed is open-sourced or generated from open-source data using the Machine Translation models at [BSC](https://huggingface.co/collections/projecte-aina/mt-models-655e154668c6dd132159081c)
62
+ * **Large Language Model Foundation:** Built on Salamandra 2B, providing a strong language understanding and generation capability.
63
+ * **Multilingual Support:** Capable of translating between 30 european languages, including low-resource languages.
64
+ * **High-Quality Translations:** Delivers accurate and fluent translations, thanks to its continual pretraining and large-scale dataset.
65
+ * **Efficient Inference:** 2 Billion parameters allow for a trade-off between performance and hardware requirements by most systems.
66
 
67
  ### Hyperparameters
68