Update README.md
Browse files
README.md
CHANGED
@@ -54,7 +54,15 @@ SalamandraTA-2B is a machine translation model that has been continually pre-tra
|
|
54 |
|
55 |
### Description
|
56 |
|
57 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
58 |
|
59 |
### Hyperparameters
|
60 |
|
|
|
54 |
|
55 |
### Description
|
56 |
|
57 |
+
This machine translation model is built upon the foundation of Salamandra 2B. By leveraging the knowledge of the base Salamandra 2B model, this model is able to perform high quality translations between **almost 900 translation directions**.
|
58 |
+
|
59 |
+
Key Features:
|
60 |
+
|
61 |
+
* **Continual Pretraining:** The model is trained on 70 Billion tokens of parallel data. All data employed is open-sourced or generated from open-source data using the Machine Translation models at [BSC](https://huggingface.co/collections/projecte-aina/mt-models-655e154668c6dd132159081c)
|
62 |
+
* **Large Language Model Foundation:** Built on Salamandra 2B, providing a strong language understanding and generation capability.
|
63 |
+
* **Multilingual Support:** Capable of translating between 30 european languages, including low-resource languages.
|
64 |
+
* **High-Quality Translations:** Delivers accurate and fluent translations, thanks to its continual pretraining and large-scale dataset.
|
65 |
+
* **Efficient Inference:** 2 Billion parameters allow for a trade-off between performance and hardware requirements by most systems.
|
66 |
|
67 |
### Hyperparameters
|
68 |
|