xixianliao
commited on
Commit
•
e136825
1
Parent(s):
7fee9d1
Upload model
Browse files
README.md
CHANGED
@@ -32,7 +32,7 @@ It is trained on a combination of Catalan-Chinese datasets
|
|
32 |
totalling 94.187.858 sentence pairs. 113.305 sentence pairs were parallel data collected from the web, while the remaining 94.074.553 sentence pairs
|
33 |
were parallel synthetic data created using the
|
34 |
[Aina Project's Spanish-Catalan machine translation model](https://huggingface.co/projecte-aina/aina-translator-es-ca) and the [Aina Project's English-Catalan machine translation model](https://huggingface.co/projecte-aina/aina-translator-en-ca).
|
35 |
-
The model was evaluated on the Flores, NTREX, and Projecte Aina's Catalan-Chinese evaluation datasets.
|
36 |
|
37 |
## Intended uses and limitations
|
38 |
|
@@ -137,7 +137,7 @@ Weights were saved every 500 updates.
|
|
137 |
Below are the evaluation results on [Flores-200](https://github.com/facebookresearch/flores/tree/main/flores200),
|
138 |
[NTREX](https://github.com/MicrosoftTranslator/NTREX), and Projecte Aina's Catalan-Chinese test sets, compared to Google Translate for the ZH-CA direction. The evaluation was conducted [`tower-eval`](https://github.com/deep-spin/tower-eval) following the standard setting (beam search with beam size 5, limiting the translation length to 200 tokens). We report the following metrics:
|
139 |
|
140 |
-
- BLEU: Sacrebleu implementation, version:2.4.0.
|
141 |
- ChrF: Sacrebleu implementation.
|
142 |
- Comet: Model checkpoint: "Unbabel/wmt22-comet-da".
|
143 |
- Comet-kiwi: Model checkpoint: "Unbabel/wmt22-cometkiwi-da".
|
|
|
32 |
totalling 94.187.858 sentence pairs. 113.305 sentence pairs were parallel data collected from the web, while the remaining 94.074.553 sentence pairs
|
33 |
were parallel synthetic data created using the
|
34 |
[Aina Project's Spanish-Catalan machine translation model](https://huggingface.co/projecte-aina/aina-translator-es-ca) and the [Aina Project's English-Catalan machine translation model](https://huggingface.co/projecte-aina/aina-translator-en-ca).
|
35 |
+
The model was evaluated on the Flores, NTREX, and Projecte Aina's Catalan-Chinese evaluation datasets, achieving results comparable to those of Google Translate.
|
36 |
|
37 |
## Intended uses and limitations
|
38 |
|
|
|
137 |
Below are the evaluation results on [Flores-200](https://github.com/facebookresearch/flores/tree/main/flores200),
|
138 |
[NTREX](https://github.com/MicrosoftTranslator/NTREX), and Projecte Aina's Catalan-Chinese test sets, compared to Google Translate for the ZH-CA direction. The evaluation was conducted [`tower-eval`](https://github.com/deep-spin/tower-eval) following the standard setting (beam search with beam size 5, limiting the translation length to 200 tokens). We report the following metrics:
|
139 |
|
140 |
+
- BLEU: Sacrebleu implementation, version: 2.4.0.
|
141 |
- ChrF: Sacrebleu implementation.
|
142 |
- Comet: Model checkpoint: "Unbabel/wmt22-comet-da".
|
143 |
- Comet-kiwi: Model checkpoint: "Unbabel/wmt22-cometkiwi-da".
|