mihaimasala
commited on
Commit
•
f740b82
1
Parent(s):
e123088
Update README.md
Browse files
README.md
CHANGED
@@ -116,6 +116,24 @@ model-index:
|
|
116 |
- name: Average macro-f1
|
117 |
type: macro-f1
|
118 |
value: 62.49
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
119 |
- task:
|
120 |
type: text-generation
|
121 |
dataset:
|
@@ -134,6 +152,24 @@ model-index:
|
|
134 |
- name: Average bleu
|
135 |
type: bleu
|
136 |
value: 23.16
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
137 |
- task:
|
138 |
type: text-generation
|
139 |
dataset:
|
@@ -152,6 +188,24 @@ model-index:
|
|
152 |
- name: Average f1
|
153 |
type: f1
|
154 |
value: 70.74
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
155 |
- task:
|
156 |
type: text-generation
|
157 |
dataset:
|
@@ -170,6 +224,24 @@ model-index:
|
|
170 |
- name: Average pearson
|
171 |
type: pearson
|
172 |
value: 77.10
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
173 |
- task:
|
174 |
type: text-generation
|
175 |
dataset:
|
@@ -417,6 +489,7 @@ model-index:
|
|
417 |
type: pearson
|
418 |
value: 71.33
|
419 |
|
|
|
420 |
---
|
421 |
|
422 |
# Model Card for Model ID
|
@@ -540,10 +613,10 @@ print(tokenizer.decode(outputs[0]))
|
|
540 |
<td><center><strong>RO-EN<br>(Bleu)</strong></center>
|
541 |
</tr>
|
542 |
<tr>
|
543 |
-
<td>gemma-2-9b-it</td><td><center>90.82</center></td><td><center>52.51</center></td><td><center
|
544 |
</tr>
|
545 |
<tr>
|
546 |
-
<td><em>RoGemma2-9b-Instruct-2024-10-09</em></td><td><center><em>96.19</em></center></td><td><center><em>62.49</em></center></td><td><center><em
|
547 |
</tr>
|
548 |
<tr>
|
549 |
<td>RoGemma2-9b-Instruct-DPO-2024-10-09</td><td><center><strong>97.74</strong></center></td><td><center><strong>67.40</strong></center></td><td><center>-</center></td><td><center>-</center></td><td><center><strong>27.32</strong></center></td><td><center>15.96</center></td><td><center>-</center></td><td><center>-</center></td>
|
@@ -578,10 +651,10 @@ print(tokenizer.decode(outputs[0]))
|
|
578 |
<td><center><strong>(Pearson)</strong></center></td>
|
579 |
</tr>
|
580 |
<tr>
|
581 |
-
<td>gemma-2-9b-it</td><td><center>37.56</center></td><td><center>57.48</center></td><td><center
|
582 |
</tr>
|
583 |
<tr>
|
584 |
-
<td><em>RoGemma2-9b-Instruct-2024-10-09</em></td><td><center><em><strong>51.37</strong></em></center></td><td><center><em><strong>70.74</strong></em></center></td><td><center><em
|
585 |
</tr>
|
586 |
<tr>
|
587 |
<td>RoGemma2-9b-Instruct-DPO-2024-10-09</td><td><center>32.42</center></td><td><center>58.68</center></td><td><center>-</center></td><td><center>-</center></td><td><center><strong>80.82</strong></center></td><td><center><strong>81.50</strong></center></td><td><center>-</center></td><td><center>-</center></td>
|
@@ -635,6 +708,7 @@ print(tokenizer.decode(outputs[0]))
|
|
635 |
</tbody>
|
636 |
</table>
|
637 |
|
|
|
638 |
## RoGemma2 Model Family
|
639 |
|
640 |
| Model | Link |
|
|
|
116 |
- name: Average macro-f1
|
117 |
type: macro-f1
|
118 |
value: 62.49
|
119 |
+
- task:
|
120 |
+
type: text-generation
|
121 |
+
dataset:
|
122 |
+
name: LaRoSeDa_binary_finetuned
|
123 |
+
type: LaRoSeDa_binary_finetuned
|
124 |
+
metrics:
|
125 |
+
- name: Average macro-f1
|
126 |
+
type: macro-f1
|
127 |
+
value: 98.93
|
128 |
+
- task:
|
129 |
+
type: text-generation
|
130 |
+
dataset:
|
131 |
+
name: LaRoSeDa_multiclass_finetuned
|
132 |
+
type: LaRoSeDa_multiclass_finetuned
|
133 |
+
metrics:
|
134 |
+
- name: Average macro-f1
|
135 |
+
type: macro-f1
|
136 |
+
value: 88.33
|
137 |
- task:
|
138 |
type: text-generation
|
139 |
dataset:
|
|
|
152 |
- name: Average bleu
|
153 |
type: bleu
|
154 |
value: 23.16
|
155 |
+
- task:
|
156 |
+
type: text-generation
|
157 |
+
dataset:
|
158 |
+
name: WMT_EN-RO_finetuned
|
159 |
+
type: WMT_EN-RO_finetuned
|
160 |
+
metrics:
|
161 |
+
- name: Average bleu
|
162 |
+
type: bleu
|
163 |
+
value: 28.43
|
164 |
+
- task:
|
165 |
+
type: text-generation
|
166 |
+
dataset:
|
167 |
+
name: WMT_RO-EN_finetuned
|
168 |
+
type: WMT_RO-EN_finetuned
|
169 |
+
metrics:
|
170 |
+
- name: Average bleu
|
171 |
+
type: bleu
|
172 |
+
value: 40.94
|
173 |
- task:
|
174 |
type: text-generation
|
175 |
dataset:
|
|
|
188 |
- name: Average f1
|
189 |
type: f1
|
190 |
value: 70.74
|
191 |
+
- task:
|
192 |
+
type: text-generation
|
193 |
+
dataset:
|
194 |
+
name: XQuAD_finetuned
|
195 |
+
type: XQuAD_finetuned
|
196 |
+
metrics:
|
197 |
+
- name: Average exact_match
|
198 |
+
type: exact_match
|
199 |
+
value: 50.00
|
200 |
+
- task:
|
201 |
+
type: text-generation
|
202 |
+
dataset:
|
203 |
+
name: XQuAD_finetuned
|
204 |
+
type: XQuAD_finetuned
|
205 |
+
metrics:
|
206 |
+
- name: Average f1
|
207 |
+
type: f1
|
208 |
+
value: 64.10
|
209 |
- task:
|
210 |
type: text-generation
|
211 |
dataset:
|
|
|
224 |
- name: Average pearson
|
225 |
type: pearson
|
226 |
value: 77.10
|
227 |
+
- task:
|
228 |
+
type: text-generation
|
229 |
+
dataset:
|
230 |
+
name: STS_finetuned
|
231 |
+
type: STS_finetuned
|
232 |
+
metrics:
|
233 |
+
- name: Average spearman
|
234 |
+
type: spearman
|
235 |
+
value: 89.45
|
236 |
+
- task:
|
237 |
+
type: text-generation
|
238 |
+
dataset:
|
239 |
+
name: STS_finetuned
|
240 |
+
type: STS_finetuned
|
241 |
+
metrics:
|
242 |
+
- name: Average pearson
|
243 |
+
type: pearson
|
244 |
+
value: 89.89
|
245 |
- task:
|
246 |
type: text-generation
|
247 |
dataset:
|
|
|
489 |
type: pearson
|
490 |
value: 71.33
|
491 |
|
492 |
+
|
493 |
---
|
494 |
|
495 |
# Model Card for Model ID
|
|
|
613 |
<td><center><strong>RO-EN<br>(Bleu)</strong></center>
|
614 |
</tr>
|
615 |
<tr>
|
616 |
+
<td>gemma-2-9b-it</td><td><center>90.82</center></td><td><center>52.51</center></td><td><center><strong>98.97</strong></center></td><td><center>86.02</center></td><td><center>19.97</center></td><td><center><strong>28.94</strong></center></td><td><center>27.94</center></td><td><center><strong>41.61</strong></center></td>
|
617 |
</tr>
|
618 |
<tr>
|
619 |
+
<td><em>RoGemma2-9b-Instruct-2024-10-09</em></td><td><center><em>96.19</em></center></td><td><center><em>62.49</em></center></td><td><center><em>98.93</em></center></td><td><center><em><strong>88.33</strong></em></center></td><td><center><em>25.74</em></center></td><td><center><em>23.16</em></center></td><td><center><em><strong>28.43</strong></em></center></td><td><center><em>40.94</em></center></td>
|
620 |
</tr>
|
621 |
<tr>
|
622 |
<td>RoGemma2-9b-Instruct-DPO-2024-10-09</td><td><center><strong>97.74</strong></center></td><td><center><strong>67.40</strong></center></td><td><center>-</center></td><td><center>-</center></td><td><center><strong>27.32</strong></center></td><td><center>15.96</center></td><td><center>-</center></td><td><center>-</center></td>
|
|
|
651 |
<td><center><strong>(Pearson)</strong></center></td>
|
652 |
</tr>
|
653 |
<tr>
|
654 |
+
<td>gemma-2-9b-it</td><td><center>37.56</center></td><td><center>57.48</center></td><td><center><strong>71.09</strong></center></td><td><center><strong>84.78</strong></center></td><td><center>71.39</center></td><td><center>71.73</center></td><td><center>89.07</center></td><td><center>89.29</center></td>
|
655 |
</tr>
|
656 |
<tr>
|
657 |
+
<td><em>RoGemma2-9b-Instruct-2024-10-09</em></td><td><center><em><strong>51.37</strong></em></center></td><td><center><em><strong>70.74</strong></em></center></td><td><center><em>50.00</em></center></td><td><center><em>64.10</em></center></td><td><center><em>77.15</em></center></td><td><center><em>77.10</em></center></td><td><center><em><strong>89.45</strong></em></center></td><td><center><em><strong>89.89</strong></em></center></td>
|
658 |
</tr>
|
659 |
<tr>
|
660 |
<td>RoGemma2-9b-Instruct-DPO-2024-10-09</td><td><center>32.42</center></td><td><center>58.68</center></td><td><center>-</center></td><td><center>-</center></td><td><center><strong>80.82</strong></center></td><td><center><strong>81.50</strong></center></td><td><center>-</center></td><td><center>-</center></td>
|
|
|
708 |
</tbody>
|
709 |
</table>
|
710 |
|
711 |
+
|
712 |
## RoGemma2 Model Family
|
713 |
|
714 |
| Model | Link |
|