Fixing some errors of the leaderboard evaluation results in the ModelCard yaml
Browse filesThe name of a few benchmarks are incorrect on the model metadata.
This commit fixes some minor errors of the [last PR](9) on the ModelCard YAML metadata.
README.md
CHANGED
@@ -167,6 +167,19 @@ model-index:
|
|
167 |
- type: f1_macro
|
168 |
value: 43.01
|
169 |
name: f1-macro
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
170 |
- type: pearson
|
171 |
value: 8.92
|
172 |
name: pearson
|
@@ -194,7 +207,7 @@ model-index:
|
|
194 |
name: Text Generation
|
195 |
dataset:
|
196 |
name: HateBR Binary
|
197 |
-
type:
|
198 |
split: test
|
199 |
args:
|
200 |
num_few_shot: 25
|
@@ -202,6 +215,19 @@ model-index:
|
|
202 |
- type: f1_macro
|
203 |
value: 50.46
|
204 |
name: f1-macro
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
205 |
- type: f1_macro
|
206 |
value: 41.19
|
207 |
name: f1-macro
|
|
|
167 |
- type: f1_macro
|
168 |
value: 43.01
|
169 |
name: f1-macro
|
170 |
+
source:
|
171 |
+
url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard?query=22h/open-cabrita3b
|
172 |
+
name: Open Portuguese LLM Leaderboard
|
173 |
+
- task:
|
174 |
+
type: text-generation
|
175 |
+
name: Text Generation
|
176 |
+
dataset:
|
177 |
+
name: Assin2 STS
|
178 |
+
type: eduagarcia/portuguese_benchmark
|
179 |
+
split: test
|
180 |
+
args:
|
181 |
+
num_few_shot: 15
|
182 |
+
metrics:
|
183 |
- type: pearson
|
184 |
value: 8.92
|
185 |
name: pearson
|
|
|
207 |
name: Text Generation
|
208 |
dataset:
|
209 |
name: HateBR Binary
|
210 |
+
type: ruanchaves/hatebr
|
211 |
split: test
|
212 |
args:
|
213 |
num_few_shot: 25
|
|
|
215 |
- type: f1_macro
|
216 |
value: 50.46
|
217 |
name: f1-macro
|
218 |
+
source:
|
219 |
+
url: https://huggingface.co/spaces/eduagarcia/open_pt_llm_leaderboard?query=22h/open-cabrita3b
|
220 |
+
name: Open Portuguese LLM Leaderboard
|
221 |
+
- task:
|
222 |
+
type: text-generation
|
223 |
+
name: Text Generation
|
224 |
+
dataset:
|
225 |
+
name: PT Hate Speech Binary
|
226 |
+
type: hate_speech_portuguese
|
227 |
+
split: test
|
228 |
+
args:
|
229 |
+
num_few_shot: 25
|
230 |
+
metrics:
|
231 |
- type: f1_macro
|
232 |
value: 41.19
|
233 |
name: f1-macro
|