Datasets:

License:
daiteng01 commited on
Commit
b6663e0
·
verified ·
1 Parent(s): 988cd4a

Delete claude-3-7-sonnet-20250219

Browse files
claude-3-7-sonnet-20250219/results_2025-01-25T10-42-53.190540.json DELETED
@@ -1,742 +0,0 @@
1
- {
2
- "config_general": {
3
- "model_name": "claude-3-7-sonnet-20250219",
4
- "model_dtype": "float16",
5
- "model_size": 0
6
- },
7
- "results": {
8
- "CMMMU": {
9
- "艺术与设计": {
10
- "num": 88,
11
- "correct": 63,
12
- "accuracy": 71.59
13
- },
14
- "overall": {
15
- "num": 900,
16
- "correct": 460,
17
- "accuracy": 51.11
18
- },
19
- "商业": {
20
- "num": 126,
21
- "correct": 42,
22
- "accuracy": 33.33
23
- },
24
- "科学": {
25
- "num": 204,
26
- "correct": 88,
27
- "accuracy": 43.14
28
- },
29
- "健康与医学": {
30
- "num": 153,
31
- "correct": 92,
32
- "accuracy": 60.13
33
- },
34
- "人文社会科学": {
35
- "num": 85,
36
- "correct": 52,
37
- "accuracy": 61.18
38
- },
39
- "技术与工程": {
40
- "num": 244,
41
- "correct": 123,
42
- "accuracy": 50.41
43
- },
44
- "accuracy": 51.11,
45
- "acc_stderr": 0,
46
- "acc": 51.11
47
- },
48
- "MMMU": {
49
- "accuracy": 54.44,
50
- "subject_score": {
51
- "Accounting": 43.33,
52
- "Agriculture": 56.67,
53
- "Architecture": 33.33,
54
- "Art": 83.33,
55
- "Basic": 76.67,
56
- "Biology": 56.67,
57
- "Chemistry": 53.33,
58
- "Clinical": 76.67,
59
- "Computer": 36.67,
60
- "Design": 83.33,
61
- "Diagnostics": 40.0,
62
- "Economics": 53.33,
63
- "Electronics": 26.67,
64
- "Energy": 23.33,
65
- "Finance": 40.0,
66
- "Geography": 40.0,
67
- "History": 70.0,
68
- "Literature": 86.67,
69
- "Manage": 66.67,
70
- "Marketing": 46.67,
71
- "Materials": 43.33,
72
- "Math": 16.67,
73
- "Mechanical": 30.0,
74
- "Music": 40.0,
75
- "Pharmacy": 60.0,
76
- "Physics": 50.0,
77
- "Psychology": 73.33,
78
- "Public": 70.0,
79
- "Sociology": 73.33
80
- },
81
- "difficulty_score": {
82
- "Medium": 55.42,
83
- "Easy": 67.12,
84
- "Hard": 31.49
85
- },
86
- "acc_stderr": 0,
87
- "acc": 54.44
88
- },
89
- "MMMU_Pro_standard": {
90
- "accuracy": 38.55,
91
- "subject_score": {
92
- "History": 48.21,
93
- "Art": 64.15,
94
- "Design": 66.67,
95
- "Literature": 71.15,
96
- "Agriculture": 35.0,
97
- "Finance": 33.33,
98
- "Sociology": 51.85,
99
- "Accounting": 39.66,
100
- "Energy_and_Power": 18.97,
101
- "Pharmacy": 43.86,
102
- "Clinical_Medicine": 44.07,
103
- "Architecture_and_Engineering": 18.33,
104
- "Public_Health": 51.72,
105
- "Physics": 31.67,
106
- "Art_Theory": 61.82,
107
- "Psychology": 45.0,
108
- "Electronics": 35.0,
109
- "Biology": 32.2,
110
- "Economics": 35.59,
111
- "Manage": 30.0,
112
- "Mechanical_Engineering": 18.64,
113
- "Diagnostics_and_Laboratory_Medicine": 31.67,
114
- "Basic_Medical_Science": 34.62,
115
- "Computer_Science": 41.67,
116
- "Math": 21.67,
117
- "Music": 21.67,
118
- "Materials": 26.67,
119
- "Marketing": 35.59,
120
- "Chemistry": 36.67,
121
- "Geography": 38.46
122
- },
123
- "difficulty_score": {
124
- "Medium": 36.08,
125
- "Easy": 50.76,
126
- "Hard": 27.43
127
- },
128
- "acc_stderr": 0,
129
- "acc": 38.55
130
- },
131
- "MMMU_Pro_vision": {
132
- "accuracy": 67.86,
133
- "subject_score": {
134
- "History": 71.43,
135
- "Art": 58.49,
136
- "Design": 68.33,
137
- "Literature": 69.23,
138
- "Agriculture": 71.67,
139
- "Sociology": 81.48,
140
- "Finance": 58.33,
141
- "Accounting": 77.59,
142
- "Energy_and_Power": 75.86,
143
- "Pharmacy": 73.68,
144
- "Clinical_Medicine": 67.8,
145
- "Architecture_and_Engineering": 56.67,
146
- "Public_Health": 63.79,
147
- "Physics": 73.33,
148
- "Art_Theory": 67.27,
149
- "Electronics": 58.33,
150
- "Psychology": 73.33,
151
- "Biology": 64.41,
152
- "Manage": 68.0,
153
- "Economics": 67.8,
154
- "Mechanical_Engineering": 72.88,
155
- "Diagnostics_and_Laboratory_Medicine": 61.67,
156
- "Basic_Medical_Science": 75.0,
157
- "Computer_Science": 70.0,
158
- "Math": 71.67,
159
- "Music": 53.33,
160
- "Materials": 71.67,
161
- "Marketing": 54.24,
162
- "Chemistry": 63.33,
163
- "Geography": 78.85
164
- },
165
- "acc_stderr": 0,
166
- "acc": 67.86
167
- },
168
- "MmvetV2": {
169
- "reject_info": {
170
- "reject_rate": 2.9,
171
- "reject_number": 15,
172
- "total_question": 517
173
- },
174
- "accuracy": 68.9044,
175
- "capability_scores": {
176
- "math": 72.94117647058825,
177
- "ocr": 71.9024390243902,
178
- "spat": 62.526315789473664,
179
- "rec": 67.20403022670028,
180
- "know": 64.45945945945944,
181
- "gen": 69.3233082706767,
182
- "seq": 72.36363636363635
183
- },
184
- "capability_detail_scores": {
185
- "math_ocr": 81.81818181818183,
186
- "math_ocr_spat": 73.33333333333333,
187
- "math_ocr_spat_rec": 45.0,
188
- "spat_rec": 63.70370370370372,
189
- "ocr_spat": 71.53846153846153,
190
- "ocr_spat_rec": 45.0,
191
- "know_ocr_spat": 75.0,
192
- "ocr_rec": 95.0,
193
- "know_spat_rec": 40.0,
194
- "ocr": 76.875,
195
- "rec": 73.38983050847457,
196
- "know_rec": 65.45454545454545,
197
- "know_rec_gen": 67.23404255319146,
198
- "know_ocr_rec_gen": 72.3076923076923,
199
- "ocr_spat_rec_gen": 71.66666666666669,
200
- "ocr_spat_gen": 85.00000000000001,
201
- "spat_math_ocr_gen_seq": 80.0,
202
- "spat_math_ocr_seq_rec": 50.0,
203
- "spat_rec_gen": 45.49999999999999,
204
- "math_ocr_spat_gen": 60.0,
205
- "seq_spat_rec": 60.0,
206
- "ocr_seq_spat_rec": 50.0,
207
- "know_spat_rec_gen": 50.0,
208
- "rec_gen": 73.23529411764707,
209
- "know_ocr_spat_rec": 25.0,
210
- "spat_know_ocr_gen_rec": 80.0,
211
- "math_ocr_rec": 100.0,
212
- "ocr_rec_gen": 82.0,
213
- "ocr_seq_rec_gen": 75.71428571428572,
214
- "ocr_gen": 75.3846153846154,
215
- "seq_rec_gen": 80.0,
216
- "seq_rec": 80.0,
217
- "seq_spat_rec_gen": 75.00000000000001,
218
- "know_seq_rec": 0.0,
219
- "know_seq_rec_gen": 75.0,
220
- "spat_ocr_gen_seq_rec": 53.333333333333336,
221
- "know_ocr_gen_seq_rec": 100.0,
222
- "know_math_rec": 50.0,
223
- "ocr_seq_rec": 100.0
224
- },
225
- "acc_stderr": 0,
226
- "acc": 68.9044
227
- },
228
- "MathVerse": {
229
- "reject_info": {
230
- "reject_rate": 0.05,
231
- "reject_number": 2,
232
- "total_question": 3940
233
- },
234
- "Vision Dominant": {
235
- "accuracy": 50.76,
236
- "correct": 400,
237
- "total": 788
238
- },
239
- "Total": {
240
- "accuracy": 53.78,
241
- "correct": 2118,
242
- "total": 3938
243
- },
244
- "Vision Intensive": {
245
- "accuracy": 51.9,
246
- "correct": 409,
247
- "total": 788
248
- },
249
- "Text Lite": {
250
- "accuracy": 55.33,
251
- "correct": 436,
252
- "total": 788
253
- },
254
- "Text Dominant": {
255
- "accuracy": 62.44,
256
- "correct": 492,
257
- "total": 788
258
- },
259
- "Vision Only": {
260
- "accuracy": 48.47,
261
- "correct": 381,
262
- "total": 786
263
- },
264
- "accuracy": 53.78,
265
- "acc_stderr": 0,
266
- "acc": 53.78
267
- },
268
- "Ocrlite": {
269
- "final_score": [
270
- 1247,
271
- 1644
272
- ],
273
- "accuracy": 75.852,
274
- "Key Information Extraction-Bookshelf": [
275
- 28,
276
- 51,
277
- 0.549,
278
- {
279
- "Default": [
280
- 28,
281
- 51,
282
- 0.549
283
- ]
284
- }
285
- ],
286
- "Scene Text-centric VQA-diet_constraints": [
287
- 58,
288
- 90,
289
- 0.644,
290
- {
291
- "Default": [
292
- 58,
293
- 90,
294
- 0.644
295
- ]
296
- }
297
- ],
298
- "Doc-oriented VQA-Control": [
299
- 150,
300
- 189,
301
- 0.794,
302
- {
303
- "Default": [
304
- 150,
305
- 189,
306
- 0.794
307
- ]
308
- }
309
- ],
310
- "Doc-oriented VQA": [
311
- 180,
312
- 204,
313
- 0.882,
314
- {
315
- "Default": [
316
- 180,
317
- 204,
318
- 0.882
319
- ]
320
- }
321
- ],
322
- "Scene Text-centric VQA-Fake_logo": [
323
- 64,
324
- 119,
325
- 0.538,
326
- {
327
- "Default": [
328
- 64,
329
- 119,
330
- 0.538
331
- ]
332
- }
333
- ],
334
- "Handwritten Mathematical Expression Recognition": [
335
- 4,
336
- 100,
337
- 0.04,
338
- {
339
- "Default": [
340
- 4,
341
- 100,
342
- 0.04
343
- ]
344
- }
345
- ],
346
- "Key Information Extraction": [
347
- 186,
348
- 209,
349
- 0.89,
350
- {
351
- "Default": [
352
- 186,
353
- 209,
354
- 0.89
355
- ]
356
- }
357
- ],
358
- "Scene Text-centric VQA-Control": [
359
- 159,
360
- 200,
361
- 0.795,
362
- {
363
- "Default": [
364
- 159,
365
- 200,
366
- 0.795
367
- ]
368
- }
369
- ],
370
- "Scene Text-centric VQA": [
371
- 245,
372
- 282,
373
- 0.869,
374
- {
375
- "Default": [
376
- 245,
377
- 282,
378
- 0.869
379
- ]
380
- }
381
- ],
382
- "Artistic Text Recognition": [
383
- 41,
384
- 50,
385
- 0.82,
386
- {
387
- "Default": [
388
- 41,
389
- 50,
390
- 0.82
391
- ]
392
- }
393
- ],
394
- "Irregular Text Recognition": [
395
- 42,
396
- 50,
397
- 0.84,
398
- {
399
- "Default": [
400
- 42,
401
- 50,
402
- 0.84
403
- ]
404
- }
405
- ],
406
- "Non-Semantic Text Recognition": [
407
- 42,
408
- 50,
409
- 0.84,
410
- {
411
- "Default": [
412
- 42,
413
- 50,
414
- 0.84
415
- ]
416
- }
417
- ],
418
- "Regular Text Recognition": [
419
- 48,
420
- 50,
421
- 0.96,
422
- {
423
- "Default": [
424
- 48,
425
- 50,
426
- 0.96
427
- ]
428
- }
429
- ],
430
- "acc_stderr": 0,
431
- "acc": 75.852
432
- },
433
- "OcrliteZh": {
434
- "final_score": [
435
- 143,
436
- 234
437
- ],
438
- "accuracy": 61.111,
439
- "Docvqa": [
440
- 9,
441
- 10,
442
- 0.9,
443
- {
444
- "Default": [
445
- 9,
446
- 10,
447
- 0.9
448
- ]
449
- }
450
- ],
451
- "Chartqa-human": [
452
- 7,
453
- 10,
454
- 0.7,
455
- {
456
- "Default": [
457
- 7,
458
- 10,
459
- 0.7
460
- ]
461
- }
462
- ],
463
- "Chartqa-au": [
464
- 8,
465
- 10,
466
- 0.8,
467
- {
468
- "Default": [
469
- 8,
470
- 10,
471
- 0.8
472
- ]
473
- }
474
- ],
475
- "infographic": [
476
- 6,
477
- 10,
478
- 0.6,
479
- {
480
- "Default": [
481
- 6,
482
- 10,
483
- 0.6
484
- ]
485
- }
486
- ],
487
- "Key Information Extraction": [
488
- 34,
489
- 45,
490
- 0.756,
491
- {
492
- "Default": [
493
- 34,
494
- 45,
495
- 0.756
496
- ]
497
- }
498
- ],
499
- "Scene Text-centric VQA": [
500
- 28,
501
- 40,
502
- 0.7,
503
- {
504
- "Default": [
505
- 28,
506
- 40,
507
- 0.7
508
- ]
509
- }
510
- ],
511
- "Artistic Text Recognition": [
512
- 3,
513
- 11,
514
- 0.273,
515
- {
516
- "Default": [
517
- 3,
518
- 11,
519
- 0.273
520
- ]
521
- }
522
- ],
523
- "IrRegular Text Recognition": [
524
- 5,
525
- 11,
526
- 0.455,
527
- {
528
- "Default": [
529
- 5,
530
- 11,
531
- 0.455
532
- ]
533
- }
534
- ],
535
- "Non-semantic Text Recognition": [
536
- 6,
537
- 12,
538
- 0.5,
539
- {
540
- "Default": [
541
- 6,
542
- 12,
543
- 0.5
544
- ]
545
- }
546
- ],
547
- "Regular Text Recognition": [
548
- 10,
549
- 11,
550
- 0.909,
551
- {
552
- "Default": [
553
- 10,
554
- 11,
555
- 0.909
556
- ]
557
- }
558
- ],
559
- "Handwriting_CN": [
560
- 6,
561
- 20,
562
- 0.3,
563
- {
564
- "Default": [
565
- 6,
566
- 20,
567
- 0.3
568
- ]
569
- }
570
- ],
571
- "Chinese Unlimited": [
572
- 21,
573
- 44,
574
- 0.477,
575
- {
576
- "Default": [
577
- 21,
578
- 44,
579
- 0.477
580
- ]
581
- }
582
- ],
583
- "acc_stderr": 0,
584
- "acc": 61.111
585
- },
586
- "CharXiv": {
587
- "reject_info": {
588
- "reject_rate": 0.08,
589
- "reject_number": 4,
590
- "total_question": 5000
591
- },
592
- "descriptive": {
593
- "Overall Score": 82.88,
594
- "By Question": {
595
- "Q1": 59.84,
596
- "Q2": 82.53,
597
- "Q3": 70.82,
598
- "Q4": 88.72,
599
- "Q5": 89.03,
600
- "Q6": 86.35,
601
- "Q7": 90.17,
602
- "Q8": 95.98,
603
- "Q9": 92.0,
604
- "Q10": 86.3,
605
- "Q11": 82.29,
606
- "Q12": 83.52,
607
- "Q13": 71.69,
608
- "Q14": 82.62,
609
- "Q15": 80.19,
610
- "Q16": 77.78,
611
- "Q17": 70.09,
612
- "Q18": 96.76,
613
- "Q19": 93.85
614
- },
615
- "By Category": {
616
- "Information Extraction": 81.11,
617
- "Enumeration": 84.01,
618
- "Pattern Recognition": 89.74,
619
- "Counting": 86.26,
620
- "Compositionality": 70.09
621
- },
622
- "By Subplot": {
623
- "1 Subplot": 85.09,
624
- "2-4 Subplots": 82.26,
625
- "5+ Subplots": 80.25
626
- },
627
- "By Subject": {
628
- "Computer Science": 80.88,
629
- "Economics": 82.94,
630
- "Electrical Engineering and Systems Science": 84.66,
631
- "Mathematics": 85.53,
632
- "Physics": 81.3,
633
- "Quantitative Biology": 81.35,
634
- "Quantitative Finance": 81.03,
635
- "Statistics": 85.4
636
- },
637
- "By Year": {
638
- "2020": 80.97,
639
- "2021": 83.21,
640
- "2022": 84.22,
641
- "2023": 83.13
642
- },
643
- "N_valid": 3996,
644
- "N_invalid": 0,
645
- "Question Type": "Descriptive"
646
- },
647
- "reasoning": {
648
- "Overall Score": 59.5,
649
- "By Answer Type": {
650
- "Text-in-Chart": 66.59,
651
- "Text-in-General": 67.68,
652
- "Number-in-Chart": 48.28,
653
- "Number-in-General": 53.71
654
- },
655
- "By Source": {
656
- "GPT-Sourced": 69.57,
657
- "GPT-Inspired": 57.41,
658
- "Completely Human": 57.17
659
- },
660
- "By Subject": {
661
- "Computer Science": 59.52,
662
- "Economics": 55.07,
663
- "Electrical Engineering and Systems Science": 57.98,
664
- "Mathematics": 62.96,
665
- "Physics": 64.57,
666
- "Quantitative Biology": 63.49,
667
- "Quantitative Finance": 52.59,
668
- "Statistics": 59.29
669
- },
670
- "By Year": {
671
- "2020": 56.68,
672
- "2021": 65.13,
673
- "2022": 57.79,
674
- "2023": 58.06
675
- },
676
- "By Subplot": {
677
- "1 Subplot": 61.14,
678
- "2-4 Subplots": 57.41,
679
- "5+ Subplots": 60.17
680
- },
681
- "N_valid": 1000,
682
- "N_invalid": 0,
683
- "Question Type": "Reasoning"
684
- },
685
- "accuracy": 71.19,
686
- "acc_stderr": 0,
687
- "acc": 71.19
688
- },
689
- "MathVision": {
690
- "accuracy": 44.8,
691
- "acc_stderr": 0,
692
- "acc": 44.8
693
- },
694
- "CII-Bench": {
695
- "reject_info": {
696
- "reject_rate": 0.52,
697
- "reject_number": 4,
698
- "total_question": 765
699
- },
700
- "accuracy": 60.97,
701
- "domain_score": {
702
- "Art": 58.09,
703
- "CTC": 58.21,
704
- "Life": 65.37,
705
- "Society": 58.92,
706
- "Env.": 60.78,
707
- "Politics": 66.67
708
- },
709
- "emotion_score": {
710
- "Negative": 60.38,
711
- "Positive": 60.17,
712
- "Neutral": 62.26
713
- },
714
- "acc_stderr": 0,
715
- "acc": 60.97
716
- },
717
- "Blink": {
718
- "reject_info": {
719
- "reject_rate": 0.05,
720
- "reject_number": 1,
721
- "total_question": 1901
722
- },
723
- "accuracy": 64.37,
724
- "Art Style": 88.03,
725
- "Counting": 67.5,
726
- "Forensic Detection": 58.33,
727
- "Functional Correspondence": 53.85,
728
- "IQ Test": 32.0,
729
- "Jigsaw": 61.33,
730
- "Multi-view Reasoning": 51.13,
731
- "Object Localization": 63.11,
732
- "Relative Depth": 77.42,
733
- "Relative Reflectance": 31.34,
734
- "Semantic Correspondence": 55.8,
735
- "Spatial Relation": 79.02,
736
- "Visual Correspondence": 91.86,
737
- "Visual Similarity": 89.63,
738
- "acc_stderr": 0,
739
- "acc": 64.37
740
- }
741
- }
742
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
claude-3-7-sonnet-20250219/results_2025-03-10T11-01-07.004431.json DELETED
@@ -1,386 +0,0 @@
1
- {
2
- "config_general": {
3
- "model_name": "claude-3-7-sonnet-20250219",
4
- "model_dtype": "float16",
5
- "model_size": 0
6
- },
7
- "results": {
8
- "ChartQA": {
9
- "acc": 27.56,
10
- "acc_stderr": 0,
11
- "accuracy": 27.56,
12
- "human_test": {
13
- "total": 1250,
14
- "correct": 390,
15
- "accuracy": 31.2
16
- },
17
- "augmented_test": {
18
- "total": 1250,
19
- "correct": 299,
20
- "accuracy": 23.92
21
- }
22
- },
23
- "CMMMU": {
24
- "acc": 49.11,
25
- "acc_stderr": 0,
26
- "\u5546\u4e1a": {
27
- "num": 126,
28
- "correct": 43,
29
- "accuracy": 34.13
30
- },
31
- "\u79d1\u5b66": {
32
- "num": 204,
33
- "correct": 83,
34
- "accuracy": 40.69
35
- },
36
- "overall": {
37
- "num": 900,
38
- "correct": 442,
39
- "accuracy": 49.11
40
- },
41
- "accuracy": 49.11,
42
- "\u5065\u5eb7\u4e0e\u533b\u5b66": {
43
- "num": 153,
44
- "correct": 89,
45
- "accuracy": 58.17
46
- },
47
- "\u6280\u672f\u4e0e\u5de5\u7a0b": {
48
- "num": 244,
49
- "correct": 112,
50
- "accuracy": 45.9
51
- },
52
- "\u827a\u672f\u4e0e\u8bbe\u8ba1": {
53
- "num": 88,
54
- "correct": 64,
55
- "accuracy": 72.73
56
- },
57
- "\u4eba\u6587\u793e\u4f1a\u79d1\u5b66": {
58
- "num": 85,
59
- "correct": 51,
60
- "accuracy": 60.0
61
- }
62
- },
63
- "CMMU": {
64
- "acc": 25.36,
65
- "acc_stderr": 0,
66
- "val": {
67
- "multiple-choice": {
68
- "hard": {
69
- "total": 150,
70
- "correct": 15,
71
- "accuracy": 10.0
72
- },
73
- "normal": {
74
- "total": 1205,
75
- "correct": 300,
76
- "accuracy": 24.9
77
- }
78
- },
79
- "fill-in-the-blank": {
80
- "hard": {
81
- "total": 300,
82
- "correct": 87,
83
- "accuracy": 29.0
84
- },
85
- "normal": {
86
- "total": 507,
87
- "correct": 171,
88
- "accuracy": 33.73
89
- }
90
- },
91
- "multiple-response": {
92
- "hard": {
93
- "total": 94,
94
- "correct": 11,
95
- "accuracy": 11.7
96
- },
97
- "normal": {
98
- "total": 33,
99
- "correct": 7,
100
- "accuracy": 21.21
101
- }
102
- }
103
- },
104
- "test": {
105
- "multiple-choice": {
106
- "hard": {
107
- "total": 150,
108
- "correct": 21,
109
- "accuracy": 14.0
110
- },
111
- "normal": {
112
- "total": 1205,
113
- "correct": 294,
114
- "accuracy": 24.4
115
- }
116
- },
117
- "fill-in-the-blank": {
118
- "hard": {
119
- "total": 296,
120
- "correct": 84,
121
- "accuracy": 28.38
122
- },
123
- "normal": {
124
- "total": 529,
125
- "correct": 165,
126
- "accuracy": 31.19
127
- }
128
- },
129
- "multiple-response": {
130
- "hard": {
131
- "total": 95,
132
- "correct": 13,
133
- "accuracy": 13.68
134
- },
135
- "normal": {
136
- "total": 32,
137
- "correct": 8,
138
- "accuracy": 25.0
139
- }
140
- }
141
- },
142
- "val-overall": {
143
- "total": 2289,
144
- "correct": 591,
145
- "accuracy": 25.82,
146
- "bias_rate": 58.07
147
- },
148
- "test-overall": {
149
- "total": 2307,
150
- "correct": 585,
151
- "accuracy": 25.36,
152
- "bias_rate": 46.77
153
- }
154
- },
155
- "MMMU": {
156
- "acc": 52.33,
157
- "acc_stderr": 0,
158
- "accuracy": 52.33,
159
- "subject_score": {
160
- "Art": 81.67,
161
- "Math": 10.0,
162
- "Basic": 73.33,
163
- "Music": 30.0,
164
- "Design": 80.0,
165
- "Energy": 43.33,
166
- "Manage": 53.33,
167
- "Public": 56.67,
168
- "Biology": 60.0,
169
- "Finance": 33.33,
170
- "History": 66.67,
171
- "Physics": 60.0,
172
- "Clinical": 63.33,
173
- "Computer": 40.0,
174
- "Pharmacy": 70.0,
175
- "Chemistry": 40.0,
176
- "Economics": 53.33,
177
- "Geography": 43.33,
178
- "Marketing": 46.67,
179
- "Materials": 40.0,
180
- "Sociology": 73.33,
181
- "Accounting": 50.0,
182
- "Literature": 83.33,
183
- "Mechanical": 20.0,
184
- "Psychology": 76.67,
185
- "Agriculture": 56.67,
186
- "Diagnostics": 40.0,
187
- "Electronics": 13.33,
188
- "Architecture": 30.0
189
- },
190
- "difficulty_score": {
191
- "Easy": 65.08,
192
- "Hard": 28.73,
193
- "Medium": 53.54
194
- }
195
- },
196
- "MMMU_Pro_standard": {
197
- "acc": 38.61,
198
- "acc_stderr": 0,
199
- "accuracy": 38.61,
200
- "subject_score": {
201
- "Art": 66.04,
202
- "Math": 21.67,
203
- "Music": 20.0,
204
- "Design": 63.33,
205
- "Manage": 30.0,
206
- "Biology": 32.2,
207
- "Finance": 33.33,
208
- "History": 48.21,
209
- "Physics": 36.67,
210
- "Pharmacy": 43.86,
211
- "Chemistry": 40.0,
212
- "Economics": 38.98,
213
- "Geography": 38.46,
214
- "Marketing": 32.2,
215
- "Materials": 16.67,
216
- "Sociology": 53.7,
217
- "Accounting": 32.76,
218
- "Art_Theory": 61.82,
219
- "Literature": 71.15,
220
- "Psychology": 41.67,
221
- "Agriculture": 35.0,
222
- "Electronics": 43.33,
223
- "Public_Health": 50.0,
224
- "Computer_Science": 35.0,
225
- "Energy_and_Power": 24.14,
226
- "Clinical_Medicine": 47.46,
227
- "Basic_Medical_Science": 36.54,
228
- "Mechanical_Engineering": 20.34,
229
- "Architecture_and_Engineering": 25.0,
230
- "Diagnostics_and_Laboratory_Medicine": 28.33
231
- },
232
- "difficulty_score": {
233
- "Easy": 52.65,
234
- "Hard": 26.68,
235
- "Medium": 35.33
236
- }
237
- },
238
- "MMMU_Pro_vision": {
239
- "acc": 68.03,
240
- "acc_stderr": 0,
241
- "accuracy": 68.03,
242
- "subject_score": {
243
- "Art": 66.04,
244
- "Math": 70.0,
245
- "Music": 61.67,
246
- "Design": 70.0,
247
- "Manage": 70.0,
248
- "Biology": 66.1,
249
- "Finance": 53.33,
250
- "History": 69.64,
251
- "Physics": 65.0,
252
- "Pharmacy": 70.18,
253
- "Chemistry": 70.0,
254
- "Economics": 64.41,
255
- "Geography": 76.92,
256
- "Marketing": 57.63,
257
- "Materials": 65.0,
258
- "Sociology": 68.52,
259
- "Accounting": 72.41,
260
- "Art_Theory": 63.64,
261
- "Literature": 69.23,
262
- "Psychology": 83.33,
263
- "Agriculture": 70.0,
264
- "Electronics": 55.0,
265
- "Public_Health": 63.79,
266
- "Computer_Science": 71.67,
267
- "Energy_and_Power": 82.76,
268
- "Clinical_Medicine": 74.58,
269
- "Basic_Medical_Science": 76.92,
270
- "Mechanical_Engineering": 76.27,
271
- "Architecture_and_Engineering": 60.0,
272
- "Diagnostics_and_Laboratory_Medicine": 60.0
273
- }
274
- },
275
- "OCRBench": {
276
- "acc": 79.819,
277
- "acc_stderr": 0,
278
- "accuracy": 79.819,
279
- "final_score": [
280
- 795,
281
- 996
282
- ],
283
- "reject_info": {
284
- "reject_rate": 0.4,
285
- "reject_number": 4,
286
- "total_question": 1000
287
- },
288
- "Doc-oriented VQA": [
289
- 173,
290
- 196
291
- ],
292
- "Scene Text-centric VQA": [
293
- 169,
294
- 200
295
- ],
296
- "Handwriting Recognition": [
297
- 38,
298
- 50
299
- ],
300
- "Digit String Recognition": [
301
- 33,
302
- 50
303
- ],
304
- "Regular Text Recognition": [
305
- 49,
306
- 50
307
- ],
308
- "Artistic Text Recognition": [
309
- 44,
310
- 50
311
- ],
312
- "Irregular Text Recognition": [
313
- 45,
314
- 50
315
- ],
316
- "Key Information Extraction": [
317
- 176,
318
- 200
319
- ],
320
- "Non-Semantic Text Recognition": [
321
- 47,
322
- 50
323
- ],
324
- "Handwritten Mathematical Expression Recognition": [
325
- 21,
326
- 100
327
- ]
328
- },
329
- "MathVision": {
330
- "acc": 45.38,
331
- "acc_stderr": 0,
332
- "accuracy": 45.38,
333
- "reject_info": {
334
- "reject_rate": 0.03,
335
- "reject_number": 1,
336
- "total_question": 3040
337
- }
338
- },
339
- "CII-Bench": {
340
- "acc": 59.26,
341
- "acc_stderr": 0,
342
- "accuracy": 59.26,
343
- "reject_info": {
344
- "reject_rate": 0.52,
345
- "reject_number": 4,
346
- "total_question": 765
347
- },
348
- "domain_score": {
349
- "Art": 55.15,
350
- "CTC": 52.24,
351
- "Env.": 62.75,
352
- "Life": 63.64,
353
- "Society": 59.46,
354
- "Politics": 70.83
355
- },
356
- "emotion_score": {
357
- "Neutral": 59.62,
358
- "Negative": 59.25,
359
- "Positive": 58.87
360
- }
361
- },
362
- "Blink": {
363
- "acc": 63.34,
364
- "acc_stderr": 0,
365
- "Jigsaw": 63.33,
366
- "IQ Test": 26.67,
367
- "Counting": 65.0,
368
- "accuracy": 63.34,
369
- "Art Style": 83.76,
370
- "Relative Depth": 75.81,
371
- "Spatial Relation": 75.52,
372
- "Visual Similarity": 86.67,
373
- "Forensic Detection": 57.58,
374
- "Object Localization": 60.66,
375
- "Multi-view Reasoning": 54.14,
376
- "Relative Reflectance": 31.34,
377
- "Visual Correspondence": 91.28,
378
- "Semantic Correspondence": 50.36,
379
- "Functional Correspondence": 63.85
380
- }
381
- },
382
- "versions": {},
383
- "config_tasks": {},
384
- "summary_tasks": {},
385
- "summary_general": {}
386
- }