|
{ |
|
"config_general": { |
|
"model_name": "Qwen2.5-VL-7B-Instruct", |
|
"model_dtype": "float16", |
|
"model_size": 0 |
|
}, |
|
"results": { |
|
"CMMMU": { |
|
"艺术与设计": { |
|
"num": 88, |
|
"correct": 56, |
|
"accuracy": 63.64 |
|
}, |
|
"overall": { |
|
"num": 900, |
|
"correct": 406, |
|
"accuracy": 45.11 |
|
}, |
|
"商业": { |
|
"num": 126, |
|
"correct": 35, |
|
"accuracy": 27.78 |
|
}, |
|
"科学": { |
|
"num": 204, |
|
"correct": 83, |
|
"accuracy": 40.69 |
|
}, |
|
"健康与医学": { |
|
"num": 153, |
|
"correct": 72, |
|
"accuracy": 47.06 |
|
}, |
|
"人文社会科学": { |
|
"num": 85, |
|
"correct": 49, |
|
"accuracy": 57.65 |
|
}, |
|
"技术与工程": { |
|
"num": 244, |
|
"correct": 111, |
|
"accuracy": 45.49 |
|
}, |
|
"accuracy": 45.11, |
|
"acc_stderr": 0, |
|
"acc": 45.11 |
|
}, |
|
"MMMU": { |
|
"accuracy": 50.67, |
|
"subject_score": { |
|
"Accounting": 33.33, |
|
"Agriculture": 40.0, |
|
"Architecture": 50.0, |
|
"Art": 78.33, |
|
"Basic": 53.33, |
|
"Biology": 50.0, |
|
"Chemistry": 36.67, |
|
"Clinical": 56.67, |
|
"Computer": 50.0, |
|
"Design": 73.33, |
|
"Diagnostics": 40.0, |
|
"Economics": 43.33, |
|
"Electronics": 33.33, |
|
"Energy": 30.0, |
|
"Finance": 33.33, |
|
"Geography": 46.67, |
|
"History": 66.67, |
|
"Literature": 80.0, |
|
"Manage": 33.33, |
|
"Marketing": 56.67, |
|
"Materials": 40.0, |
|
"Math": 43.33, |
|
"Mechanical": 46.67, |
|
"Music": 30.0, |
|
"Pharmacy": 70.0, |
|
"Physics": 33.33, |
|
"Psychology": 73.33, |
|
"Public": 60.0, |
|
"Sociology": 60.0 |
|
}, |
|
"difficulty_score": { |
|
"Medium": 46.46, |
|
"Easy": 64.07, |
|
"Hard": 38.67 |
|
}, |
|
"acc_stderr": 0, |
|
"acc": 50.67 |
|
}, |
|
"MMMU_Pro_standard": { |
|
"accuracy": 36.76, |
|
"subject_score": { |
|
"History": 48.21, |
|
"Finance": 21.67, |
|
"Design": 63.33, |
|
"Literature": 65.38, |
|
"Agriculture": 25.0, |
|
"Clinical_Medicine": 28.81, |
|
"Accounting": 25.86, |
|
"Sociology": 38.89, |
|
"Art": 54.72, |
|
"Physics": 25.0, |
|
"Public_Health": 20.69, |
|
"Energy_and_Power": 20.69, |
|
"Pharmacy": 49.12, |
|
"Electronics": 53.33, |
|
"Architecture_and_Engineering": 31.67, |
|
"Art_Theory": 67.27, |
|
"Psychology": 38.33, |
|
"Economics": 33.9, |
|
"Biology": 33.9, |
|
"Diagnostics_and_Laboratory_Medicine": 28.33, |
|
"Manage": 26.0, |
|
"Mechanical_Engineering": 35.59, |
|
"Basic_Medical_Science": 36.54, |
|
"Computer_Science": 45.0, |
|
"Math": 25.0, |
|
"Music": 26.67, |
|
"Materials": 23.33, |
|
"Marketing": 30.51, |
|
"Chemistry": 36.67, |
|
"Geography": 51.92 |
|
}, |
|
"difficulty_score": { |
|
"Medium": 34.46, |
|
"Easy": 46.59, |
|
"Hard": 28.43 |
|
}, |
|
"acc_stderr": 0, |
|
"acc": 36.76 |
|
}, |
|
"MMMU_Pro_vision": { |
|
"accuracy": 34.91, |
|
"subject_score": { |
|
"Design": 46.67, |
|
"History": 53.57, |
|
"Sociology": 33.33, |
|
"Art": 50.94, |
|
"Literature": 69.23, |
|
"Agriculture": 23.33, |
|
"Pharmacy": 40.35, |
|
"Clinical_Medicine": 30.51, |
|
"Architecture_and_Engineering": 20.0, |
|
"Accounting": 43.1, |
|
"Physics": 33.33, |
|
"Art_Theory": 50.91, |
|
"Energy_and_Power": 18.97, |
|
"Psychology": 26.67, |
|
"Biology": 30.51, |
|
"Manage": 22.0, |
|
"Economics": 37.29, |
|
"Public_Health": 46.55, |
|
"Mechanical_Engineering": 23.73, |
|
"Diagnostics_and_Laboratory_Medicine": 26.67, |
|
"Electronics": 41.67, |
|
"Basic_Medical_Science": 36.54, |
|
"Finance": 45.0, |
|
"Computer_Science": 38.33, |
|
"Math": 23.33, |
|
"Music": 28.33, |
|
"Marketing": 32.2, |
|
"Materials": 15.0, |
|
"Chemistry": 36.67, |
|
"Geography": 28.85 |
|
}, |
|
"acc_stderr": 0, |
|
"acc": 34.91 |
|
}, |
|
"MmvetV2": { |
|
"reject_info": { |
|
"reject_rate": 1.35, |
|
"reject_number": 7, |
|
"total_question": 517 |
|
}, |
|
"accuracy": 61.7843, |
|
"capability_scores": { |
|
"ocr": 68.5365853658537, |
|
"math": 69.11764705882352, |
|
"spat": 57.70408163265303, |
|
"rec": 57.30864197530875, |
|
"know": 57.56410256410256, |
|
"gen": 59.22509225092255, |
|
"seq": 44.08163265306123 |
|
}, |
|
"capability_detail_scores": { |
|
"ocr_math": 86.36363636363636, |
|
"spat_ocr_math": 70.0, |
|
"spat_rec_ocr_math": 45.0, |
|
"spat_rec": 56.07142857142857, |
|
"spat_ocr": 77.3076923076923, |
|
"spat_rec_ocr": 44.166666666666664, |
|
"spat_know_ocr": 100.0, |
|
"rec_ocr": 75.0, |
|
"spat_know_rec": 40.0, |
|
"ocr": 83.125, |
|
"rec": 67.28813559322033, |
|
"know_rec": 57.692307692307686, |
|
"know_rec_gen": 57.999999999999964, |
|
"know_rec_ocr_gen": 67.6923076923077, |
|
"spat_rec_ocr_gen": 61.86046511627909, |
|
"spat_ocr_gen": 80.0, |
|
"seq_gen_math_spat_ocr": 20.0, |
|
"seq_math_spat_ocr_rec": 0.0, |
|
"spat_rec_gen": 51.81818181818182, |
|
"spat_gen_ocr_math": 40.0, |
|
"spat_rec_seq": 38.33333333333333, |
|
"spat_rec_ocr_seq": 0.0, |
|
"spat_know_rec_gen": 40.00000000000001, |
|
"rec_gen": 60.882352941176464, |
|
"spat_know_rec_ocr": 12.5, |
|
"know_gen_spat_ocr_rec": 90.0, |
|
"rec_ocr_math": 100.0, |
|
"rec_ocr_gen": 72.0, |
|
"rec_ocr_seq_gen": 45.0, |
|
"ocr_gen": 77.6923076923077, |
|
"rec_seq_gen": 36.15384615384615, |
|
"rec_seq": 50.0, |
|
"spat_rec_seq_gen": 63.74999999999999, |
|
"know_rec_seq": 0.0, |
|
"know_rec_seq_gen": 65.0, |
|
"seq_gen_spat_ocr_rec": 50.0, |
|
"know_seq_gen_ocr_rec": 85.00000000000001, |
|
"know_rec_math": 50.0, |
|
"rec_ocr_seq": 100.0 |
|
}, |
|
"acc_stderr": 0, |
|
"acc": 61.7843 |
|
}, |
|
"MathVerse": { |
|
"Vision Intensive": { |
|
"accuracy": 42.64, |
|
"correct": 336, |
|
"total": 788 |
|
}, |
|
"Total": { |
|
"accuracy": 45.38, |
|
"correct": 1788, |
|
"total": 3940 |
|
}, |
|
"Text Dominant": { |
|
"accuracy": 53.81, |
|
"correct": 424, |
|
"total": 788 |
|
}, |
|
"Text Lite": { |
|
"accuracy": 47.34, |
|
"correct": 373, |
|
"total": 788 |
|
}, |
|
"Vision Dominant": { |
|
"accuracy": 45.43, |
|
"correct": 358, |
|
"total": 788 |
|
}, |
|
"Vision Only": { |
|
"accuracy": 37.69, |
|
"correct": 297, |
|
"total": 788 |
|
}, |
|
"accuracy": 45.38, |
|
"acc_stderr": 0, |
|
"acc": 45.38 |
|
}, |
|
"Ocrlite": { |
|
"final_score": [ |
|
1247, |
|
1644 |
|
], |
|
"accuracy": 75.852, |
|
"Key Information Extraction-Bookshelf": [ |
|
26, |
|
51, |
|
0.51, |
|
{ |
|
"Default": [ |
|
26, |
|
51, |
|
0.51 |
|
] |
|
} |
|
], |
|
"Scene Text-centric VQA-diet_constraints": [ |
|
52, |
|
90, |
|
0.578, |
|
{ |
|
"Default": [ |
|
52, |
|
90, |
|
0.578 |
|
] |
|
} |
|
], |
|
"Doc-oriented VQA-Control": [ |
|
142, |
|
189, |
|
0.751, |
|
{ |
|
"Default": [ |
|
142, |
|
189, |
|
0.751 |
|
] |
|
} |
|
], |
|
"Doc-oriented VQA": [ |
|
171, |
|
204, |
|
0.838, |
|
{ |
|
"Default": [ |
|
171, |
|
204, |
|
0.838 |
|
] |
|
} |
|
], |
|
"Scene Text-centric VQA-Fake_logo": [ |
|
54, |
|
119, |
|
0.454, |
|
{ |
|
"Default": [ |
|
54, |
|
119, |
|
0.454 |
|
] |
|
} |
|
], |
|
"Handwritten Mathematical Expression Recognition": [ |
|
1, |
|
100, |
|
0.01, |
|
{ |
|
"Default": [ |
|
1, |
|
100, |
|
0.01 |
|
] |
|
} |
|
], |
|
"Key Information Extraction": [ |
|
198, |
|
209, |
|
0.947, |
|
{ |
|
"Default": [ |
|
198, |
|
209, |
|
0.947 |
|
] |
|
} |
|
], |
|
"Scene Text-centric VQA-Control": [ |
|
173, |
|
200, |
|
0.865, |
|
{ |
|
"Default": [ |
|
173, |
|
200, |
|
0.865 |
|
] |
|
} |
|
], |
|
"Scene Text-centric VQA": [ |
|
247, |
|
282, |
|
0.876, |
|
{ |
|
"Default": [ |
|
247, |
|
282, |
|
0.876 |
|
] |
|
} |
|
], |
|
"Artistic Text Recognition": [ |
|
42, |
|
50, |
|
0.84, |
|
{ |
|
"Default": [ |
|
42, |
|
50, |
|
0.84 |
|
] |
|
} |
|
], |
|
"Irregular Text Recognition": [ |
|
47, |
|
50, |
|
0.94, |
|
{ |
|
"Default": [ |
|
47, |
|
50, |
|
0.94 |
|
] |
|
} |
|
], |
|
"Non-Semantic Text Recognition": [ |
|
45, |
|
50, |
|
0.9, |
|
{ |
|
"Default": [ |
|
45, |
|
50, |
|
0.9 |
|
] |
|
} |
|
], |
|
"Regular Text Recognition": [ |
|
49, |
|
50, |
|
0.98, |
|
{ |
|
"Default": [ |
|
49, |
|
50, |
|
0.98 |
|
] |
|
} |
|
], |
|
"acc_stderr": 0, |
|
"acc": 75.852 |
|
}, |
|
"OcrliteZh": { |
|
"final_score": [ |
|
161, |
|
234 |
|
], |
|
"accuracy": 68.803, |
|
"Docvqa": [ |
|
6, |
|
10, |
|
0.6, |
|
{ |
|
"Default": [ |
|
6, |
|
10, |
|
0.6 |
|
] |
|
} |
|
], |
|
"Chartqa-human": [ |
|
4, |
|
10, |
|
0.4, |
|
{ |
|
"Default": [ |
|
4, |
|
10, |
|
0.4 |
|
] |
|
} |
|
], |
|
"Chartqa-au": [ |
|
7, |
|
10, |
|
0.7, |
|
{ |
|
"Default": [ |
|
7, |
|
10, |
|
0.7 |
|
] |
|
} |
|
], |
|
"infographic": [ |
|
7, |
|
10, |
|
0.7, |
|
{ |
|
"Default": [ |
|
7, |
|
10, |
|
0.7 |
|
] |
|
} |
|
], |
|
"Key Information Extraction": [ |
|
39, |
|
45, |
|
0.867, |
|
{ |
|
"Default": [ |
|
39, |
|
45, |
|
0.867 |
|
] |
|
} |
|
], |
|
"Scene Text-centric VQA": [ |
|
28, |
|
40, |
|
0.7, |
|
{ |
|
"Default": [ |
|
28, |
|
40, |
|
0.7 |
|
] |
|
} |
|
], |
|
"Artistic Text Recognition": [ |
|
7, |
|
11, |
|
0.636, |
|
{ |
|
"Default": [ |
|
7, |
|
11, |
|
0.636 |
|
] |
|
} |
|
], |
|
"IrRegular Text Recognition": [ |
|
8, |
|
11, |
|
0.727, |
|
{ |
|
"Default": [ |
|
8, |
|
11, |
|
0.727 |
|
] |
|
} |
|
], |
|
"Non-semantic Text Recognition": [ |
|
10, |
|
12, |
|
0.833, |
|
{ |
|
"Default": [ |
|
10, |
|
12, |
|
0.833 |
|
] |
|
} |
|
], |
|
"Regular Text Recognition": [ |
|
10, |
|
11, |
|
0.909, |
|
{ |
|
"Default": [ |
|
10, |
|
11, |
|
0.909 |
|
] |
|
} |
|
], |
|
"Handwriting_CN": [ |
|
16, |
|
20, |
|
0.8, |
|
{ |
|
"Default": [ |
|
16, |
|
20, |
|
0.8 |
|
] |
|
} |
|
], |
|
"Chinese Unlimited": [ |
|
19, |
|
44, |
|
0.432, |
|
{ |
|
"Default": [ |
|
19, |
|
44, |
|
0.432 |
|
] |
|
} |
|
], |
|
"acc_stderr": 0, |
|
"acc": 68.803 |
|
}, |
|
"CharXiv": { |
|
"descriptive": { |
|
"Overall Score": 38.55, |
|
"By Question": { |
|
"Q1": 83.2, |
|
"Q2": 77.83, |
|
"Q3": 69.1, |
|
"Q4": 85.99, |
|
"Q5": 81.17, |
|
"Q6": 75.1, |
|
"Q7": 76.5, |
|
"Q8": 62.95, |
|
"Q9": 38.31, |
|
"Q10": 0.0, |
|
"Q11": 0.0, |
|
"Q12": 0.0, |
|
"Q13": 0.0, |
|
"Q14": 0.0, |
|
"Q15": 0.0, |
|
"Q16": 0.0, |
|
"Q17": 0.0, |
|
"Q18": 0.0, |
|
"Q19": 0.0 |
|
}, |
|
"By Category": { |
|
"Information Extraction": 78.53, |
|
"Enumeration": 17.59, |
|
"Pattern Recognition": 0.0, |
|
"Counting": 0.0, |
|
"Compositionality": 0.0 |
|
}, |
|
"By Subplot": { |
|
"1 Subplot": 41.77, |
|
"2-4 Subplots": 37.1, |
|
"5+ Subplots": 35.59 |
|
}, |
|
"By Subject": { |
|
"Computer Science": 38.49, |
|
"Economics": 38.59, |
|
"Electrical Engineering and Systems Science": 44.12, |
|
"Mathematics": 37.04, |
|
"Physics": 38.39, |
|
"Quantitative Biology": 31.94, |
|
"Quantitative Finance": 38.36, |
|
"Statistics": 42.26 |
|
}, |
|
"By Year": { |
|
"2020": 38.46, |
|
"2021": 39.08, |
|
"2022": 39.34, |
|
"2023": 37.3 |
|
}, |
|
"N_valid": 4000, |
|
"N_invalid": 1975, |
|
"Question Type": "Descriptive" |
|
}, |
|
"reasoning": { |
|
"Overall Score": 37.7, |
|
"By Answer Type": { |
|
"Text-in-Chart": 38.18, |
|
"Text-in-General": 41.41, |
|
"Number-in-Chart": 40.52, |
|
"Number-in-General": 32.31 |
|
}, |
|
"By Source": { |
|
"GPT-Sourced": 40.22, |
|
"GPT-Inspired": 37.96, |
|
"Completely Human": 36.83 |
|
}, |
|
"By Subject": { |
|
"Computer Science": 26.19, |
|
"Economics": 39.13, |
|
"Electrical Engineering and Systems Science": 39.5, |
|
"Mathematics": 41.48, |
|
"Physics": 42.52, |
|
"Quantitative Biology": 39.68, |
|
"Quantitative Finance": 36.21, |
|
"Statistics": 36.28 |
|
}, |
|
"By Year": { |
|
"2020": 30.77, |
|
"2021": 41.76, |
|
"2022": 36.89, |
|
"2023": 41.13 |
|
}, |
|
"By Subplot": { |
|
"1 Subplot": 39.9, |
|
"2-4 Subplots": 37.04, |
|
"5+ Subplots": 35.17 |
|
}, |
|
"N_valid": 1000, |
|
"N_invalid": 34, |
|
"Question Type": "Reasoning" |
|
}, |
|
"accuracy": 38.12, |
|
"acc_stderr": 0, |
|
"acc": 38.12 |
|
}, |
|
"MathVision": { |
|
"accuracy": 18.65, |
|
"acc_stderr": 0, |
|
"acc": 18.65 |
|
}, |
|
"CII-Bench": { |
|
"accuracy": 48.89, |
|
"domain_score": { |
|
"CTC": 52.59, |
|
"Society": 50.81, |
|
"Life": 41.56, |
|
"Art": 47.79, |
|
"Env.": 61.11, |
|
"Politics": 62.5 |
|
}, |
|
"emotion_score": { |
|
"Positive": 49.57, |
|
"Negative": 48.3, |
|
"Neutral": 48.87 |
|
}, |
|
"acc_stderr": 0, |
|
"acc": 48.89 |
|
}, |
|
"Blink": { |
|
"accuracy": 56.08, |
|
"Art Style": 59.83, |
|
"Counting": 63.33, |
|
"Forensic Detection": 58.33, |
|
"Functional Correspondence": 31.54, |
|
"IQ Test": 24.0, |
|
"Jigsaw": 59.33, |
|
"Multi-view Reasoning": 54.89, |
|
"Object Localization": 55.74, |
|
"Relative Depth": 76.61, |
|
"Relative Reflectance": 27.61, |
|
"Semantic Correspondence": 39.57, |
|
"Spatial Relation": 80.42, |
|
"Visual Correspondence": 68.02, |
|
"Visual Similarity": 86.67, |
|
"acc_stderr": 0, |
|
"acc": 56.08 |
|
} |
|
} |
|
} |