|
{ |
|
"config_general": { |
|
"model_name": "LLaVA-OneVision-7B", |
|
"model_dtype": "float16", |
|
"model_size": 0 |
|
}, |
|
"results": { |
|
"CMMMU": { |
|
"accuracy": 37.11, |
|
"acc_stderr": 0, |
|
"acc": 37.11 |
|
}, |
|
"MMMU": { |
|
"accuracy": 45.33, |
|
"acc_stderr": 0, |
|
"acc": 45.33 |
|
}, |
|
"MMMU_Pro_standard": { |
|
"accuracy": 28.67, |
|
"acc_stderr": 0, |
|
"acc": 28.67 |
|
}, |
|
"MMMU_Pro_vision": { |
|
"accuracy": 11.39, |
|
"subject_score": { |
|
"Literature": 17.31, |
|
"Sociology": 9.26, |
|
"History": 10.71, |
|
"Finance": 15.0, |
|
"Art": 11.32, |
|
"Design": 8.33, |
|
"Agriculture": 8.33, |
|
"Clinical_Medicine": 13.56, |
|
"Accounting": 8.62, |
|
"Energy_and_Power": 13.79, |
|
"Architecture_and_Engineering": 5.0, |
|
"Electronics": 6.67, |
|
"Art_Theory": 9.09, |
|
"Physics": 13.33, |
|
"Public_Health": 5.17, |
|
"Manage": 22.0, |
|
"Biology": 13.56, |
|
"Psychology": 16.67, |
|
"Economics": 16.95, |
|
"Diagnostics_and_Laboratory_Medicine": 8.33, |
|
"Mechanical_Engineering": 5.08, |
|
"Basic_Medical_Science": 7.69, |
|
"Computer_Science": 11.67, |
|
"Math": 8.33, |
|
"Pharmacy": 15.79, |
|
"Music": 16.67, |
|
"Materials": 5.0, |
|
"Marketing": 15.25, |
|
"Chemistry": 8.33, |
|
"Geography": 17.31 |
|
}, |
|
"acc_stderr": 0, |
|
"acc": 11.39 |
|
}, |
|
"MmvetV2": { |
|
"reject_info": { |
|
"reject_rate": 0.19, |
|
"reject_number": 1, |
|
"total_question": 517 |
|
}, |
|
"accuracy": 44.9806, |
|
"capability_scores": { |
|
"math": 42.94117647058823, |
|
"ocr": 43.07692307692308, |
|
"spat": 39.441624365482234, |
|
"rec": 44.014598540146, |
|
"know": 42.43589743589742, |
|
"gen": 42.51824817518247, |
|
"seq": 30.545454545454554 |
|
}, |
|
"capability_detail_scores": { |
|
"math_ocr": 66.36363636363636, |
|
"spat_math_ocr": 36.0, |
|
"rec_spat_math_ocr": 0.0, |
|
"rec_spat": 51.42857142857142, |
|
"spat_ocr": 43.84615384615384, |
|
"rec_spat_ocr": 12.5, |
|
"know_spat_ocr": 87.5, |
|
"rec_ocr": 67.5, |
|
"rec_know_spat": 30.0, |
|
"ocr": 56.56250000000001, |
|
"rec": 60.84745762711864, |
|
"rec_know": 34.61538461538461, |
|
"rec_know_gen": 43.299999999999976, |
|
"rec_know_gen_ocr": 54.61538461538461, |
|
"rec_gen_spat_ocr": 45.813953488372086, |
|
"gen_spat_ocr": 55.00000000000001, |
|
"gen_spat_seq_math_ocr": 0.0, |
|
"spat_rec_seq_math_ocr": 50.0, |
|
"rec_gen_spat": 33.18181818181819, |
|
"gen_spat_math_ocr": 40.0, |
|
"rec_seq_spat": 40.0, |
|
"rec_seq_spat_ocr": 13.333333333333334, |
|
"rec_know_gen_spat": 23.333333333333332, |
|
"rec_gen": 54.70588235294118, |
|
"rec_know_spat_ocr": 17.5, |
|
"gen_know_spat_rec_ocr": 50.0, |
|
"rec_math_ocr": 0.0, |
|
"rec_gen_ocr": 54.0, |
|
"rec_seq_gen_ocr": 18.571428571428573, |
|
"gen_ocr": 30.76923076923077, |
|
"rec_seq_gen": 30.76923076923077, |
|
"rec_seq": 41.66666666666667, |
|
"rec_seq_gen_spat": 43.75, |
|
"rec_seq_know": 0.0, |
|
"rec_seq_know_gen": 50.0, |
|
"gen_spat_rec_seq_ocr": 13.333333333333334, |
|
"gen_know_rec_seq_ocr": 20.0, |
|
"rec_know_math": 50.0, |
|
"rec_seq_ocr": 0.0 |
|
}, |
|
"acc_stderr": 0, |
|
"acc": 44.9806 |
|
}, |
|
"MathVerse": { |
|
"Vision Intensive": { |
|
"accuracy": 30.46, |
|
"correct": 240, |
|
"total": 788 |
|
}, |
|
"Total": { |
|
"accuracy": 27.84, |
|
"correct": 1097, |
|
"total": 3940 |
|
}, |
|
"Text Lite": { |
|
"accuracy": 31.98, |
|
"correct": 252, |
|
"total": 788 |
|
}, |
|
"Vision Dominant": { |
|
"accuracy": 25.51, |
|
"correct": 201, |
|
"total": 788 |
|
}, |
|
"Vision Only": { |
|
"accuracy": 13.96, |
|
"correct": 110, |
|
"total": 788 |
|
}, |
|
"Text Dominant": { |
|
"accuracy": 37.31, |
|
"correct": 294, |
|
"total": 788 |
|
}, |
|
"accuracy": 27.84, |
|
"acc_stderr": 0, |
|
"acc": 27.84 |
|
}, |
|
"Ocrlite": { |
|
"reject_info": { |
|
"reject_rate": 0.06, |
|
"reject_number": 1, |
|
"total_question": 1645 |
|
}, |
|
"final_score": [ |
|
1042, |
|
1644 |
|
], |
|
"accuracy": 63.382, |
|
"Key Information Extraction-Bookshelf": [ |
|
13, |
|
52 |
|
], |
|
"Scene Text-centric VQA-diet_constraints": [ |
|
39, |
|
89 |
|
], |
|
"Doc-oriented VQA-Control": [ |
|
124, |
|
189 |
|
], |
|
"Doc-oriented VQA": [ |
|
118, |
|
204 |
|
], |
|
"Scene Text-centric VQA-Fake_logo": [ |
|
43, |
|
119 |
|
], |
|
"Handwritten Mathematical Expression Recognition": [ |
|
19, |
|
100 |
|
], |
|
"Key Information Extraction": [ |
|
166, |
|
209 |
|
], |
|
"Scene Text-centric VQA-Control": [ |
|
162, |
|
200 |
|
], |
|
"Scene Text-centric VQA": [ |
|
218, |
|
282 |
|
], |
|
"Artistic Text Recognition": [ |
|
36, |
|
50 |
|
], |
|
"Irregular Text Recognition": [ |
|
38, |
|
50 |
|
], |
|
"Non-Semantic Text Recognition": [ |
|
20, |
|
50 |
|
], |
|
"Regular Text Recognition": [ |
|
46, |
|
50 |
|
], |
|
"acc_stderr": 0, |
|
"acc": 63.382 |
|
}, |
|
"OcrliteZh": { |
|
"final_score": [ |
|
42, |
|
234 |
|
], |
|
"accuracy": 17.949, |
|
"Docvqa": [ |
|
1, |
|
10 |
|
], |
|
"Chartqa-human": [ |
|
2, |
|
10 |
|
], |
|
"Chartqa-au": [ |
|
1, |
|
10 |
|
], |
|
"infographic": [ |
|
3, |
|
10 |
|
], |
|
"Key Information Extraction": [ |
|
11, |
|
45 |
|
], |
|
"Scene Text-centric VQA": [ |
|
10, |
|
40 |
|
], |
|
"Artistic Text Recognition": [ |
|
0, |
|
11 |
|
], |
|
"IrRegular Text Recognition": [ |
|
1, |
|
11 |
|
], |
|
"Non-semantic Text Recognition": [ |
|
0, |
|
12 |
|
], |
|
"Regular Text Recognition": [ |
|
0, |
|
11 |
|
], |
|
"Handwriting_CN": [ |
|
2, |
|
20 |
|
], |
|
"Chinese Unlimited": [ |
|
11, |
|
44 |
|
], |
|
"acc_stderr": 0, |
|
"acc": 17.949 |
|
}, |
|
"CharXiv": { |
|
"reject_info": { |
|
"reject_rate": 0.02, |
|
"reject_number": 1, |
|
"total_question": 5000 |
|
}, |
|
"descriptive": { |
|
"Overall Score": 39.08, |
|
"By Question": { |
|
"Q1": 34.43, |
|
"Q2": 35.65, |
|
"Q3": 24.03, |
|
"Q4": 57.2, |
|
"Q5": 63.6, |
|
"Q6": 41.37, |
|
"Q7": 50.85, |
|
"Q8": 54.46, |
|
"Q9": 50.25, |
|
"Q10": 30.14, |
|
"Q11": 42.29, |
|
"Q12": 29.67, |
|
"Q13": 27.06, |
|
"Q14": 27.3, |
|
"Q15": 21.41, |
|
"Q16": 63.89, |
|
"Q17": 7.59, |
|
"Q18": 53.85, |
|
"Q19": 75.38 |
|
}, |
|
"By Category": { |
|
"Information Extraction": 44.07, |
|
"Enumeration": 34.41, |
|
"Pattern Recognition": 50.22, |
|
"Counting": 37.4, |
|
"Compositionality": 7.59 |
|
}, |
|
"By Subplot": { |
|
"1 Subplot": 46.05, |
|
"2-4 Subplots": 37.37, |
|
"5+ Subplots": 30.43 |
|
}, |
|
"By Subject": { |
|
"Computer Science": 38.69, |
|
"Economics": 39.86, |
|
"Electrical Engineering and Systems Science": 43.07, |
|
"Mathematics": 44.44, |
|
"Physics": 36.61, |
|
"Quantitative Biology": 34.19, |
|
"Quantitative Finance": 37.93, |
|
"Statistics": 37.39 |
|
}, |
|
"By Year": { |
|
"2020": 41.34, |
|
"2021": 38.31, |
|
"2022": 38.42, |
|
"2023": 38.31 |
|
}, |
|
"N_valid": 3999, |
|
"N_invalid": 50, |
|
"Question Type": "Descriptive" |
|
}, |
|
"reasoning": { |
|
"Overall Score": 20.0, |
|
"By Answer Type": { |
|
"Text-in-Chart": 18.18, |
|
"Text-in-General": 34.34, |
|
"Number-in-Chart": 24.57, |
|
"Number-in-General": 12.66 |
|
}, |
|
"By Source": { |
|
"GPT-Sourced": 17.39, |
|
"GPT-Inspired": 19.44, |
|
"Completely Human": 21.0 |
|
}, |
|
"By Subject": { |
|
"Computer Science": 16.67, |
|
"Economics": 19.57, |
|
"Electrical Engineering and Systems Science": 21.85, |
|
"Mathematics": 22.22, |
|
"Physics": 23.62, |
|
"Quantitative Biology": 18.25, |
|
"Quantitative Finance": 14.66, |
|
"Statistics": 23.01 |
|
}, |
|
"By Year": { |
|
"2020": 20.65, |
|
"2021": 25.67, |
|
"2022": 12.3, |
|
"2023": 20.97 |
|
}, |
|
"By Subplot": { |
|
"1 Subplot": 20.47, |
|
"2-4 Subplots": 19.84, |
|
"5+ Subplots": 19.49 |
|
}, |
|
"N_valid": 1000, |
|
"N_invalid": 1, |
|
"Question Type": "Reasoning" |
|
}, |
|
"accuracy": 29.54, |
|
"acc_stderr": 0, |
|
"acc": 29.54 |
|
}, |
|
"MathVision": { |
|
"accuracy": 16.68, |
|
"acc_stderr": 0, |
|
"acc": 16.68 |
|
}, |
|
"CII-Bench": { |
|
"accuracy": 42.88, |
|
"domain_score": { |
|
"CTC": 42.22, |
|
"Society": 49.19, |
|
"Art": 43.38, |
|
"Life": 34.2, |
|
"Env.": 53.7, |
|
"Politics": 54.17 |
|
}, |
|
"emotion_score": { |
|
"Positive": 37.18, |
|
"Negative": 46.04, |
|
"Neutral": 44.74 |
|
}, |
|
"acc_stderr": 0, |
|
"acc": 42.88 |
|
}, |
|
"Blink": { |
|
"accuracy": 49.03, |
|
"Art Style": 58.12, |
|
"Counting": 65.83, |
|
"Forensic Detection": 25.0, |
|
"Functional Correspondence": 33.08, |
|
"IQ Test": 26.67, |
|
"Jigsaw": 51.33, |
|
"Multi-view Reasoning": 50.38, |
|
"Object Localization": 56.56, |
|
"Relative Depth": 75.81, |
|
"Relative Reflectance": 23.88, |
|
"Semantic Correspondence": 32.37, |
|
"Spatial Relation": 79.72, |
|
"Visual Correspondence": 38.37, |
|
"Visual Similarity": 77.78, |
|
"acc_stderr": 0, |
|
"acc": 49.03 |
|
} |
|
} |
|
} |