Datasets:

License:
vlm_results / GLM-4V-Plus /results_2025-01-25T10-42-53.190540.json
daiteng01's picture
Upload 52 files
7498b1c verified
{
"config_general": {
"model_name": "GLM-4V-Plus",
"model_dtype": "float16",
"model_size": 0
},
"results": {
"CMMMU": {
"accuracy": 43.56,
"acc_stderr": 0,
"acc": 43.56
},
"MMMU": {
"accuracy": 54.44,
"acc_stderr": 0,
"acc": 54.44
},
"MMMU_Pro_standard": {
"reject_info": {
"reject_rate": 0.06,
"reject_number": 1,
"total_question": 1730
},
"accuracy": 37.19,
"acc_stderr": 0,
"acc": 37.19
},
"MMMU_Pro_vision": {
"accuracy": 23.35,
"subject_score": {
"History": 35.71,
"Art": 47.17,
"Design": 45.0,
"Literature": 59.62,
"Agriculture": 18.33,
"Finance": 10.0,
"Sociology": 42.59,
"Accounting": 15.52,
"Energy_and_Power": 12.07,
"Pharmacy": 36.84,
"Architecture_and_Engineering": 15.0,
"Clinical_Medicine": 13.56,
"Public_Health": 13.79,
"Physics": 11.67,
"Art_Theory": 47.27,
"Electronics": 8.33,
"Psychology": 33.33,
"Biology": 23.73,
"Manage": 18.0,
"Economics": 6.78,
"Mechanical_Engineering": 10.17,
"Diagnostics_and_Laboratory_Medicine": 25.0,
"Basic_Medical_Science": 32.69,
"Computer_Science": 20.0,
"Math": 20.0,
"Music": 21.67,
"Materials": 8.33,
"Marketing": 16.95,
"Chemistry": 25.0,
"Geography": 17.31
},
"acc_stderr": 0,
"acc": 23.35
},
"MmvetV2": {
"reject_info": {
"reject_rate": 2.9,
"reject_number": 15,
"total_question": 517
},
"accuracy": 56.7131,
"capability_scores": {
"ocr": 62.079207920792065,
"math": 62.05882352941177,
"spat": 51.18556701030926,
"rec": 53.77833753148617,
"know": 54.285714285714285,
"gen": 53.901515151515135,
"seq": 36.09756097560975
},
"capability_detail_scores": {
"ocr_math": 67.27272727272727,
"ocr_spat_math": 68.66666666666667,
"ocr_rec_spat_math": 45.0,
"rec_spat": 48.214285714285715,
"ocr_spat": 59.61538461538461,
"ocr_rec_spat": 45.83333333333333,
"ocr_know_spat": 65.0,
"ocr_rec": 100.0,
"rec_know_spat": 57.99999999999999,
"ocr": 80.9375,
"rec": 64.57627118644068,
"rec_know": 46.92307692307692,
"rec_know_gen": 53.400000000000006,
"ocr_rec_know_gen": 74.61538461538463,
"ocr_rec_spat_gen": 59.76744186046512,
"ocr_spat_gen": 60.0,
"ocr_seq_spat_math_gen": 0.0,
"ocr_seq_spat_rec_math": 0.0,
"rec_spat_gen": 38.18181818181819,
"ocr_spat_math_gen": 50.0,
"rec_spat_seq": 28.333333333333332,
"ocr_rec_spat_seq": 36.66666666666667,
"rec_know_spat_gen": 50.0,
"rec_gen": 59.11764705882353,
"ocr_rec_spat_know": 25.0,
"ocr_spat_know": 65.0,
"ocr_spat_rec_know_gen": 40.0,
"ocr_rec_math": 0.0,
"ocr_rec_gen": 44.00000000000001,
"ocr_rec_gen_seq": 65.0,
"ocr_gen": 60.0,
"rec_gen_seq": 51.818181818181806,
"rec_seq": 0.0,
"rec_spat_gen_seq": 50.0,
"rec_know_seq": 0,
"rec_know_gen_seq": 0.0,
"ocr_seq_spat_rec_gen": 10.0,
"ocr_seq_rec_know_gen": 70.0,
"rec_know_math": 100.0,
"ocr_rec_seq": 100.0,
"rec_spat_know": 57.99999999999999
},
"acc_stderr": 0,
"acc": 56.7131
},
"MathVerse": {
"Text Dominant": {
"accuracy": 33.63,
"correct": 265,
"total": 788
},
"Total": {
"accuracy": 29.44,
"correct": 1160,
"total": 3940
},
"Text Lite": {
"accuracy": 30.33,
"correct": 239,
"total": 788
},
"Vision Intensive": {
"accuracy": 29.19,
"correct": 230,
"total": 788
},
"Vision Dominant": {
"accuracy": 29.31,
"correct": 231,
"total": 788
},
"Vision Only": {
"accuracy": 24.75,
"correct": 195,
"total": 788
},
"accuracy": 29.44,
"acc_stderr": 0,
"acc": 29.44
},
"Ocrlite": {
"final_score": [
1188,
1645
],
"accuracy": 72.219,
"Key Information Extraction-Bookshelf": [
31,
52
],
"Scene Text-centric VQA-diet_constraints": [
69,
90
],
"Doc-oriented VQA-Control": [
138,
189
],
"Doc-oriented VQA": [
144,
204
],
"Scene Text-centric VQA-Fake_logo": [
42,
119
],
"Handwritten Mathematical Expression Recognition": [
49,
100
],
"Key Information Extraction": [
167,
209
],
"Scene Text-centric VQA-Control": [
155,
200
],
"Scene Text-centric VQA": [
228,
282
],
"Artistic Text Recognition": [
39,
50
],
"Irregular Text Recognition": [
40,
50
],
"Non-Semantic Text Recognition": [
38,
50
],
"Regular Text Recognition": [
48,
50
],
"acc_stderr": 0,
"acc": 72.219
},
"OcrliteZh": {
"final_score": [
150,
234
],
"accuracy": 64.103,
"Docvqa": [
7,
10
],
"Chartqa-human": [
3,
10
],
"Chartqa-au": [
8,
10
],
"infographic": [
7,
10
],
"Key Information Extraction": [
36,
45
],
"Scene Text-centric VQA": [
29,
40
],
"Artistic Text Recognition": [
5,
11
],
"IrRegular Text Recognition": [
7,
11
],
"Non-semantic Text Recognition": [
8,
12
],
"Regular Text Recognition": [
10,
11
],
"Handwriting_CN": [
14,
20
],
"Chinese Unlimited": [
16,
44
],
"acc_stderr": 0,
"acc": 64.103
},
"CharXiv": {
"reject_info": {
"reject_rate": 0.06,
"reject_number": 3,
"total_question": 5000
},
"descriptive": {
"Overall Score": 58.06,
"By Question": {
"Q1": 67.49,
"Q2": 81.3,
"Q3": 67.38,
"Q4": 86.38,
"Q5": 79.92,
"Q6": 69.88,
"Q7": 61.97,
"Q8": 67.86,
"Q9": 61.19,
"Q10": 57.53,
"Q11": 32.57,
"Q12": 60.44,
"Q13": 69.86,
"Q14": 15.25,
"Q15": 37.38,
"Q16": 66.67,
"Q17": 8.04,
"Q18": 60.73,
"Q19": 78.46
},
"By Category": {
"Information Extraction": 73.59,
"Enumeration": 47.46,
"Pattern Recognition": 50.44,
"Counting": 62.34,
"Compositionality": 8.04
},
"By Subplot": {
"1 Subplot": 69.43,
"2-4 Subplots": 55.33,
"5+ Subplots": 43.86
},
"By Subject": {
"Computer Science": 56.55,
"Economics": 59.24,
"Electrical Engineering and Systems Science": 68.63,
"Mathematics": 60.19,
"Physics": 54.13,
"Quantitative Biology": 49.6,
"Quantitative Finance": 60.13,
"Statistics": 56.42
},
"By Year": {
"2020": 58.2,
"2021": 57.91,
"2022": 57.48,
"2023": 58.67
},
"N_valid": 3999,
"N_invalid": 0,
"Question Type": "Descriptive"
},
"reasoning": {
"Overall Score": 31.86,
"By Answer Type": {
"Text-in-Chart": 33.49,
"Text-in-General": 37.76,
"Number-in-Chart": 33.19,
"Number-in-General": 24.89
},
"By Source": {
"GPT-Sourced": 33.7,
"GPT-Inspired": 25.93,
"Completely Human": 33.44
},
"By Subject": {
"Computer Science": 35.71,
"Economics": 32.61,
"Electrical Engineering and Systems Science": 32.2,
"Mathematics": 30.37,
"Physics": 37.3,
"Quantitative Biology": 26.98,
"Quantitative Finance": 30.17,
"Statistics": 29.2
},
"By Year": {
"2020": 32.79,
"2021": 34.36,
"2022": 29.1,
"2023": 31.05
},
"By Subplot": {
"1 Subplot": 32.21,
"2-4 Subplots": 35.01,
"5+ Subplots": 26.27
},
"N_valid": 998,
"N_invalid": 1,
"Question Type": "Reasoning"
},
"accuracy": 44.96,
"acc_stderr": 0,
"acc": 44.96
},
"MathVision": {
"accuracy": 17.34,
"acc_stderr": 0,
"acc": 17.34
},
"CII-Bench": {
"accuracy": 60.86,
"domain_score": {
"Life": 55.41,
"Art": 59.56,
"CTC": 62.96,
"Society": 64.67,
"Env.": 66.67,
"Politics": 66.67
},
"emotion_score": {
"Neutral": 60.38,
"Negative": 60.75,
"Positive": 61.54
},
"acc_stderr": 0,
"acc": 60.86
},
"Blink": {
"accuracy": 55.44,
"Art Style": 51.28,
"Counting": 75.0,
"Forensic Detection": 13.64,
"Functional Correspondence": 35.38,
"IQ Test": 24.67,
"Jigsaw": 69.33,
"Multi-view Reasoning": 52.63,
"Object Localization": 63.93,
"Relative Depth": 69.35,
"Relative Reflectance": 28.36,
"Semantic Correspondence": 48.92,
"Spatial Relation": 86.71,
"Visual Correspondence": 72.09,
"Visual Similarity": 82.22,
"acc_stderr": 0,
"acc": 55.44
}
}
}