{ "config_general": { "model_name": "InternVL2-2B", "model_dtype": "float16", "model_size": 0 }, "results": { "CMMMU": { "艺术与设计": { "num": 88, "correct": 35, "accuracy": 39.77 }, "overall": { "num": 900, "correct": 263, "accuracy": 29.22 }, "商业": { "num": 126, "correct": 22, "accuracy": 17.46 }, "科学": { "num": 204, "correct": 43, "accuracy": 21.08 }, "健康与医学": { "num": 153, "correct": 52, "accuracy": 33.99 }, "人文社会科学": { "num": 85, "correct": 37, "accuracy": 43.53 }, "技术与工程": { "num": 244, "correct": 74, "accuracy": 30.33 }, "accuracy": 29.22, "acc_stderr": 0, "acc": 29.22 }, "MMMU": { "accuracy": 32.89, "acc_stderr": 0, "acc": 32.89 }, "MMMU_Pro_standard": { "accuracy": 20.4, "acc_stderr": 0, "acc": 20.4 }, "MMMU_Pro_vision": { "accuracy": 12.2, "subject_score": { "Finance": 6.67, "Agriculture": 11.67, "Design": 16.67, "Art": 9.43, "Literature": 19.23, "Sociology": 14.81, "History": 8.93, "Pharmacy": 22.81, "Clinical_Medicine": 8.47, "Energy_and_Power": 5.17, "Accounting": 12.07, "Public_Health": 5.17, "Architecture_and_Engineering": 11.67, "Psychology": 10.0, "Art_Theory": 7.27, "Electronics": 8.33, "Physics": 11.67, "Biology": 16.95, "Mechanical_Engineering": 18.64, "Manage": 20.0, "Economics": 8.47, "Basic_Medical_Science": 23.08, "Diagnostics_and_Laboratory_Medicine": 13.33, "Computer_Science": 11.67, "Math": 8.33, "Music": 11.67, "Marketing": 6.78, "Materials": 16.67, "Chemistry": 10.0, "Geography": 13.46 }, "acc_stderr": 0, "acc": 12.2 }, "MmvetV2": { "accuracy": 39.4584, "capability_scores": { "ocr": 40.14423076923076, "math": 21.764705882352942, "spat": 34.82233502538071, "rec": 37.93689320388349, "know": 34.615384615384606, "gen": 37.92727272727271, "seq": 40.178571428571445 }, "capability_detail_scores": { "ocr_math": 22.727272727272727, "ocr_math_spat": 16.666666666666664, "math_ocr_rec_spat": 0.0, "rec_spat": 46.07142857142858, "ocr_spat": 48.07692307692308, "ocr_rec_spat": 8.333333333333332, "ocr_know_spat": 42.5, "ocr_rec": 62.5, "know_rec_spat": 30.0, "ocr": 65.0, "rec": 47.96610169491527, "know_rec": 14.615384615384613, "know_gen_rec": 35.99999999999999, "ocr_know_gen_rec": 45.38461538461539, "ocr_gen_rec_spat": 33.25581395348837, "ocr_gen_spat": 70.0, "gen_seq_spat_math_ocr": 0.0, "seq_spat_math_rec_ocr": 0.0, "gen_rec_spat": 20.45454545454546, "ocr_gen_math_spat": 40.0, "seq_rec_spat": 48.57142857142858, "ocr_seq_rec_spat": 50.0, "know_gen_rec_spat": 33.33333333333333, "gen_rec": 47.94117647058824, "ocr_know_rec_spat": 12.5, "gen_spat_know_rec_ocr": 75.0, "math_ocr_rec": 100.0, "ocr_gen_rec": 44.00000000000001, "ocr_gen_rec_seq": 45.714285714285715, "ocr_gen": 45.38461538461538, "seq_gen_rec": 26.42857142857143, "seq_rec": 44.99999999999999, "seq_gen_rec_spat": 61.24999999999999, "know_seq_rec": 0.0, "gen_seq_rec": 26.42857142857143, "seq_know_gen_rec": 45.0, "gen_seq_spat_rec_ocr": 53.333333333333336, "gen_seq_know_rec_ocr": 30.0, "math_know_rec": 50.0, "ocr_seq_rec": 0.0 }, "acc_stderr": 0, "acc": 39.4584 }, "MathVerse": { "Text Lite": { "accuracy": 23.73, "correct": 187, "total": 788 }, "Total": { "accuracy": 21.62, "correct": 852, "total": 3940 }, "Vision Intensive": { "accuracy": 24.11, "correct": 190, "total": 788 }, "Text Dominant": { "accuracy": 24.75, "correct": 195, "total": 788 }, "Vision Dominant": { "accuracy": 20.43, "correct": 161, "total": 788 }, "Vision Only": { "accuracy": 15.1, "correct": 119, "total": 788 }, "accuracy": 21.62, "acc_stderr": 0, "acc": 21.62 }, "Ocrlite": { "final_score": [ 944, 1645 ], "accuracy": 57.386, "Key Information Extraction-Bookshelf": [ 8, 52 ], "Scene Text-centric VQA-diet_constraints": [ 20, 90 ], "Doc-oriented VQA-Control": [ 98, 189 ], "Doc-oriented VQA": [ 88, 204 ], "Scene Text-centric VQA-Fake_logo": [ 53, 119 ], "Handwritten Mathematical Expression Recognition": [ 40, 100 ], "Key Information Extraction": [ 152, 209 ], "Scene Text-centric VQA-Control": [ 155, 200 ], "Scene Text-centric VQA": [ 170, 282 ], "Artistic Text Recognition": [ 38, 50 ], "Irregular Text Recognition": [ 36, 50 ], "Non-Semantic Text Recognition": [ 40, 50 ], "Regular Text Recognition": [ 46, 50 ], "acc_stderr": 0, "acc": 57.386 }, "OcrliteZh": { "final_score": [ 99, 234 ], "accuracy": 42.308, "Docvqa": [ 2, 10 ], "Chartqa-human": [ 3, 10 ], "Chartqa-au": [ 3, 10 ], "infographic": [ 3, 10 ], "Key Information Extraction": [ 24, 45 ], "Scene Text-centric VQA": [ 16, 40 ], "Artistic Text Recognition": [ 2, 11 ], "IrRegular Text Recognition": [ 5, 11 ], "Non-semantic Text Recognition": [ 10, 12 ], "Regular Text Recognition": [ 8, 11 ], "Handwriting_CN": [ 11, 20 ], "Chinese Unlimited": [ 12, 44 ], "acc_stderr": 0, "acc": 42.308 }, "CharXiv": { "descriptive": { "Overall Score": 36.6, "By Question": { "Q1": 52.05, "Q2": 62.61, "Q3": 38.2, "Q4": 54.47, "Q5": 43.51, "Q6": 28.92, "Q7": 32.91, "Q8": 31.7, "Q9": 26.87, "Q10": 40.41, "Q11": 21.14, "Q12": 47.8, "Q13": 37.9, "Q14": 42.91, "Q15": 14.38, "Q16": 36.11, "Q17": 4.46, "Q18": 42.51, "Q19": 40.0 }, "By Category": { "Information Extraction": 44.66, "Enumeration": 30.19, "Pattern Recognition": 33.84, "Counting": 43.77, "Compositionality": 4.46 }, "By Subplot": { "1 Subplot": 48.83, "2-4 Subplots": 31.81, "5+ Subplots": 24.26 }, "By Subject": { "Computer Science": 38.89, "Economics": 37.68, "Electrical Engineering and Systems Science": 41.18, "Mathematics": 37.78, "Physics": 33.27, "Quantitative Biology": 34.33, "Quantitative Finance": 33.41, "Statistics": 36.06 }, "By Year": { "2020": 37.96, "2021": 34.77, "2022": 37.6, "2023": 36.19 }, "N_valid": 4000, "N_invalid": 0, "Question Type": "Descriptive" }, "reasoning": { "Overall Score": 20.1, "By Answer Type": { "Text-in-Chart": 22.95, "Text-in-General": 24.24, "Number-in-Chart": 22.84, "Number-in-General": 10.04 }, "By Source": { "GPT-Sourced": 25.54, "GPT-Inspired": 18.52, "Completely Human": 19.0 }, "By Subject": { "Computer Science": 21.43, "Economics": 18.12, "Electrical Engineering and Systems Science": 21.01, "Mathematics": 22.22, "Physics": 18.9, "Quantitative Biology": 22.22, "Quantitative Finance": 18.97, "Statistics": 17.7 }, "By Year": { "2020": 19.03, "2021": 22.61, "2022": 16.8, "2023": 21.77 }, "By Subplot": { "1 Subplot": 20.98, "2-4 Subplots": 21.96, "5+ Subplots": 15.68 }, "N_valid": 1000, "N_invalid": 0, "Question Type": "Reasoning" }, "accuracy": 28.35, "acc_stderr": 0, "acc": 28.35 }, "MathVision": { "accuracy": 13.88, "acc_stderr": 0, "acc": 13.88 }, "CII-Bench": { "accuracy": 38.95, "domain_score": { "Life": 29.0, "CTC": 40.74, "Art": 42.65, "Society": 41.62, "Env.": 55.56, "Politics": 45.83 }, "emotion_score": { "Neutral": 42.11, "Positive": 39.32, "Negative": 35.47 }, "acc_stderr": 0, "acc": 38.95 }, "Blink": { "accuracy": 40.35, "Art Style": 52.99, "Counting": 46.67, "Forensic Detection": 18.94, "Functional Correspondence": 18.46, "IQ Test": 28.67, "Jigsaw": 42.0, "Multi-view Reasoning": 46.62, "Object Localization": 45.08, "Relative Depth": 51.61, "Relative Reflectance": 30.6, "Semantic Correspondence": 28.06, "Spatial Relation": 72.73, "Visual Correspondence": 33.14, "Visual Similarity": 53.33, "acc_stderr": 0, "acc": 40.35 } } }