{ "config_general": { "model_name": "LLaVA-Onevision-72B", "model_dtype": "float16", "model_size": 0 }, "results": { "CMMMU": { "艺术与设计": { "num": 88, "correct": 66, "accuracy": 75.0 }, "overall": { "num": 900, "correct": 430, "accuracy": 47.78 }, "商业": { "num": 126, "correct": 35, "accuracy": 27.78 }, "科学": { "num": 204, "correct": 80, "accuracy": 39.22 }, "健康与医学": { "num": 153, "correct": 86, "accuracy": 56.21 }, "人文社会科学": { "num": 85, "correct": 50, "accuracy": 58.82 }, "技术与工程": { "num": 244, "correct": 113, "accuracy": 46.31 }, "accuracy": 47.78, "acc_stderr": 0, "acc": 47.78 }, "MMMU": { "accuracy": 56.0, "acc_stderr": 0, "acc": 56.0 }, "MMMU_Pro_standard": { "reject_info": { "reject_rate": 0.12, "reject_number": 2, "total_question": 1730 }, "accuracy": 37.21, "acc_stderr": 0, "acc": 37.21 }, "MMMU_Pro_vision": { "reject_info": { "reject_rate": 0.12, "reject_number": 2, "total_question": 1730 }, "accuracy": 23.9, "subject_score": { "History": 26.79, "Literature": 55.77, "Sociology": 35.19, "Art": 28.3, "Agriculture": 16.67, "Design": 35.0, "Pharmacy": 38.6, "Energy_and_Power": 15.52, "Architecture_and_Engineering": 11.67, "Art_Theory": 38.18, "Electronics": 23.33, "Accounting": 21.43, "Psychology": 16.67, "Biology": 16.95, "Manage": 20.0, "Public_Health": 17.24, "Economics": 28.81, "Diagnostics_and_Laboratory_Medicine": 11.67, "Clinical_Medicine": 18.64, "Physics": 18.33, "Mechanical_Engineering": 25.42, "Finance": 26.67, "Computer_Science": 30.0, "Math": 20.0, "Basic_Medical_Science": 17.31, "Marketing": 22.03, "Music": 25.0, "Materials": 21.67, "Chemistry": 16.67, "Geography": 23.08 }, "acc_stderr": 0, "acc": 23.9 }, "MmvetV2": { "reject_info": { "reject_rate": 0.19, "reject_number": 1, "total_question": 517 }, "accuracy": 58.9147, "capability_scores": { "math": 61.1764705882353, "ocr": 62.06730769230771, "spat": 55.025380710659874, "rec": 56.64233576642339, "know": 57.179487179487154, "gen": 59.34306569343065, "seq": 48.545454545454554 }, "capability_detail_scores": { "math_ocr": 78.18181818181817, "spat_math_ocr": 56.666666666666664, "spat_math_ocr_rec": 40.0, "spat_rec": 63.57142857142858, "spat_ocr": 61.92307692307693, "spat_ocr_rec": 23.333333333333332, "spat_know_ocr": 75.0, "ocr_rec": 82.5, "spat_know_rec": 50.0, "ocr": 75.62500000000001, "rec": 57.28813559322033, "know_rec": 70.76923076923076, "know_gen_rec": 58.19999999999995, "know_ocr_gen_rec": 67.6923076923077, "spat_ocr_gen_rec": 62.7906976744186, "spat_ocr_gen": 80.0, "spat_math_ocr_gen_seq": 100.0, "spat_math_ocr_rec_seq": 0.0, "spat_gen_rec": 50.90909090909093, "spat_math_ocr_gen": 40.0, "spat_seq_rec": 28.57142857142857, "spat_ocr_rec_seq": 40.0, "spat_know_gen_rec": 36.66666666666667, "gen_rec": 64.70588235294117, "spat_know_ocr_rec": 0.0, "know_spat_ocr_gen_rec": 75.0, "math_ocr_rec": 50.0, "ocr_gen_rec": 61.999999999999986, "ocr_gen_rec_seq": 72.85714285714285, "ocr_gen": 60.0, "seq_gen_rec": 38.46153846153847, "seq_rec": 60.0, "spat_seq_gen_rec": 62.499999999999986, "know_seq_rec": 0.0, "know_seq_gen_rec": 0.0, "spat_ocr_gen_rec_seq": 80.0, "know_ocr_gen_rec_seq": 70.0, "know_math_rec": 50.0, "ocr_rec_seq": 0.0 }, "acc_stderr": 0, "acc": 58.9147 }, "MathVerse": { "reject_info": { "reject_rate": 0.03, "reject_number": 1, "total_question": 3940 }, "Vision Intensive": { "accuracy": 42.13, "correct": 332, "total": 788 }, "Total": { "accuracy": 40.7, "correct": 1603, "total": 3939 }, "Text Lite": { "accuracy": 43.15, "correct": 340, "total": 788 }, "Vision Dominant": { "accuracy": 42.39, "correct": 334, "total": 788 }, "Text Dominant": { "accuracy": 47.59, "correct": 375, "total": 788 }, "Vision Only": { "accuracy": 28.21, "correct": 222, "total": 787 }, "accuracy": 40.7, "acc_stderr": 0, "acc": 40.7 }, "Ocrlite": { "reject_info": { "reject_rate": 0.06, "reject_number": 1, "total_question": 1645 }, "final_score": [ 1128, 1644 ], "accuracy": 68.613, "Key Information Extraction-Bookshelf": [ 20, 52 ], "Scene Text-centric VQA-diet_constraints": [ 62, 89 ], "Doc-oriented VQA-Control": [ 114, 189 ], "Doc-oriented VQA": [ 142, 204 ], "Scene Text-centric VQA-Fake_logo": [ 39, 119 ], "Handwritten Mathematical Expression Recognition": [ 47, 100 ], "Key Information Extraction": [ 162, 209 ], "Scene Text-centric VQA-Control": [ 159, 200 ], "Scene Text-centric VQA": [ 231, 282 ], "Artistic Text Recognition": [ 41, 50 ], "Irregular Text Recognition": [ 37, 50 ], "Non-Semantic Text Recognition": [ 27, 50 ], "Regular Text Recognition": [ 47, 50 ], "acc_stderr": 0, "acc": 68.613 }, "OcrliteZh": { "final_score": [ 71, 234 ], "accuracy": 30.342, "Docvqa": [ 2, 10 ], "Chartqa-human": [ 2, 10 ], "Chartqa-au": [ 2, 10 ], "infographic": [ 1, 10 ], "Key Information Extraction": [ 25, 45 ], "Scene Text-centric VQA": [ 17, 40 ], "Artistic Text Recognition": [ 0, 11 ], "IrRegular Text Recognition": [ 0, 11 ], "Non-semantic Text Recognition": [ 2, 12 ], "Regular Text Recognition": [ 2, 11 ], "Handwriting_CN": [ 5, 20 ], "Chinese Unlimited": [ 13, 44 ], "acc_stderr": 0, "acc": 30.342 }, "CharXiv": { "reject_info": { "reject_rate": 0.02, "reject_number": 1, "total_question": 5000 }, "descriptive": { "Overall Score": 70.97, "By Question": { "Q1": 79.51, "Q2": 73.48, "Q3": 56.65, "Q4": 82.49, "Q5": 79.08, "Q6": 65.06, "Q7": 68.38, "Q8": 70.98, "Q9": 67.66, "Q10": 76.03, "Q11": 64.0, "Q12": 70.33, "Q13": 63.01, "Q14": 86.17, "Q15": 89.42, "Q16": 80.56, "Q17": 10.71, "Q18": 82.59, "Q19": 87.69 }, "By Category": { "Information Extraction": 72.24, "Enumeration": 77.14, "Pattern Recognition": 75.33, "Counting": 75.32, "Compositionality": 10.71 }, "By Subplot": { "1 Subplot": 76.42, "2-4 Subplots": 71.54, "5+ Subplots": 61.12 }, "By Subject": { "Computer Science": 70.44, "Economics": 72.1, "Electrical Engineering and Systems Science": 76.05, "Mathematics": 73.52, "Physics": 66.47, "Quantitative Biology": 65.67, "Quantitative Finance": 71.98, "Statistics": 71.68 }, "By Year": { "2020": 72.47, "2021": 68.3, "2022": 71.9, "2023": 71.37 }, "N_valid": 3999, "N_invalid": 0, "Question Type": "Descriptive" }, "reasoning": { "Overall Score": 33.2, "By Answer Type": { "Text-in-Chart": 37.73, "Text-in-General": 41.41, "Number-in-Chart": 30.6, "Number-in-General": 23.58 }, "By Source": { "GPT-Sourced": 39.67, "GPT-Inspired": 30.09, "Completely Human": 32.33 }, "By Subject": { "Computer Science": 29.37, "Economics": 34.06, "Electrical Engineering and Systems Science": 36.97, "Mathematics": 36.3, "Physics": 33.86, "Quantitative Biology": 31.75, "Quantitative Finance": 31.03, "Statistics": 31.86 }, "By Year": { "2020": 35.22, "2021": 35.63, "2022": 30.33, "2023": 31.45 }, "By Subplot": { "1 Subplot": 33.94, "2-4 Subplots": 34.39, "5+ Subplots": 30.08 }, "N_valid": 1000, "N_invalid": 2, "Question Type": "Reasoning" }, "accuracy": 52.09, "acc_stderr": 0, "acc": 52.09 }, "MathVision": { "accuracy": 25.03, "acc_stderr": 0, "acc": 25.03 }, "CII-Bench": { "accuracy": 57.78, "domain_score": { "CTC": 53.33, "Society": 59.46, "Art": 60.29, "Env.": 68.52, "Life": 54.11, "Politics": 66.67 }, "emotion_score": { "Positive": 55.13, "Negative": 61.13, "Neutral": 56.77 }, "acc_stderr": 0, "acc": 57.78 }, "Blink": { "reject_info": { "reject_rate": 0.05, "reject_number": 1, "total_question": 1901 }, "accuracy": 56.95, "Art Style": 75.0, "Counting": 73.33, "Forensic Detection": 43.94, "Functional Correspondence": 31.54, "IQ Test": 20.67, "Jigsaw": 70.0, "Multi-view Reasoning": 38.35, "Object Localization": 63.11, "Relative Depth": 76.61, "Relative Reflectance": 38.81, "Semantic Correspondence": 43.17, "Spatial Relation": 84.62, "Visual Correspondence": 59.3, "Visual Similarity": 84.44, "acc_stderr": 0, "acc": 56.95 } } }