Model,Completeness,Precision,Relevance,Sum InfiMM (Zephyr-7B),0.77,1.08,1.71,3.58 Emu2-Chat (LLaMA-33B),1.07,1.24,1.88,4.19 Fuyu-8B (Persimmon-8B),0.88,0.83,1.82,3.53 BakLLava (Mistral-7B),1.0,0.77,1.61,3.38 SPHINX,0.79,1.14,1.72,3.65 mPLUG-Owl2 (LLaMA-7B),1.06,1.24,1.36,3.67 LLaVA-v1.5 (Vicuna-v1.5-7B),0.9,1.13,1.18,3.21 LLaVA-v1.5 (Vicuna-v1.5-13B),0.91,1.28,1.29,3.47 InternLM-XComposer-VL (InternLM),1.08,1.26,1.87,4.21 IDEFICS-Instruct (LLaMA-7B),0.83,1.03,1.33,3.18 Qwen-VL (QwenLM),0.98,0.75,1.63,3.36 Shikra (Vicuna-7B),0.89,1.11,1.33,3.34 Otter-v1 (MPT-7B),0.96,0.83,1.83,3.61 Kosmos-2,1.12,1.06,1.85,4.03 InstructBLIP (Flan-T5-XL),0.87,1.04,1.11,3.02 InstructBLIP (Vicuna-7B),0.79,1.21,0.84,2.84 VisualGLM-6B (GLM-6B),0.82,0.97,1.21,2.99 mPLUG-Owl (LLaMA-7B),1.06,1.28,1.6,3.94 LLaMA-Adapter-V2,0.85,1.15,1.44,3.45 LLaVA-v1 (Vicuna-13B),0.91,1.25,1.6,3.76 MiniGPT-4 (Vicuna-13B),1.0,1.26,1.41,3.67