|
--- |
|
language: |
|
- en |
|
license: apache-2.0 |
|
size_categories: |
|
- n<1K |
|
task_categories: |
|
- question-answering |
|
pretty_name: Mantis-Eval |
|
dataset_info: |
|
- config_name: mantis_eval |
|
features: |
|
- name: id |
|
dtype: string |
|
- name: question_type |
|
dtype: string |
|
- name: question |
|
dtype: string |
|
- name: images |
|
sequence: image |
|
- name: options |
|
sequence: string |
|
- name: answer |
|
dtype: string |
|
- name: data_source |
|
dtype: string |
|
- name: category |
|
dtype: string |
|
splits: |
|
- name: test |
|
num_bytes: 479770102 |
|
num_examples: 217 |
|
download_size: 473031413 |
|
dataset_size: 479770102 |
|
configs: |
|
- config_name: mantis_eval |
|
data_files: |
|
- split: test |
|
path: mantis_eval/test-* |
|
--- |
|
|
|
## Overview |
|
This is a newly curated dataset to evaluate multimodal language models' capability to reason over multiple images. More details are shown in https://tiger-ai-lab.github.io/Mantis/. |
|
|
|
### Statistics |
|
This evaluation dataset contains 217 human-annotated challenging multi-image reasoning problems. |
|
|
|
### Leaderboard |
|
We list the current results as follows: |
|
|
|
| Models | Size | Mantis-Eval | |
|
|:------------------|:-----|:------------| |
|
| LLaVA OneVision | 72B | 77.60 | |
|
| LLaVA OneVision | 7B | 64.20 | |
|
| GPT-4V | - | 62.67 | |
|
| Mantis-SigLIP | 8B | 59.45 | |
|
| Mantis-Idefics2 | 8B | 57.14 | |
|
| Mantis-CLIP | 8B | 55.76 | |
|
| VILA | 8B | 51.15 | |
|
| BLIP-2 | 13B | 49.77 | |
|
| Idefics2 | 8B | 48.85 | |
|
| InstructBLIP | 13B | 45.62 | |
|
| LLaVA-V1.6 | 7B | 45.62 | |
|
| CogVLM | 17B | 45.16 | |
|
| LLaVA OneVision | 0.5B | 39.60 | |
|
| Qwen-VL-Chat | 7B | 39.17 | |
|
| Emu2-Chat | 37B | 37.79 | |
|
| VideoLLaVA | 7B | 35.04 | |
|
| Mantis-Flamingo | 9B | 32.72 | |
|
| LLaVA-v1.5 | 7B | 31.34 | |
|
| Kosmos2 | 1.6B | 30.41 | |
|
| Idefics1 | 9B | 28.11 | |
|
| Fuyu | 8B | 27.19 | |
|
| OpenFlamingo | 9B | 12.44 | |
|
| Otter-Image | 9B | 14.29 | |
|
|
|
### Citation |
|
If you are using this dataset, please cite our work with |
|
``` |
|
@inproceedings{Jiang2024MANTISIM, |
|
title={MANTIS: Interleaved Multi-Image Instruction Tuning}, |
|
author={Dongfu Jiang and Xuan He and Huaye Zeng and Cong Wei and Max W.F. Ku and Qian Liu and Wenhu Chen}, |
|
publisher={arXiv2405.01483} |
|
year={2024}, |
|
} |
|
``` |
|
|