|
--- |
|
dataset_info: |
|
- config_name: challenge_data |
|
features: |
|
- name: pos_item_ids |
|
sequence: string |
|
- name: pos_item_contents |
|
sequence: string |
|
- name: question |
|
dtype: string |
|
- name: question_id |
|
dtype: string |
|
- name: instruction |
|
dtype: string |
|
- name: img_path |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 890417 |
|
num_examples: 6415 |
|
download_size: 169300 |
|
dataset_size: 890417 |
|
- config_name: challenge_passage |
|
features: |
|
- name: passage_id |
|
dtype: string |
|
- name: passage_content |
|
dtype: string |
|
- name: page_screenshot |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 44091445 |
|
num_examples: 47318 |
|
download_size: 24786149 |
|
dataset_size: 44091445 |
|
configs: |
|
- config_name: challenge_data |
|
data_files: |
|
- split: train |
|
path: challenge_data/train-* |
|
- config_name: challenge_passage |
|
data_files: |
|
- split: train |
|
path: challenge_passage/train-* |
|
--- |
|
# M2KR-Challenge Dataset |
|
|
|
A multimodal retrieval dataset for image-to-document and image+text-to-document matching tasks. |
|
|
|
## Dataset Overview |
|
|
|
This dataset contains two main subsets designed for multimodal retrieval challenges: |
|
- **challenge_data**: Query data with images and optional text questions (6.42k samples) |
|
- **challenge_passage**: Document collection with textual passages and associated web screenshot path (47.3k passages) |
|
|
|
## Dataset Structure |
|
|
|
### challenge_data (6,420 rows) |
|
**Columns**: |
|
- `img_path`: Image filename (string) |
|
- `instruction`: Task instruction for description generation |
|
- `question`: Optional text query (53% populated, if exist then it is a image+text-to-document retrieval task, else, it is a image-to-document task ) |
|
- `question_id`: Unique identifier (string) |
|
- `pos_item_ids`: Sequence of positive item IDs, the ground truth passage_id (empty, removed for private test set) |
|
- `pos_item_contents`: Sequence of positive item contents (empty, removed for private test set) |
|
|
|
**Task Types**: |
|
1. **Image-to-Document Retrieval**: When `question` is empty (image query) |
|
2. **Multimodal Retrieval**: When `question` contains text (image + text query) |
|
|
|
### challenge_passage (47,300 rows) |
|
**Columns**: |
|
- `passage_id`: Unique passage identifier (string) |
|
- `passage_content`: Textual description containing: |
|
- Image description |
|
- Structured details about persons (birth/death dates, occupations, locations, etc.) |
|
- `page_screenshot`: Associated image filename (string) |
|
|
|
For the retrieval task, you will need to retrieve the corresponding passage from the 47K-passage pool for each sample in challenge_data. |
|
|
|
## Images |
|
|
|
The image data is provided in separate archives: |
|
|
|
- **Web_Image.zip.001** |
|
- **Web_Image.zip.002** |
|
- **Web_Image.zip.003** |
|
|
|
These archives contain the web screenshots corresponding to the document passages. |
|
|
|
- **query_images.zip** |
|
Contains the query images used in the challenge. |
|
|
|
--- |
|
|
|
|
|
## References |
|
**Paper or resources for more information:** |
|
- **Paper:** https://arxiv.org/abs/2402.08327 |
|
- **Project Page:** https://preflmr.github.io/ |
|
- **Huggingface Implementation:** https://github.com/LinWeizheDragon/FLMR |
|
|
|
|
|
**Citation** |
|
If our work helped your research, please kindly cite our paper for PreFLMR. |
|
``` |
|
|
|
@inproceedings{lin-etal-2024-preflmr, |
|
title = "{P}re{FLMR}: Scaling Up Fine-Grained Late-Interaction Multi-modal Retrievers", |
|
author = "Lin, Weizhe and |
|
Mei, Jingbiao and |
|
Chen, Jinghong and |
|
Byrne, Bill", |
|
editor = "Ku, Lun-Wei and |
|
Martins, Andre and |
|
Srikumar, Vivek", |
|
booktitle = "Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", |
|
month = aug, |
|
year = "2024", |
|
address = "Bangkok, Thailand", |
|
publisher = "Association for Computational Linguistics", |
|
url = "https://aclanthology.org/2024.acl-long.289", |
|
pages = "5294--5316", |
|
abstract = "Large Multimodal Models (LMMs) excel in natural language and visual understanding but are challenged by exacting tasks such as Knowledge-based Visual Question Answering (KB-VQA) which involve the retrieval of relevant information from document collections to use in shaping answers to questions. We present an extensive training and evaluation framework, M2KR, for KB-VQA. M2KR contains a collection of vision and language tasks which we have incorporated into a single suite of benchmark tasks for training and evaluating general-purpose multi-modal retrievers. We use M2KR to develop PreFLMR, a pre-trained version of the recently developed Fine-grained Late-interaction Multi-modal Retriever (FLMR) approach to KB-VQA, and we report new state-of-the-art results across a range of tasks. We also present investigations into the scaling behaviors of PreFLMR intended to be useful in future developments in general-purpose multi-modal retrievers.", |
|
} |
|
|
|
``` |