abilibili's picture
Upload folder using huggingface_hub
4772a8d verified

A newer version of the Gradio SDK is available: 5.6.0

Upgrade

Evaluation Instruction for MiniGPT-v2

Data preparation

Images download

Image source Download path
OKVQA annotations    images
gqa annotations    images
hateful meme images and annotations
iconqa images and annotation
vizwiz images and annotation
RefCOCO annotations
RefCOCO+ annotations
RefCOCOg annotations

Evaluation dataset structure

${MINIGPTv2_EVALUATION_DATASET}
β”œβ”€β”€ gqa
β”‚   └── test_balanced_questions.json
β”‚   β”œβ”€β”€ testdev_balanced_questions.json
β”‚   β”œβ”€β”€ gqa_images
β”œβ”€β”€ hateful_meme
β”‚   └── hm_images
β”‚   β”œβ”€β”€ dev.jsonl
β”œβ”€β”€ iconvqa
β”‚   └── iconvqa_images
β”‚   β”œβ”€β”€ choose_text_val.json
β”œβ”€β”€ vizwiz
β”‚   └── vizwiz_images
β”‚   β”œβ”€β”€ val.json
β”œβ”€β”€ vsr
β”‚   └── vsr_images
β”œβ”€β”€ okvqa
β”‚   β”œβ”€β”€ okvqa_test_split.json
β”‚   β”œβ”€β”€ mscoco_val2014_annotations_clean.json
β”‚   β”œβ”€β”€ OpenEnded_mscoco_val2014_questions_clean.json
β”œβ”€β”€ refcoco
β”‚   └── instances.json
β”‚   β”œβ”€β”€ refs(google).p
β”‚   β”œβ”€β”€ refs(unc).p
β”œβ”€β”€ refcoco+
β”‚   └── instances.json
β”‚   β”œβ”€β”€ refs(unc).p
β”œβ”€β”€ refercocog
β”‚   └── instances.json
β”‚   β”œβ”€β”€ refs(google).p
β”‚   β”œβ”€β”€ refs(und).p
...

environment setup

export PYTHONPATH=$PYTHONPATH:/path/to/directory/of/MiniGPT-4

config file setup

Set llama_model to the path of LLaMA model.
Set ckpt to the path of our pretrained model.
Set eval_file_path to the path of the annotation files for each evaluation data.
Set img_path to the img_path for each evaluation dataset.
Set save_path to the save_path for each evaluation dataset.

in eval_configs/minigptv2_benchmark_evaluation.yaml

start evalauting RefCOCO, RefCOCO+, RefCOCOg

port=port_number
cfg_path=/path/to/eval_configs/minigptv2_benchmark_evaluation.yaml

dataset names:

refcoco refcoco+ refcocog
torchrun --master-port ${port} --nproc_per_node 1 eval_ref.py \
 --cfg-path ${cfg_path} --dataset refcoco,refcoco+,refcocog --resample

start evaluating visual question answering

port=port_number
cfg_path=/path/to/eval_configs/minigptv2_benchmark_evaluation.yaml

dataset names:

okvqa vizwiz iconvqa gqa vsr hm
torchrun --master-port ${port} --nproc_per_node 1 eval_vqa.py \
 --cfg-path ${cfg_path} --dataset okvqa,vizwiz,iconvqa,gqa,vsr,hm