Datasets:

maplm_v2 / maplm_v2.py
maysonma's picture
fix loader
85a1d8a
raw
history blame
5.3 kB
import json
import os
import datasets
import numpy as np
_DESCRIPTION = """
MAPLM: A Real-World Large-Scale Vision-Language Benchmark for Map and Traffic Scene Understanding
"""
_HOMEPAGE = "https://github.com/llvm-ad/maplm"
_LICENSE = "https://github.com/LLVM-AD/MAPLM/blob/main/LICENSE"
_CITATION = """\
@inproceedings{cao_maplm_2024,
title = {{MAPLM}: {A} {Real}-{World} {Large}-{Scale} {Vision}-{Language} {Dataset} for {Map} and {Traffic} {Scene} {Understanding}},
booktitle = {{CVPR}},
author = {Cao, Xu and Zhou, Tong and Ma, Yunsheng and Ye, Wenqian and Cui, Can and Tang, Kun and Cao, Zhipeng and Liang, Kaizhao and Wang, Ziran and Rehg, James M. and Zheng, Chao},
year = {2024},
}
"""
class MapLMBuilderConfig(datasets.BuilderConfig):
"""BuilderConfig for MapLM dataset."""
def __init__(self, name, splits):
super(MapLMBuilderConfig, self).__init__(name=name)
self.splits = splits
class MapLMDataset(datasets.GeneratorBasedBuilder):
BUILDER_CONFIG_CLASS = MapLMBuilderConfig
BUILDER_CONFIGS = [
MapLMBuilderConfig(
name="v2.0",
splits=["train", "val", "test"],
)
]
DEFAULT_CONFIG_NAME = "v2.0"
def _info(self):
# info stores information about your dataset like its description, license, and features.
feature_dict = {
"frame_id": datasets.Value("string"),
"images": datasets.Sequence(datasets.Value("string")),
"question": datasets.Sequence(datasets.Value("string")),
"options": datasets.Sequence(datasets.Sequence(datasets.Value("string"))),
"answer": datasets.Sequence(datasets.Sequence(datasets.Value("string"))),
"tag": datasets.Sequence(datasets.Value("string")),
}
return datasets.DatasetInfo(
# This is the description that will appear on the datasets page.
description=_DESCRIPTION,
# This defines the different columns of the dataset and their types
features=datasets.Features(feature_dict),
homepage=_HOMEPAGE,
license=_LICENSE,
citation=_CITATION,
)
def _split_generators(self, dl_manager):
splits = []
data_root = dl_manager.download("data/")
for split in self.config.splits:
annotation_file = os.path.join(data_root, f"{split}_v2.json")
annotations = json.load(open(annotation_file))
if split == "test":
generator = datasets.SplitGenerator(
name=datasets.Split.TEST,
gen_kwargs={"annotations": annotations},
)
elif split == "train":
generator = datasets.SplitGenerator(
name=datasets.Split.TRAIN,
gen_kwargs={"annotations": annotations},
)
elif split == "val":
generator = datasets.SplitGenerator(
name=datasets.Split.VALIDATION,
gen_kwargs={"annotations": annotations},
)
else:
continue
splits.append(generator)
return splits
# method parameters are unpacked from `gen_kwargs` as given in `_split_generators`
def _generate_examples(self, annotations):
for i, anno_key in enumerate(annotations):
data_item = {}
data_item["frame_id"] = annotations[anno_key]["id"]
data_item["images"] = list(annotations[anno_key]["image_paths"].values())
data_item["question"] = []
data_item["options"] = []
data_item["answer"] = []
data_item["tag"] = []
for perception_key in annotations[anno_key]["QA"]["perception"]:
data_item["question"].append(
annotations[anno_key]["QA"]["perception"][perception_key][
"question"
]
)
data_item["options"].append(
annotations[anno_key]["QA"]["perception"][perception_key]["option"]
)
anno_answer = annotations[anno_key]["QA"]["perception"][perception_key][
"answer"
]
if isinstance(anno_answer, list):
data_item["answer"].append(anno_answer)
else:
data_item["answer"].append([anno_answer])
data_item["tag"].append(
annotations[anno_key]["QA"]["perception"][perception_key]["tag"]
)
for behavior_key in annotations[anno_key]["QA"]["behavior"]:
data_item["question"].append(
annotations[anno_key]["QA"]["behavior"][behavior_key]["question"]
)
data_item["options"].append(
annotations[anno_key]["QA"]["behavior"][behavior_key]["option"]
)
data_item["answer"].append(
annotations[anno_key]["QA"]["behavior"][behavior_key]["answer"]
)
data_item["tag"].append(
annotations[anno_key]["QA"]["behavior"][behavior_key]["tag"]
)
yield i, data_item