Datasets:
File size: 3,635 Bytes
808187d eadf52e 17acea3 808187d 17acea3 62e9177 51e6200 e774f45 808187d 3371f7f 808187d e774f45 808187d d11e825 808187d 5c728e6 d11e825 808187d ba7d78e 808187d 17acea3 d11e825 f875dbb e774f45 d11e825 15fe087 d11e825 ba7d78e f875dbb 4767376 f875dbb f6ef9d8 f875dbb 4767376 e774f45 f875dbb e774f45 65d5145 0dd571a e774f45 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 |
"""CC6204-Hackaton-Cub-Dataset: Multimodal"""
import os
import re
import datasets
import pandas as pd
from requests import get
datasets.logging.set_verbosity_debug()
logger = datasets.logging.get_logger(__name__)
#datasets.logging.set_verbosity_info()
datasets.logging.set_verbosity_debug()
_DESCRIPTION = "XYZ"
_CITATION = "XYZ"
_HOMEPAGE = "https://github.com/ivansipiran/CC6204-Deep-Learning/blob/main/Hackaton/hackaton.md"
_REPO = "https://huggingface.co/datasets/alkzar90/CC6204-Hackaton-Cub-Dataset/resolve/main/data"
_URLS = {
"train_test_split": f"{_REPO}/train_test_split.txt",
"classes": f"{_REPO}/classes.txt",
"image_class_labels": f"{_REPO}/image_class_labels.txt",
"images": f"{_REPO}/images.txt",
"image_urls": f"{_REPO}/images.zip",
"text_urls": f"{_REPO}/text.zip",
}
# Create id-to-label dictionary using the classes file
classes = get(_URLS["classes"]).iter_lines()
_ID2LABEL = {}
for row in classes:
row = row.decode("UTF8")
if row != "":
idx, label = row.split(" ")
_ID2LABEL[idx] = re.search("[^\d\.\_+].+", label).group(0).replace("_", " ")
_NAMES = list(_ID2LABEL.values())
# build from images.txt: a mapping from image_file_name -> id
imgpath_to_ids = get(_URLS["images"]).iter_lines()
_IMGNAME2ID = {}
for row in imgpath_to_ids:
row = row.decode("UTF8")
if row != "":
idx, img_name = row.split(" ")
_IMGNAME2ID[os.path.basename(img_name)] = idx
# Create TRAIN_IDX_SET
train_test_split = get(_URLS["train_test_split"]).iter_lines()
_TRAIN_IDX_SET = []
for row in train_test_split:
row = row.decode("UTF8")
if row != "":
idx, train_bool = row.split(" ")
# 1: train, 0: test
if train_bool == "1":
_TRAIN_IDX_SET.append(idx)
_TRAIN_IDX_SET = set(_TRAIN_IDX_SET)
class CubDataset(datasets.GeneratorBasedBuilder):
"""Cub Dataset"""
def _info(self):
features = datasets.Features({
"image": datasets.Image(),
"labels": datasets.features.ClassLabel(names=_NAMES),
})
keys = ("image", "labels")
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=features,
supervised_keys=keys,
homepage=_HOMEPAGE,
citation=_CITATION,
)
def _split_generators(self, dl_manager):
train_files = []
test_files = []
# Download images
data_files = dl_manager.download_and_extract(_URLS["image_urls"])
for batch in data_files:
path_files = dl_manager.iter_files(batch)
for img in path_files:
if img.endswith("\d+.jpg"):
img_idx = _IMGNAME2ID[img]
if img_idx in _TRAIN_IDX_SET:
train_files.append(img)
else:
test_files.append(img)
return [
datasets.SplitGenerator(
name=datasets.Split.TRAIN,
gen_kwargs={
"files": train_files
}
),
datasets.SplitGenerator(
name=datasets.Split.TEST,
gen_kwargs={
"files": test_files
}
)
]
def _generate_examples(self, files):
for i, path in enumerate(files):
file_name = os.path.basename(path)
if file_name.endswith(".jpg"):
yield i, {
"image": path,
"labels": os.path.basename(os.path.dirname(path)).lower(),
}
|