Datasets:
Tasks:
Text Classification
Languages:
Persian
File size: 2,979 Bytes
331f4c7 2e8cf3d 331f4c7 2e8cf3d 331f4c7 57fa245 2e8cf3d fcdaa9f 331f4c7 e7b4280 2e8cf3d e7b4280 2e8cf3d e7b4280 1742452 2e8cf3d 331f4c7 2e8cf3d 331f4c7 2e8cf3d 331f4c7 fcdaa9f 2e8cf3d 331f4c7 2e8cf3d 331f4c7 2e8cf3d 331f4c7 2e8cf3d 331f4c7 2e8cf3d 331f4c7 2e8cf3d 331f4c7 2e8cf3d 331f4c7 2e8cf3d 331f4c7 2e8cf3d 331f4c7 2e8cf3d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 |
import csv
import datasets
from datasets.tasks import TextClassification
logger = datasets.logging.get_logger(__name__)
_CITATION = """Citation"""
_DESCRIPTION = """Description"""
_DOWNLOAD_URLS = {
"train": "https://huggingface.co/datasets/mahdiyehebrahimi/utc/raw/main/utc_train.csv",
"test": "https://huggingface.co/datasets/mahdiyehebrahimi/utc/raw/main/utc_test.csv",
}
class DatasetNameConfig(datasets.BuilderConfig):
def __init__(self, **kwargs):
super(DatasetNameConfig, self).__init__(**kwargs)
class DatasetName(datasets.GeneratorBasedBuilder):
BUILDER_CONFIGS = [
DatasetNameConfig(
name="utc",
version=datasets.Version("1.1.1"),
description=_DESCRIPTION,
),
]
def _info(self):
text_column = "text"
label_column = "label"
# TODO PROVIDE THE LABELS HERE
label_names = ["drop_withdraw", "centralauthentication_email","supervisor_seminar_proposal_defense", "registration"]
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=datasets.Features(
{text_column: datasets.Value("string"), label_column: datasets.features.ClassLabel(names=label_names)}
),
homepage="https://huggingface.co/datasets/mahdiyehebrahimi/utc",
citation=_CITATION,
task_templates=[TextClassification(text_column=text_column, label_column=label_column)],
)
def _split_generators(self, dl_manager):
"""
Return SplitGenerators.
"""
train_path = dl_manager.download_and_extract(_DOWNLOAD_URLS["train"])
test_path = dl_manager.download_and_extract(_DOWNLOAD_URLS["test"])
return [
datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepath": train_path}),
datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={"filepath": test_path}),
]
# TODO
def _generate_examples(self, filepath):
"""
Per each file_path read the csv file and iterate it.
For each row yield a tuple of (id, {"text": ..., "label": ..., ...})
Each call to this method yields an output like below:
```
(123, {"text": "I liked it", "label": "positive"})
```
"""
label2id = self.info.features[self.info.task_templates[0].label_column].str2int
logger.info("⏳ Generating examples from = %s", filepath)
with open(filepath, encoding="utf-8") as csv_file:
csv_reader = csv.reader(csv_file, quotechar='"', skipinitialspace=True)
# Uncomment below line to skip the first row if your csv file has a header row
next(csv_reader, None)
for id_, row in enumerate(csv_reader):
text, label = row
label = label2id(label)
# Optional preprocessing here
yield id_, {"text": text, "label": label} |