|
import os |
|
import shutil |
|
from os.path import dirname, join |
|
|
|
from datasets import load_dataset |
|
from underthesea import word_tokenize |
|
|
|
|
|
def create_wtk_dataset(text_dataset, output_folder): |
|
if os.path.exists(output_folder): |
|
shutil.rmtree(output_folder) |
|
os.makedirs(output_folder) |
|
for split in ["train", "validation", "test"]: |
|
sentences = text_dataset[split]["text"] |
|
with open(join(output_folder, f"{split}.txt"), "w") as f: |
|
for sentence in sentences: |
|
items = word_tokenize(sentence) |
|
for item in items: |
|
tokens = item.split() |
|
for i, token in enumerate(tokens): |
|
if i > 0: |
|
f.write(f"{token}\tI-W\n") |
|
else: |
|
f.write(f"{token}\tB-W\n") |
|
f.write("\n") |
|
|
|
|
|
pwd = dirname(__file__) |
|
data_folder = join(pwd, "data") |
|
|
|
text_dataset = load_dataset("undertheseanlp/UTS_Text", "small") |
|
create_wtk_dataset(text_dataset, join(data_folder, "small")) |
|
|
|
text_dataset = load_dataset("undertheseanlp/UTS_Text", "base") |
|
create_wtk_dataset(text_dataset, join(data_folder, "base")) |
|
|
|
text_dataset = load_dataset("undertheseanlp/UTS_Text", "large") |
|
create_wtk_dataset(text_dataset, join(data_folder, "large")) |
|
|