|
from datasets import load_dataset |
|
from underthesea import word_tokenize |
|
from underthesea.pipeline.word_tokenize.regex_tokenize import tokenize |
|
from os.path import dirname, join |
|
from underthesea.utils import logger |
|
|
|
dataset = load_dataset("undertheseanlp/UTS_Text_v1") |
|
|
|
sentences = dataset["train"]["text"] |
|
pwd = dirname(__file__) |
|
data_file = join(pwd, "data/train.txt") |
|
with open(data_file, "w") as f: |
|
f.write("") |
|
|
|
|
|
f = open(data_file, "a") |
|
content = "" |
|
for j, s in enumerate(sentences): |
|
if j % 100 == 0 and j > 0: |
|
f.write(content) |
|
content = "" |
|
logger.info(j) |
|
words = word_tokenize(s) |
|
for word in words: |
|
tokens = tokenize(word) |
|
for i, token in enumerate(tokens): |
|
if i == 0: |
|
tag = "B-W" |
|
else: |
|
tag = "I-W" |
|
content += token + "\t" + tag + "\n" |
|
content += "\n" |
|
f.close() |
|
|