|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
import json |
|
|
|
from database.database_clients import get_node_label_attribute_key |
|
from german_parser import get_declension_attributes |
|
from vocabulary_parser import GERMAN |
|
from vocabulary_parser import get_attributes |
|
from vocabulary_parser import get_definitions |
|
from vocabulary_parser import get_inferred_links |
|
from vocabulary_parser import get_vocabulary |
|
|
|
|
|
def generate_dataset(yaml_path: str, dataset_path: str): |
|
""" |
|
Generates a Hugging Face Dataset from https://github.com/QubitPi/wilhelm-vocabulary/blob/master/german.yaml |
|
|
|
:param yaml_path: The absolute or relative path (to the invoking script) to the YAML file above |
|
:param dataset_path: The absolute or relative path (to the invoking script) to the generated dataset file |
|
""" |
|
vocabulary = get_vocabulary(yaml_path) |
|
label_key = get_node_label_attribute_key() |
|
|
|
all_nodes = {} |
|
|
|
with open(dataset_path, "w") as graph: |
|
for word in vocabulary: |
|
term = word["term"] |
|
attributes = get_attributes(word, GERMAN, label_key, get_declension_attributes) |
|
source_node = attributes |
|
all_nodes[term] = source_node |
|
|
|
for definition_with_predicate in get_definitions(word): |
|
predicate = definition_with_predicate[0] |
|
definition = definition_with_predicate[1] |
|
|
|
target_node = {label_key: definition} |
|
label = {label_key: predicate if predicate else "definition"} |
|
|
|
graph.write(json.dumps({"source": source_node, "target": target_node, "link": label})) |
|
graph.write("\n") |
|
|
|
for link in get_inferred_links(vocabulary, label_key, get_declension_attributes): |
|
source_node = all_nodes[link["source_label"]] |
|
target_node = all_nodes[link["target_label"]] |
|
label = link["attributes"] |
|
|
|
graph.write(json.dumps({"source": source_node, "target": target_node, "link": label})) |
|
graph.write("\n") |
|
|