wiktionary-data / README.md
QubitPi's picture
Support Sanskrit
c23c545 verified
|
raw
history blame
5.09 kB
metadata
license: apache-2.0
pretty_name: English Wiktionary Data in JSONL
language:
  - en
  - de
  - la
  - grc
  - ko
  - peo
  - akk
  - sa
configs:
  - config_name: Languages
    data_files:
      - split: German
        path: german-wiktextract-data.jsonl
      - split: Latin
        path: latin-wiktextract-data.jsonl
      - split: AncientGreek
        path: ancient-greek-wiktextract-data.jsonl
      - split: Korean
        path: korean-wiktextract-data.jsonl
      - split: OldPersian
        path: old-persian-wiktextract-data.jsonl
      - split: Akkadian
        path: akkadian-wiktextract-data.jsonl
      - split: Sanskrit
        path: sanskrit-wiktextract-data.jsonl
  - config_name: Graph
    data_files:
      - split: AllLanguage
        path: word-definition-graph-data.jsonl
tags:
  - Natural Language Processing
  - NLP
  - Wiktionary
  - Vocabulary
  - German
  - Latin
  - Ancient Greek
  - Korean
  - Old Persian
  - Akkadian
  - Sanskrit
  - Knowledge Graph
size_categories:
  - 100M<n<1B

Wiktionary Data on Hugging Face Datasets

Hugging Face dataset badge

Python Version Badge GitHub workflow status badge Hugging Face sync status badge Apache License Badge

wiktionary-data is a sub-data extraction of the English Wiktionary that currently supports the following languages:

  • Deutsch - German
  • Latinum - Latin
  • Ἑλληνική - Ancient Greek
  • 한국어 - Korean
  • 𐎠𐎼𐎹 - Old Persian
  • 𒀝𒅗𒁺𒌑(𒌝) - Akkadian
  • संस्कृतम् - Sanskrit, or Classical Sanskrit

wiktionary-data was originally a sub-module of wilhelm-graphdb. While the dataset it's getting bigger, I noticed a wave of more exciting potentials this dataset can bring about that stretches beyond the scope of the containing project. Therefore I decided to promote it to a dedicated module; and here comes this repo.

The Wiktionary language data is available on 🤗 Hugging Face Datasets.

from datasets import load_dataset
dataset = load_dataset("QubitPi/wiktionary-data", split="German")

There are two data subsets:

  1. Languages subset that contains the sub-data extraction of the following splits:

    • German
    • Latin
    • AncientGreek
    • Korean
    • OldPersian
    • Akkadian
    • Sanskrit
  2. Graph subset that is useful for constructing knowledge graphs:

    • AllLanguage: all the languages in a giant graph

    The Graph data ontology is the following:

    Error loading ontology.png

Two words are structurally similar if and only if the two shares the same
stem

Development

Although the original Wiktionary dump is available, parsing it from scratch involves rather complicated process. We would probably do it in the future. At present, however, we would simply take the awesome works by tatuylonen which has already processed it and presented it in in JSON format. wiktionary-data takes the raw Wiktextract data (JSONL, one object per line) option.

License

The use and distribution terms for wiktionary-data are covered by the Apache License, Version 2.0.