--- license: mit language: - ce - ru - en metrics: - codeparrot/apps_metric - accuracy tags: - code --- The model uses only sign **ӏ** for explosive consonants (small cyrillic palochka letter)! The model was teached by folloving David Dale's instructions for erzya language (https://arxiv.org/abs/2209.09368) and using code from his repository. Commentaries in Russian were left untouched. ```python import torch from transformers import BertTokenizer, AutoModel import numpy as np import pandas as pd import razdel import matplotlib.pyplot as plt from tqdm.auto import tqdm, trange ``` Download the model from Huggingface repository: ```python model_name = 'NM-development/labse-en-ru-ce-prototype' tokenizer = BertTokenizer.from_pretrained(model_name) model = AutoModel.from_pretrained(model_name) ``` Assign files with the texts you want to split into parallel sentences: ```python file_ru = None file_nm = None with open(file_nm, 'r') as f1, open(file_ru, 'r') as f2: nm_text = f1.read() ru_text = f2.read() ``` In the following section define auxillary functions for parallel sentence comparison: ```python def embed(text): encoded_input = tokenizer(text, padding=True, truncation=True, max_length=128, return_tensors='pt') with torch.inference_mode(): model_output = model(**encoded_input.to(model.device)) embeddings = model_output.pooler_output embeddings = torch.nn.functional.normalize(embeddings) return embeddings[0].cpu().numpy() def get_top_mean_by_row(x, k=5): m, n = x.shape k = min(k, n) topk_indices = np.argpartition(x, -k, axis=1)[:, -k:] rows, _ = np.indices((m, k)) return x[rows, topk_indices].mean(1) def align3(sims): rewards = np.zeros_like(sims) choices = np.zeros_like(sims).astype(int) # 1: choose this pair, 2: decrease i, 3: decrease j # алгоритм, разрешающий пропускать сколько угодно пар, лишь бы была монотонность for i in range(sims.shape[0]): for j in range(0, sims.shape[1]): # вариант первый: выровнять i-тое предложение с j-тым score_add = sims[i, j] if i > 0 and j > 0: # вот как тогда выровняются предыдущие score_add += rewards[i-1, j-1] choices[i, j] = 1 best = score_add if i > 0 and rewards[i-1, j] > best: best = rewards[i-1, j] choices[i, j] = 2 if j > 0 and rewards[i, j-1] > best: best = rewards[i, j-1] choices[i, j] = 3 rewards[i, j] = best alignment = [] i = sims.shape[0] - 1 j = sims.shape[1] - 1 while i > 0 and j > 0: if choices[i, j] == 1: alignment.append([i, j]) i -= 1 j -= 1 elif choices[i, j] == 2: i -= 1 else: j -= 1 return alignment[::-1] def make_sents(text): sents = [s.text.replace('\n', ' ').strip() for p in text.split('\n\n') for s in razdel.sentenize(p)] sents = [s for s in sents if s] return sents ``` Firstly split your texts into sentences: ```python sents_nm = make_sents(nm_text) sents_ru = make_sents(ru_text) ``` Then embed all the chunks: ```python emb_ru = np.stack([embed(s) for s in tqdm(sents_ru)]) emb_nm = np.stack([embed(s) for s in tqdm(sents_nm)]) ``` Now compare sentenses' semanics vectors and build correlation heatmap: ```python pen = np.array([[min(len(x), len(y)) / max(len(x), len(y)) for x in sents_nm] for y in sents_ru]) sims = np.maximum(0, np.dot(emb_ru, emb_nm.T)) ** 1 * pen alpha = 0.2 penalty = 0.2 sims_rel = (sims.T - get_top_mean_by_row(sims) * alpha).T - get_top_mean_by_row(sims.T) * alpha - penalty alignment = align3(sims_rel) print(sum(sims[i, j] for i, j in alignment) / min(sims.shape)) plt.figure(figsize=(12, 6)) plt.subplot(1, 2, 1) plt.imshow(sims_rel) plt.subplot(1, 2, 2) plt.scatter(*list(zip(*alignment)), s=5); ``` Finally, save the parallel corpus into a json file: ```python nm_ru_parallel_corpus = pd.DataFrame({'nm_text' : [sents_nm[x[1]] for x in alignment], 'ru_text' : [sents_ru[x[0]] for x in alignment]}) corpus_filename = 'nm_ru_corpus.json' with open(corpus_filename, 'w') as f: nm_ru_parallel_corpus.to_json(f, force_ascii=False, indent=4) ```