|
--- |
|
language: |
|
- tr |
|
tags: |
|
- paraphrasing |
|
- encoder-decoder |
|
- seq2seq |
|
- bert |
|
--- |
|
|
|
#Bert2Bert Turkish Paraphrase Generation |
|
|
|
#INISTA 2021 |
|
|
|
#Comparison of Turkish Paraphrase Generation Models |
|
|
|
#Dataset |
|
|
|
The dataset used in model training was created with the combination of the translation of the QQP dataset and manually generated dataset. |
|
Dataset [Link](https://drive.google.com/file/d/1-2l9EwIzXZ7fUkNW1vdeF3lzQp2pygp_/view?usp=sharing) |
|
|
|
#How To Use |
|
```python |
|
from transformers import BertTokenizerFast,EncoderDecoderModel |
|
tokenizer=BertTokenizerFast.from_pretrained("dbmdz/bert-base-turkish-cased") |
|
model = EncoderDecoderModel.from_pretrained("ahmetbagci/bert2bert-turkish-paraphrase-generation") |
|
|
|
text="son model arabalar çevreye daha mı az zarar veriyor?" |
|
input_ids = tokenizer(text, return_tensors="pt").input_ids |
|
output_ids = model.generate(input_ids) |
|
print(tokenizer.decode(output_ids[0], skip_special_tokens=True)) |
|
#sample output |
|
#son model arabalar çevre için daha az zararlı mı? |
|
``` |
|
#Cite |
|
```bibtex |
|
|
|
@inproceedings{bagci2021paraphrase, |
|
title={Comparison of Turkish Paraphrase Generation Models}, |
|
author={Ba{\u{g}}c{\i}, Ahmet and Amasyali, Mehmet Fatih}, |
|
booktitle={2021 International Conference on INnovations in Intelligent SysTems and Applications (INISTA)}, |
|
year={2021}, |
|
organization={IEEE} |
|
} |
|
``` |