Spaces:
Running
on
Zero
Running
on
Zero
import gradio as gr | |
import spaces | |
from transformers import pipeline | |
import torch | |
DESCRIPTION=""" | |
### a Turkish encoder-decoder language model | |
Welcome to our Huggingface space, where you can explore the capabilities of TURNA. | |
**Key Features of TURNA:** | |
- **Powerful Architecture:** TURNA contains 1.1B parameters, and was pre-trained with an encoder-decoder architecture following the UL2 framework on 43B tokens from various domains. | |
- **Diverse Training Data:** Our model is trained on a varied dataset of 43 billion tokens, covering a wide array of domains. | |
- **Broad Applications:** TURNA is fine-tuned for a variety of generation and understanding tasks, including: | |
- Summarization | |
- Paraphrasing | |
- News title generation | |
- Sentiment classification | |
- Text categorization | |
- Named entity recognition | |
- Part-of-speech tagging | |
- Semantic textual similarity | |
- Natural language inference | |
**Note:** First inference might take time as the models are downloaded on-the-go. | |
*TURNA can generate toxic content or provide erroneous information. Double-check before usage.* | |
""" | |
CITATION = """ | |
Refer to our [paper](https://arxiv.org/abs/2401.14373) for more details. | |
### Citation | |
```bibtex | |
@misc{uludogan2024turna, | |
title={TURNA: A Turkish Encoder-Decoder Language Model for Enhanced Understanding and Generation}, | |
author={Gökçe Uludoğan and Zeynep Yirmibeşoğlu Balal and Furkan Akkurt and Melikşah Türker and Onur Güngör and Susan Üsküdarlı}, | |
year={2024}, | |
eprint={2401.14373}, | |
archivePrefix={arXiv}, | |
primaryClass={cs.CL} | |
} | |
``` | |
""" | |
sentiment_example = [["Bu üründen çok memnun kaldım."]] | |
long_text = [["Eyfel Kulesi (Fransızca: La tour Eiffel [la tuʀ ɛˈfɛl]), Paris'teki demir kule. Kule, aynı zamanda tüm dünyada Fransa'nın sembolü halini almıştır. İsmini, inşa ettiren Fransız inşaat mühendisi Gustave Eiffel'den alır.[1] En büyük turizm cazibelerinden biri olan Eyfel Kulesi, yılda 6 milyon turist çeker. 2002 yılında toplam ziyaretçi sayısı 200 milyona ulaşmıştır."], ["Kalp krizi geçirenlerin yaklaşık üçte birinin kısa bir süre önce grip atlattığı düşünülüyor. Peki grip virüsü ne yapıyor da kalp krizine yol açıyor? Karpuz şöyle açıkladı: Grip virüsü kanın yapışkanlığını veya pıhtılaşmasını artırıyor."]] | |
ner_example = [["Benim adım Turna."]] | |
t2t_example = [["Paraphrase: Bu üründen çok memnun kaldım."]] | |
nli_example = [["Bunu çok beğendim.", "Bunu çok sevdim."]] | |
text_category_example = [[" anadolu_efes e 18 lik star ! beko_basketbol_ligi nde iddialı bir kadroyla sezona giren anadolu_efes transfer harekatına devam ediyor"]] | |
def nli(first_input, second_input, model_choice="turna_nli_nli_tr"): | |
if model_choice=="turna_nli_nli_tr": | |
input = f"hipotez: {first_input} önerme: {second_input}" | |
nli_model = pipeline(model="boun-tabi-LMG/turna_nli_nli_tr", device=0) | |
return nli_model(input)[0]["generated_text"] | |
else: | |
input = f"ilk cümle: {first_input} ikinci cümle: {second_input}" | |
stsb_model = pipeline(model="boun-tabi-LMG/turna_semantic_similarity_stsb_tr", device=0) | |
return stsb_model(input)[0]["generated_text"] | |
def sentiment_analysis(input, model_choice="turna_classification_17bintweet_sentiment"): | |
sentiment_model = pipeline(model=f"boun-tabi-LMG/{model_choice}", device=0) | |
return sentiment_model(input, max_new_tokens = 4)[0]["generated_text"] | |
def pos(input, model_choice, max_new_tokens, length_penalty, no_repeat_ngram_size): | |
if model_choice=="turna_pos_imst": | |
pos_imst = pipeline(model="boun-tabi-LMG/turna_pos_imst", device=0) | |
return pos_imst(input, max_new_tokens = max_new_tokens, length_penalty=length_penalty, no_repeat_ngram_size=no_repeat_ngram_size)[0]["generated_text"] | |
else: | |
pos_boun = pipeline(model="boun-tabi-LMG/turna_pos_boun", device=0) | |
return pos_boun(input, max_new_tokens = max_new_tokens, length_penalty=length_penalty, no_repeat_ngram_size=no_repeat_ngram_size)[0]["generated_text"] | |
def ner(input, model_choice, max_new_tokens, length_penalty, no_repeat_ngram_size): | |
if model_choice=="turna_ner_wikiann": | |
ner_wikiann = pipeline(model="boun-tabi-LMG/turna_ner_wikiann", device=0) | |
return ner_wikiann(input, max_new_tokens = max_new_tokens, length_penalty=length_penalty, no_repeat_ngram_size=no_repeat_ngram_size)[0]["generated_text"] | |
else: | |
ner_model = pipeline(model="boun-tabi-LMG/turna_ner_milliyet", device=0) | |
return ner_model(input, max_new_tokens = max_new_tokens, length_penalty=length_penalty, no_repeat_ngram_size=no_repeat_ngram_size)[0]["generated_text"] | |
def paraphrase(input, model_choice, max_new_tokens): | |
if model_choice=="turna_paraphrasing_tatoeba": | |
paraphrasing = pipeline(model="boun-tabi-LMG/turna_paraphrasing_tatoeba", device=0) | |
return paraphrasing(input, max_new_tokens = max_new_tokens)[0]["generated_text"] | |
else: | |
paraphrasing_sub = pipeline(model="boun-tabi-LMG/turna_paraphrasing_opensubtitles", device=0) | |
return paraphrasing_sub(input, max_new_tokens = max_new_tokens)[0]["generated_text"] | |
def summarize(input, model_choice, max_new_tokens, length_penalty, no_repeat_ngram_size): | |
model_mapping = {"turna_summarization_tr_news": "boun-tabi-LMG/turna_summarization_tr_news", | |
"turna_summarization_mlsum": "boun-tabi-LMG/turna_summarization_mlsum"} | |
summarization_model = pipeline(model=model_mapping[model_choice], device=0) | |
return summarization_model(input, max_new_tokens = max_new_tokens, length_penalty=length_penalty, no_repeat_ngram_size=no_repeat_ngram_size)[0]["generated_text"] | |
def generate_title(input, model_choice, max_new_tokens, length_penalty, no_repeat_ngram_size): | |
model_mapping = {"turna_title_generation_tr_news": "boun-tabi-LMG/turna_title_generation_tr_news", | |
"turna_title_generation_mlsum": "boun-tabi-LMG/turna_title_generation_mlsum"} | |
summarization_model = pipeline(model=model_mapping[model_choice], device=0) | |
return summarization_model(input, max_new_tokens = max_new_tokens, length_penalty=length_penalty, no_repeat_ngram_size=no_repeat_ngram_size)[0]["generated_text"] | |
def categorize(input): | |
ttc = pipeline(model="boun-tabi-LMG/turna_classification_ttc4900", device=0) | |
return ttc(input, max_new_tokens = 8)[0]["generated_text"] | |
def turna(input, max_new_tokens, length_penalty, | |
top_k, top_p, temp, num_beams, | |
do_sample, no_repeat_ngram_size, repetition_penalty, turna_model_version): | |
turna = pipeline(model=f"boun-tabi-LMG/{turna_model_version}", device=0) | |
input = f"[S2S] {input}<EOS>" | |
return turna(input, max_new_tokens = max_new_tokens, length_penalty=length_penalty, | |
top_k=top_k, top_p=top_p, temperature=temp, num_beams=num_beams, | |
do_sample = do_sample, no_repeat_ngram_size=no_repeat_ngram_size, repetition_penalty=repetition_penalty)[0]["generated_text"] | |
with gr.Blocks(theme="abidlabs/Lime") as demo: | |
gr.Markdown("# TURNA") | |
gr.Image("images/turna-logo.png", width=100, show_label=False, show_download_button=False, show_share_button=False) | |
with gr.Tab("TURNA"): | |
gr.Markdown(DESCRIPTION) | |
with gr.Tab("Sentiment Analysis"): | |
gr.Markdown("TURNA fine-tuned on sentiment analysis. Enter text to analyse sentiment and pick the model (tweets or product reviews).") | |
with gr.Column(): | |
with gr.Row(): | |
with gr.Column(): | |
sentiment_choice = gr.Radio(choices = ["turna_classification_17bintweet_sentiment", "turna_classification_tr_product_reviews"], label ="Model", value="turna_classification_17bintweet_sentiment") | |
sentiment_input = gr.Textbox(label="Sentiment Analysis Input") | |
sentiment_submit = gr.Button() | |
sentiment_output = gr.Textbox(label="Sentiment Analysis Output") | |
sentiment_submit.click(sentiment_analysis, inputs=[sentiment_input, sentiment_choice], outputs=sentiment_output) | |
sentiment_examples = gr.Examples(examples = sentiment_example, inputs = [sentiment_input, sentiment_choice], outputs=sentiment_output, fn=sentiment_analysis) | |
with gr.Tab("Text Categorization"): | |
gr.Markdown("TURNA fine-tuned on text categorization. Enter text to categorize text or try the example.") | |
with gr.Column(): | |
with gr.Row(): | |
with gr.Column(): | |
text_input = gr.Textbox(label="Text Categorization Input") | |
text_submit = gr.Button() | |
text_output = gr.Textbox(label="Text Categorization Output") | |
text_submit.click(categorize, inputs=[text_input], outputs=text_output) | |
text_examples = gr.Examples(examples = text_category_example,inputs=[text_input], outputs=text_output, fn=categorize) | |
with gr.Tab("NLI & STS"): | |
gr.Markdown("TURNA fine-tuned on natural language inference or semantic textual similarity. Enter text to infer entailment or measure semantic similarity. ") | |
with gr.Column(): | |
with gr.Row(): | |
with gr.Column(): | |
nli_choice = gr.Radio(choices = ["turna_nli_nli_tr", "turna_semantic_similarity_stsb_tr"], label ="Model", value="turna_nli_nli_tr") | |
nli_first_input = gr.Textbox(label="First Sentence") | |
nli_second_input = gr.Textbox(label="Second Sentence") | |
nli_submit = gr.Button() | |
nli_output = gr.Textbox(label="NLI Output") | |
nli_submit.click(nli, inputs=[nli_first_input, nli_second_input, nli_choice], outputs=nli_output) | |
nli_examples = gr.Examples(examples = nli_example, inputs = [nli_first_input, nli_second_input, nli_choice], outputs=nli_output, fn=nli) | |
with gr.Tab("POS"): | |
gr.Markdown("TURNA fine-tuned on part-of-speech-tagging. Enter text to parse parts of speech and pick the model.") | |
with gr.Column(): | |
with gr.Row(): | |
with gr.Column(): | |
pos_choice = gr.Radio(choices = ["turna_pos_imst", "turna_pos_boun"], label ="Model", value="turna_pos_imst") | |
with gr.Accordion("Advanced Generation Parameters"): | |
max_new_tokens = gr.Slider(label = "Maximum length", | |
minimum = 0, | |
maximum = 64, | |
value = 64) | |
length_penalty = gr.Slider(label = "Length penalty", | |
minimum = -10, | |
maximum = 10, | |
value=2.0) | |
no_repeat_ngram_size =gr.Slider(label="No Repeat N-Gram Size", minimum=0,value=3,) | |
with gr.Column(): | |
pos_input = gr.Textbox(label="POS Input") | |
pos_submit = gr.Button() | |
pos_output = gr.Textbox(label="POS Output") | |
pos_submit.click(pos, inputs=[pos_input, pos_choice, max_new_tokens, length_penalty, no_repeat_ngram_size], outputs=pos_output) | |
pos_examples = gr.Examples(examples = ner_example, inputs = [pos_input, pos_choice, max_new_tokens, length_penalty, no_repeat_ngram_size], outputs=pos_output, fn=pos) | |
with gr.Tab("NER"): | |
gr.Markdown("TURNA fine-tuned on named entity recognition. Enter text to parse named entities and pick the model.") | |
with gr.Column(): | |
with gr.Row(): | |
with gr.Column(): | |
ner_choice = gr.Radio(choices = ["turna_ner_wikiann", "turna_ner_milliyet"], label ="Model", value="turna_ner_wikiann") | |
with gr.Accordion("Advanced Generation Parameters"): | |
max_new_tokens = gr.Slider(label = "Maximum length", | |
minimum = 0, | |
maximum = 64, | |
value = 64) | |
length_penalty = gr.Slider(label = "Length penalty", | |
minimum = -10, | |
maximum = 10, | |
value=2.0) | |
no_repeat_ngram_size =gr.Slider(label="No Repeat N-Gram Size", minimum=0,value=3,) | |
with gr.Column(): | |
ner_input = gr.Textbox(label="NER Input") | |
ner_submit = gr.Button() | |
ner_output = gr.Textbox(label="NER Output") | |
ner_submit.click(ner, inputs=[ner_input, ner_choice, max_new_tokens, length_penalty, no_repeat_ngram_size], outputs=ner_output) | |
ner_examples = gr.Examples(examples = ner_example, inputs = [ner_input, ner_choice, max_new_tokens, length_penalty, no_repeat_ngram_size], outputs=ner_output, fn=ner) | |
with gr.Tab("Paraphrase"): | |
gr.Markdown("TURNA fine-tuned on paraphrasing. Enter text to paraphrase and pick the model.") | |
with gr.Column(): | |
with gr.Row(): | |
with gr.Column(): | |
paraphrasing_choice = gr.Radio(choices = ["turna_paraphrasing_tatoeba", "turna_paraphrasing_opensubtitles"], label ="Model", value="turna_paraphrasing_tatoeba") | |
with gr.Accordion("Advanced Generation Parameters"): | |
max_new_tokens = gr.Slider(label = "Maximum length", | |
minimum = 0, | |
maximum = 20, | |
value = 20) | |
with gr.Column(): | |
paraphrasing_input = gr.Textbox(label = "Paraphrasing Input") | |
paraphrasing_submit = gr.Button() | |
paraphrasing_output = gr.Text(label="Paraphrasing Output") | |
paraphrasing_submit.click(paraphrase, inputs=[paraphrasing_input, paraphrasing_choice, max_new_tokens], outputs=paraphrasing_output) | |
paraphrase_examples = gr.Examples(examples = long_text, inputs = [paraphrasing_input, paraphrasing_choice, max_new_tokens], outputs=paraphrasing_output, fn=paraphrase) | |
with gr.Tab("Summarization"): | |
gr.Markdown("TURNA fine-tuned on summarization. Enter text to summarize and pick the model.") | |
with gr.Column(): | |
with gr.Row(): | |
with gr.Column(): | |
sum_choice = gr.Radio(choices = ["turna_summarization_mlsum", "turna_summarization_tr_news"], label ="Model", value="turna_summarization_mlsum") | |
with gr.Accordion("Advanced Generation Parameters"): | |
max_new_tokens = gr.Slider(label = "Maximum length", | |
minimum = 0, | |
maximum = 512, | |
value = 128) | |
length_penalty = gr.Slider(label = "Length penalty", | |
minimum = -10, | |
maximum = 10, | |
value=2.0) | |
no_repeat_ngram_size =gr.Slider(label="No Repeat N-Gram Size", minimum=0,value=3,) | |
with gr.Column(): | |
sum_input = gr.Textbox(label = "Summarization Input") | |
sum_submit = gr.Button() | |
sum_output = gr.Textbox(label = "Summarization Output") | |
sum_submit.click(summarize, inputs=[sum_input, sum_choice, max_new_tokens, length_penalty, no_repeat_ngram_size], outputs=sum_output) | |
sum_examples = gr.Examples(examples = long_text, inputs = [sum_input, sum_choice, max_new_tokens, length_penalty, no_repeat_ngram_size], outputs=sum_output, fn=summarize) | |
with gr.Tab("Title Generation"): | |
gr.Markdown("TURNA fine-tuned on news title generation. Enter news text to generate a title.") | |
with gr.Column(): | |
with gr.Row(): | |
with gr.Column(): | |
title_choice = gr.Radio(choices = ["turna_title_generation_tr_news", "turna_title_generation_mlsum"], label ="Model", value="turna_title_generation_tr_news") | |
with gr.Accordion("Advanced Generation Parameters"): | |
max_new_tokens = gr.Slider(label = "Maximum length", | |
minimum = 0, | |
maximum = 64, | |
value = 64) | |
length_penalty = gr.Slider(label = "Length penalty", | |
minimum = -10, | |
maximum = 10, | |
value=2.0) | |
no_repeat_ngram_size =gr.Slider(label="No Repeat N-Gram Size", minimum=0,value=3,) | |
with gr.Column(): | |
title_input = gr.Textbox(label = "News Title Generation Input") | |
title_submit = gr.Button() | |
title_output = gr.Textbox(label = "News Title Generation Output") | |
title_submit.click(generate_title, inputs=[title_input, title_choice, max_new_tokens, length_penalty, no_repeat_ngram_size], outputs=title_output) | |
title_examples = gr.Examples(examples = long_text, inputs = [title_input, title_choice, max_new_tokens, length_penalty, no_repeat_ngram_size], outputs=title_output, fn=generate_title) | |
with gr.Tab("Text Generation"): | |
gr.Markdown("Pre-trained TURNA. Enter text to start generating.") | |
with gr.Column(): | |
with gr.Row(): | |
with gr.Column(): | |
with gr.Accordion("Advanced Generation Parameters"): | |
max_new_tokens = gr.Slider(label = "Maximum length", | |
minimum = 0, | |
maximum = 512, | |
value = 128) | |
length_penalty = gr.Slider(label = "Length penalty", | |
value=1.0) | |
top_k = gr.Slider(label = "Top-k", value=10) | |
top_p = gr.Slider(label = "Top-p", value=0.95) | |
temp = gr.Slider(label = "Temperature", value=1.0, minimum=0.1, maximum=100.0) | |
no_repeat_ngram_size =gr.Slider(label="No Repeat N-Gram Size", minimum=0,value=3,) | |
repetition_penalty = gr.Slider(label = "Repetition Penalty", minimum=0.0, value=3.1, step=0.1) | |
num_beams = gr.Slider(label = "Number of beams", minimum=1, | |
maximum=10, value=3) | |
do_sample = gr.Radio(choices = [True, False], value = True, label = "Sampling") | |
turna_model_version = gr.Radio(choices = ["TURNA", "TURNA-2850K", "TURNA-4350K"], value = "TURNA", label = "Choose TURNA model version") | |
with gr.Column(): | |
text_gen_input = gr.Textbox(label="Text Generation Input") | |
text_gen_submit = gr.Button() | |
text_gen_output = gr.Textbox(label="Text Generation Output") | |
text_gen_submit.click(turna, inputs=[text_gen_input, max_new_tokens, length_penalty, | |
top_k, top_p, temp, num_beams, | |
do_sample, no_repeat_ngram_size, repetition_penalty, turna_model_version], outputs=text_gen_output) | |
text_gen_example = [["Bir varmış, bir yokmuş, evvel zaman içinde, kalbur saman içinde, uzak diyarların birinde bir turna"]] | |
text_gen_examples = gr.Examples(examples = text_gen_example, inputs = [text_gen_input, max_new_tokens, length_penalty, | |
top_k, top_p, temp, num_beams, do_sample, no_repeat_ngram_size, repetition_penalty, turna_model_version], outputs=text_gen_output, fn=turna) | |
gr.Markdown(CITATION) | |
demo.launch() |