Spaces:
Sleeping
Sleeping
File size: 3,860 Bytes
f0f5479 f8c1b24 e0a4ce0 0c05d31 700ece2 0c05d31 e0a4ce0 700ece2 e0a4ce0 0c05d31 700ece2 0c05d31 700ece2 0c05d31 700ece2 0c05d31 700ece2 0c05d31 700ece2 0c05d31 700ece2 0c05d31 700ece2 f8c1b24 e0a4ce0 f8c1b24 e0a4ce0 0c05d31 700ece2 e0a4ce0 700ece2 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 |
# app.py
import os
os.system('python download.py')
from transformers import T5Tokenizer, T5ForConditionalGeneration
import gradio as gr
import nltk
from nltk.tokenize import sent_tokenize, word_tokenize
from nltk.corpus import wordnet as wn
from difflib import SequenceMatcher
# Load a pre-trained T5 model specifically fine-tuned for grammar correction
tokenizer = T5Tokenizer.from_pretrained("prithivida/grammar_error_correcter_v1")
model = T5ForConditionalGeneration.from_pretrained("prithivida/grammar_error_correcter_v1")
# Function to get the base form (lemma) of verbs
def get_base_form(word, tag):
wn_tag = {'VBD': wn.VERB, 'VBG': wn.VERB, 'VBN': wn.VERB, 'VBP': wn.VERB, 'VBZ': wn.VERB, 'VB': wn.VERB}
if tag in wn_tag:
lemma = nltk.WordNetLemmatizer().lemmatize(word, wn_tag[tag])
return lemma
return word
# Function to extract verbs from a sentence
def extract_verbs(sentence):
words = word_tokenize(sentence)
tagged = nltk.pos_tag(words)
verbs = [(word, tag) for word, tag in tagged if tag.startswith('VB')]
return verbs
# Function to perform grammar correction and generate verb forms list
def grammar_check(text):
sentences = sent_tokenize(text)
corrected_sentences = []
original_verbs = []
corrected_verbs = []
for sentence in sentences:
original_verbs.extend(extract_verbs(sentence))
input_text = f"gec: {sentence}"
input_ids = tokenizer.encode(input_text, return_tensors="pt", max_length=512, truncation=True)
outputs = model.generate(input_ids, max_length=512, num_beams=4, early_stopping=True)
corrected_sentence = tokenizer.decode(outputs[0], skip_special_tokens=True)
corrected_sentences.append(corrected_sentence)
corrected_verbs.extend(extract_verbs(corrected_sentence))
# Function to underline and color revised parts
def underline_and_color_revisions(original, corrected):
diff = SequenceMatcher(None, original.split(), corrected.split())
result = []
for tag, i1, i2, j1, j2 in diff.get_opcodes():
if tag == 'insert':
result.append(f"<u style='color:red;'>{' '.join(corrected.split()[j1:j2])}</u>")
elif tag == 'replace':
result.append(f"<u style='color:red;'>{' '.join(corrected.split()[j1:j2])}</u>")
elif tag == 'equal':
result.append(' '.join(original.split()[i1:i2]))
return " ".join(result)
corrected_text = " ".join(
underline_and_color_revisions(orig, corr) for orig, corr in zip(sentences, corrected_sentences)
)
# Generate verb forms list
verb_forms_list = []
for orig, corr in zip(original_verbs, corrected_verbs):
base_orig = get_base_form(orig[0], orig[1])
base_corr = get_base_form(corr[0], corr[1])
if base_orig != base_corr:
verb_forms_list.append(f"{base_orig}-{corr[0]}-{base_corr}")
verb_forms_str = "\n".join(verb_forms_list)
return corrected_text, verb_forms_str
# Create Gradio interface with a writing prompt
interface = gr.Interface(
fn=grammar_check,
inputs="text",
outputs=["html", "text"], # Two output boxes: HTML for corrected text, Text for verb forms list
title="Grammar Checker",
description=(
"Enter text to check for grammar mistakes.\n\n"
"Writing Prompt:\n"
"In the story, Alex and his friends discovered an ancient treasure in Whispering Hollow and decided to donate the artifacts to the local museum.\n\n"
"In the past, did you have a similar experience where you found something valuable or interesting? Tell the story. Describe what you found, what you did with it, and how you felt about your decision.\n\n"
"Remember to use past tense in your writing."
)
)
# Launch the interface
interface.launch()
|