abdulmatinomotoso's picture
Create app.py
4d45cee
raw
history blame
1.97 kB
#importing the necessary libraries
import gradio as gr
import numpy as np
import pandas as pd
import re
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
#Defining the labels of the models
labels = ["Explicit", "Not_Explicit"]
#Defining the models and tokenuzer
model_name = 'valurank/finetuned-distilbert-explicit_content_detection'
model = AutoModelForSequenceClassification.from_pretrained(model_name)
tokenizer = AutoTokenizer.from_pretrained(model_name)
#Reading in the text file
def read_in_text(url):
with open(url, 'r') as file:
article = file.read()
return article
def clean_text(url):
text = url
text = text.encode("ascii", errors="ignore").decode(
"ascii"
) # remove non-ascii, Chinese characters
text = re.sub(r"\n", " ", text)
text = re.sub(r"\n\n", " ", text)
text = re.sub(r"\t", " ", text)
text = text.strip(" ")
text = re.sub(
" +", " ", text
).strip() # get rid of multiple spaces and replace with a single
text = re.sub(r'Date\s\d{1,2}\/\d{1,2}\/\d{4}', '', text) #remove date
text = re.sub(r'\d{1,2}:\d{2}\s[A-Z]+\s[A-Z]+', '', text) #remove time
return text
#Defining a function to get the category of the news article
def get_category(file):
text = clean_text(file)
input_tensor = tokenizer.encode(text, return_tensors='pt', truncation=True)
logits = model(input_tensor).logits
softmax = torch.nn.Softmax(dim=1)
probs = softmax(logits)[0]
probs = probs.cpu().detach().numpy()
max_index = np.argmax(probs)
emotion = labels[max_index]
return emotion
#Creating the interface for the radio app
demo = gr.Interface(get_category, inputs=gr.inputs.Textbox(label='Drop your articles here'),
outputs = 'text',
title='Explicit Content Detection')
#Launching the gradio app
if __name__ == '__main__':
demo.launch(debug=True)