Ayushdavidkushwahaaaa's picture
Create app.py
7054162 verified
import gradio as gr
import torch
import whisper
import warnings
warnings.filterwarnings('ignore')
from transformers import pipeline
import os
MODEL_NAME = "openai/whisper-small"
BATCH_SIZE = 8
device = 0 if torch.cuda.is_available() else "cpu"
pipe = pipeline(
task="automatic-speech-recognition",
model=MODEL_NAME,
chunk_length_s=30,
device=device)
emotion_classifier = pipeline("text-classification",model='MilaNLProc/xlm-emo-t', return_all_scores=True)
def transcribe(microphone, file_upload, task):
output = ""
if (microphone is not None) and (file_upload is not None):
warn_output = (
"WARNING: You've uploaded an audio file and used the microphone. "
"The recorded file from the microphone will be used and the uploaded audio will be discarded.\n"
)
elif (microphone is None) and (file_upload is None):
raise gr.Error("You have to either use the microphone or upload an audio file")
file = microphone if microphone is not None else file_upload
text = pipe(file, batch_size=BATCH_SIZE, generate_kwargs={"task": task})["text"]
return output + text
def translate_and_classify(audio):
text_result = transcribe(audio, None, "transcribe")
emotion = emotion_classifier(text_result)
detected_emotion = {}
for emotion in emotion[0]:
detected_emotion[emotion["label"]] = emotion["score"]
return text_result, detected_emotion
with gr.Blocks() as demo:
gr.Markdown(
""" # Emotion Detection from Speech
##### Detection of anger, sadness, joy, fear in speech using OpenAI Whisper and XLM-RoBERTa
""")
with gr.Column():
with gr.Tab("Record Audio"):
# The 'source' argument is no longer supported, use 'sources' instead
audio_input_r = gr.Audio(label = 'Record Audio Input',sources=["microphone"],type="filepath")
transcribe_audio_r = gr.Button('Transcribe')
with gr.Tab("Upload Audio as File"):
# The 'source' argument is no longer supported, use 'sources' instead
audio_input_u = gr.Audio(label = 'Upload Audio',sources=["upload"],type="filepath")
transcribe_audio_u = gr.Button('Transcribe')
with gr.Row():
transcript_output = gr.Textbox(label="Transcription in the language of speech/audio", lines = 3)
emotion_output = gr.Label(label = "Detected Emotion")
transcribe_audio_r.click(translate_and_classify, inputs = audio_input_r, outputs = [transcript_output,emotion_output])
transcribe_audio_u.click(translate_and_classify, inputs = audio_input_u, outputs = [transcript_output,emotion_output])
demo.launch(share=True)