explorewithai's picture
Update app.py
f0a68fe verified
raw
history blame
662 Bytes
import gradio as gr
from transformers import pipeline
import torch
import spaces
device = 0 if torch.cuda.is_available() else -1
@spaces.GPU
def generate_response(user_input, history):
pipe = pipeline("text-generation", model="explorewithai/ChatFrame-Uncensored-Instruct-Small", device = device)
messages = [
{"role": "user", "content": user_input},
]
response = pipe(messages, max_length=512)
return response[0]['generated_text'][1]["content"]
iface = gr.ChatInterface(
fn=generate_response,
title="Text Generation Chatbot",
description="Enter your text and get a generated response from the model."
)
iface.launch()