Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -52,9 +52,13 @@ def _load_model_tokenizer():
|
|
52 |
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
|
53 |
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto",trust_remote_code=True, fp16=True).eval()
|
54 |
return model, tokenizer
|
55 |
-
|
56 |
-
|
57 |
-
|
|
|
|
|
|
|
|
|
58 |
def postprocess(self, y):
|
59 |
if y is None:
|
60 |
return []
|
@@ -96,7 +100,7 @@ def _parse_text(text):
|
|
96 |
lines[i] = "<br>" + line
|
97 |
text = "".join(lines)
|
98 |
return text
|
99 |
-
|
100 |
def predict(_query, _chatbot, _task_history):
|
101 |
print(f"User: {_parse_text(_query)}")
|
102 |
_chatbot.append((_parse_text(_query), ""))
|
|
|
52 |
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
|
53 |
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto",trust_remote_code=True, fp16=True).eval()
|
54 |
return model, tokenizer
|
55 |
+
|
56 |
+
if torch.cuda.is_available():
|
57 |
+
whisper_model = load_whisper()
|
58 |
+
tts_model, tokenizer_tss = load_tts()
|
59 |
+
model, tokenizer = _load_model_tokenizer()
|
60 |
+
|
61 |
+
|
62 |
def postprocess(self, y):
|
63 |
if y is None:
|
64 |
return []
|
|
|
100 |
lines[i] = "<br>" + line
|
101 |
text = "".join(lines)
|
102 |
return text
|
103 |
+
@spaces.GPU
|
104 |
def predict(_query, _chatbot, _task_history):
|
105 |
print(f"User: {_parse_text(_query)}")
|
106 |
_chatbot.append((_parse_text(_query), ""))
|