Spaces:
Sleeping
Sleeping
toaster61
commited on
Commit
·
18aac0c
1
Parent(s):
021692e
oops fixing silly mistakes
Browse files- README.md +1 -1
- gradio_app.py +0 -35
README.md
CHANGED
@@ -7,4 +7,4 @@ sdk: docker
|
|
7 |
pinned: true
|
8 |
---
|
9 |
|
10 |
-
I build this just for fun. Please clone this space to personal use. Test API you can here.
|
|
|
7 |
pinned: true
|
8 |
---
|
9 |
|
10 |
+
I build this just for fun. Please clone this space to personal use. Test API you can here.
|
gradio_app.py
CHANGED
@@ -36,41 +36,6 @@ Or you can once follow steps in Dockerfile and try it on your machine, not in Do
|
|
36 |
with open('system.prompt', 'r', encoding='utf-8') as f:
|
37 |
prompt = f.read()
|
38 |
|
39 |
-
# this model was loaded from https://hf.co/models
|
40 |
-
model = AutoModelForSeq2SeqLM.from_pretrained("facebook/nllb-200-distilled-600M")
|
41 |
-
tokenizer = AutoTokenizer.from_pretrained("facebook/nllb-200-distilled-600M")
|
42 |
-
device = 0 if torch.cuda.is_available() else -1
|
43 |
-
LANGS = ["ace_Arab", "eng_Latn", "fra_Latn", "spa_Latn"]
|
44 |
-
|
45 |
-
def t1ranslate(text, src_lang, tgt_lang):
|
46 |
-
try:
|
47 |
-
maxTokens = data.get("max_tokens", 64)
|
48 |
-
if isinstance(data.get("system_prompt"), str):
|
49 |
-
userPrompt = data.get("system_prompt") + "\n\nUser: " + data['request'] + "\nAssistant: "
|
50 |
-
else:
|
51 |
-
userPrompt = prompt + "\n\nUser: " + data['request'] + "\nAssistant: "
|
52 |
-
except:
|
53 |
-
return {"error": "Not enough data", "output": "Oops! Error occured! If you're a developer, using this API, check 'error' key."}, 400
|
54 |
-
try:
|
55 |
-
output = llm(userPrompt, max_tokens=maxTokens, stop=["User:", "\n"], echo=False)
|
56 |
-
text = output["choices"][0]["text"]
|
57 |
-
# i allowed only certain languages:
|
58 |
-
# russian (ru), ukranian (uk), chinese (zh)
|
59 |
-
if isinstance(data.get("target_lang"), str) and data.get("target_lang").lower() in ["ru", "uk", "zh"]:
|
60 |
-
encoded_input = translator_tokenizer(output, return_tensors="pt")
|
61 |
-
generated_tokens = translator_model.generate(
|
62 |
-
**encoded_input, forced_bos_token_id=translator_tokenizer.get_lang_id(data.get("target_lang"))
|
63 |
-
)
|
64 |
-
translated_text = translator_tokenizer.batch_decode(
|
65 |
-
generated_tokens, skip_special_tokens=True
|
66 |
-
)[0]
|
67 |
-
return {"output": text, "translated_output": translated_text}
|
68 |
-
|
69 |
-
return {"output": text}
|
70 |
-
except Exception as e:
|
71 |
-
print(e)
|
72 |
-
return {"error": str(e), "output": "Oops! Internal server error. Check the logs. If you're a developer, using this API, check 'error' key."}, 500
|
73 |
-
|
74 |
def translate(request: str, max_tokens: int = 256, language: str = "en", custom_prompt: str = None):
|
75 |
try:
|
76 |
maxTokens = max_tokens if 16 <= max_tokens <= 256 else 64
|
|
|
36 |
with open('system.prompt', 'r', encoding='utf-8') as f:
|
37 |
prompt = f.read()
|
38 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
39 |
def translate(request: str, max_tokens: int = 256, language: str = "en", custom_prompt: str = None):
|
40 |
try:
|
41 |
maxTokens = max_tokens if 16 <= max_tokens <= 256 else 64
|