ctranslate2-4you
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -58,19 +58,19 @@ import ctranslate2
|
|
58 |
from transformers import AutoTokenizer
|
59 |
import sys
|
60 |
|
61 |
-
def generate_response(prompt, system_message, model_path):
|
62 |
generator = ctranslate2.Generator(model_path, device="cuda", compute_type="int8")
|
63 |
tokenizer = AutoTokenizer.from_pretrained(model_path)
|
64 |
formatted_prompt = f"""<|im_start|>system<|im_sep|>{system_message}<|im_end|>
|
65 |
<|im_start|>user<|im_sep|>{prompt}<|im_end|>
|
66 |
<|im_start|>assistant<|im_sep|>"""
|
67 |
-
tokens = tokenizer.tokenize(formatted_prompt)
|
68 |
for step in generator.generate_tokens([tokens], max_length=1024, sampling_temperature=0.7):
|
69 |
-
token = step.token
|
70 |
-
decoded_token = tokenizer.decode([step.token_id])
|
71 |
-
print(decoded_token, end="", flush=True)
|
72 |
if token in tokenizer.eos_token or token in tokenizer.all_special_tokens:
|
73 |
break
|
|
|
|
|
74 |
|
75 |
if __name__ == "__main__":
|
76 |
model_path = "path/to/your/phi-4-ct2-model"
|
@@ -78,6 +78,5 @@ if __name__ == "__main__":
|
|
78 |
user_prompt = "Write a short poem about a cat."
|
79 |
print("\nGenerating response:")
|
80 |
generate_response(user_prompt, system_message, model_path)
|
81 |
-
|
82 |
```
|
83 |
</details>
|
|
|
58 |
from transformers import AutoTokenizer
|
59 |
import sys
|
60 |
|
61 |
+
def generate_response(prompt: str, system_message: str, model_path: str) -> None:
|
62 |
generator = ctranslate2.Generator(model_path, device="cuda", compute_type="int8")
|
63 |
tokenizer = AutoTokenizer.from_pretrained(model_path)
|
64 |
formatted_prompt = f"""<|im_start|>system<|im_sep|>{system_message}<|im_end|>
|
65 |
<|im_start|>user<|im_sep|>{prompt}<|im_end|>
|
66 |
<|im_start|>assistant<|im_sep|>"""
|
67 |
+
tokens: List[str] = tokenizer.tokenize(formatted_prompt)
|
68 |
for step in generator.generate_tokens([tokens], max_length=1024, sampling_temperature=0.7):
|
69 |
+
token: str = step.token
|
|
|
|
|
70 |
if token in tokenizer.eos_token or token in tokenizer.all_special_tokens:
|
71 |
break
|
72 |
+
decoded_token: str = tokenizer.decode([step.token_id])
|
73 |
+
print(decoded_token, end="", flush=True)
|
74 |
|
75 |
if __name__ == "__main__":
|
76 |
model_path = "path/to/your/phi-4-ct2-model"
|
|
|
78 |
user_prompt = "Write a short poem about a cat."
|
79 |
print("\nGenerating response:")
|
80 |
generate_response(user_prompt, system_message, model_path)
|
|
|
81 |
```
|
82 |
</details>
|