ctranslate2-4you commited on
Commit
4db1aeb
·
verified ·
1 Parent(s): 0699039

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +54 -0
README.md ADDED
@@ -0,0 +1,54 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: ctranslate2
3
+ license: mit
4
+ base_model:
5
+ - microsoft/phi-4
6
+ base_model_relation: quantized
7
+ tags:
8
+ - ctranslate2
9
+ - phi-4
10
+ - chat
11
+ ---
12
+
13
+ Sample Script:
14
+
15
+ ```python
16
+ import ctranslate2
17
+ from transformers import AutoTokenizer
18
+
19
+ def generate_response(prompt, system_message, model_path):
20
+ # Initialize the model and tokenizer
21
+ generator = ctranslate2.Generator(
22
+ model_path,
23
+ device="cuda",
24
+ compute_type="int8"
25
+ )
26
+ tokenizer = AutoTokenizer.from_pretrained(model_path)
27
+
28
+ # Format the prompt
29
+ formatted_prompt = f"""<|im_start|>system<|im_sep|>{system_message}<|im_end|>
30
+ <|im_start|>user<|im_sep|>{prompt}<|im_end|>
31
+ <|im_start|>assistant<|im_sep|>"""
32
+
33
+ # Tokenize and generate
34
+ tokens = tokenizer.convert_ids_to_tokens(tokenizer.encode(formatted_prompt))
35
+ results = generator.generate_batch(
36
+ [tokens],
37
+ max_batch_size=8192,
38
+ max_length=1024,
39
+ sampling_temperature=0.7
40
+ )
41
+
42
+ # Decode and return the response
43
+ response = tokenizer.decode(results[0].sequences_ids[0])
44
+ return response
45
+
46
+ if __name__ == "__main__":
47
+ model_path = "path/to/your/phi-4-ct2-model"
48
+ system_message = "You are a helpful AI assistant."
49
+ user_prompt = "Write a short poem about a cat."
50
+
51
+ response = generate_response(user_prompt, system_message, model_path)
52
+ print("\nGenerated response:")
53
+ print(response)
54
+ ```