TheBloke commited on
Commit
c5e738b
1 Parent(s): c2a2aba

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -0
README.md CHANGED
@@ -34,6 +34,22 @@ Please note that these GGMLs are **not compatible with llama.cpp, or currently w
34
  * [4, 5, and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/starchat-beta-GGML)
35
  * [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/HuggingFaceH4/starchat-beta)
36
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
37
  <!-- compatibility_ggml start -->
38
  ## Compatibilty
39
 
 
34
  * [4, 5, and 8-bit GGML models for CPU+GPU inference](https://huggingface.co/TheBloke/starchat-beta-GGML)
35
  * [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/HuggingFaceH4/starchat-beta)
36
 
37
+ ## Prompt template
38
+
39
+ ```
40
+ <|system|> system message goes here <|end|>
41
+ <|user|> prompt goes here <|end|>
42
+ <|assistant|>
43
+ ```
44
+
45
+ Example:
46
+
47
+ ```
48
+ <|system|> Below is a conversation between a human user and a helpful AI coding assistant. <|end|>
49
+ <|user|> How do I sort a list in Python? <|end|>
50
+ <|assistant|>
51
+ ```
52
+
53
  <!-- compatibility_ggml start -->
54
  ## Compatibilty
55