Update README.md
Browse files
README.md
CHANGED
@@ -70,7 +70,7 @@ I have tested CUDA acceleration and it works great. I have not yet tested other
|
|
70 |
## Prompt template: Mistral
|
71 |
|
72 |
```
|
73 |
-
|
74 |
```
|
75 |
|
76 |
<!-- prompt-template end -->
|
@@ -171,7 +171,7 @@ Windows Command Line users: You can set the environment variable by running `set
|
|
171 |
Make sure you are using `llama.cpp` from commit [d0cee0d](https://github.com/ggerganov/llama.cpp/commit/d0cee0d36d5be95a0d9088b674dbb27354107221) or later.
|
172 |
|
173 |
```shell
|
174 |
-
./main -ngl 35 -m mixtral-8x7b-instruct-v0.1.Q4_K_M.gguf --color -c 32768 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "
|
175 |
```
|
176 |
|
177 |
Change `-ngl 32` to the number of layers to offload to GPU. Remove it if you don't have GPU acceleration.
|
|
|
70 |
## Prompt template: Mistral
|
71 |
|
72 |
```
|
73 |
+
[INST] {prompt} [/INST]
|
74 |
```
|
75 |
|
76 |
<!-- prompt-template end -->
|
|
|
171 |
Make sure you are using `llama.cpp` from commit [d0cee0d](https://github.com/ggerganov/llama.cpp/commit/d0cee0d36d5be95a0d9088b674dbb27354107221) or later.
|
172 |
|
173 |
```shell
|
174 |
+
./main -ngl 35 -m mixtral-8x7b-instruct-v0.1.Q4_K_M.gguf --color -c 32768 --temp 0.7 --repeat_penalty 1.1 -n -1 -p "[INST] {prompt} [/INST]"
|
175 |
```
|
176 |
|
177 |
Change `-ngl 32` to the number of layers to offload to GPU. Remove it if you don't have GPU acceleration.
|