Update README.md
Browse files
README.md
CHANGED
@@ -27,9 +27,10 @@ I have the following Vicuna 1.1 repositories available:
|
|
27 |
* [GPTQ quantized 4bit 7B 1.1 for GPU - `safetensors` and `pt` formats](https://huggingface.co/TheBloke/vicuna-7B-1.1-GPTQ-4bit-128g)
|
28 |
|
29 |
**GGMLs for CPU inference**
|
|
|
30 |
I removed the GGMLs I originally made for Vicuna 1.1 because they were directly converted GPTQ -> GGML and this seemed to give poor results
|
31 |
|
32 |
-
Instead I recommend you
|
33 |
* [eachadea's Vicuna 13B 1.1 GGML format for `llama.cpp`](https://huggingface.co/eachadea/ggml-vicuna-13b-1.1)
|
34 |
* [eachadea's Vicuna 7B 1.1 GGML format for `llama.cpp`](https://huggingface.co/eachadea/ggml-vicuna-7b-1.1)
|
35 |
|
|
|
27 |
* [GPTQ quantized 4bit 7B 1.1 for GPU - `safetensors` and `pt` formats](https://huggingface.co/TheBloke/vicuna-7B-1.1-GPTQ-4bit-128g)
|
28 |
|
29 |
**GGMLs for CPU inference**
|
30 |
+
|
31 |
I removed the GGMLs I originally made for Vicuna 1.1 because they were directly converted GPTQ -> GGML and this seemed to give poor results
|
32 |
|
33 |
+
Instead I recommend you use eachadea's GGMLs:
|
34 |
* [eachadea's Vicuna 13B 1.1 GGML format for `llama.cpp`](https://huggingface.co/eachadea/ggml-vicuna-13b-1.1)
|
35 |
* [eachadea's Vicuna 7B 1.1 GGML format for `llama.cpp`](https://huggingface.co/eachadea/ggml-vicuna-7b-1.1)
|
36 |
|