Update README.md
Browse files
README.md
CHANGED
@@ -15,6 +15,8 @@ This is GGML format quantised 4-bit, 5-bit and 8-bit [MosaicML's MPT-7B-Instruct
|
|
15 |
|
16 |
This repo is the result of converting to GGML and quantising.
|
17 |
|
|
|
|
|
18 |
## Repositories available
|
19 |
|
20 |
* [MPT-7B: 4-bit, 5-bit and 8-bit GGML models for CPU (+CUDA) inference](https://huggingface.co/TheBloke/MPT-7B-GGML).
|
|
|
15 |
|
16 |
This repo is the result of converting to GGML and quantising.
|
17 |
|
18 |
+
Please note that these MPT GGMLs are **not compatbile with llama.cpp**. Right now they can only be used in a couple of tools. This is bound to improve in the near future.
|
19 |
+
|
20 |
## Repositories available
|
21 |
|
22 |
* [MPT-7B: 4-bit, 5-bit and 8-bit GGML models for CPU (+CUDA) inference](https://huggingface.co/TheBloke/MPT-7B-GGML).
|