readme: TODO implement llamafile
Browse files
README.md
CHANGED
@@ -19,6 +19,8 @@ Quantizised from [https://huggingface.co/deepseek-ai/DeepSeek-V2-Chat](https://h
|
|
19 |
|
20 |
Using llama.cpp fork: [https://github.com/fairydreaming/llama.cpp/tree/deepseek-v2](https://github.com/fairydreaming/llama.cpp/tree/deepseek-v2)
|
21 |
|
|
|
|
|
22 |
# Warning: This will not work unless you compile llama.cpp from the repo provided (and set metadata KV overrides)!
|
23 |
|
24 |
# How to use:
|
|
|
19 |
|
20 |
Using llama.cpp fork: [https://github.com/fairydreaming/llama.cpp/tree/deepseek-v2](https://github.com/fairydreaming/llama.cpp/tree/deepseek-v2)
|
21 |
|
22 |
+
TODO: Make llamafile for Q2_K and Q4_K_M
|
23 |
+
|
24 |
# Warning: This will not work unless you compile llama.cpp from the repo provided (and set metadata KV overrides)!
|
25 |
|
26 |
# How to use:
|