ibalampanis
commited on
Commit
•
2c891bd
1
Parent(s):
e32a928
Update README.md
Browse files
README.md
CHANGED
@@ -28,9 +28,9 @@ This repository contains GGUF format model files for [ilsp's Meltemi 7B Instruct
|
|
28 |
|
29 |
| Name | Quantization Method | Precision (Bits) | File Size | Max RAM Required | Use Case |
|
30 |
| --------------------------------------------------------------------------------------------------------------------------------------- | ------------------- | ---------------- | --------- | ---------------- | ------------------------------------------------------------- |
|
31 |
-
| [meltemi-7b-instruct-v1_q8_0.gguf](https://huggingface.co/SPAHE/Meltemi-7B-Instruct-v1-GGUF/blob/main/meltemi-7b-instruct-v1_q8_0.gguf) | Q8_0 | 8 | 7.
|
32 |
-
| [meltemi-7b-instruct-v1_f16.gguf](https://huggingface.co/SPAHE/Meltemi-7B-Instruct-v1-GGUF/blob/main/meltemi-7b-instruct-v1_f16.gguf) | F16 | 16 |
|
33 |
-
| [meltemi-7b-instruct-v1_f32.gguf](https://huggingface.co/SPAHE/Meltemi-7B-Instruct-v1-GGUF/blob/main/meltemi-7b-instruct-v1_f32.gguf) | F32 | 32 | 27.
|
34 |
|
35 |
**Note**: The above RAM figures assume no GPU offloading. If layers are offloaded to the GPU, this will reduce RAM usage and use VRAM instead.
|
36 |
|
|
|
28 |
|
29 |
| Name | Quantization Method | Precision (Bits) | File Size | Max RAM Required | Use Case |
|
30 |
| --------------------------------------------------------------------------------------------------------------------------------------- | ------------------- | ---------------- | --------- | ---------------- | ------------------------------------------------------------- |
|
31 |
+
| [meltemi-7b-instruct-v1_q8_0.gguf](https://huggingface.co/SPAHE/Meltemi-7B-Instruct-v1-GGUF/blob/main/meltemi-7b-instruct-v1_q8_0.gguf) | Q8_0 | 8 | 7.95 GB | 7.30 GB | Low quality loss - recommended |
|
32 |
+
| [meltemi-7b-instruct-v1_f16.gguf](https://huggingface.co/SPAHE/Meltemi-7B-Instruct-v1-GGUF/blob/main/meltemi-7b-instruct-v1_f16.gguf) | F16 | 16 | 15.00 GB | 14.20 GB | Very large, extremely low quality loss - recommended |
|
33 |
+
| [meltemi-7b-instruct-v1_f32.gguf](https://huggingface.co/SPAHE/Meltemi-7B-Instruct-v1-GGUF/blob/main/meltemi-7b-instruct-v1_f32.gguf) | F32 | 32 | 27.90 GB | 29.30 GB | Very very large, extremely low quality loss - not recommended |
|
34 |
|
35 |
**Note**: The above RAM figures assume no GPU offloading. If layers are offloaded to the GPU, this will reduce RAM usage and use VRAM instead.
|
36 |
|