updated readme
Browse files
README.md
CHANGED
@@ -56,7 +56,7 @@ source repo: [BSC-LT/salamandra-2b-instruct](https://huggingface.co/BSC-LT/salam
|
|
56 |
### **Notes:**
|
57 |
|
58 |
- **Recommended Quantizations:**
|
59 |
-
- **Q4_K_S:** Although it offers good size reduction with minimal PPL impact, it is superseded by more optimal choices like Q5_K_M and Q6_K.
|
60 |
- **Q5_K_M:** Offers the best balance between low perplexity and reduced file size above Q4, making it ideal for most applications.
|
61 |
- **Q6_K:** Delivers nearly lossless performance compared to bf16 with a reduced file size (2.4G vs. 4.2G). Ideal for scenarios requiring maximum accuracy with some size savings.
|
62 |
- **Non-recommended Quantizations:**
|
|
|
56 |
### **Notes:**
|
57 |
|
58 |
- **Recommended Quantizations:**
|
59 |
+
- **Q4_K_S:** Although it offers good size reduction with minimal PPL impact, it is superseded by more optimal choices like Q5_K_M and Q6_K, but it is the only model with minimal PPL impact below 2GB.
|
60 |
- **Q5_K_M:** Offers the best balance between low perplexity and reduced file size above Q4, making it ideal for most applications.
|
61 |
- **Q6_K:** Delivers nearly lossless performance compared to bf16 with a reduced file size (2.4G vs. 4.2G). Ideal for scenarios requiring maximum accuracy with some size savings.
|
62 |
- **Non-recommended Quantizations:**
|