Hampetiudo
commited on
Commit
•
0be2dae
1
Parent(s):
11a4136
Update README.md
Browse files
README.md
CHANGED
@@ -3,4 +3,14 @@ license: gemma
|
|
3 |
base_model:
|
4 |
- ifable/gemma-2-Ifable-9B
|
5 |
pipeline_tag: text-generation
|
6 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
base_model:
|
4 |
- ifable/gemma-2-Ifable-9B
|
5 |
pipeline_tag: text-generation
|
6 |
+
---
|
7 |
+
|
8 |
+
## Llama.cpp imatrix quants of gemma-2-Ifable-9B
|
9 |
+
|
10 |
+
Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/b3804">b3804</a> for quantization.
|
11 |
+
|
12 |
+
Original model: https://huggingface.co/ifable/gemma-2-Ifable-9B
|
13 |
+
|
14 |
+
All quants made using the imatrix option with the dataset from [here](https://gist.github.com/tristandruyen/9e207a95c7d75ddf37525d353e00659c) and with a context size of 8192.
|
15 |
+
|
16 |
+
I also uploaded the BF16 GGUF because I'm too lazy to make every single quant and also in case the original model gets taken down.
|