Update README.md
Browse files
README.md
CHANGED
@@ -13,20 +13,22 @@ quantized_by: bartowski
|
|
13 |
|
14 |
## Llamacpp imatrix Quantizations of gemma-2-27b-it
|
15 |
|
16 |
-
Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/
|
17 |
|
18 |
Original model: https://huggingface.co/google/gemma-2-27b-it
|
19 |
|
20 |
All quants made using imatrix option with dataset from [here](https://gist.github.com/bartowski1182/eb213dccb3571f863da82e99418f81e8)
|
21 |
|
|
|
|
|
|
|
|
|
22 |
## Prompt format
|
23 |
|
24 |
```
|
25 |
-
<
|
26 |
{prompt}<end_of_turn>
|
27 |
<start_of_turn>model
|
28 |
-
<end_of_turn>
|
29 |
-
<start_of_turn>model
|
30 |
|
31 |
```
|
32 |
|
|
|
13 |
|
14 |
## Llamacpp imatrix Quantizations of gemma-2-27b-it
|
15 |
|
16 |
+
Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/b3278">b3278</a> for quantization.
|
17 |
|
18 |
Original model: https://huggingface.co/google/gemma-2-27b-it
|
19 |
|
20 |
All quants made using imatrix option with dataset from [here](https://gist.github.com/bartowski1182/eb213dccb3571f863da82e99418f81e8)
|
21 |
|
22 |
+
## What's new
|
23 |
+
|
24 |
+
- July 21 2024: Contains latest tokenizer fixes, which addressed a few oddities from the original fix, should be closest to correct performance yet. Also has metadata for SWA and logit softcapping.
|
25 |
+
|
26 |
## Prompt format
|
27 |
|
28 |
```
|
29 |
+
<start_of_turn>user
|
30 |
{prompt}<end_of_turn>
|
31 |
<start_of_turn>model
|
|
|
|
|
32 |
|
33 |
```
|
34 |
|