Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,16 @@
|
|
1 |
-
---
|
2 |
-
license: llama3.2
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: llama3.2
|
3 |
+
base_model: meta-llama/Llama-3.2-1B-Instruct
|
4 |
+
pipeline_tag: text-generation
|
5 |
+
quanted_by: grimjim
|
6 |
+
---
|
7 |
+
EXL2 quants of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct/tree/main) by branch:
|
8 |
+
- 4_0 : [4.0 bits per weight](https://huggingface.co/grimjim/meta-llama-Llama-3.2-1B-Instruct-exl2/tree/4_0)
|
9 |
+
- 5_0 : [5.0 bits per weight](https://huggingface.co/grimjim/meta-llama-Llama-3.2-1B-Instruct-exl2/tree/5_0)
|
10 |
+
- 6_0 : [6.0 bits per weight](https://huggingface.co/grimjim/meta-llama-Llama-3.2-1B-Instruct-exl2/tree/6_0)
|
11 |
+
- 8_0 : [8.0 bits per weight](https://huggingface.co/grimjim/meta-llama-Llama-3.2-1B-Instruct-exl2/tree/8_0)
|
12 |
+
|
13 |
+
Make your own EXL2 quants with
|
14 |
+
[measurement.json](https://huggingface.co/grimjim/meta-llama-Llama-3.2-1B-Instruct-exl2/blob/main/measurement.json).
|
15 |
+
|
16 |
+
Quanted with [exllamav2](https://github.com/turboderp/exllamav2) v0.2.4.
|