|
--- |
|
license: llama3.2 |
|
base_model: meta-llama/Llama-3.2-1B-Instruct |
|
pipeline_tag: text-generation |
|
quanted_by: grimjim |
|
--- |
|
EXL2 quants of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct/tree/main) by branch: |
|
- 4_0 : [4.0 bits per weight](https://huggingface.co/grimjim/meta-llama-Llama-3.2-1B-Instruct-exl2/tree/4_0) |
|
- 5_0 : [5.0 bits per weight](https://huggingface.co/grimjim/meta-llama-Llama-3.2-1B-Instruct-exl2/tree/5_0) |
|
- 6_0 : [6.0 bits per weight](https://huggingface.co/grimjim/meta-llama-Llama-3.2-1B-Instruct-exl2/tree/6_0) |
|
- 8_0 : [8.0 bits per weight](https://huggingface.co/grimjim/meta-llama-Llama-3.2-1B-Instruct-exl2/tree/8_0) |
|
|
|
Make your own EXL2 quants with |
|
[measurement.json](https://huggingface.co/grimjim/meta-llama-Llama-3.2-1B-Instruct-exl2/blob/main/measurement.json). |
|
|
|
Quanted with [exllamav2](https://github.com/turboderp/exllamav2) v0.2.4. |