Edit model card

GGUFs PLUS:

Q8 and Q6 GGUFs with critical parts of the model in F16 / Full precision.

File sizes will be slightly larger than standard, but should yeild higher quality results under all tasks and conditions.

Downloads last month
27
GGUF
Model size
10.7B params
Architecture
llama

6-bit

8-bit

Inference Examples
Unable to determine this model's library. Check the docs .

Collection including DavidAU/LemonadeRP-4.5.3-11B-GGUF-Plus