Edit model card

The data used to train the model are on Huggingface under siacus/cap_pe_verified_final-and-last

F16 version from merged weights created with llama.cpp on a CUDA GPU and the 4bit quantized version created on a Mac M2 Ultra Metal architecture. If you want to use the 4bit quantized version on CUDA, please quantize it directly from the F16 version.

Downloads last month
34
GGUF
Model size
6.74B params
Architecture
llama

4-bit

16-bit

Inference API
Unable to determine this model's library. Check the docs .

Model tree for siacus/llama-2-7b-cap_verified-final-and-last

Quantized
(53)
this model

Dataset used to train siacus/llama-2-7b-cap_verified-final-and-last