Edit model card

GGUF quantizations for ChaoticNeutrals/Prima-LelantaclesV5-7b.

If you want any specific quantization to be added, feel free to ask.

All credits belong to the respective creators.

Base⇢ GGUF(F16)⇢ GGUF(Quants)

Using llama.cpp-b2222. For --imatrix, included reference imatrix-Q8_0.dat was used.

Original model information:

image/png

image/jpeg

https://huggingface.co/ChaoticNeutrals/Prima-LelantaclesV5-7b/tree/main/ST%20presets

This model was merged using the DARE TIES merge method.

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

merge_method: dare_ties
base_model: Test157t/Prima-LelantaclesV4-7b-16k
parameters:
  normalize: true
models:
  - model: Test157t/Pasta-Lake-7b
    parameters:
      weight: 1
  - model: Test157t/Prima-LelantaclesV4-7b-16k
    parameters:
      weight: 1
dtype: float16
Downloads last month
55
GGUF
Model size
7.24B params
Architecture
llama

3-bit

4-bit

5-bit

6-bit

Inference Examples
Inference API (serverless) has been turned off for this model.

Model tree for Lewdiculous/Prima-LelantaclesV5-7b-GGUF

Quantized
(3)
this model

Collection including Lewdiculous/Prima-LelantaclesV5-7b-GGUF