Edit model card

lang-uk/dragoman-4bit

This model was converted to MLX format from the lang-uk/dragoman adapter fused into the mistralai/Mistral-7b-v0.1 base model and quantized into 4 bits using mlx-lm version 0.4.0. Refer to the original model card for more details on the model.

Use with mlx

pip install mlx-lm
from mlx_lm import load, generate

model, tokenizer = load("lang-uk/dragoman-4bit")
response = generate(model, tokenizer, prompt="[INST] who holds this neighborhood? [/INST]", verbose=True)

Or use from your shell:

python -m mlx_lm.generate --model lang-uk/dragoman-4bit --prompt '[INST] who holds this neighborhood? [/INST]' --temp 0 --max-tokens 100
Downloads last month
0
Safetensors
Model size
1.24B params
Tensor type
FP16
·
U32
·
Inference Examples
Inference API (serverless) has been turned off for this model.

Model tree for lang-uk/dragoman-4bit

Adapter
(1172)
this model

Datasets used to train lang-uk/dragoman-4bit

Collection including lang-uk/dragoman-4bit

Evaluation results