Qwarkstar 4B Instruct (Preview)

Training complete!

This model is fine-tuned using Supervised Fine-Tuning (SFT) on 100k samples from the HuggingFaceTB/smoltalk dataset.
It follows the ChatML input-output formatting template.

Training Details:

  • Base Model: qingy2024/Qwarkstar-4B
  • Batch Size: 32 (2 H100s x 8 per GPU)
  • Max Gradient Norm: 1.0
  • Final Loss: ~0.59
Downloads last month
44
Safetensors
Model size
4.47B params
Tensor type
BF16
·
Inference Examples
Unable to determine this model's library. Check the docs .

Model tree for qingy2024/Qwarkstar-4B-Instruct-Preview

Base model

Qwen/Qwen2.5-3B
Finetuned
qingy2024/Qwark-4B
Finetuned
(2)
this model
Quantizations
1 model

Dataset used to train qingy2024/Qwarkstar-4B-Instruct-Preview