Model Card
Model Description
This is a Large Language Model (LLM) trained on a subset of the dataset "mlabonne/orpo-dpo-mix-40k".
Evaluation Results
Hellaswag
Metric | Value |
---|---|
Accuracy | 0.4517 |
How to Use
To use this model, simply download the checkpoint and load it into your preferred deep learning framework.
- Downloads last month
- 177
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the model is not deployed on the HF Inference API.
Model tree for d4niel92/llama-3.2-1B-orpo
Base model
meta-llama/Llama-3.2-1B