|
--- |
|
base_model: |
|
- meta-llama/Llama-3.2-1B |
|
--- |
|
|
|
## Model Description |
|
|
|
This is the meta-llama/Llama-3.2-1B base model fine tuned on the mlabonne/orpo-dpo-mix-40k dataset. |
|
|
|
## Evaluation Results |
|
|
|
We used lm-evalutation-harness from EleutherAI to evaluate this fine-tuned version of meta-llama/Llama-3.2-1B on the 'Hellaswag' benchmark. |
|
|
|
### Hellaswag |
|
|
|
| Tasks |Version|Filter|n-shot| Metric | |Value | |Stderr| |
|
|---------|------:|------|-----:|--------|---|-----:|---|-----:| |
|
|hellaswag| 1|none | 0|acc |↑ |0.4773|± |0.0050| |
|
| | |none | 0|acc_norm|↑ |0.6358|± |0.0048| |