Model Card

The Pythia-160M model is designed for research on language model behavior and interpretability, trained on the Pile dataset. Here we've evaluated it on HELLASWAG and can be fine-tuned for further experimentation.

Hellaswag Eval

Evaluated on the Eleuther evaluation harness, revision 100,000 steps

Tasks Version Filter n-shot Metric Value Stderr
hellaswag 1 none 0 acc ↑ 0.2872 ± 0.0045
none 0 acc_norm ↑ 0.3082 ± 0.0046

How to Use

Done just an exercise - not intended for deployment or human-facing interactions.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for illeto/finetunning-week1

Finetuned
(82)
this model