Context-awareness in instruction finetuning
Collection
15 items
•
Updated
This model is a fine-tuned version of meta-llama/Llama-2-7b-hf on the yihanwang617/WizardLM_70k_processed_indicator_unfiltered_4k dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
0.693 | 0.2225 | 200 | 0.6798 |
0.6836 | 0.4450 | 400 | 0.6580 |
0.683 | 0.6675 | 600 | 0.6481 |
0.6654 | 0.8900 | 800 | 0.6413 |
0.6446 | 1.1125 | 1000 | 0.6380 |
0.629 | 1.3350 | 1200 | 0.6334 |
0.6241 | 1.5575 | 1400 | 0.6312 |
0.6087 | 1.7800 | 1600 | 0.6299 |
Base model
meta-llama/Llama-2-7b-hf