gemma2b-summarize-claude3sonnet
Collection
9 items
•
Updated
This model is a fine-tuned version of google/gemma-2b on the llama-duo/synth_summarize_dataset_dedup dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
1.0192 | 1.0 | 402 | 2.4514 |
0.9424 | 2.0 | 804 | 2.4604 |
0.8955 | 3.0 | 1206 | 2.5064 |
0.8659 | 4.0 | 1608 | 2.5306 |
0.8359 | 5.0 | 2010 | 2.5706 |
0.7986 | 6.0 | 2412 | 2.6196 |
0.7778 | 7.0 | 2814 | 2.6583 |
0.7562 | 8.0 | 3216 | 2.6846 |
0.7563 | 9.0 | 3618 | 2.6927 |
0.7461 | 10.0 | 4020 | 2.6928 |
Base model
google/gemma-2b