collapse_gemma-2-27b_hs2_accumulate_iter3_sftsd0
This model is a fine-tuned version of google/gemma-2-27b on an unknown dataset. It achieves the following results on the evaluation set:
- Loss: 0.9289
- Num Input Tokens Seen: 13382208
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 8e-06
- train_batch_size: 4
- eval_batch_size: 16
- seed: 0
- gradient_accumulation_steps: 32
- total_train_batch_size: 128
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: constant_with_warmup
- lr_scheduler_warmup_ratio: 0.05
- num_epochs: 1
Training results
Training Loss | Epoch | Step | Validation Loss | Input Tokens Seen |
---|---|---|---|---|
No log | 0 | 0 | 1.1282 | 0 |
2.5273 | 0.0186 | 5 | 1.0497 | 239432 |
2.2642 | 0.0371 | 10 | 0.9938 | 490472 |
2.1944 | 0.0557 | 15 | 0.9799 | 738476 |
2.0449 | 0.0742 | 20 | 0.9761 | 991768 |
1.7622 | 0.0928 | 25 | 0.9788 | 1234816 |
1.6823 | 0.1113 | 30 | 0.9860 | 1486428 |
1.5237 | 0.1299 | 35 | 0.9862 | 1735404 |
1.4638 | 0.1484 | 40 | 0.9833 | 1983880 |
1.2775 | 0.1670 | 45 | 0.9803 | 2226820 |
1.246 | 0.1855 | 50 | 0.9762 | 2471660 |
1.1798 | 0.2041 | 55 | 0.9701 | 2723564 |
1.1618 | 0.2226 | 60 | 0.9658 | 2969216 |
1.1255 | 0.2412 | 65 | 0.9656 | 3218648 |
0.902 | 0.2597 | 70 | 0.9609 | 3474940 |
0.873 | 0.2783 | 75 | 0.9577 | 3721068 |
0.7585 | 0.2968 | 80 | 0.9560 | 3977036 |
0.9329 | 0.3154 | 85 | 0.9542 | 4227848 |
0.9888 | 0.3340 | 90 | 0.9544 | 4471040 |
0.8856 | 0.3525 | 95 | 0.9510 | 4719044 |
0.8959 | 0.3711 | 100 | 0.9519 | 4966088 |
0.707 | 0.3896 | 105 | 0.9476 | 5210868 |
0.8089 | 0.4082 | 110 | 0.9476 | 5470016 |
0.7476 | 0.4267 | 115 | 0.9459 | 5718420 |
0.6473 | 0.4453 | 120 | 0.9438 | 5972536 |
0.758 | 0.4638 | 125 | 0.9435 | 6221248 |
0.8454 | 0.4824 | 130 | 0.9403 | 6475340 |
0.7976 | 0.5009 | 135 | 0.9412 | 6727528 |
0.8476 | 0.5195 | 140 | 0.9400 | 6982388 |
0.7554 | 0.5380 | 145 | 0.9387 | 7218200 |
0.7193 | 0.5566 | 150 | 0.9386 | 7466484 |
0.6614 | 0.5751 | 155 | 0.9378 | 7709588 |
0.7586 | 0.5937 | 160 | 0.9344 | 7958964 |
0.769 | 0.6122 | 165 | 0.9353 | 8214680 |
0.6696 | 0.6308 | 170 | 0.9347 | 8457832 |
0.8566 | 0.6494 | 175 | 0.9377 | 8710088 |
0.8531 | 0.6679 | 180 | 0.9346 | 8959260 |
0.8454 | 0.6865 | 185 | 0.9346 | 9216248 |
0.7314 | 0.7050 | 190 | 0.9330 | 9465964 |
0.914 | 0.7236 | 195 | 0.9326 | 9718276 |
0.6292 | 0.7421 | 200 | 0.9335 | 9963556 |
0.683 | 0.7607 | 205 | 0.9348 | 10204596 |
0.5968 | 0.7792 | 210 | 0.9338 | 10460212 |
0.7731 | 0.7978 | 215 | 0.9338 | 10712008 |
0.707 | 0.8163 | 220 | 0.9318 | 10955092 |
0.7059 | 0.8349 | 225 | 0.9348 | 11197300 |
0.6878 | 0.8534 | 230 | 0.9301 | 11440440 |
0.6978 | 0.8720 | 235 | 0.9312 | 11685992 |
0.8379 | 0.8905 | 240 | 0.9294 | 11928976 |
0.8208 | 0.9091 | 245 | 0.9331 | 12185160 |
0.7653 | 0.9276 | 250 | 0.9314 | 12430192 |
0.7021 | 0.9462 | 255 | 0.9295 | 12684252 |
0.78 | 0.9647 | 260 | 0.9327 | 12932032 |
0.6731 | 0.9833 | 265 | 0.9279 | 13180768 |
Framework versions
- Transformers 4.44.0
- Pytorch 2.4.0+cu121
- Datasets 2.20.0
- Tokenizers 0.19.1
- Downloads last month
- 11
Model tree for RylanSchaeffer/collapse_gemma-2-27b_hs2_accumulate_iter3_sftsd0
Base model
google/gemma-2-27b