farmery commited on
Commit
cc5114b
1 Parent(s): 8af613b

Training in progress, step 99, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0ffb0b2511916105682cf80bf9696833163afb4bee41004ff0b8200c9f08b191
3
  size 237402
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc138d6658e8e095c787a05ea8c7acb9723acdbbf5b49bfb28685fe7044e575d
3
  size 237402
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:791ddabfb59f09149d5becaccac457c6ba26c528908751e157b0ccc437888f6c
3
  size 222294
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b32541ba1a16f052d7a3db977d9ece8dc9ce41f7f9b6ec83f2008a1b8e73123
3
  size 222294
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1318a700856bcdf76961c92e318a3b39bde15057dc94458041a6ee7452060e66
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7db5f040100ef639879df4e7ff2bd3ab5f6e11208b015560d33f3f4bb39c680d
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:020fbe0f82efc7984cb137980f674aa1a26f24fc1774e038e8433aa642c4aa86
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a03cc7a398e34f4833c81504da7e98a819dc8ef2649b1ba38633a959d049481f
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:346ae7ad1ffbb1ba7d9f71113f7ade9f0a382da141487c6347a82436c307f147
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a263fdde13fc7b695eb316b07ca5835d2c7b10e46642c1a58a0ca11d5a739f0b
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68333bc873842e58e67b8ea867c281a4c5484cc61d73e761bb48e26bc6301317
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b31521eec184753dbe4c30573b992c5bd063a8acd795f7c130b28632b36a9f18
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:355b85b545659c2cc1d9c5605cde8c2d283fa57cb41a54787534d12dd02cbb89
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2377833f26b6f1902c6fcca29fee5c63db83bdd037d63afc31b5aa4a26de7f6d
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 6.883584976196289,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-75",
4
- "epoch": 2.3000949667616335,
5
  "eval_steps": 25,
6
- "global_step": 75,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -564,6 +564,174 @@
564
  "eval_samples_per_second": 724.277,
565
  "eval_steps_per_second": 188.312,
566
  "step": 75
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
567
  }
568
  ],
569
  "logging_steps": 1,
@@ -587,12 +755,12 @@
587
  "should_evaluate": false,
588
  "should_log": false,
589
  "should_save": true,
590
- "should_training_stop": false
591
  },
592
  "attributes": {}
593
  }
594
  },
595
- "total_flos": 8558936064000.0,
596
  "train_batch_size": 1,
597
  "trial_name": null,
598
  "trial_params": null
 
1
  {
2
  "best_metric": 6.883584976196289,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-75",
4
+ "epoch": 3.03988603988604,
5
  "eval_steps": 25,
6
+ "global_step": 99,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
564
  "eval_samples_per_second": 724.277,
565
  "eval_steps_per_second": 188.312,
566
  "step": 75
567
+ },
568
+ {
569
+ "epoch": 2.3304843304843303,
570
+ "grad_norm": 0.09302153438329697,
571
+ "learning_rate": 2.2400865784401e-05,
572
+ "loss": 6.88,
573
+ "step": 76
574
+ },
575
+ {
576
+ "epoch": 2.3608736942070276,
577
+ "grad_norm": 0.09038078039884567,
578
+ "learning_rate": 2.1393033535713093e-05,
579
+ "loss": 6.9026,
580
+ "step": 77
581
+ },
582
+ {
583
+ "epoch": 2.3912630579297245,
584
+ "grad_norm": 0.09370548278093338,
585
+ "learning_rate": 2.0421950011441354e-05,
586
+ "loss": 6.9711,
587
+ "step": 78
588
+ },
589
+ {
590
+ "epoch": 2.421652421652422,
591
+ "grad_norm": 0.09348244965076447,
592
+ "learning_rate": 1.9488677077162295e-05,
593
+ "loss": 6.8911,
594
+ "step": 79
595
+ },
596
+ {
597
+ "epoch": 2.4520417853751186,
598
+ "grad_norm": 0.09923094511032104,
599
+ "learning_rate": 1.8594235253127375e-05,
600
+ "loss": 6.91,
601
+ "step": 80
602
+ },
603
+ {
604
+ "epoch": 2.482431149097816,
605
+ "grad_norm": 0.12247008085250854,
606
+ "learning_rate": 1.77396025983391e-05,
607
+ "loss": 7.1999,
608
+ "step": 81
609
+ },
610
+ {
611
+ "epoch": 2.5128205128205128,
612
+ "grad_norm": 0.08389267325401306,
613
+ "learning_rate": 1.6925713641057904e-05,
614
+ "loss": 6.4161,
615
+ "step": 82
616
+ },
617
+ {
618
+ "epoch": 2.5432098765432096,
619
+ "grad_norm": 0.08750791847705841,
620
+ "learning_rate": 1.6153458356909176e-05,
621
+ "loss": 6.8745,
622
+ "step": 83
623
+ },
624
+ {
625
+ "epoch": 2.573599240265907,
626
+ "grad_norm": 0.08594594895839691,
627
+ "learning_rate": 1.5423681195707997e-05,
628
+ "loss": 6.9089,
629
+ "step": 84
630
+ },
631
+ {
632
+ "epoch": 2.603988603988604,
633
+ "grad_norm": 0.09055335074663162,
634
+ "learning_rate": 1.4737180158065644e-05,
635
+ "loss": 6.8928,
636
+ "step": 85
637
+ },
638
+ {
639
+ "epoch": 2.634377967711301,
640
+ "grad_norm": 0.09193196147680283,
641
+ "learning_rate": 1.4094705922787687e-05,
642
+ "loss": 6.8778,
643
+ "step": 86
644
+ },
645
+ {
646
+ "epoch": 2.664767331433998,
647
+ "grad_norm": 0.09884382039308548,
648
+ "learning_rate": 1.3496961026017687e-05,
649
+ "loss": 7.1493,
650
+ "step": 87
651
+ },
652
+ {
653
+ "epoch": 2.695156695156695,
654
+ "grad_norm": 0.09690383821725845,
655
+ "learning_rate": 1.2944599093024267e-05,
656
+ "loss": 6.6664,
657
+ "step": 88
658
+ },
659
+ {
660
+ "epoch": 2.725546058879392,
661
+ "grad_norm": 0.12492170184850693,
662
+ "learning_rate": 1.2438224123471442e-05,
663
+ "loss": 7.3624,
664
+ "step": 89
665
+ },
666
+ {
667
+ "epoch": 2.7559354226020893,
668
+ "grad_norm": 0.08144571632146835,
669
+ "learning_rate": 1.1978389830953907e-05,
670
+ "loss": 6.2938,
671
+ "step": 90
672
+ },
673
+ {
674
+ "epoch": 2.786324786324786,
675
+ "grad_norm": 0.08170946687459946,
676
+ "learning_rate": 1.1565599037519316e-05,
677
+ "loss": 6.8895,
678
+ "step": 91
679
+ },
680
+ {
681
+ "epoch": 2.8167141500474835,
682
+ "grad_norm": 0.08227747678756714,
683
+ "learning_rate": 1.1200303123839742e-05,
684
+ "loss": 6.9282,
685
+ "step": 92
686
+ },
687
+ {
688
+ "epoch": 2.8471035137701803,
689
+ "grad_norm": 0.08924560993909836,
690
+ "learning_rate": 1.088290153563358e-05,
691
+ "loss": 6.8725,
692
+ "step": 93
693
+ },
694
+ {
695
+ "epoch": 2.8774928774928776,
696
+ "grad_norm": 0.09153356403112411,
697
+ "learning_rate": 1.0613741346877497e-05,
698
+ "loss": 6.9255,
699
+ "step": 94
700
+ },
701
+ {
702
+ "epoch": 2.9078822412155745,
703
+ "grad_norm": 0.09689171612262726,
704
+ "learning_rate": 1.0393116880286118e-05,
705
+ "loss": 7.0523,
706
+ "step": 95
707
+ },
708
+ {
709
+ "epoch": 2.9382716049382713,
710
+ "grad_norm": 0.09343353658914566,
711
+ "learning_rate": 1.0221269385474488e-05,
712
+ "loss": 6.7672,
713
+ "step": 96
714
+ },
715
+ {
716
+ "epoch": 2.9686609686609686,
717
+ "grad_norm": 0.1134086549282074,
718
+ "learning_rate": 1.0098386775155147e-05,
719
+ "loss": 7.2979,
720
+ "step": 97
721
+ },
722
+ {
723
+ "epoch": 3.009496676163343,
724
+ "grad_norm": 0.11922775208950043,
725
+ "learning_rate": 1.0024603419658329e-05,
726
+ "loss": 9.4261,
727
+ "step": 98
728
+ },
729
+ {
730
+ "epoch": 3.03988603988604,
731
+ "grad_norm": 0.08165828138589859,
732
+ "learning_rate": 1e-05,
733
+ "loss": 6.7951,
734
+ "step": 99
735
  }
736
  ],
737
  "logging_steps": 1,
 
755
  "should_evaluate": false,
756
  "should_log": false,
757
  "should_save": true,
758
+ "should_training_stop": true
759
  },
760
  "attributes": {}
761
  }
762
  },
763
+ "total_flos": 11297795604480.0,
764
  "train_batch_size": 1,
765
  "trial_name": null,
766
  "trial_params": null