lesso06 commited on
Commit
fafce8a
1 Parent(s): 800205f

Training in progress, step 100, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:94d0cff8548dc5af4b8cb758aab0529c37901481cf2e17faf1cc41fb1a3fe719
3
  size 767856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73c287d1cd15c20c08bd3e715cc99d240da6ddbf663b952f1b105d32a6da7fe4
3
  size 767856
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2c1ee2215d0456fa6debc3e9eda3e896e5e7527e920a574eb85cd4655f4ec72f
3
  size 1601338
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:862bff549a9bee8077d171c8ec84dc86fc6009d31084960526bd3377714ebd63
3
  size 1601338
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6a488ec6d3f5e5a798465f5dc551e4b485f1f571042f5124861158cd9d793a28
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37bc9f091951863adc5f18668a30a9e6aac3f21f8c892b02b33c031686cfb847
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f23e2214bcafb439ebc7528dcc283ef6218d509a276c0baff0743503ecbe3d92
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49d60a69e2379be2053e816cbaff31e6c931b5922dd86c71c9eaf473299cbf62
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.0028740587457607636,
5
  "eval_steps": 9,
6
- "global_step": 75,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -604,6 +604,205 @@
604
  "learning_rate": 1.7860619515673033e-05,
605
  "loss": 8.4948,
606
  "step": 75
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
607
  }
608
  ],
609
  "logging_steps": 1,
@@ -618,12 +817,12 @@
618
  "should_evaluate": false,
619
  "should_log": false,
620
  "should_save": true,
621
- "should_training_stop": false
622
  },
623
  "attributes": {}
624
  }
625
  },
626
- "total_flos": 20362140057600.0,
627
  "train_batch_size": 8,
628
  "trial_name": null,
629
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.003832078327681018,
5
  "eval_steps": 9,
6
+ "global_step": 100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
604
  "learning_rate": 1.7860619515673033e-05,
605
  "loss": 8.4948,
606
  "step": 75
607
+ },
608
+ {
609
+ "epoch": 0.0029123795290375734,
610
+ "grad_norm": 1.3906676769256592,
611
+ "learning_rate": 1.6543469682057106e-05,
612
+ "loss": 8.5319,
613
+ "step": 76
614
+ },
615
+ {
616
+ "epoch": 0.0029507003123143837,
617
+ "grad_norm": 1.549890160560608,
618
+ "learning_rate": 1.526708147705013e-05,
619
+ "loss": 8.3946,
620
+ "step": 77
621
+ },
622
+ {
623
+ "epoch": 0.002989021095591194,
624
+ "grad_norm": 1.4431867599487305,
625
+ "learning_rate": 1.4033009983067452e-05,
626
+ "loss": 8.3487,
627
+ "step": 78
628
+ },
629
+ {
630
+ "epoch": 0.003027341878868004,
631
+ "grad_norm": 1.5942152738571167,
632
+ "learning_rate": 1.2842758726130283e-05,
633
+ "loss": 8.3657,
634
+ "step": 79
635
+ },
636
+ {
637
+ "epoch": 0.003065662662144814,
638
+ "grad_norm": 1.3526906967163086,
639
+ "learning_rate": 1.1697777844051105e-05,
640
+ "loss": 8.5344,
641
+ "step": 80
642
+ },
643
+ {
644
+ "epoch": 0.0031039834454216243,
645
+ "grad_norm": 1.568245530128479,
646
+ "learning_rate": 1.0599462319663905e-05,
647
+ "loss": 8.4987,
648
+ "step": 81
649
+ },
650
+ {
651
+ "epoch": 0.0031039834454216243,
652
+ "eval_loss": 8.4481782913208,
653
+ "eval_runtime": 197.166,
654
+ "eval_samples_per_second": 111.459,
655
+ "eval_steps_per_second": 13.932,
656
+ "step": 81
657
+ },
658
+ {
659
+ "epoch": 0.0031423042286984346,
660
+ "grad_norm": 1.4625965356826782,
661
+ "learning_rate": 9.549150281252633e-06,
662
+ "loss": 8.3553,
663
+ "step": 82
664
+ },
665
+ {
666
+ "epoch": 0.003180625011975245,
667
+ "grad_norm": 1.6190887689590454,
668
+ "learning_rate": 8.548121372247918e-06,
669
+ "loss": 8.4586,
670
+ "step": 83
671
+ },
672
+ {
673
+ "epoch": 0.003218945795252055,
674
+ "grad_norm": 1.4923970699310303,
675
+ "learning_rate": 7.597595192178702e-06,
676
+ "loss": 8.3606,
677
+ "step": 84
678
+ },
679
+ {
680
+ "epoch": 0.003257266578528865,
681
+ "grad_norm": 1.470651626586914,
682
+ "learning_rate": 6.698729810778065e-06,
683
+ "loss": 8.3707,
684
+ "step": 85
685
+ },
686
+ {
687
+ "epoch": 0.0032955873618056753,
688
+ "grad_norm": 1.3880239725112915,
689
+ "learning_rate": 5.852620357053651e-06,
690
+ "loss": 8.5658,
691
+ "step": 86
692
+ },
693
+ {
694
+ "epoch": 0.0033339081450824855,
695
+ "grad_norm": 1.5443851947784424,
696
+ "learning_rate": 5.060297685041659e-06,
697
+ "loss": 8.381,
698
+ "step": 87
699
+ },
700
+ {
701
+ "epoch": 0.003372228928359296,
702
+ "grad_norm": 1.3600224256515503,
703
+ "learning_rate": 4.322727117869951e-06,
704
+ "loss": 8.4857,
705
+ "step": 88
706
+ },
707
+ {
708
+ "epoch": 0.0034105497116361056,
709
+ "grad_norm": 1.3307414054870605,
710
+ "learning_rate": 3.6408072716606346e-06,
711
+ "loss": 8.3026,
712
+ "step": 89
713
+ },
714
+ {
715
+ "epoch": 0.003448870494912916,
716
+ "grad_norm": 1.3035300970077515,
717
+ "learning_rate": 3.0153689607045845e-06,
718
+ "loss": 8.3625,
719
+ "step": 90
720
+ },
721
+ {
722
+ "epoch": 0.003448870494912916,
723
+ "eval_loss": 8.436622619628906,
724
+ "eval_runtime": 197.0498,
725
+ "eval_samples_per_second": 111.525,
726
+ "eval_steps_per_second": 13.941,
727
+ "step": 90
728
+ },
729
+ {
730
+ "epoch": 0.003487191278189726,
731
+ "grad_norm": 1.4547570943832397,
732
+ "learning_rate": 2.4471741852423237e-06,
733
+ "loss": 8.4409,
734
+ "step": 91
735
+ },
736
+ {
737
+ "epoch": 0.0035255120614665365,
738
+ "grad_norm": 1.3213475942611694,
739
+ "learning_rate": 1.9369152030840556e-06,
740
+ "loss": 8.4133,
741
+ "step": 92
742
+ },
743
+ {
744
+ "epoch": 0.0035638328447433467,
745
+ "grad_norm": 1.3285075426101685,
746
+ "learning_rate": 1.4852136862001764e-06,
747
+ "loss": 8.4509,
748
+ "step": 93
749
+ },
750
+ {
751
+ "epoch": 0.0036021536280201566,
752
+ "grad_norm": 1.4316322803497314,
753
+ "learning_rate": 1.0926199633097157e-06,
754
+ "loss": 8.4731,
755
+ "step": 94
756
+ },
757
+ {
758
+ "epoch": 0.003640474411296967,
759
+ "grad_norm": 1.4853651523590088,
760
+ "learning_rate": 7.596123493895991e-07,
761
+ "loss": 8.4,
762
+ "step": 95
763
+ },
764
+ {
765
+ "epoch": 0.003678795194573777,
766
+ "grad_norm": 1.407329797744751,
767
+ "learning_rate": 4.865965629214819e-07,
768
+ "loss": 8.4689,
769
+ "step": 96
770
+ },
771
+ {
772
+ "epoch": 0.0037171159778505874,
773
+ "grad_norm": 1.4351682662963867,
774
+ "learning_rate": 2.7390523158633554e-07,
775
+ "loss": 8.3587,
776
+ "step": 97
777
+ },
778
+ {
779
+ "epoch": 0.0037554367611273972,
780
+ "grad_norm": 1.4511561393737793,
781
+ "learning_rate": 1.2179748700879012e-07,
782
+ "loss": 8.4413,
783
+ "step": 98
784
+ },
785
+ {
786
+ "epoch": 0.0037937575444042075,
787
+ "grad_norm": 1.4957984685897827,
788
+ "learning_rate": 3.04586490452119e-08,
789
+ "loss": 8.3962,
790
+ "step": 99
791
+ },
792
+ {
793
+ "epoch": 0.0037937575444042075,
794
+ "eval_loss": 8.434378623962402,
795
+ "eval_runtime": 197.6671,
796
+ "eval_samples_per_second": 111.177,
797
+ "eval_steps_per_second": 13.897,
798
+ "step": 99
799
+ },
800
+ {
801
+ "epoch": 0.003832078327681018,
802
+ "grad_norm": 1.5082837343215942,
803
+ "learning_rate": 0.0,
804
+ "loss": 8.2469,
805
+ "step": 100
806
  }
807
  ],
808
  "logging_steps": 1,
 
817
  "should_evaluate": false,
818
  "should_log": false,
819
  "should_save": true,
820
+ "should_training_stop": true
821
  },
822
  "attributes": {}
823
  }
824
  },
825
+ "total_flos": 27149520076800.0,
826
  "train_batch_size": 8,
827
  "trial_name": null,
828
  "trial_params": null