Nexspear commited on
Commit
3b6020f
·
verified ·
1 Parent(s): c24c7f4

Training in progress, step 336, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:930e0e949e58701597d874e209b7344a4d706b043639fc115daabd8ebbeab0fe
3
  size 100966336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90941e1693a558421377af5791061bc37e7dd6f4ed3b75ca4bda99a4e94b2488
3
  size 100966336
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:709f78d5e19655cbd94a1c21408c442f8f92f8b4fdef1ad9958db57bcf19cf98
3
  size 51613668
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18e92072ad5cb96ea116208da591f4f9a7a38b1a697688a159b5227c548ac55d
3
  size 51613668
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5e926e6553b71955bfbf13bceb3502fb7e3ab2b01b876537fd15965365d03680
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1628639e06cf0da96dea52ad1bb1fc7d510f41ec927e99fe48623eb85119f343
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0316ab07671fffcd24f1d6dedeb5b41d6e5808ab64506af6cae3d569237843af
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73411ca8ec7e5959b18da4fe8d2de8084ec9d3eedda030beb9eadece4c950464
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.044043294258642,
5
  "eval_steps": 42,
6
- "global_step": 294,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -757,6 +757,112 @@
757
  "eval_samples_per_second": 35.967,
758
  "eval_steps_per_second": 4.498,
759
  "step": 294
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
760
  }
761
  ],
762
  "logging_steps": 3,
@@ -776,7 +882,7 @@
776
  "attributes": {}
777
  }
778
  },
779
- "total_flos": 9.381143263366349e+16,
780
  "train_batch_size": 8,
781
  "trial_name": null,
782
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.050335193438448,
5
  "eval_steps": 42,
6
+ "global_step": 336,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
757
  "eval_samples_per_second": 35.967,
758
  "eval_steps_per_second": 4.498,
759
  "step": 294
760
+ },
761
+ {
762
+ "epoch": 0.044492715628628145,
763
+ "grad_norm": 0.18710507452487946,
764
+ "learning_rate": 1.8349078860833123e-05,
765
+ "loss": 1.0978,
766
+ "step": 297
767
+ },
768
+ {
769
+ "epoch": 0.04494213699861428,
770
+ "grad_norm": 0.1569843292236328,
771
+ "learning_rate": 1.7886810334224192e-05,
772
+ "loss": 1.0783,
773
+ "step": 300
774
+ },
775
+ {
776
+ "epoch": 0.045391558368600426,
777
+ "grad_norm": 0.14208896458148956,
778
+ "learning_rate": 1.74271732938007e-05,
779
+ "loss": 1.0512,
780
+ "step": 303
781
+ },
782
+ {
783
+ "epoch": 0.04584097973858657,
784
+ "grad_norm": 0.16557104885578156,
785
+ "learning_rate": 1.6970337779808862e-05,
786
+ "loss": 1.143,
787
+ "step": 306
788
+ },
789
+ {
790
+ "epoch": 0.046290401108572715,
791
+ "grad_norm": 0.17679089307785034,
792
+ "learning_rate": 1.6516472796085315e-05,
793
+ "loss": 1.0428,
794
+ "step": 309
795
+ },
796
+ {
797
+ "epoch": 0.04673982247855885,
798
+ "grad_norm": 0.21449750661849976,
799
+ "learning_rate": 1.6065746247534984e-05,
800
+ "loss": 1.0541,
801
+ "step": 312
802
+ },
803
+ {
804
+ "epoch": 0.047189243848545,
805
+ "grad_norm": 0.19140197336673737,
806
+ "learning_rate": 1.561832487801565e-05,
807
+ "loss": 1.0943,
808
+ "step": 315
809
+ },
810
+ {
811
+ "epoch": 0.04763866521853114,
812
+ "grad_norm": 0.21146325767040253,
813
+ "learning_rate": 1.5174374208651912e-05,
814
+ "loss": 1.1286,
815
+ "step": 318
816
+ },
817
+ {
818
+ "epoch": 0.048088086588517286,
819
+ "grad_norm": 0.16576792299747467,
820
+ "learning_rate": 1.4734058476601553e-05,
821
+ "loss": 1.1543,
822
+ "step": 321
823
+ },
824
+ {
825
+ "epoch": 0.04853750795850343,
826
+ "grad_norm": 0.17730730772018433,
827
+ "learning_rate": 1.4297540574296869e-05,
828
+ "loss": 1.0809,
829
+ "step": 324
830
+ },
831
+ {
832
+ "epoch": 0.04898692932848957,
833
+ "grad_norm": 0.16303078830242157,
834
+ "learning_rate": 1.386498198918352e-05,
835
+ "loss": 1.0781,
836
+ "step": 327
837
+ },
838
+ {
839
+ "epoch": 0.04943635069847571,
840
+ "grad_norm": 0.15803277492523193,
841
+ "learning_rate": 1.3436542743979125e-05,
842
+ "loss": 1.1138,
843
+ "step": 330
844
+ },
845
+ {
846
+ "epoch": 0.049885772068461856,
847
+ "grad_norm": 0.161447212100029,
848
+ "learning_rate": 1.3012381337473656e-05,
849
+ "loss": 1.0087,
850
+ "step": 333
851
+ },
852
+ {
853
+ "epoch": 0.050335193438448,
854
+ "grad_norm": 0.14916126430034637,
855
+ "learning_rate": 1.2592654685893757e-05,
856
+ "loss": 1.0916,
857
+ "step": 336
858
+ },
859
+ {
860
+ "epoch": 0.050335193438448,
861
+ "eval_loss": 1.0906805992126465,
862
+ "eval_runtime": 312.9768,
863
+ "eval_samples_per_second": 35.923,
864
+ "eval_steps_per_second": 4.492,
865
+ "step": 336
866
  }
867
  ],
868
  "logging_steps": 3,
 
882
  "attributes": {}
883
  }
884
  },
885
+ "total_flos": 1.0751071457850163e+17,
886
  "train_batch_size": 8,
887
  "trial_name": null,
888
  "trial_params": null