leixa commited on
Commit
694d64d
·
verified ·
1 Parent(s): 3ce3ff9

Training in progress, step 336, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7cc948873caa08921327726f20f406c613f84022f629a8b1be0e9b9d5c8d1f31
3
  size 93608
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6c7a70d48890b36af3199b1750d7e5fd17b21aa11b89db9652dda5db64438a8
3
  size 93608
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:77ad83b8581d98d467b489581929046033ba13e2931665a83fafd2a14bfad6bf
3
  size 197158
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:369e20b3070de99f6b2dbbf54cc6cbaa52f220bb8233b806b9dfba33eafacff4
3
  size 197158
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:37abdf82049403a03d10ecacdbbeb17537569cc71638264b2a476090c2ddcbf8
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f469bffe8acf737c47c0ff1bea10bc59a435d371a4846175c68c32944596f40
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:318b114b83c26bcba11815378a88e0015bce044b0002c702e0a2627e1a1d1e56
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10affc9ced28dcfaf0d40e3497a97c8e7416bd057324538f99a7e1756fd84408
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.8316831683168316,
5
  "eval_steps": 42,
6
- "global_step": 294,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -757,6 +757,112 @@
757
  "eval_samples_per_second": 90.248,
758
  "eval_steps_per_second": 11.357,
759
  "step": 294
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
760
  }
761
  ],
762
  "logging_steps": 3,
@@ -776,7 +882,7 @@
776
  "attributes": {}
777
  }
778
  },
779
- "total_flos": 1802989731840.0,
780
  "train_batch_size": 8,
781
  "trial_name": null,
782
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.9504950495049505,
5
  "eval_steps": 42,
6
+ "global_step": 336,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
757
  "eval_samples_per_second": 90.248,
758
  "eval_steps_per_second": 11.357,
759
  "step": 294
760
+ },
761
+ {
762
+ "epoch": 0.8401697312588402,
763
+ "grad_norm": 0.11685353517532349,
764
+ "learning_rate": 3.6698157721666246e-05,
765
+ "loss": 11.8211,
766
+ "step": 297
767
+ },
768
+ {
769
+ "epoch": 0.8486562942008486,
770
+ "grad_norm": 0.10467521101236343,
771
+ "learning_rate": 3.5773620668448384e-05,
772
+ "loss": 11.8131,
773
+ "step": 300
774
+ },
775
+ {
776
+ "epoch": 0.8571428571428571,
777
+ "grad_norm": 0.08696059882640839,
778
+ "learning_rate": 3.48543465876014e-05,
779
+ "loss": 11.8151,
780
+ "step": 303
781
+ },
782
+ {
783
+ "epoch": 0.8656294200848657,
784
+ "grad_norm": 0.21422286331653595,
785
+ "learning_rate": 3.3940675559617724e-05,
786
+ "loss": 11.817,
787
+ "step": 306
788
+ },
789
+ {
790
+ "epoch": 0.8741159830268741,
791
+ "grad_norm": 0.09098684042692184,
792
+ "learning_rate": 3.303294559217063e-05,
793
+ "loss": 11.8177,
794
+ "step": 309
795
+ },
796
+ {
797
+ "epoch": 0.8826025459688827,
798
+ "grad_norm": 0.11003749072551727,
799
+ "learning_rate": 3.213149249506997e-05,
800
+ "loss": 11.8127,
801
+ "step": 312
802
+ },
803
+ {
804
+ "epoch": 0.8910891089108911,
805
+ "grad_norm": 0.10718018561601639,
806
+ "learning_rate": 3.12366497560313e-05,
807
+ "loss": 11.8152,
808
+ "step": 315
809
+ },
810
+ {
811
+ "epoch": 0.8995756718528995,
812
+ "grad_norm": 0.058545950800180435,
813
+ "learning_rate": 3.0348748417303823e-05,
814
+ "loss": 11.8117,
815
+ "step": 318
816
+ },
817
+ {
818
+ "epoch": 0.9080622347949081,
819
+ "grad_norm": 0.15917275846004486,
820
+ "learning_rate": 2.9468116953203107e-05,
821
+ "loss": 11.816,
822
+ "step": 321
823
+ },
824
+ {
825
+ "epoch": 0.9165487977369166,
826
+ "grad_norm": 0.08903782814741135,
827
+ "learning_rate": 2.8595081148593738e-05,
828
+ "loss": 11.8119,
829
+ "step": 324
830
+ },
831
+ {
832
+ "epoch": 0.925035360678925,
833
+ "grad_norm": 0.10488823056221008,
834
+ "learning_rate": 2.772996397836704e-05,
835
+ "loss": 11.8185,
836
+ "step": 327
837
+ },
838
+ {
839
+ "epoch": 0.9335219236209336,
840
+ "grad_norm": 0.09897799789905548,
841
+ "learning_rate": 2.687308548795825e-05,
842
+ "loss": 11.8194,
843
+ "step": 330
844
+ },
845
+ {
846
+ "epoch": 0.942008486562942,
847
+ "grad_norm": 0.10232014954090118,
848
+ "learning_rate": 2.6024762674947313e-05,
849
+ "loss": 11.8139,
850
+ "step": 333
851
+ },
852
+ {
853
+ "epoch": 0.9504950495049505,
854
+ "grad_norm": 0.07098989188671112,
855
+ "learning_rate": 2.5185309371787513e-05,
856
+ "loss": 11.8159,
857
+ "step": 336
858
+ },
859
+ {
860
+ "epoch": 0.9504950495049505,
861
+ "eval_loss": 11.817768096923828,
862
+ "eval_runtime": 6.6321,
863
+ "eval_samples_per_second": 89.866,
864
+ "eval_steps_per_second": 11.309,
865
+ "step": 336
866
  }
867
  ],
868
  "logging_steps": 3,
 
882
  "attributes": {}
883
  }
884
  },
885
+ "total_flos": 2060904038400.0,
886
  "train_batch_size": 8,
887
  "trial_name": null,
888
  "trial_params": null