leixa commited on
Commit
2c1d1e3
·
verified ·
1 Parent(s): a6a245b

Training in progress, step 294, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4c4d1afd543174e3ab9a2fe34af04ea8af62f4a7862fc78925dd62a536e9feaf
3
  size 201892112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcc7c1bb49029978b38479c4be7f1ef61c6a2f919bd26fa335b25ed148c05645
3
  size 201892112
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c67f2182ec15d55237558a16b27d2d849447fdca6258247107fbf31c03b1c68d
3
- size 102864548
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc3812fdb507dc890ad3dfeb461b18c142361e92db1a8d19d6b3dbcdb50cb254
3
+ size 102864868
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa5a85cd5fb1a43d1a3608369f309632ab765e2fc493019f6bf624d685f85850
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1fd586e160e28252df6d0804518bcfa9955584c079b4b4506e38e98dfc07096c
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ee19ddad9c4c375a1de2d74fb4c1cf5e15d36c1ed47a2cb80f7cb0fbacb3b29e
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:318b114b83c26bcba11815378a88e0015bce044b0002c702e0a2627e1a1d1e56
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.05282465150403522,
5
  "eval_steps": 42,
6
- "global_step": 252,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -651,6 +651,112 @@
651
  "eval_samples_per_second": 48.7,
652
  "eval_steps_per_second": 6.091,
653
  "step": 252
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
654
  }
655
  ],
656
  "logging_steps": 3,
@@ -670,7 +776,7 @@
670
  "attributes": {}
671
  }
672
  },
673
- "total_flos": 5.375532588230246e+16,
674
  "train_batch_size": 8,
675
  "trial_name": null,
676
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.061628760088041086,
5
  "eval_steps": 42,
6
+ "global_step": 294,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
651
  "eval_samples_per_second": 48.7,
652
  "eval_steps_per_second": 6.091,
653
  "step": 252
654
+ },
655
+ {
656
+ "epoch": 0.05345351640289278,
657
+ "grad_norm": 0.36929386854171753,
658
+ "learning_rate": 5e-05,
659
+ "loss": 1.0279,
660
+ "step": 255
661
+ },
662
+ {
663
+ "epoch": 0.05408238130175034,
664
+ "grad_norm": 0.34936413168907166,
665
+ "learning_rate": 4.903834726061565e-05,
666
+ "loss": 1.0564,
667
+ "step": 258
668
+ },
669
+ {
670
+ "epoch": 0.0547112462006079,
671
+ "grad_norm": 0.33937400579452515,
672
+ "learning_rate": 4.807705027948008e-05,
673
+ "loss": 1.0349,
674
+ "step": 261
675
+ },
676
+ {
677
+ "epoch": 0.05534011109946547,
678
+ "grad_norm": 0.3877185583114624,
679
+ "learning_rate": 4.711646468323129e-05,
680
+ "loss": 1.0518,
681
+ "step": 264
682
+ },
683
+ {
684
+ "epoch": 0.055968975998323026,
685
+ "grad_norm": 0.36920610070228577,
686
+ "learning_rate": 4.6156945835334184e-05,
687
+ "loss": 1.0907,
688
+ "step": 267
689
+ },
690
+ {
691
+ "epoch": 0.05659784089718059,
692
+ "grad_norm": 0.3572535216808319,
693
+ "learning_rate": 4.5198848704615914e-05,
694
+ "loss": 1.0811,
695
+ "step": 270
696
+ },
697
+ {
698
+ "epoch": 0.05722670579603815,
699
+ "grad_norm": 0.3565980792045593,
700
+ "learning_rate": 4.424252773394704e-05,
701
+ "loss": 1.0598,
702
+ "step": 273
703
+ },
704
+ {
705
+ "epoch": 0.05785557069489571,
706
+ "grad_norm": 0.3740125298500061,
707
+ "learning_rate": 4.328833670911724e-05,
708
+ "loss": 1.0403,
709
+ "step": 276
710
+ },
711
+ {
712
+ "epoch": 0.05848443559375328,
713
+ "grad_norm": 0.382012277841568,
714
+ "learning_rate": 4.23366286279542e-05,
715
+ "loss": 1.0877,
716
+ "step": 279
717
+ },
718
+ {
719
+ "epoch": 0.059113300492610835,
720
+ "grad_norm": 0.359355092048645,
721
+ "learning_rate": 4.138775556973406e-05,
722
+ "loss": 1.0644,
723
+ "step": 282
724
+ },
725
+ {
726
+ "epoch": 0.0597421653914684,
727
+ "grad_norm": 0.35905787348747253,
728
+ "learning_rate": 4.04420685649314e-05,
729
+ "loss": 1.0783,
730
+ "step": 285
731
+ },
732
+ {
733
+ "epoch": 0.060371030290325964,
734
+ "grad_norm": 0.4150475263595581,
735
+ "learning_rate": 3.9499917465357534e-05,
736
+ "loss": 1.0996,
737
+ "step": 288
738
+ },
739
+ {
740
+ "epoch": 0.06099989518918352,
741
+ "grad_norm": 0.3718733489513397,
742
+ "learning_rate": 3.856165081473474e-05,
743
+ "loss": 1.0191,
744
+ "step": 291
745
+ },
746
+ {
747
+ "epoch": 0.061628760088041086,
748
+ "grad_norm": 0.36830800771713257,
749
+ "learning_rate": 3.762761571975429e-05,
750
+ "loss": 1.1456,
751
+ "step": 294
752
+ },
753
+ {
754
+ "epoch": 0.061628760088041086,
755
+ "eval_loss": 1.0759488344192505,
756
+ "eval_runtime": 164.9861,
757
+ "eval_samples_per_second": 48.701,
758
+ "eval_steps_per_second": 6.091,
759
+ "step": 294
760
  }
761
  ],
762
  "logging_steps": 3,
 
776
  "attributes": {}
777
  }
778
  },
779
+ "total_flos": 6.271454686268621e+16,
780
  "train_batch_size": 8,
781
  "trial_name": null,
782
  "trial_params": null