diaenra commited on
Commit
68be929
1 Parent(s): e7c42cc

Training in progress, step 100, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:198d342ef19c8d689e7d5807eef3d0db68d68033424117e91cd08740f64cb48e
3
  size 159967880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e94000994e49a5b7ea7e66dd28d1a7859c2a40464068833c5c3660070b2f5cd0
3
  size 159967880
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:56acf5affbb1287f1122bd505d97a6ef8790e1fbf13e5129a98d6ed4d987c9e0
3
  size 81730196
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0c235426aff5e7816c3778a2a10f2600144918273ca7af95ecdf7e1f3637895
3
  size 81730196
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:997b37e8e4d4313266e5aafe8f4d418b086633a9c94b4c13db86ae839aa09b15
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9246553064bea11894d1cdc342a12358c62b3d7caceee2321170ee68aed86db4
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:beb167787cdd9594eed637547cf4a56f4a4d8ea359757a120c76bb5d91190782
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a60c7d771c1fd156acee762fba03c724cb41829a3f71df370ecd1d20b134982
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.0633445945945946,
5
  "eval_steps": 25,
6
- "global_step": 75,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -564,6 +564,189 @@
564
  "eval_samples_per_second": 3.931,
565
  "eval_steps_per_second": 1.969,
566
  "step": 75
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
567
  }
568
  ],
569
  "logging_steps": 1,
@@ -578,12 +761,12 @@
578
  "should_evaluate": false,
579
  "should_log": false,
580
  "should_save": true,
581
- "should_training_stop": false
582
  },
583
  "attributes": {}
584
  }
585
  },
586
- "total_flos": 9.9164117532672e+16,
587
  "train_batch_size": 2,
588
  "trial_name": null,
589
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.08445945945945946,
5
  "eval_steps": 25,
6
+ "global_step": 100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
564
  "eval_samples_per_second": 3.931,
565
  "eval_steps_per_second": 1.969,
566
  "step": 75
567
+ },
568
+ {
569
+ "epoch": 0.06418918918918919,
570
+ "grad_norm": 0.20781630277633667,
571
+ "learning_rate": 3.308693936411421e-05,
572
+ "loss": 0.0065,
573
+ "step": 76
574
+ },
575
+ {
576
+ "epoch": 0.06503378378378379,
577
+ "grad_norm": 0.7386759519577026,
578
+ "learning_rate": 3.053416295410026e-05,
579
+ "loss": 0.0431,
580
+ "step": 77
581
+ },
582
+ {
583
+ "epoch": 0.06587837837837837,
584
+ "grad_norm": 3.0190958976745605,
585
+ "learning_rate": 2.8066019966134904e-05,
586
+ "loss": 0.2471,
587
+ "step": 78
588
+ },
589
+ {
590
+ "epoch": 0.06672297297297297,
591
+ "grad_norm": 4.477311134338379,
592
+ "learning_rate": 2.5685517452260567e-05,
593
+ "loss": 0.8406,
594
+ "step": 79
595
+ },
596
+ {
597
+ "epoch": 0.06756756756756757,
598
+ "grad_norm": 0.9087244272232056,
599
+ "learning_rate": 2.339555568810221e-05,
600
+ "loss": 0.0187,
601
+ "step": 80
602
+ },
603
+ {
604
+ "epoch": 0.06841216216216216,
605
+ "grad_norm": 2.032402276992798,
606
+ "learning_rate": 2.119892463932781e-05,
607
+ "loss": 0.3885,
608
+ "step": 81
609
+ },
610
+ {
611
+ "epoch": 0.06925675675675676,
612
+ "grad_norm": 5.099620819091797,
613
+ "learning_rate": 1.9098300562505266e-05,
614
+ "loss": 0.1469,
615
+ "step": 82
616
+ },
617
+ {
618
+ "epoch": 0.07010135135135136,
619
+ "grad_norm": 1.4657068252563477,
620
+ "learning_rate": 1.7096242744495837e-05,
621
+ "loss": 0.0909,
622
+ "step": 83
623
+ },
624
+ {
625
+ "epoch": 0.07094594594594594,
626
+ "grad_norm": 0.7036009430885315,
627
+ "learning_rate": 1.5195190384357404e-05,
628
+ "loss": 0.0208,
629
+ "step": 84
630
+ },
631
+ {
632
+ "epoch": 0.07179054054054054,
633
+ "grad_norm": 7.542021751403809,
634
+ "learning_rate": 1.339745962155613e-05,
635
+ "loss": 1.0755,
636
+ "step": 85
637
+ },
638
+ {
639
+ "epoch": 0.07263513513513513,
640
+ "grad_norm": 0.7535510063171387,
641
+ "learning_rate": 1.1705240714107302e-05,
642
+ "loss": 0.0495,
643
+ "step": 86
644
+ },
645
+ {
646
+ "epoch": 0.07347972972972973,
647
+ "grad_norm": 2.493562936782837,
648
+ "learning_rate": 1.0120595370083318e-05,
649
+ "loss": 0.2186,
650
+ "step": 87
651
+ },
652
+ {
653
+ "epoch": 0.07432432432432433,
654
+ "grad_norm": 3.849121332168579,
655
+ "learning_rate": 8.645454235739903e-06,
656
+ "loss": 0.2773,
657
+ "step": 88
658
+ },
659
+ {
660
+ "epoch": 0.07516891891891891,
661
+ "grad_norm": 4.295147895812988,
662
+ "learning_rate": 7.281614543321269e-06,
663
+ "loss": 0.448,
664
+ "step": 89
665
+ },
666
+ {
667
+ "epoch": 0.07601351351351351,
668
+ "grad_norm": 3.61863112449646,
669
+ "learning_rate": 6.030737921409169e-06,
670
+ "loss": 0.329,
671
+ "step": 90
672
+ },
673
+ {
674
+ "epoch": 0.07685810810810811,
675
+ "grad_norm": 5.776922702789307,
676
+ "learning_rate": 4.8943483704846475e-06,
677
+ "loss": 0.4583,
678
+ "step": 91
679
+ },
680
+ {
681
+ "epoch": 0.0777027027027027,
682
+ "grad_norm": 2.3311469554901123,
683
+ "learning_rate": 3.873830406168111e-06,
684
+ "loss": 0.2086,
685
+ "step": 92
686
+ },
687
+ {
688
+ "epoch": 0.0785472972972973,
689
+ "grad_norm": 0.32096993923187256,
690
+ "learning_rate": 2.970427372400353e-06,
691
+ "loss": 0.0184,
692
+ "step": 93
693
+ },
694
+ {
695
+ "epoch": 0.07939189189189189,
696
+ "grad_norm": 0.9880800843238831,
697
+ "learning_rate": 2.1852399266194314e-06,
698
+ "loss": 0.0559,
699
+ "step": 94
700
+ },
701
+ {
702
+ "epoch": 0.08023648648648649,
703
+ "grad_norm": 5.171363830566406,
704
+ "learning_rate": 1.5192246987791981e-06,
705
+ "loss": 0.4847,
706
+ "step": 95
707
+ },
708
+ {
709
+ "epoch": 0.08108108108108109,
710
+ "grad_norm": 3.456587314605713,
711
+ "learning_rate": 9.731931258429638e-07,
712
+ "loss": 0.2026,
713
+ "step": 96
714
+ },
715
+ {
716
+ "epoch": 0.08192567567567567,
717
+ "grad_norm": 3.7345874309539795,
718
+ "learning_rate": 5.478104631726711e-07,
719
+ "loss": 0.5642,
720
+ "step": 97
721
+ },
722
+ {
723
+ "epoch": 0.08277027027027027,
724
+ "grad_norm": 4.7123541831970215,
725
+ "learning_rate": 2.4359497401758024e-07,
726
+ "loss": 0.4377,
727
+ "step": 98
728
+ },
729
+ {
730
+ "epoch": 0.08361486486486487,
731
+ "grad_norm": 2.566887617111206,
732
+ "learning_rate": 6.09172980904238e-08,
733
+ "loss": 0.2207,
734
+ "step": 99
735
+ },
736
+ {
737
+ "epoch": 0.08445945945945946,
738
+ "grad_norm": 1.7636704444885254,
739
+ "learning_rate": 0.0,
740
+ "loss": 0.0955,
741
+ "step": 100
742
+ },
743
+ {
744
+ "epoch": 0.08445945945945946,
745
+ "eval_loss": 0.14757972955703735,
746
+ "eval_runtime": 126.9347,
747
+ "eval_samples_per_second": 3.931,
748
+ "eval_steps_per_second": 1.97,
749
+ "step": 100
750
  }
751
  ],
752
  "logging_steps": 1,
 
761
  "should_evaluate": false,
762
  "should_log": false,
763
  "should_save": true,
764
+ "should_training_stop": true
765
  },
766
  "attributes": {}
767
  }
768
  },
769
+ "total_flos": 1.32218823376896e+17,
770
  "train_batch_size": 2,
771
  "trial_name": null,
772
  "trial_params": null