Seosnaps commited on
Commit
62017e0
1 Parent(s): 0667581

Training in progress, step 2355, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:511578a0d482675e3ff17b4d731393ce3c283abd8d16f8a85a3f15d02a56e39f
3
  size 966995080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f1338dfd6f18752b4013e43c300716a64bf71b503d43f1c8bbbfc90371d4bd2
3
  size 966995080
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d6ed9b6da13b19371de47f3a2b44715717ead8b5eeed3a91c863518d2a661a22
3
  size 1925070764
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eee60cb97d325b98f6bb5681ff52f9706d140347a9459d3b7c365c4b570b64f5
3
  size 1925070764
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:811d52f8a469bde4a4138f47efe6e2b676d318980b3df9610e6fcc5abea00325
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9526eb78c9d5120112457d446047c713782319439e09b5cb1a2c5b9b035bb069
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f3ea9d13baff2282d300ceb3c3984a3388d1450303ffc8640c73967fa3325903
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:617bb927196c8fc850b17a9f4bc1faa1c723902215e41b7b836f9d3bc37475b7
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 78.86568308105001,
3
  "best_model_checkpoint": "./whisper-small-ha-adam-v4/checkpoint-2000",
4
- "epoch": 12.738853503184714,
5
  "eval_steps": 500,
6
- "global_step": 2000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -607,6 +607,104 @@
607
  "eval_wer": 78.86568308105001,
608
  "eval_wer_ortho": 81.0546875,
609
  "step": 2000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
610
  }
611
  ],
612
  "logging_steps": 25,
@@ -621,12 +719,12 @@
621
  "should_evaluate": false,
622
  "should_log": false,
623
  "should_save": true,
624
- "should_training_stop": false
625
  },
626
  "attributes": {}
627
  }
628
  },
629
- "total_flos": 9.22088071102464e+18,
630
  "train_batch_size": 16,
631
  "trial_name": null,
632
  "trial_params": null
 
1
  {
2
  "best_metric": 78.86568308105001,
3
  "best_model_checkpoint": "./whisper-small-ha-adam-v4/checkpoint-2000",
4
+ "epoch": 15.0,
5
  "eval_steps": 500,
6
+ "global_step": 2355,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
607
  "eval_wer": 78.86568308105001,
608
  "eval_wer_ortho": 81.0546875,
609
  "step": 2000
610
+ },
611
+ {
612
+ "epoch": 12.898089171974522,
613
+ "grad_norm": 2.6695284843444824,
614
+ "learning_rate": 5e-05,
615
+ "loss": 0.0364,
616
+ "step": 2025
617
+ },
618
+ {
619
+ "epoch": 13.05732484076433,
620
+ "grad_norm": 1.3920303583145142,
621
+ "learning_rate": 5e-05,
622
+ "loss": 0.0314,
623
+ "step": 2050
624
+ },
625
+ {
626
+ "epoch": 13.21656050955414,
627
+ "grad_norm": 2.2318477630615234,
628
+ "learning_rate": 5e-05,
629
+ "loss": 0.0226,
630
+ "step": 2075
631
+ },
632
+ {
633
+ "epoch": 13.375796178343949,
634
+ "grad_norm": 2.457688093185425,
635
+ "learning_rate": 5e-05,
636
+ "loss": 0.0374,
637
+ "step": 2100
638
+ },
639
+ {
640
+ "epoch": 13.535031847133759,
641
+ "grad_norm": 3.3562824726104736,
642
+ "learning_rate": 5e-05,
643
+ "loss": 0.0263,
644
+ "step": 2125
645
+ },
646
+ {
647
+ "epoch": 13.694267515923567,
648
+ "grad_norm": 2.763430118560791,
649
+ "learning_rate": 5e-05,
650
+ "loss": 0.0359,
651
+ "step": 2150
652
+ },
653
+ {
654
+ "epoch": 13.853503184713375,
655
+ "grad_norm": 5.378473281860352,
656
+ "learning_rate": 5e-05,
657
+ "loss": 0.039,
658
+ "step": 2175
659
+ },
660
+ {
661
+ "epoch": 14.012738853503185,
662
+ "grad_norm": 2.072021007537842,
663
+ "learning_rate": 5e-05,
664
+ "loss": 0.0373,
665
+ "step": 2200
666
+ },
667
+ {
668
+ "epoch": 14.171974522292993,
669
+ "grad_norm": 1.6606969833374023,
670
+ "learning_rate": 5e-05,
671
+ "loss": 0.0263,
672
+ "step": 2225
673
+ },
674
+ {
675
+ "epoch": 14.331210191082803,
676
+ "grad_norm": 3.090102195739746,
677
+ "learning_rate": 5e-05,
678
+ "loss": 0.0257,
679
+ "step": 2250
680
+ },
681
+ {
682
+ "epoch": 14.490445859872612,
683
+ "grad_norm": 6.045629024505615,
684
+ "learning_rate": 5e-05,
685
+ "loss": 0.0232,
686
+ "step": 2275
687
+ },
688
+ {
689
+ "epoch": 14.64968152866242,
690
+ "grad_norm": 0.8971702456474304,
691
+ "learning_rate": 5e-05,
692
+ "loss": 0.0263,
693
+ "step": 2300
694
+ },
695
+ {
696
+ "epoch": 14.80891719745223,
697
+ "grad_norm": 0.5613566637039185,
698
+ "learning_rate": 5e-05,
699
+ "loss": 0.0295,
700
+ "step": 2325
701
+ },
702
+ {
703
+ "epoch": 14.968152866242038,
704
+ "grad_norm": 3.578268051147461,
705
+ "learning_rate": 5e-05,
706
+ "loss": 0.0246,
707
+ "step": 2350
708
  }
709
  ],
710
  "logging_steps": 25,
 
719
  "should_evaluate": false,
720
  "should_log": false,
721
  "should_save": true,
722
+ "should_training_stop": true
723
  },
724
  "attributes": {}
725
  }
726
  },
727
+ "total_flos": 1.08565827600384e+19,
728
  "train_batch_size": 16,
729
  "trial_name": null,
730
  "trial_params": null