Femboyuwu2000 commited on
Commit
44b3a10
1 Parent(s): 6837900

Training in progress, step 2740, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9d5da627a83cc018d4bf0bfc59c1098cf129633a1abc41bd7d51fb13114c0b1
3
  size 13982248
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d9e8b32b966f871c511b5657fa29716b505e3b46707edada57780def8a3a4f0
3
  size 13982248
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:682f21df5b619b8fc7a877dafd3f697aa66e4f2b393f8b846f0e3646f7ba1e83
3
  size 7062522
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09d1dd512e5173901c2f4b3e632283257f9e9c5e3f7e6ab6ae19723b4371d0b6
3
  size 7062522
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cfdb4aa20ae8b96f35fdee422b168c0d1ffe465638f08b59f076576138bbe18e
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91283a9e0978c1c4026d0596abe9afca642afabfe3852b23f40302b2b4a58b48
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d6164b3c850dce9a36a31f30919cac3506bfad9cadb5f9ef73ee2c76d96b38ff
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a07583ff04916149d2b98a04a5a3645539e81ed096b0a41ed14e70c156517bb
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.2128,
5
  "eval_steps": 500,
6
- "global_step": 2660,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -938,6 +938,34 @@
938
  "learning_rate": 2.9173085201008144e-05,
939
  "loss": 3.5365,
940
  "step": 2660
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
941
  }
942
  ],
943
  "logging_steps": 20,
@@ -945,7 +973,7 @@
945
  "num_input_tokens_seen": 0,
946
  "num_train_epochs": 2,
947
  "save_steps": 20,
948
- "total_flos": 6299096079630336.0,
949
  "train_batch_size": 8,
950
  "trial_name": null,
951
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.2192,
5
  "eval_steps": 500,
6
+ "global_step": 2740,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
938
  "learning_rate": 2.9173085201008144e-05,
939
  "loss": 3.5365,
940
  "step": 2660
941
+ },
942
+ {
943
+ "epoch": 0.21,
944
+ "grad_norm": 26.394916534423828,
945
+ "learning_rate": 2.9157103476060547e-05,
946
+ "loss": 3.5944,
947
+ "step": 2680
948
+ },
949
+ {
950
+ "epoch": 0.22,
951
+ "grad_norm": 23.64137077331543,
952
+ "learning_rate": 2.914097324979651e-05,
953
+ "loss": 3.5048,
954
+ "step": 2700
955
+ },
956
+ {
957
+ "epoch": 0.22,
958
+ "grad_norm": 35.18840408325195,
959
+ "learning_rate": 2.9124694691414485e-05,
960
+ "loss": 3.5769,
961
+ "step": 2720
962
+ },
963
+ {
964
+ "epoch": 0.22,
965
+ "grad_norm": 25.55812644958496,
966
+ "learning_rate": 2.9108267971668828e-05,
967
+ "loss": 3.5082,
968
+ "step": 2740
969
  }
970
  ],
971
  "logging_steps": 20,
 
973
  "num_input_tokens_seen": 0,
974
  "num_train_epochs": 2,
975
  "save_steps": 20,
976
+ "total_flos": 6492126069227520.0,
977
  "train_batch_size": 8,
978
  "trial_name": null,
979
  "trial_params": null