Femboyuwu2000 commited on
Commit
5b17dcf
·
verified ·
1 Parent(s): 43b97da

Training in progress, step 820, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a7b375ea6f5e70ef64e4641abb6a08df506c86a9a7bb439340b989e59fe64c0c
3
  size 13982248
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c685aafc65181370b17db36243c0d171817f938725ed77fa0a1e03b53e030014
3
  size 13982248
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:033fa7ccae359fb38b200fbf2e73770b5a80ea7b2b9e5686853f4c0453109b9b
3
  size 7062522
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:036773597bfd897fa517e6ad196b274fca905a2d4b3b02935e43e14210ec8f79
3
  size 7062522
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f17d400e8d0a3279d48c33167ab70e99da7e17a6d2f06962e7ee39cc165c3506
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ad9df02f521bb1e4003457689d3b3dba0522b12dae9613bb1e420ae84cb87c5
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2f0d32846dc851291d36978371f88c4a2e97295576d6f59868838a673ab4bcbf
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7d3370081bd8faebed56ed9d5e5383e284bccc346e41b56e56c69c853ca9ca1
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.064,
5
  "eval_steps": 500,
6
- "global_step": 800,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -287,6 +287,13 @@
287
  "learning_rate": 2.9992133535682725e-05,
288
  "loss": 3.6919,
289
  "step": 800
 
 
 
 
 
 
 
290
  }
291
  ],
292
  "logging_steps": 20,
@@ -294,7 +301,7 @@
294
  "num_input_tokens_seen": 0,
295
  "num_train_epochs": 2,
296
  "save_steps": 20,
297
- "total_flos": 1896312970543104.0,
298
  "train_batch_size": 8,
299
  "trial_name": null,
300
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.0656,
5
  "eval_steps": 500,
6
+ "global_step": 820,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
287
  "learning_rate": 2.9992133535682725e-05,
288
  "loss": 3.6919,
289
  "step": 800
290
+ },
291
+ {
292
+ "epoch": 0.07,
293
+ "grad_norm": 32.69138717651367,
294
+ "learning_rate": 2.9990481752904566e-05,
295
+ "loss": 3.6855,
296
+ "step": 820
297
  }
298
  ],
299
  "logging_steps": 20,
 
301
  "num_input_tokens_seen": 0,
302
  "num_train_epochs": 2,
303
  "save_steps": 20,
304
+ "total_flos": 1942798948663296.0,
305
  "train_batch_size": 8,
306
  "trial_name": null,
307
  "trial_params": null