Femboyuwu2000 commited on
Commit
843a8b6
·
verified ·
1 Parent(s): 9257edc

Training in progress, step 9140, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a26110e7834c1f75bf667cb78a27934275699342535f70d91611de4b759e11f
3
  size 13982248
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66f5cc7dc8a42efe36ea76d8ca34e1a3c950c58533fa878ecad8d052380ac9f8
3
  size 13982248
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b9b1ac0cee5529382bf0af2e23e74569e060fdad0a5c8ce28b8a4575e5bbbd4a
3
  size 7062522
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84699a0b76749e08a4088d70c72da5ffb07c2ecf5da72c70b97e96ec7eb934f3
3
  size 7062522
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d92688ab149ae3d2e8540b107e36e29f9346aed504282ef57b5b6880174ef198
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:481dc395f467fef8db19044242527752744d919a7271d6ece20ea5b5cd541aa2
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8394cc03bc74834b66f0ae2abaadc4e104cc02be3d2024bd053349a26e0c7ddd
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:290442ea872222629cafe79f666d304861e371480c9ea6270813c4faff62a31b
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.7296,
5
  "eval_steps": 500,
6
- "global_step": 9120,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -3199,6 +3199,13 @@
3199
  "learning_rate": 1.7848886898095356e-05,
3200
  "loss": 3.4606,
3201
  "step": 9120
 
 
 
 
 
 
 
3202
  }
3203
  ],
3204
  "logging_steps": 20,
@@ -3206,7 +3213,7 @@
3206
  "num_input_tokens_seen": 0,
3207
  "num_train_epochs": 2,
3208
  "save_steps": 20,
3209
- "total_flos": 2.156355597759283e+16,
3210
  "train_batch_size": 8,
3211
  "trial_name": null,
3212
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.7312,
5
  "eval_steps": 500,
6
+ "global_step": 9140,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
3199
  "learning_rate": 1.7848886898095356e-05,
3200
  "loss": 3.4606,
3201
  "step": 9120
3202
+ },
3203
+ {
3204
+ "epoch": 0.73,
3205
+ "grad_norm": 31.924890518188477,
3206
+ "learning_rate": 1.7801174965633898e-05,
3207
+ "loss": 3.3977,
3208
+ "step": 9140
3209
  }
3210
  ],
3211
  "logging_steps": 20,
 
3213
  "num_input_tokens_seen": 0,
3214
  "num_train_epochs": 2,
3215
  "save_steps": 20,
3216
+ "total_flos": 2.161509406624973e+16,
3217
  "train_batch_size": 8,
3218
  "trial_name": null,
3219
  "trial_params": null