Femboyuwu2000 commited on
Commit
cb4194d
1 Parent(s): a7a6f22

Training in progress, step 9160, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:66f5cc7dc8a42efe36ea76d8ca34e1a3c950c58533fa878ecad8d052380ac9f8
3
  size 13982248
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20f1a714eb67c73b6569a5f7e5dedf54237b0ad2716eb3b0475c33a5c2632e6b
3
  size 13982248
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:84699a0b76749e08a4088d70c72da5ffb07c2ecf5da72c70b97e96ec7eb934f3
3
  size 7062522
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9730bd1fd19af1573bb86be461d8d2283246a50b424518f179c57f93b8e62ac4
3
  size 7062522
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:481dc395f467fef8db19044242527752744d919a7271d6ece20ea5b5cd541aa2
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:626b6a087fd94d790b874f54e98174d3e59d634f407178aef84daa0bee2f4f38
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:290442ea872222629cafe79f666d304861e371480c9ea6270813c4faff62a31b
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff87507bf7d087bfecc1378d173f4afe91a57804efc7c9f721f922c781ee2ed5
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.7312,
5
  "eval_steps": 500,
6
- "global_step": 9140,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -3206,6 +3206,13 @@
3206
  "learning_rate": 1.7801174965633898e-05,
3207
  "loss": 3.3977,
3208
  "step": 9140
 
 
 
 
 
 
 
3209
  }
3210
  ],
3211
  "logging_steps": 20,
@@ -3213,7 +3220,7 @@
3213
  "num_input_tokens_seen": 0,
3214
  "num_train_epochs": 2,
3215
  "save_steps": 20,
3216
- "total_flos": 2.161509406624973e+16,
3217
  "train_batch_size": 8,
3218
  "trial_name": null,
3219
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.7328,
5
  "eval_steps": 500,
6
+ "global_step": 9160,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
3206
  "learning_rate": 1.7801174965633898e-05,
3207
  "loss": 3.3977,
3208
  "step": 9140
3209
+ },
3210
+ {
3211
+ "epoch": 0.73,
3212
+ "grad_norm": 32.8968620300293,
3213
+ "learning_rate": 1.775343365017254e-05,
3214
+ "loss": 3.5308,
3215
+ "step": 9160
3216
  }
3217
  ],
3218
  "logging_steps": 20,
 
3220
  "num_input_tokens_seen": 0,
3221
  "num_train_epochs": 2,
3222
  "save_steps": 20,
3223
+ "total_flos": 2.166026780786688e+16,
3224
  "train_batch_size": 8,
3225
  "trial_name": null,
3226
  "trial_params": null