dimasik87 commited on
Commit
2ebeb10
1 Parent(s): 054aee6

Training in progress, step 36, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:128525095c9c4aa2360441c5622c7962547ed143337bfac52c83e3c618dc4e84
3
  size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15e78fa49fb00271ad118e01dad1e6674dfce4334a69cbe9e3cf1f3e76ece4a4
3
  size 167832240
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dfbc3e8f45768fc4da85d1bfec94316cf2bc55ecb2cfaf6f30ff35e6567ace72
3
  size 335922386
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e12cf4a8339ad6e6ccc2e4a2b586e7d84e101d5a489f8144ee3b6d4413b9990
3
  size 335922386
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fd8ec78ff15268c9da253ecb4b146058c0df0371aefad94e4b756060686ce26d
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5daf23a304609686c8d20bb6fce1b79761fd95eb6abbd30d197b1a5200625adb
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:06c69db2db6de56f38ba12b474a491d20087e27dc2893a95d6ac7716476ca645
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0aa0f75a1f3e346be25756b578158b09a68943f0b9f1cfe29f97939687f864ef
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.08613728129205922,
5
  "eval_steps": 4,
6
- "global_step": 32,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -303,6 +303,42 @@
303
  "eval_samples_per_second": 7.797,
304
  "eval_steps_per_second": 7.797,
305
  "step": 32
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
306
  }
307
  ],
308
  "logging_steps": 1,
@@ -322,7 +358,7 @@
322
  "attributes": {}
323
  }
324
  },
325
- "total_flos": 1.3168499283197952e+16,
326
  "train_batch_size": 1,
327
  "trial_name": null,
328
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.09690444145356662,
5
  "eval_steps": 4,
6
+ "global_step": 36,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
303
  "eval_samples_per_second": 7.797,
304
  "eval_steps_per_second": 7.797,
305
  "step": 32
306
+ },
307
+ {
308
+ "epoch": 0.08882907133243607,
309
+ "grad_norm": 0.9806796312332153,
310
+ "learning_rate": 7.66554636144095e-05,
311
+ "loss": 0.3062,
312
+ "step": 33
313
+ },
314
+ {
315
+ "epoch": 0.09152086137281291,
316
+ "grad_norm": 1.1644279956817627,
317
+ "learning_rate": 6.909830056250527e-05,
318
+ "loss": 0.3511,
319
+ "step": 34
320
+ },
321
+ {
322
+ "epoch": 0.09421265141318977,
323
+ "grad_norm": 0.9300962090492249,
324
+ "learning_rate": 6.173165676349103e-05,
325
+ "loss": 0.2276,
326
+ "step": 35
327
+ },
328
+ {
329
+ "epoch": 0.09690444145356662,
330
+ "grad_norm": 0.9256501793861389,
331
+ "learning_rate": 5.4600950026045326e-05,
332
+ "loss": 0.3099,
333
+ "step": 36
334
+ },
335
+ {
336
+ "epoch": 0.09690444145356662,
337
+ "eval_loss": 0.27339035272598267,
338
+ "eval_runtime": 10.1945,
339
+ "eval_samples_per_second": 7.749,
340
+ "eval_steps_per_second": 7.749,
341
+ "step": 36
342
  }
343
  ],
344
  "logging_steps": 1,
 
358
  "attributes": {}
359
  }
360
  },
361
+ "total_flos": 1.4745009760763904e+16,
362
  "train_batch_size": 1,
363
  "trial_name": null,
364
  "trial_params": null