Nexspear commited on
Commit
9b98fd9
·
verified ·
1 Parent(s): c42c553

Training in progress, step 168, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:30d12559217c7f23d41501f47dca546b26251a84fd89ee55f980fc13aa93f12b
3
  size 100966336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ffd098374c4829deef4e86bb45b6d99e299ec4d785b2818c11701cc0184f9b4
3
  size 100966336
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b2fa88fe8784ca0c128965c9ad8a5a8e29ef38b586f101c90b586476e3ed4558
3
  size 51613348
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d430efc5b96ee5cc101a4108edaa9ac66a363cbaf91802f71917f34d4f30221f
3
  size 51613348
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ac024c06a7fafa2800668c036cb57c23ab71073198d0a72d5ca171de9c65f87b
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e477b2f97b5d76e20fb3241eba6754e5c05c0a9ad01e9b25e3c321898b520d6
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d6789ad2e7141c857db1a1b89c580a274f669224bfc6f48fb1745262d15c182
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79783dc253db8561cbce9475c59c8adb4831f213eda27032651f04af31d6a595
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.018875697539418,
5
  "eval_steps": 42,
6
- "global_step": 126,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -333,6 +333,112 @@
333
  "eval_samples_per_second": 35.97,
334
  "eval_steps_per_second": 4.498,
335
  "step": 126
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
336
  }
337
  ],
338
  "logging_steps": 3,
@@ -352,7 +458,7 @@
352
  "attributes": {}
353
  }
354
  },
355
- "total_flos": 4.03686064914432e+16,
356
  "train_batch_size": 8,
357
  "trial_name": null,
358
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.025167596719224,
5
  "eval_steps": 42,
6
+ "global_step": 168,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
333
  "eval_samples_per_second": 35.97,
334
  "eval_steps_per_second": 4.498,
335
  "step": 126
336
+ },
337
+ {
338
+ "epoch": 0.01932511890940414,
339
+ "grad_norm": 0.13838274776935577,
340
+ "learning_rate": 4.306987159568479e-05,
341
+ "loss": 0.9996,
342
+ "step": 129
343
+ },
344
+ {
345
+ "epoch": 0.019774540279390285,
346
+ "grad_norm": 0.14386090636253357,
347
+ "learning_rate": 4.273424802513145e-05,
348
+ "loss": 0.9947,
349
+ "step": 132
350
+ },
351
+ {
352
+ "epoch": 0.020223961649376426,
353
+ "grad_norm": 0.14473505318164825,
354
+ "learning_rate": 4.239206376508717e-05,
355
+ "loss": 1.1779,
356
+ "step": 135
357
+ },
358
+ {
359
+ "epoch": 0.02067338301936257,
360
+ "grad_norm": 0.15975573658943176,
361
+ "learning_rate": 4.204344540477499e-05,
362
+ "loss": 1.2125,
363
+ "step": 138
364
+ },
365
+ {
366
+ "epoch": 0.021122804389348715,
367
+ "grad_norm": 0.12576735019683838,
368
+ "learning_rate": 4.16885219136787e-05,
369
+ "loss": 1.1312,
370
+ "step": 141
371
+ },
372
+ {
373
+ "epoch": 0.021572225759334856,
374
+ "grad_norm": 0.16916967928409576,
375
+ "learning_rate": 4.132742459383122e-05,
376
+ "loss": 1.0823,
377
+ "step": 144
378
+ },
379
+ {
380
+ "epoch": 0.022021647129321,
381
+ "grad_norm": 0.17954471707344055,
382
+ "learning_rate": 4.096028703124014e-05,
383
+ "loss": 1.0728,
384
+ "step": 147
385
+ },
386
+ {
387
+ "epoch": 0.02247106849930714,
388
+ "grad_norm": 0.11972087621688843,
389
+ "learning_rate": 4.058724504646834e-05,
390
+ "loss": 1.0886,
391
+ "step": 150
392
+ },
393
+ {
394
+ "epoch": 0.022920489869293285,
395
+ "grad_norm": 0.14640676975250244,
396
+ "learning_rate": 4.0208436644387834e-05,
397
+ "loss": 1.0472,
398
+ "step": 153
399
+ },
400
+ {
401
+ "epoch": 0.023369911239279426,
402
+ "grad_norm": 0.13775284588336945,
403
+ "learning_rate": 3.982400196312564e-05,
404
+ "loss": 1.0984,
405
+ "step": 156
406
+ },
407
+ {
408
+ "epoch": 0.02381933260926557,
409
+ "grad_norm": 0.14184071123600006,
410
+ "learning_rate": 3.943408322222049e-05,
411
+ "loss": 1.1431,
412
+ "step": 159
413
+ },
414
+ {
415
+ "epoch": 0.024268753979251715,
416
+ "grad_norm": 0.1685701608657837,
417
+ "learning_rate": 3.903882467000937e-05,
418
+ "loss": 1.1531,
419
+ "step": 162
420
+ },
421
+ {
422
+ "epoch": 0.024718175349237856,
423
+ "grad_norm": 0.1468999981880188,
424
+ "learning_rate": 3.8638372530263715e-05,
425
+ "loss": 1.1069,
426
+ "step": 165
427
+ },
428
+ {
429
+ "epoch": 0.025167596719224,
430
+ "grad_norm": 0.13278649747371674,
431
+ "learning_rate": 3.823287494809469e-05,
432
+ "loss": 1.1124,
433
+ "step": 168
434
+ },
435
+ {
436
+ "epoch": 0.025167596719224,
437
+ "eval_loss": 1.1012423038482666,
438
+ "eval_runtime": 312.1008,
439
+ "eval_samples_per_second": 36.024,
440
+ "eval_steps_per_second": 4.505,
441
+ "step": 168
442
  }
443
  ],
444
  "logging_steps": 3,
 
458
  "attributes": {}
459
  }
460
  },
461
+ "total_flos": 5.365118024024064e+16,
462
  "train_batch_size": 8,
463
  "trial_name": null,
464
  "trial_params": null