Nexspear commited on
Commit
f915a4c
·
verified ·
1 Parent(s): 0abd7ba

Training in progress, step 210, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2ffd098374c4829deef4e86bb45b6d99e299ec4d785b2818c11701cc0184f9b4
3
  size 100966336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:971e54de63920eed6700963a65c28021c96f5ce4caf96cc351d0df00fb6d803c
3
  size 100966336
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d430efc5b96ee5cc101a4108edaa9ac66a363cbaf91802f71917f34d4f30221f
3
  size 51613348
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54c1dc870cb0fc047cba8029354bda92e92efcf53a26b597de339883b64b86e3
3
  size 51613348
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e477b2f97b5d76e20fb3241eba6754e5c05c0a9ad01e9b25e3c321898b520d6
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:792d8588f6659622a3919c79527db3d98a5bae8354f3a9892e08b902b222082b
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:79783dc253db8561cbce9475c59c8adb4831f213eda27032651f04af31d6a595
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d485fc873443ef7ad597c6f4a82e93694f8fe8522b8ffecf4d60075246020043
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.025167596719224,
5
  "eval_steps": 42,
6
- "global_step": 168,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -439,6 +439,112 @@
439
  "eval_samples_per_second": 36.024,
440
  "eval_steps_per_second": 4.505,
441
  "step": 168
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
442
  }
443
  ],
444
  "logging_steps": 3,
@@ -458,7 +564,7 @@
458
  "attributes": {}
459
  }
460
  },
461
- "total_flos": 5.365118024024064e+16,
462
  "train_batch_size": 8,
463
  "trial_name": null,
464
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.03145949589903,
5
  "eval_steps": 42,
6
+ "global_step": 210,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
439
  "eval_samples_per_second": 36.024,
440
  "eval_steps_per_second": 4.505,
441
  "step": 168
442
+ },
443
+ {
444
+ "epoch": 0.02561701808921014,
445
+ "grad_norm": 0.180728018283844,
446
+ "learning_rate": 3.782248193514766e-05,
447
+ "loss": 1.0867,
448
+ "step": 171
449
+ },
450
+ {
451
+ "epoch": 0.026066439459196285,
452
+ "grad_norm": 0.15069565176963806,
453
+ "learning_rate": 3.740734531410626e-05,
454
+ "loss": 1.0624,
455
+ "step": 174
456
+ },
457
+ {
458
+ "epoch": 0.026515860829182426,
459
+ "grad_norm": 0.13892242312431335,
460
+ "learning_rate": 3.698761866252635e-05,
461
+ "loss": 1.0351,
462
+ "step": 177
463
+ },
464
+ {
465
+ "epoch": 0.02696528219916857,
466
+ "grad_norm": 0.1399199515581131,
467
+ "learning_rate": 3.656345725602089e-05,
468
+ "loss": 1.1609,
469
+ "step": 180
470
+ },
471
+ {
472
+ "epoch": 0.027414703569154715,
473
+ "grad_norm": 0.14930486679077148,
474
+ "learning_rate": 3.6135018010816477e-05,
475
+ "loss": 1.1117,
476
+ "step": 183
477
+ },
478
+ {
479
+ "epoch": 0.027864124939140856,
480
+ "grad_norm": 0.15556196868419647,
481
+ "learning_rate": 3.570245942570315e-05,
482
+ "loss": 1.1169,
483
+ "step": 186
484
+ },
485
+ {
486
+ "epoch": 0.028313546309127,
487
+ "grad_norm": 0.17272590100765228,
488
+ "learning_rate": 3.526594152339845e-05,
489
+ "loss": 1.115,
490
+ "step": 189
491
+ },
492
+ {
493
+ "epoch": 0.02876296767911314,
494
+ "grad_norm": 0.17533355951309204,
495
+ "learning_rate": 3.4825625791348096e-05,
496
+ "loss": 1.1298,
497
+ "step": 192
498
+ },
499
+ {
500
+ "epoch": 0.029212389049099285,
501
+ "grad_norm": 0.14778710901737213,
502
+ "learning_rate": 3.438167512198436e-05,
503
+ "loss": 1.1183,
504
+ "step": 195
505
+ },
506
+ {
507
+ "epoch": 0.029661810419085426,
508
+ "grad_norm": 0.14693984389305115,
509
+ "learning_rate": 3.393425375246503e-05,
510
+ "loss": 1.0647,
511
+ "step": 198
512
+ },
513
+ {
514
+ "epoch": 0.03011123178907157,
515
+ "grad_norm": 0.14994005858898163,
516
+ "learning_rate": 3.348352720391469e-05,
517
+ "loss": 1.0008,
518
+ "step": 201
519
+ },
520
+ {
521
+ "epoch": 0.03056065315905771,
522
+ "grad_norm": 0.1611510068178177,
523
+ "learning_rate": 3.3029662220191144e-05,
524
+ "loss": 1.094,
525
+ "step": 204
526
+ },
527
+ {
528
+ "epoch": 0.031010074529043856,
529
+ "grad_norm": 0.19615799188613892,
530
+ "learning_rate": 3.2572826706199305e-05,
531
+ "loss": 1.051,
532
+ "step": 207
533
+ },
534
+ {
535
+ "epoch": 0.03145949589903,
536
+ "grad_norm": 0.15789468586444855,
537
+ "learning_rate": 3.211318966577581e-05,
538
+ "loss": 1.0302,
539
+ "step": 210
540
+ },
541
+ {
542
+ "epoch": 0.03145949589903,
543
+ "eval_loss": 1.0971506834030151,
544
+ "eval_runtime": 312.684,
545
+ "eval_samples_per_second": 35.956,
546
+ "eval_steps_per_second": 4.497,
547
+ "step": 210
548
  }
549
  ],
550
  "logging_steps": 3,
 
564
  "attributes": {}
565
  }
566
  },
567
+ "total_flos": 6.724628513606861e+16,
568
  "train_batch_size": 8,
569
  "trial_name": null,
570
  "trial_params": null