leixa commited on
Commit
48b2ebe
·
verified ·
1 Parent(s): e384846

Training in progress, step 210, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4cc759cf13661b4c5b335690bc8b11943275e9a8dd062881721370264ed868f6
3
  size 93608
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc8adbf18c3ea34bb1571bf5de0ce009ebe24517b8660f6e89dfa0cdf4072767
3
  size 93608
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4638d466f282e677c317e35edb73d3ec79883b19369b6efef3161077916f349e
3
  size 197158
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4652b9efe4fd390bf1401ab6edba1f9c127e67fe335840814c044c07d2681a5
3
  size 197158
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:780a66f050a6446e42e4531bfc89697709babbf1afa47eb4223703c667289080
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4aab3b1615a709a9704390ada7ef9edf3edc2d75f6ff7af351c654f587750225
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2f516a6e4e8a8eba956b80cb2ea416b9fd98f0dec12d7d9d9a36274d0eef4a63
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4bd95a1db0e917ddf11b12a343f06e907fcec4b81104002e2471b4778587b465
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.4752475247524752,
5
  "eval_steps": 42,
6
- "global_step": 168,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -439,6 +439,112 @@
439
  "eval_samples_per_second": 92.895,
440
  "eval_steps_per_second": 11.69,
441
  "step": 168
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
442
  }
443
  ],
444
  "logging_steps": 3,
@@ -458,7 +564,7 @@
458
  "attributes": {}
459
  }
460
  },
461
- "total_flos": 1025631191040.0,
462
  "train_batch_size": 8,
463
  "trial_name": null,
464
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.594059405940594,
5
  "eval_steps": 42,
6
+ "global_step": 210,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
439
  "eval_samples_per_second": 92.895,
440
  "eval_steps_per_second": 11.69,
441
  "step": 168
442
+ },
443
+ {
444
+ "epoch": 0.4837340876944837,
445
+ "grad_norm": 0.13749000430107117,
446
+ "learning_rate": 7.564496387029532e-05,
447
+ "loss": 11.8156,
448
+ "step": 171
449
+ },
450
+ {
451
+ "epoch": 0.49222065063649223,
452
+ "grad_norm": 0.07802052795886993,
453
+ "learning_rate": 7.481469062821252e-05,
454
+ "loss": 11.8182,
455
+ "step": 174
456
+ },
457
+ {
458
+ "epoch": 0.5007072135785007,
459
+ "grad_norm": 0.150814950466156,
460
+ "learning_rate": 7.39752373250527e-05,
461
+ "loss": 11.8179,
462
+ "step": 177
463
+ },
464
+ {
465
+ "epoch": 0.5091937765205092,
466
+ "grad_norm": 0.1514790952205658,
467
+ "learning_rate": 7.312691451204178e-05,
468
+ "loss": 11.8099,
469
+ "step": 180
470
+ },
471
+ {
472
+ "epoch": 0.5176803394625177,
473
+ "grad_norm": 0.13362684845924377,
474
+ "learning_rate": 7.227003602163295e-05,
475
+ "loss": 11.8172,
476
+ "step": 183
477
+ },
478
+ {
479
+ "epoch": 0.5261669024045261,
480
+ "grad_norm": 0.09337490051984787,
481
+ "learning_rate": 7.14049188514063e-05,
482
+ "loss": 11.8184,
483
+ "step": 186
484
+ },
485
+ {
486
+ "epoch": 0.5346534653465347,
487
+ "grad_norm": 0.08015663921833038,
488
+ "learning_rate": 7.05318830467969e-05,
489
+ "loss": 11.8158,
490
+ "step": 189
491
+ },
492
+ {
493
+ "epoch": 0.5431400282885431,
494
+ "grad_norm": 0.16405069828033447,
495
+ "learning_rate": 6.965125158269619e-05,
496
+ "loss": 11.816,
497
+ "step": 192
498
+ },
499
+ {
500
+ "epoch": 0.5516265912305516,
501
+ "grad_norm": 0.14057497680187225,
502
+ "learning_rate": 6.876335024396872e-05,
503
+ "loss": 11.8147,
504
+ "step": 195
505
+ },
506
+ {
507
+ "epoch": 0.5601131541725601,
508
+ "grad_norm": 0.1409187614917755,
509
+ "learning_rate": 6.786850750493006e-05,
510
+ "loss": 11.8157,
511
+ "step": 198
512
+ },
513
+ {
514
+ "epoch": 0.5685997171145686,
515
+ "grad_norm": 0.1987845003604889,
516
+ "learning_rate": 6.696705440782938e-05,
517
+ "loss": 11.8185,
518
+ "step": 201
519
+ },
520
+ {
521
+ "epoch": 0.5770862800565771,
522
+ "grad_norm": 0.10339465737342834,
523
+ "learning_rate": 6.605932444038229e-05,
524
+ "loss": 11.815,
525
+ "step": 204
526
+ },
527
+ {
528
+ "epoch": 0.5855728429985856,
529
+ "grad_norm": 0.12926329672336578,
530
+ "learning_rate": 6.514565341239861e-05,
531
+ "loss": 11.818,
532
+ "step": 207
533
+ },
534
+ {
535
+ "epoch": 0.594059405940594,
536
+ "grad_norm": 0.07949727028608322,
537
+ "learning_rate": 6.422637933155162e-05,
538
+ "loss": 11.8151,
539
+ "step": 210
540
+ },
541
+ {
542
+ "epoch": 0.594059405940594,
543
+ "eval_loss": 11.818094253540039,
544
+ "eval_runtime": 6.6444,
545
+ "eval_samples_per_second": 89.7,
546
+ "eval_steps_per_second": 11.288,
547
+ "step": 210
548
  }
549
  ],
550
  "logging_steps": 3,
 
564
  "attributes": {}
565
  }
566
  },
567
+ "total_flos": 1284750704640.0,
568
  "train_batch_size": 8,
569
  "trial_name": null,
570
  "trial_params": null