dimasik87 commited on
Commit
96f3e88
·
verified ·
1 Parent(s): 6888a31

Training in progress, step 36, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c10cd879656c14f6c67706e5ef6a6ed5392b7455783b83226008557ab6c1c197
3
  size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7173f2095eaefc6173f2296bc5ca0608b2758d7af43d3295f6d0b0fa07c89c2f
3
  size 167832240
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:727c0144258ef51a1f515989e494415a1427822d9138284e64c8b3c60045bfff
3
  size 335922386
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe914b1d9853545e2ace71d2ea9f0c4da3dfd1c4d91fbd902fd559e493e92959
3
  size 335922386
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cc78d73481e3ba72a3d6fb04344d343a7f6ed3f4dbca1980cc2dfd7e0b74a18a
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed20cb1e030baf9351dd9705735eda93875b2582f4380baf52cb1cb6e628a469
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:06c69db2db6de56f38ba12b474a491d20087e27dc2893a95d6ac7716476ca645
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0aa0f75a1f3e346be25756b578158b09a68943f0b9f1cfe29f97939687f864ef
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.00958155550565162,
5
  "eval_steps": 4,
6
- "global_step": 32,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -303,6 +303,42 @@
303
  "eval_samples_per_second": 8.439,
304
  "eval_steps_per_second": 8.439,
305
  "step": 32
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
306
  }
307
  ],
308
  "logging_steps": 1,
@@ -322,7 +358,7 @@
322
  "attributes": {}
323
  }
324
  },
325
- "total_flos": 1.1870196536967168e+16,
326
  "train_batch_size": 1,
327
  "trial_name": null,
328
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.010779249943858073,
5
  "eval_steps": 4,
6
+ "global_step": 36,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
303
  "eval_samples_per_second": 8.439,
304
  "eval_steps_per_second": 8.439,
305
  "step": 32
306
+ },
307
+ {
308
+ "epoch": 0.009880979115203234,
309
+ "grad_norm": 41.4135627746582,
310
+ "learning_rate": 7.66554636144095e-05,
311
+ "loss": 3.9928,
312
+ "step": 33
313
+ },
314
+ {
315
+ "epoch": 0.010180402724754846,
316
+ "grad_norm": 29.63922882080078,
317
+ "learning_rate": 6.909830056250527e-05,
318
+ "loss": 2.1258,
319
+ "step": 34
320
+ },
321
+ {
322
+ "epoch": 0.01047982633430646,
323
+ "grad_norm": 34.11370086669922,
324
+ "learning_rate": 6.173165676349103e-05,
325
+ "loss": 2.8219,
326
+ "step": 35
327
+ },
328
+ {
329
+ "epoch": 0.010779249943858073,
330
+ "grad_norm": 19.921571731567383,
331
+ "learning_rate": 5.4600950026045326e-05,
332
+ "loss": 4.0769,
333
+ "step": 36
334
+ },
335
+ {
336
+ "epoch": 0.010779249943858073,
337
+ "eval_loss": 3.3533554077148438,
338
+ "eval_runtime": 83.2266,
339
+ "eval_samples_per_second": 8.459,
340
+ "eval_steps_per_second": 8.459,
341
+ "step": 36
342
  }
343
  ],
344
  "logging_steps": 1,
 
358
  "attributes": {}
359
  }
360
  },
361
+ "total_flos": 1.3353971104088064e+16,
362
  "train_batch_size": 1,
363
  "trial_name": null,
364
  "trial_params": null