dimasik87 commited on
Commit
4ddc78a
·
verified ·
1 Parent(s): 0d36278

Training in progress, step 32, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8a7b38936076e5b11391ceccd250e1610af1e6bc87425413dd6f9433acbb5e7a
3
  size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c10cd879656c14f6c67706e5ef6a6ed5392b7455783b83226008557ab6c1c197
3
  size 167832240
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f6bb764c63ad42972788691621061a20dccf3a6ff14d4f48e016cc4376f18ad
3
  size 335922386
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:727c0144258ef51a1f515989e494415a1427822d9138284e64c8b3c60045bfff
3
  size 335922386
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7e654dc294f85bc7763a518d429cb3f8268e4e0d46f7022676840843e0297900
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc78d73481e3ba72a3d6fb04344d343a7f6ed3f4dbca1980cc2dfd7e0b74a18a
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ff994fffd2fb6fe21545e6fbc55baa2a1474438a89b2d40605678f7de701427c
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06c69db2db6de56f38ba12b474a491d20087e27dc2893a95d6ac7716476ca645
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.008383861067445169,
5
  "eval_steps": 4,
6
- "global_step": 28,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -267,6 +267,42 @@
267
  "eval_samples_per_second": 8.468,
268
  "eval_steps_per_second": 8.468,
269
  "step": 28
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
270
  }
271
  ],
272
  "logging_steps": 1,
@@ -286,7 +322,7 @@
286
  "attributes": {}
287
  }
288
  },
289
- "total_flos": 1.0386421969846272e+16,
290
  "train_batch_size": 1,
291
  "trial_name": null,
292
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.00958155550565162,
5
  "eval_steps": 4,
6
+ "global_step": 32,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
267
  "eval_samples_per_second": 8.468,
268
  "eval_steps_per_second": 8.468,
269
  "step": 28
270
+ },
271
+ {
272
+ "epoch": 0.008683284676996781,
273
+ "grad_norm": 16.283584594726562,
274
+ "learning_rate": 0.0001078459095727845,
275
+ "loss": 3.5697,
276
+ "step": 29
277
+ },
278
+ {
279
+ "epoch": 0.008982708286548395,
280
+ "grad_norm": 32.41582489013672,
281
+ "learning_rate": 0.0001,
282
+ "loss": 3.4125,
283
+ "step": 30
284
+ },
285
+ {
286
+ "epoch": 0.009282131896100007,
287
+ "grad_norm": 26.877944946289062,
288
+ "learning_rate": 9.215409042721552e-05,
289
+ "loss": 3.3671,
290
+ "step": 31
291
+ },
292
+ {
293
+ "epoch": 0.00958155550565162,
294
+ "grad_norm": 27.856897354125977,
295
+ "learning_rate": 8.435655349597689e-05,
296
+ "loss": 3.5875,
297
+ "step": 32
298
+ },
299
+ {
300
+ "epoch": 0.00958155550565162,
301
+ "eval_loss": 3.401846408843994,
302
+ "eval_runtime": 83.4215,
303
+ "eval_samples_per_second": 8.439,
304
+ "eval_steps_per_second": 8.439,
305
+ "step": 32
306
  }
307
  ],
308
  "logging_steps": 1,
 
322
  "attributes": {}
323
  }
324
  },
325
+ "total_flos": 1.1870196536967168e+16,
326
  "train_batch_size": 1,
327
  "trial_name": null,
328
  "trial_params": null