leixa commited on
Commit
10f0301
·
verified ·
1 Parent(s): cd67608

Training in progress, step 126, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f96fe0f77223bf21e72bb8cec824d6e8118b382aeb4fe8d6ab68534e260cbf8d
3
  size 93608
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d06192d758f599c81e887b2a13e19ff974602cb65fdc40c23689ba90f5a0cc51
3
  size 93608
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b9ecd8d4edbc1eb01635b7969b0317e2e2ac7ee4e46b9c2667bf035cf5a6df33
3
  size 197158
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ef7a8706f17569c6dea0871a2304a6f7d337bb45555726571afc0a2d7edb4eb
3
  size 197158
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:103b84660003cbb55211ca46c75dea4086b05940b67e924565c8120dfc91f67a
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59a28cc54d01a2cacfb06a120854babf31065aeadae401292bfd08b1b5f75518
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2df224011d0e75c4a97901f6c1b2930bba4bc3a9aa7c877e6c91e796bec6013f
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e00328545513b592b6bf311a002e087bb048e374276594cf37ee8782d28887fc
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.2376237623762376,
5
  "eval_steps": 42,
6
- "global_step": 84,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -227,6 +227,112 @@
227
  "eval_samples_per_second": 93.097,
228
  "eval_steps_per_second": 11.715,
229
  "step": 84
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
230
  }
231
  ],
232
  "logging_steps": 3,
@@ -246,7 +352,7 @@
246
  "attributes": {}
247
  }
248
  },
249
- "total_flos": 503776542720.0,
250
  "train_batch_size": 8,
251
  "trial_name": null,
252
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.3564356435643564,
5
  "eval_steps": 42,
6
+ "global_step": 126,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
227
  "eval_samples_per_second": 93.097,
228
  "eval_steps_per_second": 11.715,
229
  "step": 84
230
+ },
231
+ {
232
+ "epoch": 0.24611032531824611,
233
+ "grad_norm": 0.28861185908317566,
234
+ "learning_rate": 9.40297765928369e-05,
235
+ "loss": 11.8219,
236
+ "step": 87
237
+ },
238
+ {
239
+ "epoch": 0.2545968882602546,
240
+ "grad_norm": 0.13039568066596985,
241
+ "learning_rate": 9.356593520616948e-05,
242
+ "loss": 11.8245,
243
+ "step": 90
244
+ },
245
+ {
246
+ "epoch": 0.26308345120226306,
247
+ "grad_norm": 0.1711033284664154,
248
+ "learning_rate": 9.308597683653975e-05,
249
+ "loss": 11.8246,
250
+ "step": 93
251
+ },
252
+ {
253
+ "epoch": 0.27157001414427157,
254
+ "grad_norm": 0.2108013778924942,
255
+ "learning_rate": 9.259007904196023e-05,
256
+ "loss": 11.8228,
257
+ "step": 96
258
+ },
259
+ {
260
+ "epoch": 0.28005657708628007,
261
+ "grad_norm": 0.1204076036810875,
262
+ "learning_rate": 9.207842527714767e-05,
263
+ "loss": 11.82,
264
+ "step": 99
265
+ },
266
+ {
267
+ "epoch": 0.28854314002828857,
268
+ "grad_norm": 0.1477670669555664,
269
+ "learning_rate": 9.155120482565521e-05,
270
+ "loss": 11.8189,
271
+ "step": 102
272
+ },
273
+ {
274
+ "epoch": 0.297029702970297,
275
+ "grad_norm": 0.10705884546041489,
276
+ "learning_rate": 9.10086127298478e-05,
277
+ "loss": 11.8169,
278
+ "step": 105
279
+ },
280
+ {
281
+ "epoch": 0.3055162659123055,
282
+ "grad_norm": 0.29956066608428955,
283
+ "learning_rate": 9.045084971874738e-05,
284
+ "loss": 11.8205,
285
+ "step": 108
286
+ },
287
+ {
288
+ "epoch": 0.314002828854314,
289
+ "grad_norm": 0.13027921319007874,
290
+ "learning_rate": 8.987812213377424e-05,
291
+ "loss": 11.8168,
292
+ "step": 111
293
+ },
294
+ {
295
+ "epoch": 0.32248939179632247,
296
+ "grad_norm": 0.13590934872627258,
297
+ "learning_rate": 8.929064185241213e-05,
298
+ "loss": 11.819,
299
+ "step": 114
300
+ },
301
+ {
302
+ "epoch": 0.33097595473833097,
303
+ "grad_norm": 0.09333682060241699,
304
+ "learning_rate": 8.868862620982534e-05,
305
+ "loss": 11.8267,
306
+ "step": 117
307
+ },
308
+ {
309
+ "epoch": 0.33946251768033947,
310
+ "grad_norm": 0.12400602549314499,
311
+ "learning_rate": 8.807229791845673e-05,
312
+ "loss": 11.818,
313
+ "step": 120
314
+ },
315
+ {
316
+ "epoch": 0.347949080622348,
317
+ "grad_norm": 0.12071343511343002,
318
+ "learning_rate": 8.744188498563641e-05,
319
+ "loss": 11.8166,
320
+ "step": 123
321
+ },
322
+ {
323
+ "epoch": 0.3564356435643564,
324
+ "grad_norm": 0.1693616658449173,
325
+ "learning_rate": 8.679762062923175e-05,
326
+ "loss": 11.8183,
327
+ "step": 126
328
+ },
329
+ {
330
+ "epoch": 0.3564356435643564,
331
+ "eval_loss": 11.819117546081543,
332
+ "eval_runtime": 6.7651,
333
+ "eval_samples_per_second": 88.1,
334
+ "eval_steps_per_second": 11.086,
335
+ "step": 126
336
  }
337
  ],
338
  "logging_steps": 3,
 
352
  "attributes": {}
353
  }
354
  },
355
+ "total_flos": 765306470400.0,
356
  "train_batch_size": 8,
357
  "trial_name": null,
358
  "trial_params": null