dimasik87 commited on
Commit
9df09b4
·
verified ·
1 Parent(s): 9a15458

Training in progress, step 50, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bb435ae44e1260bf001862b4712cedf6e7ccb6e2947659ca26e077ee9f463da7
3
  size 493712
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4de474944576d9ce13db25076c3fb637b6714a154785e9424e8d7d628b3e7bf9
3
  size 493712
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:11d672426f8a4e05ea1ed343c9f10c28fb8ba064956e5b138f61d885146b0673
3
  size 997038
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b7aab4b77062896ba7df2e1f26e5d1679be401c5426dd31c303cd282f95aa0c
3
  size 997038
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8ccb34b93699f078ddb82b9a167e79da1e720b4cd2112f0de5c0f8f14e38bd22
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eba8ef7ef80dbd60fbb694437f7d04ba5db6a3825fc93a9958bf4f0903e7337d
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6676fe28230ae15b45fb334c871c6fdf1a7984a935952b9f8650896c37a8c106
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1df0528620c07325b8faa7567e59b0c1e86a1f1ee6af1245a69c6c0463fe4e2
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.006348668895754857,
5
  "eval_steps": 6,
6
- "global_step": 25,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -222,6 +222,213 @@
222
  "learning_rate": 6.91341716182545e-05,
223
  "loss": 66.0543,
224
  "step": 25
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
225
  }
226
  ],
227
  "logging_steps": 1,
@@ -236,12 +443,12 @@
236
  "should_evaluate": false,
237
  "should_log": false,
238
  "should_save": true,
239
- "should_training_stop": false
240
  },
241
  "attributes": {}
242
  }
243
  },
244
- "total_flos": 3572917862400.0,
245
  "train_batch_size": 4,
246
  "trial_name": null,
247
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.012697337791509713,
5
  "eval_steps": 6,
6
+ "global_step": 50,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
222
  "learning_rate": 6.91341716182545e-05,
223
  "loss": 66.0543,
224
  "step": 25
225
+ },
226
+ {
227
+ "epoch": 0.006602615651585051,
228
+ "grad_norm": 2.5673084259033203,
229
+ "learning_rate": 6.545084971874738e-05,
230
+ "loss": 66.0834,
231
+ "step": 26
232
+ },
233
+ {
234
+ "epoch": 0.006856562407415245,
235
+ "grad_norm": 2.866518259048462,
236
+ "learning_rate": 6.167226819279528e-05,
237
+ "loss": 66.044,
238
+ "step": 27
239
+ },
240
+ {
241
+ "epoch": 0.00711050916324544,
242
+ "grad_norm": 2.417178153991699,
243
+ "learning_rate": 5.782172325201155e-05,
244
+ "loss": 66.1386,
245
+ "step": 28
246
+ },
247
+ {
248
+ "epoch": 0.007364455919075634,
249
+ "grad_norm": 2.504648447036743,
250
+ "learning_rate": 5.392295478639225e-05,
251
+ "loss": 66.046,
252
+ "step": 29
253
+ },
254
+ {
255
+ "epoch": 0.007618402674905828,
256
+ "grad_norm": 2.539602279663086,
257
+ "learning_rate": 5e-05,
258
+ "loss": 66.0429,
259
+ "step": 30
260
+ },
261
+ {
262
+ "epoch": 0.007618402674905828,
263
+ "eval_loss": 11.006031036376953,
264
+ "eval_runtime": 18.2105,
265
+ "eval_samples_per_second": 273.194,
266
+ "eval_steps_per_second": 68.312,
267
+ "step": 30
268
+ },
269
+ {
270
+ "epoch": 0.007872349430736022,
271
+ "grad_norm": 2.809628963470459,
272
+ "learning_rate": 4.607704521360776e-05,
273
+ "loss": 65.9401,
274
+ "step": 31
275
+ },
276
+ {
277
+ "epoch": 0.008126296186566217,
278
+ "grad_norm": 2.533236026763916,
279
+ "learning_rate": 4.2178276747988446e-05,
280
+ "loss": 66.0943,
281
+ "step": 32
282
+ },
283
+ {
284
+ "epoch": 0.00838024294239641,
285
+ "grad_norm": 2.7511136531829834,
286
+ "learning_rate": 3.832773180720475e-05,
287
+ "loss": 65.9569,
288
+ "step": 33
289
+ },
290
+ {
291
+ "epoch": 0.008634189698226605,
292
+ "grad_norm": 2.529318332672119,
293
+ "learning_rate": 3.4549150281252636e-05,
294
+ "loss": 65.9841,
295
+ "step": 34
296
+ },
297
+ {
298
+ "epoch": 0.008888136454056799,
299
+ "grad_norm": 2.60366153717041,
300
+ "learning_rate": 3.086582838174551e-05,
301
+ "loss": 65.9559,
302
+ "step": 35
303
+ },
304
+ {
305
+ "epoch": 0.009142083209886994,
306
+ "grad_norm": 2.6384119987487793,
307
+ "learning_rate": 2.7300475013022663e-05,
308
+ "loss": 65.9892,
309
+ "step": 36
310
+ },
311
+ {
312
+ "epoch": 0.009142083209886994,
313
+ "eval_loss": 10.992709159851074,
314
+ "eval_runtime": 18.2216,
315
+ "eval_samples_per_second": 273.028,
316
+ "eval_steps_per_second": 68.271,
317
+ "step": 36
318
+ },
319
+ {
320
+ "epoch": 0.009396029965717189,
321
+ "grad_norm": 2.68752121925354,
322
+ "learning_rate": 2.3875071764202563e-05,
323
+ "loss": 65.8811,
324
+ "step": 37
325
+ },
326
+ {
327
+ "epoch": 0.009649976721547382,
328
+ "grad_norm": 2.529193162918091,
329
+ "learning_rate": 2.061073738537635e-05,
330
+ "loss": 65.97,
331
+ "step": 38
332
+ },
333
+ {
334
+ "epoch": 0.009903923477377577,
335
+ "grad_norm": 2.684138059616089,
336
+ "learning_rate": 1.7527597583490822e-05,
337
+ "loss": 65.8963,
338
+ "step": 39
339
+ },
340
+ {
341
+ "epoch": 0.01015787023320777,
342
+ "grad_norm": 2.5320022106170654,
343
+ "learning_rate": 1.4644660940672627e-05,
344
+ "loss": 65.9401,
345
+ "step": 40
346
+ },
347
+ {
348
+ "epoch": 0.010411816989037965,
349
+ "grad_norm": 2.516324281692505,
350
+ "learning_rate": 1.1979701719998453e-05,
351
+ "loss": 65.9055,
352
+ "step": 41
353
+ },
354
+ {
355
+ "epoch": 0.010665763744868158,
356
+ "grad_norm": 2.543747901916504,
357
+ "learning_rate": 9.549150281252633e-06,
358
+ "loss": 65.8852,
359
+ "step": 42
360
+ },
361
+ {
362
+ "epoch": 0.010665763744868158,
363
+ "eval_loss": 10.98641300201416,
364
+ "eval_runtime": 18.1929,
365
+ "eval_samples_per_second": 273.459,
366
+ "eval_steps_per_second": 68.378,
367
+ "step": 42
368
+ },
369
+ {
370
+ "epoch": 0.010919710500698353,
371
+ "grad_norm": 2.3728582859039307,
372
+ "learning_rate": 7.367991782295391e-06,
373
+ "loss": 66.1291,
374
+ "step": 43
375
+ },
376
+ {
377
+ "epoch": 0.011173657256528548,
378
+ "grad_norm": 2.4906511306762695,
379
+ "learning_rate": 5.449673790581611e-06,
380
+ "loss": 65.9421,
381
+ "step": 44
382
+ },
383
+ {
384
+ "epoch": 0.011427604012358742,
385
+ "grad_norm": 2.665292978286743,
386
+ "learning_rate": 3.8060233744356633e-06,
387
+ "loss": 65.9262,
388
+ "step": 45
389
+ },
390
+ {
391
+ "epoch": 0.011681550768188937,
392
+ "grad_norm": 2.6886563301086426,
393
+ "learning_rate": 2.4471741852423237e-06,
394
+ "loss": 65.881,
395
+ "step": 46
396
+ },
397
+ {
398
+ "epoch": 0.01193549752401913,
399
+ "grad_norm": 2.69474196434021,
400
+ "learning_rate": 1.3815039801161721e-06,
401
+ "loss": 65.8305,
402
+ "step": 47
403
+ },
404
+ {
405
+ "epoch": 0.012189444279849325,
406
+ "grad_norm": 2.3172616958618164,
407
+ "learning_rate": 6.15582970243117e-07,
408
+ "loss": 65.9682,
409
+ "step": 48
410
+ },
411
+ {
412
+ "epoch": 0.012189444279849325,
413
+ "eval_loss": 10.984763145446777,
414
+ "eval_runtime": 18.4652,
415
+ "eval_samples_per_second": 269.425,
416
+ "eval_steps_per_second": 67.37,
417
+ "step": 48
418
+ },
419
+ {
420
+ "epoch": 0.01244339103567952,
421
+ "grad_norm": 2.542539119720459,
422
+ "learning_rate": 1.5413331334360182e-07,
423
+ "loss": 66.0153,
424
+ "step": 49
425
+ },
426
+ {
427
+ "epoch": 0.012697337791509713,
428
+ "grad_norm": 2.630923271179199,
429
+ "learning_rate": 0.0,
430
+ "loss": 65.7818,
431
+ "step": 50
432
  }
433
  ],
434
  "logging_steps": 1,
 
443
  "should_evaluate": false,
444
  "should_log": false,
445
  "should_save": true,
446
+ "should_training_stop": true
447
  },
448
  "attributes": {}
449
  }
450
  },
451
+ "total_flos": 7145835724800.0,
452
  "train_batch_size": 4,
453
  "trial_name": null,
454
  "trial_params": null