dimasik1987 commited on
Commit
e79d4cd
1 Parent(s): e9a0126

Training in progress, step 50, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:75a6cf38fef33765a1f5520bedfaac160cbb9223015e7a3ab3786bd8ceaa432e
3
  size 335604696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3458b1c788586fb9c34a610545b1a3d8f52bc0a66dc193ccd31e105065a9a90
3
  size 335604696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:59d1dc215800f312848df48efb337364e9be3d58be6fd5a6b494aaa343453313
3
  size 671466706
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b887cbcf5af4c529c39d41f4fdc0940a1f8b3ca1b02a94ace29737bd618856d7
3
  size 671466706
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cbd18a87b1347fb9cc79cc3548bb35011c7972900ba3127f9f74a79279cd6e26
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f533f534d4800805436810cfc9380e579dd299a63eee9349424167eb9af0e4a
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6676fe28230ae15b45fb334c871c6fdf1a7984a935952b9f8650896c37a8c106
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1df0528620c07325b8faa7567e59b0c1e86a1f1ee6af1245a69c6c0463fe4e2
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.0410958904109589,
5
  "eval_steps": 6,
6
- "global_step": 25,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -222,6 +222,213 @@
222
  "learning_rate": 6.91341716182545e-05,
223
  "loss": 0.7425,
224
  "step": 25
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
225
  }
226
  ],
227
  "logging_steps": 1,
@@ -236,12 +443,12 @@
236
  "should_evaluate": false,
237
  "should_log": false,
238
  "should_save": true,
239
- "should_training_stop": false
240
  },
241
  "attributes": {}
242
  }
243
  },
244
- "total_flos": 1.119015678246912e+17,
245
  "train_batch_size": 4,
246
  "trial_name": null,
247
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.0821917808219178,
5
  "eval_steps": 6,
6
+ "global_step": 50,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
222
  "learning_rate": 6.91341716182545e-05,
223
  "loss": 0.7425,
224
  "step": 25
225
+ },
226
+ {
227
+ "epoch": 0.042739726027397264,
228
+ "grad_norm": 1.6898382902145386,
229
+ "learning_rate": 6.545084971874738e-05,
230
+ "loss": 0.7092,
231
+ "step": 26
232
+ },
233
+ {
234
+ "epoch": 0.04438356164383562,
235
+ "grad_norm": 1.588605523109436,
236
+ "learning_rate": 6.167226819279528e-05,
237
+ "loss": 0.7136,
238
+ "step": 27
239
+ },
240
+ {
241
+ "epoch": 0.046027397260273974,
242
+ "grad_norm": 2.993543863296509,
243
+ "learning_rate": 5.782172325201155e-05,
244
+ "loss": 0.7542,
245
+ "step": 28
246
+ },
247
+ {
248
+ "epoch": 0.04767123287671233,
249
+ "grad_norm": 2.4193227291107178,
250
+ "learning_rate": 5.392295478639225e-05,
251
+ "loss": 0.7245,
252
+ "step": 29
253
+ },
254
+ {
255
+ "epoch": 0.049315068493150684,
256
+ "grad_norm": 0.8837425708770752,
257
+ "learning_rate": 5e-05,
258
+ "loss": 0.6908,
259
+ "step": 30
260
+ },
261
+ {
262
+ "epoch": 0.049315068493150684,
263
+ "eval_loss": 0.7084067463874817,
264
+ "eval_runtime": 401.4002,
265
+ "eval_samples_per_second": 1.916,
266
+ "eval_steps_per_second": 0.481,
267
+ "step": 30
268
+ },
269
+ {
270
+ "epoch": 0.05095890410958904,
271
+ "grad_norm": 1.4400691986083984,
272
+ "learning_rate": 4.607704521360776e-05,
273
+ "loss": 0.6718,
274
+ "step": 31
275
+ },
276
+ {
277
+ "epoch": 0.0526027397260274,
278
+ "grad_norm": 1.405909538269043,
279
+ "learning_rate": 4.2178276747988446e-05,
280
+ "loss": 0.7016,
281
+ "step": 32
282
+ },
283
+ {
284
+ "epoch": 0.054246575342465755,
285
+ "grad_norm": 1.128387689590454,
286
+ "learning_rate": 3.832773180720475e-05,
287
+ "loss": 0.7065,
288
+ "step": 33
289
+ },
290
+ {
291
+ "epoch": 0.05589041095890411,
292
+ "grad_norm": 0.8258553147315979,
293
+ "learning_rate": 3.4549150281252636e-05,
294
+ "loss": 0.68,
295
+ "step": 34
296
+ },
297
+ {
298
+ "epoch": 0.057534246575342465,
299
+ "grad_norm": 1.6371694803237915,
300
+ "learning_rate": 3.086582838174551e-05,
301
+ "loss": 0.7027,
302
+ "step": 35
303
+ },
304
+ {
305
+ "epoch": 0.05917808219178082,
306
+ "grad_norm": 1.7670177221298218,
307
+ "learning_rate": 2.7300475013022663e-05,
308
+ "loss": 0.7175,
309
+ "step": 36
310
+ },
311
+ {
312
+ "epoch": 0.05917808219178082,
313
+ "eval_loss": 0.7042004466056824,
314
+ "eval_runtime": 401.6237,
315
+ "eval_samples_per_second": 1.915,
316
+ "eval_steps_per_second": 0.481,
317
+ "step": 36
318
+ },
319
+ {
320
+ "epoch": 0.060821917808219175,
321
+ "grad_norm": 4.805910110473633,
322
+ "learning_rate": 2.3875071764202563e-05,
323
+ "loss": 0.7525,
324
+ "step": 37
325
+ },
326
+ {
327
+ "epoch": 0.06246575342465754,
328
+ "grad_norm": 2.293574810028076,
329
+ "learning_rate": 2.061073738537635e-05,
330
+ "loss": 0.7078,
331
+ "step": 38
332
+ },
333
+ {
334
+ "epoch": 0.06410958904109589,
335
+ "grad_norm": 1.8330103158950806,
336
+ "learning_rate": 1.7527597583490822e-05,
337
+ "loss": 0.6966,
338
+ "step": 39
339
+ },
340
+ {
341
+ "epoch": 0.06575342465753424,
342
+ "grad_norm": 2.2366013526916504,
343
+ "learning_rate": 1.4644660940672627e-05,
344
+ "loss": 0.6836,
345
+ "step": 40
346
+ },
347
+ {
348
+ "epoch": 0.06739726027397261,
349
+ "grad_norm": 1.588487982749939,
350
+ "learning_rate": 1.1979701719998453e-05,
351
+ "loss": 0.6875,
352
+ "step": 41
353
+ },
354
+ {
355
+ "epoch": 0.06904109589041096,
356
+ "grad_norm": 0.7167198061943054,
357
+ "learning_rate": 9.549150281252633e-06,
358
+ "loss": 0.695,
359
+ "step": 42
360
+ },
361
+ {
362
+ "epoch": 0.06904109589041096,
363
+ "eval_loss": 0.6984082460403442,
364
+ "eval_runtime": 401.5322,
365
+ "eval_samples_per_second": 1.915,
366
+ "eval_steps_per_second": 0.481,
367
+ "step": 42
368
+ },
369
+ {
370
+ "epoch": 0.07068493150684932,
371
+ "grad_norm": 1.4269287586212158,
372
+ "learning_rate": 7.367991782295391e-06,
373
+ "loss": 0.6566,
374
+ "step": 43
375
+ },
376
+ {
377
+ "epoch": 0.07232876712328767,
378
+ "grad_norm": 1.8338453769683838,
379
+ "learning_rate": 5.449673790581611e-06,
380
+ "loss": 0.7158,
381
+ "step": 44
382
+ },
383
+ {
384
+ "epoch": 0.07397260273972603,
385
+ "grad_norm": 1.369246244430542,
386
+ "learning_rate": 3.8060233744356633e-06,
387
+ "loss": 0.7047,
388
+ "step": 45
389
+ },
390
+ {
391
+ "epoch": 0.07561643835616438,
392
+ "grad_norm": 0.5873619318008423,
393
+ "learning_rate": 2.4471741852423237e-06,
394
+ "loss": 0.6859,
395
+ "step": 46
396
+ },
397
+ {
398
+ "epoch": 0.07726027397260274,
399
+ "grad_norm": 0.5974377989768982,
400
+ "learning_rate": 1.3815039801161721e-06,
401
+ "loss": 0.7013,
402
+ "step": 47
403
+ },
404
+ {
405
+ "epoch": 0.0789041095890411,
406
+ "grad_norm": 1.1578266620635986,
407
+ "learning_rate": 6.15582970243117e-07,
408
+ "loss": 0.6891,
409
+ "step": 48
410
+ },
411
+ {
412
+ "epoch": 0.0789041095890411,
413
+ "eval_loss": 0.7000875473022461,
414
+ "eval_runtime": 401.4631,
415
+ "eval_samples_per_second": 1.915,
416
+ "eval_steps_per_second": 0.481,
417
+ "step": 48
418
+ },
419
+ {
420
+ "epoch": 0.08054794520547945,
421
+ "grad_norm": 2.2058608531951904,
422
+ "learning_rate": 1.5413331334360182e-07,
423
+ "loss": 0.7191,
424
+ "step": 49
425
+ },
426
+ {
427
+ "epoch": 0.0821917808219178,
428
+ "grad_norm": 1.0864945650100708,
429
+ "learning_rate": 0.0,
430
+ "loss": 0.7048,
431
+ "step": 50
432
  }
433
  ],
434
  "logging_steps": 1,
 
443
  "should_evaluate": false,
444
  "should_log": false,
445
  "should_save": true,
446
+ "should_training_stop": true
447
  },
448
  "attributes": {}
449
  }
450
  },
451
+ "total_flos": 2.238031356493824e+17,
452
  "train_batch_size": 4,
453
  "trial_name": null,
454
  "trial_params": null