dada22231 commited on
Commit
4358469
1 Parent(s): 863c06d

Training in progress, step 50, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c7c606f3ec6092e9f6754a19d19da41c2db88b148858acdb23dd44d550ac9329
3
  size 319876032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6404eb21f64f0fd0495017d0d23caaec61252653ab90a1588bbb3c531463c3c
3
  size 319876032
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5f29de210c45c07b3170ad15d40db2c3310ed36a0a77f6f99511fab8814e569a
3
  size 640009682
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6704fcb92df52ce2a4fd1ee2c475016dae08da655885600db67dc4c7b475ab86
3
  size 640009682
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2edfa45f33fab1e4ef439a0c57573a48a6cce909aaf4e89c07ffc7665c5dbf2f
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c453a58c4bf4312ef0a30c0b43d19fc75e51ae316fb27c948d9633e8580b938
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:590775605cd813196fc85dac96b9894ed963fd457008ecd542374b5377cc4ff2
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e7dff3eab3ccb4a7c3120630c486dd2cff179895632db0ffcb893f2176156c6
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cbb3c72c1591cfb4e4980f314657d27cdfd68f1dd571892a8ea18b5a54757edf
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a827522647025d34e47663465b058aa44ac5e109e57377e97716e2276fb8fe36
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:015c6ec189245107367b9e11c21b45829c2a1895662daaf501436ab85e386106
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dda0c85b48e6d17862dee43bbd5a51b1e593b03910bbd2f556f8212c0007227e
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0f5c305e4a92be904895c02f0f0a1da666e6e7555f6043a8f089990c87f4ce88
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df19ed1a9610a5422497073697cbf4575f80de47fbb46ef0cdd2779386b031fa
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 0.7101312875747681,
3
- "best_model_checkpoint": "miner_id_24/checkpoint-25",
4
- "epoch": 0.03241885156218341,
5
  "eval_steps": 25,
6
- "global_step": 25,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -198,6 +198,189 @@
198
  "eval_samples_per_second": 16.84,
199
  "eval_steps_per_second": 4.378,
200
  "step": 25
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
201
  }
202
  ],
203
  "logging_steps": 1,
@@ -226,7 +409,7 @@
226
  "attributes": {}
227
  }
228
  },
229
- "total_flos": 2.6303741225336832e+17,
230
  "train_batch_size": 1,
231
  "trial_name": null,
232
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.652353048324585,
3
+ "best_model_checkpoint": "miner_id_24/checkpoint-50",
4
+ "epoch": 0.06483770312436682,
5
  "eval_steps": 25,
6
+ "global_step": 50,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
198
  "eval_samples_per_second": 16.84,
199
  "eval_steps_per_second": 4.378,
200
  "step": 25
201
+ },
202
+ {
203
+ "epoch": 0.03371560562467075,
204
+ "grad_norm": 9.792278289794922,
205
+ "learning_rate": 8.681980515339464e-05,
206
+ "loss": 36.8223,
207
+ "step": 26
208
+ },
209
+ {
210
+ "epoch": 0.03501235968715808,
211
+ "grad_norm": 15.209010124206543,
212
+ "learning_rate": 8.571489144483944e-05,
213
+ "loss": 28.8723,
214
+ "step": 27
215
+ },
216
+ {
217
+ "epoch": 0.03630911374964542,
218
+ "grad_norm": 15.312687873840332,
219
+ "learning_rate": 8.457416554680877e-05,
220
+ "loss": 20.8671,
221
+ "step": 28
222
+ },
223
+ {
224
+ "epoch": 0.03760586781213276,
225
+ "grad_norm": 11.13228988647461,
226
+ "learning_rate": 8.339895749467238e-05,
227
+ "loss": 15.874,
228
+ "step": 29
229
+ },
230
+ {
231
+ "epoch": 0.03890262187462009,
232
+ "grad_norm": 12.170778274536133,
233
+ "learning_rate": 8.219063752844926e-05,
234
+ "loss": 9.841,
235
+ "step": 30
236
+ },
237
+ {
238
+ "epoch": 0.04019937593710743,
239
+ "grad_norm": 8.809076309204102,
240
+ "learning_rate": 8.095061449516903e-05,
241
+ "loss": 12.2626,
242
+ "step": 31
243
+ },
244
+ {
245
+ "epoch": 0.04149612999959477,
246
+ "grad_norm": 8.72727108001709,
247
+ "learning_rate": 7.968033420621935e-05,
248
+ "loss": 9.9618,
249
+ "step": 32
250
+ },
251
+ {
252
+ "epoch": 0.0427928840620821,
253
+ "grad_norm": 8.521458625793457,
254
+ "learning_rate": 7.838127775159452e-05,
255
+ "loss": 11.9306,
256
+ "step": 33
257
+ },
258
+ {
259
+ "epoch": 0.04408963812456944,
260
+ "grad_norm": 10.72695255279541,
261
+ "learning_rate": 7.705495977301078e-05,
262
+ "loss": 17.8682,
263
+ "step": 34
264
+ },
265
+ {
266
+ "epoch": 0.04538639218705677,
267
+ "grad_norm": 12.048954963684082,
268
+ "learning_rate": 7.570292669790186e-05,
269
+ "loss": 24.6443,
270
+ "step": 35
271
+ },
272
+ {
273
+ "epoch": 0.04668314624954411,
274
+ "grad_norm": 14.231986045837402,
275
+ "learning_rate": 7.43267549363537e-05,
276
+ "loss": 30.6334,
277
+ "step": 36
278
+ },
279
+ {
280
+ "epoch": 0.04797990031203145,
281
+ "grad_norm": 16.270370483398438,
282
+ "learning_rate": 7.292804904308087e-05,
283
+ "loss": 36.3834,
284
+ "step": 37
285
+ },
286
+ {
287
+ "epoch": 0.04927665437451878,
288
+ "grad_norm": 14.480351448059082,
289
+ "learning_rate": 7.150843984658754e-05,
290
+ "loss": 41.1031,
291
+ "step": 38
292
+ },
293
+ {
294
+ "epoch": 0.05057340843700612,
295
+ "grad_norm": 9.69229507446289,
296
+ "learning_rate": 7.006958254769438e-05,
297
+ "loss": 30.1462,
298
+ "step": 39
299
+ },
300
+ {
301
+ "epoch": 0.05187016249949346,
302
+ "grad_norm": 10.705002784729004,
303
+ "learning_rate": 6.861315478964841e-05,
304
+ "loss": 23.4604,
305
+ "step": 40
306
+ },
307
+ {
308
+ "epoch": 0.05316691656198079,
309
+ "grad_norm": 11.107697486877441,
310
+ "learning_rate": 6.714085470206609e-05,
311
+ "loss": 13.1305,
312
+ "step": 41
313
+ },
314
+ {
315
+ "epoch": 0.05446367062446813,
316
+ "grad_norm": 7.914254665374756,
317
+ "learning_rate": 6.56543989209901e-05,
318
+ "loss": 10.4824,
319
+ "step": 42
320
+ },
321
+ {
322
+ "epoch": 0.05576042468695547,
323
+ "grad_norm": 7.241426467895508,
324
+ "learning_rate": 6.415552058736854e-05,
325
+ "loss": 12.7372,
326
+ "step": 43
327
+ },
328
+ {
329
+ "epoch": 0.0570571787494428,
330
+ "grad_norm": 7.243592262268066,
331
+ "learning_rate": 6.264596732629e-05,
332
+ "loss": 8.7167,
333
+ "step": 44
334
+ },
335
+ {
336
+ "epoch": 0.05835393281193014,
337
+ "grad_norm": 8.968466758728027,
338
+ "learning_rate": 6.112749920933111e-05,
339
+ "loss": 11.5413,
340
+ "step": 45
341
+ },
342
+ {
343
+ "epoch": 0.05965068687441747,
344
+ "grad_norm": 10.867255210876465,
345
+ "learning_rate": 5.960188670239154e-05,
346
+ "loss": 14.4485,
347
+ "step": 46
348
+ },
349
+ {
350
+ "epoch": 0.06094744093690481,
351
+ "grad_norm": 10.397505760192871,
352
+ "learning_rate": 5.80709086014102e-05,
353
+ "loss": 14.4408,
354
+ "step": 47
355
+ },
356
+ {
357
+ "epoch": 0.06224419499939215,
358
+ "grad_norm": 11.32922077178955,
359
+ "learning_rate": 5.653634995836856e-05,
360
+ "loss": 21.16,
361
+ "step": 48
362
+ },
363
+ {
364
+ "epoch": 0.06354094906187949,
365
+ "grad_norm": 15.780669212341309,
366
+ "learning_rate": 5.500000000000001e-05,
367
+ "loss": 29.0683,
368
+ "step": 49
369
+ },
370
+ {
371
+ "epoch": 0.06483770312436682,
372
+ "grad_norm": 21.16714096069336,
373
+ "learning_rate": 5.346365004163145e-05,
374
+ "loss": 35.8908,
375
+ "step": 50
376
+ },
377
+ {
378
+ "epoch": 0.06483770312436682,
379
+ "eval_loss": 0.652353048324585,
380
+ "eval_runtime": 2.9375,
381
+ "eval_samples_per_second": 17.021,
382
+ "eval_steps_per_second": 4.426,
383
+ "step": 50
384
  }
385
  ],
386
  "logging_steps": 1,
 
409
  "attributes": {}
410
  }
411
  },
412
+ "total_flos": 5.259926510449459e+17,
413
  "train_batch_size": 1,
414
  "trial_name": null,
415
  "trial_params": null