valadapt-llama-3.1-8B-it-greek / trainer_state.json
faridlazuarda's picture
Replace with latest checkpoint files
1e0a75f verified
raw
history blame
11.3 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 6.0,
"eval_steps": 500,
"global_step": 16506,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.09087604507451835,
"grad_norm": 1.2130728960037231,
"learning_rate": 0.0002,
"loss": 0.9008,
"step": 250
},
{
"epoch": 0.1817520901490367,
"grad_norm": 1.1032902002334595,
"learning_rate": 0.0002,
"loss": 0.5317,
"step": 500
},
{
"epoch": 0.27262813522355506,
"grad_norm": 1.0168739557266235,
"learning_rate": 0.0002,
"loss": 0.4231,
"step": 750
},
{
"epoch": 0.3635041802980734,
"grad_norm": 1.888968825340271,
"learning_rate": 0.0002,
"loss": 0.3506,
"step": 1000
},
{
"epoch": 0.45438022537259176,
"grad_norm": 0.7193501591682434,
"learning_rate": 0.0002,
"loss": 0.3049,
"step": 1250
},
{
"epoch": 0.5452562704471101,
"grad_norm": 0.958330512046814,
"learning_rate": 0.0002,
"loss": 0.2575,
"step": 1500
},
{
"epoch": 0.6361323155216285,
"grad_norm": 0.9223587512969971,
"learning_rate": 0.0002,
"loss": 0.2395,
"step": 1750
},
{
"epoch": 0.7270083605961468,
"grad_norm": 0.639679491519928,
"learning_rate": 0.0002,
"loss": 0.2213,
"step": 2000
},
{
"epoch": 0.8178844056706652,
"grad_norm": 0.5571266412734985,
"learning_rate": 0.0002,
"loss": 0.2057,
"step": 2250
},
{
"epoch": 0.9087604507451835,
"grad_norm": 0.5572911500930786,
"learning_rate": 0.0002,
"loss": 0.1926,
"step": 2500
},
{
"epoch": 0.9996364958197019,
"grad_norm": 0.7683030366897583,
"learning_rate": 0.0002,
"loss": 0.1848,
"step": 2750
},
{
"epoch": 1.0905125408942202,
"grad_norm": 0.6649575233459473,
"learning_rate": 0.0002,
"loss": 0.1757,
"step": 3000
},
{
"epoch": 1.1813885859687385,
"grad_norm": 1.3259624242782593,
"learning_rate": 0.0002,
"loss": 0.1724,
"step": 3250
},
{
"epoch": 1.272264631043257,
"grad_norm": 0.5268917083740234,
"learning_rate": 0.0002,
"loss": 0.1696,
"step": 3500
},
{
"epoch": 1.3631406761177753,
"grad_norm": 0.7535058856010437,
"learning_rate": 0.0002,
"loss": 0.1676,
"step": 3750
},
{
"epoch": 1.4540167211922936,
"grad_norm": 0.3919288218021393,
"learning_rate": 0.0002,
"loss": 0.1615,
"step": 4000
},
{
"epoch": 1.5448927662668122,
"grad_norm": 0.2522818446159363,
"learning_rate": 0.0002,
"loss": 0.1596,
"step": 4250
},
{
"epoch": 1.6357688113413305,
"grad_norm": 0.574805498123169,
"learning_rate": 0.0002,
"loss": 0.1579,
"step": 4500
},
{
"epoch": 1.7266448564158487,
"grad_norm": 0.35280880331993103,
"learning_rate": 0.0002,
"loss": 0.1572,
"step": 4750
},
{
"epoch": 1.8175209014903673,
"grad_norm": 0.343710333108902,
"learning_rate": 0.0002,
"loss": 0.1529,
"step": 5000
},
{
"epoch": 1.9083969465648853,
"grad_norm": 0.6308000683784485,
"learning_rate": 0.0002,
"loss": 0.1514,
"step": 5250
},
{
"epoch": 1.9992729916394039,
"grad_norm": 0.7428322434425354,
"learning_rate": 0.0002,
"loss": 0.1504,
"step": 5500
},
{
"epoch": 2.0901490367139224,
"grad_norm": 0.816819965839386,
"learning_rate": 0.0002,
"loss": 0.1455,
"step": 5750
},
{
"epoch": 2.1810250817884405,
"grad_norm": 0.31729263067245483,
"learning_rate": 0.0002,
"loss": 0.1467,
"step": 6000
},
{
"epoch": 2.271901126862959,
"grad_norm": 0.3655836582183838,
"learning_rate": 0.0002,
"loss": 0.1447,
"step": 6250
},
{
"epoch": 2.362777171937477,
"grad_norm": 0.34890076518058777,
"learning_rate": 0.0002,
"loss": 0.1449,
"step": 6500
},
{
"epoch": 2.4536532170119956,
"grad_norm": 0.25036028027534485,
"learning_rate": 0.0002,
"loss": 0.1446,
"step": 6750
},
{
"epoch": 2.544529262086514,
"grad_norm": 0.37913835048675537,
"learning_rate": 0.0002,
"loss": 0.1429,
"step": 7000
},
{
"epoch": 2.635405307161032,
"grad_norm": 0.6284835934638977,
"learning_rate": 0.0002,
"loss": 0.1427,
"step": 7250
},
{
"epoch": 2.7262813522355507,
"grad_norm": 0.18296529352664948,
"learning_rate": 0.0002,
"loss": 0.1425,
"step": 7500
},
{
"epoch": 2.817157397310069,
"grad_norm": 0.31024783849716187,
"learning_rate": 0.0002,
"loss": 0.1405,
"step": 7750
},
{
"epoch": 2.9080334423845873,
"grad_norm": 0.2511535584926605,
"learning_rate": 0.0002,
"loss": 0.1393,
"step": 8000
},
{
"epoch": 2.998909487459106,
"grad_norm": 0.2819952070713043,
"learning_rate": 0.0002,
"loss": 0.1394,
"step": 8250
},
{
"epoch": 3.0897855325336243,
"grad_norm": 0.2572026252746582,
"learning_rate": 0.0002,
"loss": 0.1352,
"step": 8500
},
{
"epoch": 3.1806615776081424,
"grad_norm": 0.6162417531013489,
"learning_rate": 0.0002,
"loss": 0.1367,
"step": 8750
},
{
"epoch": 3.271537622682661,
"grad_norm": 0.6815608143806458,
"learning_rate": 0.0002,
"loss": 0.136,
"step": 9000
},
{
"epoch": 3.3624136677571794,
"grad_norm": 0.4514828324317932,
"learning_rate": 0.0002,
"loss": 0.1387,
"step": 9250
},
{
"epoch": 3.4532897128316975,
"grad_norm": 0.14196833968162537,
"learning_rate": 0.0002,
"loss": 0.1354,
"step": 9500
},
{
"epoch": 3.544165757906216,
"grad_norm": 0.2632617950439453,
"learning_rate": 0.0002,
"loss": 0.1358,
"step": 9750
},
{
"epoch": 3.6350418029807345,
"grad_norm": 0.15455371141433716,
"learning_rate": 0.0002,
"loss": 0.1363,
"step": 10000
},
{
"epoch": 3.7259178480552526,
"grad_norm": 0.14311546087265015,
"learning_rate": 0.0002,
"loss": 0.136,
"step": 10250
},
{
"epoch": 3.816793893129771,
"grad_norm": 0.21806533634662628,
"learning_rate": 0.0002,
"loss": 0.1336,
"step": 10500
},
{
"epoch": 3.907669938204289,
"grad_norm": 0.6075801253318787,
"learning_rate": 0.0002,
"loss": 0.1339,
"step": 10750
},
{
"epoch": 3.9985459832788077,
"grad_norm": 0.38970544934272766,
"learning_rate": 0.0002,
"loss": 0.134,
"step": 11000
},
{
"epoch": 4.089422028353326,
"grad_norm": 0.12462722510099411,
"learning_rate": 0.0002,
"loss": 0.1341,
"step": 11250
},
{
"epoch": 4.180298073427845,
"grad_norm": 0.12478788197040558,
"learning_rate": 0.0002,
"loss": 0.1311,
"step": 11500
},
{
"epoch": 4.271174118502363,
"grad_norm": 0.17592966556549072,
"learning_rate": 0.0002,
"loss": 0.1334,
"step": 11750
},
{
"epoch": 4.362050163576881,
"grad_norm": 0.16919384896755219,
"learning_rate": 0.0002,
"loss": 0.1315,
"step": 12000
},
{
"epoch": 4.4529262086514,
"grad_norm": 0.232040673494339,
"learning_rate": 0.0002,
"loss": 0.1324,
"step": 12250
},
{
"epoch": 4.543802253725918,
"grad_norm": 0.14165472984313965,
"learning_rate": 0.0002,
"loss": 0.1336,
"step": 12500
},
{
"epoch": 4.634678298800436,
"grad_norm": 0.17017216980457306,
"learning_rate": 0.0002,
"loss": 0.1336,
"step": 12750
},
{
"epoch": 4.725554343874954,
"grad_norm": 0.15150970220565796,
"learning_rate": 0.0002,
"loss": 0.1316,
"step": 13000
},
{
"epoch": 4.816430388949473,
"grad_norm": 0.7836107611656189,
"learning_rate": 0.0002,
"loss": 0.1328,
"step": 13250
},
{
"epoch": 4.907306434023991,
"grad_norm": 0.28059402108192444,
"learning_rate": 0.0002,
"loss": 0.1323,
"step": 13500
},
{
"epoch": 4.99818247909851,
"grad_norm": 0.5553152561187744,
"learning_rate": 0.0002,
"loss": 0.1317,
"step": 13750
},
{
"epoch": 5.089058524173028,
"grad_norm": 0.18768849968910217,
"learning_rate": 0.0002,
"loss": 0.1307,
"step": 14000
},
{
"epoch": 5.179934569247546,
"grad_norm": 0.12853963673114777,
"learning_rate": 0.0002,
"loss": 0.1305,
"step": 14250
},
{
"epoch": 5.270810614322064,
"grad_norm": 0.13940949738025665,
"learning_rate": 0.0002,
"loss": 0.1314,
"step": 14500
},
{
"epoch": 5.361686659396583,
"grad_norm": 0.12992176413536072,
"learning_rate": 0.0002,
"loss": 0.1291,
"step": 14750
},
{
"epoch": 5.452562704471101,
"grad_norm": 0.17497240006923676,
"learning_rate": 0.0002,
"loss": 0.1309,
"step": 15000
},
{
"epoch": 5.543438749545619,
"grad_norm": 0.1531759798526764,
"learning_rate": 0.0002,
"loss": 0.1303,
"step": 15250
},
{
"epoch": 5.634314794620138,
"grad_norm": 0.1523718535900116,
"learning_rate": 0.0002,
"loss": 0.1295,
"step": 15500
},
{
"epoch": 5.7251908396946565,
"grad_norm": 0.17395178973674774,
"learning_rate": 0.0002,
"loss": 0.1313,
"step": 15750
},
{
"epoch": 5.8160668847691745,
"grad_norm": 0.10413607209920883,
"learning_rate": 0.0002,
"loss": 0.1328,
"step": 16000
},
{
"epoch": 5.9069429298436935,
"grad_norm": 0.24993538856506348,
"learning_rate": 0.0002,
"loss": 0.1297,
"step": 16250
},
{
"epoch": 5.997818974918212,
"grad_norm": 0.30852416157722473,
"learning_rate": 0.0002,
"loss": 0.13,
"step": 16500
}
],
"logging_steps": 250,
"max_steps": 16506,
"num_input_tokens_seen": 0,
"num_train_epochs": 6,
"save_steps": 250,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2.098309604079698e+17,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}