{ "best_metric": 0.9108405113220215, "best_model_checkpoint": "miner_id_24/checkpoint-50", "epoch": 0.7188885604493054, "eval_steps": 25, "global_step": 76, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.009459060005911913, "grad_norm": 0.9699360728263855, "learning_rate": 0.00015, "loss": 1.3474, "step": 1 }, { "epoch": 0.009459060005911913, "eval_loss": 1.4942742586135864, "eval_runtime": 11.3224, "eval_samples_per_second": 4.416, "eval_steps_per_second": 4.416, "step": 1 }, { "epoch": 0.018918120011823827, "grad_norm": 0.9874446392059326, "learning_rate": 0.0003, "loss": 1.4419, "step": 2 }, { "epoch": 0.02837718001773574, "grad_norm": 0.8229495286941528, "learning_rate": 0.000299878360437632, "loss": 1.2491, "step": 3 }, { "epoch": 0.03783624002364765, "grad_norm": 0.6731188297271729, "learning_rate": 0.00029951366095324104, "loss": 1.0968, "step": 4 }, { "epoch": 0.04729530002955956, "grad_norm": 0.7858128547668457, "learning_rate": 0.00029890655875994835, "loss": 1.0792, "step": 5 }, { "epoch": 0.05675436003547148, "grad_norm": 0.7995796799659729, "learning_rate": 0.0002980581478969406, "loss": 1.1033, "step": 6 }, { "epoch": 0.06621342004138339, "grad_norm": 0.616423487663269, "learning_rate": 0.00029696995725793764, "loss": 0.942, "step": 7 }, { "epoch": 0.0756724800472953, "grad_norm": 0.8099045753479004, "learning_rate": 0.00029564394783602234, "loss": 1.0823, "step": 8 }, { "epoch": 0.08513154005320721, "grad_norm": 0.7018675208091736, "learning_rate": 0.0002940825091897988, "loss": 1.1322, "step": 9 }, { "epoch": 0.09459060005911912, "grad_norm": 0.6773211359977722, "learning_rate": 0.00029228845513724634, "loss": 1.1547, "step": 10 }, { "epoch": 0.10404966006503104, "grad_norm": 0.6189546585083008, "learning_rate": 0.00029026501868502873, "loss": 1.0361, "step": 11 }, { "epoch": 0.11350872007094295, "grad_norm": 0.623130202293396, "learning_rate": 0.0002880158462023983, "loss": 1.0837, "step": 12 }, { "epoch": 0.12296778007685487, "grad_norm": 0.55083829164505, "learning_rate": 0.0002855449908501917, "loss": 0.8752, "step": 13 }, { "epoch": 0.13242684008276678, "grad_norm": 0.5632800459861755, "learning_rate": 0.00028285690527676035, "loss": 1.0141, "step": 14 }, { "epoch": 0.14188590008867868, "grad_norm": 0.5698175430297852, "learning_rate": 0.000279956433593997, "loss": 0.9583, "step": 15 }, { "epoch": 0.1513449600945906, "grad_norm": 0.6257957220077515, "learning_rate": 0.00027684880264791867, "loss": 1.1261, "step": 16 }, { "epoch": 0.16080402010050251, "grad_norm": 0.6792426109313965, "learning_rate": 0.00027353961259953696, "loss": 1.076, "step": 17 }, { "epoch": 0.17026308010641442, "grad_norm": 0.5756276249885559, "learning_rate": 0.00027003482683298933, "loss": 0.9779, "step": 18 }, { "epoch": 0.17972214011232635, "grad_norm": 0.6053583025932312, "learning_rate": 0.00026634076120911777, "loss": 1.168, "step": 19 }, { "epoch": 0.18918120011823825, "grad_norm": 0.5745776891708374, "learning_rate": 0.0002624640726838608, "loss": 0.997, "step": 20 }, { "epoch": 0.19864026012415018, "grad_norm": 0.5367314219474792, "learning_rate": 0.00025841174731196877, "loss": 0.9194, "step": 21 }, { "epoch": 0.20809932013006208, "grad_norm": 0.5325618386268616, "learning_rate": 0.000254191087657661, "loss": 1.0697, "step": 22 }, { "epoch": 0.21755838013597398, "grad_norm": 0.6518679857254028, "learning_rate": 0.0002498096996349117, "loss": 1.1852, "step": 23 }, { "epoch": 0.2270174401418859, "grad_norm": 0.6570560336112976, "learning_rate": 0.0002452754788010787, "loss": 1.1949, "step": 24 }, { "epoch": 0.2364765001477978, "grad_norm": 0.6248823404312134, "learning_rate": 0.00024059659612857536, "loss": 1.1139, "step": 25 }, { "epoch": 0.2364765001477978, "eval_loss": 0.9684370160102844, "eval_runtime": 11.3298, "eval_samples_per_second": 4.413, "eval_steps_per_second": 4.413, "step": 25 }, { "epoch": 0.24593556015370974, "grad_norm": 0.6149452924728394, "learning_rate": 0.00023578148328022626, "loss": 1.2, "step": 26 }, { "epoch": 0.2553946201596216, "grad_norm": 0.5431950688362122, "learning_rate": 0.00023083881741484068, "loss": 1.1115, "step": 27 }, { "epoch": 0.26485368016553357, "grad_norm": 0.5499697923660278, "learning_rate": 0.00022577750555038587, "loss": 0.8515, "step": 28 }, { "epoch": 0.27431274017144547, "grad_norm": 0.45134755969047546, "learning_rate": 0.000220606668512939, "loss": 0.8734, "step": 29 }, { "epoch": 0.28377180017735737, "grad_norm": 0.539884626865387, "learning_rate": 0.00021533562450034164, "loss": 0.9292, "step": 30 }, { "epoch": 0.29323086018326927, "grad_norm": 0.4615366458892822, "learning_rate": 0.00020997387229017774, "loss": 0.8075, "step": 31 }, { "epoch": 0.3026899201891812, "grad_norm": 0.4927918612957001, "learning_rate": 0.00020453107412233428, "loss": 0.8667, "step": 32 }, { "epoch": 0.31214898019509313, "grad_norm": 0.49951159954071045, "learning_rate": 0.0001990170382869919, "loss": 0.977, "step": 33 }, { "epoch": 0.32160804020100503, "grad_norm": 0.4911053478717804, "learning_rate": 0.00019344170144942302, "loss": 0.8319, "step": 34 }, { "epoch": 0.33106710020691693, "grad_norm": 0.5060461163520813, "learning_rate": 0.00018781511074344962, "loss": 0.9303, "step": 35 }, { "epoch": 0.34052616021282883, "grad_norm": 0.554681658744812, "learning_rate": 0.0001821474056658286, "loss": 0.9563, "step": 36 }, { "epoch": 0.3499852202187408, "grad_norm": 0.5921431183815002, "learning_rate": 0.00017644879980419374, "loss": 1.0516, "step": 37 }, { "epoch": 0.3594442802246527, "grad_norm": 0.4988238215446472, "learning_rate": 0.00017072956243148002, "loss": 0.9642, "step": 38 }, { "epoch": 0.3689033402305646, "grad_norm": 0.4554699659347534, "learning_rate": 0.000165, "loss": 0.8114, "step": 39 }, { "epoch": 0.3783624002364765, "grad_norm": 0.4806513488292694, "learning_rate": 0.00015927043756852, "loss": 1.0024, "step": 40 }, { "epoch": 0.3878214602423884, "grad_norm": 0.5541513562202454, "learning_rate": 0.0001535512001958063, "loss": 1.0891, "step": 41 }, { "epoch": 0.39728052024830035, "grad_norm": 0.4677070081233978, "learning_rate": 0.00014785259433417133, "loss": 0.8085, "step": 42 }, { "epoch": 0.40673958025421225, "grad_norm": 0.4648246467113495, "learning_rate": 0.00014218488925655037, "loss": 0.8908, "step": 43 }, { "epoch": 0.41619864026012415, "grad_norm": 0.5301900506019592, "learning_rate": 0.00013655829855057698, "loss": 0.9752, "step": 44 }, { "epoch": 0.42565770026603605, "grad_norm": 0.5299180746078491, "learning_rate": 0.00013098296171300814, "loss": 1.0093, "step": 45 }, { "epoch": 0.43511676027194796, "grad_norm": 0.5281239151954651, "learning_rate": 0.0001254689258776657, "loss": 0.9709, "step": 46 }, { "epoch": 0.4445758202778599, "grad_norm": 0.5601954460144043, "learning_rate": 0.00012002612770982222, "loss": 1.0698, "step": 47 }, { "epoch": 0.4540348802837718, "grad_norm": 0.5613210201263428, "learning_rate": 0.00011466437549965834, "loss": 1.0451, "step": 48 }, { "epoch": 0.4634939402896837, "grad_norm": 0.6079455018043518, "learning_rate": 0.00010939333148706099, "loss": 1.0703, "step": 49 }, { "epoch": 0.4729530002955956, "grad_norm": 0.6424407362937927, "learning_rate": 0.00010422249444961407, "loss": 1.2829, "step": 50 }, { "epoch": 0.4729530002955956, "eval_loss": 0.9108405113220215, "eval_runtime": 11.325, "eval_samples_per_second": 4.415, "eval_steps_per_second": 4.415, "step": 50 }, { "epoch": 0.4824120603015075, "grad_norm": 0.629723846912384, "learning_rate": 9.916118258515936e-05, "loss": 1.1712, "step": 51 }, { "epoch": 0.4918711203074195, "grad_norm": 0.7443132996559143, "learning_rate": 9.421851671977372e-05, "loss": 1.1707, "step": 52 }, { "epoch": 0.5013301803133313, "grad_norm": 0.46892011165618896, "learning_rate": 8.940340387142462e-05, "loss": 0.8093, "step": 53 }, { "epoch": 0.5107892403192432, "grad_norm": 0.43437010049819946, "learning_rate": 8.47245211989213e-05, "loss": 0.8652, "step": 54 }, { "epoch": 0.5202483003251552, "grad_norm": 0.3998465836048126, "learning_rate": 8.019030036508827e-05, "loss": 0.8075, "step": 55 }, { "epoch": 0.5297073603310671, "grad_norm": 0.4319959282875061, "learning_rate": 7.580891234233904e-05, "loss": 0.8248, "step": 56 }, { "epoch": 0.539166420336979, "grad_norm": 0.4736616015434265, "learning_rate": 7.158825268803127e-05, "loss": 0.9148, "step": 57 }, { "epoch": 0.5486254803428909, "grad_norm": 0.3903331458568573, "learning_rate": 6.75359273161392e-05, "loss": 0.8549, "step": 58 }, { "epoch": 0.5580845403488028, "grad_norm": 0.428145170211792, "learning_rate": 6.365923879088219e-05, "loss": 0.8272, "step": 59 }, { "epoch": 0.5675436003547147, "grad_norm": 0.5180810689926147, "learning_rate": 5.996517316701069e-05, "loss": 0.9229, "step": 60 }, { "epoch": 0.5770026603606266, "grad_norm": 0.47326382994651794, "learning_rate": 5.646038740046304e-05, "loss": 0.768, "step": 61 }, { "epoch": 0.5864617203665385, "grad_norm": 0.47181880474090576, "learning_rate": 5.315119735208132e-05, "loss": 0.973, "step": 62 }, { "epoch": 0.5959207803724504, "grad_norm": 0.48134690523147583, "learning_rate": 5.004356640600297e-05, "loss": 0.8445, "step": 63 }, { "epoch": 0.6053798403783625, "grad_norm": 0.4601708650588989, "learning_rate": 4.71430947232396e-05, "loss": 0.898, "step": 64 }, { "epoch": 0.6148389003842744, "grad_norm": 0.4865597188472748, "learning_rate": 4.4455009149808265e-05, "loss": 0.922, "step": 65 }, { "epoch": 0.6242979603901863, "grad_norm": 0.47632795572280884, "learning_rate": 4.1984153797601665e-05, "loss": 0.936, "step": 66 }, { "epoch": 0.6337570203960982, "grad_norm": 0.5297707319259644, "learning_rate": 3.9734981314971234e-05, "loss": 0.9437, "step": 67 }, { "epoch": 0.6432160804020101, "grad_norm": 0.46655818819999695, "learning_rate": 3.771154486275363e-05, "loss": 0.7435, "step": 68 }, { "epoch": 0.652675140407922, "grad_norm": 0.513863205909729, "learning_rate": 3.591749081020113e-05, "loss": 0.9687, "step": 69 }, { "epoch": 0.6621342004138339, "grad_norm": 0.5340255498886108, "learning_rate": 3.435605216397765e-05, "loss": 1.0353, "step": 70 }, { "epoch": 0.6715932604197458, "grad_norm": 0.45835715532302856, "learning_rate": 3.303004274206237e-05, "loss": 0.864, "step": 71 }, { "epoch": 0.6810523204256577, "grad_norm": 0.4819653034210205, "learning_rate": 3.194185210305936e-05, "loss": 1.0022, "step": 72 }, { "epoch": 0.6905113804315696, "grad_norm": 0.5713886618614197, "learning_rate": 3.1093441240051626e-05, "loss": 1.0681, "step": 73 }, { "epoch": 0.6999704404374816, "grad_norm": 0.5911141633987427, "learning_rate": 3.048633904675892e-05, "loss": 0.9129, "step": 74 }, { "epoch": 0.7094295004433935, "grad_norm": 0.5698217749595642, "learning_rate": 3.012163956236801e-05, "loss": 1.0666, "step": 75 }, { "epoch": 0.7094295004433935, "eval_loss": 0.8948501348495483, "eval_runtime": 11.3345, "eval_samples_per_second": 4.411, "eval_steps_per_second": 4.411, "step": 75 }, { "epoch": 0.7188885604493054, "grad_norm": 0.5917747616767883, "learning_rate": 2.9999999999999997e-05, "loss": 1.0924, "step": 76 } ], "logging_steps": 1, "max_steps": 76, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 50, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 1, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 0 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.002568490422108e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }