{ "best_metric": 0.925627350509402, "best_model_checkpoint": "checkpoints/BEE-spoke-data-bert-plus-L8-v1.0-allNLI_matryoshka-synthetic-text-similarity-Mar-07_22-56/checkpoint-3600", "epoch": 0.35376884422110555, "eval_steps": 300, "global_step": 4400, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 7.8406219482421875, "learning_rate": 8.038585209003216e-08, "loss": 0.1372, "step": 5 }, { "epoch": 0.0, "grad_norm": 6.115014553070068, "learning_rate": 1.6077170418006432e-07, "loss": 0.1134, "step": 10 }, { "epoch": 0.0, "grad_norm": 2.3239176273345947, "learning_rate": 2.4115755627009647e-07, "loss": 0.0902, "step": 15 }, { "epoch": 0.0, "grad_norm": 4.643126964569092, "learning_rate": 3.2154340836012864e-07, "loss": 0.0907, "step": 20 }, { "epoch": 0.0, "grad_norm": 2.8138389587402344, "learning_rate": 4.019292604501608e-07, "loss": 0.087, "step": 25 }, { "epoch": 0.0, "grad_norm": 2.0662753582000732, "learning_rate": 4.823151125401929e-07, "loss": 0.0713, "step": 30 }, { "epoch": 0.0, "grad_norm": 2.2391765117645264, "learning_rate": 5.627009646302252e-07, "loss": 0.0724, "step": 35 }, { "epoch": 0.0, "grad_norm": 2.7336483001708984, "learning_rate": 6.430868167202573e-07, "loss": 0.0711, "step": 40 }, { "epoch": 0.0, "grad_norm": 0.5051072239875793, "learning_rate": 7.234726688102894e-07, "loss": 0.0386, "step": 45 }, { "epoch": 0.0, "grad_norm": 1.8692330121994019, "learning_rate": 8.038585209003216e-07, "loss": 0.07, "step": 50 }, { "epoch": 0.0, "grad_norm": 2.106194257736206, "learning_rate": 8.842443729903538e-07, "loss": 0.0908, "step": 55 }, { "epoch": 0.0, "grad_norm": 2.4909865856170654, "learning_rate": 9.646302250803859e-07, "loss": 0.0632, "step": 60 }, { "epoch": 0.01, "grad_norm": 1.8705946207046509, "learning_rate": 1.045016077170418e-06, "loss": 0.0436, "step": 65 }, { "epoch": 0.01, "grad_norm": 1.9496204853057861, "learning_rate": 1.1254019292604503e-06, "loss": 0.0481, "step": 70 }, { "epoch": 0.01, "grad_norm": 2.2997841835021973, "learning_rate": 1.2057877813504825e-06, "loss": 0.0552, "step": 75 }, { "epoch": 0.01, "grad_norm": 2.0006277561187744, "learning_rate": 1.2861736334405146e-06, "loss": 0.0249, "step": 80 }, { "epoch": 0.01, "grad_norm": 1.1965364217758179, "learning_rate": 1.3665594855305468e-06, "loss": 0.0261, "step": 85 }, { "epoch": 0.01, "grad_norm": 0.8660438656806946, "learning_rate": 1.4469453376205788e-06, "loss": 0.0383, "step": 90 }, { "epoch": 0.01, "grad_norm": 0.8808113932609558, "learning_rate": 1.527331189710611e-06, "loss": 0.0254, "step": 95 }, { "epoch": 0.01, "grad_norm": 1.918907880783081, "learning_rate": 1.6077170418006432e-06, "loss": 0.0656, "step": 100 }, { "epoch": 0.01, "grad_norm": 2.735292673110962, "learning_rate": 1.6881028938906755e-06, "loss": 0.0518, "step": 105 }, { "epoch": 0.01, "grad_norm": 3.8118743896484375, "learning_rate": 1.7684887459807077e-06, "loss": 0.0908, "step": 110 }, { "epoch": 0.01, "grad_norm": 2.0685555934906006, "learning_rate": 1.84887459807074e-06, "loss": 0.0332, "step": 115 }, { "epoch": 0.01, "grad_norm": 0.8618605732917786, "learning_rate": 1.9292604501607717e-06, "loss": 0.0456, "step": 120 }, { "epoch": 0.01, "grad_norm": 1.7448190450668335, "learning_rate": 2.0096463022508037e-06, "loss": 0.0349, "step": 125 }, { "epoch": 0.01, "grad_norm": 1.6762810945510864, "learning_rate": 2.090032154340836e-06, "loss": 0.025, "step": 130 }, { "epoch": 0.01, "grad_norm": 2.172909736633301, "learning_rate": 2.170418006430868e-06, "loss": 0.0334, "step": 135 }, { "epoch": 0.01, "grad_norm": 5.036996364593506, "learning_rate": 2.2508038585209006e-06, "loss": 0.0548, "step": 140 }, { "epoch": 0.01, "grad_norm": 2.498236894607544, "learning_rate": 2.3311897106109326e-06, "loss": 0.0492, "step": 145 }, { "epoch": 0.01, "grad_norm": 2.8583147525787354, "learning_rate": 2.411575562700965e-06, "loss": 0.05, "step": 150 }, { "epoch": 0.01, "grad_norm": 1.241025686264038, "learning_rate": 2.491961414790997e-06, "loss": 0.0264, "step": 155 }, { "epoch": 0.01, "grad_norm": 2.9725112915039062, "learning_rate": 2.572347266881029e-06, "loss": 0.0321, "step": 160 }, { "epoch": 0.01, "grad_norm": 0.9365035891532898, "learning_rate": 2.6527331189710615e-06, "loss": 0.0149, "step": 165 }, { "epoch": 0.01, "grad_norm": 0.9742757081985474, "learning_rate": 2.7331189710610936e-06, "loss": 0.0186, "step": 170 }, { "epoch": 0.01, "grad_norm": 0.9915574193000793, "learning_rate": 2.813504823151126e-06, "loss": 0.0308, "step": 175 }, { "epoch": 0.01, "grad_norm": 3.514364719390869, "learning_rate": 2.8938906752411576e-06, "loss": 0.0288, "step": 180 }, { "epoch": 0.01, "grad_norm": 2.717421293258667, "learning_rate": 2.97427652733119e-06, "loss": 0.0371, "step": 185 }, { "epoch": 0.02, "grad_norm": 4.118988990783691, "learning_rate": 3.054662379421222e-06, "loss": 0.0494, "step": 190 }, { "epoch": 0.02, "grad_norm": 2.1433117389678955, "learning_rate": 3.135048231511254e-06, "loss": 0.0515, "step": 195 }, { "epoch": 0.02, "grad_norm": 0.93222975730896, "learning_rate": 3.2154340836012865e-06, "loss": 0.0276, "step": 200 }, { "epoch": 0.02, "grad_norm": 0.6795055270195007, "learning_rate": 3.2958199356913185e-06, "loss": 0.0318, "step": 205 }, { "epoch": 0.02, "grad_norm": 2.950875997543335, "learning_rate": 3.376205787781351e-06, "loss": 0.0335, "step": 210 }, { "epoch": 0.02, "grad_norm": 2.8429243564605713, "learning_rate": 3.456591639871383e-06, "loss": 0.0434, "step": 215 }, { "epoch": 0.02, "grad_norm": 2.2567358016967773, "learning_rate": 3.5369774919614154e-06, "loss": 0.0324, "step": 220 }, { "epoch": 0.02, "grad_norm": 1.407845139503479, "learning_rate": 3.617363344051447e-06, "loss": 0.0309, "step": 225 }, { "epoch": 0.02, "grad_norm": 3.4068539142608643, "learning_rate": 3.69774919614148e-06, "loss": 0.0454, "step": 230 }, { "epoch": 0.02, "grad_norm": 1.8196914196014404, "learning_rate": 3.7781350482315114e-06, "loss": 0.0255, "step": 235 }, { "epoch": 0.02, "grad_norm": 1.0097465515136719, "learning_rate": 3.8585209003215434e-06, "loss": 0.0263, "step": 240 }, { "epoch": 0.02, "grad_norm": 1.995619773864746, "learning_rate": 3.938906752411576e-06, "loss": 0.0331, "step": 245 }, { "epoch": 0.02, "grad_norm": 2.168858528137207, "learning_rate": 4.0192926045016075e-06, "loss": 0.0416, "step": 250 }, { "epoch": 0.02, "grad_norm": 2.8158152103424072, "learning_rate": 4.09967845659164e-06, "loss": 0.0385, "step": 255 }, { "epoch": 0.02, "grad_norm": 1.8158726692199707, "learning_rate": 4.180064308681672e-06, "loss": 0.0385, "step": 260 }, { "epoch": 0.02, "grad_norm": 0.6891674995422363, "learning_rate": 4.260450160771704e-06, "loss": 0.0182, "step": 265 }, { "epoch": 0.02, "grad_norm": 1.5722671747207642, "learning_rate": 4.340836012861736e-06, "loss": 0.0323, "step": 270 }, { "epoch": 0.02, "grad_norm": 2.8213624954223633, "learning_rate": 4.421221864951769e-06, "loss": 0.0351, "step": 275 }, { "epoch": 0.02, "grad_norm": 1.2886526584625244, "learning_rate": 4.501607717041801e-06, "loss": 0.0401, "step": 280 }, { "epoch": 0.02, "grad_norm": 2.6343109607696533, "learning_rate": 4.581993569131833e-06, "loss": 0.0343, "step": 285 }, { "epoch": 0.02, "grad_norm": 0.7094342708587646, "learning_rate": 4.662379421221865e-06, "loss": 0.0261, "step": 290 }, { "epoch": 0.02, "grad_norm": 2.2275326251983643, "learning_rate": 4.742765273311897e-06, "loss": 0.0183, "step": 295 }, { "epoch": 0.02, "grad_norm": 0.9329387545585632, "learning_rate": 4.82315112540193e-06, "loss": 0.0651, "step": 300 }, { "epoch": 0.02, "eval_loss": 0.03527999296784401, "eval_pearson_cosine": 0.8237974983947806, "eval_pearson_dot": 0.79982187310788, "eval_pearson_euclidean": 0.8049934987510711, "eval_pearson_manhattan": 0.802979519881207, "eval_pearson_max": 0.8237974983947806, "eval_runtime": 426.4216, "eval_samples_per_second": 1.173, "eval_spearman_cosine": 0.8749033396133584, "eval_spearman_dot": 0.8475867183468735, "eval_spearman_euclidean": 0.8703331453325813, "eval_spearman_manhattan": 0.8672266689066755, "eval_spearman_max": 0.8749033396133584, "eval_steps_per_second": 1.173, "step": 300 }, { "epoch": 0.02, "grad_norm": 0.6371708512306213, "learning_rate": 4.903536977491961e-06, "loss": 0.019, "step": 305 }, { "epoch": 0.02, "grad_norm": 2.065721035003662, "learning_rate": 4.983922829581994e-06, "loss": 0.0193, "step": 310 }, { "epoch": 0.03, "grad_norm": 3.1992969512939453, "learning_rate": 5.064308681672026e-06, "loss": 0.0579, "step": 315 }, { "epoch": 0.03, "grad_norm": 2.422349214553833, "learning_rate": 5.144694533762058e-06, "loss": 0.0342, "step": 320 }, { "epoch": 0.03, "grad_norm": 2.09092378616333, "learning_rate": 5.22508038585209e-06, "loss": 0.0321, "step": 325 }, { "epoch": 0.03, "grad_norm": 1.6249920129776, "learning_rate": 5.305466237942123e-06, "loss": 0.05, "step": 330 }, { "epoch": 0.03, "grad_norm": 1.4493677616119385, "learning_rate": 5.385852090032154e-06, "loss": 0.0371, "step": 335 }, { "epoch": 0.03, "grad_norm": 1.1457599401474, "learning_rate": 5.466237942122187e-06, "loss": 0.0191, "step": 340 }, { "epoch": 0.03, "grad_norm": 0.7196341156959534, "learning_rate": 5.546623794212219e-06, "loss": 0.0653, "step": 345 }, { "epoch": 0.03, "grad_norm": 0.7609484791755676, "learning_rate": 5.627009646302252e-06, "loss": 0.0234, "step": 350 }, { "epoch": 0.03, "grad_norm": 1.9668811559677124, "learning_rate": 5.707395498392283e-06, "loss": 0.043, "step": 355 }, { "epoch": 0.03, "grad_norm": 1.4503982067108154, "learning_rate": 5.787781350482315e-06, "loss": 0.0423, "step": 360 }, { "epoch": 0.03, "grad_norm": 2.6571695804595947, "learning_rate": 5.868167202572348e-06, "loss": 0.0463, "step": 365 }, { "epoch": 0.03, "grad_norm": 1.0262566804885864, "learning_rate": 5.94855305466238e-06, "loss": 0.0362, "step": 370 }, { "epoch": 0.03, "grad_norm": 3.1059563159942627, "learning_rate": 6.028938906752412e-06, "loss": 0.0291, "step": 375 }, { "epoch": 0.03, "grad_norm": 1.3710355758666992, "learning_rate": 6.109324758842444e-06, "loss": 0.0286, "step": 380 }, { "epoch": 0.03, "grad_norm": 0.9536277055740356, "learning_rate": 6.189710610932477e-06, "loss": 0.048, "step": 385 }, { "epoch": 0.03, "grad_norm": 0.9812152981758118, "learning_rate": 6.270096463022508e-06, "loss": 0.0211, "step": 390 }, { "epoch": 0.03, "grad_norm": 1.8256497383117676, "learning_rate": 6.350482315112541e-06, "loss": 0.0248, "step": 395 }, { "epoch": 0.03, "grad_norm": 1.5534074306488037, "learning_rate": 6.430868167202573e-06, "loss": 0.0406, "step": 400 }, { "epoch": 0.03, "grad_norm": 2.0551486015319824, "learning_rate": 6.511254019292606e-06, "loss": 0.0299, "step": 405 }, { "epoch": 0.03, "grad_norm": 1.8460248708724976, "learning_rate": 6.591639871382637e-06, "loss": 0.0182, "step": 410 }, { "epoch": 0.03, "grad_norm": 0.7704716920852661, "learning_rate": 6.672025723472669e-06, "loss": 0.0223, "step": 415 }, { "epoch": 0.03, "grad_norm": 8.103775024414062, "learning_rate": 6.752411575562702e-06, "loss": 0.0296, "step": 420 }, { "epoch": 0.03, "grad_norm": 1.514302134513855, "learning_rate": 6.832797427652733e-06, "loss": 0.0302, "step": 425 }, { "epoch": 0.03, "grad_norm": 1.5787880420684814, "learning_rate": 6.913183279742766e-06, "loss": 0.0229, "step": 430 }, { "epoch": 0.03, "grad_norm": 1.478020429611206, "learning_rate": 6.993569131832798e-06, "loss": 0.03, "step": 435 }, { "epoch": 0.04, "grad_norm": 0.8203696608543396, "learning_rate": 7.073954983922831e-06, "loss": 0.0333, "step": 440 }, { "epoch": 0.04, "grad_norm": 1.000648856163025, "learning_rate": 7.154340836012862e-06, "loss": 0.0177, "step": 445 }, { "epoch": 0.04, "grad_norm": 1.5386018753051758, "learning_rate": 7.234726688102894e-06, "loss": 0.0351, "step": 450 }, { "epoch": 0.04, "grad_norm": 1.0292292833328247, "learning_rate": 7.315112540192927e-06, "loss": 0.0395, "step": 455 }, { "epoch": 0.04, "grad_norm": 0.5625340938568115, "learning_rate": 7.39549839228296e-06, "loss": 0.0299, "step": 460 }, { "epoch": 0.04, "grad_norm": 3.6210525035858154, "learning_rate": 7.475884244372991e-06, "loss": 0.0321, "step": 465 }, { "epoch": 0.04, "grad_norm": 1.8112176656723022, "learning_rate": 7.556270096463023e-06, "loss": 0.0316, "step": 470 }, { "epoch": 0.04, "grad_norm": 1.0322524309158325, "learning_rate": 7.636655948553056e-06, "loss": 0.0253, "step": 475 }, { "epoch": 0.04, "grad_norm": 3.308305263519287, "learning_rate": 7.717041800643087e-06, "loss": 0.0387, "step": 480 }, { "epoch": 0.04, "grad_norm": 1.2584173679351807, "learning_rate": 7.79742765273312e-06, "loss": 0.0353, "step": 485 }, { "epoch": 0.04, "grad_norm": 0.6924129128456116, "learning_rate": 7.877813504823153e-06, "loss": 0.0157, "step": 490 }, { "epoch": 0.04, "grad_norm": 2.4199492931365967, "learning_rate": 7.958199356913184e-06, "loss": 0.0401, "step": 495 }, { "epoch": 0.04, "grad_norm": 0.35832083225250244, "learning_rate": 8.038585209003215e-06, "loss": 0.0227, "step": 500 }, { "epoch": 0.04, "grad_norm": 1.6043404340744019, "learning_rate": 8.118971061093248e-06, "loss": 0.0199, "step": 505 }, { "epoch": 0.04, "grad_norm": 2.0943973064422607, "learning_rate": 8.19935691318328e-06, "loss": 0.0213, "step": 510 }, { "epoch": 0.04, "grad_norm": 0.8544149398803711, "learning_rate": 8.279742765273314e-06, "loss": 0.0226, "step": 515 }, { "epoch": 0.04, "grad_norm": 4.021656036376953, "learning_rate": 8.360128617363345e-06, "loss": 0.0574, "step": 520 }, { "epoch": 0.04, "grad_norm": 2.3117215633392334, "learning_rate": 8.440514469453378e-06, "loss": 0.019, "step": 525 }, { "epoch": 0.04, "grad_norm": 1.2881953716278076, "learning_rate": 8.520900321543409e-06, "loss": 0.0267, "step": 530 }, { "epoch": 0.04, "grad_norm": 2.115152597427368, "learning_rate": 8.601286173633442e-06, "loss": 0.0233, "step": 535 }, { "epoch": 0.04, "grad_norm": 1.7420244216918945, "learning_rate": 8.681672025723473e-06, "loss": 0.0216, "step": 540 }, { "epoch": 0.04, "grad_norm": 1.2626619338989258, "learning_rate": 8.762057877813506e-06, "loss": 0.0182, "step": 545 }, { "epoch": 0.04, "grad_norm": 1.5143030881881714, "learning_rate": 8.842443729903538e-06, "loss": 0.0261, "step": 550 }, { "epoch": 0.04, "grad_norm": 1.1002925634384155, "learning_rate": 8.92282958199357e-06, "loss": 0.022, "step": 555 }, { "epoch": 0.05, "grad_norm": 1.9754141569137573, "learning_rate": 9.003215434083602e-06, "loss": 0.0252, "step": 560 }, { "epoch": 0.05, "grad_norm": 1.3871146440505981, "learning_rate": 9.083601286173634e-06, "loss": 0.0456, "step": 565 }, { "epoch": 0.05, "grad_norm": 1.471267580986023, "learning_rate": 9.163987138263667e-06, "loss": 0.0384, "step": 570 }, { "epoch": 0.05, "grad_norm": 0.8994916677474976, "learning_rate": 9.244372990353698e-06, "loss": 0.036, "step": 575 }, { "epoch": 0.05, "grad_norm": 1.7490404844284058, "learning_rate": 9.32475884244373e-06, "loss": 0.0221, "step": 580 }, { "epoch": 0.05, "grad_norm": 1.3971872329711914, "learning_rate": 9.405144694533763e-06, "loss": 0.0302, "step": 585 }, { "epoch": 0.05, "grad_norm": 0.9706817865371704, "learning_rate": 9.485530546623795e-06, "loss": 0.0291, "step": 590 }, { "epoch": 0.05, "grad_norm": 2.0623135566711426, "learning_rate": 9.565916398713827e-06, "loss": 0.0332, "step": 595 }, { "epoch": 0.05, "grad_norm": 2.330348253250122, "learning_rate": 9.64630225080386e-06, "loss": 0.031, "step": 600 }, { "epoch": 0.05, "eval_loss": 0.03203292191028595, "eval_pearson_cosine": 0.8607192041230748, "eval_pearson_dot": 0.8492954726274867, "eval_pearson_euclidean": 0.8425777905216318, "eval_pearson_manhattan": 0.8394006469079365, "eval_pearson_max": 0.8607192041230748, "eval_runtime": 425.6285, "eval_samples_per_second": 1.175, "eval_spearman_cosine": 0.8971169644678579, "eval_spearman_dot": 0.8784386497545991, "eval_spearman_euclidean": 0.894069240276961, "eval_spearman_manhattan": 0.8926532346129386, "eval_spearman_max": 0.8971169644678579, "eval_steps_per_second": 1.175, "step": 600 }, { "epoch": 0.05, "grad_norm": 2.334221124649048, "learning_rate": 9.726688102893891e-06, "loss": 0.0273, "step": 605 }, { "epoch": 0.05, "grad_norm": 0.7424949407577515, "learning_rate": 9.807073954983923e-06, "loss": 0.0647, "step": 610 }, { "epoch": 0.05, "grad_norm": 1.4894630908966064, "learning_rate": 9.887459807073955e-06, "loss": 0.0361, "step": 615 }, { "epoch": 0.05, "grad_norm": 0.4873441159725189, "learning_rate": 9.967845659163988e-06, "loss": 0.0264, "step": 620 }, { "epoch": 0.05, "grad_norm": 1.5973559617996216, "learning_rate": 1.0048231511254021e-05, "loss": 0.029, "step": 625 }, { "epoch": 0.05, "grad_norm": 1.1919150352478027, "learning_rate": 1.0128617363344052e-05, "loss": 0.0328, "step": 630 }, { "epoch": 0.05, "grad_norm": 1.3969812393188477, "learning_rate": 1.0209003215434084e-05, "loss": 0.043, "step": 635 }, { "epoch": 0.05, "grad_norm": 2.33707332611084, "learning_rate": 1.0289389067524116e-05, "loss": 0.0448, "step": 640 }, { "epoch": 0.05, "grad_norm": 1.3516573905944824, "learning_rate": 1.0369774919614148e-05, "loss": 0.0328, "step": 645 }, { "epoch": 0.05, "grad_norm": 3.014158248901367, "learning_rate": 1.045016077170418e-05, "loss": 0.0326, "step": 650 }, { "epoch": 0.05, "grad_norm": 1.1477378606796265, "learning_rate": 1.0530546623794213e-05, "loss": 0.025, "step": 655 }, { "epoch": 0.05, "grad_norm": 0.8221299648284912, "learning_rate": 1.0610932475884246e-05, "loss": 0.0425, "step": 660 }, { "epoch": 0.05, "grad_norm": 1.68177330493927, "learning_rate": 1.0691318327974279e-05, "loss": 0.025, "step": 665 }, { "epoch": 0.05, "grad_norm": 1.4552159309387207, "learning_rate": 1.0771704180064308e-05, "loss": 0.0398, "step": 670 }, { "epoch": 0.05, "grad_norm": 1.352478265762329, "learning_rate": 1.0852090032154341e-05, "loss": 0.0155, "step": 675 }, { "epoch": 0.05, "grad_norm": 0.9538177251815796, "learning_rate": 1.0932475884244374e-05, "loss": 0.0299, "step": 680 }, { "epoch": 0.06, "grad_norm": 2.0173323154449463, "learning_rate": 1.1012861736334405e-05, "loss": 0.0268, "step": 685 }, { "epoch": 0.06, "grad_norm": 2.0908379554748535, "learning_rate": 1.1093247588424438e-05, "loss": 0.0308, "step": 690 }, { "epoch": 0.06, "grad_norm": 0.6585212349891663, "learning_rate": 1.1173633440514471e-05, "loss": 0.025, "step": 695 }, { "epoch": 0.06, "grad_norm": 0.959600567817688, "learning_rate": 1.1254019292604504e-05, "loss": 0.0219, "step": 700 }, { "epoch": 0.06, "grad_norm": 1.2033023834228516, "learning_rate": 1.1334405144694535e-05, "loss": 0.0315, "step": 705 }, { "epoch": 0.06, "grad_norm": 1.735729455947876, "learning_rate": 1.1414790996784566e-05, "loss": 0.0249, "step": 710 }, { "epoch": 0.06, "grad_norm": 1.850893497467041, "learning_rate": 1.14951768488746e-05, "loss": 0.0248, "step": 715 }, { "epoch": 0.06, "grad_norm": 0.902856171131134, "learning_rate": 1.157556270096463e-05, "loss": 0.0683, "step": 720 }, { "epoch": 0.06, "grad_norm": 0.6975336670875549, "learning_rate": 1.1655948553054663e-05, "loss": 0.0315, "step": 725 }, { "epoch": 0.06, "grad_norm": 3.3982033729553223, "learning_rate": 1.1736334405144696e-05, "loss": 0.0479, "step": 730 }, { "epoch": 0.06, "grad_norm": 2.775419235229492, "learning_rate": 1.1816720257234729e-05, "loss": 0.068, "step": 735 }, { "epoch": 0.06, "grad_norm": 1.9351024627685547, "learning_rate": 1.189710610932476e-05, "loss": 0.0254, "step": 740 }, { "epoch": 0.06, "grad_norm": 0.5352950096130371, "learning_rate": 1.1977491961414791e-05, "loss": 0.0193, "step": 745 }, { "epoch": 0.06, "grad_norm": 1.5528688430786133, "learning_rate": 1.2057877813504824e-05, "loss": 0.039, "step": 750 }, { "epoch": 0.06, "grad_norm": 1.6287634372711182, "learning_rate": 1.2138263665594855e-05, "loss": 0.017, "step": 755 }, { "epoch": 0.06, "grad_norm": 1.0720481872558594, "learning_rate": 1.2218649517684888e-05, "loss": 0.0261, "step": 760 }, { "epoch": 0.06, "grad_norm": 0.8808945417404175, "learning_rate": 1.2299035369774921e-05, "loss": 0.0224, "step": 765 }, { "epoch": 0.06, "grad_norm": 1.175379753112793, "learning_rate": 1.2379421221864954e-05, "loss": 0.0157, "step": 770 }, { "epoch": 0.06, "grad_norm": 0.7339066863059998, "learning_rate": 1.2459807073954987e-05, "loss": 0.0242, "step": 775 }, { "epoch": 0.06, "grad_norm": 0.5107578039169312, "learning_rate": 1.2540192926045016e-05, "loss": 0.0245, "step": 780 }, { "epoch": 0.06, "grad_norm": 0.5132850408554077, "learning_rate": 1.2620578778135049e-05, "loss": 0.0191, "step": 785 }, { "epoch": 0.06, "grad_norm": 0.7692009210586548, "learning_rate": 1.2700964630225082e-05, "loss": 0.019, "step": 790 }, { "epoch": 0.06, "grad_norm": 1.1860913038253784, "learning_rate": 1.2781350482315113e-05, "loss": 0.0121, "step": 795 }, { "epoch": 0.06, "grad_norm": 0.9162280559539795, "learning_rate": 1.2861736334405146e-05, "loss": 0.0195, "step": 800 }, { "epoch": 0.06, "grad_norm": 1.2460323572158813, "learning_rate": 1.2942122186495179e-05, "loss": 0.0223, "step": 805 }, { "epoch": 0.07, "grad_norm": 0.7504421472549438, "learning_rate": 1.3022508038585212e-05, "loss": 0.0325, "step": 810 }, { "epoch": 0.07, "grad_norm": 1.3735928535461426, "learning_rate": 1.3102893890675241e-05, "loss": 0.0213, "step": 815 }, { "epoch": 0.07, "grad_norm": 1.1271485090255737, "learning_rate": 1.3183279742765274e-05, "loss": 0.0219, "step": 820 }, { "epoch": 0.07, "grad_norm": 0.8849589228630066, "learning_rate": 1.3263665594855307e-05, "loss": 0.0143, "step": 825 }, { "epoch": 0.07, "grad_norm": 0.7527520060539246, "learning_rate": 1.3344051446945338e-05, "loss": 0.0241, "step": 830 }, { "epoch": 0.07, "grad_norm": 1.0900169610977173, "learning_rate": 1.3424437299035371e-05, "loss": 0.0414, "step": 835 }, { "epoch": 0.07, "grad_norm": 1.1919902563095093, "learning_rate": 1.3504823151125404e-05, "loss": 0.0234, "step": 840 }, { "epoch": 0.07, "grad_norm": 0.8385984301567078, "learning_rate": 1.3585209003215437e-05, "loss": 0.0239, "step": 845 }, { "epoch": 0.07, "grad_norm": 0.9777655005455017, "learning_rate": 1.3665594855305466e-05, "loss": 0.0145, "step": 850 }, { "epoch": 0.07, "grad_norm": 1.2803690433502197, "learning_rate": 1.3745980707395499e-05, "loss": 0.0274, "step": 855 }, { "epoch": 0.07, "grad_norm": 1.1460819244384766, "learning_rate": 1.3826366559485532e-05, "loss": 0.0373, "step": 860 }, { "epoch": 0.07, "grad_norm": 0.8751824498176575, "learning_rate": 1.3906752411575563e-05, "loss": 0.0204, "step": 865 }, { "epoch": 0.07, "grad_norm": 0.566417932510376, "learning_rate": 1.3987138263665596e-05, "loss": 0.0139, "step": 870 }, { "epoch": 0.07, "grad_norm": 0.6487017869949341, "learning_rate": 1.4067524115755629e-05, "loss": 0.0458, "step": 875 }, { "epoch": 0.07, "grad_norm": 1.9419571161270142, "learning_rate": 1.4147909967845662e-05, "loss": 0.0215, "step": 880 }, { "epoch": 0.07, "grad_norm": 0.7554964423179626, "learning_rate": 1.4228295819935693e-05, "loss": 0.0219, "step": 885 }, { "epoch": 0.07, "grad_norm": 1.240870475769043, "learning_rate": 1.4308681672025724e-05, "loss": 0.0291, "step": 890 }, { "epoch": 0.07, "grad_norm": 1.2490875720977783, "learning_rate": 1.4389067524115757e-05, "loss": 0.0256, "step": 895 }, { "epoch": 0.07, "grad_norm": 0.9301912188529968, "learning_rate": 1.4469453376205788e-05, "loss": 0.0303, "step": 900 }, { "epoch": 0.07, "eval_loss": 0.0409804992377758, "eval_pearson_cosine": 0.8657545592954407, "eval_pearson_dot": 0.8604614808110218, "eval_pearson_euclidean": 0.849931010036339, "eval_pearson_manhattan": 0.84985721810182, "eval_pearson_max": 0.8657545592954407, "eval_runtime": 425.4053, "eval_samples_per_second": 1.175, "eval_spearman_cosine": 0.9034730538922157, "eval_spearman_dot": 0.8936158304633218, "eval_spearman_euclidean": 0.9027816591266363, "eval_spearman_manhattan": 0.9029234516938067, "eval_spearman_max": 0.9034730538922157, "eval_steps_per_second": 1.175, "step": 900 }, { "epoch": 0.07, "grad_norm": 0.6353436708450317, "learning_rate": 1.454983922829582e-05, "loss": 0.0306, "step": 905 }, { "epoch": 0.07, "grad_norm": 3.5071334838867188, "learning_rate": 1.4630225080385854e-05, "loss": 0.0529, "step": 910 }, { "epoch": 0.07, "grad_norm": 2.511518716812134, "learning_rate": 1.4710610932475886e-05, "loss": 0.0351, "step": 915 }, { "epoch": 0.07, "grad_norm": 0.8150439858436584, "learning_rate": 1.479099678456592e-05, "loss": 0.0304, "step": 920 }, { "epoch": 0.07, "grad_norm": 0.909027636051178, "learning_rate": 1.4871382636655949e-05, "loss": 0.0306, "step": 925 }, { "epoch": 0.07, "grad_norm": 2.117823362350464, "learning_rate": 1.4951768488745982e-05, "loss": 0.0444, "step": 930 }, { "epoch": 0.08, "grad_norm": 1.0820565223693848, "learning_rate": 1.5032154340836015e-05, "loss": 0.0147, "step": 935 }, { "epoch": 0.08, "grad_norm": 1.0721031427383423, "learning_rate": 1.5112540192926046e-05, "loss": 0.0379, "step": 940 }, { "epoch": 0.08, "grad_norm": 0.6273934841156006, "learning_rate": 1.5192926045016079e-05, "loss": 0.0202, "step": 945 }, { "epoch": 0.08, "grad_norm": 1.4175567626953125, "learning_rate": 1.527331189710611e-05, "loss": 0.0265, "step": 950 }, { "epoch": 0.08, "grad_norm": 1.7840615510940552, "learning_rate": 1.5353697749196143e-05, "loss": 0.0194, "step": 955 }, { "epoch": 0.08, "grad_norm": 2.6450772285461426, "learning_rate": 1.5434083601286174e-05, "loss": 0.0271, "step": 960 }, { "epoch": 0.08, "grad_norm": 1.1013175249099731, "learning_rate": 1.5514469453376205e-05, "loss": 0.035, "step": 965 }, { "epoch": 0.08, "grad_norm": 0.6244901418685913, "learning_rate": 1.559485530546624e-05, "loss": 0.0183, "step": 970 }, { "epoch": 0.08, "grad_norm": 1.0193443298339844, "learning_rate": 1.567524115755627e-05, "loss": 0.0234, "step": 975 }, { "epoch": 0.08, "grad_norm": 1.1802617311477661, "learning_rate": 1.5755627009646305e-05, "loss": 0.0302, "step": 980 }, { "epoch": 0.08, "grad_norm": 1.1835167407989502, "learning_rate": 1.5836012861736336e-05, "loss": 0.0253, "step": 985 }, { "epoch": 0.08, "grad_norm": 1.4406228065490723, "learning_rate": 1.5916398713826368e-05, "loss": 0.0316, "step": 990 }, { "epoch": 0.08, "grad_norm": 0.9229764938354492, "learning_rate": 1.59967845659164e-05, "loss": 0.0408, "step": 995 }, { "epoch": 0.08, "grad_norm": 0.6133905649185181, "learning_rate": 1.607717041800643e-05, "loss": 0.0163, "step": 1000 }, { "epoch": 0.08, "grad_norm": 1.388720154762268, "learning_rate": 1.6157556270096464e-05, "loss": 0.0221, "step": 1005 }, { "epoch": 0.08, "grad_norm": 0.9162605404853821, "learning_rate": 1.6237942122186496e-05, "loss": 0.0292, "step": 1010 }, { "epoch": 0.08, "grad_norm": 1.0816845893859863, "learning_rate": 1.631832797427653e-05, "loss": 0.0284, "step": 1015 }, { "epoch": 0.08, "grad_norm": 0.974607527256012, "learning_rate": 1.639871382636656e-05, "loss": 0.0245, "step": 1020 }, { "epoch": 0.08, "grad_norm": 0.7858002185821533, "learning_rate": 1.6479099678456592e-05, "loss": 0.0208, "step": 1025 }, { "epoch": 0.08, "grad_norm": 0.8541093468666077, "learning_rate": 1.6559485530546627e-05, "loss": 0.0402, "step": 1030 }, { "epoch": 0.08, "grad_norm": 0.9254925847053528, "learning_rate": 1.6639871382636655e-05, "loss": 0.0277, "step": 1035 }, { "epoch": 0.08, "grad_norm": 1.1855261325836182, "learning_rate": 1.672025723472669e-05, "loss": 0.0289, "step": 1040 }, { "epoch": 0.08, "grad_norm": 1.1872082948684692, "learning_rate": 1.680064308681672e-05, "loss": 0.0255, "step": 1045 }, { "epoch": 0.08, "grad_norm": 1.9865713119506836, "learning_rate": 1.6881028938906755e-05, "loss": 0.0331, "step": 1050 }, { "epoch": 0.08, "grad_norm": 1.5213806629180908, "learning_rate": 1.6961414790996786e-05, "loss": 0.0385, "step": 1055 }, { "epoch": 0.09, "grad_norm": 1.5804238319396973, "learning_rate": 1.7041800643086817e-05, "loss": 0.0447, "step": 1060 }, { "epoch": 0.09, "grad_norm": 1.2709295749664307, "learning_rate": 1.7122186495176852e-05, "loss": 0.0184, "step": 1065 }, { "epoch": 0.09, "grad_norm": 0.6574845314025879, "learning_rate": 1.7202572347266883e-05, "loss": 0.0201, "step": 1070 }, { "epoch": 0.09, "grad_norm": 2.590877056121826, "learning_rate": 1.7282958199356914e-05, "loss": 0.0456, "step": 1075 }, { "epoch": 0.09, "grad_norm": 1.060797929763794, "learning_rate": 1.7363344051446945e-05, "loss": 0.0375, "step": 1080 }, { "epoch": 0.09, "grad_norm": 0.38614028692245483, "learning_rate": 1.744372990353698e-05, "loss": 0.0371, "step": 1085 }, { "epoch": 0.09, "grad_norm": 1.6158710718154907, "learning_rate": 1.752411575562701e-05, "loss": 0.0403, "step": 1090 }, { "epoch": 0.09, "grad_norm": 1.5008140802383423, "learning_rate": 1.7604501607717042e-05, "loss": 0.0296, "step": 1095 }, { "epoch": 0.09, "grad_norm": 0.41157564520835876, "learning_rate": 1.7684887459807077e-05, "loss": 0.0255, "step": 1100 }, { "epoch": 0.09, "grad_norm": 1.3567136526107788, "learning_rate": 1.7765273311897108e-05, "loss": 0.0217, "step": 1105 }, { "epoch": 0.09, "grad_norm": 0.698455810546875, "learning_rate": 1.784565916398714e-05, "loss": 0.0188, "step": 1110 }, { "epoch": 0.09, "grad_norm": 1.4908955097198486, "learning_rate": 1.792604501607717e-05, "loss": 0.0188, "step": 1115 }, { "epoch": 0.09, "grad_norm": 2.2162156105041504, "learning_rate": 1.8006430868167205e-05, "loss": 0.02, "step": 1120 }, { "epoch": 0.09, "grad_norm": 0.3919101655483246, "learning_rate": 1.8086816720257236e-05, "loss": 0.0133, "step": 1125 }, { "epoch": 0.09, "grad_norm": 0.4801468551158905, "learning_rate": 1.8167202572347267e-05, "loss": 0.0303, "step": 1130 }, { "epoch": 0.09, "grad_norm": 0.5365796685218811, "learning_rate": 1.8247588424437302e-05, "loss": 0.0275, "step": 1135 }, { "epoch": 0.09, "grad_norm": 0.777245819568634, "learning_rate": 1.8327974276527333e-05, "loss": 0.0273, "step": 1140 }, { "epoch": 0.09, "grad_norm": 1.5682530403137207, "learning_rate": 1.8408360128617364e-05, "loss": 0.0261, "step": 1145 }, { "epoch": 0.09, "grad_norm": 1.0468765497207642, "learning_rate": 1.8488745980707395e-05, "loss": 0.0288, "step": 1150 }, { "epoch": 0.09, "grad_norm": 1.2527729272842407, "learning_rate": 1.856913183279743e-05, "loss": 0.0239, "step": 1155 }, { "epoch": 0.09, "grad_norm": 0.5002691149711609, "learning_rate": 1.864951768488746e-05, "loss": 0.0219, "step": 1160 }, { "epoch": 0.09, "grad_norm": 1.9675867557525635, "learning_rate": 1.8729903536977496e-05, "loss": 0.0476, "step": 1165 }, { "epoch": 0.09, "grad_norm": 0.8520388007164001, "learning_rate": 1.8810289389067527e-05, "loss": 0.0258, "step": 1170 }, { "epoch": 0.09, "grad_norm": 0.5552634596824646, "learning_rate": 1.8890675241157558e-05, "loss": 0.0237, "step": 1175 }, { "epoch": 0.09, "grad_norm": 0.6992123126983643, "learning_rate": 1.897106109324759e-05, "loss": 0.0265, "step": 1180 }, { "epoch": 0.1, "grad_norm": 1.1183786392211914, "learning_rate": 1.905144694533762e-05, "loss": 0.027, "step": 1185 }, { "epoch": 0.1, "grad_norm": 2.0854079723358154, "learning_rate": 1.9131832797427655e-05, "loss": 0.0252, "step": 1190 }, { "epoch": 0.1, "grad_norm": 1.8095922470092773, "learning_rate": 1.9212218649517686e-05, "loss": 0.037, "step": 1195 }, { "epoch": 0.1, "grad_norm": 1.6434860229492188, "learning_rate": 1.929260450160772e-05, "loss": 0.0277, "step": 1200 }, { "epoch": 0.1, "eval_loss": 0.030386893078684807, "eval_pearson_cosine": 0.8622394577970579, "eval_pearson_dot": 0.8570136493582737, "eval_pearson_euclidean": 0.849571905885398, "eval_pearson_manhattan": 0.8505210712126229, "eval_pearson_max": 0.8622394577970579, "eval_runtime": 425.8405, "eval_samples_per_second": 1.174, "eval_spearman_cosine": 0.8911063804255216, "eval_spearman_dot": 0.8816442945771783, "eval_spearman_euclidean": 0.8895569342277367, "eval_spearman_manhattan": 0.890007464029856, "eval_spearman_max": 0.8911063804255216, "eval_steps_per_second": 1.174, "step": 1200 }, { "epoch": 0.1, "grad_norm": 0.8423680067062378, "learning_rate": 1.9372990353697752e-05, "loss": 0.0313, "step": 1205 }, { "epoch": 0.1, "grad_norm": 1.2395281791687012, "learning_rate": 1.9453376205787783e-05, "loss": 0.0297, "step": 1210 }, { "epoch": 0.1, "grad_norm": 1.149891972541809, "learning_rate": 1.9533762057877814e-05, "loss": 0.0208, "step": 1215 }, { "epoch": 0.1, "grad_norm": 2.4256904125213623, "learning_rate": 1.9614147909967845e-05, "loss": 0.0437, "step": 1220 }, { "epoch": 0.1, "grad_norm": 0.7631718516349792, "learning_rate": 1.969453376205788e-05, "loss": 0.0335, "step": 1225 }, { "epoch": 0.1, "grad_norm": 1.0927650928497314, "learning_rate": 1.977491961414791e-05, "loss": 0.0349, "step": 1230 }, { "epoch": 0.1, "grad_norm": 0.7310745120048523, "learning_rate": 1.9855305466237946e-05, "loss": 0.0305, "step": 1235 }, { "epoch": 0.1, "grad_norm": 2.624340534210205, "learning_rate": 1.9935691318327977e-05, "loss": 0.0456, "step": 1240 }, { "epoch": 0.1, "grad_norm": 1.09121835231781, "learning_rate": 1.999821316894488e-05, "loss": 0.0119, "step": 1245 }, { "epoch": 0.1, "grad_norm": 1.6216334104537964, "learning_rate": 1.998927901366926e-05, "loss": 0.0419, "step": 1250 }, { "epoch": 0.1, "grad_norm": 1.0775346755981445, "learning_rate": 1.998034485839364e-05, "loss": 0.0317, "step": 1255 }, { "epoch": 0.1, "grad_norm": 1.044417381286621, "learning_rate": 1.997141070311802e-05, "loss": 0.0245, "step": 1260 }, { "epoch": 0.1, "grad_norm": 1.581361174583435, "learning_rate": 1.9962476547842403e-05, "loss": 0.0246, "step": 1265 }, { "epoch": 0.1, "grad_norm": 0.4875144362449646, "learning_rate": 1.9953542392566784e-05, "loss": 0.0218, "step": 1270 }, { "epoch": 0.1, "grad_norm": 1.2498222589492798, "learning_rate": 1.9944608237291165e-05, "loss": 0.04, "step": 1275 }, { "epoch": 0.1, "grad_norm": 0.562926173210144, "learning_rate": 1.993567408201555e-05, "loss": 0.0268, "step": 1280 }, { "epoch": 0.1, "grad_norm": 0.5474754571914673, "learning_rate": 1.992673992673993e-05, "loss": 0.0514, "step": 1285 }, { "epoch": 0.1, "grad_norm": 0.6975357532501221, "learning_rate": 1.9917805771464308e-05, "loss": 0.0345, "step": 1290 }, { "epoch": 0.1, "grad_norm": 2.0237133502960205, "learning_rate": 1.9908871616188692e-05, "loss": 0.0523, "step": 1295 }, { "epoch": 0.1, "grad_norm": 0.604995846748352, "learning_rate": 1.9899937460913073e-05, "loss": 0.0279, "step": 1300 }, { "epoch": 0.1, "grad_norm": 1.3883243799209595, "learning_rate": 1.9891003305637454e-05, "loss": 0.0227, "step": 1305 }, { "epoch": 0.11, "grad_norm": 0.5331462621688843, "learning_rate": 1.9882069150361835e-05, "loss": 0.0326, "step": 1310 }, { "epoch": 0.11, "grad_norm": 0.44855189323425293, "learning_rate": 1.9873134995086216e-05, "loss": 0.0352, "step": 1315 }, { "epoch": 0.11, "grad_norm": 2.213827610015869, "learning_rate": 1.9864200839810597e-05, "loss": 0.0432, "step": 1320 }, { "epoch": 0.11, "grad_norm": 0.5794016718864441, "learning_rate": 1.985526668453498e-05, "loss": 0.0245, "step": 1325 }, { "epoch": 0.11, "grad_norm": 0.6042608618736267, "learning_rate": 1.984633252925936e-05, "loss": 0.0168, "step": 1330 }, { "epoch": 0.11, "grad_norm": 1.935172200202942, "learning_rate": 1.983739837398374e-05, "loss": 0.0347, "step": 1335 }, { "epoch": 0.11, "grad_norm": 1.000755786895752, "learning_rate": 1.982846421870812e-05, "loss": 0.0302, "step": 1340 }, { "epoch": 0.11, "grad_norm": 2.054868698120117, "learning_rate": 1.9819530063432503e-05, "loss": 0.0622, "step": 1345 }, { "epoch": 0.11, "grad_norm": 0.873913049697876, "learning_rate": 1.9810595908156884e-05, "loss": 0.0178, "step": 1350 }, { "epoch": 0.11, "grad_norm": 0.6479474306106567, "learning_rate": 1.9801661752881268e-05, "loss": 0.0204, "step": 1355 }, { "epoch": 0.11, "grad_norm": 1.0146560668945312, "learning_rate": 1.979272759760565e-05, "loss": 0.0312, "step": 1360 }, { "epoch": 0.11, "grad_norm": 0.7326614856719971, "learning_rate": 1.978379344233003e-05, "loss": 0.0444, "step": 1365 }, { "epoch": 0.11, "grad_norm": 0.8839195370674133, "learning_rate": 1.977485928705441e-05, "loss": 0.0301, "step": 1370 }, { "epoch": 0.11, "grad_norm": 0.485784113407135, "learning_rate": 1.9765925131778792e-05, "loss": 0.0349, "step": 1375 }, { "epoch": 0.11, "grad_norm": 0.7656601071357727, "learning_rate": 1.9756990976503173e-05, "loss": 0.0282, "step": 1380 }, { "epoch": 0.11, "grad_norm": 0.6562867164611816, "learning_rate": 1.9748056821227554e-05, "loss": 0.033, "step": 1385 }, { "epoch": 0.11, "grad_norm": 1.971908688545227, "learning_rate": 1.9739122665951935e-05, "loss": 0.0358, "step": 1390 }, { "epoch": 0.11, "grad_norm": 1.4427505731582642, "learning_rate": 1.9730188510676316e-05, "loss": 0.029, "step": 1395 }, { "epoch": 0.11, "grad_norm": 1.177832841873169, "learning_rate": 1.9721254355400697e-05, "loss": 0.0296, "step": 1400 }, { "epoch": 0.11, "grad_norm": 0.5108907222747803, "learning_rate": 1.971232020012508e-05, "loss": 0.0239, "step": 1405 }, { "epoch": 0.11, "grad_norm": 1.7967244386672974, "learning_rate": 1.9703386044849463e-05, "loss": 0.0336, "step": 1410 }, { "epoch": 0.11, "grad_norm": 1.008859395980835, "learning_rate": 1.9694451889573844e-05, "loss": 0.0197, "step": 1415 }, { "epoch": 0.11, "grad_norm": 1.1244670152664185, "learning_rate": 1.968551773429822e-05, "loss": 0.0407, "step": 1420 }, { "epoch": 0.11, "grad_norm": 1.6256637573242188, "learning_rate": 1.9676583579022606e-05, "loss": 0.0464, "step": 1425 }, { "epoch": 0.11, "grad_norm": 0.6180126667022705, "learning_rate": 1.9667649423746987e-05, "loss": 0.0185, "step": 1430 }, { "epoch": 0.12, "grad_norm": 0.8708995580673218, "learning_rate": 1.9658715268471368e-05, "loss": 0.0206, "step": 1435 }, { "epoch": 0.12, "grad_norm": 1.933128833770752, "learning_rate": 1.964978111319575e-05, "loss": 0.0427, "step": 1440 }, { "epoch": 0.12, "grad_norm": 0.588142991065979, "learning_rate": 1.964084695792013e-05, "loss": 0.0242, "step": 1445 }, { "epoch": 0.12, "grad_norm": 0.38248518109321594, "learning_rate": 1.963191280264451e-05, "loss": 0.0252, "step": 1450 }, { "epoch": 0.12, "grad_norm": 0.8211073279380798, "learning_rate": 1.9622978647368892e-05, "loss": 0.0159, "step": 1455 }, { "epoch": 0.12, "grad_norm": 0.4323325455188751, "learning_rate": 1.9614044492093273e-05, "loss": 0.0383, "step": 1460 }, { "epoch": 0.12, "grad_norm": 1.7777843475341797, "learning_rate": 1.9605110336817658e-05, "loss": 0.0399, "step": 1465 }, { "epoch": 0.12, "grad_norm": 0.7342023253440857, "learning_rate": 1.9596176181542035e-05, "loss": 0.0232, "step": 1470 }, { "epoch": 0.12, "grad_norm": 0.43286553025245667, "learning_rate": 1.9587242026266416e-05, "loss": 0.03, "step": 1475 }, { "epoch": 0.12, "grad_norm": 0.4914606809616089, "learning_rate": 1.95783078709908e-05, "loss": 0.0253, "step": 1480 }, { "epoch": 0.12, "grad_norm": 1.6109514236450195, "learning_rate": 1.9569373715715182e-05, "loss": 0.0244, "step": 1485 }, { "epoch": 0.12, "grad_norm": 1.099238395690918, "learning_rate": 1.9560439560439563e-05, "loss": 0.0219, "step": 1490 }, { "epoch": 0.12, "grad_norm": 0.672016978263855, "learning_rate": 1.9551505405163944e-05, "loss": 0.0156, "step": 1495 }, { "epoch": 0.12, "grad_norm": 0.9910028576850891, "learning_rate": 1.9542571249888325e-05, "loss": 0.025, "step": 1500 }, { "epoch": 0.12, "eval_loss": 0.029246093705296516, "eval_pearson_cosine": 0.8662779121023975, "eval_pearson_dot": 0.8609972822698294, "eval_pearson_euclidean": 0.8525960773878268, "eval_pearson_manhattan": 0.8525524127391249, "eval_pearson_max": 0.8662779121023975, "eval_runtime": 425.5831, "eval_samples_per_second": 1.175, "eval_spearman_cosine": 0.8936378145512582, "eval_spearman_dot": 0.8876585266341065, "eval_spearman_euclidean": 0.891318445273781, "eval_spearman_manhattan": 0.889811431245725, "eval_spearman_max": 0.8936378145512582, "eval_steps_per_second": 1.175, "step": 1500 }, { "epoch": 0.12, "grad_norm": 0.7004162073135376, "learning_rate": 1.9533637094612706e-05, "loss": 0.0248, "step": 1505 }, { "epoch": 0.12, "grad_norm": 1.0799355506896973, "learning_rate": 1.9524702939337087e-05, "loss": 0.0268, "step": 1510 }, { "epoch": 0.12, "grad_norm": 0.8577656745910645, "learning_rate": 1.9515768784061468e-05, "loss": 0.0363, "step": 1515 }, { "epoch": 0.12, "grad_norm": 1.811858892440796, "learning_rate": 1.950683462878585e-05, "loss": 0.0408, "step": 1520 }, { "epoch": 0.12, "grad_norm": 1.2648321390151978, "learning_rate": 1.949790047351023e-05, "loss": 0.0273, "step": 1525 }, { "epoch": 0.12, "grad_norm": 0.6332646012306213, "learning_rate": 1.948896631823461e-05, "loss": 0.0234, "step": 1530 }, { "epoch": 0.12, "grad_norm": 1.1494219303131104, "learning_rate": 1.9480032162958992e-05, "loss": 0.0396, "step": 1535 }, { "epoch": 0.12, "grad_norm": 1.5537444353103638, "learning_rate": 1.9471098007683377e-05, "loss": 0.0392, "step": 1540 }, { "epoch": 0.12, "grad_norm": 1.7606208324432373, "learning_rate": 1.9462163852407754e-05, "loss": 0.0468, "step": 1545 }, { "epoch": 0.12, "grad_norm": 0.3266965448856354, "learning_rate": 1.945322969713214e-05, "loss": 0.0164, "step": 1550 }, { "epoch": 0.13, "grad_norm": 0.4910846948623657, "learning_rate": 1.944429554185652e-05, "loss": 0.0181, "step": 1555 }, { "epoch": 0.13, "grad_norm": 0.621263861656189, "learning_rate": 1.94353613865809e-05, "loss": 0.021, "step": 1560 }, { "epoch": 0.13, "grad_norm": 0.8153178691864014, "learning_rate": 1.9426427231305282e-05, "loss": 0.019, "step": 1565 }, { "epoch": 0.13, "grad_norm": 0.9617846012115479, "learning_rate": 1.9417493076029663e-05, "loss": 0.0186, "step": 1570 }, { "epoch": 0.13, "grad_norm": 0.6482746005058289, "learning_rate": 1.9408558920754044e-05, "loss": 0.0226, "step": 1575 }, { "epoch": 0.13, "grad_norm": 0.6985549926757812, "learning_rate": 1.9399624765478425e-05, "loss": 0.0267, "step": 1580 }, { "epoch": 0.13, "grad_norm": 0.9519606828689575, "learning_rate": 1.9390690610202806e-05, "loss": 0.0187, "step": 1585 }, { "epoch": 0.13, "grad_norm": 0.9981301426887512, "learning_rate": 1.9381756454927187e-05, "loss": 0.0195, "step": 1590 }, { "epoch": 0.13, "grad_norm": 0.5453444719314575, "learning_rate": 1.937282229965157e-05, "loss": 0.0185, "step": 1595 }, { "epoch": 0.13, "grad_norm": 1.3481849431991577, "learning_rate": 1.936388814437595e-05, "loss": 0.0439, "step": 1600 }, { "epoch": 0.13, "grad_norm": 0.49602463841438293, "learning_rate": 1.935495398910033e-05, "loss": 0.0158, "step": 1605 }, { "epoch": 0.13, "grad_norm": 0.477524071931839, "learning_rate": 1.9346019833824715e-05, "loss": 0.0276, "step": 1610 }, { "epoch": 0.13, "grad_norm": 0.8374057412147522, "learning_rate": 1.9337085678549096e-05, "loss": 0.0259, "step": 1615 }, { "epoch": 0.13, "grad_norm": 0.9524512887001038, "learning_rate": 1.9328151523273477e-05, "loss": 0.0403, "step": 1620 }, { "epoch": 0.13, "grad_norm": 0.6074572205543518, "learning_rate": 1.9319217367997858e-05, "loss": 0.0217, "step": 1625 }, { "epoch": 0.13, "grad_norm": 0.7999153137207031, "learning_rate": 1.931028321272224e-05, "loss": 0.0234, "step": 1630 }, { "epoch": 0.13, "grad_norm": 0.9614180326461792, "learning_rate": 1.930134905744662e-05, "loss": 0.0231, "step": 1635 }, { "epoch": 0.13, "grad_norm": 1.1887145042419434, "learning_rate": 1.9292414902171e-05, "loss": 0.0414, "step": 1640 }, { "epoch": 0.13, "grad_norm": 0.8795381784439087, "learning_rate": 1.9283480746895382e-05, "loss": 0.0208, "step": 1645 }, { "epoch": 0.13, "grad_norm": 1.3685334920883179, "learning_rate": 1.9274546591619766e-05, "loss": 0.054, "step": 1650 }, { "epoch": 0.13, "grad_norm": 1.7518516778945923, "learning_rate": 1.9265612436344144e-05, "loss": 0.0274, "step": 1655 }, { "epoch": 0.13, "grad_norm": 1.029801607131958, "learning_rate": 1.9256678281068525e-05, "loss": 0.0174, "step": 1660 }, { "epoch": 0.13, "grad_norm": 1.7995644807815552, "learning_rate": 1.924774412579291e-05, "loss": 0.037, "step": 1665 }, { "epoch": 0.13, "grad_norm": 1.5098012685775757, "learning_rate": 1.923880997051729e-05, "loss": 0.0216, "step": 1670 }, { "epoch": 0.13, "grad_norm": 0.5958812832832336, "learning_rate": 1.9229875815241668e-05, "loss": 0.0142, "step": 1675 }, { "epoch": 0.14, "grad_norm": 0.7016432285308838, "learning_rate": 1.9220941659966053e-05, "loss": 0.0416, "step": 1680 }, { "epoch": 0.14, "grad_norm": 0.5232900977134705, "learning_rate": 1.9212007504690434e-05, "loss": 0.0264, "step": 1685 }, { "epoch": 0.14, "grad_norm": 0.4362630248069763, "learning_rate": 1.9203073349414815e-05, "loss": 0.022, "step": 1690 }, { "epoch": 0.14, "grad_norm": 1.451072335243225, "learning_rate": 1.9194139194139196e-05, "loss": 0.0262, "step": 1695 }, { "epoch": 0.14, "grad_norm": 0.5889037251472473, "learning_rate": 1.9185205038863577e-05, "loss": 0.0267, "step": 1700 }, { "epoch": 0.14, "grad_norm": 0.49965912103652954, "learning_rate": 1.9176270883587958e-05, "loss": 0.0159, "step": 1705 }, { "epoch": 0.14, "grad_norm": 1.920385479927063, "learning_rate": 1.916733672831234e-05, "loss": 0.0357, "step": 1710 }, { "epoch": 0.14, "grad_norm": 0.5249290466308594, "learning_rate": 1.915840257303672e-05, "loss": 0.0213, "step": 1715 }, { "epoch": 0.14, "grad_norm": 0.5141476988792419, "learning_rate": 1.91494684177611e-05, "loss": 0.0196, "step": 1720 }, { "epoch": 0.14, "grad_norm": 0.5003548264503479, "learning_rate": 1.9140534262485485e-05, "loss": 0.0157, "step": 1725 }, { "epoch": 0.14, "grad_norm": 0.4535447657108307, "learning_rate": 1.9131600107209863e-05, "loss": 0.019, "step": 1730 }, { "epoch": 0.14, "grad_norm": 0.4489072263240814, "learning_rate": 1.9122665951934247e-05, "loss": 0.0168, "step": 1735 }, { "epoch": 0.14, "grad_norm": 0.5539095997810364, "learning_rate": 1.911373179665863e-05, "loss": 0.02, "step": 1740 }, { "epoch": 0.14, "grad_norm": 1.3125584125518799, "learning_rate": 1.910479764138301e-05, "loss": 0.0202, "step": 1745 }, { "epoch": 0.14, "grad_norm": 0.6378400921821594, "learning_rate": 1.909586348610739e-05, "loss": 0.0289, "step": 1750 }, { "epoch": 0.14, "grad_norm": 1.118848443031311, "learning_rate": 1.908692933083177e-05, "loss": 0.0242, "step": 1755 }, { "epoch": 0.14, "grad_norm": 0.8065999746322632, "learning_rate": 1.9077995175556153e-05, "loss": 0.0277, "step": 1760 }, { "epoch": 0.14, "grad_norm": 0.9087148904800415, "learning_rate": 1.9069061020280534e-05, "loss": 0.0239, "step": 1765 }, { "epoch": 0.14, "grad_norm": 0.6891593933105469, "learning_rate": 1.9060126865004915e-05, "loss": 0.0217, "step": 1770 }, { "epoch": 0.14, "grad_norm": 0.7434559464454651, "learning_rate": 1.9051192709729296e-05, "loss": 0.0145, "step": 1775 }, { "epoch": 0.14, "grad_norm": 0.585472047328949, "learning_rate": 1.904225855445368e-05, "loss": 0.0287, "step": 1780 }, { "epoch": 0.14, "grad_norm": 1.561167597770691, "learning_rate": 1.9033324399178058e-05, "loss": 0.0206, "step": 1785 }, { "epoch": 0.14, "grad_norm": 0.9487882852554321, "learning_rate": 1.902439024390244e-05, "loss": 0.0212, "step": 1790 }, { "epoch": 0.14, "grad_norm": 0.7181324362754822, "learning_rate": 1.9015456088626823e-05, "loss": 0.0241, "step": 1795 }, { "epoch": 0.14, "grad_norm": 0.5083106160163879, "learning_rate": 1.9006521933351204e-05, "loss": 0.0159, "step": 1800 }, { "epoch": 0.14, "eval_loss": 0.030221089720726013, "eval_pearson_cosine": 0.8857723004603547, "eval_pearson_dot": 0.874834502431298, "eval_pearson_euclidean": 0.8749650474368084, "eval_pearson_manhattan": 0.8746243487782971, "eval_pearson_max": 0.8857723004603547, "eval_runtime": 426.1522, "eval_samples_per_second": 1.173, "eval_spearman_cosine": 0.9000108960435841, "eval_spearman_dot": 0.8859179436717747, "eval_spearman_euclidean": 0.9029959319837279, "eval_spearman_manhattan": 0.9019226476905907, "eval_spearman_max": 0.9029959319837279, "eval_steps_per_second": 1.173, "step": 1800 }, { "epoch": 0.15, "grad_norm": 0.43535247445106506, "learning_rate": 1.8997587778075582e-05, "loss": 0.0153, "step": 1805 }, { "epoch": 0.15, "grad_norm": 2.1544151306152344, "learning_rate": 1.8988653622799966e-05, "loss": 0.0241, "step": 1810 }, { "epoch": 0.15, "grad_norm": 1.9339686632156372, "learning_rate": 1.8979719467524347e-05, "loss": 0.0439, "step": 1815 }, { "epoch": 0.15, "grad_norm": 0.8600350618362427, "learning_rate": 1.897078531224873e-05, "loss": 0.0268, "step": 1820 }, { "epoch": 0.15, "grad_norm": 0.7343426942825317, "learning_rate": 1.896185115697311e-05, "loss": 0.0134, "step": 1825 }, { "epoch": 0.15, "grad_norm": 1.237486481666565, "learning_rate": 1.895291700169749e-05, "loss": 0.0387, "step": 1830 }, { "epoch": 0.15, "grad_norm": 0.8042169213294983, "learning_rate": 1.8943982846421875e-05, "loss": 0.0218, "step": 1835 }, { "epoch": 0.15, "grad_norm": 0.5267782211303711, "learning_rate": 1.8935048691146253e-05, "loss": 0.0277, "step": 1840 }, { "epoch": 0.15, "grad_norm": 0.7091891169548035, "learning_rate": 1.8926114535870634e-05, "loss": 0.0196, "step": 1845 }, { "epoch": 0.15, "grad_norm": 0.3862394094467163, "learning_rate": 1.8917180380595018e-05, "loss": 0.0173, "step": 1850 }, { "epoch": 0.15, "grad_norm": 0.8204643726348877, "learning_rate": 1.89082462253194e-05, "loss": 0.0174, "step": 1855 }, { "epoch": 0.15, "grad_norm": 0.6407317519187927, "learning_rate": 1.8899312070043777e-05, "loss": 0.0152, "step": 1860 }, { "epoch": 0.15, "grad_norm": 0.6078154444694519, "learning_rate": 1.889037791476816e-05, "loss": 0.0287, "step": 1865 }, { "epoch": 0.15, "grad_norm": 0.8564119338989258, "learning_rate": 1.8881443759492542e-05, "loss": 0.0185, "step": 1870 }, { "epoch": 0.15, "grad_norm": 0.7499640583992004, "learning_rate": 1.8872509604216923e-05, "loss": 0.04, "step": 1875 }, { "epoch": 0.15, "grad_norm": 0.5337907671928406, "learning_rate": 1.8863575448941304e-05, "loss": 0.0157, "step": 1880 }, { "epoch": 0.15, "grad_norm": 0.525449275970459, "learning_rate": 1.8854641293665685e-05, "loss": 0.0252, "step": 1885 }, { "epoch": 0.15, "grad_norm": 1.069151759147644, "learning_rate": 1.8845707138390066e-05, "loss": 0.0261, "step": 1890 }, { "epoch": 0.15, "grad_norm": 0.5704050660133362, "learning_rate": 1.8836772983114447e-05, "loss": 0.0228, "step": 1895 }, { "epoch": 0.15, "grad_norm": 0.8183658719062805, "learning_rate": 1.882783882783883e-05, "loss": 0.0293, "step": 1900 }, { "epoch": 0.15, "grad_norm": 0.5661863088607788, "learning_rate": 1.881890467256321e-05, "loss": 0.0214, "step": 1905 }, { "epoch": 0.15, "grad_norm": 0.24864475429058075, "learning_rate": 1.8809970517287594e-05, "loss": 0.0169, "step": 1910 }, { "epoch": 0.15, "grad_norm": 0.622249960899353, "learning_rate": 1.880103636201197e-05, "loss": 0.0267, "step": 1915 }, { "epoch": 0.15, "grad_norm": 1.6661076545715332, "learning_rate": 1.8792102206736356e-05, "loss": 0.0287, "step": 1920 }, { "epoch": 0.15, "grad_norm": 0.737390398979187, "learning_rate": 1.8783168051460737e-05, "loss": 0.0179, "step": 1925 }, { "epoch": 0.16, "grad_norm": 0.6081177592277527, "learning_rate": 1.8774233896185118e-05, "loss": 0.0194, "step": 1930 }, { "epoch": 0.16, "grad_norm": 1.8305076360702515, "learning_rate": 1.87652997409095e-05, "loss": 0.0333, "step": 1935 }, { "epoch": 0.16, "grad_norm": 0.7024769186973572, "learning_rate": 1.875636558563388e-05, "loss": 0.0272, "step": 1940 }, { "epoch": 0.16, "grad_norm": 0.7568246722221375, "learning_rate": 1.874743143035826e-05, "loss": 0.0182, "step": 1945 }, { "epoch": 0.16, "grad_norm": 1.3228060007095337, "learning_rate": 1.8738497275082642e-05, "loss": 0.0223, "step": 1950 }, { "epoch": 0.16, "grad_norm": 1.272178053855896, "learning_rate": 1.8729563119807023e-05, "loss": 0.0299, "step": 1955 }, { "epoch": 0.16, "grad_norm": 0.30096063017845154, "learning_rate": 1.8720628964531404e-05, "loss": 0.017, "step": 1960 }, { "epoch": 0.16, "grad_norm": 0.5879131555557251, "learning_rate": 1.871169480925579e-05, "loss": 0.0265, "step": 1965 }, { "epoch": 0.16, "grad_norm": 0.6623579263687134, "learning_rate": 1.8702760653980166e-05, "loss": 0.024, "step": 1970 }, { "epoch": 0.16, "grad_norm": 0.7402109503746033, "learning_rate": 1.8693826498704547e-05, "loss": 0.0248, "step": 1975 }, { "epoch": 0.16, "grad_norm": 0.46998223662376404, "learning_rate": 1.8684892343428932e-05, "loss": 0.0367, "step": 1980 }, { "epoch": 0.16, "grad_norm": 0.8414738178253174, "learning_rate": 1.8675958188153313e-05, "loss": 0.0352, "step": 1985 }, { "epoch": 0.16, "grad_norm": 0.5672045350074768, "learning_rate": 1.866702403287769e-05, "loss": 0.0286, "step": 1990 }, { "epoch": 0.16, "grad_norm": 0.4115860164165497, "learning_rate": 1.8658089877602075e-05, "loss": 0.0235, "step": 1995 }, { "epoch": 0.16, "grad_norm": 0.6932913064956665, "learning_rate": 1.8649155722326456e-05, "loss": 0.0128, "step": 2000 }, { "epoch": 0.16, "grad_norm": 0.6872785687446594, "learning_rate": 1.8640221567050837e-05, "loss": 0.0205, "step": 2005 }, { "epoch": 0.16, "grad_norm": 0.590305745601654, "learning_rate": 1.8631287411775218e-05, "loss": 0.0207, "step": 2010 }, { "epoch": 0.16, "grad_norm": 0.5067042708396912, "learning_rate": 1.86223532564996e-05, "loss": 0.038, "step": 2015 }, { "epoch": 0.16, "grad_norm": 0.46255382895469666, "learning_rate": 1.861341910122398e-05, "loss": 0.0237, "step": 2020 }, { "epoch": 0.16, "grad_norm": 0.4533158540725708, "learning_rate": 1.860448494594836e-05, "loss": 0.0197, "step": 2025 }, { "epoch": 0.16, "grad_norm": 0.919363796710968, "learning_rate": 1.8595550790672742e-05, "loss": 0.023, "step": 2030 }, { "epoch": 0.16, "grad_norm": 0.42132195830345154, "learning_rate": 1.8586616635397127e-05, "loss": 0.0119, "step": 2035 }, { "epoch": 0.16, "grad_norm": 0.9792714715003967, "learning_rate": 1.8577682480121508e-05, "loss": 0.022, "step": 2040 }, { "epoch": 0.16, "grad_norm": 1.0622044801712036, "learning_rate": 1.8568748324845885e-05, "loss": 0.0191, "step": 2045 }, { "epoch": 0.16, "grad_norm": 0.3776015341281891, "learning_rate": 1.855981416957027e-05, "loss": 0.0173, "step": 2050 }, { "epoch": 0.17, "grad_norm": 0.7663435935974121, "learning_rate": 1.855088001429465e-05, "loss": 0.0257, "step": 2055 }, { "epoch": 0.17, "grad_norm": 0.5264787673950195, "learning_rate": 1.8541945859019032e-05, "loss": 0.023, "step": 2060 }, { "epoch": 0.17, "grad_norm": 1.0220805406570435, "learning_rate": 1.8533011703743413e-05, "loss": 0.0215, "step": 2065 }, { "epoch": 0.17, "grad_norm": 0.47214239835739136, "learning_rate": 1.8524077548467794e-05, "loss": 0.0368, "step": 2070 }, { "epoch": 0.17, "grad_norm": 0.40313026309013367, "learning_rate": 1.8515143393192175e-05, "loss": 0.0199, "step": 2075 }, { "epoch": 0.17, "grad_norm": 0.7139260768890381, "learning_rate": 1.8506209237916556e-05, "loss": 0.0182, "step": 2080 }, { "epoch": 0.17, "grad_norm": 1.0911624431610107, "learning_rate": 1.8497275082640937e-05, "loss": 0.0215, "step": 2085 }, { "epoch": 0.17, "grad_norm": 0.5224277973175049, "learning_rate": 1.8488340927365318e-05, "loss": 0.0203, "step": 2090 }, { "epoch": 0.17, "grad_norm": 0.6372708678245544, "learning_rate": 1.84794067720897e-05, "loss": 0.0363, "step": 2095 }, { "epoch": 0.17, "grad_norm": 1.0599619150161743, "learning_rate": 1.847047261681408e-05, "loss": 0.0343, "step": 2100 }, { "epoch": 0.17, "eval_loss": 0.023791618645191193, "eval_pearson_cosine": 0.8906459396898687, "eval_pearson_dot": 0.8866654230263398, "eval_pearson_euclidean": 0.8789258913375184, "eval_pearson_manhattan": 0.8774504409979331, "eval_pearson_max": 0.8906459396898687, "eval_runtime": 425.9801, "eval_samples_per_second": 1.174, "eval_spearman_cosine": 0.9069881079524318, "eval_spearman_dot": 0.8997297109188437, "eval_spearman_euclidean": 0.9081995367981472, "eval_spearman_manhattan": 0.90537923751695, "eval_spearman_max": 0.9081995367981472, "eval_steps_per_second": 1.174, "step": 2100 }, { "epoch": 0.17, "grad_norm": 1.0371475219726562, "learning_rate": 1.8461538461538465e-05, "loss": 0.0205, "step": 2105 }, { "epoch": 0.17, "grad_norm": 0.7635193467140198, "learning_rate": 1.8452604306262846e-05, "loss": 0.0236, "step": 2110 }, { "epoch": 0.17, "grad_norm": 0.4593610167503357, "learning_rate": 1.8443670150987227e-05, "loss": 0.0223, "step": 2115 }, { "epoch": 0.17, "grad_norm": 0.9037809371948242, "learning_rate": 1.8434735995711608e-05, "loss": 0.0348, "step": 2120 }, { "epoch": 0.17, "grad_norm": 1.4961464405059814, "learning_rate": 1.842580184043599e-05, "loss": 0.0278, "step": 2125 }, { "epoch": 0.17, "grad_norm": 1.0911996364593506, "learning_rate": 1.841686768516037e-05, "loss": 0.0124, "step": 2130 }, { "epoch": 0.17, "grad_norm": 0.8265781998634338, "learning_rate": 1.840793352988475e-05, "loss": 0.0336, "step": 2135 }, { "epoch": 0.17, "grad_norm": 1.2838935852050781, "learning_rate": 1.8398999374609132e-05, "loss": 0.0271, "step": 2140 }, { "epoch": 0.17, "grad_norm": 0.8582805395126343, "learning_rate": 1.8390065219333513e-05, "loss": 0.0466, "step": 2145 }, { "epoch": 0.17, "grad_norm": 2.2625269889831543, "learning_rate": 1.8381131064057894e-05, "loss": 0.0338, "step": 2150 }, { "epoch": 0.17, "grad_norm": 0.7238175272941589, "learning_rate": 1.8372196908782275e-05, "loss": 0.0149, "step": 2155 }, { "epoch": 0.17, "grad_norm": 0.4642048478126526, "learning_rate": 1.8363262753506656e-05, "loss": 0.017, "step": 2160 }, { "epoch": 0.17, "grad_norm": 0.6727991700172424, "learning_rate": 1.835432859823104e-05, "loss": 0.0302, "step": 2165 }, { "epoch": 0.17, "grad_norm": 1.1912602186203003, "learning_rate": 1.834539444295542e-05, "loss": 0.0244, "step": 2170 }, { "epoch": 0.17, "grad_norm": 0.8099949955940247, "learning_rate": 1.83364602876798e-05, "loss": 0.013, "step": 2175 }, { "epoch": 0.18, "grad_norm": 0.6302483081817627, "learning_rate": 1.8327526132404184e-05, "loss": 0.0245, "step": 2180 }, { "epoch": 0.18, "grad_norm": 0.44703614711761475, "learning_rate": 1.8318591977128565e-05, "loss": 0.0179, "step": 2185 }, { "epoch": 0.18, "grad_norm": 0.4171731472015381, "learning_rate": 1.8309657821852946e-05, "loss": 0.0155, "step": 2190 }, { "epoch": 0.18, "grad_norm": 0.9061782360076904, "learning_rate": 1.8300723666577327e-05, "loss": 0.0222, "step": 2195 }, { "epoch": 0.18, "grad_norm": 0.616051197052002, "learning_rate": 1.8291789511301708e-05, "loss": 0.0167, "step": 2200 }, { "epoch": 0.18, "grad_norm": 1.4685133695602417, "learning_rate": 1.828285535602609e-05, "loss": 0.0259, "step": 2205 }, { "epoch": 0.18, "grad_norm": 0.9447664022445679, "learning_rate": 1.827392120075047e-05, "loss": 0.0351, "step": 2210 }, { "epoch": 0.18, "grad_norm": 1.0720714330673218, "learning_rate": 1.826498704547485e-05, "loss": 0.0164, "step": 2215 }, { "epoch": 0.18, "grad_norm": 0.7534575462341309, "learning_rate": 1.8256052890199235e-05, "loss": 0.0161, "step": 2220 }, { "epoch": 0.18, "grad_norm": 0.8038508892059326, "learning_rate": 1.8247118734923613e-05, "loss": 0.0205, "step": 2225 }, { "epoch": 0.18, "grad_norm": 1.0298678874969482, "learning_rate": 1.8238184579647994e-05, "loss": 0.019, "step": 2230 }, { "epoch": 0.18, "grad_norm": 1.3235474824905396, "learning_rate": 1.822925042437238e-05, "loss": 0.0265, "step": 2235 }, { "epoch": 0.18, "grad_norm": 0.6873222589492798, "learning_rate": 1.822031626909676e-05, "loss": 0.0258, "step": 2240 }, { "epoch": 0.18, "grad_norm": 0.4005903899669647, "learning_rate": 1.821138211382114e-05, "loss": 0.0185, "step": 2245 }, { "epoch": 0.18, "grad_norm": 0.47836175560951233, "learning_rate": 1.820244795854552e-05, "loss": 0.0229, "step": 2250 }, { "epoch": 0.18, "grad_norm": 0.653409481048584, "learning_rate": 1.8193513803269903e-05, "loss": 0.0155, "step": 2255 }, { "epoch": 0.18, "grad_norm": 0.2501707077026367, "learning_rate": 1.8184579647994284e-05, "loss": 0.0253, "step": 2260 }, { "epoch": 0.18, "grad_norm": 0.5146343111991882, "learning_rate": 1.8175645492718665e-05, "loss": 0.0355, "step": 2265 }, { "epoch": 0.18, "grad_norm": 0.7069603204727173, "learning_rate": 1.8166711337443046e-05, "loss": 0.0245, "step": 2270 }, { "epoch": 0.18, "grad_norm": 0.3391192853450775, "learning_rate": 1.8157777182167427e-05, "loss": 0.0125, "step": 2275 }, { "epoch": 0.18, "grad_norm": 1.3187044858932495, "learning_rate": 1.8148843026891808e-05, "loss": 0.0314, "step": 2280 }, { "epoch": 0.18, "grad_norm": 0.3739178478717804, "learning_rate": 1.813990887161619e-05, "loss": 0.0254, "step": 2285 }, { "epoch": 0.18, "grad_norm": 0.9314795732498169, "learning_rate": 1.8130974716340573e-05, "loss": 0.0264, "step": 2290 }, { "epoch": 0.18, "grad_norm": 0.9376474022865295, "learning_rate": 1.8122040561064954e-05, "loss": 0.0172, "step": 2295 }, { "epoch": 0.18, "grad_norm": 0.9479859471321106, "learning_rate": 1.8113106405789332e-05, "loss": 0.0321, "step": 2300 }, { "epoch": 0.19, "grad_norm": 0.8741356730461121, "learning_rate": 1.8104172250513716e-05, "loss": 0.0216, "step": 2305 }, { "epoch": 0.19, "grad_norm": 0.7016437649726868, "learning_rate": 1.8095238095238097e-05, "loss": 0.0181, "step": 2310 }, { "epoch": 0.19, "grad_norm": 1.6568303108215332, "learning_rate": 1.808630393996248e-05, "loss": 0.0374, "step": 2315 }, { "epoch": 0.19, "grad_norm": 0.4468625783920288, "learning_rate": 1.807736978468686e-05, "loss": 0.021, "step": 2320 }, { "epoch": 0.19, "grad_norm": 0.8952460885047913, "learning_rate": 1.806843562941124e-05, "loss": 0.0215, "step": 2325 }, { "epoch": 0.19, "grad_norm": 0.5987122654914856, "learning_rate": 1.805950147413562e-05, "loss": 0.0182, "step": 2330 }, { "epoch": 0.19, "grad_norm": 0.8986082077026367, "learning_rate": 1.8050567318860003e-05, "loss": 0.0224, "step": 2335 }, { "epoch": 0.19, "grad_norm": 0.4435628652572632, "learning_rate": 1.8041633163584384e-05, "loss": 0.0246, "step": 2340 }, { "epoch": 0.19, "grad_norm": 0.5219196081161499, "learning_rate": 1.8032699008308765e-05, "loss": 0.0222, "step": 2345 }, { "epoch": 0.19, "grad_norm": 1.8987548351287842, "learning_rate": 1.802376485303315e-05, "loss": 0.0416, "step": 2350 }, { "epoch": 0.19, "grad_norm": 0.3122328817844391, "learning_rate": 1.8014830697757527e-05, "loss": 0.0162, "step": 2355 }, { "epoch": 0.19, "grad_norm": 0.45380455255508423, "learning_rate": 1.8005896542481908e-05, "loss": 0.0348, "step": 2360 }, { "epoch": 0.19, "grad_norm": 0.7789851427078247, "learning_rate": 1.7996962387206292e-05, "loss": 0.0207, "step": 2365 }, { "epoch": 0.19, "grad_norm": 0.7289481163024902, "learning_rate": 1.7988028231930673e-05, "loss": 0.0154, "step": 2370 }, { "epoch": 0.19, "grad_norm": 1.5287164449691772, "learning_rate": 1.7979094076655054e-05, "loss": 0.0306, "step": 2375 }, { "epoch": 0.19, "grad_norm": 1.236247181892395, "learning_rate": 1.7970159921379435e-05, "loss": 0.0346, "step": 2380 }, { "epoch": 0.19, "grad_norm": 0.6785498857498169, "learning_rate": 1.7961225766103816e-05, "loss": 0.0191, "step": 2385 }, { "epoch": 0.19, "grad_norm": 0.7532624006271362, "learning_rate": 1.7952291610828197e-05, "loss": 0.0247, "step": 2390 }, { "epoch": 0.19, "grad_norm": 0.8474083542823792, "learning_rate": 1.794335745555258e-05, "loss": 0.028, "step": 2395 }, { "epoch": 0.19, "grad_norm": 0.8650068640708923, "learning_rate": 1.793442330027696e-05, "loss": 0.0234, "step": 2400 }, { "epoch": 0.19, "eval_loss": 0.02186727523803711, "eval_pearson_cosine": 0.9016899300379176, "eval_pearson_dot": 0.8979110419976675, "eval_pearson_euclidean": 0.892515905459987, "eval_pearson_manhattan": 0.8905747154959989, "eval_pearson_max": 0.9016899300379176, "eval_runtime": 426.0444, "eval_samples_per_second": 1.174, "eval_spearman_cosine": 0.9066876267505071, "eval_spearman_dot": 0.9012341329365315, "eval_spearman_euclidean": 0.9076169104676419, "eval_spearman_manhattan": 0.9052689330757322, "eval_spearman_max": 0.9076169104676419, "eval_steps_per_second": 1.174, "step": 2400 }, { "epoch": 0.19, "grad_norm": 0.7486416101455688, "learning_rate": 1.7925489145001344e-05, "loss": 0.0158, "step": 2405 }, { "epoch": 0.19, "grad_norm": 0.20279864966869354, "learning_rate": 1.791655498972572e-05, "loss": 0.0141, "step": 2410 }, { "epoch": 0.19, "grad_norm": 0.4669971466064453, "learning_rate": 1.7907620834450103e-05, "loss": 0.019, "step": 2415 }, { "epoch": 0.19, "grad_norm": 0.6175631880760193, "learning_rate": 1.7898686679174487e-05, "loss": 0.0148, "step": 2420 }, { "epoch": 0.19, "grad_norm": 0.5767547488212585, "learning_rate": 1.7889752523898868e-05, "loss": 0.0252, "step": 2425 }, { "epoch": 0.2, "grad_norm": 0.5371144413948059, "learning_rate": 1.7880818368623246e-05, "loss": 0.0121, "step": 2430 }, { "epoch": 0.2, "grad_norm": 0.6145671606063843, "learning_rate": 1.787188421334763e-05, "loss": 0.0195, "step": 2435 }, { "epoch": 0.2, "grad_norm": 1.262970209121704, "learning_rate": 1.786295005807201e-05, "loss": 0.0298, "step": 2440 }, { "epoch": 0.2, "grad_norm": 0.3572503328323364, "learning_rate": 1.7854015902796392e-05, "loss": 0.0152, "step": 2445 }, { "epoch": 0.2, "grad_norm": 0.6049114465713501, "learning_rate": 1.7845081747520773e-05, "loss": 0.0153, "step": 2450 }, { "epoch": 0.2, "grad_norm": 0.40164369344711304, "learning_rate": 1.7836147592245154e-05, "loss": 0.0171, "step": 2455 }, { "epoch": 0.2, "grad_norm": 0.33443671464920044, "learning_rate": 1.7827213436969535e-05, "loss": 0.0174, "step": 2460 }, { "epoch": 0.2, "grad_norm": 0.37140563130378723, "learning_rate": 1.7818279281693916e-05, "loss": 0.0147, "step": 2465 }, { "epoch": 0.2, "grad_norm": 0.44236427545547485, "learning_rate": 1.7809345126418297e-05, "loss": 0.019, "step": 2470 }, { "epoch": 0.2, "grad_norm": 1.1675889492034912, "learning_rate": 1.7800410971142682e-05, "loss": 0.0196, "step": 2475 }, { "epoch": 0.2, "grad_norm": 0.7172742486000061, "learning_rate": 1.7791476815867063e-05, "loss": 0.0262, "step": 2480 }, { "epoch": 0.2, "grad_norm": 0.9804342985153198, "learning_rate": 1.778254266059144e-05, "loss": 0.0337, "step": 2485 }, { "epoch": 0.2, "grad_norm": 1.989030361175537, "learning_rate": 1.7773608505315825e-05, "loss": 0.0238, "step": 2490 }, { "epoch": 0.2, "grad_norm": 0.754467785358429, "learning_rate": 1.7764674350040206e-05, "loss": 0.0224, "step": 2495 }, { "epoch": 0.2, "grad_norm": 0.9083816409111023, "learning_rate": 1.7755740194764587e-05, "loss": 0.0165, "step": 2500 }, { "epoch": 0.2, "grad_norm": 0.5486024022102356, "learning_rate": 1.7746806039488968e-05, "loss": 0.0173, "step": 2505 }, { "epoch": 0.2, "grad_norm": 0.4576135575771332, "learning_rate": 1.773787188421335e-05, "loss": 0.0177, "step": 2510 }, { "epoch": 0.2, "grad_norm": 0.9265009760856628, "learning_rate": 1.772893772893773e-05, "loss": 0.0216, "step": 2515 }, { "epoch": 0.2, "grad_norm": 0.506006121635437, "learning_rate": 1.772000357366211e-05, "loss": 0.0135, "step": 2520 }, { "epoch": 0.2, "grad_norm": 0.2627745270729065, "learning_rate": 1.7711069418386492e-05, "loss": 0.0244, "step": 2525 }, { "epoch": 0.2, "grad_norm": 0.7158712148666382, "learning_rate": 1.7702135263110873e-05, "loss": 0.0238, "step": 2530 }, { "epoch": 0.2, "grad_norm": 0.8237607479095459, "learning_rate": 1.7693201107835258e-05, "loss": 0.0272, "step": 2535 }, { "epoch": 0.2, "grad_norm": 0.266671359539032, "learning_rate": 1.7684266952559635e-05, "loss": 0.0202, "step": 2540 }, { "epoch": 0.2, "grad_norm": 0.6004403829574585, "learning_rate": 1.7675332797284016e-05, "loss": 0.0414, "step": 2545 }, { "epoch": 0.21, "grad_norm": 0.9549585580825806, "learning_rate": 1.76663986420084e-05, "loss": 0.0209, "step": 2550 }, { "epoch": 0.21, "grad_norm": 0.9587855935096741, "learning_rate": 1.7657464486732782e-05, "loss": 0.0239, "step": 2555 }, { "epoch": 0.21, "grad_norm": 0.5804222226142883, "learning_rate": 1.7648530331457163e-05, "loss": 0.0183, "step": 2560 }, { "epoch": 0.21, "grad_norm": 0.506087601184845, "learning_rate": 1.7639596176181544e-05, "loss": 0.0202, "step": 2565 }, { "epoch": 0.21, "grad_norm": 0.8496732711791992, "learning_rate": 1.7630662020905925e-05, "loss": 0.0181, "step": 2570 }, { "epoch": 0.21, "grad_norm": 0.3317871391773224, "learning_rate": 1.7621727865630306e-05, "loss": 0.0218, "step": 2575 }, { "epoch": 0.21, "grad_norm": 0.3938535451889038, "learning_rate": 1.7612793710354687e-05, "loss": 0.0112, "step": 2580 }, { "epoch": 0.21, "grad_norm": 1.3293476104736328, "learning_rate": 1.7603859555079068e-05, "loss": 0.0181, "step": 2585 }, { "epoch": 0.21, "grad_norm": 0.28195351362228394, "learning_rate": 1.7594925399803452e-05, "loss": 0.0135, "step": 2590 }, { "epoch": 0.21, "grad_norm": 0.7966617941856384, "learning_rate": 1.758599124452783e-05, "loss": 0.0145, "step": 2595 }, { "epoch": 0.21, "grad_norm": 0.5484746694564819, "learning_rate": 1.757705708925221e-05, "loss": 0.0207, "step": 2600 }, { "epoch": 0.21, "grad_norm": 0.576448380947113, "learning_rate": 1.7568122933976596e-05, "loss": 0.0127, "step": 2605 }, { "epoch": 0.21, "grad_norm": 0.28207531571388245, "learning_rate": 1.7559188778700977e-05, "loss": 0.0132, "step": 2610 }, { "epoch": 0.21, "grad_norm": 0.29697030782699585, "learning_rate": 1.7550254623425354e-05, "loss": 0.0127, "step": 2615 }, { "epoch": 0.21, "grad_norm": 1.122375249862671, "learning_rate": 1.754132046814974e-05, "loss": 0.0297, "step": 2620 }, { "epoch": 0.21, "grad_norm": 0.5798119902610779, "learning_rate": 1.753238631287412e-05, "loss": 0.0171, "step": 2625 }, { "epoch": 0.21, "grad_norm": 0.6546443104743958, "learning_rate": 1.75234521575985e-05, "loss": 0.0436, "step": 2630 }, { "epoch": 0.21, "grad_norm": 0.8825979828834534, "learning_rate": 1.7514518002322882e-05, "loss": 0.0138, "step": 2635 }, { "epoch": 0.21, "grad_norm": 0.46849188208580017, "learning_rate": 1.7505583847047263e-05, "loss": 0.0328, "step": 2640 }, { "epoch": 0.21, "grad_norm": 1.324487566947937, "learning_rate": 1.7496649691771644e-05, "loss": 0.0241, "step": 2645 }, { "epoch": 0.21, "grad_norm": 0.5214132070541382, "learning_rate": 1.7487715536496025e-05, "loss": 0.0344, "step": 2650 }, { "epoch": 0.21, "grad_norm": 0.791459858417511, "learning_rate": 1.7478781381220406e-05, "loss": 0.0167, "step": 2655 }, { "epoch": 0.21, "grad_norm": 0.888538658618927, "learning_rate": 1.746984722594479e-05, "loss": 0.0216, "step": 2660 }, { "epoch": 0.21, "grad_norm": 0.31034013628959656, "learning_rate": 1.746091307066917e-05, "loss": 0.0327, "step": 2665 }, { "epoch": 0.21, "grad_norm": 0.4363711476325989, "learning_rate": 1.745197891539355e-05, "loss": 0.0203, "step": 2670 }, { "epoch": 0.22, "grad_norm": 1.0624545812606812, "learning_rate": 1.7443044760117934e-05, "loss": 0.0216, "step": 2675 }, { "epoch": 0.22, "grad_norm": 1.4411804676055908, "learning_rate": 1.7434110604842315e-05, "loss": 0.0294, "step": 2680 }, { "epoch": 0.22, "grad_norm": 0.7424313426017761, "learning_rate": 1.7425176449566696e-05, "loss": 0.0233, "step": 2685 }, { "epoch": 0.22, "grad_norm": 0.6678118705749512, "learning_rate": 1.7416242294291077e-05, "loss": 0.0129, "step": 2690 }, { "epoch": 0.22, "grad_norm": 0.43497857451438904, "learning_rate": 1.7407308139015458e-05, "loss": 0.0272, "step": 2695 }, { "epoch": 0.22, "grad_norm": 0.3903897702693939, "learning_rate": 1.739837398373984e-05, "loss": 0.0209, "step": 2700 }, { "epoch": 0.22, "eval_loss": 0.027375079691410065, "eval_pearson_cosine": 0.9001051843718942, "eval_pearson_dot": 0.9006651139179915, "eval_pearson_euclidean": 0.8930651682478097, "eval_pearson_manhattan": 0.891765654050478, "eval_pearson_max": 0.9006651139179915, "eval_runtime": 426.2183, "eval_samples_per_second": 1.173, "eval_spearman_cosine": 0.9177515750063, "eval_spearman_dot": 0.9167129468517873, "eval_spearman_euclidean": 0.9167275389101556, "eval_spearman_manhattan": 0.9159309277237108, "eval_spearman_max": 0.9177515750063, "eval_steps_per_second": 1.173, "step": 2700 }, { "epoch": 0.22, "grad_norm": 1.3496129512786865, "learning_rate": 1.738943982846422e-05, "loss": 0.0259, "step": 2705 }, { "epoch": 0.22, "grad_norm": 1.1606944799423218, "learning_rate": 1.73805056731886e-05, "loss": 0.0181, "step": 2710 }, { "epoch": 0.22, "grad_norm": 0.5323413610458374, "learning_rate": 1.7371571517912982e-05, "loss": 0.0137, "step": 2715 }, { "epoch": 0.22, "grad_norm": 0.5756454467773438, "learning_rate": 1.7362637362637363e-05, "loss": 0.0181, "step": 2720 }, { "epoch": 0.22, "grad_norm": 0.3727392256259918, "learning_rate": 1.7353703207361744e-05, "loss": 0.0123, "step": 2725 }, { "epoch": 0.22, "grad_norm": 0.441544771194458, "learning_rate": 1.7344769052086125e-05, "loss": 0.0157, "step": 2730 }, { "epoch": 0.22, "grad_norm": 0.6806672811508179, "learning_rate": 1.733583489681051e-05, "loss": 0.0196, "step": 2735 }, { "epoch": 0.22, "grad_norm": 0.5537866353988647, "learning_rate": 1.732690074153489e-05, "loss": 0.0163, "step": 2740 }, { "epoch": 0.22, "grad_norm": 0.5595599412918091, "learning_rate": 1.731796658625927e-05, "loss": 0.0229, "step": 2745 }, { "epoch": 0.22, "grad_norm": 0.48301759362220764, "learning_rate": 1.7309032430983652e-05, "loss": 0.0197, "step": 2750 }, { "epoch": 0.22, "grad_norm": 1.2349361181259155, "learning_rate": 1.7300098275708034e-05, "loss": 0.0198, "step": 2755 }, { "epoch": 0.22, "grad_norm": 0.3702278137207031, "learning_rate": 1.7291164120432415e-05, "loss": 0.0165, "step": 2760 }, { "epoch": 0.22, "grad_norm": 0.5885152220726013, "learning_rate": 1.7282229965156796e-05, "loss": 0.0126, "step": 2765 }, { "epoch": 0.22, "grad_norm": 1.2845672369003296, "learning_rate": 1.7273295809881177e-05, "loss": 0.0232, "step": 2770 }, { "epoch": 0.22, "grad_norm": 1.0059493780136108, "learning_rate": 1.7264361654605558e-05, "loss": 0.0102, "step": 2775 }, { "epoch": 0.22, "grad_norm": 0.621356189250946, "learning_rate": 1.725542749932994e-05, "loss": 0.0149, "step": 2780 }, { "epoch": 0.22, "grad_norm": 0.8701572418212891, "learning_rate": 1.724649334405432e-05, "loss": 0.0147, "step": 2785 }, { "epoch": 0.22, "grad_norm": 0.35073745250701904, "learning_rate": 1.7237559188778704e-05, "loss": 0.0176, "step": 2790 }, { "epoch": 0.22, "grad_norm": 0.6895755529403687, "learning_rate": 1.7228625033503085e-05, "loss": 0.0212, "step": 2795 }, { "epoch": 0.23, "grad_norm": 0.9424052834510803, "learning_rate": 1.7219690878227463e-05, "loss": 0.0213, "step": 2800 }, { "epoch": 0.23, "grad_norm": 0.45782536268234253, "learning_rate": 1.7210756722951847e-05, "loss": 0.0122, "step": 2805 }, { "epoch": 0.23, "grad_norm": 0.9347144365310669, "learning_rate": 1.720182256767623e-05, "loss": 0.0221, "step": 2810 }, { "epoch": 0.23, "grad_norm": 0.4957163333892822, "learning_rate": 1.719288841240061e-05, "loss": 0.0154, "step": 2815 }, { "epoch": 0.23, "grad_norm": 0.40040886402130127, "learning_rate": 1.718395425712499e-05, "loss": 0.018, "step": 2820 }, { "epoch": 0.23, "grad_norm": 0.4707423150539398, "learning_rate": 1.717502010184937e-05, "loss": 0.0122, "step": 2825 }, { "epoch": 0.23, "grad_norm": 0.6660599708557129, "learning_rate": 1.7166085946573752e-05, "loss": 0.0103, "step": 2830 }, { "epoch": 0.23, "grad_norm": 0.5606631636619568, "learning_rate": 1.7157151791298134e-05, "loss": 0.0246, "step": 2835 }, { "epoch": 0.23, "grad_norm": 1.0515074729919434, "learning_rate": 1.7148217636022515e-05, "loss": 0.0173, "step": 2840 }, { "epoch": 0.23, "grad_norm": 1.8896968364715576, "learning_rate": 1.71392834807469e-05, "loss": 0.0357, "step": 2845 }, { "epoch": 0.23, "grad_norm": 0.8183419108390808, "learning_rate": 1.7130349325471277e-05, "loss": 0.0125, "step": 2850 }, { "epoch": 0.23, "grad_norm": 0.28626278042793274, "learning_rate": 1.7121415170195658e-05, "loss": 0.011, "step": 2855 }, { "epoch": 0.23, "grad_norm": 0.5537832975387573, "learning_rate": 1.7112481014920042e-05, "loss": 0.0155, "step": 2860 }, { "epoch": 0.23, "grad_norm": 0.3938451111316681, "learning_rate": 1.7103546859644423e-05, "loss": 0.0163, "step": 2865 }, { "epoch": 0.23, "grad_norm": 0.27609241008758545, "learning_rate": 1.7094612704368804e-05, "loss": 0.0193, "step": 2870 }, { "epoch": 0.23, "grad_norm": 1.3002891540527344, "learning_rate": 1.7085678549093185e-05, "loss": 0.0142, "step": 2875 }, { "epoch": 0.23, "grad_norm": 1.8881809711456299, "learning_rate": 1.7076744393817566e-05, "loss": 0.0296, "step": 2880 }, { "epoch": 0.23, "grad_norm": 0.99460768699646, "learning_rate": 1.7067810238541947e-05, "loss": 0.0157, "step": 2885 }, { "epoch": 0.23, "grad_norm": 1.14095938205719, "learning_rate": 1.705887608326633e-05, "loss": 0.0164, "step": 2890 }, { "epoch": 0.23, "grad_norm": 0.2840583026409149, "learning_rate": 1.704994192799071e-05, "loss": 0.0139, "step": 2895 }, { "epoch": 0.23, "grad_norm": 0.3837971091270447, "learning_rate": 1.704100777271509e-05, "loss": 0.016, "step": 2900 }, { "epoch": 0.23, "grad_norm": 0.4202588200569153, "learning_rate": 1.703207361743947e-05, "loss": 0.0161, "step": 2905 }, { "epoch": 0.23, "grad_norm": 0.9836485385894775, "learning_rate": 1.7023139462163853e-05, "loss": 0.0336, "step": 2910 }, { "epoch": 0.23, "grad_norm": 0.28357017040252686, "learning_rate": 1.7014205306888234e-05, "loss": 0.0151, "step": 2915 }, { "epoch": 0.23, "grad_norm": 0.3464276194572449, "learning_rate": 1.7005271151612618e-05, "loss": 0.0151, "step": 2920 }, { "epoch": 0.24, "grad_norm": 0.6251973509788513, "learning_rate": 1.6996336996336996e-05, "loss": 0.0275, "step": 2925 }, { "epoch": 0.24, "grad_norm": 2.1420838832855225, "learning_rate": 1.698740284106138e-05, "loss": 0.0339, "step": 2930 }, { "epoch": 0.24, "grad_norm": 0.43490684032440186, "learning_rate": 1.697846868578576e-05, "loss": 0.0124, "step": 2935 }, { "epoch": 0.24, "grad_norm": 0.576763391494751, "learning_rate": 1.6969534530510142e-05, "loss": 0.0162, "step": 2940 }, { "epoch": 0.24, "grad_norm": 0.8549007773399353, "learning_rate": 1.6960600375234523e-05, "loss": 0.0217, "step": 2945 }, { "epoch": 0.24, "grad_norm": 0.3743274509906769, "learning_rate": 1.6951666219958904e-05, "loss": 0.0165, "step": 2950 }, { "epoch": 0.24, "grad_norm": 0.7152208089828491, "learning_rate": 1.6942732064683285e-05, "loss": 0.0158, "step": 2955 }, { "epoch": 0.24, "grad_norm": 0.5613319873809814, "learning_rate": 1.6933797909407666e-05, "loss": 0.0232, "step": 2960 }, { "epoch": 0.24, "grad_norm": 0.3946327269077301, "learning_rate": 1.6924863754132047e-05, "loss": 0.0156, "step": 2965 }, { "epoch": 0.24, "grad_norm": 0.5846483707427979, "learning_rate": 1.691592959885643e-05, "loss": 0.0118, "step": 2970 }, { "epoch": 0.24, "grad_norm": 0.6644983291625977, "learning_rate": 1.6906995443580813e-05, "loss": 0.0156, "step": 2975 }, { "epoch": 0.24, "grad_norm": 0.7113365530967712, "learning_rate": 1.689806128830519e-05, "loss": 0.0131, "step": 2980 }, { "epoch": 0.24, "grad_norm": 0.7063427567481995, "learning_rate": 1.688912713302957e-05, "loss": 0.0196, "step": 2985 }, { "epoch": 0.24, "grad_norm": 0.8692206740379333, "learning_rate": 1.6880192977753956e-05, "loss": 0.02, "step": 2990 }, { "epoch": 0.24, "grad_norm": 0.6556200385093689, "learning_rate": 1.6871258822478337e-05, "loss": 0.0273, "step": 2995 }, { "epoch": 0.24, "grad_norm": 0.9636979103088379, "learning_rate": 1.6862324667202718e-05, "loss": 0.0192, "step": 3000 }, { "epoch": 0.24, "eval_loss": 0.0198129303753376, "eval_pearson_cosine": 0.9064648118738065, "eval_pearson_dot": 0.9033815480817863, "eval_pearson_euclidean": 0.8939116848670458, "eval_pearson_manhattan": 0.8927147647429707, "eval_pearson_max": 0.9064648118738065, "eval_runtime": 426.1029, "eval_samples_per_second": 1.173, "eval_spearman_cosine": 0.9222772251089003, "eval_spearman_dot": 0.9168591554366217, "eval_spearman_euclidean": 0.9222438169752679, "eval_spearman_manhattan": 0.9220269521078085, "eval_spearman_max": 0.9222772251089003, "eval_steps_per_second": 1.173, "step": 3000 }, { "epoch": 0.24, "grad_norm": 0.5286144614219666, "learning_rate": 1.68533905119271e-05, "loss": 0.0228, "step": 3005 }, { "epoch": 0.24, "grad_norm": 0.7542079091072083, "learning_rate": 1.684445635665148e-05, "loss": 0.0156, "step": 3010 }, { "epoch": 0.24, "grad_norm": 0.6064092516899109, "learning_rate": 1.683552220137586e-05, "loss": 0.0258, "step": 3015 }, { "epoch": 0.24, "grad_norm": 0.7690675258636475, "learning_rate": 1.6826588046100242e-05, "loss": 0.0201, "step": 3020 }, { "epoch": 0.24, "grad_norm": 0.4456731975078583, "learning_rate": 1.6817653890824623e-05, "loss": 0.0159, "step": 3025 }, { "epoch": 0.24, "grad_norm": 0.6881704330444336, "learning_rate": 1.6808719735549008e-05, "loss": 0.0191, "step": 3030 }, { "epoch": 0.24, "grad_norm": 1.515715479850769, "learning_rate": 1.6799785580273385e-05, "loss": 0.0272, "step": 3035 }, { "epoch": 0.24, "grad_norm": 0.9520334601402283, "learning_rate": 1.6790851424997766e-05, "loss": 0.0164, "step": 3040 }, { "epoch": 0.24, "grad_norm": 1.4020353555679321, "learning_rate": 1.678191726972215e-05, "loss": 0.0224, "step": 3045 }, { "epoch": 0.25, "grad_norm": 1.868922472000122, "learning_rate": 1.6772983114446532e-05, "loss": 0.0343, "step": 3050 }, { "epoch": 0.25, "grad_norm": 0.5014618635177612, "learning_rate": 1.676404895917091e-05, "loss": 0.0318, "step": 3055 }, { "epoch": 0.25, "grad_norm": 0.820292592048645, "learning_rate": 1.6755114803895294e-05, "loss": 0.0264, "step": 3060 }, { "epoch": 0.25, "grad_norm": 0.6107310056686401, "learning_rate": 1.6746180648619675e-05, "loss": 0.0188, "step": 3065 }, { "epoch": 0.25, "grad_norm": 0.3835718631744385, "learning_rate": 1.6737246493344056e-05, "loss": 0.012, "step": 3070 }, { "epoch": 0.25, "grad_norm": 0.46898066997528076, "learning_rate": 1.6728312338068437e-05, "loss": 0.0333, "step": 3075 }, { "epoch": 0.25, "grad_norm": 0.6556499004364014, "learning_rate": 1.6719378182792818e-05, "loss": 0.0184, "step": 3080 }, { "epoch": 0.25, "grad_norm": 0.8477081060409546, "learning_rate": 1.67104440275172e-05, "loss": 0.0327, "step": 3085 }, { "epoch": 0.25, "grad_norm": 0.6761387586593628, "learning_rate": 1.670150987224158e-05, "loss": 0.0178, "step": 3090 }, { "epoch": 0.25, "grad_norm": 0.4892427623271942, "learning_rate": 1.669257571696596e-05, "loss": 0.0266, "step": 3095 }, { "epoch": 0.25, "grad_norm": 0.7155781388282776, "learning_rate": 1.6683641561690342e-05, "loss": 0.0242, "step": 3100 }, { "epoch": 0.25, "grad_norm": 0.25152644515037537, "learning_rate": 1.6674707406414727e-05, "loss": 0.0102, "step": 3105 }, { "epoch": 0.25, "grad_norm": 0.8297644853591919, "learning_rate": 1.6665773251139104e-05, "loss": 0.0209, "step": 3110 }, { "epoch": 0.25, "grad_norm": 0.36103129386901855, "learning_rate": 1.665683909586349e-05, "loss": 0.0166, "step": 3115 }, { "epoch": 0.25, "grad_norm": 0.8184224367141724, "learning_rate": 1.664790494058787e-05, "loss": 0.0215, "step": 3120 }, { "epoch": 0.25, "grad_norm": 0.5235499739646912, "learning_rate": 1.663897078531225e-05, "loss": 0.0223, "step": 3125 }, { "epoch": 0.25, "grad_norm": 0.6001405715942383, "learning_rate": 1.6630036630036632e-05, "loss": 0.0177, "step": 3130 }, { "epoch": 0.25, "grad_norm": 0.5021240711212158, "learning_rate": 1.6621102474761013e-05, "loss": 0.0199, "step": 3135 }, { "epoch": 0.25, "grad_norm": 0.2629213035106659, "learning_rate": 1.6612168319485394e-05, "loss": 0.0375, "step": 3140 }, { "epoch": 0.25, "grad_norm": 0.6333101987838745, "learning_rate": 1.6603234164209775e-05, "loss": 0.0138, "step": 3145 }, { "epoch": 0.25, "grad_norm": 0.5239858627319336, "learning_rate": 1.6594300008934156e-05, "loss": 0.0211, "step": 3150 }, { "epoch": 0.25, "grad_norm": 0.5263399481773376, "learning_rate": 1.6585365853658537e-05, "loss": 0.0281, "step": 3155 }, { "epoch": 0.25, "grad_norm": 0.7547926902770996, "learning_rate": 1.657643169838292e-05, "loss": 0.0325, "step": 3160 }, { "epoch": 0.25, "grad_norm": 1.098939061164856, "learning_rate": 1.65674975431073e-05, "loss": 0.0136, "step": 3165 }, { "epoch": 0.25, "grad_norm": 0.5231286883354187, "learning_rate": 1.655856338783168e-05, "loss": 0.0241, "step": 3170 }, { "epoch": 0.26, "grad_norm": 0.6786376237869263, "learning_rate": 1.6549629232556064e-05, "loss": 0.0264, "step": 3175 }, { "epoch": 0.26, "grad_norm": 1.402637243270874, "learning_rate": 1.6540695077280446e-05, "loss": 0.0214, "step": 3180 }, { "epoch": 0.26, "grad_norm": 1.169063687324524, "learning_rate": 1.6531760922004823e-05, "loss": 0.038, "step": 3185 }, { "epoch": 0.26, "grad_norm": 0.9597774147987366, "learning_rate": 1.6522826766729208e-05, "loss": 0.0289, "step": 3190 }, { "epoch": 0.26, "grad_norm": 0.5611427426338196, "learning_rate": 1.651389261145359e-05, "loss": 0.0155, "step": 3195 }, { "epoch": 0.26, "grad_norm": 0.9891091585159302, "learning_rate": 1.650495845617797e-05, "loss": 0.0261, "step": 3200 }, { "epoch": 0.26, "grad_norm": 1.445168137550354, "learning_rate": 1.649602430090235e-05, "loss": 0.0192, "step": 3205 }, { "epoch": 0.26, "grad_norm": 1.0663169622421265, "learning_rate": 1.6487090145626732e-05, "loss": 0.0252, "step": 3210 }, { "epoch": 0.26, "grad_norm": 0.9698647856712341, "learning_rate": 1.6478155990351116e-05, "loss": 0.0223, "step": 3215 }, { "epoch": 0.26, "grad_norm": 0.3728877305984497, "learning_rate": 1.6469221835075494e-05, "loss": 0.0194, "step": 3220 }, { "epoch": 0.26, "grad_norm": 0.654509961605072, "learning_rate": 1.6460287679799875e-05, "loss": 0.0255, "step": 3225 }, { "epoch": 0.26, "grad_norm": 0.8245360851287842, "learning_rate": 1.645135352452426e-05, "loss": 0.0151, "step": 3230 }, { "epoch": 0.26, "grad_norm": 0.6417776942253113, "learning_rate": 1.644241936924864e-05, "loss": 0.0153, "step": 3235 }, { "epoch": 0.26, "grad_norm": 0.23088638484477997, "learning_rate": 1.6433485213973018e-05, "loss": 0.0258, "step": 3240 }, { "epoch": 0.26, "grad_norm": 0.4798796474933624, "learning_rate": 1.6424551058697402e-05, "loss": 0.0152, "step": 3245 }, { "epoch": 0.26, "grad_norm": 0.42295366525650024, "learning_rate": 1.6415616903421783e-05, "loss": 0.0153, "step": 3250 }, { "epoch": 0.26, "grad_norm": 0.3053552210330963, "learning_rate": 1.6406682748146165e-05, "loss": 0.0098, "step": 3255 }, { "epoch": 0.26, "grad_norm": 0.37020963430404663, "learning_rate": 1.6397748592870546e-05, "loss": 0.0151, "step": 3260 }, { "epoch": 0.26, "grad_norm": 0.5717061758041382, "learning_rate": 1.6388814437594927e-05, "loss": 0.0157, "step": 3265 }, { "epoch": 0.26, "grad_norm": 0.5504093766212463, "learning_rate": 1.6379880282319308e-05, "loss": 0.0397, "step": 3270 }, { "epoch": 0.26, "grad_norm": 0.6604989767074585, "learning_rate": 1.637094612704369e-05, "loss": 0.0158, "step": 3275 }, { "epoch": 0.26, "grad_norm": 0.7319514155387878, "learning_rate": 1.636201197176807e-05, "loss": 0.0118, "step": 3280 }, { "epoch": 0.26, "grad_norm": 0.5383831262588501, "learning_rate": 1.6353077816492454e-05, "loss": 0.0196, "step": 3285 }, { "epoch": 0.26, "grad_norm": 0.9546700716018677, "learning_rate": 1.6344143661216835e-05, "loss": 0.0204, "step": 3290 }, { "epoch": 0.26, "grad_norm": 0.929811954498291, "learning_rate": 1.6335209505941213e-05, "loss": 0.0286, "step": 3295 }, { "epoch": 0.27, "grad_norm": 0.545769453048706, "learning_rate": 1.6326275350665597e-05, "loss": 0.0178, "step": 3300 }, { "epoch": 0.27, "eval_loss": 0.02022329904139042, "eval_pearson_cosine": 0.9200674333472122, "eval_pearson_dot": 0.9176731158869308, "eval_pearson_euclidean": 0.9061361146439217, "eval_pearson_manhattan": 0.9047424977345989, "eval_pearson_max": 0.9200674333472122, "eval_runtime": 426.1218, "eval_samples_per_second": 1.173, "eval_spearman_cosine": 0.9299045516182064, "eval_spearman_dot": 0.9254067416269663, "eval_spearman_euclidean": 0.9302610970443881, "eval_spearman_manhattan": 0.9284666578666315, "eval_spearman_max": 0.9302610970443881, "eval_steps_per_second": 1.173, "step": 3300 }, { "epoch": 0.27, "grad_norm": 0.5426648259162903, "learning_rate": 1.6317341195389978e-05, "loss": 0.0123, "step": 3305 }, { "epoch": 0.27, "grad_norm": 0.4093814492225647, "learning_rate": 1.630840704011436e-05, "loss": 0.0176, "step": 3310 }, { "epoch": 0.27, "grad_norm": 0.9546383619308472, "learning_rate": 1.629947288483874e-05, "loss": 0.0195, "step": 3315 }, { "epoch": 0.27, "grad_norm": 0.6893852949142456, "learning_rate": 1.629053872956312e-05, "loss": 0.0253, "step": 3320 }, { "epoch": 0.27, "grad_norm": 0.3756393790245056, "learning_rate": 1.6281604574287502e-05, "loss": 0.0106, "step": 3325 }, { "epoch": 0.27, "grad_norm": 0.563850998878479, "learning_rate": 1.6272670419011883e-05, "loss": 0.017, "step": 3330 }, { "epoch": 0.27, "grad_norm": 1.3767915964126587, "learning_rate": 1.6263736263736265e-05, "loss": 0.0231, "step": 3335 }, { "epoch": 0.27, "grad_norm": 0.3786735236644745, "learning_rate": 1.6254802108460646e-05, "loss": 0.0121, "step": 3340 }, { "epoch": 0.27, "grad_norm": 0.2710884213447571, "learning_rate": 1.624586795318503e-05, "loss": 0.0121, "step": 3345 }, { "epoch": 0.27, "grad_norm": 0.4059496819972992, "learning_rate": 1.6236933797909408e-05, "loss": 0.0138, "step": 3350 }, { "epoch": 0.27, "grad_norm": 0.5236977338790894, "learning_rate": 1.622799964263379e-05, "loss": 0.023, "step": 3355 }, { "epoch": 0.27, "grad_norm": 1.2646647691726685, "learning_rate": 1.6219065487358173e-05, "loss": 0.026, "step": 3360 }, { "epoch": 0.27, "grad_norm": 0.5769827365875244, "learning_rate": 1.6210131332082554e-05, "loss": 0.0145, "step": 3365 }, { "epoch": 0.27, "grad_norm": 0.8552572727203369, "learning_rate": 1.6201197176806932e-05, "loss": 0.0132, "step": 3370 }, { "epoch": 0.27, "grad_norm": 1.690263271331787, "learning_rate": 1.6192263021531316e-05, "loss": 0.04, "step": 3375 }, { "epoch": 0.27, "grad_norm": 0.6928017735481262, "learning_rate": 1.6183328866255697e-05, "loss": 0.0184, "step": 3380 }, { "epoch": 0.27, "grad_norm": 0.31537023186683655, "learning_rate": 1.6174394710980078e-05, "loss": 0.0156, "step": 3385 }, { "epoch": 0.27, "grad_norm": 0.826603889465332, "learning_rate": 1.616546055570446e-05, "loss": 0.0253, "step": 3390 }, { "epoch": 0.27, "grad_norm": 1.0481098890304565, "learning_rate": 1.615652640042884e-05, "loss": 0.019, "step": 3395 }, { "epoch": 0.27, "grad_norm": 0.719532310962677, "learning_rate": 1.614759224515322e-05, "loss": 0.0152, "step": 3400 }, { "epoch": 0.27, "grad_norm": 0.7920011281967163, "learning_rate": 1.6138658089877602e-05, "loss": 0.0142, "step": 3405 }, { "epoch": 0.27, "grad_norm": 0.9310560822486877, "learning_rate": 1.6129723934601983e-05, "loss": 0.0187, "step": 3410 }, { "epoch": 0.27, "grad_norm": 0.4473055303096771, "learning_rate": 1.6120789779326368e-05, "loss": 0.0317, "step": 3415 }, { "epoch": 0.27, "grad_norm": 0.3835084140300751, "learning_rate": 1.611185562405075e-05, "loss": 0.0143, "step": 3420 }, { "epoch": 0.28, "grad_norm": 0.6586607098579407, "learning_rate": 1.6102921468775127e-05, "loss": 0.027, "step": 3425 }, { "epoch": 0.28, "grad_norm": 0.510007917881012, "learning_rate": 1.609398731349951e-05, "loss": 0.0213, "step": 3430 }, { "epoch": 0.28, "grad_norm": 0.7229037284851074, "learning_rate": 1.6085053158223892e-05, "loss": 0.014, "step": 3435 }, { "epoch": 0.28, "grad_norm": 0.35988929867744446, "learning_rate": 1.6076119002948273e-05, "loss": 0.0166, "step": 3440 }, { "epoch": 0.28, "grad_norm": 0.9715976119041443, "learning_rate": 1.6067184847672654e-05, "loss": 0.0202, "step": 3445 }, { "epoch": 0.28, "grad_norm": 0.5804651975631714, "learning_rate": 1.6058250692397035e-05, "loss": 0.0177, "step": 3450 }, { "epoch": 0.28, "grad_norm": 0.6329138278961182, "learning_rate": 1.6049316537121416e-05, "loss": 0.0184, "step": 3455 }, { "epoch": 0.28, "grad_norm": 0.46104130148887634, "learning_rate": 1.6040382381845797e-05, "loss": 0.0208, "step": 3460 }, { "epoch": 0.28, "grad_norm": 1.040407657623291, "learning_rate": 1.6031448226570178e-05, "loss": 0.0281, "step": 3465 }, { "epoch": 0.28, "grad_norm": 0.7342686057090759, "learning_rate": 1.6022514071294563e-05, "loss": 0.0163, "step": 3470 }, { "epoch": 0.28, "grad_norm": 0.7275007367134094, "learning_rate": 1.601357991601894e-05, "loss": 0.0193, "step": 3475 }, { "epoch": 0.28, "grad_norm": 0.30385661125183105, "learning_rate": 1.600464576074332e-05, "loss": 0.0153, "step": 3480 }, { "epoch": 0.28, "grad_norm": 0.5035312175750732, "learning_rate": 1.5995711605467706e-05, "loss": 0.0189, "step": 3485 }, { "epoch": 0.28, "grad_norm": 0.2530502676963806, "learning_rate": 1.5986777450192087e-05, "loss": 0.017, "step": 3490 }, { "epoch": 0.28, "grad_norm": 1.068692922592163, "learning_rate": 1.5977843294916468e-05, "loss": 0.0178, "step": 3495 }, { "epoch": 0.28, "grad_norm": 0.40073007345199585, "learning_rate": 1.596890913964085e-05, "loss": 0.0165, "step": 3500 }, { "epoch": 0.28, "grad_norm": 6.783413887023926, "learning_rate": 1.595997498436523e-05, "loss": 0.0235, "step": 3505 }, { "epoch": 0.28, "grad_norm": 0.46688321232795715, "learning_rate": 1.595104082908961e-05, "loss": 0.0188, "step": 3510 }, { "epoch": 0.28, "grad_norm": 0.6832770705223083, "learning_rate": 1.5942106673813992e-05, "loss": 0.0202, "step": 3515 }, { "epoch": 0.28, "grad_norm": 0.39423632621765137, "learning_rate": 1.5933172518538373e-05, "loss": 0.0118, "step": 3520 }, { "epoch": 0.28, "grad_norm": 0.8265882730484009, "learning_rate": 1.5924238363262754e-05, "loss": 0.0175, "step": 3525 }, { "epoch": 0.28, "grad_norm": 0.8301231265068054, "learning_rate": 1.5915304207987135e-05, "loss": 0.0396, "step": 3530 }, { "epoch": 0.28, "grad_norm": 0.45745956897735596, "learning_rate": 1.5906370052711516e-05, "loss": 0.0195, "step": 3535 }, { "epoch": 0.28, "grad_norm": 1.464941382408142, "learning_rate": 1.5897435897435897e-05, "loss": 0.0207, "step": 3540 }, { "epoch": 0.29, "grad_norm": 0.7965016961097717, "learning_rate": 1.588850174216028e-05, "loss": 0.0178, "step": 3545 }, { "epoch": 0.29, "grad_norm": 0.6353152394294739, "learning_rate": 1.5879567586884663e-05, "loss": 0.019, "step": 3550 }, { "epoch": 0.29, "grad_norm": 0.38456305861473083, "learning_rate": 1.587063343160904e-05, "loss": 0.0169, "step": 3555 }, { "epoch": 0.29, "grad_norm": 1.2627872228622437, "learning_rate": 1.5861699276333425e-05, "loss": 0.0226, "step": 3560 }, { "epoch": 0.29, "grad_norm": 0.5265514254570007, "learning_rate": 1.5852765121057806e-05, "loss": 0.0141, "step": 3565 }, { "epoch": 0.29, "grad_norm": 0.5938783288002014, "learning_rate": 1.5843830965782187e-05, "loss": 0.0178, "step": 3570 }, { "epoch": 0.29, "grad_norm": 0.39797863364219666, "learning_rate": 1.5834896810506568e-05, "loss": 0.0123, "step": 3575 }, { "epoch": 0.29, "grad_norm": 0.6035088300704956, "learning_rate": 1.582596265523095e-05, "loss": 0.0162, "step": 3580 }, { "epoch": 0.29, "grad_norm": 0.34465670585632324, "learning_rate": 1.581702849995533e-05, "loss": 0.0135, "step": 3585 }, { "epoch": 0.29, "grad_norm": 0.40939584374427795, "learning_rate": 1.580809434467971e-05, "loss": 0.0123, "step": 3590 }, { "epoch": 0.29, "grad_norm": 0.4972923994064331, "learning_rate": 1.5799160189404092e-05, "loss": 0.0178, "step": 3595 }, { "epoch": 0.29, "grad_norm": 0.3403087556362152, "learning_rate": 1.5790226034128477e-05, "loss": 0.0104, "step": 3600 }, { "epoch": 0.29, "eval_loss": 0.01895288936793804, "eval_pearson_cosine": 0.9183723675584331, "eval_pearson_dot": 0.9165141926322535, "eval_pearson_euclidean": 0.9111139441737433, "eval_pearson_manhattan": 0.9105498844088082, "eval_pearson_max": 0.9183723675584331, "eval_runtime": 425.5624, "eval_samples_per_second": 1.175, "eval_spearman_cosine": 0.925627350509402, "eval_spearman_dot": 0.9219613838455353, "eval_spearman_euclidean": 0.9258841515366061, "eval_spearman_manhattan": 0.9260336241344965, "eval_spearman_max": 0.9260336241344965, "eval_steps_per_second": 1.175, "step": 3600 }, { "epoch": 0.29, "grad_norm": 0.2765352129936218, "learning_rate": 1.5781291878852854e-05, "loss": 0.0154, "step": 3605 }, { "epoch": 0.29, "grad_norm": 3.1084377765655518, "learning_rate": 1.5772357723577235e-05, "loss": 0.0214, "step": 3610 }, { "epoch": 0.29, "grad_norm": 0.42401209473609924, "learning_rate": 1.576342356830162e-05, "loss": 0.0111, "step": 3615 }, { "epoch": 0.29, "grad_norm": 0.4498884081840515, "learning_rate": 1.5754489413026e-05, "loss": 0.0144, "step": 3620 }, { "epoch": 0.29, "grad_norm": 0.7784323692321777, "learning_rate": 1.574555525775038e-05, "loss": 0.0164, "step": 3625 }, { "epoch": 0.29, "grad_norm": 0.792809247970581, "learning_rate": 1.5736621102474763e-05, "loss": 0.0226, "step": 3630 }, { "epoch": 0.29, "grad_norm": 0.5559731721878052, "learning_rate": 1.5727686947199144e-05, "loss": 0.0157, "step": 3635 }, { "epoch": 0.29, "grad_norm": 0.5122278928756714, "learning_rate": 1.5718752791923525e-05, "loss": 0.0114, "step": 3640 }, { "epoch": 0.29, "grad_norm": 0.579097330570221, "learning_rate": 1.5709818636647906e-05, "loss": 0.0176, "step": 3645 }, { "epoch": 0.29, "grad_norm": 0.5125916004180908, "learning_rate": 1.5700884481372287e-05, "loss": 0.0112, "step": 3650 }, { "epoch": 0.29, "grad_norm": 0.6206502318382263, "learning_rate": 1.569195032609667e-05, "loss": 0.0129, "step": 3655 }, { "epoch": 0.29, "grad_norm": 0.578705906867981, "learning_rate": 1.568301617082105e-05, "loss": 0.0133, "step": 3660 }, { "epoch": 0.29, "grad_norm": 1.273260235786438, "learning_rate": 1.567408201554543e-05, "loss": 0.0238, "step": 3665 }, { "epoch": 0.3, "grad_norm": 0.6316016912460327, "learning_rate": 1.5665147860269814e-05, "loss": 0.0107, "step": 3670 }, { "epoch": 0.3, "grad_norm": 0.6251145005226135, "learning_rate": 1.5656213704994195e-05, "loss": 0.0115, "step": 3675 }, { "epoch": 0.3, "grad_norm": 0.48484379053115845, "learning_rate": 1.5647279549718573e-05, "loss": 0.0171, "step": 3680 }, { "epoch": 0.3, "grad_norm": 0.4522140920162201, "learning_rate": 1.5638345394442958e-05, "loss": 0.0222, "step": 3685 }, { "epoch": 0.3, "grad_norm": 0.6267346739768982, "learning_rate": 1.562941123916734e-05, "loss": 0.0214, "step": 3690 }, { "epoch": 0.3, "grad_norm": 0.9123010635375977, "learning_rate": 1.562047708389172e-05, "loss": 0.0214, "step": 3695 }, { "epoch": 0.3, "grad_norm": 1.2564802169799805, "learning_rate": 1.56115429286161e-05, "loss": 0.0209, "step": 3700 }, { "epoch": 0.3, "grad_norm": 0.4140920341014862, "learning_rate": 1.5602608773340482e-05, "loss": 0.0194, "step": 3705 }, { "epoch": 0.3, "grad_norm": 1.2953025102615356, "learning_rate": 1.5593674618064863e-05, "loss": 0.0175, "step": 3710 }, { "epoch": 0.3, "grad_norm": 0.19866344332695007, "learning_rate": 1.5584740462789244e-05, "loss": 0.017, "step": 3715 }, { "epoch": 0.3, "grad_norm": 0.5271120071411133, "learning_rate": 1.5575806307513625e-05, "loss": 0.0137, "step": 3720 }, { "epoch": 0.3, "grad_norm": 0.6619897484779358, "learning_rate": 1.5566872152238006e-05, "loss": 0.0184, "step": 3725 }, { "epoch": 0.3, "grad_norm": 0.7720787525177002, "learning_rate": 1.555793799696239e-05, "loss": 0.0198, "step": 3730 }, { "epoch": 0.3, "grad_norm": 1.2014611959457397, "learning_rate": 1.5549003841686768e-05, "loss": 0.0176, "step": 3735 }, { "epoch": 0.3, "grad_norm": 0.42123496532440186, "learning_rate": 1.554006968641115e-05, "loss": 0.0136, "step": 3740 }, { "epoch": 0.3, "grad_norm": 2.8967819213867188, "learning_rate": 1.5531135531135533e-05, "loss": 0.0202, "step": 3745 }, { "epoch": 0.3, "grad_norm": 0.4060193598270416, "learning_rate": 1.5522201375859914e-05, "loss": 0.0183, "step": 3750 }, { "epoch": 0.3, "grad_norm": 0.6353384256362915, "learning_rate": 1.5513267220584295e-05, "loss": 0.0141, "step": 3755 }, { "epoch": 0.3, "grad_norm": 0.3105901777744293, "learning_rate": 1.5504333065308677e-05, "loss": 0.0123, "step": 3760 }, { "epoch": 0.3, "grad_norm": 0.33913782238960266, "learning_rate": 1.5495398910033058e-05, "loss": 0.0176, "step": 3765 }, { "epoch": 0.3, "grad_norm": 0.664908766746521, "learning_rate": 1.548646475475744e-05, "loss": 0.0135, "step": 3770 }, { "epoch": 0.3, "grad_norm": 0.7868657112121582, "learning_rate": 1.547753059948182e-05, "loss": 0.0122, "step": 3775 }, { "epoch": 0.3, "grad_norm": 0.6869320869445801, "learning_rate": 1.54685964442062e-05, "loss": 0.0135, "step": 3780 }, { "epoch": 0.3, "grad_norm": 1.1669894456863403, "learning_rate": 1.5459662288930585e-05, "loss": 0.024, "step": 3785 }, { "epoch": 0.3, "grad_norm": 0.32140159606933594, "learning_rate": 1.5450728133654963e-05, "loss": 0.0109, "step": 3790 }, { "epoch": 0.31, "grad_norm": 0.5477967262268066, "learning_rate": 1.5441793978379344e-05, "loss": 0.0132, "step": 3795 }, { "epoch": 0.31, "grad_norm": 0.2863134741783142, "learning_rate": 1.5432859823103728e-05, "loss": 0.0225, "step": 3800 }, { "epoch": 0.31, "grad_norm": 0.7895782589912415, "learning_rate": 1.542392566782811e-05, "loss": 0.0265, "step": 3805 }, { "epoch": 0.31, "grad_norm": 0.36137330532073975, "learning_rate": 1.5414991512552487e-05, "loss": 0.0114, "step": 3810 }, { "epoch": 0.31, "grad_norm": 0.5505108833312988, "learning_rate": 1.540605735727687e-05, "loss": 0.0244, "step": 3815 }, { "epoch": 0.31, "grad_norm": 1.2253755331039429, "learning_rate": 1.5397123202001252e-05, "loss": 0.0253, "step": 3820 }, { "epoch": 0.31, "grad_norm": 0.25381264090538025, "learning_rate": 1.5388189046725633e-05, "loss": 0.0215, "step": 3825 }, { "epoch": 0.31, "grad_norm": 0.59675133228302, "learning_rate": 1.5379254891450014e-05, "loss": 0.0229, "step": 3830 }, { "epoch": 0.31, "grad_norm": 0.4737885296344757, "learning_rate": 1.5370320736174395e-05, "loss": 0.0172, "step": 3835 }, { "epoch": 0.31, "grad_norm": 0.7271224856376648, "learning_rate": 1.536138658089878e-05, "loss": 0.0155, "step": 3840 }, { "epoch": 0.31, "grad_norm": 0.3435806930065155, "learning_rate": 1.5352452425623158e-05, "loss": 0.0115, "step": 3845 }, { "epoch": 0.31, "grad_norm": 0.37383875250816345, "learning_rate": 1.534351827034754e-05, "loss": 0.0175, "step": 3850 }, { "epoch": 0.31, "grad_norm": 0.2769816219806671, "learning_rate": 1.5334584115071923e-05, "loss": 0.012, "step": 3855 }, { "epoch": 0.31, "grad_norm": 0.421824187040329, "learning_rate": 1.5325649959796304e-05, "loss": 0.0168, "step": 3860 }, { "epoch": 0.31, "grad_norm": 0.48985883593559265, "learning_rate": 1.5316715804520682e-05, "loss": 0.0121, "step": 3865 }, { "epoch": 0.31, "grad_norm": 0.4765077829360962, "learning_rate": 1.5307781649245066e-05, "loss": 0.0181, "step": 3870 }, { "epoch": 0.31, "grad_norm": 0.48202696442604065, "learning_rate": 1.5298847493969447e-05, "loss": 0.0157, "step": 3875 }, { "epoch": 0.31, "grad_norm": 2.033970594406128, "learning_rate": 1.5289913338693828e-05, "loss": 0.0136, "step": 3880 }, { "epoch": 0.31, "grad_norm": 0.33205658197402954, "learning_rate": 1.528097918341821e-05, "loss": 0.012, "step": 3885 }, { "epoch": 0.31, "grad_norm": 0.4964057207107544, "learning_rate": 1.527204502814259e-05, "loss": 0.0109, "step": 3890 }, { "epoch": 0.31, "grad_norm": 0.9996275901794434, "learning_rate": 1.526311087286697e-05, "loss": 0.0153, "step": 3895 }, { "epoch": 0.31, "grad_norm": 0.5008103251457214, "learning_rate": 1.5254176717591354e-05, "loss": 0.0194, "step": 3900 }, { "epoch": 0.31, "eval_loss": 0.019112512469291687, "eval_pearson_cosine": 0.9187946359889192, "eval_pearson_dot": 0.916935359224543, "eval_pearson_euclidean": 0.9078268066209557, "eval_pearson_manhattan": 0.9068373107836385, "eval_pearson_max": 0.9187946359889192, "eval_runtime": 426.0704, "eval_samples_per_second": 1.174, "eval_spearman_cosine": 0.9320740642962572, "eval_spearman_dot": 0.9297467269869079, "eval_spearman_euclidean": 0.9318172632690529, "eval_spearman_manhattan": 0.9317702230808924, "eval_spearman_max": 0.9320740642962572, "eval_steps_per_second": 1.174, "step": 3900 }, { "epoch": 0.31, "grad_norm": 0.5863137245178223, "learning_rate": 1.5245242562315733e-05, "loss": 0.0155, "step": 3905 }, { "epoch": 0.31, "grad_norm": 0.9501858353614807, "learning_rate": 1.5236308407040114e-05, "loss": 0.023, "step": 3910 }, { "epoch": 0.31, "grad_norm": 1.0070725679397583, "learning_rate": 1.5227374251764497e-05, "loss": 0.0159, "step": 3915 }, { "epoch": 0.32, "grad_norm": 1.1535245180130005, "learning_rate": 1.5218440096488878e-05, "loss": 0.0271, "step": 3920 }, { "epoch": 0.32, "grad_norm": 0.3500936031341553, "learning_rate": 1.5209505941213261e-05, "loss": 0.0185, "step": 3925 }, { "epoch": 0.32, "grad_norm": 0.5063101649284363, "learning_rate": 1.5200571785937642e-05, "loss": 0.0092, "step": 3930 }, { "epoch": 0.32, "grad_norm": 0.48119908571243286, "learning_rate": 1.5191637630662021e-05, "loss": 0.0156, "step": 3935 }, { "epoch": 0.32, "grad_norm": 0.7710624933242798, "learning_rate": 1.5182703475386404e-05, "loss": 0.0182, "step": 3940 }, { "epoch": 0.32, "grad_norm": 0.5304814577102661, "learning_rate": 1.5173769320110785e-05, "loss": 0.0127, "step": 3945 }, { "epoch": 0.32, "grad_norm": 0.5894374251365662, "learning_rate": 1.5164835164835166e-05, "loss": 0.012, "step": 3950 }, { "epoch": 0.32, "grad_norm": 0.48509135842323303, "learning_rate": 1.5155901009559549e-05, "loss": 0.0164, "step": 3955 }, { "epoch": 0.32, "grad_norm": 1.00001060962677, "learning_rate": 1.5146966854283928e-05, "loss": 0.0105, "step": 3960 }, { "epoch": 0.32, "grad_norm": 0.5003077983856201, "learning_rate": 1.513803269900831e-05, "loss": 0.0191, "step": 3965 }, { "epoch": 0.32, "grad_norm": 0.4220547378063202, "learning_rate": 1.5129098543732692e-05, "loss": 0.0182, "step": 3970 }, { "epoch": 0.32, "grad_norm": 0.353935182094574, "learning_rate": 1.5120164388457073e-05, "loss": 0.0105, "step": 3975 }, { "epoch": 0.32, "grad_norm": 0.7862482070922852, "learning_rate": 1.5111230233181452e-05, "loss": 0.0153, "step": 3980 }, { "epoch": 0.32, "grad_norm": 0.4707240164279938, "learning_rate": 1.5102296077905835e-05, "loss": 0.0106, "step": 3985 }, { "epoch": 0.32, "grad_norm": 0.5706127882003784, "learning_rate": 1.5093361922630216e-05, "loss": 0.0193, "step": 3990 }, { "epoch": 0.32, "grad_norm": 0.5268707275390625, "learning_rate": 1.5084427767354597e-05, "loss": 0.0115, "step": 3995 }, { "epoch": 0.32, "grad_norm": 0.42871159315109253, "learning_rate": 1.507549361207898e-05, "loss": 0.0111, "step": 4000 }, { "epoch": 0.32, "grad_norm": 0.47120198607444763, "learning_rate": 1.5066559456803361e-05, "loss": 0.0148, "step": 4005 }, { "epoch": 0.32, "grad_norm": 0.6913847327232361, "learning_rate": 1.505762530152774e-05, "loss": 0.0156, "step": 4010 }, { "epoch": 0.32, "grad_norm": 0.40243229269981384, "learning_rate": 1.5048691146252123e-05, "loss": 0.021, "step": 4015 }, { "epoch": 0.32, "grad_norm": 0.4630833566188812, "learning_rate": 1.5039756990976504e-05, "loss": 0.0142, "step": 4020 }, { "epoch": 0.32, "grad_norm": 0.7278119325637817, "learning_rate": 1.5030822835700887e-05, "loss": 0.0221, "step": 4025 }, { "epoch": 0.32, "grad_norm": 0.5856547951698303, "learning_rate": 1.5021888680425268e-05, "loss": 0.0127, "step": 4030 }, { "epoch": 0.32, "grad_norm": 0.3612341582775116, "learning_rate": 1.5012954525149647e-05, "loss": 0.0142, "step": 4035 }, { "epoch": 0.32, "grad_norm": 0.3373230993747711, "learning_rate": 1.500402036987403e-05, "loss": 0.013, "step": 4040 }, { "epoch": 0.33, "grad_norm": 0.3555310070514679, "learning_rate": 1.4995086214598411e-05, "loss": 0.0188, "step": 4045 }, { "epoch": 0.33, "grad_norm": 0.27012017369270325, "learning_rate": 1.4986152059322792e-05, "loss": 0.016, "step": 4050 }, { "epoch": 0.33, "grad_norm": 0.7879682183265686, "learning_rate": 1.4977217904047175e-05, "loss": 0.019, "step": 4055 }, { "epoch": 0.33, "grad_norm": 0.4750874638557434, "learning_rate": 1.4968283748771554e-05, "loss": 0.0156, "step": 4060 }, { "epoch": 0.33, "grad_norm": 0.5031574368476868, "learning_rate": 1.4959349593495935e-05, "loss": 0.0206, "step": 4065 }, { "epoch": 0.33, "grad_norm": 0.5057447552680969, "learning_rate": 1.4950415438220318e-05, "loss": 0.0144, "step": 4070 }, { "epoch": 0.33, "grad_norm": 1.0315029621124268, "learning_rate": 1.4941481282944699e-05, "loss": 0.0179, "step": 4075 }, { "epoch": 0.33, "grad_norm": 0.5417433977127075, "learning_rate": 1.493254712766908e-05, "loss": 0.0126, "step": 4080 }, { "epoch": 0.33, "grad_norm": 0.667726993560791, "learning_rate": 1.4923612972393463e-05, "loss": 0.0143, "step": 4085 }, { "epoch": 0.33, "grad_norm": 0.8723002672195435, "learning_rate": 1.4914678817117842e-05, "loss": 0.012, "step": 4090 }, { "epoch": 0.33, "grad_norm": 0.8944787979125977, "learning_rate": 1.4905744661842223e-05, "loss": 0.0205, "step": 4095 }, { "epoch": 0.33, "grad_norm": 0.5145695805549622, "learning_rate": 1.4896810506566606e-05, "loss": 0.0118, "step": 4100 }, { "epoch": 0.33, "grad_norm": 0.9632664322853088, "learning_rate": 1.4887876351290987e-05, "loss": 0.0233, "step": 4105 }, { "epoch": 0.33, "grad_norm": 0.4608353078365326, "learning_rate": 1.487894219601537e-05, "loss": 0.0192, "step": 4110 }, { "epoch": 0.33, "grad_norm": 0.8873416781425476, "learning_rate": 1.4870008040739749e-05, "loss": 0.0196, "step": 4115 }, { "epoch": 0.33, "grad_norm": 0.7420912384986877, "learning_rate": 1.486107388546413e-05, "loss": 0.0114, "step": 4120 }, { "epoch": 0.33, "grad_norm": 0.3144592344760895, "learning_rate": 1.4852139730188513e-05, "loss": 0.0136, "step": 4125 }, { "epoch": 0.33, "grad_norm": 0.4143444001674652, "learning_rate": 1.4843205574912894e-05, "loss": 0.0102, "step": 4130 }, { "epoch": 0.33, "grad_norm": 1.3450114727020264, "learning_rate": 1.4834271419637275e-05, "loss": 0.0273, "step": 4135 }, { "epoch": 0.33, "grad_norm": 0.6474159955978394, "learning_rate": 1.4825337264361657e-05, "loss": 0.0125, "step": 4140 }, { "epoch": 0.33, "grad_norm": 0.3440127372741699, "learning_rate": 1.4816403109086037e-05, "loss": 0.016, "step": 4145 }, { "epoch": 0.33, "grad_norm": 0.6236218214035034, "learning_rate": 1.4807468953810418e-05, "loss": 0.019, "step": 4150 }, { "epoch": 0.33, "grad_norm": 0.9320889115333557, "learning_rate": 1.47985347985348e-05, "loss": 0.015, "step": 4155 }, { "epoch": 0.33, "grad_norm": 0.36292609572410583, "learning_rate": 1.4789600643259182e-05, "loss": 0.0155, "step": 4160 }, { "epoch": 0.33, "grad_norm": 1.278648018836975, "learning_rate": 1.4780666487983561e-05, "loss": 0.0229, "step": 4165 }, { "epoch": 0.34, "grad_norm": 0.24334418773651123, "learning_rate": 1.4771732332707944e-05, "loss": 0.0099, "step": 4170 }, { "epoch": 0.34, "grad_norm": 0.7839992046356201, "learning_rate": 1.4762798177432325e-05, "loss": 0.0167, "step": 4175 }, { "epoch": 0.34, "grad_norm": 1.3292075395584106, "learning_rate": 1.4753864022156706e-05, "loss": 0.0168, "step": 4180 }, { "epoch": 0.34, "grad_norm": 0.4192187488079071, "learning_rate": 1.4744929866881089e-05, "loss": 0.0177, "step": 4185 }, { "epoch": 0.34, "grad_norm": 0.5623775124549866, "learning_rate": 1.4735995711605468e-05, "loss": 0.0103, "step": 4190 }, { "epoch": 0.34, "grad_norm": 0.5713854432106018, "learning_rate": 1.4727061556329849e-05, "loss": 0.011, "step": 4195 }, { "epoch": 0.34, "grad_norm": 0.6574892997741699, "learning_rate": 1.4718127401054232e-05, "loss": 0.0169, "step": 4200 }, { "epoch": 0.34, "eval_loss": 0.01563582755625248, "eval_pearson_cosine": 0.93071357698338, "eval_pearson_dot": 0.9303618828257327, "eval_pearson_euclidean": 0.9158033535152222, "eval_pearson_manhattan": 0.9136122746167228, "eval_pearson_max": 0.93071357698338, "eval_runtime": 426.316, "eval_samples_per_second": 1.173, "eval_spearman_cosine": 0.9357050868203473, "eval_spearman_dot": 0.9326605306421225, "eval_spearman_euclidean": 0.9346635786543147, "eval_spearman_manhattan": 0.9334011736046943, "eval_spearman_max": 0.9357050868203473, "eval_steps_per_second": 1.173, "step": 4200 }, { "epoch": 0.34, "grad_norm": 0.5762566328048706, "learning_rate": 1.4709193245778613e-05, "loss": 0.0296, "step": 4205 }, { "epoch": 0.34, "grad_norm": 0.6294378638267517, "learning_rate": 1.4700259090502995e-05, "loss": 0.0158, "step": 4210 }, { "epoch": 0.34, "grad_norm": 0.5651215314865112, "learning_rate": 1.4691324935227376e-05, "loss": 0.0131, "step": 4215 }, { "epoch": 0.34, "grad_norm": 0.6574472784996033, "learning_rate": 1.4682390779951756e-05, "loss": 0.0166, "step": 4220 }, { "epoch": 0.34, "grad_norm": 0.4540180563926697, "learning_rate": 1.4673456624676139e-05, "loss": 0.024, "step": 4225 }, { "epoch": 0.34, "grad_norm": 0.3493760824203491, "learning_rate": 1.466452246940052e-05, "loss": 0.0149, "step": 4230 }, { "epoch": 0.34, "grad_norm": 1.0167644023895264, "learning_rate": 1.46555883141249e-05, "loss": 0.0149, "step": 4235 }, { "epoch": 0.34, "grad_norm": 0.9241525530815125, "learning_rate": 1.4646654158849283e-05, "loss": 0.0184, "step": 4240 }, { "epoch": 0.34, "grad_norm": 0.6745620965957642, "learning_rate": 1.4637720003573663e-05, "loss": 0.0158, "step": 4245 }, { "epoch": 0.34, "grad_norm": 0.4359780251979828, "learning_rate": 1.4628785848298044e-05, "loss": 0.0098, "step": 4250 }, { "epoch": 0.34, "grad_norm": 0.4030389189720154, "learning_rate": 1.4619851693022426e-05, "loss": 0.0184, "step": 4255 }, { "epoch": 0.34, "grad_norm": 0.5261355638504028, "learning_rate": 1.4610917537746807e-05, "loss": 0.0141, "step": 4260 }, { "epoch": 0.34, "grad_norm": 0.38097327947616577, "learning_rate": 1.4601983382471187e-05, "loss": 0.0083, "step": 4265 }, { "epoch": 0.34, "grad_norm": 0.41486430168151855, "learning_rate": 1.459304922719557e-05, "loss": 0.0202, "step": 4270 }, { "epoch": 0.34, "grad_norm": 0.5645167827606201, "learning_rate": 1.458411507191995e-05, "loss": 0.0126, "step": 4275 }, { "epoch": 0.34, "grad_norm": 0.4771600365638733, "learning_rate": 1.4575180916644332e-05, "loss": 0.0103, "step": 4280 }, { "epoch": 0.34, "grad_norm": 0.8267874717712402, "learning_rate": 1.4566246761368714e-05, "loss": 0.0182, "step": 4285 }, { "epoch": 0.34, "grad_norm": 0.7216476202011108, "learning_rate": 1.4557312606093095e-05, "loss": 0.0155, "step": 4290 }, { "epoch": 0.35, "grad_norm": 1.0095362663269043, "learning_rate": 1.4548378450817478e-05, "loss": 0.0199, "step": 4295 }, { "epoch": 0.35, "grad_norm": 0.36790329217910767, "learning_rate": 1.4539444295541857e-05, "loss": 0.0231, "step": 4300 }, { "epoch": 0.35, "grad_norm": 0.4778492748737335, "learning_rate": 1.4530510140266239e-05, "loss": 0.0101, "step": 4305 }, { "epoch": 0.35, "grad_norm": 0.3736858665943146, "learning_rate": 1.4521575984990621e-05, "loss": 0.0157, "step": 4310 }, { "epoch": 0.35, "grad_norm": 0.5705660581588745, "learning_rate": 1.4512641829715002e-05, "loss": 0.0133, "step": 4315 }, { "epoch": 0.35, "grad_norm": 0.7104588747024536, "learning_rate": 1.4503707674439382e-05, "loss": 0.0209, "step": 4320 }, { "epoch": 0.35, "grad_norm": 0.4687543511390686, "learning_rate": 1.4494773519163764e-05, "loss": 0.0072, "step": 4325 }, { "epoch": 0.35, "grad_norm": 0.5142909288406372, "learning_rate": 1.4485839363888145e-05, "loss": 0.0127, "step": 4330 }, { "epoch": 0.35, "grad_norm": 0.3397541046142578, "learning_rate": 1.4476905208612526e-05, "loss": 0.0122, "step": 4335 }, { "epoch": 0.35, "grad_norm": 0.4727453589439392, "learning_rate": 1.446797105333691e-05, "loss": 0.0212, "step": 4340 }, { "epoch": 0.35, "grad_norm": 0.7717282176017761, "learning_rate": 1.445903689806129e-05, "loss": 0.0277, "step": 4345 }, { "epoch": 0.35, "grad_norm": 0.4375796914100647, "learning_rate": 1.445010274278567e-05, "loss": 0.0119, "step": 4350 }, { "epoch": 0.35, "grad_norm": 0.45401817560195923, "learning_rate": 1.4441168587510052e-05, "loss": 0.0128, "step": 4355 }, { "epoch": 0.35, "grad_norm": 0.4388526976108551, "learning_rate": 1.4432234432234433e-05, "loss": 0.0181, "step": 4360 }, { "epoch": 0.35, "grad_norm": 0.4506477415561676, "learning_rate": 1.4423300276958814e-05, "loss": 0.0118, "step": 4365 }, { "epoch": 0.35, "grad_norm": 0.7086319923400879, "learning_rate": 1.4414366121683197e-05, "loss": 0.0111, "step": 4370 }, { "epoch": 0.35, "grad_norm": 0.185771182179451, "learning_rate": 1.4405431966407576e-05, "loss": 0.0151, "step": 4375 }, { "epoch": 0.35, "grad_norm": 0.26904481649398804, "learning_rate": 1.4396497811131958e-05, "loss": 0.0116, "step": 4380 }, { "epoch": 0.35, "grad_norm": 0.8145022392272949, "learning_rate": 1.438756365585634e-05, "loss": 0.0097, "step": 4385 }, { "epoch": 0.35, "grad_norm": 1.1931172609329224, "learning_rate": 1.4378629500580721e-05, "loss": 0.0158, "step": 4390 }, { "epoch": 0.35, "grad_norm": 0.37602362036705017, "learning_rate": 1.4369695345305104e-05, "loss": 0.0145, "step": 4395 }, { "epoch": 0.35, "grad_norm": 0.5753958225250244, "learning_rate": 1.4360761190029483e-05, "loss": 0.0156, "step": 4400 } ], "logging_steps": 5, "max_steps": 12437, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 400, "total_flos": 0.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }