{ "best_metric": 1.7320468425750732, "best_model_checkpoint": "miner_id_24/checkpoint-200", "epoch": 0.9673518742442564, "eval_steps": 50, "global_step": 200, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0048367593712212815, "grad_norm": 0.9632701277732849, "learning_rate": 6e-06, "loss": 2.4015, "step": 1 }, { "epoch": 0.0048367593712212815, "eval_loss": 2.2109501361846924, "eval_runtime": 17.8916, "eval_samples_per_second": 19.45, "eval_steps_per_second": 4.863, "step": 1 }, { "epoch": 0.009673518742442563, "grad_norm": 0.9858382940292358, "learning_rate": 1.2e-05, "loss": 2.3818, "step": 2 }, { "epoch": 0.014510278113663845, "grad_norm": 0.9527080059051514, "learning_rate": 1.8e-05, "loss": 2.2582, "step": 3 }, { "epoch": 0.019347037484885126, "grad_norm": 0.8105273246765137, "learning_rate": 2.4e-05, "loss": 2.1767, "step": 4 }, { "epoch": 0.02418379685610641, "grad_norm": 0.7631579041481018, "learning_rate": 3e-05, "loss": 2.0332, "step": 5 }, { "epoch": 0.02902055622732769, "grad_norm": 0.8983649015426636, "learning_rate": 3.6e-05, "loss": 2.2603, "step": 6 }, { "epoch": 0.03385731559854897, "grad_norm": 0.7237122058868408, "learning_rate": 4.2e-05, "loss": 2.1122, "step": 7 }, { "epoch": 0.03869407496977025, "grad_norm": 0.6886872053146362, "learning_rate": 4.8e-05, "loss": 2.1978, "step": 8 }, { "epoch": 0.04353083434099154, "grad_norm": 0.6846583485603333, "learning_rate": 5.4000000000000005e-05, "loss": 2.2006, "step": 9 }, { "epoch": 0.04836759371221282, "grad_norm": 0.6080830097198486, "learning_rate": 6e-05, "loss": 2.0522, "step": 10 }, { "epoch": 0.053204353083434096, "grad_norm": 0.7755049467086792, "learning_rate": 5.999589914977407e-05, "loss": 2.123, "step": 11 }, { "epoch": 0.05804111245465538, "grad_norm": 0.7986268401145935, "learning_rate": 5.998359772022778e-05, "loss": 2.0668, "step": 12 }, { "epoch": 0.06287787182587666, "grad_norm": 0.5952561497688293, "learning_rate": 5.996309907444915e-05, "loss": 1.9486, "step": 13 }, { "epoch": 0.06771463119709795, "grad_norm": 0.7420691251754761, "learning_rate": 5.9934408816563236e-05, "loss": 2.0322, "step": 14 }, { "epoch": 0.07255139056831923, "grad_norm": 0.6334537267684937, "learning_rate": 5.98975347902001e-05, "loss": 1.8305, "step": 15 }, { "epoch": 0.0773881499395405, "grad_norm": 0.5374141931533813, "learning_rate": 5.9852487076350345e-05, "loss": 2.0767, "step": 16 }, { "epoch": 0.08222490931076179, "grad_norm": 0.54233717918396, "learning_rate": 5.979927799060915e-05, "loss": 1.9312, "step": 17 }, { "epoch": 0.08706166868198308, "grad_norm": 0.47292783856391907, "learning_rate": 5.9737922079809257e-05, "loss": 1.8548, "step": 18 }, { "epoch": 0.09189842805320435, "grad_norm": 0.5466502904891968, "learning_rate": 5.9668436118044054e-05, "loss": 2.0625, "step": 19 }, { "epoch": 0.09673518742442563, "grad_norm": 0.5066276788711548, "learning_rate": 5.959083910208167e-05, "loss": 1.8547, "step": 20 }, { "epoch": 0.10157194679564692, "grad_norm": 0.49660593271255493, "learning_rate": 5.9505152246171474e-05, "loss": 1.9956, "step": 21 }, { "epoch": 0.10640870616686819, "grad_norm": 0.5254217982292175, "learning_rate": 5.941139897624428e-05, "loss": 1.85, "step": 22 }, { "epoch": 0.11124546553808948, "grad_norm": 0.4571835994720459, "learning_rate": 5.9309604923507984e-05, "loss": 1.8309, "step": 23 }, { "epoch": 0.11608222490931076, "grad_norm": 0.4921605587005615, "learning_rate": 5.9199797917440176e-05, "loss": 1.8846, "step": 24 }, { "epoch": 0.12091898428053205, "grad_norm": 0.4969245195388794, "learning_rate": 5.908200797817991e-05, "loss": 1.9426, "step": 25 }, { "epoch": 0.12575574365175332, "grad_norm": 0.5094956159591675, "learning_rate": 5.895626730832046e-05, "loss": 2.0461, "step": 26 }, { "epoch": 0.13059250302297462, "grad_norm": 0.48882463574409485, "learning_rate": 5.882261028410545e-05, "loss": 1.892, "step": 27 }, { "epoch": 0.1354292623941959, "grad_norm": 0.47303664684295654, "learning_rate": 5.8681073446030734e-05, "loss": 2.0175, "step": 28 }, { "epoch": 0.14026602176541716, "grad_norm": 0.5476979613304138, "learning_rate": 5.853169548885461e-05, "loss": 1.8809, "step": 29 }, { "epoch": 0.14510278113663846, "grad_norm": 0.4879865050315857, "learning_rate": 5.8374517251019035e-05, "loss": 1.8718, "step": 30 }, { "epoch": 0.14993954050785974, "grad_norm": 0.465002179145813, "learning_rate": 5.820958170348484e-05, "loss": 1.8935, "step": 31 }, { "epoch": 0.154776299879081, "grad_norm": 0.46331265568733215, "learning_rate": 5.8036933937983825e-05, "loss": 1.9589, "step": 32 }, { "epoch": 0.1596130592503023, "grad_norm": 0.4595167934894562, "learning_rate": 5.7856621154691217e-05, "loss": 1.8369, "step": 33 }, { "epoch": 0.16444981862152358, "grad_norm": 0.4678586721420288, "learning_rate": 5.766869264932154e-05, "loss": 1.9281, "step": 34 }, { "epoch": 0.16928657799274485, "grad_norm": 0.46128371357917786, "learning_rate": 5.747319979965172e-05, "loss": 1.8761, "step": 35 }, { "epoch": 0.17412333736396615, "grad_norm": 0.5178216695785522, "learning_rate": 5.727019605147488e-05, "loss": 1.9501, "step": 36 }, { "epoch": 0.17896009673518742, "grad_norm": 0.4670916497707367, "learning_rate": 5.7059736903988775e-05, "loss": 1.9176, "step": 37 }, { "epoch": 0.1837968561064087, "grad_norm": 0.4867013990879059, "learning_rate": 5.684187989462291e-05, "loss": 1.8645, "step": 38 }, { "epoch": 0.18863361547763, "grad_norm": 0.43132489919662476, "learning_rate": 5.661668458330836e-05, "loss": 1.7651, "step": 39 }, { "epoch": 0.19347037484885127, "grad_norm": 0.4480936527252197, "learning_rate": 5.638421253619467e-05, "loss": 2.0129, "step": 40 }, { "epoch": 0.19830713422007254, "grad_norm": 0.4475274384021759, "learning_rate": 5.614452730881832e-05, "loss": 1.9084, "step": 41 }, { "epoch": 0.20314389359129384, "grad_norm": 0.4610969126224518, "learning_rate": 5.589769442872722e-05, "loss": 1.878, "step": 42 }, { "epoch": 0.2079806529625151, "grad_norm": 0.467438668012619, "learning_rate": 5.5643781377566175e-05, "loss": 1.9073, "step": 43 }, { "epoch": 0.21281741233373638, "grad_norm": 0.47766032814979553, "learning_rate": 5.538285757262806e-05, "loss": 1.8999, "step": 44 }, { "epoch": 0.21765417170495768, "grad_norm": 0.5315982103347778, "learning_rate": 5.5114994347875856e-05, "loss": 1.9372, "step": 45 }, { "epoch": 0.22249093107617895, "grad_norm": 0.6379398703575134, "learning_rate": 5.48402649344406e-05, "loss": 1.9203, "step": 46 }, { "epoch": 0.22732769044740025, "grad_norm": 0.6935904026031494, "learning_rate": 5.455874444060078e-05, "loss": 1.9296, "step": 47 }, { "epoch": 0.23216444981862153, "grad_norm": 0.7248342633247375, "learning_rate": 5.427050983124843e-05, "loss": 1.9293, "step": 48 }, { "epoch": 0.2370012091898428, "grad_norm": 0.8179011940956116, "learning_rate": 5.397563990684774e-05, "loss": 2.0154, "step": 49 }, { "epoch": 0.2418379685610641, "grad_norm": 0.9732866883277893, "learning_rate": 5.367421528189181e-05, "loss": 2.0389, "step": 50 }, { "epoch": 0.2418379685610641, "eval_loss": 1.8738901615142822, "eval_runtime": 18.4039, "eval_samples_per_second": 18.909, "eval_steps_per_second": 4.727, "step": 50 }, { "epoch": 0.24667472793228537, "grad_norm": 0.47695332765579224, "learning_rate": 5.336631836286338e-05, "loss": 2.0234, "step": 51 }, { "epoch": 0.25151148730350664, "grad_norm": 0.4575190842151642, "learning_rate": 5.3052033325705774e-05, "loss": 1.9453, "step": 52 }, { "epoch": 0.25634824667472794, "grad_norm": 0.46312448382377625, "learning_rate": 5.2731446092810044e-05, "loss": 1.9204, "step": 53 }, { "epoch": 0.26118500604594924, "grad_norm": 0.4436776340007782, "learning_rate": 5.240464430952462e-05, "loss": 1.9036, "step": 54 }, { "epoch": 0.2660217654171705, "grad_norm": 0.42219528555870056, "learning_rate": 5.207171732019395e-05, "loss": 1.6588, "step": 55 }, { "epoch": 0.2708585247883918, "grad_norm": 0.5274652242660522, "learning_rate": 5.1732756143732675e-05, "loss": 1.9097, "step": 56 }, { "epoch": 0.2756952841596131, "grad_norm": 0.5278433561325073, "learning_rate": 5.1387853448741916e-05, "loss": 1.9225, "step": 57 }, { "epoch": 0.28053204353083433, "grad_norm": 0.46141645312309265, "learning_rate": 5.103710352817465e-05, "loss": 1.7575, "step": 58 }, { "epoch": 0.28536880290205563, "grad_norm": 0.37459704279899597, "learning_rate": 5.068060227355698e-05, "loss": 1.8215, "step": 59 }, { "epoch": 0.29020556227327693, "grad_norm": 0.4300343096256256, "learning_rate": 5.0318447148772234e-05, "loss": 1.7966, "step": 60 }, { "epoch": 0.2950423216444982, "grad_norm": 0.44032105803489685, "learning_rate": 4.995073716341545e-05, "loss": 1.827, "step": 61 }, { "epoch": 0.2998790810157195, "grad_norm": 0.4639372229576111, "learning_rate": 4.957757284572506e-05, "loss": 1.8347, "step": 62 }, { "epoch": 0.3047158403869408, "grad_norm": 0.41214707493782043, "learning_rate": 4.91990562150995e-05, "loss": 1.7148, "step": 63 }, { "epoch": 0.309552599758162, "grad_norm": 0.3841429054737091, "learning_rate": 4.881529075420611e-05, "loss": 1.8549, "step": 64 }, { "epoch": 0.3143893591293833, "grad_norm": 0.4167156517505646, "learning_rate": 4.8426381380690036e-05, "loss": 1.8203, "step": 65 }, { "epoch": 0.3192261185006046, "grad_norm": 0.4834750294685364, "learning_rate": 4.8032434418490753e-05, "loss": 1.8424, "step": 66 }, { "epoch": 0.32406287787182586, "grad_norm": 0.42098328471183777, "learning_rate": 4.7633557568774194e-05, "loss": 1.8602, "step": 67 }, { "epoch": 0.32889963724304716, "grad_norm": 0.43436023592948914, "learning_rate": 4.722985988048831e-05, "loss": 1.7259, "step": 68 }, { "epoch": 0.33373639661426846, "grad_norm": 0.4097272753715515, "learning_rate": 4.6821451720550184e-05, "loss": 1.8614, "step": 69 }, { "epoch": 0.3385731559854897, "grad_norm": 0.40257489681243896, "learning_rate": 4.640844474367282e-05, "loss": 1.7213, "step": 70 }, { "epoch": 0.343409915356711, "grad_norm": 0.39686596393585205, "learning_rate": 4.5990951861839815e-05, "loss": 1.7731, "step": 71 }, { "epoch": 0.3482466747279323, "grad_norm": 0.4416908323764801, "learning_rate": 4.5569087213436455e-05, "loss": 1.8476, "step": 72 }, { "epoch": 0.35308343409915355, "grad_norm": 0.42743927240371704, "learning_rate": 4.514296613204532e-05, "loss": 1.8935, "step": 73 }, { "epoch": 0.35792019347037485, "grad_norm": 0.4807327687740326, "learning_rate": 4.471270511491525e-05, "loss": 1.7099, "step": 74 }, { "epoch": 0.36275695284159615, "grad_norm": 0.48261377215385437, "learning_rate": 4.427842179111221e-05, "loss": 2.0033, "step": 75 }, { "epoch": 0.3675937122128174, "grad_norm": 0.4463306665420532, "learning_rate": 4.3840234889360634e-05, "loss": 1.7878, "step": 76 }, { "epoch": 0.3724304715840387, "grad_norm": 0.41338077187538147, "learning_rate": 4.33982642055842e-05, "loss": 1.7056, "step": 77 }, { "epoch": 0.37726723095526, "grad_norm": 0.48970669507980347, "learning_rate": 4.2952630570154785e-05, "loss": 1.767, "step": 78 }, { "epoch": 0.38210399032648124, "grad_norm": 0.47082316875457764, "learning_rate": 4.250345581485871e-05, "loss": 1.8836, "step": 79 }, { "epoch": 0.38694074969770254, "grad_norm": 0.40101462602615356, "learning_rate": 4.205086273958909e-05, "loss": 1.7996, "step": 80 }, { "epoch": 0.39177750906892383, "grad_norm": 0.49976804852485657, "learning_rate": 4.1594975078773565e-05, "loss": 1.8354, "step": 81 }, { "epoch": 0.3966142684401451, "grad_norm": 0.43082740902900696, "learning_rate": 4.113591746754662e-05, "loss": 1.7727, "step": 82 }, { "epoch": 0.4014510278113664, "grad_norm": 0.430978924036026, "learning_rate": 4.06738154076755e-05, "loss": 1.6518, "step": 83 }, { "epoch": 0.4062877871825877, "grad_norm": 0.5494586825370789, "learning_rate": 4.020879523324929e-05, "loss": 1.8428, "step": 84 }, { "epoch": 0.4111245465538089, "grad_norm": 0.5254581570625305, "learning_rate": 3.974098407614051e-05, "loss": 1.7367, "step": 85 }, { "epoch": 0.4159613059250302, "grad_norm": 0.45177075266838074, "learning_rate": 3.927050983124842e-05, "loss": 1.8426, "step": 86 }, { "epoch": 0.4207980652962515, "grad_norm": 0.5090339779853821, "learning_rate": 3.8797501121533946e-05, "loss": 1.8293, "step": 87 }, { "epoch": 0.42563482466747277, "grad_norm": 0.4986412525177002, "learning_rate": 3.832208726285534e-05, "loss": 1.7828, "step": 88 }, { "epoch": 0.43047158403869407, "grad_norm": 0.42650216817855835, "learning_rate": 3.784439822861459e-05, "loss": 1.7109, "step": 89 }, { "epoch": 0.43530834340991537, "grad_norm": 0.4212074279785156, "learning_rate": 3.7364564614223976e-05, "loss": 1.8452, "step": 90 }, { "epoch": 0.44014510278113667, "grad_norm": 0.4842492938041687, "learning_rate": 3.688271760140255e-05, "loss": 1.8276, "step": 91 }, { "epoch": 0.4449818621523579, "grad_norm": 0.4585839509963989, "learning_rate": 3.6398988922312406e-05, "loss": 1.8796, "step": 92 }, { "epoch": 0.4498186215235792, "grad_norm": 0.48161616921424866, "learning_rate": 3.591351082354441e-05, "loss": 1.8683, "step": 93 }, { "epoch": 0.4546553808948005, "grad_norm": 0.5221933722496033, "learning_rate": 3.54264160299633e-05, "loss": 1.8533, "step": 94 }, { "epoch": 0.45949214026602175, "grad_norm": 0.5291529893875122, "learning_rate": 3.493783770842202e-05, "loss": 1.7899, "step": 95 }, { "epoch": 0.46432889963724305, "grad_norm": 0.6196396350860596, "learning_rate": 3.444790943135526e-05, "loss": 1.7449, "step": 96 }, { "epoch": 0.46916565900846435, "grad_norm": 0.7176190614700317, "learning_rate": 3.3956765140262074e-05, "loss": 1.8663, "step": 97 }, { "epoch": 0.4740024183796856, "grad_norm": 0.6267790198326111, "learning_rate": 3.346453910908759e-05, "loss": 1.7183, "step": 98 }, { "epoch": 0.4788391777509069, "grad_norm": 0.9203118085861206, "learning_rate": 3.297136590751389e-05, "loss": 1.8954, "step": 99 }, { "epoch": 0.4836759371221282, "grad_norm": 0.8572964072227478, "learning_rate": 3.247738036416998e-05, "loss": 1.7169, "step": 100 }, { "epoch": 0.4836759371221282, "eval_loss": 1.7830288410186768, "eval_runtime": 18.3415, "eval_samples_per_second": 18.973, "eval_steps_per_second": 4.743, "step": 100 }, { "epoch": 0.48851269649334944, "grad_norm": 0.4275000989437103, "learning_rate": 3.1982717529770985e-05, "loss": 1.9067, "step": 101 }, { "epoch": 0.49334945586457074, "grad_norm": 0.48399701714515686, "learning_rate": 3.148751264019667e-05, "loss": 1.9673, "step": 102 }, { "epoch": 0.49818621523579204, "grad_norm": 0.45135805010795593, "learning_rate": 3.099190107951924e-05, "loss": 1.8383, "step": 103 }, { "epoch": 0.5030229746070133, "grad_norm": 0.476682186126709, "learning_rate": 3.049601834299076e-05, "loss": 1.8095, "step": 104 }, { "epoch": 0.5078597339782346, "grad_norm": 0.46085941791534424, "learning_rate": 3e-05, "loss": 1.6057, "step": 105 }, { "epoch": 0.5126964933494559, "grad_norm": 0.5784292221069336, "learning_rate": 2.9503981657009246e-05, "loss": 1.7255, "step": 106 }, { "epoch": 0.5175332527206772, "grad_norm": 0.5619342923164368, "learning_rate": 2.9008098920480752e-05, "loss": 1.9666, "step": 107 }, { "epoch": 0.5223700120918985, "grad_norm": 0.5032961368560791, "learning_rate": 2.851248735980333e-05, "loss": 1.8762, "step": 108 }, { "epoch": 0.5272067714631197, "grad_norm": 0.4964190721511841, "learning_rate": 2.801728247022902e-05, "loss": 1.9043, "step": 109 }, { "epoch": 0.532043530834341, "grad_norm": 0.5014073848724365, "learning_rate": 2.7522619635830034e-05, "loss": 1.6769, "step": 110 }, { "epoch": 0.5368802902055623, "grad_norm": 0.5301333665847778, "learning_rate": 2.702863409248612e-05, "loss": 1.9171, "step": 111 }, { "epoch": 0.5417170495767836, "grad_norm": 0.5182745456695557, "learning_rate": 2.6535460890912416e-05, "loss": 1.8716, "step": 112 }, { "epoch": 0.5465538089480049, "grad_norm": 0.46317240595817566, "learning_rate": 2.604323485973793e-05, "loss": 1.7914, "step": 113 }, { "epoch": 0.5513905683192262, "grad_norm": 0.43662911653518677, "learning_rate": 2.555209056864474e-05, "loss": 1.7344, "step": 114 }, { "epoch": 0.5562273276904474, "grad_norm": 0.48569121956825256, "learning_rate": 2.5062162291577978e-05, "loss": 1.7216, "step": 115 }, { "epoch": 0.5610640870616687, "grad_norm": 0.4334430992603302, "learning_rate": 2.4573583970036712e-05, "loss": 1.7228, "step": 116 }, { "epoch": 0.56590084643289, "grad_norm": 0.4561009705066681, "learning_rate": 2.4086489176455595e-05, "loss": 1.7068, "step": 117 }, { "epoch": 0.5707376058041113, "grad_norm": 0.42410483956336975, "learning_rate": 2.36010110776876e-05, "loss": 1.7405, "step": 118 }, { "epoch": 0.5755743651753326, "grad_norm": 0.46949705481529236, "learning_rate": 2.3117282398597456e-05, "loss": 1.7154, "step": 119 }, { "epoch": 0.5804111245465539, "grad_norm": 0.4945320785045624, "learning_rate": 2.263543538577603e-05, "loss": 1.7313, "step": 120 }, { "epoch": 0.585247883917775, "grad_norm": 0.49296748638153076, "learning_rate": 2.215560177138541e-05, "loss": 1.8307, "step": 121 }, { "epoch": 0.5900846432889963, "grad_norm": 0.5368390083312988, "learning_rate": 2.167791273714467e-05, "loss": 1.7453, "step": 122 }, { "epoch": 0.5949214026602176, "grad_norm": 0.47453513741493225, "learning_rate": 2.1202498878466062e-05, "loss": 1.7756, "step": 123 }, { "epoch": 0.599758162031439, "grad_norm": 0.5000871419906616, "learning_rate": 2.072949016875158e-05, "loss": 1.628, "step": 124 }, { "epoch": 0.6045949214026602, "grad_norm": 0.5026735663414001, "learning_rate": 2.0259015923859498e-05, "loss": 1.8197, "step": 125 }, { "epoch": 0.6094316807738815, "grad_norm": 0.47887665033340454, "learning_rate": 1.979120476675071e-05, "loss": 1.7951, "step": 126 }, { "epoch": 0.6142684401451027, "grad_norm": 0.46891364455223083, "learning_rate": 1.9326184592324503e-05, "loss": 1.6736, "step": 127 }, { "epoch": 0.619105199516324, "grad_norm": 0.5452415943145752, "learning_rate": 1.8864082532453373e-05, "loss": 1.8137, "step": 128 }, { "epoch": 0.6239419588875453, "grad_norm": 0.5273683667182922, "learning_rate": 1.840502492122644e-05, "loss": 1.7407, "step": 129 }, { "epoch": 0.6287787182587666, "grad_norm": 0.47598665952682495, "learning_rate": 1.7949137260410924e-05, "loss": 1.7605, "step": 130 }, { "epoch": 0.6336154776299879, "grad_norm": 0.5396198034286499, "learning_rate": 1.7496544185141295e-05, "loss": 1.7761, "step": 131 }, { "epoch": 0.6384522370012092, "grad_norm": 0.48490336537361145, "learning_rate": 1.7047369429845216e-05, "loss": 1.6718, "step": 132 }, { "epoch": 0.6432889963724304, "grad_norm": 0.4524966776371002, "learning_rate": 1.6601735794415806e-05, "loss": 1.6303, "step": 133 }, { "epoch": 0.6481257557436517, "grad_norm": 0.4709789454936981, "learning_rate": 1.615976511063937e-05, "loss": 1.7838, "step": 134 }, { "epoch": 0.652962515114873, "grad_norm": 0.46935707330703735, "learning_rate": 1.5721578208887793e-05, "loss": 1.6582, "step": 135 }, { "epoch": 0.6577992744860943, "grad_norm": 0.5257735252380371, "learning_rate": 1.5287294885084766e-05, "loss": 1.7776, "step": 136 }, { "epoch": 0.6626360338573156, "grad_norm": 0.5525801777839661, "learning_rate": 1.4857033867954697e-05, "loss": 1.8144, "step": 137 }, { "epoch": 0.6674727932285369, "grad_norm": 0.4242516756057739, "learning_rate": 1.4430912786563554e-05, "loss": 1.7099, "step": 138 }, { "epoch": 0.6723095525997581, "grad_norm": 0.5075966715812683, "learning_rate": 1.4009048138160195e-05, "loss": 1.7698, "step": 139 }, { "epoch": 0.6771463119709794, "grad_norm": 0.5571949481964111, "learning_rate": 1.3591555256327199e-05, "loss": 1.7397, "step": 140 }, { "epoch": 0.6819830713422007, "grad_norm": 0.5257452726364136, "learning_rate": 1.3178548279449822e-05, "loss": 1.7735, "step": 141 }, { "epoch": 0.686819830713422, "grad_norm": 0.4458635747432709, "learning_rate": 1.2770140119511693e-05, "loss": 1.8301, "step": 142 }, { "epoch": 0.6916565900846433, "grad_norm": 0.5735492706298828, "learning_rate": 1.2366442431225809e-05, "loss": 1.7515, "step": 143 }, { "epoch": 0.6964933494558646, "grad_norm": 0.5404819250106812, "learning_rate": 1.1967565581509248e-05, "loss": 1.8213, "step": 144 }, { "epoch": 0.7013301088270859, "grad_norm": 0.6705479621887207, "learning_rate": 1.1573618619309965e-05, "loss": 1.7866, "step": 145 }, { "epoch": 0.7061668681983071, "grad_norm": 0.7792295217514038, "learning_rate": 1.1184709245793889e-05, "loss": 1.7902, "step": 146 }, { "epoch": 0.7110036275695284, "grad_norm": 0.8956598043441772, "learning_rate": 1.0800943784900502e-05, "loss": 1.8193, "step": 147 }, { "epoch": 0.7158403869407497, "grad_norm": 0.7993762493133545, "learning_rate": 1.042242715427494e-05, "loss": 1.688, "step": 148 }, { "epoch": 0.720677146311971, "grad_norm": 0.8420385122299194, "learning_rate": 1.004926283658455e-05, "loss": 1.8126, "step": 149 }, { "epoch": 0.7255139056831923, "grad_norm": 0.9537421464920044, "learning_rate": 9.681552851227774e-06, "loss": 1.4474, "step": 150 }, { "epoch": 0.7255139056831923, "eval_loss": 1.7413969039916992, "eval_runtime": 18.3528, "eval_samples_per_second": 18.962, "eval_steps_per_second": 4.74, "step": 150 }, { "epoch": 0.7303506650544136, "grad_norm": 0.46916308999061584, "learning_rate": 9.319397726443026e-06, "loss": 2.028, "step": 151 }, { "epoch": 0.7351874244256348, "grad_norm": 0.43807369470596313, "learning_rate": 8.962896471825342e-06, "loss": 1.7917, "step": 152 }, { "epoch": 0.7400241837968561, "grad_norm": 0.530551016330719, "learning_rate": 8.61214655125809e-06, "loss": 1.8804, "step": 153 }, { "epoch": 0.7448609431680774, "grad_norm": 0.5489689707756042, "learning_rate": 8.267243856267331e-06, "loss": 1.8999, "step": 154 }, { "epoch": 0.7496977025392987, "grad_norm": 0.47269207239151, "learning_rate": 7.928282679806052e-06, "loss": 1.6923, "step": 155 }, { "epoch": 0.75453446191052, "grad_norm": 0.5164920687675476, "learning_rate": 7.595355690475393e-06, "loss": 1.7004, "step": 156 }, { "epoch": 0.7593712212817413, "grad_norm": 0.4851786494255066, "learning_rate": 7.268553907189964e-06, "loss": 1.6288, "step": 157 }, { "epoch": 0.7642079806529625, "grad_norm": 0.5359386801719666, "learning_rate": 6.947966674294236e-06, "loss": 1.6777, "step": 158 }, { "epoch": 0.7690447400241838, "grad_norm": 0.5173937678337097, "learning_rate": 6.6336816371366305e-06, "loss": 1.6535, "step": 159 }, { "epoch": 0.7738814993954051, "grad_norm": 0.4466940462589264, "learning_rate": 6.325784718108196e-06, "loss": 1.8034, "step": 160 }, { "epoch": 0.7787182587666264, "grad_norm": 0.4748440086841583, "learning_rate": 6.0243600931522595e-06, "loss": 1.7022, "step": 161 }, { "epoch": 0.7835550181378477, "grad_norm": 0.49979591369628906, "learning_rate": 5.72949016875158e-06, "loss": 1.6983, "step": 162 }, { "epoch": 0.788391777509069, "grad_norm": 0.47810453176498413, "learning_rate": 5.44125555939923e-06, "loss": 1.7931, "step": 163 }, { "epoch": 0.7932285368802902, "grad_norm": 0.444297194480896, "learning_rate": 5.159735065559399e-06, "loss": 1.785, "step": 164 }, { "epoch": 0.7980652962515115, "grad_norm": 0.4740820825099945, "learning_rate": 4.885005652124144e-06, "loss": 1.7403, "step": 165 }, { "epoch": 0.8029020556227328, "grad_norm": 0.4951646625995636, "learning_rate": 4.617142427371934e-06, "loss": 1.7834, "step": 166 }, { "epoch": 0.8077388149939541, "grad_norm": 0.49473491311073303, "learning_rate": 4.3562186224338265e-06, "loss": 1.7891, "step": 167 }, { "epoch": 0.8125755743651754, "grad_norm": 0.5567054152488708, "learning_rate": 4.102305571272783e-06, "loss": 1.742, "step": 168 }, { "epoch": 0.8174123337363967, "grad_norm": 0.531502366065979, "learning_rate": 3.855472691181678e-06, "loss": 1.7225, "step": 169 }, { "epoch": 0.8222490931076178, "grad_norm": 0.4598694443702698, "learning_rate": 3.615787463805331e-06, "loss": 1.6636, "step": 170 }, { "epoch": 0.8270858524788391, "grad_norm": 0.46727102994918823, "learning_rate": 3.383315416691646e-06, "loss": 1.6728, "step": 171 }, { "epoch": 0.8319226118500604, "grad_norm": 0.4816998541355133, "learning_rate": 3.158120105377096e-06, "loss": 1.7808, "step": 172 }, { "epoch": 0.8367593712212817, "grad_norm": 0.4520528316497803, "learning_rate": 2.940263096011233e-06, "loss": 1.573, "step": 173 }, { "epoch": 0.841596130592503, "grad_norm": 0.4842596650123596, "learning_rate": 2.729803948525125e-06, "loss": 1.838, "step": 174 }, { "epoch": 0.8464328899637243, "grad_norm": 0.5535080432891846, "learning_rate": 2.526800200348275e-06, "loss": 1.7461, "step": 175 }, { "epoch": 0.8512696493349455, "grad_norm": 0.5143581032752991, "learning_rate": 2.3313073506784575e-06, "loss": 1.6, "step": 176 }, { "epoch": 0.8561064087061668, "grad_norm": 0.5159276127815247, "learning_rate": 2.143378845308791e-06, "loss": 1.6323, "step": 177 }, { "epoch": 0.8609431680773881, "grad_norm": 0.4979780316352844, "learning_rate": 1.9630660620161777e-06, "loss": 1.779, "step": 178 }, { "epoch": 0.8657799274486094, "grad_norm": 0.5675414800643921, "learning_rate": 1.790418296515165e-06, "loss": 1.7153, "step": 179 }, { "epoch": 0.8706166868198307, "grad_norm": 0.4874797761440277, "learning_rate": 1.625482748980961e-06, "loss": 1.7539, "step": 180 }, { "epoch": 0.875453446191052, "grad_norm": 0.5073657631874084, "learning_rate": 1.4683045111453942e-06, "loss": 1.6124, "step": 181 }, { "epoch": 0.8802902055622733, "grad_norm": 0.4889651834964752, "learning_rate": 1.3189265539692707e-06, "loss": 1.6506, "step": 182 }, { "epoch": 0.8851269649334945, "grad_norm": 0.6806184649467468, "learning_rate": 1.1773897158945557e-06, "loss": 1.6727, "step": 183 }, { "epoch": 0.8899637243047158, "grad_norm": 0.5414454936981201, "learning_rate": 1.0437326916795432e-06, "loss": 1.7342, "step": 184 }, { "epoch": 0.8948004836759371, "grad_norm": 0.42585650086402893, "learning_rate": 9.179920218200888e-07, "loss": 1.6879, "step": 185 }, { "epoch": 0.8996372430471584, "grad_norm": 0.45129069685935974, "learning_rate": 8.002020825598277e-07, "loss": 1.6605, "step": 186 }, { "epoch": 0.9044740024183797, "grad_norm": 0.6295913457870483, "learning_rate": 6.90395076492022e-07, "loss": 1.7574, "step": 187 }, { "epoch": 0.909310761789601, "grad_norm": 0.4469881057739258, "learning_rate": 5.886010237557194e-07, "loss": 1.5844, "step": 188 }, { "epoch": 0.9141475211608222, "grad_norm": 0.5988472700119019, "learning_rate": 4.94847753828529e-07, "loss": 1.9071, "step": 189 }, { "epoch": 0.9189842805320435, "grad_norm": 0.4462520182132721, "learning_rate": 4.091608979183303e-07, "loss": 1.7049, "step": 190 }, { "epoch": 0.9238210399032648, "grad_norm": 0.5181970596313477, "learning_rate": 3.315638819559452e-07, "loss": 1.8397, "step": 191 }, { "epoch": 0.9286577992744861, "grad_norm": 0.5435971617698669, "learning_rate": 2.6207792019074414e-07, "loss": 1.6886, "step": 192 }, { "epoch": 0.9334945586457074, "grad_norm": 0.5636361241340637, "learning_rate": 2.0072200939085573e-07, "loss": 1.6915, "step": 193 }, { "epoch": 0.9383313180169287, "grad_norm": 0.5563094615936279, "learning_rate": 1.475129236496575e-07, "loss": 1.7518, "step": 194 }, { "epoch": 0.9431680773881499, "grad_norm": 0.5630325078964233, "learning_rate": 1.0246520979990459e-07, "loss": 1.8014, "step": 195 }, { "epoch": 0.9480048367593712, "grad_norm": 0.6778345108032227, "learning_rate": 6.559118343676396e-08, "loss": 1.6761, "step": 196 }, { "epoch": 0.9528415961305925, "grad_norm": 0.8598151803016663, "learning_rate": 3.690092555085789e-08, "loss": 1.7755, "step": 197 }, { "epoch": 0.9576783555018138, "grad_norm": 0.8544036149978638, "learning_rate": 1.640227977221853e-08, "loss": 1.7744, "step": 198 }, { "epoch": 0.9625151148730351, "grad_norm": 0.988280177116394, "learning_rate": 4.1008502259298755e-09, "loss": 1.751, "step": 199 }, { "epoch": 0.9673518742442564, "grad_norm": 1.538628101348877, "learning_rate": 0.0, "loss": 1.4039, "step": 200 }, { "epoch": 0.9673518742442564, "eval_loss": 1.7320468425750732, "eval_runtime": 18.3522, "eval_samples_per_second": 18.962, "eval_steps_per_second": 4.741, "step": 200 } ], "logging_steps": 1, "max_steps": 200, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 50, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 4, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 0 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.582807765745664e+17, "train_batch_size": 8, "trial_name": null, "trial_params": null }