|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.9950576606260295, |
|
"eval_steps": 500, |
|
"global_step": 909, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0032948929159802307, |
|
"grad_norm": 0.8727272666018974, |
|
"learning_rate": 2.197802197802198e-06, |
|
"loss": 1.2966, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.006589785831960461, |
|
"grad_norm": 0.9006388360232964, |
|
"learning_rate": 4.395604395604396e-06, |
|
"loss": 1.3333, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.009884678747940691, |
|
"grad_norm": 0.9149925703883882, |
|
"learning_rate": 6.5934065934065935e-06, |
|
"loss": 1.3275, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.013179571663920923, |
|
"grad_norm": 0.9362223066609972, |
|
"learning_rate": 8.791208791208792e-06, |
|
"loss": 1.3456, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.016474464579901153, |
|
"grad_norm": 0.8362157088104918, |
|
"learning_rate": 1.0989010989010989e-05, |
|
"loss": 1.2551, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.019769357495881382, |
|
"grad_norm": 0.8409114192243606, |
|
"learning_rate": 1.3186813186813187e-05, |
|
"loss": 1.2888, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.023064250411861616, |
|
"grad_norm": 0.8323916919747559, |
|
"learning_rate": 1.5384615384615387e-05, |
|
"loss": 1.2524, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.026359143327841845, |
|
"grad_norm": 0.7826776515036836, |
|
"learning_rate": 1.7582417582417584e-05, |
|
"loss": 1.2206, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.029654036243822075, |
|
"grad_norm": 0.7679461778818735, |
|
"learning_rate": 1.978021978021978e-05, |
|
"loss": 1.238, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.032948929159802305, |
|
"grad_norm": 0.6218001630318019, |
|
"learning_rate": 2.1978021978021977e-05, |
|
"loss": 1.144, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.036243822075782535, |
|
"grad_norm": 0.550844391887277, |
|
"learning_rate": 2.4175824175824177e-05, |
|
"loss": 1.1091, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.039538714991762765, |
|
"grad_norm": 0.5376512267561467, |
|
"learning_rate": 2.6373626373626374e-05, |
|
"loss": 1.1048, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.042833607907743, |
|
"grad_norm": 0.5061475154817852, |
|
"learning_rate": 2.857142857142857e-05, |
|
"loss": 1.0239, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.04612850082372323, |
|
"grad_norm": 0.5440235934771381, |
|
"learning_rate": 3.0769230769230774e-05, |
|
"loss": 0.9955, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.04942339373970346, |
|
"grad_norm": 0.5710584188811608, |
|
"learning_rate": 3.296703296703297e-05, |
|
"loss": 0.9453, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.05271828665568369, |
|
"grad_norm": 0.5629646386751581, |
|
"learning_rate": 3.516483516483517e-05, |
|
"loss": 0.8922, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.05601317957166392, |
|
"grad_norm": 0.5787661460532081, |
|
"learning_rate": 3.7362637362637365e-05, |
|
"loss": 0.853, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.05930807248764415, |
|
"grad_norm": 0.5219200821941683, |
|
"learning_rate": 3.956043956043956e-05, |
|
"loss": 0.8057, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.06260296540362438, |
|
"grad_norm": 0.4949320198088838, |
|
"learning_rate": 4.1758241758241765e-05, |
|
"loss": 0.7203, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.06589785831960461, |
|
"grad_norm": 0.5174702935923171, |
|
"learning_rate": 4.3956043956043955e-05, |
|
"loss": 0.7028, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.06919275123558484, |
|
"grad_norm": 0.4399438447558331, |
|
"learning_rate": 4.615384615384616e-05, |
|
"loss": 0.6712, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.07248764415156507, |
|
"grad_norm": 0.3850468864162824, |
|
"learning_rate": 4.8351648351648355e-05, |
|
"loss": 0.6205, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.0757825370675453, |
|
"grad_norm": 0.327866579681946, |
|
"learning_rate": 5.054945054945055e-05, |
|
"loss": 0.5923, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.07907742998352553, |
|
"grad_norm": 0.24857639537783982, |
|
"learning_rate": 5.274725274725275e-05, |
|
"loss": 0.5706, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.08237232289950576, |
|
"grad_norm": 0.24508430082340485, |
|
"learning_rate": 5.494505494505495e-05, |
|
"loss": 0.5534, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.085667215815486, |
|
"grad_norm": 0.21401457324982626, |
|
"learning_rate": 5.714285714285714e-05, |
|
"loss": 0.5343, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.08896210873146623, |
|
"grad_norm": 0.19432840487037037, |
|
"learning_rate": 5.9340659340659345e-05, |
|
"loss": 0.5205, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.09225700164744646, |
|
"grad_norm": 0.25188173455273966, |
|
"learning_rate": 6.153846153846155e-05, |
|
"loss": 0.517, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.09555189456342669, |
|
"grad_norm": 0.19175435794663873, |
|
"learning_rate": 6.373626373626373e-05, |
|
"loss": 0.5177, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.09884678747940692, |
|
"grad_norm": 0.22909359876581833, |
|
"learning_rate": 6.593406593406594e-05, |
|
"loss": 0.507, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.10214168039538715, |
|
"grad_norm": 0.21953419980838787, |
|
"learning_rate": 6.813186813186814e-05, |
|
"loss": 0.4935, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.10543657331136738, |
|
"grad_norm": 0.20750358706332223, |
|
"learning_rate": 7.032967032967034e-05, |
|
"loss": 0.4987, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.10873146622734761, |
|
"grad_norm": 0.19232398536892564, |
|
"learning_rate": 7.252747252747253e-05, |
|
"loss": 0.4993, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.11202635914332784, |
|
"grad_norm": 0.1775783275207957, |
|
"learning_rate": 7.472527472527473e-05, |
|
"loss": 0.4854, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.11532125205930807, |
|
"grad_norm": 0.14336113752224602, |
|
"learning_rate": 7.692307692307693e-05, |
|
"loss": 0.4659, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.1186161449752883, |
|
"grad_norm": 0.14129976475916678, |
|
"learning_rate": 7.912087912087912e-05, |
|
"loss": 0.4634, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.12191103789126853, |
|
"grad_norm": 0.12988205851707507, |
|
"learning_rate": 8.131868131868132e-05, |
|
"loss": 0.4648, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.12520593080724876, |
|
"grad_norm": 0.13738530280987127, |
|
"learning_rate": 8.351648351648353e-05, |
|
"loss": 0.4772, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.128500823723229, |
|
"grad_norm": 0.11545159705042944, |
|
"learning_rate": 8.571428571428571e-05, |
|
"loss": 0.4728, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.13179571663920922, |
|
"grad_norm": 0.11033726958873326, |
|
"learning_rate": 8.791208791208791e-05, |
|
"loss": 0.4621, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.13509060955518945, |
|
"grad_norm": 0.1201472253052247, |
|
"learning_rate": 9.010989010989012e-05, |
|
"loss": 0.4576, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.13838550247116968, |
|
"grad_norm": 0.11436096928615554, |
|
"learning_rate": 9.230769230769232e-05, |
|
"loss": 0.4371, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.1416803953871499, |
|
"grad_norm": 0.11149821850019738, |
|
"learning_rate": 9.450549450549451e-05, |
|
"loss": 0.4484, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.14497528830313014, |
|
"grad_norm": 0.1058121629816968, |
|
"learning_rate": 9.670329670329671e-05, |
|
"loss": 0.4389, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.14827018121911037, |
|
"grad_norm": 0.10501103645473217, |
|
"learning_rate": 9.89010989010989e-05, |
|
"loss": 0.4335, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.1515650741350906, |
|
"grad_norm": 0.0964684052416727, |
|
"learning_rate": 0.0001010989010989011, |
|
"loss": 0.4394, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.15485996705107083, |
|
"grad_norm": 0.10135506801778058, |
|
"learning_rate": 0.00010329670329670331, |
|
"loss": 0.4284, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.15815485996705106, |
|
"grad_norm": 0.10029097763419231, |
|
"learning_rate": 0.0001054945054945055, |
|
"loss": 0.4371, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.1614497528830313, |
|
"grad_norm": 0.10052795766573124, |
|
"learning_rate": 0.0001076923076923077, |
|
"loss": 0.4293, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.16474464579901152, |
|
"grad_norm": 0.10420997329264613, |
|
"learning_rate": 0.0001098901098901099, |
|
"loss": 0.4279, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.16803953871499178, |
|
"grad_norm": 0.1217798536526884, |
|
"learning_rate": 0.0001120879120879121, |
|
"loss": 0.4141, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.171334431630972, |
|
"grad_norm": 0.10417010160655436, |
|
"learning_rate": 0.00011428571428571428, |
|
"loss": 0.41, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.17462932454695224, |
|
"grad_norm": 0.10296977506815383, |
|
"learning_rate": 0.0001164835164835165, |
|
"loss": 0.4299, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.17792421746293247, |
|
"grad_norm": 0.10122420990330976, |
|
"learning_rate": 0.00011868131868131869, |
|
"loss": 0.4134, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.1812191103789127, |
|
"grad_norm": 0.09986057816216841, |
|
"learning_rate": 0.00012087912087912087, |
|
"loss": 0.4144, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.18451400329489293, |
|
"grad_norm": 0.10113082205051421, |
|
"learning_rate": 0.0001230769230769231, |
|
"loss": 0.4154, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.18780889621087316, |
|
"grad_norm": 0.09878123227751724, |
|
"learning_rate": 0.00012527472527472527, |
|
"loss": 0.4181, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.19110378912685339, |
|
"grad_norm": 0.10418790342701288, |
|
"learning_rate": 0.00012747252747252746, |
|
"loss": 0.4204, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.19439868204283361, |
|
"grad_norm": 0.10614228869769458, |
|
"learning_rate": 0.0001296703296703297, |
|
"loss": 0.428, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.19769357495881384, |
|
"grad_norm": 0.09964224792215227, |
|
"learning_rate": 0.00013186813186813188, |
|
"loss": 0.3888, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.20098846787479407, |
|
"grad_norm": 0.1060800549771199, |
|
"learning_rate": 0.00013406593406593405, |
|
"loss": 0.4045, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.2042833607907743, |
|
"grad_norm": 0.10220349513246392, |
|
"learning_rate": 0.00013626373626373628, |
|
"loss": 0.407, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.20757825370675453, |
|
"grad_norm": 0.10499038993641154, |
|
"learning_rate": 0.00013846153846153847, |
|
"loss": 0.4042, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.21087314662273476, |
|
"grad_norm": 0.10008875757336905, |
|
"learning_rate": 0.00014065934065934067, |
|
"loss": 0.3968, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.214168039538715, |
|
"grad_norm": 0.10767447090788189, |
|
"learning_rate": 0.00014285714285714287, |
|
"loss": 0.4096, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.21746293245469522, |
|
"grad_norm": 0.10409029011454787, |
|
"learning_rate": 0.00014505494505494506, |
|
"loss": 0.3925, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.22075782537067545, |
|
"grad_norm": 0.10006210224544987, |
|
"learning_rate": 0.00014725274725274726, |
|
"loss": 0.3733, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.22405271828665568, |
|
"grad_norm": 0.10154215775645745, |
|
"learning_rate": 0.00014945054945054946, |
|
"loss": 0.4008, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.2273476112026359, |
|
"grad_norm": 0.10280194055666982, |
|
"learning_rate": 0.00015164835164835165, |
|
"loss": 0.4074, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.23064250411861614, |
|
"grad_norm": 0.10449991568283325, |
|
"learning_rate": 0.00015384615384615385, |
|
"loss": 0.4007, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.23393739703459637, |
|
"grad_norm": 0.09867099523694993, |
|
"learning_rate": 0.00015604395604395605, |
|
"loss": 0.3945, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.2372322899505766, |
|
"grad_norm": 0.10364637285833565, |
|
"learning_rate": 0.00015824175824175824, |
|
"loss": 0.3756, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.24052718286655683, |
|
"grad_norm": 0.10500702289024652, |
|
"learning_rate": 0.00016043956043956044, |
|
"loss": 0.3976, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.24382207578253706, |
|
"grad_norm": 0.10508505128672692, |
|
"learning_rate": 0.00016263736263736264, |
|
"loss": 0.3936, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.2471169686985173, |
|
"grad_norm": 0.11077255112820204, |
|
"learning_rate": 0.00016483516483516484, |
|
"loss": 0.3844, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.2504118616144975, |
|
"grad_norm": 0.11497427991652769, |
|
"learning_rate": 0.00016703296703296706, |
|
"loss": 0.4002, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.25370675453047775, |
|
"grad_norm": 0.10361858110894616, |
|
"learning_rate": 0.00016923076923076923, |
|
"loss": 0.3877, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.257001647446458, |
|
"grad_norm": 0.10689216566545769, |
|
"learning_rate": 0.00017142857142857143, |
|
"loss": 0.375, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.2602965403624382, |
|
"grad_norm": 0.10379603284636893, |
|
"learning_rate": 0.00017362637362637365, |
|
"loss": 0.3855, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.26359143327841844, |
|
"grad_norm": 0.11008235355031443, |
|
"learning_rate": 0.00017582417582417582, |
|
"loss": 0.3928, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.26688632619439867, |
|
"grad_norm": 0.10801759571590709, |
|
"learning_rate": 0.00017802197802197802, |
|
"loss": 0.3796, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.2701812191103789, |
|
"grad_norm": 0.1072207208697183, |
|
"learning_rate": 0.00018021978021978024, |
|
"loss": 0.3806, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.27347611202635913, |
|
"grad_norm": 0.11074742721907231, |
|
"learning_rate": 0.0001824175824175824, |
|
"loss": 0.3759, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.27677100494233936, |
|
"grad_norm": 0.10332829028464724, |
|
"learning_rate": 0.00018461538461538463, |
|
"loss": 0.4013, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.2800658978583196, |
|
"grad_norm": 0.12121573096651414, |
|
"learning_rate": 0.00018681318681318683, |
|
"loss": 0.3718, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.2833607907742998, |
|
"grad_norm": 0.11029471534257396, |
|
"learning_rate": 0.00018901098901098903, |
|
"loss": 0.3957, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.28665568369028005, |
|
"grad_norm": 0.10611352834090172, |
|
"learning_rate": 0.00019120879120879122, |
|
"loss": 0.3775, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.2899505766062603, |
|
"grad_norm": 0.11444959045215879, |
|
"learning_rate": 0.00019340659340659342, |
|
"loss": 0.3911, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.2932454695222405, |
|
"grad_norm": 0.11698929258412147, |
|
"learning_rate": 0.00019560439560439562, |
|
"loss": 0.3801, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.29654036243822074, |
|
"grad_norm": 0.11887085945257751, |
|
"learning_rate": 0.0001978021978021978, |
|
"loss": 0.3824, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.29983525535420097, |
|
"grad_norm": 0.12394486608854069, |
|
"learning_rate": 0.0002, |
|
"loss": 0.3713, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.3031301482701812, |
|
"grad_norm": 0.11524526961416616, |
|
"learning_rate": 0.00019999926249900844, |
|
"loss": 0.375, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.30642504118616143, |
|
"grad_norm": 0.10954878706910288, |
|
"learning_rate": 0.00019999705000691188, |
|
"loss": 0.373, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.30971993410214166, |
|
"grad_norm": 0.11180610022301836, |
|
"learning_rate": 0.00019999336255634465, |
|
"loss": 0.3767, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.3130148270181219, |
|
"grad_norm": 0.10423320337270955, |
|
"learning_rate": 0.00019998820020169668, |
|
"loss": 0.3726, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.3163097199341021, |
|
"grad_norm": 0.10584925582232764, |
|
"learning_rate": 0.00019998156301911284, |
|
"loss": 0.3955, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.31960461285008235, |
|
"grad_norm": 0.10917585284115876, |
|
"learning_rate": 0.00019997345110649167, |
|
"loss": 0.38, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.3228995057660626, |
|
"grad_norm": 0.11250538116569028, |
|
"learning_rate": 0.0001999638645834841, |
|
"loss": 0.3659, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.3261943986820428, |
|
"grad_norm": 0.10144424444540087, |
|
"learning_rate": 0.00019995280359149149, |
|
"loss": 0.3724, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.32948929159802304, |
|
"grad_norm": 0.11312360619640419, |
|
"learning_rate": 0.0001999402682936637, |
|
"loss": 0.3675, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.33278418451400327, |
|
"grad_norm": 0.11072192514968879, |
|
"learning_rate": 0.0001999262588748966, |
|
"loss": 0.3734, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.33607907742998355, |
|
"grad_norm": 0.10852569603049844, |
|
"learning_rate": 0.00019991077554182943, |
|
"loss": 0.3729, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.3393739703459638, |
|
"grad_norm": 0.10138437779435457, |
|
"learning_rate": 0.00019989381852284166, |
|
"loss": 0.368, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.342668863261944, |
|
"grad_norm": 0.11109618855106276, |
|
"learning_rate": 0.00019987538806804964, |
|
"loss": 0.3728, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.34596375617792424, |
|
"grad_norm": 0.10566572499718903, |
|
"learning_rate": 0.00019985548444930294, |
|
"loss": 0.3625, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.34925864909390447, |
|
"grad_norm": 0.10329823166968895, |
|
"learning_rate": 0.00019983410796018032, |
|
"loss": 0.3736, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.3525535420098847, |
|
"grad_norm": 0.10526340058049245, |
|
"learning_rate": 0.00019981125891598546, |
|
"loss": 0.3776, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.35584843492586493, |
|
"grad_norm": 0.10634239695176183, |
|
"learning_rate": 0.00019978693765374217, |
|
"loss": 0.3627, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.35914332784184516, |
|
"grad_norm": 0.10473623073599642, |
|
"learning_rate": 0.0001997611445321896, |
|
"loss": 0.3651, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.3624382207578254, |
|
"grad_norm": 0.1038114878244784, |
|
"learning_rate": 0.00019973387993177673, |
|
"loss": 0.3795, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.3657331136738056, |
|
"grad_norm": 0.097631843964531, |
|
"learning_rate": 0.00019970514425465705, |
|
"loss": 0.3758, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.36902800658978585, |
|
"grad_norm": 0.105170092111551, |
|
"learning_rate": 0.00019967493792468228, |
|
"loss": 0.3689, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.3723228995057661, |
|
"grad_norm": 0.10400530821586597, |
|
"learning_rate": 0.00019964326138739645, |
|
"loss": 0.3772, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.3756177924217463, |
|
"grad_norm": 0.09711231346227854, |
|
"learning_rate": 0.0001996101151100291, |
|
"loss": 0.3568, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.37891268533772654, |
|
"grad_norm": 0.10219375076601767, |
|
"learning_rate": 0.00019957549958148841, |
|
"loss": 0.3733, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.38220757825370677, |
|
"grad_norm": 0.10620115483842468, |
|
"learning_rate": 0.00019953941531235424, |
|
"loss": 0.3813, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.385502471169687, |
|
"grad_norm": 0.09890990719776012, |
|
"learning_rate": 0.0001995018628348702, |
|
"loss": 0.3588, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.38879736408566723, |
|
"grad_norm": 0.10332397343960774, |
|
"learning_rate": 0.00019946284270293605, |
|
"loss": 0.3745, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.39209225700164746, |
|
"grad_norm": 0.0962153448347187, |
|
"learning_rate": 0.00019942235549209953, |
|
"loss": 0.3691, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.3953871499176277, |
|
"grad_norm": 0.10293197017899637, |
|
"learning_rate": 0.00019938040179954785, |
|
"loss": 0.3629, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.3986820428336079, |
|
"grad_norm": 0.10120783240834415, |
|
"learning_rate": 0.00019933698224409876, |
|
"loss": 0.3583, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.40197693574958815, |
|
"grad_norm": 0.09823996102922933, |
|
"learning_rate": 0.0001992920974661916, |
|
"loss": 0.3551, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.4052718286655684, |
|
"grad_norm": 0.11608055320274326, |
|
"learning_rate": 0.00019924574812787768, |
|
"loss": 0.3776, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.4085667215815486, |
|
"grad_norm": 0.1015390883033684, |
|
"learning_rate": 0.00019919793491281069, |
|
"loss": 0.3658, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.41186161449752884, |
|
"grad_norm": 0.11305236549565593, |
|
"learning_rate": 0.0001991486585262365, |
|
"loss": 0.3698, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.41515650741350907, |
|
"grad_norm": 0.09858293841585426, |
|
"learning_rate": 0.0001990979196949828, |
|
"loss": 0.3655, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.4184514003294893, |
|
"grad_norm": 0.09867163357687168, |
|
"learning_rate": 0.00019904571916744837, |
|
"loss": 0.3589, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.42174629324546953, |
|
"grad_norm": 0.099878248246959, |
|
"learning_rate": 0.00019899205771359198, |
|
"loss": 0.3694, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.42504118616144976, |
|
"grad_norm": 0.09933371254810067, |
|
"learning_rate": 0.00019893693612492116, |
|
"loss": 0.3698, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.42833607907743, |
|
"grad_norm": 0.09374415766405315, |
|
"learning_rate": 0.00019888035521448044, |
|
"loss": 0.3618, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.4316309719934102, |
|
"grad_norm": 0.10862350391810786, |
|
"learning_rate": 0.00019882231581683937, |
|
"loss": 0.3597, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.43492586490939045, |
|
"grad_norm": 0.09819927590552285, |
|
"learning_rate": 0.0001987628187880802, |
|
"loss": 0.3592, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.4382207578253707, |
|
"grad_norm": 0.09799469540398538, |
|
"learning_rate": 0.0001987018650057853, |
|
"loss": 0.3548, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.4415156507413509, |
|
"grad_norm": 0.10446031673946489, |
|
"learning_rate": 0.0001986394553690242, |
|
"loss": 0.3739, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.44481054365733114, |
|
"grad_norm": 0.09368559137183911, |
|
"learning_rate": 0.00019857559079834022, |
|
"loss": 0.3646, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.44810543657331137, |
|
"grad_norm": 0.09651140610849232, |
|
"learning_rate": 0.0001985102722357371, |
|
"loss": 0.3546, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.4514003294892916, |
|
"grad_norm": 0.10543606439531415, |
|
"learning_rate": 0.00019844350064466486, |
|
"loss": 0.3644, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.4546952224052718, |
|
"grad_norm": 0.0949282387849649, |
|
"learning_rate": 0.00019837527701000587, |
|
"loss": 0.3604, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.45799011532125206, |
|
"grad_norm": 0.09906601410462178, |
|
"learning_rate": 0.00019830560233806006, |
|
"loss": 0.3648, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.4612850082372323, |
|
"grad_norm": 0.1027605621404487, |
|
"learning_rate": 0.00019823447765653023, |
|
"loss": 0.3482, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.4645799011532125, |
|
"grad_norm": 0.09699652288591568, |
|
"learning_rate": 0.0001981619040145068, |
|
"loss": 0.3545, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.46787479406919275, |
|
"grad_norm": 0.10545828026299639, |
|
"learning_rate": 0.00019808788248245249, |
|
"loss": 0.3719, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.471169686985173, |
|
"grad_norm": 0.0969238306005527, |
|
"learning_rate": 0.00019801241415218636, |
|
"loss": 0.356, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.4744645799011532, |
|
"grad_norm": 0.09451741240756485, |
|
"learning_rate": 0.00019793550013686773, |
|
"loss": 0.3463, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.47775947281713343, |
|
"grad_norm": 0.09646227177006125, |
|
"learning_rate": 0.0001978571415709799, |
|
"loss": 0.3644, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.48105436573311366, |
|
"grad_norm": 0.09993416821340104, |
|
"learning_rate": 0.00019777733961031326, |
|
"loss": 0.3498, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.4843492586490939, |
|
"grad_norm": 0.09746839978917406, |
|
"learning_rate": 0.00019769609543194826, |
|
"loss": 0.3428, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.4876441515650741, |
|
"grad_norm": 0.10241864087501074, |
|
"learning_rate": 0.00019761341023423826, |
|
"loss": 0.3677, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.49093904448105435, |
|
"grad_norm": 0.0930243618471461, |
|
"learning_rate": 0.00019752928523679143, |
|
"loss": 0.357, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.4942339373970346, |
|
"grad_norm": 0.09542173403956193, |
|
"learning_rate": 0.00019744372168045324, |
|
"loss": 0.3623, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.4975288303130148, |
|
"grad_norm": 0.092951368800601, |
|
"learning_rate": 0.00019735672082728782, |
|
"loss": 0.3502, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.500823723228995, |
|
"grad_norm": 0.09468069312701037, |
|
"learning_rate": 0.00019726828396055948, |
|
"loss": 0.3422, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.5041186161449753, |
|
"grad_norm": 0.09372935130232003, |
|
"learning_rate": 0.00019717841238471375, |
|
"loss": 0.3533, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.5074135090609555, |
|
"grad_norm": 0.09623089008115779, |
|
"learning_rate": 0.00019708710742535814, |
|
"loss": 0.3521, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.5107084019769358, |
|
"grad_norm": 0.09433147739759222, |
|
"learning_rate": 0.00019699437042924265, |
|
"loss": 0.376, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.514003294892916, |
|
"grad_norm": 0.08740337875902095, |
|
"learning_rate": 0.0001969002027642398, |
|
"loss": 0.3419, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.5172981878088962, |
|
"grad_norm": 0.09306098636137253, |
|
"learning_rate": 0.00019680460581932447, |
|
"loss": 0.3578, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.5205930807248764, |
|
"grad_norm": 0.0901779549976545, |
|
"learning_rate": 0.00019670758100455356, |
|
"loss": 0.3597, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.5238879736408567, |
|
"grad_norm": 0.09418423029537448, |
|
"learning_rate": 0.000196609129751045, |
|
"loss": 0.3608, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.5271828665568369, |
|
"grad_norm": 0.09028368053330069, |
|
"learning_rate": 0.0001965092535109567, |
|
"loss": 0.3592, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.5304777594728172, |
|
"grad_norm": 0.0935932711587901, |
|
"learning_rate": 0.00019640795375746518, |
|
"loss": 0.3619, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.5337726523887973, |
|
"grad_norm": 0.09991159935467303, |
|
"learning_rate": 0.00019630523198474386, |
|
"loss": 0.3493, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.5370675453047776, |
|
"grad_norm": 0.09135500404430619, |
|
"learning_rate": 0.00019620108970794088, |
|
"loss": 0.3504, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.5403624382207578, |
|
"grad_norm": 0.09436717385152184, |
|
"learning_rate": 0.00019609552846315694, |
|
"loss": 0.3489, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.5436573311367381, |
|
"grad_norm": 0.10184053426464197, |
|
"learning_rate": 0.0001959885498074224, |
|
"loss": 0.3443, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.5469522240527183, |
|
"grad_norm": 0.09470012639193129, |
|
"learning_rate": 0.00019588015531867464, |
|
"loss": 0.3611, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.5502471169686985, |
|
"grad_norm": 0.0909557452721919, |
|
"learning_rate": 0.0001957703465957345, |
|
"loss": 0.3663, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.5535420098846787, |
|
"grad_norm": 0.09051693863828568, |
|
"learning_rate": 0.0001956591252582828, |
|
"loss": 0.3539, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.556836902800659, |
|
"grad_norm": 0.09596952767327611, |
|
"learning_rate": 0.0001955464929468365, |
|
"loss": 0.3432, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.5601317957166392, |
|
"grad_norm": 0.0908006040293162, |
|
"learning_rate": 0.00019543245132272441, |
|
"loss": 0.3518, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.5634266886326195, |
|
"grad_norm": 0.09261485040615038, |
|
"learning_rate": 0.00019531700206806274, |
|
"loss": 0.346, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.5667215815485996, |
|
"grad_norm": 0.08952858341904249, |
|
"learning_rate": 0.0001952001468857303, |
|
"loss": 0.341, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.5700164744645799, |
|
"grad_norm": 0.0912632841864883, |
|
"learning_rate": 0.00019508188749934333, |
|
"loss": 0.3369, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.5733113673805601, |
|
"grad_norm": 0.09208815220485178, |
|
"learning_rate": 0.00019496222565323015, |
|
"loss": 0.356, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.5766062602965404, |
|
"grad_norm": 0.08856220347094375, |
|
"learning_rate": 0.00019484116311240532, |
|
"loss": 0.3593, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.5799011532125206, |
|
"grad_norm": 0.09046349122843272, |
|
"learning_rate": 0.00019471870166254377, |
|
"loss": 0.3469, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.5831960461285008, |
|
"grad_norm": 0.09233531995170684, |
|
"learning_rate": 0.0001945948431099543, |
|
"loss": 0.3541, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.586490939044481, |
|
"grad_norm": 0.090106620935963, |
|
"learning_rate": 0.00019446958928155298, |
|
"loss": 0.3603, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.5897858319604613, |
|
"grad_norm": 0.0875097489689974, |
|
"learning_rate": 0.00019434294202483633, |
|
"loss": 0.359, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.5930807248764415, |
|
"grad_norm": 0.09123195997573252, |
|
"learning_rate": 0.00019421490320785384, |
|
"loss": 0.3565, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.5963756177924218, |
|
"grad_norm": 0.09114904646800603, |
|
"learning_rate": 0.00019408547471918061, |
|
"loss": 0.35, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.5996705107084019, |
|
"grad_norm": 0.08497423438629802, |
|
"learning_rate": 0.00019395465846788946, |
|
"loss": 0.3476, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.6029654036243822, |
|
"grad_norm": 0.09604855511067968, |
|
"learning_rate": 0.00019382245638352262, |
|
"loss": 0.3501, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.6062602965403624, |
|
"grad_norm": 0.0869157197559244, |
|
"learning_rate": 0.0001936888704160635, |
|
"loss": 0.3526, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.6095551894563427, |
|
"grad_norm": 0.08477176252928098, |
|
"learning_rate": 0.00019355390253590775, |
|
"loss": 0.3541, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.6128500823723229, |
|
"grad_norm": 0.089378043032419, |
|
"learning_rate": 0.00019341755473383432, |
|
"loss": 0.3558, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.6161449752883031, |
|
"grad_norm": 0.0844453554389038, |
|
"learning_rate": 0.00019327982902097595, |
|
"loss": 0.3505, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.6194398682042833, |
|
"grad_norm": 0.09223714783810881, |
|
"learning_rate": 0.00019314072742878963, |
|
"loss": 0.3467, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.6227347611202636, |
|
"grad_norm": 0.09154045104035775, |
|
"learning_rate": 0.00019300025200902666, |
|
"loss": 0.3485, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.6260296540362438, |
|
"grad_norm": 0.08978112849737604, |
|
"learning_rate": 0.0001928584048337022, |
|
"loss": 0.3535, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.6293245469522241, |
|
"grad_norm": 0.09219618310798985, |
|
"learning_rate": 0.00019271518799506492, |
|
"loss": 0.3492, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.6326194398682042, |
|
"grad_norm": 0.0878840125521925, |
|
"learning_rate": 0.00019257060360556606, |
|
"loss": 0.3517, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.6359143327841845, |
|
"grad_norm": 0.09421976864427346, |
|
"learning_rate": 0.00019242465379782823, |
|
"loss": 0.3592, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.6392092257001647, |
|
"grad_norm": 0.09096356033322509, |
|
"learning_rate": 0.00019227734072461392, |
|
"loss": 0.3576, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.642504118616145, |
|
"grad_norm": 0.08546178334325298, |
|
"learning_rate": 0.00019212866655879396, |
|
"loss": 0.3443, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.6457990115321252, |
|
"grad_norm": 0.08815077311380441, |
|
"learning_rate": 0.00019197863349331522, |
|
"loss": 0.35, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.6490939044481054, |
|
"grad_norm": 0.0909243070087766, |
|
"learning_rate": 0.00019182724374116838, |
|
"loss": 0.3477, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.6523887973640856, |
|
"grad_norm": 0.08716295830994963, |
|
"learning_rate": 0.0001916744995353553, |
|
"loss": 0.3425, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.6556836902800659, |
|
"grad_norm": 0.08437105411786357, |
|
"learning_rate": 0.00019152040312885604, |
|
"loss": 0.3473, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.6589785831960461, |
|
"grad_norm": 0.08999850132612741, |
|
"learning_rate": 0.00019136495679459564, |
|
"loss": 0.3446, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.6622734761120264, |
|
"grad_norm": 0.09295157213742279, |
|
"learning_rate": 0.00019120816282541063, |
|
"loss": 0.3509, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.6655683690280065, |
|
"grad_norm": 0.09096038036728704, |
|
"learning_rate": 0.00019105002353401516, |
|
"loss": 0.3462, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.6688632619439868, |
|
"grad_norm": 0.09051478487792422, |
|
"learning_rate": 0.0001908905412529669, |
|
"loss": 0.3448, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.6721581548599671, |
|
"grad_norm": 0.0932355722348614, |
|
"learning_rate": 0.00019072971833463269, |
|
"loss": 0.3519, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.6754530477759473, |
|
"grad_norm": 0.09198493368242532, |
|
"learning_rate": 0.00019056755715115374, |
|
"loss": 0.3474, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.6787479406919276, |
|
"grad_norm": 0.09105873669312843, |
|
"learning_rate": 0.00019040406009441073, |
|
"loss": 0.3446, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.6820428336079077, |
|
"grad_norm": 0.09474895757250051, |
|
"learning_rate": 0.00019023922957598846, |
|
"loss": 0.3549, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.685337726523888, |
|
"grad_norm": 0.09113820085850777, |
|
"learning_rate": 0.0001900730680271404, |
|
"loss": 0.3517, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.6886326194398682, |
|
"grad_norm": 0.08952952284278057, |
|
"learning_rate": 0.00018990557789875265, |
|
"loss": 0.3446, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.6919275123558485, |
|
"grad_norm": 0.08504291347151309, |
|
"learning_rate": 0.00018973676166130795, |
|
"loss": 0.3499, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.6952224052718287, |
|
"grad_norm": 0.08682346989926243, |
|
"learning_rate": 0.00018956662180484913, |
|
"loss": 0.3461, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.6985172981878089, |
|
"grad_norm": 0.09135947097159414, |
|
"learning_rate": 0.00018939516083894248, |
|
"loss": 0.3507, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.7018121911037891, |
|
"grad_norm": 0.08817857221763035, |
|
"learning_rate": 0.0001892223812926406, |
|
"loss": 0.34, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.7051070840197694, |
|
"grad_norm": 0.08660281169520562, |
|
"learning_rate": 0.00018904828571444525, |
|
"loss": 0.3467, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.7084019769357496, |
|
"grad_norm": 0.08873620967277855, |
|
"learning_rate": 0.00018887287667226964, |
|
"loss": 0.3359, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.7116968698517299, |
|
"grad_norm": 0.09198572626736658, |
|
"learning_rate": 0.00018869615675340068, |
|
"loss": 0.3419, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.71499176276771, |
|
"grad_norm": 0.08907612433680563, |
|
"learning_rate": 0.0001885181285644606, |
|
"loss": 0.3587, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.7182866556836903, |
|
"grad_norm": 0.08850895851919431, |
|
"learning_rate": 0.00018833879473136877, |
|
"loss": 0.3477, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.7215815485996705, |
|
"grad_norm": 0.0846389728816628, |
|
"learning_rate": 0.00018815815789930275, |
|
"loss": 0.3366, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.7248764415156508, |
|
"grad_norm": 0.09005118866716647, |
|
"learning_rate": 0.00018797622073265946, |
|
"loss": 0.3602, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.728171334431631, |
|
"grad_norm": 0.08535147302632334, |
|
"learning_rate": 0.00018779298591501564, |
|
"loss": 0.3422, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.7314662273476112, |
|
"grad_norm": 0.0854751704724938, |
|
"learning_rate": 0.0001876084561490885, |
|
"loss": 0.3426, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.7347611202635914, |
|
"grad_norm": 0.08664582706109708, |
|
"learning_rate": 0.00018742263415669582, |
|
"loss": 0.3532, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.7380560131795717, |
|
"grad_norm": 0.08458101578511544, |
|
"learning_rate": 0.00018723552267871555, |
|
"loss": 0.3438, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.7413509060955519, |
|
"grad_norm": 0.08431293605775432, |
|
"learning_rate": 0.0001870471244750458, |
|
"loss": 0.3502, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.7446457990115322, |
|
"grad_norm": 0.08369525987098254, |
|
"learning_rate": 0.00018685744232456374, |
|
"loss": 0.3333, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.7479406919275123, |
|
"grad_norm": 0.08784634286641266, |
|
"learning_rate": 0.00018666647902508494, |
|
"loss": 0.3401, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.7512355848434926, |
|
"grad_norm": 0.08421464642591442, |
|
"learning_rate": 0.00018647423739332175, |
|
"loss": 0.3275, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.7545304777594728, |
|
"grad_norm": 0.08154049872789887, |
|
"learning_rate": 0.00018628072026484214, |
|
"loss": 0.3333, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.7578253706754531, |
|
"grad_norm": 0.08114999530616501, |
|
"learning_rate": 0.00018608593049402754, |
|
"loss": 0.3434, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.7611202635914333, |
|
"grad_norm": 0.08374578563091178, |
|
"learning_rate": 0.0001858898709540309, |
|
"loss": 0.3449, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.7644151565074135, |
|
"grad_norm": 0.07916589630276719, |
|
"learning_rate": 0.00018569254453673444, |
|
"loss": 0.3268, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.7677100494233937, |
|
"grad_norm": 0.0867833099769683, |
|
"learning_rate": 0.00018549395415270664, |
|
"loss": 0.3292, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.771004942339374, |
|
"grad_norm": 0.08554090940119152, |
|
"learning_rate": 0.00018529410273115962, |
|
"loss": 0.3476, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.7742998352553542, |
|
"grad_norm": 0.08576144914828523, |
|
"learning_rate": 0.0001850929932199058, |
|
"loss": 0.3454, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.7775947281713345, |
|
"grad_norm": 0.09025913774937107, |
|
"learning_rate": 0.0001848906285853145, |
|
"loss": 0.347, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.7808896210873146, |
|
"grad_norm": 0.08519936716335193, |
|
"learning_rate": 0.00018468701181226804, |
|
"loss": 0.3397, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.7841845140032949, |
|
"grad_norm": 0.08499222661359229, |
|
"learning_rate": 0.0001844821459041179, |
|
"loss": 0.3458, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.7874794069192751, |
|
"grad_norm": 0.08609729000174243, |
|
"learning_rate": 0.00018427603388264025, |
|
"loss": 0.3332, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.7907742998352554, |
|
"grad_norm": 0.09181680148202499, |
|
"learning_rate": 0.00018406867878799154, |
|
"loss": 0.3499, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.7940691927512356, |
|
"grad_norm": 0.08755042199814178, |
|
"learning_rate": 0.0001838600836786635, |
|
"loss": 0.3442, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.7973640856672158, |
|
"grad_norm": 0.08357114463901723, |
|
"learning_rate": 0.00018365025163143814, |
|
"loss": 0.3496, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.800658978583196, |
|
"grad_norm": 0.08851618055895055, |
|
"learning_rate": 0.0001834391857413423, |
|
"loss": 0.3513, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.8039538714991763, |
|
"grad_norm": 0.09323076567865188, |
|
"learning_rate": 0.0001832268891216021, |
|
"loss": 0.3611, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.8072487644151565, |
|
"grad_norm": 0.0835634009338884, |
|
"learning_rate": 0.0001830133649035968, |
|
"loss": 0.3387, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.8105436573311368, |
|
"grad_norm": 0.08610843703599567, |
|
"learning_rate": 0.0001827986162368129, |
|
"loss": 0.3465, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.8138385502471169, |
|
"grad_norm": 0.08371169841943385, |
|
"learning_rate": 0.00018258264628879752, |
|
"loss": 0.3535, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.8171334431630972, |
|
"grad_norm": 0.08370848845452357, |
|
"learning_rate": 0.00018236545824511168, |
|
"loss": 0.3267, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.8204283360790774, |
|
"grad_norm": 0.09272034869991407, |
|
"learning_rate": 0.0001821470553092832, |
|
"loss": 0.3484, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.8237232289950577, |
|
"grad_norm": 0.08921657004712517, |
|
"learning_rate": 0.0001819274407027599, |
|
"loss": 0.3373, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.8270181219110379, |
|
"grad_norm": 0.08576896779949927, |
|
"learning_rate": 0.0001817066176648615, |
|
"loss": 0.3533, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.8303130148270181, |
|
"grad_norm": 0.08520255452310015, |
|
"learning_rate": 0.00018148458945273213, |
|
"loss": 0.3542, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.8336079077429983, |
|
"grad_norm": 0.08556329735314366, |
|
"learning_rate": 0.0001812613593412924, |
|
"loss": 0.3312, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.8369028006589786, |
|
"grad_norm": 0.08644232163750781, |
|
"learning_rate": 0.0001810369306231909, |
|
"loss": 0.3454, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.8401976935749588, |
|
"grad_norm": 0.08165688948658312, |
|
"learning_rate": 0.00018081130660875557, |
|
"loss": 0.3444, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.8434925864909391, |
|
"grad_norm": 0.08092566240118927, |
|
"learning_rate": 0.0001805844906259452, |
|
"loss": 0.341, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.8467874794069192, |
|
"grad_norm": 0.08175414891463337, |
|
"learning_rate": 0.00018035648602029997, |
|
"loss": 0.3474, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.8500823723228995, |
|
"grad_norm": 0.0852653315237305, |
|
"learning_rate": 0.00018012729615489236, |
|
"loss": 0.3445, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.8533772652388797, |
|
"grad_norm": 0.0839021217620115, |
|
"learning_rate": 0.00017989692441027744, |
|
"loss": 0.3478, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.85667215815486, |
|
"grad_norm": 0.08924662802136864, |
|
"learning_rate": 0.000179665374184443, |
|
"loss": 0.3392, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.8599670510708401, |
|
"grad_norm": 0.08424677502777078, |
|
"learning_rate": 0.00017943264889275944, |
|
"loss": 0.3451, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.8632619439868204, |
|
"grad_norm": 0.08560296301440222, |
|
"learning_rate": 0.00017919875196792948, |
|
"loss": 0.3438, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.8665568369028006, |
|
"grad_norm": 0.08330737239164673, |
|
"learning_rate": 0.00017896368685993736, |
|
"loss": 0.3499, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.8698517298187809, |
|
"grad_norm": 0.08043192435705314, |
|
"learning_rate": 0.00017872745703599808, |
|
"loss": 0.3326, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.8731466227347611, |
|
"grad_norm": 0.08601789750278353, |
|
"learning_rate": 0.00017849006598050625, |
|
"loss": 0.3377, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.8764415156507414, |
|
"grad_norm": 0.084248493499374, |
|
"learning_rate": 0.00017825151719498466, |
|
"loss": 0.3394, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.8797364085667215, |
|
"grad_norm": 0.08413208325626391, |
|
"learning_rate": 0.00017801181419803256, |
|
"loss": 0.3463, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.8830313014827018, |
|
"grad_norm": 0.08490917843632152, |
|
"learning_rate": 0.00017777096052527398, |
|
"loss": 0.3395, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.886326194398682, |
|
"grad_norm": 0.08529218362151314, |
|
"learning_rate": 0.00017752895972930537, |
|
"loss": 0.3432, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.8896210873146623, |
|
"grad_norm": 0.08406202028031955, |
|
"learning_rate": 0.0001772858153796432, |
|
"loss": 0.3382, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.8929159802306426, |
|
"grad_norm": 0.08192359095153046, |
|
"learning_rate": 0.0001770415310626715, |
|
"loss": 0.3438, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.8962108731466227, |
|
"grad_norm": 0.08667337369960532, |
|
"learning_rate": 0.0001767961103815888, |
|
"loss": 0.3467, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.899505766062603, |
|
"grad_norm": 0.09030944746983277, |
|
"learning_rate": 0.00017654955695635497, |
|
"loss": 0.3353, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.9028006589785832, |
|
"grad_norm": 0.08866625882756005, |
|
"learning_rate": 0.00017630187442363798, |
|
"loss": 0.3436, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.9060955518945635, |
|
"grad_norm": 0.08505719283151414, |
|
"learning_rate": 0.00017605306643676008, |
|
"loss": 0.3391, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.9093904448105437, |
|
"grad_norm": 0.08322038992732612, |
|
"learning_rate": 0.00017580313666564395, |
|
"loss": 0.3393, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.9126853377265239, |
|
"grad_norm": 0.0866505683085061, |
|
"learning_rate": 0.00017555208879675875, |
|
"loss": 0.3409, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.9159802306425041, |
|
"grad_norm": 0.08205474422496563, |
|
"learning_rate": 0.00017529992653306548, |
|
"loss": 0.3429, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.9192751235584844, |
|
"grad_norm": 0.08286735294573946, |
|
"learning_rate": 0.00017504665359396255, |
|
"loss": 0.3457, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.9225700164744646, |
|
"grad_norm": 0.0844618640503933, |
|
"learning_rate": 0.00017479227371523082, |
|
"loss": 0.3447, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.9258649093904449, |
|
"grad_norm": 0.07920730838964726, |
|
"learning_rate": 0.0001745367906489786, |
|
"loss": 0.3297, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.929159802306425, |
|
"grad_norm": 0.08334771190565178, |
|
"learning_rate": 0.00017428020816358605, |
|
"loss": 0.3474, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.9324546952224053, |
|
"grad_norm": 0.08786988637856916, |
|
"learning_rate": 0.00017402253004365008, |
|
"loss": 0.3302, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.9357495881383855, |
|
"grad_norm": 0.07938060869152685, |
|
"learning_rate": 0.00017376376008992797, |
|
"loss": 0.3215, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.9390444810543658, |
|
"grad_norm": 0.08542382095743929, |
|
"learning_rate": 0.00017350390211928166, |
|
"loss": 0.3391, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.942339373970346, |
|
"grad_norm": 0.08832893571241651, |
|
"learning_rate": 0.00017324295996462146, |
|
"loss": 0.3475, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.9456342668863262, |
|
"grad_norm": 0.08963369742867137, |
|
"learning_rate": 0.00017298093747484923, |
|
"loss": 0.3516, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.9489291598023064, |
|
"grad_norm": 0.08105453925768774, |
|
"learning_rate": 0.00017271783851480194, |
|
"loss": 0.3387, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.9522240527182867, |
|
"grad_norm": 0.08806421113907988, |
|
"learning_rate": 0.00017245366696519448, |
|
"loss": 0.3384, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.9555189456342669, |
|
"grad_norm": 0.08613965774858882, |
|
"learning_rate": 0.0001721884267225624, |
|
"loss": 0.3402, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.9588138385502472, |
|
"grad_norm": 0.08154607108470488, |
|
"learning_rate": 0.00017192212169920459, |
|
"loss": 0.3379, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.9621087314662273, |
|
"grad_norm": 0.08288246340352515, |
|
"learning_rate": 0.00017165475582312537, |
|
"loss": 0.3375, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.9654036243822076, |
|
"grad_norm": 0.09926330786301062, |
|
"learning_rate": 0.00017138633303797674, |
|
"loss": 0.346, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.9686985172981878, |
|
"grad_norm": 0.08416129865679868, |
|
"learning_rate": 0.00017111685730300015, |
|
"loss": 0.3397, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.9719934102141681, |
|
"grad_norm": 0.08420994658478355, |
|
"learning_rate": 0.00017084633259296797, |
|
"loss": 0.344, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.9752883031301482, |
|
"grad_norm": 0.0798569849854742, |
|
"learning_rate": 0.00017057476289812504, |
|
"loss": 0.3346, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.9785831960461285, |
|
"grad_norm": 0.08247914082738622, |
|
"learning_rate": 0.0001703021522241298, |
|
"loss": 0.3327, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.9818780889621087, |
|
"grad_norm": 0.08351637496453113, |
|
"learning_rate": 0.00017002850459199505, |
|
"loss": 0.3465, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.985172981878089, |
|
"grad_norm": 0.08746419940698986, |
|
"learning_rate": 0.00016975382403802878, |
|
"loss": 0.3511, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.9884678747940692, |
|
"grad_norm": 0.08467784721413267, |
|
"learning_rate": 0.00016947811461377467, |
|
"loss": 0.3482, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.9917627677100495, |
|
"grad_norm": 0.07930487498006872, |
|
"learning_rate": 0.00016920138038595216, |
|
"loss": 0.3347, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.9950576606260296, |
|
"grad_norm": 0.08974209141023505, |
|
"learning_rate": 0.00016892362543639654, |
|
"loss": 0.3435, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.9983525535420099, |
|
"grad_norm": 0.08015781337535312, |
|
"learning_rate": 0.00016864485386199892, |
|
"loss": 0.337, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.9983525535420099, |
|
"eval_loss": 0.33799201250076294, |
|
"eval_runtime": 167.3264, |
|
"eval_samples_per_second": 30.509, |
|
"eval_steps_per_second": 0.956, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 1.00164744645799, |
|
"grad_norm": 0.08187065929331228, |
|
"learning_rate": 0.0001683650697746455, |
|
"loss": 0.331, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 1.0049423393739703, |
|
"grad_norm": 0.08344895940789815, |
|
"learning_rate": 0.00016808427730115715, |
|
"loss": 0.3212, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 1.0082372322899507, |
|
"grad_norm": 0.0812051492859621, |
|
"learning_rate": 0.0001678024805832284, |
|
"loss": 0.321, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 1.0115321252059308, |
|
"grad_norm": 0.082485666654798, |
|
"learning_rate": 0.00016751968377736639, |
|
"loss": 0.3291, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 1.014827018121911, |
|
"grad_norm": 0.08852023762319512, |
|
"learning_rate": 0.00016723589105482967, |
|
"loss": 0.3243, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 1.0181219110378912, |
|
"grad_norm": 0.08541502987915972, |
|
"learning_rate": 0.00016695110660156653, |
|
"loss": 0.3179, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 1.0214168039538716, |
|
"grad_norm": 0.08162657768023696, |
|
"learning_rate": 0.00016666533461815326, |
|
"loss": 0.3128, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.0247116968698518, |
|
"grad_norm": 0.08866859604447454, |
|
"learning_rate": 0.0001663785793197323, |
|
"loss": 0.3263, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 1.028006589785832, |
|
"grad_norm": 0.08377699516728787, |
|
"learning_rate": 0.00016609084493595, |
|
"loss": 0.3172, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 1.031301482701812, |
|
"grad_norm": 0.08249601913350711, |
|
"learning_rate": 0.00016580213571089426, |
|
"loss": 0.3273, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 1.0345963756177925, |
|
"grad_norm": 0.08916625891991832, |
|
"learning_rate": 0.00016551245590303178, |
|
"loss": 0.3306, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 1.0378912685337727, |
|
"grad_norm": 0.08580016654634558, |
|
"learning_rate": 0.00016522180978514555, |
|
"loss": 0.322, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 1.0411861614497528, |
|
"grad_norm": 0.08571304387115758, |
|
"learning_rate": 0.00016493020164427152, |
|
"loss": 0.3275, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 1.044481054365733, |
|
"grad_norm": 0.08216233263790286, |
|
"learning_rate": 0.00016463763578163562, |
|
"loss": 0.3312, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 1.0477759472817134, |
|
"grad_norm": 0.0875991270266792, |
|
"learning_rate": 0.00016434411651259007, |
|
"loss": 0.3313, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 1.0510708401976936, |
|
"grad_norm": 0.08448691383529304, |
|
"learning_rate": 0.00016404964816654993, |
|
"loss": 0.3255, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 1.0543657331136738, |
|
"grad_norm": 0.0845703455173748, |
|
"learning_rate": 0.00016375423508692912, |
|
"loss": 0.3359, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.057660626029654, |
|
"grad_norm": 0.0882361267271713, |
|
"learning_rate": 0.00016345788163107646, |
|
"loss": 0.3341, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 1.0609555189456343, |
|
"grad_norm": 0.0849311944593189, |
|
"learning_rate": 0.00016316059217021125, |
|
"loss": 0.3239, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 1.0642504118616145, |
|
"grad_norm": 0.08152300437502101, |
|
"learning_rate": 0.000162862371089359, |
|
"loss": 0.3243, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 1.0675453047775947, |
|
"grad_norm": 0.08857601001972923, |
|
"learning_rate": 0.0001625632227872865, |
|
"loss": 0.333, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 1.0708401976935749, |
|
"grad_norm": 0.08801532835686154, |
|
"learning_rate": 0.00016226315167643723, |
|
"loss": 0.3322, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 1.0741350906095553, |
|
"grad_norm": 0.08525184904271058, |
|
"learning_rate": 0.0001619621621828659, |
|
"loss": 0.3167, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 1.0774299835255354, |
|
"grad_norm": 0.08135029197370526, |
|
"learning_rate": 0.0001616602587461736, |
|
"loss": 0.3269, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 1.0807248764415156, |
|
"grad_norm": 0.08424072997426707, |
|
"learning_rate": 0.000161357445819442, |
|
"loss": 0.3239, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 1.084019769357496, |
|
"grad_norm": 0.08258821495595987, |
|
"learning_rate": 0.00016105372786916775, |
|
"loss": 0.3183, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 1.0873146622734762, |
|
"grad_norm": 0.08498350432029378, |
|
"learning_rate": 0.00016074910937519663, |
|
"loss": 0.3328, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.0906095551894563, |
|
"grad_norm": 0.08667493093004443, |
|
"learning_rate": 0.0001604435948306575, |
|
"loss": 0.3298, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 1.0939044481054365, |
|
"grad_norm": 0.08869319837110255, |
|
"learning_rate": 0.00016013718874189595, |
|
"loss": 0.3336, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 1.0971993410214167, |
|
"grad_norm": 0.08379372154609564, |
|
"learning_rate": 0.00015982989562840784, |
|
"loss": 0.3206, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 1.100494233937397, |
|
"grad_norm": 0.08172530708530337, |
|
"learning_rate": 0.0001595217200227727, |
|
"loss": 0.3212, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 1.1037891268533773, |
|
"grad_norm": 0.08535886274333539, |
|
"learning_rate": 0.0001592126664705868, |
|
"loss": 0.3243, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 1.1070840197693574, |
|
"grad_norm": 0.08360595473099759, |
|
"learning_rate": 0.0001589027395303962, |
|
"loss": 0.3233, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 1.1103789126853378, |
|
"grad_norm": 0.08265998271604205, |
|
"learning_rate": 0.0001585919437736294, |
|
"loss": 0.3271, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 1.113673805601318, |
|
"grad_norm": 0.08237810191987908, |
|
"learning_rate": 0.00015828028378452998, |
|
"loss": 0.3238, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 1.1169686985172982, |
|
"grad_norm": 0.08730291854461844, |
|
"learning_rate": 0.00015796776416008898, |
|
"loss": 0.3195, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 1.1202635914332784, |
|
"grad_norm": 0.08634254849058057, |
|
"learning_rate": 0.00015765438950997705, |
|
"loss": 0.323, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.1235584843492585, |
|
"grad_norm": 0.08081122456257506, |
|
"learning_rate": 0.0001573401644564764, |
|
"loss": 0.3102, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 1.126853377265239, |
|
"grad_norm": 0.08841633690696776, |
|
"learning_rate": 0.00015702509363441295, |
|
"loss": 0.3345, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 1.130148270181219, |
|
"grad_norm": 0.08830337626554721, |
|
"learning_rate": 0.00015670918169108752, |
|
"loss": 0.3282, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 1.1334431630971993, |
|
"grad_norm": 0.08099355585017688, |
|
"learning_rate": 0.00015639243328620744, |
|
"loss": 0.3058, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 1.1367380560131797, |
|
"grad_norm": 0.08076571239439685, |
|
"learning_rate": 0.00015607485309181813, |
|
"loss": 0.3182, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 1.1400329489291599, |
|
"grad_norm": 0.0829246184070146, |
|
"learning_rate": 0.00015575644579223362, |
|
"loss": 0.323, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 1.14332784184514, |
|
"grad_norm": 0.08668721952043845, |
|
"learning_rate": 0.00015543721608396796, |
|
"loss": 0.3163, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 1.1466227347611202, |
|
"grad_norm": 0.08394538492198296, |
|
"learning_rate": 0.0001551171686756657, |
|
"loss": 0.3116, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 1.1499176276771004, |
|
"grad_norm": 0.08597909584931814, |
|
"learning_rate": 0.00015479630828803235, |
|
"loss": 0.3205, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 1.1532125205930808, |
|
"grad_norm": 0.0840149508071814, |
|
"learning_rate": 0.0001544746396537651, |
|
"loss": 0.3093, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.156507413509061, |
|
"grad_norm": 0.08658284232618915, |
|
"learning_rate": 0.00015415216751748264, |
|
"loss": 0.316, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 1.1598023064250411, |
|
"grad_norm": 0.08152435596529481, |
|
"learning_rate": 0.0001538288966356554, |
|
"loss": 0.3177, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 1.1630971993410215, |
|
"grad_norm": 0.08209820514345408, |
|
"learning_rate": 0.00015350483177653526, |
|
"loss": 0.3261, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 1.1663920922570017, |
|
"grad_norm": 0.08272819539416575, |
|
"learning_rate": 0.00015317997772008537, |
|
"loss": 0.3226, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 1.1696869851729819, |
|
"grad_norm": 0.08660002845038178, |
|
"learning_rate": 0.00015285433925790945, |
|
"loss": 0.3191, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 1.172981878088962, |
|
"grad_norm": 0.08632540443067051, |
|
"learning_rate": 0.0001525279211931813, |
|
"loss": 0.3263, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 1.1762767710049424, |
|
"grad_norm": 0.08565341191026239, |
|
"learning_rate": 0.00015220072834057387, |
|
"loss": 0.3192, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 1.1795716639209226, |
|
"grad_norm": 0.07902224269764019, |
|
"learning_rate": 0.00015187276552618817, |
|
"loss": 0.3052, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 1.1828665568369028, |
|
"grad_norm": 0.08704133599468743, |
|
"learning_rate": 0.0001515440375874823, |
|
"loss": 0.3129, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 1.186161449752883, |
|
"grad_norm": 0.08910719635327755, |
|
"learning_rate": 0.00015121454937319976, |
|
"loss": 0.3393, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.1894563426688634, |
|
"grad_norm": 0.08936796334171324, |
|
"learning_rate": 0.00015088430574329836, |
|
"loss": 0.3237, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 1.1927512355848435, |
|
"grad_norm": 0.08635637992605481, |
|
"learning_rate": 0.0001505533115688781, |
|
"loss": 0.3317, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 1.1960461285008237, |
|
"grad_norm": 0.08610779702269872, |
|
"learning_rate": 0.00015022157173210968, |
|
"loss": 0.3293, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 1.1993410214168039, |
|
"grad_norm": 0.08526653923014106, |
|
"learning_rate": 0.0001498890911261622, |
|
"loss": 0.3245, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 1.2026359143327843, |
|
"grad_norm": 0.08300876143938384, |
|
"learning_rate": 0.0001495558746551313, |
|
"loss": 0.3341, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 1.2059308072487644, |
|
"grad_norm": 0.08284607312666305, |
|
"learning_rate": 0.00014922192723396645, |
|
"loss": 0.3219, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 1.2092257001647446, |
|
"grad_norm": 0.08002802793230862, |
|
"learning_rate": 0.00014888725378839877, |
|
"loss": 0.3143, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 1.2125205930807248, |
|
"grad_norm": 0.08140933851788057, |
|
"learning_rate": 0.00014855185925486818, |
|
"loss": 0.3243, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 1.2158154859967052, |
|
"grad_norm": 0.08664085115985616, |
|
"learning_rate": 0.00014821574858045074, |
|
"loss": 0.3229, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 1.2191103789126854, |
|
"grad_norm": 0.08612508394887794, |
|
"learning_rate": 0.00014787892672278556, |
|
"loss": 0.3259, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.2224052718286655, |
|
"grad_norm": 0.08770343929357799, |
|
"learning_rate": 0.00014754139865000168, |
|
"loss": 0.3315, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 1.2257001647446457, |
|
"grad_norm": 0.08425286738586797, |
|
"learning_rate": 0.00014720316934064496, |
|
"loss": 0.3348, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 1.2289950576606261, |
|
"grad_norm": 0.08560046261080007, |
|
"learning_rate": 0.00014686424378360433, |
|
"loss": 0.3278, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 1.2322899505766063, |
|
"grad_norm": 0.08692213775944008, |
|
"learning_rate": 0.00014652462697803848, |
|
"loss": 0.3289, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 1.2355848434925865, |
|
"grad_norm": 0.0850172377986241, |
|
"learning_rate": 0.0001461843239333021, |
|
"loss": 0.3289, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 1.2388797364085666, |
|
"grad_norm": 0.08538674165797111, |
|
"learning_rate": 0.00014584333966887177, |
|
"loss": 0.3077, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 1.242174629324547, |
|
"grad_norm": 0.08225612872698992, |
|
"learning_rate": 0.0001455016792142722, |
|
"loss": 0.3207, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 1.2454695222405272, |
|
"grad_norm": 0.0824992791408092, |
|
"learning_rate": 0.00014515934760900184, |
|
"loss": 0.3123, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 1.2487644151565074, |
|
"grad_norm": 0.08817004563803918, |
|
"learning_rate": 0.0001448163499024587, |
|
"loss": 0.331, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 1.2520593080724876, |
|
"grad_norm": 0.08988900733842042, |
|
"learning_rate": 0.00014447269115386573, |
|
"loss": 0.3171, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.255354200988468, |
|
"grad_norm": 0.08242288142785008, |
|
"learning_rate": 0.00014412837643219625, |
|
"loss": 0.3195, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 1.2586490939044481, |
|
"grad_norm": 0.08150045524497893, |
|
"learning_rate": 0.00014378341081609927, |
|
"loss": 0.3287, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 1.2619439868204283, |
|
"grad_norm": 0.08343807243424793, |
|
"learning_rate": 0.00014343779939382452, |
|
"loss": 0.3111, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 1.2652388797364087, |
|
"grad_norm": 0.08374911663870195, |
|
"learning_rate": 0.0001430915472631472, |
|
"loss": 0.3278, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 1.2685337726523889, |
|
"grad_norm": 0.08428186467144357, |
|
"learning_rate": 0.00014274465953129325, |
|
"loss": 0.3186, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 1.271828665568369, |
|
"grad_norm": 0.08176159742747383, |
|
"learning_rate": 0.00014239714131486348, |
|
"loss": 0.3196, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 1.2751235584843492, |
|
"grad_norm": 0.08350565101214433, |
|
"learning_rate": 0.00014204899773975855, |
|
"loss": 0.3235, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 1.2784184514003294, |
|
"grad_norm": 0.08179249241721015, |
|
"learning_rate": 0.00014170023394110306, |
|
"loss": 0.3137, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 1.2817133443163098, |
|
"grad_norm": 0.0837263180330502, |
|
"learning_rate": 0.00014135085506316997, |
|
"loss": 0.3229, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 1.28500823723229, |
|
"grad_norm": 0.08607305244666162, |
|
"learning_rate": 0.00014100086625930464, |
|
"loss": 0.3231, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.2883031301482701, |
|
"grad_norm": 0.08575599533724002, |
|
"learning_rate": 0.00014065027269184887, |
|
"loss": 0.3231, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 1.2915980230642505, |
|
"grad_norm": 0.08064829044874067, |
|
"learning_rate": 0.00014029907953206475, |
|
"loss": 0.3199, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 1.2948929159802307, |
|
"grad_norm": 0.08408021103636296, |
|
"learning_rate": 0.0001399472919600584, |
|
"loss": 0.3273, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 1.2981878088962109, |
|
"grad_norm": 0.08114989710998861, |
|
"learning_rate": 0.00013959491516470334, |
|
"loss": 0.3162, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 1.301482701812191, |
|
"grad_norm": 0.0834997194534925, |
|
"learning_rate": 0.00013924195434356442, |
|
"loss": 0.3185, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 1.3047775947281712, |
|
"grad_norm": 0.08216535139972479, |
|
"learning_rate": 0.0001388884147028207, |
|
"loss": 0.3318, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 1.3080724876441516, |
|
"grad_norm": 0.08151769664701901, |
|
"learning_rate": 0.0001385343014571889, |
|
"loss": 0.3242, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 1.3113673805601318, |
|
"grad_norm": 0.0807742083114109, |
|
"learning_rate": 0.00013817961982984643, |
|
"loss": 0.3205, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 1.314662273476112, |
|
"grad_norm": 0.08184040624493487, |
|
"learning_rate": 0.0001378243750523543, |
|
"loss": 0.3149, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 1.3179571663920924, |
|
"grad_norm": 0.08102459745060192, |
|
"learning_rate": 0.00013746857236458007, |
|
"loss": 0.319, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.3212520593080725, |
|
"grad_norm": 0.08059247007108714, |
|
"learning_rate": 0.00013711221701462036, |
|
"loss": 0.3127, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 1.3245469522240527, |
|
"grad_norm": 0.0793438468854946, |
|
"learning_rate": 0.0001367553142587237, |
|
"loss": 0.3251, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 1.327841845140033, |
|
"grad_norm": 0.08216864917484416, |
|
"learning_rate": 0.00013639786936121286, |
|
"loss": 0.3197, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 1.331136738056013, |
|
"grad_norm": 0.08321227130258257, |
|
"learning_rate": 0.0001360398875944071, |
|
"loss": 0.321, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 1.3344316309719935, |
|
"grad_norm": 0.08148653549162627, |
|
"learning_rate": 0.00013568137423854458, |
|
"loss": 0.3238, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 1.3377265238879736, |
|
"grad_norm": 0.08624309126721959, |
|
"learning_rate": 0.00013532233458170444, |
|
"loss": 0.3239, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 1.3410214168039538, |
|
"grad_norm": 0.0797835803516585, |
|
"learning_rate": 0.00013496277391972873, |
|
"loss": 0.3269, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 1.3443163097199342, |
|
"grad_norm": 0.08094589206278008, |
|
"learning_rate": 0.00013460269755614435, |
|
"loss": 0.3041, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 1.3476112026359144, |
|
"grad_norm": 0.0852983943023146, |
|
"learning_rate": 0.00013424211080208479, |
|
"loss": 0.3118, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 1.3509060955518946, |
|
"grad_norm": 0.08749111120896899, |
|
"learning_rate": 0.00013388101897621182, |
|
"loss": 0.3271, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.3542009884678747, |
|
"grad_norm": 0.07952649573183727, |
|
"learning_rate": 0.00013351942740463707, |
|
"loss": 0.3081, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 1.357495881383855, |
|
"grad_norm": 0.08578795160366025, |
|
"learning_rate": 0.00013315734142084335, |
|
"loss": 0.32, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 1.3607907742998353, |
|
"grad_norm": 0.08928715903816203, |
|
"learning_rate": 0.00013279476636560608, |
|
"loss": 0.3227, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 1.3640856672158155, |
|
"grad_norm": 0.08229016926007318, |
|
"learning_rate": 0.0001324317075869146, |
|
"loss": 0.3275, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 1.3673805601317957, |
|
"grad_norm": 0.0887890717135067, |
|
"learning_rate": 0.00013206817043989302, |
|
"loss": 0.3323, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 1.370675453047776, |
|
"grad_norm": 0.08699973758850778, |
|
"learning_rate": 0.0001317041602867215, |
|
"loss": 0.325, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 1.3739703459637562, |
|
"grad_norm": 0.08467845536010866, |
|
"learning_rate": 0.000131339682496557, |
|
"loss": 0.3078, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 1.3772652388797364, |
|
"grad_norm": 0.08776785203585662, |
|
"learning_rate": 0.0001309747424454542, |
|
"loss": 0.3418, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 1.3805601317957166, |
|
"grad_norm": 0.08658565822289628, |
|
"learning_rate": 0.00013060934551628603, |
|
"loss": 0.3328, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 1.3838550247116967, |
|
"grad_norm": 0.08764833043201878, |
|
"learning_rate": 0.00013024349709866447, |
|
"loss": 0.309, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.3871499176276771, |
|
"grad_norm": 0.08756434828387893, |
|
"learning_rate": 0.00012987720258886095, |
|
"loss": 0.3306, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 1.3904448105436573, |
|
"grad_norm": 0.08469701138029223, |
|
"learning_rate": 0.00012951046738972672, |
|
"loss": 0.323, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 1.3937397034596375, |
|
"grad_norm": 0.08504023280141097, |
|
"learning_rate": 0.00012914329691061328, |
|
"loss": 0.3211, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 1.3970345963756179, |
|
"grad_norm": 0.08487408883622825, |
|
"learning_rate": 0.00012877569656729243, |
|
"loss": 0.3313, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 1.400329489291598, |
|
"grad_norm": 0.07939354397337003, |
|
"learning_rate": 0.00012840767178187655, |
|
"loss": 0.3102, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.4036243822075782, |
|
"grad_norm": 0.0829139296391848, |
|
"learning_rate": 0.00012803922798273852, |
|
"loss": 0.316, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 1.4069192751235584, |
|
"grad_norm": 0.0823703090760657, |
|
"learning_rate": 0.00012767037060443172, |
|
"loss": 0.3251, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 1.4102141680395386, |
|
"grad_norm": 0.08066102303361472, |
|
"learning_rate": 0.00012730110508760975, |
|
"loss": 0.3238, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 1.413509060955519, |
|
"grad_norm": 0.0844768933412219, |
|
"learning_rate": 0.0001269314368789463, |
|
"loss": 0.3242, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 1.4168039538714992, |
|
"grad_norm": 0.0815698910146405, |
|
"learning_rate": 0.00012656137143105483, |
|
"loss": 0.3242, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.4200988467874793, |
|
"grad_norm": 0.08017268325963865, |
|
"learning_rate": 0.00012619091420240793, |
|
"loss": 0.3034, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 1.4233937397034597, |
|
"grad_norm": 0.08083122415122998, |
|
"learning_rate": 0.0001258200706572572, |
|
"loss": 0.3169, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 1.42668863261944, |
|
"grad_norm": 0.08337187987916603, |
|
"learning_rate": 0.00012544884626555225, |
|
"loss": 0.3193, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 1.42998352553542, |
|
"grad_norm": 0.08129518922024724, |
|
"learning_rate": 0.00012507724650286014, |
|
"loss": 0.3071, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 1.4332784184514002, |
|
"grad_norm": 0.08367220188501179, |
|
"learning_rate": 0.00012470527685028482, |
|
"loss": 0.3238, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 1.4365733113673804, |
|
"grad_norm": 0.08412225464889816, |
|
"learning_rate": 0.00012433294279438602, |
|
"loss": 0.3157, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 1.4398682042833608, |
|
"grad_norm": 0.0831281998958449, |
|
"learning_rate": 0.00012396024982709843, |
|
"loss": 0.3141, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 1.443163097199341, |
|
"grad_norm": 0.08365116413450047, |
|
"learning_rate": 0.00012358720344565077, |
|
"loss": 0.3159, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 1.4464579901153214, |
|
"grad_norm": 0.08191433963107164, |
|
"learning_rate": 0.00012321380915248446, |
|
"loss": 0.3098, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 1.4497528830313016, |
|
"grad_norm": 0.08348553478357966, |
|
"learning_rate": 0.0001228400724551728, |
|
"loss": 0.3129, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.4530477759472817, |
|
"grad_norm": 0.0881498231342535, |
|
"learning_rate": 0.0001224659988663395, |
|
"loss": 0.3372, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 1.456342668863262, |
|
"grad_norm": 0.08112710228760632, |
|
"learning_rate": 0.0001220915939035774, |
|
"loss": 0.3169, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 1.459637561779242, |
|
"grad_norm": 0.0836876463479396, |
|
"learning_rate": 0.00012171686308936709, |
|
"loss": 0.3192, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 1.4629324546952225, |
|
"grad_norm": 0.08499306017819758, |
|
"learning_rate": 0.00012134181195099555, |
|
"loss": 0.3309, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 1.4662273476112027, |
|
"grad_norm": 0.07947051276130836, |
|
"learning_rate": 0.00012096644602047447, |
|
"loss": 0.3137, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 1.4695222405271828, |
|
"grad_norm": 0.08141505738594229, |
|
"learning_rate": 0.0001205907708344588, |
|
"loss": 0.3183, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 1.4728171334431632, |
|
"grad_norm": 0.08267307338524962, |
|
"learning_rate": 0.000120214791934165, |
|
"loss": 0.3195, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 1.4761120263591434, |
|
"grad_norm": 0.08366455798813263, |
|
"learning_rate": 0.00011983851486528925, |
|
"loss": 0.3252, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 1.4794069192751236, |
|
"grad_norm": 0.08075885404123719, |
|
"learning_rate": 0.00011946194517792584, |
|
"loss": 0.3255, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 1.4827018121911038, |
|
"grad_norm": 0.0812184444903526, |
|
"learning_rate": 0.00011908508842648506, |
|
"loss": 0.3196, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.485996705107084, |
|
"grad_norm": 0.08188793010828326, |
|
"learning_rate": 0.00011870795016961156, |
|
"loss": 0.3191, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 1.4892915980230643, |
|
"grad_norm": 0.08074626674132604, |
|
"learning_rate": 0.00011833053597010201, |
|
"loss": 0.3082, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 1.4925864909390445, |
|
"grad_norm": 0.0799989877308315, |
|
"learning_rate": 0.00011795285139482342, |
|
"loss": 0.3185, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 1.4958813838550247, |
|
"grad_norm": 0.07888888308680399, |
|
"learning_rate": 0.00011757490201463065, |
|
"loss": 0.3158, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 1.499176276771005, |
|
"grad_norm": 0.07859416029173805, |
|
"learning_rate": 0.00011719669340428472, |
|
"loss": 0.316, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 1.5024711696869852, |
|
"grad_norm": 0.07954941297717674, |
|
"learning_rate": 0.00011681823114237, |
|
"loss": 0.3058, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 1.5057660626029654, |
|
"grad_norm": 0.07977327521145369, |
|
"learning_rate": 0.00011643952081121238, |
|
"loss": 0.3125, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 1.5090609555189456, |
|
"grad_norm": 0.08507806643743294, |
|
"learning_rate": 0.00011606056799679684, |
|
"loss": 0.3286, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 1.5123558484349258, |
|
"grad_norm": 0.08398623989677817, |
|
"learning_rate": 0.00011568137828868477, |
|
"loss": 0.3242, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 1.515650741350906, |
|
"grad_norm": 0.07796453705937313, |
|
"learning_rate": 0.000115301957279932, |
|
"loss": 0.3042, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.5189456342668863, |
|
"grad_norm": 0.08177587861673677, |
|
"learning_rate": 0.0001149223105670059, |
|
"loss": 0.3088, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 1.5222405271828665, |
|
"grad_norm": 0.07924829822850359, |
|
"learning_rate": 0.00011454244374970297, |
|
"loss": 0.3203, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 1.525535420098847, |
|
"grad_norm": 0.07918582660420821, |
|
"learning_rate": 0.00011416236243106637, |
|
"loss": 0.3176, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 1.528830313014827, |
|
"grad_norm": 0.08010698180871471, |
|
"learning_rate": 0.00011378207221730301, |
|
"loss": 0.3124, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 1.5321252059308073, |
|
"grad_norm": 0.08128777832664769, |
|
"learning_rate": 0.00011340157871770117, |
|
"loss": 0.3044, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 1.5354200988467874, |
|
"grad_norm": 0.08036561662806507, |
|
"learning_rate": 0.00011302088754454744, |
|
"loss": 0.3129, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 1.5387149917627676, |
|
"grad_norm": 0.08224346354578807, |
|
"learning_rate": 0.00011264000431304422, |
|
"loss": 0.3259, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 1.5420098846787478, |
|
"grad_norm": 0.08411964168011243, |
|
"learning_rate": 0.00011225893464122673, |
|
"loss": 0.3242, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 1.5453047775947282, |
|
"grad_norm": 0.08217977092177307, |
|
"learning_rate": 0.00011187768414988014, |
|
"loss": 0.3133, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 1.5485996705107083, |
|
"grad_norm": 0.07968019018419265, |
|
"learning_rate": 0.00011149625846245682, |
|
"loss": 0.3129, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.5518945634266887, |
|
"grad_norm": 0.08282232295877028, |
|
"learning_rate": 0.00011111466320499317, |
|
"loss": 0.3238, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 1.555189456342669, |
|
"grad_norm": 0.07977651680710264, |
|
"learning_rate": 0.00011073290400602681, |
|
"loss": 0.3152, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 1.558484349258649, |
|
"grad_norm": 0.0796874941168346, |
|
"learning_rate": 0.00011035098649651355, |
|
"loss": 0.3189, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 1.5617792421746293, |
|
"grad_norm": 0.07896439490671953, |
|
"learning_rate": 0.00010996891630974415, |
|
"loss": 0.315, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 1.5650741350906094, |
|
"grad_norm": 0.08381587983527669, |
|
"learning_rate": 0.00010958669908126151, |
|
"loss": 0.3265, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 1.5683690280065898, |
|
"grad_norm": 0.08244330689629394, |
|
"learning_rate": 0.00010920434044877732, |
|
"loss": 0.3256, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 1.57166392092257, |
|
"grad_norm": 0.0826167817539002, |
|
"learning_rate": 0.00010882184605208894, |
|
"loss": 0.313, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 1.5749588138385504, |
|
"grad_norm": 0.08267352564043831, |
|
"learning_rate": 0.00010843922153299637, |
|
"loss": 0.329, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 1.5782537067545306, |
|
"grad_norm": 0.08070973779200152, |
|
"learning_rate": 0.0001080564725352188, |
|
"loss": 0.3196, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 1.5815485996705108, |
|
"grad_norm": 0.08132267986915792, |
|
"learning_rate": 0.00010767360470431158, |
|
"loss": 0.3235, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.584843492586491, |
|
"grad_norm": 0.07850537692603574, |
|
"learning_rate": 0.00010729062368758278, |
|
"loss": 0.3181, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 1.588138385502471, |
|
"grad_norm": 0.07881922957573682, |
|
"learning_rate": 0.00010690753513401003, |
|
"loss": 0.3205, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 1.5914332784184513, |
|
"grad_norm": 0.08487734661734532, |
|
"learning_rate": 0.00010652434469415705, |
|
"loss": 0.3256, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 1.5947281713344317, |
|
"grad_norm": 0.08101732034272957, |
|
"learning_rate": 0.00010614105802009044, |
|
"loss": 0.3275, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 1.5980230642504119, |
|
"grad_norm": 0.08118895741694347, |
|
"learning_rate": 0.00010575768076529626, |
|
"loss": 0.3153, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 1.6013179571663922, |
|
"grad_norm": 0.08250314923552278, |
|
"learning_rate": 0.00010537421858459661, |
|
"loss": 0.3227, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 1.6046128500823724, |
|
"grad_norm": 0.07731877431714403, |
|
"learning_rate": 0.00010499067713406623, |
|
"loss": 0.3099, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 1.6079077429983526, |
|
"grad_norm": 0.08521321410108347, |
|
"learning_rate": 0.0001046070620709492, |
|
"loss": 0.3207, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 1.6112026359143328, |
|
"grad_norm": 0.08226505701423492, |
|
"learning_rate": 0.00010422337905357523, |
|
"loss": 0.317, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 1.614497528830313, |
|
"grad_norm": 0.08182233558893907, |
|
"learning_rate": 0.00010383963374127645, |
|
"loss": 0.3153, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.6177924217462931, |
|
"grad_norm": 0.08005303314312309, |
|
"learning_rate": 0.00010345583179430388, |
|
"loss": 0.3255, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 1.6210873146622735, |
|
"grad_norm": 0.08168467422688232, |
|
"learning_rate": 0.00010307197887374376, |
|
"loss": 0.3149, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 1.6243822075782537, |
|
"grad_norm": 0.0794434626126838, |
|
"learning_rate": 0.00010268808064143438, |
|
"loss": 0.3135, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 1.627677100494234, |
|
"grad_norm": 0.08056080525788428, |
|
"learning_rate": 0.00010230414275988218, |
|
"loss": 0.3167, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 1.6309719934102143, |
|
"grad_norm": 0.08147329845434119, |
|
"learning_rate": 0.00010192017089217862, |
|
"loss": 0.3232, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 1.6342668863261944, |
|
"grad_norm": 0.07839262012603093, |
|
"learning_rate": 0.00010153617070191632, |
|
"loss": 0.3123, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 1.6375617792421746, |
|
"grad_norm": 0.07951939016035157, |
|
"learning_rate": 0.00010115214785310568, |
|
"loss": 0.3116, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 1.6408566721581548, |
|
"grad_norm": 0.07912600699062161, |
|
"learning_rate": 0.00010076810801009137, |
|
"loss": 0.3142, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 1.644151565074135, |
|
"grad_norm": 0.08317226545954248, |
|
"learning_rate": 0.00010038405683746867, |
|
"loss": 0.3189, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 1.6474464579901154, |
|
"grad_norm": 0.08011582289342602, |
|
"learning_rate": 0.0001, |
|
"loss": 0.3173, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.6507413509060955, |
|
"grad_norm": 0.08299947106298346, |
|
"learning_rate": 9.961594316253134e-05, |
|
"loss": 0.3317, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 1.654036243822076, |
|
"grad_norm": 0.08048632067214515, |
|
"learning_rate": 9.923189198990868e-05, |
|
"loss": 0.3155, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 1.657331136738056, |
|
"grad_norm": 0.08095937151716077, |
|
"learning_rate": 9.884785214689435e-05, |
|
"loss": 0.3125, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 1.6606260296540363, |
|
"grad_norm": 0.08711970279631695, |
|
"learning_rate": 9.84638292980837e-05, |
|
"loss": 0.3231, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 1.6639209225700164, |
|
"grad_norm": 0.08339433460792585, |
|
"learning_rate": 9.807982910782141e-05, |
|
"loss": 0.3176, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 1.6672158154859966, |
|
"grad_norm": 0.08207683826613399, |
|
"learning_rate": 9.769585724011783e-05, |
|
"loss": 0.3218, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 1.6705107084019768, |
|
"grad_norm": 0.08030351365951861, |
|
"learning_rate": 9.731191935856565e-05, |
|
"loss": 0.3266, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 1.6738056013179572, |
|
"grad_norm": 0.08063540809733169, |
|
"learning_rate": 9.692802112625623e-05, |
|
"loss": 0.3109, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 1.6771004942339374, |
|
"grad_norm": 0.08217788739453986, |
|
"learning_rate": 9.654416820569617e-05, |
|
"loss": 0.3212, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 1.6803953871499178, |
|
"grad_norm": 0.08104865270695036, |
|
"learning_rate": 9.616036625872357e-05, |
|
"loss": 0.3159, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.683690280065898, |
|
"grad_norm": 0.07916408331958537, |
|
"learning_rate": 9.577662094642478e-05, |
|
"loss": 0.3175, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 1.6869851729818781, |
|
"grad_norm": 0.0778681677011439, |
|
"learning_rate": 9.539293792905083e-05, |
|
"loss": 0.3004, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 1.6902800658978583, |
|
"grad_norm": 0.08479179086459149, |
|
"learning_rate": 9.500932286593377e-05, |
|
"loss": 0.3218, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 1.6935749588138385, |
|
"grad_norm": 0.08491747525792748, |
|
"learning_rate": 9.462578141540341e-05, |
|
"loss": 0.3187, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 1.6968698517298186, |
|
"grad_norm": 0.08428239314838194, |
|
"learning_rate": 9.424231923470377e-05, |
|
"loss": 0.3261, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 1.700164744645799, |
|
"grad_norm": 0.0776528664851824, |
|
"learning_rate": 9.385894197990957e-05, |
|
"loss": 0.3154, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 1.7034596375617792, |
|
"grad_norm": 0.08160169468766518, |
|
"learning_rate": 9.347565530584298e-05, |
|
"loss": 0.3208, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 1.7067545304777596, |
|
"grad_norm": 0.08219835239653911, |
|
"learning_rate": 9.309246486598999e-05, |
|
"loss": 0.3185, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 1.7100494233937398, |
|
"grad_norm": 0.07844970543869025, |
|
"learning_rate": 9.270937631241723e-05, |
|
"loss": 0.3095, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 1.71334431630972, |
|
"grad_norm": 0.0804533791298634, |
|
"learning_rate": 9.232639529568843e-05, |
|
"loss": 0.3062, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.7166392092257001, |
|
"grad_norm": 0.07915250160793909, |
|
"learning_rate": 9.194352746478123e-05, |
|
"loss": 0.3193, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 1.7199341021416803, |
|
"grad_norm": 0.08112939736857154, |
|
"learning_rate": 9.156077846700367e-05, |
|
"loss": 0.3246, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 1.7232289950576605, |
|
"grad_norm": 0.0838606481003645, |
|
"learning_rate": 9.117815394791107e-05, |
|
"loss": 0.3147, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 1.7265238879736409, |
|
"grad_norm": 0.08242793925392256, |
|
"learning_rate": 9.07956595512227e-05, |
|
"loss": 0.3038, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 1.729818780889621, |
|
"grad_norm": 0.08176093816973408, |
|
"learning_rate": 9.041330091873853e-05, |
|
"loss": 0.3104, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.7331136738056014, |
|
"grad_norm": 0.08767477644679618, |
|
"learning_rate": 9.003108369025586e-05, |
|
"loss": 0.3317, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 1.7364085667215816, |
|
"grad_norm": 0.07767826934356571, |
|
"learning_rate": 8.964901350348648e-05, |
|
"loss": 0.3109, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 1.7397034596375618, |
|
"grad_norm": 0.08145456958981293, |
|
"learning_rate": 8.926709599397318e-05, |
|
"loss": 0.3189, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 1.742998352553542, |
|
"grad_norm": 0.08009880884270958, |
|
"learning_rate": 8.888533679500688e-05, |
|
"loss": 0.3148, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 1.7462932454695221, |
|
"grad_norm": 0.07944592757175745, |
|
"learning_rate": 8.850374153754322e-05, |
|
"loss": 0.3205, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.7495881383855023, |
|
"grad_norm": 0.08077078414479961, |
|
"learning_rate": 8.812231585011986e-05, |
|
"loss": 0.3109, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 1.7528830313014827, |
|
"grad_norm": 0.07868302874406433, |
|
"learning_rate": 8.774106535877331e-05, |
|
"loss": 0.3115, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 1.7561779242174629, |
|
"grad_norm": 0.07987610438576559, |
|
"learning_rate": 8.735999568695579e-05, |
|
"loss": 0.3129, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 1.7594728171334433, |
|
"grad_norm": 0.08073251563008862, |
|
"learning_rate": 8.697911245545257e-05, |
|
"loss": 0.3113, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 1.7627677100494235, |
|
"grad_norm": 0.07944843744740043, |
|
"learning_rate": 8.659842128229887e-05, |
|
"loss": 0.3155, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 1.7660626029654036, |
|
"grad_norm": 0.07780637939142554, |
|
"learning_rate": 8.6217927782697e-05, |
|
"loss": 0.3, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 1.7693574958813838, |
|
"grad_norm": 0.08113477442420164, |
|
"learning_rate": 8.583763756893365e-05, |
|
"loss": 0.3167, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 1.772652388797364, |
|
"grad_norm": 0.08057722076446969, |
|
"learning_rate": 8.545755625029704e-05, |
|
"loss": 0.3175, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 1.7759472817133442, |
|
"grad_norm": 0.08313615782876255, |
|
"learning_rate": 8.507768943299415e-05, |
|
"loss": 0.3151, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 1.7792421746293245, |
|
"grad_norm": 0.08066682962003371, |
|
"learning_rate": 8.469804272006801e-05, |
|
"loss": 0.3113, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.782537067545305, |
|
"grad_norm": 0.0813733641664339, |
|
"learning_rate": 8.431862171131523e-05, |
|
"loss": 0.3025, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 1.7858319604612851, |
|
"grad_norm": 0.08070284448139889, |
|
"learning_rate": 8.393943200320323e-05, |
|
"loss": 0.3196, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 1.7891268533772653, |
|
"grad_norm": 0.08287982363174917, |
|
"learning_rate": 8.356047918878763e-05, |
|
"loss": 0.3148, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 1.7924217462932455, |
|
"grad_norm": 0.08019250653063249, |
|
"learning_rate": 8.318176885763002e-05, |
|
"loss": 0.3028, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 1.7957166392092256, |
|
"grad_norm": 0.08071265971974027, |
|
"learning_rate": 8.280330659571531e-05, |
|
"loss": 0.3039, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 1.7990115321252058, |
|
"grad_norm": 0.07903551548611944, |
|
"learning_rate": 8.242509798536935e-05, |
|
"loss": 0.3209, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 1.8023064250411862, |
|
"grad_norm": 0.07993774098400507, |
|
"learning_rate": 8.204714860517662e-05, |
|
"loss": 0.315, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 1.8056013179571664, |
|
"grad_norm": 0.08071601924849857, |
|
"learning_rate": 8.1669464029898e-05, |
|
"loss": 0.3103, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 1.8088962108731468, |
|
"grad_norm": 0.08034136007232408, |
|
"learning_rate": 8.129204983038847e-05, |
|
"loss": 0.3055, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 1.812191103789127, |
|
"grad_norm": 0.080811797791999, |
|
"learning_rate": 8.091491157351495e-05, |
|
"loss": 0.3106, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.8154859967051071, |
|
"grad_norm": 0.08057302103747395, |
|
"learning_rate": 8.053805482207419e-05, |
|
"loss": 0.3167, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 1.8187808896210873, |
|
"grad_norm": 0.07972513982263217, |
|
"learning_rate": 8.016148513471077e-05, |
|
"loss": 0.3195, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 1.8220757825370675, |
|
"grad_norm": 0.08065942919433094, |
|
"learning_rate": 7.978520806583502e-05, |
|
"loss": 0.3137, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 1.8253706754530477, |
|
"grad_norm": 0.07866104765728227, |
|
"learning_rate": 7.940922916554122e-05, |
|
"loss": 0.3003, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 1.828665568369028, |
|
"grad_norm": 0.08248026306411695, |
|
"learning_rate": 7.903355397952557e-05, |
|
"loss": 0.3091, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 1.8319604612850082, |
|
"grad_norm": 0.07940956558724678, |
|
"learning_rate": 7.865818804900449e-05, |
|
"loss": 0.3083, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 1.8352553542009886, |
|
"grad_norm": 0.0793106056162054, |
|
"learning_rate": 7.828313691063293e-05, |
|
"loss": 0.3148, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 1.8385502471169688, |
|
"grad_norm": 0.08268047744068477, |
|
"learning_rate": 7.79084060964226e-05, |
|
"loss": 0.2999, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 1.841845140032949, |
|
"grad_norm": 0.08118175241280721, |
|
"learning_rate": 7.753400113366051e-05, |
|
"loss": 0.3169, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 1.8451400329489291, |
|
"grad_norm": 0.0828667048207493, |
|
"learning_rate": 7.71599275448272e-05, |
|
"loss": 0.3186, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.8484349258649093, |
|
"grad_norm": 0.08283621296978765, |
|
"learning_rate": 7.678619084751553e-05, |
|
"loss": 0.3143, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 1.8517298187808895, |
|
"grad_norm": 0.08274280076295433, |
|
"learning_rate": 7.641279655434928e-05, |
|
"loss": 0.3267, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 1.8550247116968699, |
|
"grad_norm": 0.08039004247048885, |
|
"learning_rate": 7.603975017290158e-05, |
|
"loss": 0.3172, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 1.85831960461285, |
|
"grad_norm": 0.07867131124374156, |
|
"learning_rate": 7.566705720561399e-05, |
|
"loss": 0.3126, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 1.8616144975288305, |
|
"grad_norm": 0.08253187445048786, |
|
"learning_rate": 7.529472314971521e-05, |
|
"loss": 0.3115, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 1.8649093904448106, |
|
"grad_norm": 0.08202843679224964, |
|
"learning_rate": 7.492275349713988e-05, |
|
"loss": 0.3114, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 1.8682042833607908, |
|
"grad_norm": 0.08477138135829519, |
|
"learning_rate": 7.455115373444779e-05, |
|
"loss": 0.3194, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 1.871499176276771, |
|
"grad_norm": 0.08194783958889466, |
|
"learning_rate": 7.417992934274279e-05, |
|
"loss": 0.316, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 1.8747940691927512, |
|
"grad_norm": 0.07844240178385448, |
|
"learning_rate": 7.380908579759206e-05, |
|
"loss": 0.3132, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 1.8780889621087313, |
|
"grad_norm": 0.08034073534981348, |
|
"learning_rate": 7.343862856894521e-05, |
|
"loss": 0.3109, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.8813838550247117, |
|
"grad_norm": 0.07683009249076064, |
|
"learning_rate": 7.30685631210537e-05, |
|
"loss": 0.3118, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 1.884678747940692, |
|
"grad_norm": 0.078737692987615, |
|
"learning_rate": 7.26988949123903e-05, |
|
"loss": 0.3037, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 1.8879736408566723, |
|
"grad_norm": 0.08036884301385262, |
|
"learning_rate": 7.232962939556832e-05, |
|
"loss": 0.3077, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 1.8912685337726525, |
|
"grad_norm": 0.08225292912320385, |
|
"learning_rate": 7.196077201726148e-05, |
|
"loss": 0.3153, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 1.8945634266886326, |
|
"grad_norm": 0.0814304024397034, |
|
"learning_rate": 7.159232821812348e-05, |
|
"loss": 0.3222, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 1.8978583196046128, |
|
"grad_norm": 0.08201543804410394, |
|
"learning_rate": 7.122430343270761e-05, |
|
"loss": 0.3085, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 1.901153212520593, |
|
"grad_norm": 0.08252119673034763, |
|
"learning_rate": 7.085670308938675e-05, |
|
"loss": 0.3177, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 1.9044481054365732, |
|
"grad_norm": 0.08193246340388376, |
|
"learning_rate": 7.048953261027328e-05, |
|
"loss": 0.3146, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 1.9077429983525536, |
|
"grad_norm": 0.08467744748067746, |
|
"learning_rate": 7.012279741113909e-05, |
|
"loss": 0.3154, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 1.9110378912685337, |
|
"grad_norm": 0.0796890981600387, |
|
"learning_rate": 6.975650290133554e-05, |
|
"loss": 0.3089, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.9143327841845141, |
|
"grad_norm": 0.08018458516133334, |
|
"learning_rate": 6.939065448371398e-05, |
|
"loss": 0.3145, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 1.9176276771004943, |
|
"grad_norm": 0.08351184814776264, |
|
"learning_rate": 6.902525755454582e-05, |
|
"loss": 0.3072, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 1.9209225700164745, |
|
"grad_norm": 0.08470137530950465, |
|
"learning_rate": 6.866031750344302e-05, |
|
"loss": 0.3199, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 1.9242174629324547, |
|
"grad_norm": 0.08085473387051476, |
|
"learning_rate": 6.829583971327851e-05, |
|
"loss": 0.3135, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 1.9275123558484348, |
|
"grad_norm": 0.07885114940373524, |
|
"learning_rate": 6.793182956010699e-05, |
|
"loss": 0.2951, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 1.930807248764415, |
|
"grad_norm": 0.08341968429527848, |
|
"learning_rate": 6.756829241308542e-05, |
|
"loss": 0.316, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 1.9341021416803954, |
|
"grad_norm": 0.08157140843535905, |
|
"learning_rate": 6.720523363439393e-05, |
|
"loss": 0.3109, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 1.9373970345963756, |
|
"grad_norm": 0.08023033371726958, |
|
"learning_rate": 6.684265857915669e-05, |
|
"loss": 0.3035, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 1.940691927512356, |
|
"grad_norm": 0.08530632792173666, |
|
"learning_rate": 6.648057259536297e-05, |
|
"loss": 0.3179, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 1.9439868204283361, |
|
"grad_norm": 0.0816362830414679, |
|
"learning_rate": 6.611898102378818e-05, |
|
"loss": 0.3116, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.9472817133443163, |
|
"grad_norm": 0.0808945434707495, |
|
"learning_rate": 6.575788919791521e-05, |
|
"loss": 0.3233, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 1.9505766062602965, |
|
"grad_norm": 0.08188718070776142, |
|
"learning_rate": 6.539730244385568e-05, |
|
"loss": 0.319, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 1.9538714991762767, |
|
"grad_norm": 0.07841386595300816, |
|
"learning_rate": 6.503722608027128e-05, |
|
"loss": 0.306, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 1.9571663920922568, |
|
"grad_norm": 0.08232575487936705, |
|
"learning_rate": 6.467766541829557e-05, |
|
"loss": 0.306, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 1.9604612850082372, |
|
"grad_norm": 0.08054867187757639, |
|
"learning_rate": 6.431862576145546e-05, |
|
"loss": 0.3066, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 1.9637561779242174, |
|
"grad_norm": 0.07860962030970246, |
|
"learning_rate": 6.396011240559294e-05, |
|
"loss": 0.3201, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 1.9670510708401978, |
|
"grad_norm": 0.07986110172464526, |
|
"learning_rate": 6.360213063878715e-05, |
|
"loss": 0.3055, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 1.970345963756178, |
|
"grad_norm": 0.0789796364693798, |
|
"learning_rate": 6.324468574127627e-05, |
|
"loss": 0.3097, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 1.9736408566721582, |
|
"grad_norm": 0.0787082961725244, |
|
"learning_rate": 6.288778298537967e-05, |
|
"loss": 0.3109, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 1.9769357495881383, |
|
"grad_norm": 0.08085871461604628, |
|
"learning_rate": 6.253142763541996e-05, |
|
"loss": 0.3082, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.9802306425041185, |
|
"grad_norm": 0.08173480993968053, |
|
"learning_rate": 6.21756249476457e-05, |
|
"loss": 0.3027, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 1.9835255354200987, |
|
"grad_norm": 0.07951252806564535, |
|
"learning_rate": 6.182038017015359e-05, |
|
"loss": 0.3064, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 1.986820428336079, |
|
"grad_norm": 0.07908799027316071, |
|
"learning_rate": 6.14656985428111e-05, |
|
"loss": 0.2988, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 1.9901153212520593, |
|
"grad_norm": 0.07976971510017963, |
|
"learning_rate": 6.111158529717931e-05, |
|
"loss": 0.3076, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 1.9934102141680397, |
|
"grad_norm": 0.07973855612867402, |
|
"learning_rate": 6.075804565643561e-05, |
|
"loss": 0.3052, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 1.9967051070840198, |
|
"grad_norm": 0.08113426466491773, |
|
"learning_rate": 6.0405084835296674e-05, |
|
"loss": 0.3114, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.07946560263483063, |
|
"learning_rate": 6.005270803994165e-05, |
|
"loss": 0.2954, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.3216867446899414, |
|
"eval_runtime": 160.0399, |
|
"eval_samples_per_second": 31.898, |
|
"eval_steps_per_second": 1.0, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 2.00329489291598, |
|
"grad_norm": 0.07877266800911653, |
|
"learning_rate": 5.970092046793523e-05, |
|
"loss": 0.2924, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 2.0065897858319603, |
|
"grad_norm": 0.07812553246498843, |
|
"learning_rate": 5.934972730815115e-05, |
|
"loss": 0.291, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 2.0098846787479405, |
|
"grad_norm": 0.08088679451304487, |
|
"learning_rate": 5.899913374069539e-05, |
|
"loss": 0.2928, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 2.013179571663921, |
|
"grad_norm": 0.07976290454539975, |
|
"learning_rate": 5.864914493683005e-05, |
|
"loss": 0.2894, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 2.0164744645799013, |
|
"grad_norm": 0.07969314347284605, |
|
"learning_rate": 5.829976605889695e-05, |
|
"loss": 0.2982, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 2.0197693574958815, |
|
"grad_norm": 0.0815615988419636, |
|
"learning_rate": 5.795100226024145e-05, |
|
"loss": 0.2971, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 2.0230642504118617, |
|
"grad_norm": 0.08736468979222958, |
|
"learning_rate": 5.760285868513649e-05, |
|
"loss": 0.3038, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 2.026359143327842, |
|
"grad_norm": 0.08409772537245033, |
|
"learning_rate": 5.7255340468706776e-05, |
|
"loss": 0.2859, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 2.029654036243822, |
|
"grad_norm": 0.08020824554821433, |
|
"learning_rate": 5.690845273685279e-05, |
|
"loss": 0.2842, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 2.032948929159802, |
|
"grad_norm": 0.08818956404130317, |
|
"learning_rate": 5.6562200606175495e-05, |
|
"loss": 0.3075, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 2.0362438220757824, |
|
"grad_norm": 0.08599576872590033, |
|
"learning_rate": 5.6216589183900695e-05, |
|
"loss": 0.2966, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 2.039538714991763, |
|
"grad_norm": 0.08560071569717868, |
|
"learning_rate": 5.587162356780376e-05, |
|
"loss": 0.291, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 2.042833607907743, |
|
"grad_norm": 0.0862088898346008, |
|
"learning_rate": 5.552730884613429e-05, |
|
"loss": 0.2912, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 2.0461285008237233, |
|
"grad_norm": 0.08680473893419544, |
|
"learning_rate": 5.5183650097541293e-05, |
|
"loss": 0.3036, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 2.0494233937397035, |
|
"grad_norm": 0.08699770517707416, |
|
"learning_rate": 5.484065239099817e-05, |
|
"loss": 0.296, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 2.0527182866556837, |
|
"grad_norm": 0.08379594467728943, |
|
"learning_rate": 5.449832078572781e-05, |
|
"loss": 0.2921, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 2.056013179571664, |
|
"grad_norm": 0.08411666762788954, |
|
"learning_rate": 5.4156660331128225e-05, |
|
"loss": 0.2911, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 2.059308072487644, |
|
"grad_norm": 0.08606290445123865, |
|
"learning_rate": 5.381567606669794e-05, |
|
"loss": 0.2891, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 2.062602965403624, |
|
"grad_norm": 0.0878110595169254, |
|
"learning_rate": 5.347537302196153e-05, |
|
"loss": 0.3094, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 2.065897858319605, |
|
"grad_norm": 0.08473022854436194, |
|
"learning_rate": 5.313575621639568e-05, |
|
"loss": 0.2935, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 2.069192751235585, |
|
"grad_norm": 0.08277404639202951, |
|
"learning_rate": 5.279683065935505e-05, |
|
"loss": 0.3011, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 2.072487644151565, |
|
"grad_norm": 0.08576477710245138, |
|
"learning_rate": 5.245860134999831e-05, |
|
"loss": 0.298, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 2.0757825370675453, |
|
"grad_norm": 0.0861916805489822, |
|
"learning_rate": 5.212107327721445e-05, |
|
"loss": 0.2914, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 2.0790774299835255, |
|
"grad_norm": 0.08488906269091921, |
|
"learning_rate": 5.178425141954925e-05, |
|
"loss": 0.294, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 2.0823723228995057, |
|
"grad_norm": 0.08098586035137433, |
|
"learning_rate": 5.1448140745131844e-05, |
|
"loss": 0.2855, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 2.085667215815486, |
|
"grad_norm": 0.08838145724831743, |
|
"learning_rate": 5.111274621160127e-05, |
|
"loss": 0.2952, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 2.088962108731466, |
|
"grad_norm": 0.08648279828081434, |
|
"learning_rate": 5.077807276603357e-05, |
|
"loss": 0.2928, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 2.0922570016474467, |
|
"grad_norm": 0.08314911866387735, |
|
"learning_rate": 5.044412534486873e-05, |
|
"loss": 0.2945, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 2.095551894563427, |
|
"grad_norm": 0.08461362273542454, |
|
"learning_rate": 5.01109088738378e-05, |
|
"loss": 0.299, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 2.098846787479407, |
|
"grad_norm": 0.08377057965335477, |
|
"learning_rate": 4.9778428267890345e-05, |
|
"loss": 0.2891, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 2.102141680395387, |
|
"grad_norm": 0.08543795983675807, |
|
"learning_rate": 4.9446688431121944e-05, |
|
"loss": 0.2912, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 2.1054365733113674, |
|
"grad_norm": 0.0819773341791353, |
|
"learning_rate": 4.911569425670168e-05, |
|
"loss": 0.2945, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 2.1087314662273475, |
|
"grad_norm": 0.08749676356833817, |
|
"learning_rate": 4.878545062680025e-05, |
|
"loss": 0.3111, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 2.1120263591433277, |
|
"grad_norm": 0.08931336745649711, |
|
"learning_rate": 4.845596241251773e-05, |
|
"loss": 0.2967, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 2.115321252059308, |
|
"grad_norm": 0.08738403486711675, |
|
"learning_rate": 4.8127234473811855e-05, |
|
"loss": 0.2983, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 2.1186161449752885, |
|
"grad_norm": 0.08631801633541691, |
|
"learning_rate": 4.7799271659426156e-05, |
|
"loss": 0.2979, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 2.1219110378912687, |
|
"grad_norm": 0.08204476625680207, |
|
"learning_rate": 4.74720788068187e-05, |
|
"loss": 0.2911, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 2.125205930807249, |
|
"grad_norm": 0.08540431532875102, |
|
"learning_rate": 4.714566074209058e-05, |
|
"loss": 0.2934, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 2.128500823723229, |
|
"grad_norm": 0.0867243714824401, |
|
"learning_rate": 4.682002227991466e-05, |
|
"loss": 0.295, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 2.131795716639209, |
|
"grad_norm": 0.08772574559749194, |
|
"learning_rate": 4.6495168223464734e-05, |
|
"loss": 0.2992, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 2.1350906095551894, |
|
"grad_norm": 0.0862846034965142, |
|
"learning_rate": 4.617110336434464e-05, |
|
"loss": 0.2934, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 2.1383855024711695, |
|
"grad_norm": 0.08279698024636946, |
|
"learning_rate": 4.5847832482517386e-05, |
|
"loss": 0.2905, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 2.1416803953871497, |
|
"grad_norm": 0.08449602542472846, |
|
"learning_rate": 4.5525360346234916e-05, |
|
"loss": 0.2762, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 2.1449752883031303, |
|
"grad_norm": 0.08512352335672224, |
|
"learning_rate": 4.520369171196765e-05, |
|
"loss": 0.2819, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 2.1482701812191105, |
|
"grad_norm": 0.08942106754915288, |
|
"learning_rate": 4.488283132433437e-05, |
|
"loss": 0.2809, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 2.1515650741350907, |
|
"grad_norm": 0.08538361680827215, |
|
"learning_rate": 4.456278391603207e-05, |
|
"loss": 0.2875, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 2.154859967051071, |
|
"grad_norm": 0.08716447313431239, |
|
"learning_rate": 4.42435542077664e-05, |
|
"loss": 0.3049, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 2.158154859967051, |
|
"grad_norm": 0.08722819981927521, |
|
"learning_rate": 4.392514690818193e-05, |
|
"loss": 0.2947, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 2.161449752883031, |
|
"grad_norm": 0.08703119195800675, |
|
"learning_rate": 4.360756671379258e-05, |
|
"loss": 0.2899, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 2.1647446457990114, |
|
"grad_norm": 0.08554895733632949, |
|
"learning_rate": 4.329081830891253e-05, |
|
"loss": 0.3008, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 2.168039538714992, |
|
"grad_norm": 0.0855723358947586, |
|
"learning_rate": 4.2974906365587095e-05, |
|
"loss": 0.2948, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 2.171334431630972, |
|
"grad_norm": 0.08669144280813641, |
|
"learning_rate": 4.265983554352361e-05, |
|
"loss": 0.2918, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 2.1746293245469523, |
|
"grad_norm": 0.08587506440809513, |
|
"learning_rate": 4.2345610490023004e-05, |
|
"loss": 0.3007, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 2.1779242174629325, |
|
"grad_norm": 0.08325186764285471, |
|
"learning_rate": 4.203223583991103e-05, |
|
"loss": 0.2862, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 2.1812191103789127, |
|
"grad_norm": 0.0848179385036651, |
|
"learning_rate": 4.171971621547004e-05, |
|
"loss": 0.2941, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 2.184514003294893, |
|
"grad_norm": 0.08744598220624812, |
|
"learning_rate": 4.140805622637062e-05, |
|
"loss": 0.2884, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 2.187808896210873, |
|
"grad_norm": 0.08730152511128712, |
|
"learning_rate": 4.1097260469603815e-05, |
|
"loss": 0.2939, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 2.191103789126853, |
|
"grad_norm": 0.08479459661361502, |
|
"learning_rate": 4.0787333529413216e-05, |
|
"loss": 0.2907, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 2.1943986820428334, |
|
"grad_norm": 0.08748979890541152, |
|
"learning_rate": 4.0478279977227364e-05, |
|
"loss": 0.3054, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 2.197693574958814, |
|
"grad_norm": 0.08688506963459419, |
|
"learning_rate": 4.0170104371592196e-05, |
|
"loss": 0.294, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 2.200988467874794, |
|
"grad_norm": 0.08813970885177967, |
|
"learning_rate": 3.986281125810408e-05, |
|
"loss": 0.2906, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 2.2042833607907744, |
|
"grad_norm": 0.0862215239726716, |
|
"learning_rate": 3.9556405169342506e-05, |
|
"loss": 0.2919, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 2.2075782537067545, |
|
"grad_norm": 0.0864141576056617, |
|
"learning_rate": 3.925089062480339e-05, |
|
"loss": 0.2988, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 2.2108731466227347, |
|
"grad_norm": 0.08741141927551971, |
|
"learning_rate": 3.8946272130832276e-05, |
|
"loss": 0.2973, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 2.214168039538715, |
|
"grad_norm": 0.087264203309356, |
|
"learning_rate": 3.864255418055801e-05, |
|
"loss": 0.2984, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 2.217462932454695, |
|
"grad_norm": 0.0863024056441452, |
|
"learning_rate": 3.8339741253826386e-05, |
|
"loss": 0.2894, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 2.2207578253706757, |
|
"grad_norm": 0.08686962836143866, |
|
"learning_rate": 3.803783781713411e-05, |
|
"loss": 0.2943, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 2.224052718286656, |
|
"grad_norm": 0.088506180884522, |
|
"learning_rate": 3.7736848323562804e-05, |
|
"loss": 0.2936, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 2.227347611202636, |
|
"grad_norm": 0.08429663780585792, |
|
"learning_rate": 3.7436777212713484e-05, |
|
"loss": 0.2883, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 2.230642504118616, |
|
"grad_norm": 0.0861610009496725, |
|
"learning_rate": 3.7137628910641e-05, |
|
"loss": 0.3014, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 2.2339373970345964, |
|
"grad_norm": 0.08744439598418424, |
|
"learning_rate": 3.683940782978875e-05, |
|
"loss": 0.3027, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 2.2372322899505765, |
|
"grad_norm": 0.08478804953508354, |
|
"learning_rate": 3.654211836892356e-05, |
|
"loss": 0.2885, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 2.2405271828665567, |
|
"grad_norm": 0.08396188944325574, |
|
"learning_rate": 3.624576491307088e-05, |
|
"loss": 0.2898, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 2.243822075782537, |
|
"grad_norm": 0.08464943476670307, |
|
"learning_rate": 3.595035183345007e-05, |
|
"loss": 0.2932, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 2.247116968698517, |
|
"grad_norm": 0.08361972692496587, |
|
"learning_rate": 3.565588348740995e-05, |
|
"loss": 0.2846, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 2.2504118616144977, |
|
"grad_norm": 0.08473337066131743, |
|
"learning_rate": 3.536236421836438e-05, |
|
"loss": 0.2873, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 2.253706754530478, |
|
"grad_norm": 0.08560947782135529, |
|
"learning_rate": 3.5069798355728456e-05, |
|
"loss": 0.2952, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 2.257001647446458, |
|
"grad_norm": 0.08273696228936674, |
|
"learning_rate": 3.477819021485448e-05, |
|
"loss": 0.2843, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 2.260296540362438, |
|
"grad_norm": 0.0856458280242604, |
|
"learning_rate": 3.448754409696823e-05, |
|
"loss": 0.2862, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 2.2635914332784184, |
|
"grad_norm": 0.0849473074749724, |
|
"learning_rate": 3.4197864289105763e-05, |
|
"loss": 0.2844, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 2.2668863261943986, |
|
"grad_norm": 0.08752747087542179, |
|
"learning_rate": 3.390915506405e-05, |
|
"loss": 0.3009, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 2.2701812191103787, |
|
"grad_norm": 0.08515710595917833, |
|
"learning_rate": 3.36214206802677e-05, |
|
"loss": 0.2917, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 2.2734761120263594, |
|
"grad_norm": 0.08735259929721675, |
|
"learning_rate": 3.3334665381846744e-05, |
|
"loss": 0.3025, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 2.2767710049423395, |
|
"grad_norm": 0.08615519636042336, |
|
"learning_rate": 3.3048893398433465e-05, |
|
"loss": 0.293, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 2.2800658978583197, |
|
"grad_norm": 0.08824900221463129, |
|
"learning_rate": 3.2764108945170336e-05, |
|
"loss": 0.3008, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 2.2833607907743, |
|
"grad_norm": 0.08484959409699497, |
|
"learning_rate": 3.248031622263361e-05, |
|
"loss": 0.2873, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 2.28665568369028, |
|
"grad_norm": 0.08724514821548604, |
|
"learning_rate": 3.2197519416771615e-05, |
|
"loss": 0.2973, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 2.2899505766062602, |
|
"grad_norm": 0.08393767608817229, |
|
"learning_rate": 3.1915722698842877e-05, |
|
"loss": 0.2841, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 2.2932454695222404, |
|
"grad_norm": 0.08641309723191103, |
|
"learning_rate": 3.163493022535451e-05, |
|
"loss": 0.2872, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 2.2965403624382206, |
|
"grad_norm": 0.08788590177688262, |
|
"learning_rate": 3.135514613800108e-05, |
|
"loss": 0.2995, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 2.2998352553542007, |
|
"grad_norm": 0.08531593399084322, |
|
"learning_rate": 3.107637456360348e-05, |
|
"loss": 0.2894, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 2.3031301482701814, |
|
"grad_norm": 0.0872718919185223, |
|
"learning_rate": 3.079861961404789e-05, |
|
"loss": 0.2887, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 2.3064250411861615, |
|
"grad_norm": 0.08612255232765391, |
|
"learning_rate": 3.052188538622535e-05, |
|
"loss": 0.2901, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 2.3097199341021417, |
|
"grad_norm": 0.08922298918575919, |
|
"learning_rate": 3.0246175961971212e-05, |
|
"loss": 0.2937, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 2.313014827018122, |
|
"grad_norm": 0.0882010224849904, |
|
"learning_rate": 2.997149540800498e-05, |
|
"loss": 0.2931, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 2.316309719934102, |
|
"grad_norm": 0.08622719885654752, |
|
"learning_rate": 2.9697847775870224e-05, |
|
"loss": 0.2898, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 2.3196046128500822, |
|
"grad_norm": 0.08359805006615195, |
|
"learning_rate": 2.942523710187496e-05, |
|
"loss": 0.2847, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 2.3228995057660624, |
|
"grad_norm": 0.08449003074110172, |
|
"learning_rate": 2.915366740703207e-05, |
|
"loss": 0.28, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 2.326194398682043, |
|
"grad_norm": 0.08855451697096253, |
|
"learning_rate": 2.8883142696999878e-05, |
|
"loss": 0.2987, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 2.329489291598023, |
|
"grad_norm": 0.08622350256458039, |
|
"learning_rate": 2.8613666962023254e-05, |
|
"loss": 0.2968, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 2.3327841845140034, |
|
"grad_norm": 0.08652168396802937, |
|
"learning_rate": 2.8345244176874663e-05, |
|
"loss": 0.2823, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 2.3360790774299836, |
|
"grad_norm": 0.08823592083642218, |
|
"learning_rate": 2.8077878300795448e-05, |
|
"loss": 0.2912, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 2.3393739703459637, |
|
"grad_norm": 0.08507829603456869, |
|
"learning_rate": 2.7811573277437608e-05, |
|
"loss": 0.2835, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 2.342668863261944, |
|
"grad_norm": 0.08549841674015039, |
|
"learning_rate": 2.7546333034805527e-05, |
|
"loss": 0.292, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 2.345963756177924, |
|
"grad_norm": 0.08443979295008668, |
|
"learning_rate": 2.7282161485198078e-05, |
|
"loss": 0.2974, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 2.3492586490939047, |
|
"grad_norm": 0.08503200569748393, |
|
"learning_rate": 2.7019062525150784e-05, |
|
"loss": 0.3016, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 2.352553542009885, |
|
"grad_norm": 0.08504863770154245, |
|
"learning_rate": 2.6757040035378568e-05, |
|
"loss": 0.2984, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 2.355848434925865, |
|
"grad_norm": 0.08644443469983008, |
|
"learning_rate": 2.6496097880718364e-05, |
|
"loss": 0.2954, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 2.359143327841845, |
|
"grad_norm": 0.08915999578886315, |
|
"learning_rate": 2.6236239910072068e-05, |
|
"loss": 0.2988, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 2.3624382207578254, |
|
"grad_norm": 0.08416603357610633, |
|
"learning_rate": 2.5977469956349954e-05, |
|
"loss": 0.2879, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 2.3657331136738056, |
|
"grad_norm": 0.08817807859466884, |
|
"learning_rate": 2.5719791836413976e-05, |
|
"loss": 0.299, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 2.3690280065897857, |
|
"grad_norm": 0.08541077975584914, |
|
"learning_rate": 2.5463209351021456e-05, |
|
"loss": 0.2901, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 2.372322899505766, |
|
"grad_norm": 0.08470671550098466, |
|
"learning_rate": 2.5207726284769194e-05, |
|
"loss": 0.303, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 2.375617792421746, |
|
"grad_norm": 0.08819661663546222, |
|
"learning_rate": 2.495334640603746e-05, |
|
"loss": 0.2888, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 2.3789126853377267, |
|
"grad_norm": 0.08591537862829479, |
|
"learning_rate": 2.470007346693455e-05, |
|
"loss": 0.2948, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 2.382207578253707, |
|
"grad_norm": 0.0875830979203168, |
|
"learning_rate": 2.444791120324127e-05, |
|
"loss": 0.2915, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 2.385502471169687, |
|
"grad_norm": 0.08613104051905594, |
|
"learning_rate": 2.419686333435606e-05, |
|
"loss": 0.2972, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 2.3887973640856672, |
|
"grad_norm": 0.08782242765122425, |
|
"learning_rate": 2.3946933563239972e-05, |
|
"loss": 0.299, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 2.3920922570016474, |
|
"grad_norm": 0.08680612007460217, |
|
"learning_rate": 2.369812557636204e-05, |
|
"loss": 0.2943, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 2.3953871499176276, |
|
"grad_norm": 0.08603749003168497, |
|
"learning_rate": 2.3450443043645032e-05, |
|
"loss": 0.3012, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 2.3986820428336078, |
|
"grad_norm": 0.08657552256692291, |
|
"learning_rate": 2.3203889618411233e-05, |
|
"loss": 0.2906, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 2.4019769357495884, |
|
"grad_norm": 0.08704450296844311, |
|
"learning_rate": 2.2958468937328526e-05, |
|
"loss": 0.2951, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 2.4052718286655685, |
|
"grad_norm": 0.0855118301451358, |
|
"learning_rate": 2.2714184620356827e-05, |
|
"loss": 0.2931, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 2.4085667215815487, |
|
"grad_norm": 0.08724807161271898, |
|
"learning_rate": 2.2471040270694666e-05, |
|
"loss": 0.2917, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 2.411861614497529, |
|
"grad_norm": 0.08600975835485788, |
|
"learning_rate": 2.2229039474726053e-05, |
|
"loss": 0.296, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 2.415156507413509, |
|
"grad_norm": 0.08864784516812786, |
|
"learning_rate": 2.1988185801967464e-05, |
|
"loss": 0.2972, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 2.4184514003294892, |
|
"grad_norm": 0.08762596803411996, |
|
"learning_rate": 2.174848280501538e-05, |
|
"loss": 0.2906, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 2.4217462932454694, |
|
"grad_norm": 0.09035462710118229, |
|
"learning_rate": 2.150993401949376e-05, |
|
"loss": 0.2951, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 2.4250411861614496, |
|
"grad_norm": 0.08755876151137919, |
|
"learning_rate": 2.127254296400195e-05, |
|
"loss": 0.2901, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 2.4283360790774298, |
|
"grad_norm": 0.08766926155438874, |
|
"learning_rate": 2.103631314006267e-05, |
|
"loss": 0.2934, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 2.4316309719934104, |
|
"grad_norm": 0.08606023381132255, |
|
"learning_rate": 2.0801248032070542e-05, |
|
"loss": 0.2846, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 2.4349258649093906, |
|
"grad_norm": 0.08809592154677096, |
|
"learning_rate": 2.0567351107240563e-05, |
|
"loss": 0.2916, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 2.4382207578253707, |
|
"grad_norm": 0.0904592519515893, |
|
"learning_rate": 2.033462581555703e-05, |
|
"loss": 0.2819, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 2.441515650741351, |
|
"grad_norm": 0.08617703763043115, |
|
"learning_rate": 2.0103075589722575e-05, |
|
"loss": 0.2908, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 2.444810543657331, |
|
"grad_norm": 0.08964817977842639, |
|
"learning_rate": 1.9872703845107643e-05, |
|
"loss": 0.305, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 2.4481054365733113, |
|
"grad_norm": 0.09015699027576811, |
|
"learning_rate": 1.9643513979700033e-05, |
|
"loss": 0.2954, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 2.4514003294892914, |
|
"grad_norm": 0.08997440049824991, |
|
"learning_rate": 1.941550937405483e-05, |
|
"loss": 0.2926, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 2.454695222405272, |
|
"grad_norm": 0.09114828822884588, |
|
"learning_rate": 1.918869339124444e-05, |
|
"loss": 0.296, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 2.4579901153212522, |
|
"grad_norm": 0.09027991136572545, |
|
"learning_rate": 1.896306937680913e-05, |
|
"loss": 0.2977, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 2.4612850082372324, |
|
"grad_norm": 0.08660452640872197, |
|
"learning_rate": 1.8738640658707585e-05, |
|
"loss": 0.2885, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 2.4645799011532126, |
|
"grad_norm": 0.08979882751073061, |
|
"learning_rate": 1.8515410547267875e-05, |
|
"loss": 0.2906, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 2.4678747940691927, |
|
"grad_norm": 0.08618796776224673, |
|
"learning_rate": 1.829338233513853e-05, |
|
"loss": 0.2972, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 2.471169686985173, |
|
"grad_norm": 0.08735262583561819, |
|
"learning_rate": 1.80725592972401e-05, |
|
"loss": 0.2851, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.474464579901153, |
|
"grad_norm": 0.08587818073514078, |
|
"learning_rate": 1.7852944690716768e-05, |
|
"loss": 0.298, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 2.4777594728171333, |
|
"grad_norm": 0.09093449501445203, |
|
"learning_rate": 1.7634541754888367e-05, |
|
"loss": 0.2996, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 2.4810543657331134, |
|
"grad_norm": 0.08576101421203153, |
|
"learning_rate": 1.7417353711202478e-05, |
|
"loss": 0.2892, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 2.484349258649094, |
|
"grad_norm": 0.08796655318401572, |
|
"learning_rate": 1.7201383763187085e-05, |
|
"loss": 0.3064, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 2.4876441515650742, |
|
"grad_norm": 0.08917308260607959, |
|
"learning_rate": 1.6986635096403215e-05, |
|
"loss": 0.296, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 2.4909390444810544, |
|
"grad_norm": 0.08712369748650076, |
|
"learning_rate": 1.6773110878397935e-05, |
|
"loss": 0.2892, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 2.4942339373970346, |
|
"grad_norm": 0.08830732733373077, |
|
"learning_rate": 1.6560814258657686e-05, |
|
"loss": 0.2868, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 2.4975288303130148, |
|
"grad_norm": 0.08886428573765547, |
|
"learning_rate": 1.634974836856188e-05, |
|
"loss": 0.2975, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 2.500823723228995, |
|
"grad_norm": 0.09001998396318142, |
|
"learning_rate": 1.6139916321336512e-05, |
|
"loss": 0.3011, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 2.504118616144975, |
|
"grad_norm": 0.0866643459541065, |
|
"learning_rate": 1.5931321212008466e-05, |
|
"loss": 0.2791, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 2.5074135090609557, |
|
"grad_norm": 0.08969541659303033, |
|
"learning_rate": 1.5723966117359746e-05, |
|
"loss": 0.2845, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 2.510708401976936, |
|
"grad_norm": 0.08678120829747425, |
|
"learning_rate": 1.5517854095882124e-05, |
|
"loss": 0.2897, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 2.514003294892916, |
|
"grad_norm": 0.08768949528062786, |
|
"learning_rate": 1.531298818773197e-05, |
|
"loss": 0.2926, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 2.5172981878088962, |
|
"grad_norm": 0.08613630689405796, |
|
"learning_rate": 1.5109371414685513e-05, |
|
"loss": 0.2994, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 2.5205930807248764, |
|
"grad_norm": 0.08818859976681247, |
|
"learning_rate": 1.4907006780094213e-05, |
|
"loss": 0.2915, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 2.5238879736408566, |
|
"grad_norm": 0.08916911710423102, |
|
"learning_rate": 1.4705897268840396e-05, |
|
"loss": 0.3087, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 2.5271828665568368, |
|
"grad_norm": 0.08860834462061187, |
|
"learning_rate": 1.4506045847293359e-05, |
|
"loss": 0.2936, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 2.5304777594728174, |
|
"grad_norm": 0.08563535943112995, |
|
"learning_rate": 1.4307455463265573e-05, |
|
"loss": 0.2855, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 2.533772652388797, |
|
"grad_norm": 0.08826978729379603, |
|
"learning_rate": 1.411012904596909e-05, |
|
"loss": 0.3115, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 2.5370675453047777, |
|
"grad_norm": 0.0870890210971865, |
|
"learning_rate": 1.3914069505972483e-05, |
|
"loss": 0.2905, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 2.540362438220758, |
|
"grad_norm": 0.08354870518763845, |
|
"learning_rate": 1.3719279735157874e-05, |
|
"loss": 0.2916, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 2.543657331136738, |
|
"grad_norm": 0.08910519459239277, |
|
"learning_rate": 1.3525762606678271e-05, |
|
"loss": 0.3048, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 2.5469522240527183, |
|
"grad_norm": 0.08365455172188432, |
|
"learning_rate": 1.3333520974915093e-05, |
|
"loss": 0.292, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 2.5502471169686984, |
|
"grad_norm": 0.08911881943765725, |
|
"learning_rate": 1.3142557675436262e-05, |
|
"loss": 0.2983, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 2.5535420098846786, |
|
"grad_norm": 0.08661195377868255, |
|
"learning_rate": 1.2952875524954233e-05, |
|
"loss": 0.2916, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 2.556836902800659, |
|
"grad_norm": 0.08738778807939013, |
|
"learning_rate": 1.2764477321284474e-05, |
|
"loss": 0.2827, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 2.5601317957166394, |
|
"grad_norm": 0.08675909170932133, |
|
"learning_rate": 1.2577365843304212e-05, |
|
"loss": 0.2922, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 2.5634266886326196, |
|
"grad_norm": 0.08653045636242618, |
|
"learning_rate": 1.2391543850911514e-05, |
|
"loss": 0.289, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 2.5667215815485998, |
|
"grad_norm": 0.08794743727959366, |
|
"learning_rate": 1.2207014084984381e-05, |
|
"loss": 0.3006, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 2.57001647446458, |
|
"grad_norm": 0.08765042878548815, |
|
"learning_rate": 1.2023779267340562e-05, |
|
"loss": 0.3053, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 2.57331136738056, |
|
"grad_norm": 0.08506311043178476, |
|
"learning_rate": 1.1841842100697254e-05, |
|
"loss": 0.287, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 2.5766062602965403, |
|
"grad_norm": 0.08559193608270571, |
|
"learning_rate": 1.1661205268631247e-05, |
|
"loss": 0.2913, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 2.5799011532125204, |
|
"grad_norm": 0.08817756260222355, |
|
"learning_rate": 1.1481871435539415e-05, |
|
"loss": 0.2912, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 2.583196046128501, |
|
"grad_norm": 0.0871445033944061, |
|
"learning_rate": 1.1303843246599344e-05, |
|
"loss": 0.2807, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 2.586490939044481, |
|
"grad_norm": 0.08535578251400154, |
|
"learning_rate": 1.1127123327730381e-05, |
|
"loss": 0.2818, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 2.5897858319604614, |
|
"grad_norm": 0.08761486848867922, |
|
"learning_rate": 1.0951714285554782e-05, |
|
"loss": 0.3059, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 2.5930807248764416, |
|
"grad_norm": 0.08719801254842507, |
|
"learning_rate": 1.0777618707359428e-05, |
|
"loss": 0.2825, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 2.5963756177924218, |
|
"grad_norm": 0.09034119263808828, |
|
"learning_rate": 1.0604839161057567e-05, |
|
"loss": 0.2946, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 2.599670510708402, |
|
"grad_norm": 0.08954661242149643, |
|
"learning_rate": 1.043337819515089e-05, |
|
"loss": 0.2852, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 2.602965403624382, |
|
"grad_norm": 0.08724138640911316, |
|
"learning_rate": 1.0263238338692061e-05, |
|
"loss": 0.2907, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 2.6062602965403623, |
|
"grad_norm": 0.086330949719512, |
|
"learning_rate": 1.0094422101247369e-05, |
|
"loss": 0.2868, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 2.6095551894563425, |
|
"grad_norm": 0.08529314875981138, |
|
"learning_rate": 9.926931972859622e-06, |
|
"loss": 0.2862, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 2.612850082372323, |
|
"grad_norm": 0.08538099892093676, |
|
"learning_rate": 9.760770424011555e-06, |
|
"loss": 0.2842, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 2.6161449752883033, |
|
"grad_norm": 0.08346633575820249, |
|
"learning_rate": 9.595939905589301e-06, |
|
"loss": 0.2872, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 2.6194398682042834, |
|
"grad_norm": 0.09087513588899795, |
|
"learning_rate": 9.43244284884629e-06, |
|
"loss": 0.2991, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 2.6227347611202636, |
|
"grad_norm": 0.08970315413157041, |
|
"learning_rate": 9.270281665367331e-06, |
|
"loss": 0.2864, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 2.6260296540362438, |
|
"grad_norm": 0.08831508853553123, |
|
"learning_rate": 9.109458747033106e-06, |
|
"loss": 0.299, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 2.629324546952224, |
|
"grad_norm": 0.08962119945659898, |
|
"learning_rate": 8.949976465984878e-06, |
|
"loss": 0.2933, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 2.632619439868204, |
|
"grad_norm": 0.0903147303314654, |
|
"learning_rate": 8.791837174589402e-06, |
|
"loss": 0.291, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 2.6359143327841847, |
|
"grad_norm": 0.0874384283159893, |
|
"learning_rate": 8.63504320540438e-06, |
|
"loss": 0.2947, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.6392092257001645, |
|
"grad_norm": 0.09114762691001292, |
|
"learning_rate": 8.47959687114398e-06, |
|
"loss": 0.3005, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 2.642504118616145, |
|
"grad_norm": 0.08629238625513358, |
|
"learning_rate": 8.325500464644731e-06, |
|
"loss": 0.2893, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 2.6457990115321253, |
|
"grad_norm": 0.08689109034275155, |
|
"learning_rate": 8.172756258831638e-06, |
|
"loss": 0.2861, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 2.6490939044481054, |
|
"grad_norm": 0.08687013125668969, |
|
"learning_rate": 8.021366506684802e-06, |
|
"loss": 0.3018, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 2.6523887973640856, |
|
"grad_norm": 0.08733316173692361, |
|
"learning_rate": 7.871333441206053e-06, |
|
"loss": 0.2918, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 2.655683690280066, |
|
"grad_norm": 0.08666514271966166, |
|
"learning_rate": 7.722659275386101e-06, |
|
"loss": 0.2893, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 2.658978583196046, |
|
"grad_norm": 0.08572152703241656, |
|
"learning_rate": 7.575346202171818e-06, |
|
"loss": 0.2845, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 2.662273476112026, |
|
"grad_norm": 0.08616248420611884, |
|
"learning_rate": 7.429396394433952e-06, |
|
"loss": 0.2908, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 2.6655683690280068, |
|
"grad_norm": 0.08544220799292981, |
|
"learning_rate": 7.284812004935082e-06, |
|
"loss": 0.2828, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 2.668863261943987, |
|
"grad_norm": 0.08530855466026539, |
|
"learning_rate": 7.141595166297832e-06, |
|
"loss": 0.2921, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.672158154859967, |
|
"grad_norm": 0.08676630765622653, |
|
"learning_rate": 6.999747990973382e-06, |
|
"loss": 0.2824, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 2.6754530477759473, |
|
"grad_norm": 0.0858286642847653, |
|
"learning_rate": 6.859272571210385e-06, |
|
"loss": 0.2931, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 2.6787479406919275, |
|
"grad_norm": 0.08895340909580104, |
|
"learning_rate": 6.720170979024065e-06, |
|
"loss": 0.2974, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 2.6820428336079076, |
|
"grad_norm": 0.08704509442550527, |
|
"learning_rate": 6.5824452661656936e-06, |
|
"loss": 0.2946, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 2.685337726523888, |
|
"grad_norm": 0.08719224060765357, |
|
"learning_rate": 6.446097464092249e-06, |
|
"loss": 0.2825, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 2.6886326194398684, |
|
"grad_norm": 0.08505418031859806, |
|
"learning_rate": 6.311129583936504e-06, |
|
"loss": 0.2857, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 2.6919275123558486, |
|
"grad_norm": 0.08777962549058228, |
|
"learning_rate": 6.177543616477377e-06, |
|
"loss": 0.2918, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 2.6952224052718288, |
|
"grad_norm": 0.08802224386633083, |
|
"learning_rate": 6.04534153211056e-06, |
|
"loss": 0.2943, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 2.698517298187809, |
|
"grad_norm": 0.08672625342077533, |
|
"learning_rate": 5.914525280819383e-06, |
|
"loss": 0.2892, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 2.701812191103789, |
|
"grad_norm": 0.08512834708737091, |
|
"learning_rate": 5.785096792146161e-06, |
|
"loss": 0.2961, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 2.7051070840197693, |
|
"grad_norm": 0.08644060832701138, |
|
"learning_rate": 5.6570579751636825e-06, |
|
"loss": 0.2828, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 2.7084019769357495, |
|
"grad_norm": 0.08586016865471162, |
|
"learning_rate": 5.53041071844701e-06, |
|
"loss": 0.2847, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 2.71169686985173, |
|
"grad_norm": 0.08748419080120705, |
|
"learning_rate": 5.405156890045704e-06, |
|
"loss": 0.2914, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 2.71499176276771, |
|
"grad_norm": 0.08611456546743239, |
|
"learning_rate": 5.2812983374562195e-06, |
|
"loss": 0.2817, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 2.7182866556836904, |
|
"grad_norm": 0.08601404856441977, |
|
"learning_rate": 5.158836887594687e-06, |
|
"loss": 0.2838, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 2.7215815485996706, |
|
"grad_norm": 0.08519086941747171, |
|
"learning_rate": 5.037774346769874e-06, |
|
"loss": 0.2911, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 2.724876441515651, |
|
"grad_norm": 0.08802028933104723, |
|
"learning_rate": 4.91811250065668e-06, |
|
"loss": 0.2857, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 2.728171334431631, |
|
"grad_norm": 0.08817444740319591, |
|
"learning_rate": 4.799853114269725e-06, |
|
"loss": 0.2882, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 2.731466227347611, |
|
"grad_norm": 0.08474998779590631, |
|
"learning_rate": 4.682997931937283e-06, |
|
"loss": 0.2857, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 2.7347611202635913, |
|
"grad_norm": 0.08826586804083016, |
|
"learning_rate": 4.567548677275602e-06, |
|
"loss": 0.3008, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 2.7380560131795715, |
|
"grad_norm": 0.08589345490586381, |
|
"learning_rate": 4.4535070531635196e-06, |
|
"loss": 0.2792, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 2.741350906095552, |
|
"grad_norm": 0.08688465761385525, |
|
"learning_rate": 4.340874741717194e-06, |
|
"loss": 0.2821, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 2.7446457990115323, |
|
"grad_norm": 0.08775893601760283, |
|
"learning_rate": 4.2296534042654985e-06, |
|
"loss": 0.2866, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 2.7479406919275124, |
|
"grad_norm": 0.08369390812057961, |
|
"learning_rate": 4.119844681325347e-06, |
|
"loss": 0.2847, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 2.7512355848434926, |
|
"grad_norm": 0.08608061264894241, |
|
"learning_rate": 4.011450192577593e-06, |
|
"loss": 0.294, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 2.754530477759473, |
|
"grad_norm": 0.08747895302707503, |
|
"learning_rate": 3.9044715368430796e-06, |
|
"loss": 0.2925, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 2.757825370675453, |
|
"grad_norm": 0.09108934336270781, |
|
"learning_rate": 3.79891029205911e-06, |
|
"loss": 0.2986, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 2.761120263591433, |
|
"grad_norm": 0.08739799542018062, |
|
"learning_rate": 3.6947680152561448e-06, |
|
"loss": 0.2862, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 2.7644151565074138, |
|
"grad_norm": 0.08517335050043656, |
|
"learning_rate": 3.592046242534819e-06, |
|
"loss": 0.2801, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 2.7677100494233935, |
|
"grad_norm": 0.087382204616756, |
|
"learning_rate": 3.490746489043317e-06, |
|
"loss": 0.3014, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 2.771004942339374, |
|
"grad_norm": 0.0874720007338996, |
|
"learning_rate": 3.3908702489550246e-06, |
|
"loss": 0.2897, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 2.7742998352553543, |
|
"grad_norm": 0.08620311190604066, |
|
"learning_rate": 3.292418995446445e-06, |
|
"loss": 0.2917, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 2.7775947281713345, |
|
"grad_norm": 0.08850431386846441, |
|
"learning_rate": 3.195394180675526e-06, |
|
"loss": 0.2876, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 2.7808896210873146, |
|
"grad_norm": 0.08718749143095159, |
|
"learning_rate": 3.0997972357602267e-06, |
|
"loss": 0.2915, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 2.784184514003295, |
|
"grad_norm": 0.08432627868786195, |
|
"learning_rate": 3.0056295707573732e-06, |
|
"loss": 0.2807, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 2.787479406919275, |
|
"grad_norm": 0.08758768949078255, |
|
"learning_rate": 2.912892574641879e-06, |
|
"loss": 0.3013, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 2.790774299835255, |
|
"grad_norm": 0.08774352124282939, |
|
"learning_rate": 2.8215876152862695e-06, |
|
"loss": 0.2953, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 2.7940691927512358, |
|
"grad_norm": 0.08578852522202607, |
|
"learning_rate": 2.731716039440546e-06, |
|
"loss": 0.2849, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 2.797364085667216, |
|
"grad_norm": 0.08692795875841626, |
|
"learning_rate": 2.6432791727121984e-06, |
|
"loss": 0.2887, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 2.800658978583196, |
|
"grad_norm": 0.0875787452554814, |
|
"learning_rate": 2.5562783195467676e-06, |
|
"loss": 0.2876, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 2.8039538714991763, |
|
"grad_norm": 0.08941768041430048, |
|
"learning_rate": 2.4707147632085812e-06, |
|
"loss": 0.2905, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 2.8072487644151565, |
|
"grad_norm": 0.08799458088680515, |
|
"learning_rate": 2.386589765761771e-06, |
|
"loss": 0.3009, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 2.8105436573311366, |
|
"grad_norm": 0.08621413521392886, |
|
"learning_rate": 2.303904568051729e-06, |
|
"loss": 0.2853, |
|
"step": 853 |
|
}, |
|
{ |
|
"epoch": 2.813838550247117, |
|
"grad_norm": 0.08830565616085327, |
|
"learning_rate": 2.222660389686759e-06, |
|
"loss": 0.2916, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 2.8171334431630974, |
|
"grad_norm": 0.08465275130891076, |
|
"learning_rate": 2.1428584290201117e-06, |
|
"loss": 0.2936, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 2.820428336079077, |
|
"grad_norm": 0.08848516159290687, |
|
"learning_rate": 2.0644998631322743e-06, |
|
"loss": 0.2897, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 2.823723228995058, |
|
"grad_norm": 0.08667388930109197, |
|
"learning_rate": 1.9875858478136555e-06, |
|
"loss": 0.2864, |
|
"step": 857 |
|
}, |
|
{ |
|
"epoch": 2.827018121911038, |
|
"grad_norm": 0.08730797196019499, |
|
"learning_rate": 1.912117517547518e-06, |
|
"loss": 0.2921, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 2.830313014827018, |
|
"grad_norm": 0.08754556569098132, |
|
"learning_rate": 1.8380959854932046e-06, |
|
"loss": 0.2845, |
|
"step": 859 |
|
}, |
|
{ |
|
"epoch": 2.8336079077429983, |
|
"grad_norm": 0.0888004567935744, |
|
"learning_rate": 1.7655223434698055e-06, |
|
"loss": 0.2944, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 2.8369028006589785, |
|
"grad_norm": 0.08629780498350664, |
|
"learning_rate": 1.6943976619399615e-06, |
|
"loss": 0.2894, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 2.8401976935749587, |
|
"grad_norm": 0.08678492881753877, |
|
"learning_rate": 1.624722989994143e-06, |
|
"loss": 0.2922, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 2.843492586490939, |
|
"grad_norm": 0.08681999478349432, |
|
"learning_rate": 1.5564993553351393e-06, |
|
"loss": 0.2941, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 2.8467874794069195, |
|
"grad_norm": 0.08455558885488165, |
|
"learning_rate": 1.489727764262927e-06, |
|
"loss": 0.2823, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 2.8500823723228996, |
|
"grad_norm": 0.0869909676987709, |
|
"learning_rate": 1.4244092016597932e-06, |
|
"loss": 0.2897, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 2.85337726523888, |
|
"grad_norm": 0.08999029894977731, |
|
"learning_rate": 1.360544630975813e-06, |
|
"loss": 0.2925, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 2.85667215815486, |
|
"grad_norm": 0.08908986656975491, |
|
"learning_rate": 1.2981349942146947e-06, |
|
"loss": 0.2935, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 2.85996705107084, |
|
"grad_norm": 0.0852396248401669, |
|
"learning_rate": 1.2371812119198133e-06, |
|
"loss": 0.2912, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 2.8632619439868203, |
|
"grad_norm": 0.09176483266172472, |
|
"learning_rate": 1.1776841831606545e-06, |
|
"loss": 0.2938, |
|
"step": 869 |
|
}, |
|
{ |
|
"epoch": 2.8665568369028005, |
|
"grad_norm": 0.08602564538788204, |
|
"learning_rate": 1.1196447855195802e-06, |
|
"loss": 0.2901, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 2.869851729818781, |
|
"grad_norm": 0.08784529858760343, |
|
"learning_rate": 1.0630638750788623e-06, |
|
"loss": 0.2975, |
|
"step": 871 |
|
}, |
|
{ |
|
"epoch": 2.873146622734761, |
|
"grad_norm": 0.08971622964903737, |
|
"learning_rate": 1.007942286408048e-06, |
|
"loss": 0.2925, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 2.8764415156507415, |
|
"grad_norm": 0.08941711599334909, |
|
"learning_rate": 9.542808325516571e-07, |
|
"loss": 0.2985, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 2.8797364085667216, |
|
"grad_norm": 0.08771600586759773, |
|
"learning_rate": 9.020803050172055e-07, |
|
"loss": 0.2891, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 2.883031301482702, |
|
"grad_norm": 0.087619918065711, |
|
"learning_rate": 8.513414737635006e-07, |
|
"loss": 0.2993, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 2.886326194398682, |
|
"grad_norm": 0.08574010553502064, |
|
"learning_rate": 8.020650871893299e-07, |
|
"loss": 0.283, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 2.889621087314662, |
|
"grad_norm": 0.08895621856403858, |
|
"learning_rate": 7.542518721223469e-07, |
|
"loss": 0.2953, |
|
"step": 877 |
|
}, |
|
{ |
|
"epoch": 2.892915980230643, |
|
"grad_norm": 0.08752216931371197, |
|
"learning_rate": 7.079025338084356e-07, |
|
"loss": 0.3001, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 2.8962108731466225, |
|
"grad_norm": 0.08592028048367323, |
|
"learning_rate": 6.630177559012518e-07, |
|
"loss": 0.2932, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 2.899505766062603, |
|
"grad_norm": 0.08922288724108252, |
|
"learning_rate": 6.195982004521539e-07, |
|
"loss": 0.2919, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 2.9028006589785833, |
|
"grad_norm": 0.08607983338366068, |
|
"learning_rate": 5.776445079004656e-07, |
|
"loss": 0.2865, |
|
"step": 881 |
|
}, |
|
{ |
|
"epoch": 2.9060955518945635, |
|
"grad_norm": 0.08631649082660897, |
|
"learning_rate": 5.371572970639727e-07, |
|
"loss": 0.2954, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 2.9093904448105437, |
|
"grad_norm": 0.08780715085907931, |
|
"learning_rate": 4.981371651298305e-07, |
|
"loss": 0.2973, |
|
"step": 883 |
|
}, |
|
{ |
|
"epoch": 2.912685337726524, |
|
"grad_norm": 0.08653539418303985, |
|
"learning_rate": 4.605846876457709e-07, |
|
"loss": 0.287, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 2.915980230642504, |
|
"grad_norm": 0.08853983076560318, |
|
"learning_rate": 4.245004185115753e-07, |
|
"loss": 0.2926, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 2.919275123558484, |
|
"grad_norm": 0.08688701796384708, |
|
"learning_rate": 3.8988488997092623e-07, |
|
"loss": 0.2909, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 2.922570016474465, |
|
"grad_norm": 0.08707666200796532, |
|
"learning_rate": 3.5673861260355767e-07, |
|
"loss": 0.2916, |
|
"step": 887 |
|
}, |
|
{ |
|
"epoch": 2.925864909390445, |
|
"grad_norm": 0.0881348831962719, |
|
"learning_rate": 3.25062075317728e-07, |
|
"loss": 0.298, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 2.929159802306425, |
|
"grad_norm": 0.08957445515530212, |
|
"learning_rate": 2.948557453429701e-07, |
|
"loss": 0.2976, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 2.9324546952224053, |
|
"grad_norm": 0.08509653098922053, |
|
"learning_rate": 2.6612006822327454e-07, |
|
"loss": 0.2813, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 2.9357495881383855, |
|
"grad_norm": 0.08699474462722692, |
|
"learning_rate": 2.3885546781042824e-07, |
|
"loss": 0.2784, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 2.9390444810543657, |
|
"grad_norm": 0.08668396855759522, |
|
"learning_rate": 2.1306234625784182e-07, |
|
"loss": 0.2832, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 2.942339373970346, |
|
"grad_norm": 0.08908080709823125, |
|
"learning_rate": 1.8874108401456536e-07, |
|
"loss": 0.2943, |
|
"step": 893 |
|
}, |
|
{ |
|
"epoch": 2.9456342668863265, |
|
"grad_norm": 0.087253529375096, |
|
"learning_rate": 1.658920398196928e-07, |
|
"loss": 0.2955, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 2.948929159802306, |
|
"grad_norm": 0.08662330916170789, |
|
"learning_rate": 1.4451555069708856e-07, |
|
"loss": 0.2791, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 2.952224052718287, |
|
"grad_norm": 0.0874866376451616, |
|
"learning_rate": 1.2461193195038022e-07, |
|
"loss": 0.2854, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 2.955518945634267, |
|
"grad_norm": 0.08661370692532423, |
|
"learning_rate": 1.0618147715835137e-07, |
|
"loss": 0.2871, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 2.958813838550247, |
|
"grad_norm": 0.08427881713621545, |
|
"learning_rate": 8.922445817056701e-08, |
|
"loss": 0.2776, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 2.9621087314662273, |
|
"grad_norm": 0.08831048147016833, |
|
"learning_rate": 7.374112510339926e-08, |
|
"loss": 0.3043, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 2.9654036243822075, |
|
"grad_norm": 0.08457458559860552, |
|
"learning_rate": 5.973170633631897e-08, |
|
"loss": 0.2852, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 2.9686985172981877, |
|
"grad_norm": 0.08674405906715675, |
|
"learning_rate": 4.719640850852081e-08, |
|
"loss": 0.2873, |
|
"step": 901 |
|
}, |
|
{ |
|
"epoch": 2.971993410214168, |
|
"grad_norm": 0.08956957869825671, |
|
"learning_rate": 3.6135416515903356e-08, |
|
"loss": 0.3088, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 2.9752883031301485, |
|
"grad_norm": 0.08742057262425709, |
|
"learning_rate": 2.6548893508315798e-08, |
|
"loss": 0.2914, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 2.9785831960461286, |
|
"grad_norm": 0.08686074450100637, |
|
"learning_rate": 1.8436980887170942e-08, |
|
"loss": 0.2823, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 2.981878088962109, |
|
"grad_norm": 0.08958760682591677, |
|
"learning_rate": 1.1799798303335775e-08, |
|
"loss": 0.2935, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 2.985172981878089, |
|
"grad_norm": 0.08621279240787924, |
|
"learning_rate": 6.637443655366227e-09, |
|
"loss": 0.2798, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 2.988467874794069, |
|
"grad_norm": 0.08689572544052393, |
|
"learning_rate": 2.949993088130487e-09, |
|
"loss": 0.2945, |
|
"step": 907 |
|
}, |
|
{ |
|
"epoch": 2.9917627677100493, |
|
"grad_norm": 0.08696925538627061, |
|
"learning_rate": 7.375009915655539e-10, |
|
"loss": 0.2918, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 2.9950576606260295, |
|
"grad_norm": 0.08749368325157687, |
|
"learning_rate": 0.0, |
|
"loss": 0.2889, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 2.9950576606260295, |
|
"eval_loss": 0.31883931159973145, |
|
"eval_runtime": 159.817, |
|
"eval_samples_per_second": 31.943, |
|
"eval_steps_per_second": 1.001, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 2.9950576606260295, |
|
"step": 909, |
|
"total_flos": 3.2428259312245146e+17, |
|
"train_loss": 0.34365460066774367, |
|
"train_runtime": 10432.531, |
|
"train_samples_per_second": 11.155, |
|
"train_steps_per_second": 0.087 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 909, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.2428259312245146e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|