watsonchua's picture
Upload folder using huggingface_hub
046fdc8 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9994397759103641,
"eval_steps": 500,
"global_step": 446,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.002240896358543417,
"grad_norm": 0.22982779145240784,
"learning_rate": 1.1235955056179775e-06,
"loss": 1.1973,
"step": 1
},
{
"epoch": 0.004481792717086834,
"grad_norm": 0.2098255604505539,
"learning_rate": 2.247191011235955e-06,
"loss": 1.2685,
"step": 2
},
{
"epoch": 0.0067226890756302525,
"grad_norm": 0.2404344230890274,
"learning_rate": 3.3707865168539327e-06,
"loss": 1.2292,
"step": 3
},
{
"epoch": 0.008963585434173669,
"grad_norm": 0.23824891448020935,
"learning_rate": 4.49438202247191e-06,
"loss": 1.1901,
"step": 4
},
{
"epoch": 0.011204481792717087,
"grad_norm": 0.22885526716709137,
"learning_rate": 5.617977528089888e-06,
"loss": 1.2695,
"step": 5
},
{
"epoch": 0.013445378151260505,
"grad_norm": 0.2576353847980499,
"learning_rate": 6.741573033707865e-06,
"loss": 1.2872,
"step": 6
},
{
"epoch": 0.01568627450980392,
"grad_norm": 0.2190636247396469,
"learning_rate": 7.865168539325843e-06,
"loss": 1.1952,
"step": 7
},
{
"epoch": 0.017927170868347338,
"grad_norm": 0.21357473731040955,
"learning_rate": 8.98876404494382e-06,
"loss": 1.2597,
"step": 8
},
{
"epoch": 0.020168067226890758,
"grad_norm": 0.24502670764923096,
"learning_rate": 1.0112359550561798e-05,
"loss": 1.279,
"step": 9
},
{
"epoch": 0.022408963585434174,
"grad_norm": 0.25532686710357666,
"learning_rate": 1.1235955056179776e-05,
"loss": 1.2163,
"step": 10
},
{
"epoch": 0.02464985994397759,
"grad_norm": 0.250683456659317,
"learning_rate": 1.2359550561797752e-05,
"loss": 1.2735,
"step": 11
},
{
"epoch": 0.02689075630252101,
"grad_norm": 0.2523595988750458,
"learning_rate": 1.348314606741573e-05,
"loss": 1.3471,
"step": 12
},
{
"epoch": 0.029131652661064426,
"grad_norm": 0.23875640332698822,
"learning_rate": 1.4606741573033709e-05,
"loss": 1.2595,
"step": 13
},
{
"epoch": 0.03137254901960784,
"grad_norm": 0.24998927116394043,
"learning_rate": 1.5730337078651687e-05,
"loss": 1.2142,
"step": 14
},
{
"epoch": 0.03361344537815126,
"grad_norm": 0.2783028483390808,
"learning_rate": 1.6853932584269665e-05,
"loss": 1.185,
"step": 15
},
{
"epoch": 0.035854341736694675,
"grad_norm": 0.2796604633331299,
"learning_rate": 1.797752808988764e-05,
"loss": 1.226,
"step": 16
},
{
"epoch": 0.0380952380952381,
"grad_norm": 0.2899184823036194,
"learning_rate": 1.9101123595505618e-05,
"loss": 1.1672,
"step": 17
},
{
"epoch": 0.040336134453781515,
"grad_norm": 0.2601150572299957,
"learning_rate": 2.0224719101123596e-05,
"loss": 1.1127,
"step": 18
},
{
"epoch": 0.04257703081232493,
"grad_norm": 0.3052348792552948,
"learning_rate": 2.1348314606741574e-05,
"loss": 1.2865,
"step": 19
},
{
"epoch": 0.04481792717086835,
"grad_norm": 0.24783451855182648,
"learning_rate": 2.2471910112359552e-05,
"loss": 1.0888,
"step": 20
},
{
"epoch": 0.047058823529411764,
"grad_norm": 0.24703693389892578,
"learning_rate": 2.359550561797753e-05,
"loss": 1.059,
"step": 21
},
{
"epoch": 0.04929971988795518,
"grad_norm": 0.2503679096698761,
"learning_rate": 2.4719101123595505e-05,
"loss": 1.1602,
"step": 22
},
{
"epoch": 0.0515406162464986,
"grad_norm": 0.2639842927455902,
"learning_rate": 2.5842696629213486e-05,
"loss": 1.0911,
"step": 23
},
{
"epoch": 0.05378151260504202,
"grad_norm": 0.2942507266998291,
"learning_rate": 2.696629213483146e-05,
"loss": 1.0777,
"step": 24
},
{
"epoch": 0.056022408963585436,
"grad_norm": 0.28088828921318054,
"learning_rate": 2.8089887640449443e-05,
"loss": 1.1111,
"step": 25
},
{
"epoch": 0.05826330532212885,
"grad_norm": 0.26791295409202576,
"learning_rate": 2.9213483146067417e-05,
"loss": 1.0016,
"step": 26
},
{
"epoch": 0.06050420168067227,
"grad_norm": 0.2685791850090027,
"learning_rate": 3.0337078651685396e-05,
"loss": 1.1085,
"step": 27
},
{
"epoch": 0.06274509803921569,
"grad_norm": 0.2627420127391815,
"learning_rate": 3.1460674157303374e-05,
"loss": 1.0704,
"step": 28
},
{
"epoch": 0.06498599439775911,
"grad_norm": 0.3000424802303314,
"learning_rate": 3.258426966292135e-05,
"loss": 1.0786,
"step": 29
},
{
"epoch": 0.06722689075630252,
"grad_norm": 0.3018706738948822,
"learning_rate": 3.370786516853933e-05,
"loss": 1.0423,
"step": 30
},
{
"epoch": 0.06946778711484594,
"grad_norm": 0.27565667033195496,
"learning_rate": 3.483146067415731e-05,
"loss": 0.9834,
"step": 31
},
{
"epoch": 0.07170868347338935,
"grad_norm": 0.30244842171669006,
"learning_rate": 3.595505617977528e-05,
"loss": 1.0927,
"step": 32
},
{
"epoch": 0.07394957983193277,
"grad_norm": 0.3654678165912628,
"learning_rate": 3.7078651685393264e-05,
"loss": 0.9812,
"step": 33
},
{
"epoch": 0.0761904761904762,
"grad_norm": 0.36483272910118103,
"learning_rate": 3.8202247191011236e-05,
"loss": 0.9693,
"step": 34
},
{
"epoch": 0.0784313725490196,
"grad_norm": 0.2949022054672241,
"learning_rate": 3.9325842696629214e-05,
"loss": 0.9647,
"step": 35
},
{
"epoch": 0.08067226890756303,
"grad_norm": 0.36239683628082275,
"learning_rate": 4.044943820224719e-05,
"loss": 0.9725,
"step": 36
},
{
"epoch": 0.08291316526610644,
"grad_norm": 0.32511067390441895,
"learning_rate": 4.157303370786517e-05,
"loss": 1.0136,
"step": 37
},
{
"epoch": 0.08515406162464986,
"grad_norm": 0.32111090421676636,
"learning_rate": 4.269662921348315e-05,
"loss": 0.9207,
"step": 38
},
{
"epoch": 0.08739495798319327,
"grad_norm": 0.3080519735813141,
"learning_rate": 4.3820224719101126e-05,
"loss": 0.9729,
"step": 39
},
{
"epoch": 0.0896358543417367,
"grad_norm": 0.2933235764503479,
"learning_rate": 4.4943820224719104e-05,
"loss": 0.9403,
"step": 40
},
{
"epoch": 0.09187675070028012,
"grad_norm": 0.3132348358631134,
"learning_rate": 4.606741573033708e-05,
"loss": 0.9587,
"step": 41
},
{
"epoch": 0.09411764705882353,
"grad_norm": 0.3013119101524353,
"learning_rate": 4.719101123595506e-05,
"loss": 0.9162,
"step": 42
},
{
"epoch": 0.09635854341736695,
"grad_norm": 0.300485223531723,
"learning_rate": 4.831460674157304e-05,
"loss": 0.8984,
"step": 43
},
{
"epoch": 0.09859943977591036,
"grad_norm": 0.2726304233074188,
"learning_rate": 4.943820224719101e-05,
"loss": 0.987,
"step": 44
},
{
"epoch": 0.10084033613445378,
"grad_norm": 0.2733825445175171,
"learning_rate": 5.0561797752808995e-05,
"loss": 0.9438,
"step": 45
},
{
"epoch": 0.1030812324929972,
"grad_norm": 0.28922319412231445,
"learning_rate": 5.168539325842697e-05,
"loss": 0.8674,
"step": 46
},
{
"epoch": 0.10532212885154062,
"grad_norm": 0.2743085026741028,
"learning_rate": 5.2808988764044944e-05,
"loss": 0.9041,
"step": 47
},
{
"epoch": 0.10756302521008404,
"grad_norm": 0.28649550676345825,
"learning_rate": 5.393258426966292e-05,
"loss": 0.9057,
"step": 48
},
{
"epoch": 0.10980392156862745,
"grad_norm": 0.2877427339553833,
"learning_rate": 5.50561797752809e-05,
"loss": 0.9139,
"step": 49
},
{
"epoch": 0.11204481792717087,
"grad_norm": 0.27738648653030396,
"learning_rate": 5.6179775280898885e-05,
"loss": 0.8518,
"step": 50
},
{
"epoch": 0.11428571428571428,
"grad_norm": 0.2839404046535492,
"learning_rate": 5.730337078651685e-05,
"loss": 0.8634,
"step": 51
},
{
"epoch": 0.1165266106442577,
"grad_norm": 0.2748688757419586,
"learning_rate": 5.8426966292134835e-05,
"loss": 0.9504,
"step": 52
},
{
"epoch": 0.11876750700280111,
"grad_norm": 0.2953556776046753,
"learning_rate": 5.955056179775281e-05,
"loss": 0.861,
"step": 53
},
{
"epoch": 0.12100840336134454,
"grad_norm": 0.2947392165660858,
"learning_rate": 6.067415730337079e-05,
"loss": 0.8578,
"step": 54
},
{
"epoch": 0.12324929971988796,
"grad_norm": 0.3029733896255493,
"learning_rate": 6.179775280898876e-05,
"loss": 0.895,
"step": 55
},
{
"epoch": 0.12549019607843137,
"grad_norm": 0.28483161330223083,
"learning_rate": 6.292134831460675e-05,
"loss": 0.9316,
"step": 56
},
{
"epoch": 0.12773109243697478,
"grad_norm": 0.280499666929245,
"learning_rate": 6.404494382022472e-05,
"loss": 0.8763,
"step": 57
},
{
"epoch": 0.12997198879551822,
"grad_norm": 0.2687634825706482,
"learning_rate": 6.51685393258427e-05,
"loss": 0.9216,
"step": 58
},
{
"epoch": 0.13221288515406163,
"grad_norm": 0.2869422435760498,
"learning_rate": 6.629213483146067e-05,
"loss": 0.8783,
"step": 59
},
{
"epoch": 0.13445378151260504,
"grad_norm": 0.29140859842300415,
"learning_rate": 6.741573033707866e-05,
"loss": 0.9613,
"step": 60
},
{
"epoch": 0.13669467787114845,
"grad_norm": 0.29342207312583923,
"learning_rate": 6.853932584269663e-05,
"loss": 0.982,
"step": 61
},
{
"epoch": 0.13893557422969188,
"grad_norm": 0.29143285751342773,
"learning_rate": 6.966292134831462e-05,
"loss": 0.8857,
"step": 62
},
{
"epoch": 0.1411764705882353,
"grad_norm": 0.2944166660308838,
"learning_rate": 7.078651685393259e-05,
"loss": 0.8668,
"step": 63
},
{
"epoch": 0.1434173669467787,
"grad_norm": 0.27430447936058044,
"learning_rate": 7.191011235955056e-05,
"loss": 0.9403,
"step": 64
},
{
"epoch": 0.14565826330532214,
"grad_norm": 0.274800568819046,
"learning_rate": 7.303370786516854e-05,
"loss": 0.8689,
"step": 65
},
{
"epoch": 0.14789915966386555,
"grad_norm": 0.3063383102416992,
"learning_rate": 7.415730337078653e-05,
"loss": 0.9217,
"step": 66
},
{
"epoch": 0.15014005602240896,
"grad_norm": 0.26958584785461426,
"learning_rate": 7.52808988764045e-05,
"loss": 0.949,
"step": 67
},
{
"epoch": 0.1523809523809524,
"grad_norm": 0.3030094504356384,
"learning_rate": 7.640449438202247e-05,
"loss": 0.8814,
"step": 68
},
{
"epoch": 0.1546218487394958,
"grad_norm": 0.26790764927864075,
"learning_rate": 7.752808988764046e-05,
"loss": 0.9814,
"step": 69
},
{
"epoch": 0.1568627450980392,
"grad_norm": 0.28143975138664246,
"learning_rate": 7.865168539325843e-05,
"loss": 0.8667,
"step": 70
},
{
"epoch": 0.15910364145658262,
"grad_norm": 0.32539746165275574,
"learning_rate": 7.97752808988764e-05,
"loss": 1.0709,
"step": 71
},
{
"epoch": 0.16134453781512606,
"grad_norm": 0.28114452958106995,
"learning_rate": 8.089887640449438e-05,
"loss": 0.9253,
"step": 72
},
{
"epoch": 0.16358543417366947,
"grad_norm": 0.2680191993713379,
"learning_rate": 8.202247191011237e-05,
"loss": 0.9345,
"step": 73
},
{
"epoch": 0.16582633053221288,
"grad_norm": 0.2928783893585205,
"learning_rate": 8.314606741573034e-05,
"loss": 0.906,
"step": 74
},
{
"epoch": 0.16806722689075632,
"grad_norm": 0.2979438900947571,
"learning_rate": 8.426966292134831e-05,
"loss": 0.9581,
"step": 75
},
{
"epoch": 0.17030812324929973,
"grad_norm": 0.27883946895599365,
"learning_rate": 8.53932584269663e-05,
"loss": 0.9203,
"step": 76
},
{
"epoch": 0.17254901960784313,
"grad_norm": 0.3945503234863281,
"learning_rate": 8.651685393258427e-05,
"loss": 0.8136,
"step": 77
},
{
"epoch": 0.17478991596638654,
"grad_norm": 0.27328184247016907,
"learning_rate": 8.764044943820225e-05,
"loss": 0.8966,
"step": 78
},
{
"epoch": 0.17703081232492998,
"grad_norm": 0.3049486577510834,
"learning_rate": 8.876404494382022e-05,
"loss": 0.9658,
"step": 79
},
{
"epoch": 0.1792717086834734,
"grad_norm": 0.3143669366836548,
"learning_rate": 8.988764044943821e-05,
"loss": 0.9867,
"step": 80
},
{
"epoch": 0.1815126050420168,
"grad_norm": 0.30048221349716187,
"learning_rate": 9.101123595505618e-05,
"loss": 0.9249,
"step": 81
},
{
"epoch": 0.18375350140056024,
"grad_norm": 0.28177568316459656,
"learning_rate": 9.213483146067416e-05,
"loss": 0.9899,
"step": 82
},
{
"epoch": 0.18599439775910365,
"grad_norm": 0.296478807926178,
"learning_rate": 9.325842696629214e-05,
"loss": 0.8476,
"step": 83
},
{
"epoch": 0.18823529411764706,
"grad_norm": 0.2855791449546814,
"learning_rate": 9.438202247191012e-05,
"loss": 0.8869,
"step": 84
},
{
"epoch": 0.19047619047619047,
"grad_norm": 0.2829147279262543,
"learning_rate": 9.550561797752809e-05,
"loss": 0.9274,
"step": 85
},
{
"epoch": 0.1927170868347339,
"grad_norm": 0.2549046277999878,
"learning_rate": 9.662921348314608e-05,
"loss": 0.8535,
"step": 86
},
{
"epoch": 0.1949579831932773,
"grad_norm": 0.2655014991760254,
"learning_rate": 9.775280898876405e-05,
"loss": 0.8905,
"step": 87
},
{
"epoch": 0.19719887955182072,
"grad_norm": 0.27607807517051697,
"learning_rate": 9.887640449438202e-05,
"loss": 0.8788,
"step": 88
},
{
"epoch": 0.19943977591036416,
"grad_norm": 0.24970707297325134,
"learning_rate": 0.0001,
"loss": 0.7632,
"step": 89
},
{
"epoch": 0.20168067226890757,
"grad_norm": 0.2821432650089264,
"learning_rate": 0.00010112359550561799,
"loss": 0.9314,
"step": 90
},
{
"epoch": 0.20392156862745098,
"grad_norm": 0.28368303179740906,
"learning_rate": 0.00010224719101123596,
"loss": 0.8424,
"step": 91
},
{
"epoch": 0.2061624649859944,
"grad_norm": 0.3087138235569,
"learning_rate": 0.00010337078651685395,
"loss": 0.9787,
"step": 92
},
{
"epoch": 0.20840336134453782,
"grad_norm": 0.2952103614807129,
"learning_rate": 0.00010449438202247193,
"loss": 0.9679,
"step": 93
},
{
"epoch": 0.21064425770308123,
"grad_norm": 0.2875281572341919,
"learning_rate": 0.00010561797752808989,
"loss": 0.9181,
"step": 94
},
{
"epoch": 0.21288515406162464,
"grad_norm": 0.3084465265274048,
"learning_rate": 0.00010674157303370786,
"loss": 0.9615,
"step": 95
},
{
"epoch": 0.21512605042016808,
"grad_norm": 0.28005871176719666,
"learning_rate": 0.00010786516853932584,
"loss": 0.8757,
"step": 96
},
{
"epoch": 0.2173669467787115,
"grad_norm": 0.2795560657978058,
"learning_rate": 0.00010898876404494383,
"loss": 0.8266,
"step": 97
},
{
"epoch": 0.2196078431372549,
"grad_norm": 0.2671497166156769,
"learning_rate": 0.0001101123595505618,
"loss": 0.8693,
"step": 98
},
{
"epoch": 0.2218487394957983,
"grad_norm": 0.24397605657577515,
"learning_rate": 0.00011123595505617979,
"loss": 0.8334,
"step": 99
},
{
"epoch": 0.22408963585434175,
"grad_norm": 0.26632094383239746,
"learning_rate": 0.00011235955056179777,
"loss": 0.966,
"step": 100
},
{
"epoch": 0.22633053221288515,
"grad_norm": 0.2785671055316925,
"learning_rate": 0.00011348314606741574,
"loss": 0.8823,
"step": 101
},
{
"epoch": 0.22857142857142856,
"grad_norm": 0.26920077204704285,
"learning_rate": 0.0001146067415730337,
"loss": 0.9726,
"step": 102
},
{
"epoch": 0.230812324929972,
"grad_norm": 0.2633483409881592,
"learning_rate": 0.00011573033707865168,
"loss": 0.8849,
"step": 103
},
{
"epoch": 0.2330532212885154,
"grad_norm": 0.26563239097595215,
"learning_rate": 0.00011685393258426967,
"loss": 0.7896,
"step": 104
},
{
"epoch": 0.23529411764705882,
"grad_norm": 0.27475470304489136,
"learning_rate": 0.00011797752808988764,
"loss": 0.9659,
"step": 105
},
{
"epoch": 0.23753501400560223,
"grad_norm": 0.2691081166267395,
"learning_rate": 0.00011910112359550563,
"loss": 0.9758,
"step": 106
},
{
"epoch": 0.23977591036414567,
"grad_norm": 0.2714226543903351,
"learning_rate": 0.00012022471910112361,
"loss": 0.8736,
"step": 107
},
{
"epoch": 0.24201680672268908,
"grad_norm": 0.2638857960700989,
"learning_rate": 0.00012134831460674158,
"loss": 0.8792,
"step": 108
},
{
"epoch": 0.24425770308123249,
"grad_norm": 0.25696009397506714,
"learning_rate": 0.00012247191011235955,
"loss": 0.8813,
"step": 109
},
{
"epoch": 0.24649859943977592,
"grad_norm": 0.27648770809173584,
"learning_rate": 0.00012359550561797752,
"loss": 0.9341,
"step": 110
},
{
"epoch": 0.24873949579831933,
"grad_norm": 0.27543121576309204,
"learning_rate": 0.0001247191011235955,
"loss": 0.8102,
"step": 111
},
{
"epoch": 0.25098039215686274,
"grad_norm": 0.2804222106933594,
"learning_rate": 0.0001258426966292135,
"loss": 0.9004,
"step": 112
},
{
"epoch": 0.25322128851540615,
"grad_norm": 0.27446436882019043,
"learning_rate": 0.00012696629213483147,
"loss": 0.8432,
"step": 113
},
{
"epoch": 0.25546218487394956,
"grad_norm": 0.27675163745880127,
"learning_rate": 0.00012808988764044944,
"loss": 0.8868,
"step": 114
},
{
"epoch": 0.25770308123249297,
"grad_norm": 0.24772736430168152,
"learning_rate": 0.00012921348314606744,
"loss": 0.8711,
"step": 115
},
{
"epoch": 0.25994397759103643,
"grad_norm": 0.2736036479473114,
"learning_rate": 0.0001303370786516854,
"loss": 0.922,
"step": 116
},
{
"epoch": 0.26218487394957984,
"grad_norm": 0.23815171420574188,
"learning_rate": 0.00013146067415730338,
"loss": 0.8469,
"step": 117
},
{
"epoch": 0.26442577030812325,
"grad_norm": 0.2564987540245056,
"learning_rate": 0.00013258426966292135,
"loss": 0.7626,
"step": 118
},
{
"epoch": 0.26666666666666666,
"grad_norm": 0.26105883717536926,
"learning_rate": 0.00013370786516853932,
"loss": 0.7911,
"step": 119
},
{
"epoch": 0.2689075630252101,
"grad_norm": 0.2915550768375397,
"learning_rate": 0.00013483146067415732,
"loss": 0.9643,
"step": 120
},
{
"epoch": 0.2711484593837535,
"grad_norm": 0.28031301498413086,
"learning_rate": 0.0001359550561797753,
"loss": 0.8797,
"step": 121
},
{
"epoch": 0.2733893557422969,
"grad_norm": 0.2468908429145813,
"learning_rate": 0.00013707865168539326,
"loss": 0.8497,
"step": 122
},
{
"epoch": 0.27563025210084036,
"grad_norm": 0.2498752921819687,
"learning_rate": 0.00013820224719101123,
"loss": 0.842,
"step": 123
},
{
"epoch": 0.27787114845938377,
"grad_norm": 0.2517074942588806,
"learning_rate": 0.00013932584269662923,
"loss": 0.8072,
"step": 124
},
{
"epoch": 0.2801120448179272,
"grad_norm": 0.24205273389816284,
"learning_rate": 0.0001404494382022472,
"loss": 0.796,
"step": 125
},
{
"epoch": 0.2823529411764706,
"grad_norm": 0.259343683719635,
"learning_rate": 0.00014157303370786517,
"loss": 0.8609,
"step": 126
},
{
"epoch": 0.284593837535014,
"grad_norm": 0.2564795911312103,
"learning_rate": 0.00014269662921348315,
"loss": 0.8976,
"step": 127
},
{
"epoch": 0.2868347338935574,
"grad_norm": 0.2650969922542572,
"learning_rate": 0.00014382022471910112,
"loss": 0.8239,
"step": 128
},
{
"epoch": 0.28907563025210087,
"grad_norm": 0.26443368196487427,
"learning_rate": 0.00014494382022471912,
"loss": 0.8479,
"step": 129
},
{
"epoch": 0.2913165266106443,
"grad_norm": 0.26897069811820984,
"learning_rate": 0.0001460674157303371,
"loss": 0.9104,
"step": 130
},
{
"epoch": 0.2935574229691877,
"grad_norm": 0.2578631341457367,
"learning_rate": 0.00014719101123595506,
"loss": 0.8151,
"step": 131
},
{
"epoch": 0.2957983193277311,
"grad_norm": 0.2454749345779419,
"learning_rate": 0.00014831460674157306,
"loss": 0.7915,
"step": 132
},
{
"epoch": 0.2980392156862745,
"grad_norm": 0.25589731335639954,
"learning_rate": 0.00014943820224719103,
"loss": 0.8952,
"step": 133
},
{
"epoch": 0.3002801120448179,
"grad_norm": 0.2591662108898163,
"learning_rate": 0.000150561797752809,
"loss": 0.8112,
"step": 134
},
{
"epoch": 0.3025210084033613,
"grad_norm": 0.26816102862358093,
"learning_rate": 0.00015168539325842697,
"loss": 0.8893,
"step": 135
},
{
"epoch": 0.3047619047619048,
"grad_norm": 0.24405767023563385,
"learning_rate": 0.00015280898876404494,
"loss": 0.877,
"step": 136
},
{
"epoch": 0.3070028011204482,
"grad_norm": 0.2588540315628052,
"learning_rate": 0.00015393258426966294,
"loss": 0.7779,
"step": 137
},
{
"epoch": 0.3092436974789916,
"grad_norm": 0.27598896622657776,
"learning_rate": 0.0001550561797752809,
"loss": 0.7752,
"step": 138
},
{
"epoch": 0.311484593837535,
"grad_norm": 0.24622836709022522,
"learning_rate": 0.00015617977528089888,
"loss": 0.8849,
"step": 139
},
{
"epoch": 0.3137254901960784,
"grad_norm": 0.2502545118331909,
"learning_rate": 0.00015730337078651685,
"loss": 0.8012,
"step": 140
},
{
"epoch": 0.31596638655462184,
"grad_norm": 0.25384724140167236,
"learning_rate": 0.00015842696629213485,
"loss": 0.8395,
"step": 141
},
{
"epoch": 0.31820728291316525,
"grad_norm": 0.2527698874473572,
"learning_rate": 0.0001595505617977528,
"loss": 0.831,
"step": 142
},
{
"epoch": 0.3204481792717087,
"grad_norm": 0.24567513167858124,
"learning_rate": 0.0001606741573033708,
"loss": 0.7959,
"step": 143
},
{
"epoch": 0.3226890756302521,
"grad_norm": 0.24283619225025177,
"learning_rate": 0.00016179775280898877,
"loss": 0.8146,
"step": 144
},
{
"epoch": 0.32492997198879553,
"grad_norm": 0.2708129286766052,
"learning_rate": 0.00016292134831460674,
"loss": 0.905,
"step": 145
},
{
"epoch": 0.32717086834733894,
"grad_norm": 0.27091729640960693,
"learning_rate": 0.00016404494382022474,
"loss": 0.8681,
"step": 146
},
{
"epoch": 0.32941176470588235,
"grad_norm": 0.2502163350582123,
"learning_rate": 0.0001651685393258427,
"loss": 0.8859,
"step": 147
},
{
"epoch": 0.33165266106442576,
"grad_norm": 0.23066553473472595,
"learning_rate": 0.00016629213483146068,
"loss": 0.9034,
"step": 148
},
{
"epoch": 0.33389355742296917,
"grad_norm": 0.2503213882446289,
"learning_rate": 0.00016741573033707868,
"loss": 0.8348,
"step": 149
},
{
"epoch": 0.33613445378151263,
"grad_norm": 0.25126519799232483,
"learning_rate": 0.00016853932584269662,
"loss": 0.8553,
"step": 150
},
{
"epoch": 0.33837535014005604,
"grad_norm": 0.241397425532341,
"learning_rate": 0.00016966292134831462,
"loss": 0.9132,
"step": 151
},
{
"epoch": 0.34061624649859945,
"grad_norm": 0.25682827830314636,
"learning_rate": 0.0001707865168539326,
"loss": 0.9481,
"step": 152
},
{
"epoch": 0.34285714285714286,
"grad_norm": 0.24024637043476105,
"learning_rate": 0.00017191011235955056,
"loss": 0.943,
"step": 153
},
{
"epoch": 0.34509803921568627,
"grad_norm": 0.2626055181026459,
"learning_rate": 0.00017303370786516853,
"loss": 0.7863,
"step": 154
},
{
"epoch": 0.3473389355742297,
"grad_norm": 0.24571183323860168,
"learning_rate": 0.00017415730337078653,
"loss": 0.926,
"step": 155
},
{
"epoch": 0.3495798319327731,
"grad_norm": 0.2499912977218628,
"learning_rate": 0.0001752808988764045,
"loss": 0.8232,
"step": 156
},
{
"epoch": 0.35182072829131655,
"grad_norm": 0.25473934412002563,
"learning_rate": 0.00017640449438202248,
"loss": 0.9021,
"step": 157
},
{
"epoch": 0.35406162464985996,
"grad_norm": 0.25837019085884094,
"learning_rate": 0.00017752808988764045,
"loss": 0.7753,
"step": 158
},
{
"epoch": 0.3563025210084034,
"grad_norm": 0.255958616733551,
"learning_rate": 0.00017865168539325842,
"loss": 0.9155,
"step": 159
},
{
"epoch": 0.3585434173669468,
"grad_norm": 0.24315786361694336,
"learning_rate": 0.00017977528089887642,
"loss": 0.8039,
"step": 160
},
{
"epoch": 0.3607843137254902,
"grad_norm": 0.24614644050598145,
"learning_rate": 0.0001808988764044944,
"loss": 0.8857,
"step": 161
},
{
"epoch": 0.3630252100840336,
"grad_norm": 0.24280671775341034,
"learning_rate": 0.00018202247191011236,
"loss": 0.8354,
"step": 162
},
{
"epoch": 0.365266106442577,
"grad_norm": 0.2597411572933197,
"learning_rate": 0.00018314606741573036,
"loss": 0.8852,
"step": 163
},
{
"epoch": 0.3675070028011205,
"grad_norm": 0.2597702145576477,
"learning_rate": 0.00018426966292134833,
"loss": 0.9029,
"step": 164
},
{
"epoch": 0.3697478991596639,
"grad_norm": 0.23551709949970245,
"learning_rate": 0.0001853932584269663,
"loss": 0.9118,
"step": 165
},
{
"epoch": 0.3719887955182073,
"grad_norm": 0.2516990303993225,
"learning_rate": 0.00018651685393258427,
"loss": 0.7997,
"step": 166
},
{
"epoch": 0.3742296918767507,
"grad_norm": 0.2297232747077942,
"learning_rate": 0.00018764044943820224,
"loss": 0.7823,
"step": 167
},
{
"epoch": 0.3764705882352941,
"grad_norm": 0.22287945449352264,
"learning_rate": 0.00018876404494382024,
"loss": 0.8609,
"step": 168
},
{
"epoch": 0.3787114845938375,
"grad_norm": 0.2428523749113083,
"learning_rate": 0.0001898876404494382,
"loss": 0.8842,
"step": 169
},
{
"epoch": 0.38095238095238093,
"grad_norm": 0.25949063897132874,
"learning_rate": 0.00019101123595505618,
"loss": 0.7737,
"step": 170
},
{
"epoch": 0.3831932773109244,
"grad_norm": 0.2532324492931366,
"learning_rate": 0.00019213483146067416,
"loss": 0.7548,
"step": 171
},
{
"epoch": 0.3854341736694678,
"grad_norm": 0.2657643258571625,
"learning_rate": 0.00019325842696629215,
"loss": 0.8375,
"step": 172
},
{
"epoch": 0.3876750700280112,
"grad_norm": 0.2578238546848297,
"learning_rate": 0.00019438202247191013,
"loss": 0.8591,
"step": 173
},
{
"epoch": 0.3899159663865546,
"grad_norm": 0.24303969740867615,
"learning_rate": 0.0001955056179775281,
"loss": 0.8131,
"step": 174
},
{
"epoch": 0.39215686274509803,
"grad_norm": 0.271139919757843,
"learning_rate": 0.00019662921348314607,
"loss": 0.8234,
"step": 175
},
{
"epoch": 0.39439775910364144,
"grad_norm": 0.2569217085838318,
"learning_rate": 0.00019775280898876404,
"loss": 0.8907,
"step": 176
},
{
"epoch": 0.39663865546218485,
"grad_norm": 0.23687879741191864,
"learning_rate": 0.00019887640449438204,
"loss": 0.8082,
"step": 177
},
{
"epoch": 0.3988795518207283,
"grad_norm": 0.24828100204467773,
"learning_rate": 0.0002,
"loss": 0.8575,
"step": 178
},
{
"epoch": 0.4011204481792717,
"grad_norm": 0.24300478398799896,
"learning_rate": 0.00019999980867200105,
"loss": 0.7963,
"step": 179
},
{
"epoch": 0.40336134453781514,
"grad_norm": 0.2662373483181,
"learning_rate": 0.00019999923468873635,
"loss": 0.8871,
"step": 180
},
{
"epoch": 0.40560224089635855,
"grad_norm": 0.2600310146808624,
"learning_rate": 0.00019999827805240226,
"loss": 0.8477,
"step": 181
},
{
"epoch": 0.40784313725490196,
"grad_norm": 0.26801761984825134,
"learning_rate": 0.00019999693876665938,
"loss": 0.8978,
"step": 182
},
{
"epoch": 0.41008403361344536,
"grad_norm": 0.26510775089263916,
"learning_rate": 0.00019999521683663262,
"loss": 0.8387,
"step": 183
},
{
"epoch": 0.4123249299719888,
"grad_norm": 0.25927454233169556,
"learning_rate": 0.00019999311226891103,
"loss": 0.884,
"step": 184
},
{
"epoch": 0.41456582633053224,
"grad_norm": 0.2489653080701828,
"learning_rate": 0.00019999062507154784,
"loss": 0.9473,
"step": 185
},
{
"epoch": 0.41680672268907565,
"grad_norm": 0.2461009919643402,
"learning_rate": 0.0001999877552540605,
"loss": 0.8609,
"step": 186
},
{
"epoch": 0.41904761904761906,
"grad_norm": 0.24591152369976044,
"learning_rate": 0.00019998450282743052,
"loss": 0.9284,
"step": 187
},
{
"epoch": 0.42128851540616247,
"grad_norm": 0.23085853457450867,
"learning_rate": 0.00019998086780410353,
"loss": 0.8898,
"step": 188
},
{
"epoch": 0.4235294117647059,
"grad_norm": 0.2512655556201935,
"learning_rate": 0.00019997685019798912,
"loss": 0.8275,
"step": 189
},
{
"epoch": 0.4257703081232493,
"grad_norm": 0.24720723927021027,
"learning_rate": 0.0001999724500244609,
"loss": 0.8809,
"step": 190
},
{
"epoch": 0.4280112044817927,
"grad_norm": 0.25972652435302734,
"learning_rate": 0.00019996766730035642,
"loss": 0.891,
"step": 191
},
{
"epoch": 0.43025210084033616,
"grad_norm": 0.26756009459495544,
"learning_rate": 0.0001999625020439771,
"loss": 0.923,
"step": 192
},
{
"epoch": 0.43249299719887957,
"grad_norm": 0.24553890526294708,
"learning_rate": 0.000199956954275088,
"loss": 0.8222,
"step": 193
},
{
"epoch": 0.434733893557423,
"grad_norm": 0.24937503039836884,
"learning_rate": 0.0001999510240149181,
"loss": 0.788,
"step": 194
},
{
"epoch": 0.4369747899159664,
"grad_norm": 0.2749722898006439,
"learning_rate": 0.00019994471128615985,
"loss": 0.9383,
"step": 195
},
{
"epoch": 0.4392156862745098,
"grad_norm": 0.25508439540863037,
"learning_rate": 0.00019993801611296923,
"loss": 0.8234,
"step": 196
},
{
"epoch": 0.4414565826330532,
"grad_norm": 0.26953256130218506,
"learning_rate": 0.00019993093852096582,
"loss": 0.9895,
"step": 197
},
{
"epoch": 0.4436974789915966,
"grad_norm": 0.26215213537216187,
"learning_rate": 0.0001999234785372324,
"loss": 0.8432,
"step": 198
},
{
"epoch": 0.4459383753501401,
"grad_norm": 0.24869990348815918,
"learning_rate": 0.00019991563619031508,
"loss": 0.8786,
"step": 199
},
{
"epoch": 0.4481792717086835,
"grad_norm": 0.23765677213668823,
"learning_rate": 0.00019990741151022301,
"loss": 0.9402,
"step": 200
},
{
"epoch": 0.4504201680672269,
"grad_norm": 0.24099862575531006,
"learning_rate": 0.00019989880452842847,
"loss": 0.9192,
"step": 201
},
{
"epoch": 0.4526610644257703,
"grad_norm": 0.2570018172264099,
"learning_rate": 0.00019988981527786654,
"loss": 0.846,
"step": 202
},
{
"epoch": 0.4549019607843137,
"grad_norm": 0.2420913279056549,
"learning_rate": 0.00019988044379293523,
"loss": 0.9021,
"step": 203
},
{
"epoch": 0.45714285714285713,
"grad_norm": 0.25546249747276306,
"learning_rate": 0.00019987069010949496,
"loss": 0.9191,
"step": 204
},
{
"epoch": 0.45938375350140054,
"grad_norm": 0.25439539551734924,
"learning_rate": 0.00019986055426486887,
"loss": 0.9046,
"step": 205
},
{
"epoch": 0.461624649859944,
"grad_norm": 0.23050928115844727,
"learning_rate": 0.00019985003629784237,
"loss": 0.7777,
"step": 206
},
{
"epoch": 0.4638655462184874,
"grad_norm": 0.25798794627189636,
"learning_rate": 0.00019983913624866304,
"loss": 0.9232,
"step": 207
},
{
"epoch": 0.4661064425770308,
"grad_norm": 0.2508363127708435,
"learning_rate": 0.00019982785415904064,
"loss": 0.9924,
"step": 208
},
{
"epoch": 0.46834733893557423,
"grad_norm": 0.25454050302505493,
"learning_rate": 0.00019981619007214673,
"loss": 0.9056,
"step": 209
},
{
"epoch": 0.47058823529411764,
"grad_norm": 0.24120093882083893,
"learning_rate": 0.0001998041440326146,
"loss": 0.8339,
"step": 210
},
{
"epoch": 0.47282913165266105,
"grad_norm": 0.2521376609802246,
"learning_rate": 0.00019979171608653924,
"loss": 0.7907,
"step": 211
},
{
"epoch": 0.47507002801120446,
"grad_norm": 0.2513487637042999,
"learning_rate": 0.00019977890628147682,
"loss": 0.9142,
"step": 212
},
{
"epoch": 0.4773109243697479,
"grad_norm": 0.2808438241481781,
"learning_rate": 0.00019976571466644492,
"loss": 0.8708,
"step": 213
},
{
"epoch": 0.47955182072829133,
"grad_norm": 0.2564331293106079,
"learning_rate": 0.00019975214129192196,
"loss": 0.8726,
"step": 214
},
{
"epoch": 0.48179271708683474,
"grad_norm": 0.2591661214828491,
"learning_rate": 0.00019973818620984738,
"loss": 0.791,
"step": 215
},
{
"epoch": 0.48403361344537815,
"grad_norm": 0.2555089592933655,
"learning_rate": 0.00019972384947362101,
"loss": 0.9129,
"step": 216
},
{
"epoch": 0.48627450980392156,
"grad_norm": 0.2542738616466522,
"learning_rate": 0.00019970913113810334,
"loss": 0.8097,
"step": 217
},
{
"epoch": 0.48851540616246497,
"grad_norm": 0.26333680748939514,
"learning_rate": 0.0001996940312596149,
"loss": 0.8342,
"step": 218
},
{
"epoch": 0.4907563025210084,
"grad_norm": 0.24918504059314728,
"learning_rate": 0.00019967854989593633,
"loss": 0.8441,
"step": 219
},
{
"epoch": 0.49299719887955185,
"grad_norm": 0.26157376170158386,
"learning_rate": 0.00019966268710630797,
"loss": 0.8891,
"step": 220
},
{
"epoch": 0.49523809523809526,
"grad_norm": 0.277658611536026,
"learning_rate": 0.00019964644295142968,
"loss": 0.8862,
"step": 221
},
{
"epoch": 0.49747899159663866,
"grad_norm": 0.2798251211643219,
"learning_rate": 0.00019962981749346078,
"loss": 0.8742,
"step": 222
},
{
"epoch": 0.4997198879551821,
"grad_norm": 0.28391703963279724,
"learning_rate": 0.00019961281079601957,
"loss": 0.8928,
"step": 223
},
{
"epoch": 0.5019607843137255,
"grad_norm": 0.2569884955883026,
"learning_rate": 0.00019959542292418317,
"loss": 0.8289,
"step": 224
},
{
"epoch": 0.5042016806722689,
"grad_norm": 0.2689199447631836,
"learning_rate": 0.00019957765394448728,
"loss": 0.8002,
"step": 225
},
{
"epoch": 0.5064425770308123,
"grad_norm": 0.2564886510372162,
"learning_rate": 0.00019955950392492604,
"loss": 0.8817,
"step": 226
},
{
"epoch": 0.5086834733893557,
"grad_norm": 0.2417922168970108,
"learning_rate": 0.00019954097293495155,
"loss": 0.8838,
"step": 227
},
{
"epoch": 0.5109243697478991,
"grad_norm": 0.22530929744243622,
"learning_rate": 0.00019952206104547376,
"loss": 0.7967,
"step": 228
},
{
"epoch": 0.5131652661064425,
"grad_norm": 0.23360571265220642,
"learning_rate": 0.00019950276832886017,
"loss": 0.8364,
"step": 229
},
{
"epoch": 0.5154061624649859,
"grad_norm": 0.2611772418022156,
"learning_rate": 0.00019948309485893549,
"loss": 0.8786,
"step": 230
},
{
"epoch": 0.5176470588235295,
"grad_norm": 0.24788020551204681,
"learning_rate": 0.00019946304071098142,
"loss": 0.86,
"step": 231
},
{
"epoch": 0.5198879551820729,
"grad_norm": 0.23259863257408142,
"learning_rate": 0.00019944260596173641,
"loss": 0.822,
"step": 232
},
{
"epoch": 0.5221288515406163,
"grad_norm": 0.2678159773349762,
"learning_rate": 0.0001994217906893952,
"loss": 0.8522,
"step": 233
},
{
"epoch": 0.5243697478991597,
"grad_norm": 0.268537312746048,
"learning_rate": 0.00019940059497360873,
"loss": 0.8638,
"step": 234
},
{
"epoch": 0.5266106442577031,
"grad_norm": 0.26122671365737915,
"learning_rate": 0.0001993790188954836,
"loss": 0.8509,
"step": 235
},
{
"epoch": 0.5288515406162465,
"grad_norm": 0.38401761651039124,
"learning_rate": 0.00019935706253758207,
"loss": 0.8184,
"step": 236
},
{
"epoch": 0.5310924369747899,
"grad_norm": 0.2459888607263565,
"learning_rate": 0.00019933472598392138,
"loss": 0.8042,
"step": 237
},
{
"epoch": 0.5333333333333333,
"grad_norm": 0.27041634917259216,
"learning_rate": 0.0001993120093199737,
"loss": 0.9426,
"step": 238
},
{
"epoch": 0.5355742296918767,
"grad_norm": 0.2503630220890045,
"learning_rate": 0.00019928891263266578,
"loss": 0.901,
"step": 239
},
{
"epoch": 0.5378151260504201,
"grad_norm": 0.2358655333518982,
"learning_rate": 0.00019926543601037842,
"loss": 0.8165,
"step": 240
},
{
"epoch": 0.5400560224089636,
"grad_norm": 0.24630171060562134,
"learning_rate": 0.00019924157954294628,
"loss": 0.8083,
"step": 241
},
{
"epoch": 0.542296918767507,
"grad_norm": 0.2504923939704895,
"learning_rate": 0.00019921734332165766,
"loss": 0.7973,
"step": 242
},
{
"epoch": 0.5445378151260504,
"grad_norm": 0.2728249132633209,
"learning_rate": 0.00019919272743925385,
"loss": 0.8817,
"step": 243
},
{
"epoch": 0.5467787114845938,
"grad_norm": 0.2694825828075409,
"learning_rate": 0.000199167731989929,
"loss": 0.8458,
"step": 244
},
{
"epoch": 0.5490196078431373,
"grad_norm": 0.31109780073165894,
"learning_rate": 0.00019914235706932972,
"loss": 0.939,
"step": 245
},
{
"epoch": 0.5512605042016807,
"grad_norm": 0.25847890973091125,
"learning_rate": 0.0001991166027745547,
"loss": 0.8704,
"step": 246
},
{
"epoch": 0.5535014005602241,
"grad_norm": 0.2678435742855072,
"learning_rate": 0.00019909046920415423,
"loss": 0.7445,
"step": 247
},
{
"epoch": 0.5557422969187675,
"grad_norm": 0.25951921939849854,
"learning_rate": 0.00019906395645812998,
"loss": 0.8869,
"step": 248
},
{
"epoch": 0.5579831932773109,
"grad_norm": 0.25794005393981934,
"learning_rate": 0.00019903706463793462,
"loss": 0.9506,
"step": 249
},
{
"epoch": 0.5602240896358543,
"grad_norm": 0.2340375930070877,
"learning_rate": 0.00019900979384647127,
"loss": 0.811,
"step": 250
},
{
"epoch": 0.5624649859943978,
"grad_norm": 0.25867462158203125,
"learning_rate": 0.0001989821441880933,
"loss": 0.8621,
"step": 251
},
{
"epoch": 0.5647058823529412,
"grad_norm": 0.2387334555387497,
"learning_rate": 0.0001989541157686037,
"loss": 0.8514,
"step": 252
},
{
"epoch": 0.5669467787114846,
"grad_norm": 0.26006007194519043,
"learning_rate": 0.00019892570869525496,
"loss": 0.8946,
"step": 253
},
{
"epoch": 0.569187675070028,
"grad_norm": 0.2735587954521179,
"learning_rate": 0.00019889692307674845,
"loss": 0.8605,
"step": 254
},
{
"epoch": 0.5714285714285714,
"grad_norm": 0.28292015194892883,
"learning_rate": 0.00019886775902323405,
"loss": 1.0231,
"step": 255
},
{
"epoch": 0.5736694677871148,
"grad_norm": 0.2500753402709961,
"learning_rate": 0.00019883821664630976,
"loss": 0.9155,
"step": 256
},
{
"epoch": 0.5759103641456582,
"grad_norm": 0.23072321712970734,
"learning_rate": 0.00019880829605902126,
"loss": 0.8268,
"step": 257
},
{
"epoch": 0.5781512605042017,
"grad_norm": 0.25353720784187317,
"learning_rate": 0.0001987779973758615,
"loss": 0.9315,
"step": 258
},
{
"epoch": 0.5803921568627451,
"grad_norm": 0.24684756994247437,
"learning_rate": 0.00019874732071277013,
"loss": 0.8652,
"step": 259
},
{
"epoch": 0.5826330532212886,
"grad_norm": 0.2583218812942505,
"learning_rate": 0.0001987162661871333,
"loss": 0.7722,
"step": 260
},
{
"epoch": 0.584873949579832,
"grad_norm": 0.23789426684379578,
"learning_rate": 0.00019868483391778302,
"loss": 0.9084,
"step": 261
},
{
"epoch": 0.5871148459383754,
"grad_norm": 0.24503661692142487,
"learning_rate": 0.00019865302402499678,
"loss": 0.8683,
"step": 262
},
{
"epoch": 0.5893557422969188,
"grad_norm": 0.2724620997905731,
"learning_rate": 0.00019862083663049694,
"loss": 0.8323,
"step": 263
},
{
"epoch": 0.5915966386554622,
"grad_norm": 0.27704504132270813,
"learning_rate": 0.0001985882718574506,
"loss": 0.9132,
"step": 264
},
{
"epoch": 0.5938375350140056,
"grad_norm": 0.2760598063468933,
"learning_rate": 0.00019855532983046876,
"loss": 0.7416,
"step": 265
},
{
"epoch": 0.596078431372549,
"grad_norm": 0.26945289969444275,
"learning_rate": 0.00019852201067560606,
"loss": 0.965,
"step": 266
},
{
"epoch": 0.5983193277310924,
"grad_norm": 0.2512185573577881,
"learning_rate": 0.0001984883145203603,
"loss": 0.8582,
"step": 267
},
{
"epoch": 0.6005602240896358,
"grad_norm": 0.24201013147830963,
"learning_rate": 0.00019845424149367177,
"loss": 0.8433,
"step": 268
},
{
"epoch": 0.6028011204481792,
"grad_norm": 0.24099834263324738,
"learning_rate": 0.000198419791725923,
"loss": 0.9501,
"step": 269
},
{
"epoch": 0.6050420168067226,
"grad_norm": 0.24521470069885254,
"learning_rate": 0.00019838496534893806,
"loss": 0.8458,
"step": 270
},
{
"epoch": 0.6072829131652661,
"grad_norm": 0.2377120554447174,
"learning_rate": 0.00019834976249598221,
"loss": 0.8934,
"step": 271
},
{
"epoch": 0.6095238095238096,
"grad_norm": 0.2444024235010147,
"learning_rate": 0.00019831418330176125,
"loss": 0.8326,
"step": 272
},
{
"epoch": 0.611764705882353,
"grad_norm": 0.23458163440227509,
"learning_rate": 0.0001982782279024211,
"loss": 0.8743,
"step": 273
},
{
"epoch": 0.6140056022408964,
"grad_norm": 0.2571447491645813,
"learning_rate": 0.00019824189643554725,
"loss": 0.8265,
"step": 274
},
{
"epoch": 0.6162464985994398,
"grad_norm": 0.25711604952812195,
"learning_rate": 0.00019820518904016426,
"loss": 0.8418,
"step": 275
},
{
"epoch": 0.6184873949579832,
"grad_norm": 0.26878827810287476,
"learning_rate": 0.00019816810585673514,
"loss": 0.8007,
"step": 276
},
{
"epoch": 0.6207282913165266,
"grad_norm": 0.2555851638317108,
"learning_rate": 0.00019813064702716094,
"loss": 0.8536,
"step": 277
},
{
"epoch": 0.62296918767507,
"grad_norm": 0.2718588411808014,
"learning_rate": 0.00019809281269478012,
"loss": 0.8884,
"step": 278
},
{
"epoch": 0.6252100840336134,
"grad_norm": 0.25706928968429565,
"learning_rate": 0.00019805460300436803,
"loss": 0.8729,
"step": 279
},
{
"epoch": 0.6274509803921569,
"grad_norm": 0.23707084357738495,
"learning_rate": 0.00019801601810213635,
"loss": 0.8268,
"step": 280
},
{
"epoch": 0.6296918767507003,
"grad_norm": 0.24094390869140625,
"learning_rate": 0.00019797705813573245,
"loss": 0.8457,
"step": 281
},
{
"epoch": 0.6319327731092437,
"grad_norm": 0.24794656038284302,
"learning_rate": 0.00019793772325423908,
"loss": 0.8495,
"step": 282
},
{
"epoch": 0.6341736694677871,
"grad_norm": 0.2355436086654663,
"learning_rate": 0.00019789801360817346,
"loss": 0.8167,
"step": 283
},
{
"epoch": 0.6364145658263305,
"grad_norm": 0.24097253382205963,
"learning_rate": 0.00019785792934948695,
"loss": 0.8887,
"step": 284
},
{
"epoch": 0.6386554621848739,
"grad_norm": 0.23715715110301971,
"learning_rate": 0.00019781747063156435,
"loss": 0.8205,
"step": 285
},
{
"epoch": 0.6408963585434174,
"grad_norm": 0.2627374529838562,
"learning_rate": 0.00019777663760922343,
"loss": 0.9175,
"step": 286
},
{
"epoch": 0.6431372549019608,
"grad_norm": 0.25734400749206543,
"learning_rate": 0.00019773543043871412,
"loss": 0.8104,
"step": 287
},
{
"epoch": 0.6453781512605042,
"grad_norm": 0.2477787584066391,
"learning_rate": 0.0001976938492777182,
"loss": 0.9969,
"step": 288
},
{
"epoch": 0.6476190476190476,
"grad_norm": 0.2440386414527893,
"learning_rate": 0.0001976518942853484,
"loss": 0.8084,
"step": 289
},
{
"epoch": 0.6498599439775911,
"grad_norm": 0.2703987956047058,
"learning_rate": 0.00019760956562214806,
"loss": 0.7827,
"step": 290
},
{
"epoch": 0.6521008403361345,
"grad_norm": 0.29450032114982605,
"learning_rate": 0.0001975668634500904,
"loss": 0.8807,
"step": 291
},
{
"epoch": 0.6543417366946779,
"grad_norm": 0.27439677715301514,
"learning_rate": 0.00019752378793257776,
"loss": 0.8152,
"step": 292
},
{
"epoch": 0.6565826330532213,
"grad_norm": 0.2940295338630676,
"learning_rate": 0.00019748033923444122,
"loss": 0.9051,
"step": 293
},
{
"epoch": 0.6588235294117647,
"grad_norm": 0.2529405355453491,
"learning_rate": 0.00019743651752193982,
"loss": 0.7429,
"step": 294
},
{
"epoch": 0.6610644257703081,
"grad_norm": 0.26964515447616577,
"learning_rate": 0.00019739232296276003,
"loss": 0.8584,
"step": 295
},
{
"epoch": 0.6633053221288515,
"grad_norm": 0.24733929336071014,
"learning_rate": 0.00019734775572601487,
"loss": 0.8465,
"step": 296
},
{
"epoch": 0.6655462184873949,
"grad_norm": 0.2559841275215149,
"learning_rate": 0.00019730281598224364,
"loss": 0.883,
"step": 297
},
{
"epoch": 0.6677871148459383,
"grad_norm": 0.24375739693641663,
"learning_rate": 0.00019725750390341094,
"loss": 0.8042,
"step": 298
},
{
"epoch": 0.6700280112044817,
"grad_norm": 0.25603991746902466,
"learning_rate": 0.00019721181966290613,
"loss": 0.9363,
"step": 299
},
{
"epoch": 0.6722689075630253,
"grad_norm": 0.2391873151063919,
"learning_rate": 0.00019716576343554274,
"loss": 0.9364,
"step": 300
},
{
"epoch": 0.6745098039215687,
"grad_norm": 0.24241115152835846,
"learning_rate": 0.00019711933539755765,
"loss": 0.8168,
"step": 301
},
{
"epoch": 0.6767507002801121,
"grad_norm": 0.2441554218530655,
"learning_rate": 0.00019707253572661055,
"loss": 0.9179,
"step": 302
},
{
"epoch": 0.6789915966386555,
"grad_norm": 0.2645583748817444,
"learning_rate": 0.00019702536460178318,
"loss": 0.7971,
"step": 303
},
{
"epoch": 0.6812324929971989,
"grad_norm": 0.2557383179664612,
"learning_rate": 0.0001969778222035787,
"loss": 0.9112,
"step": 304
},
{
"epoch": 0.6834733893557423,
"grad_norm": 0.25000718235969543,
"learning_rate": 0.0001969299087139209,
"loss": 0.7842,
"step": 305
},
{
"epoch": 0.6857142857142857,
"grad_norm": 0.23475897312164307,
"learning_rate": 0.00019688162431615367,
"loss": 0.7806,
"step": 306
},
{
"epoch": 0.6879551820728291,
"grad_norm": 0.2505607306957245,
"learning_rate": 0.00019683296919504012,
"loss": 0.922,
"step": 307
},
{
"epoch": 0.6901960784313725,
"grad_norm": 0.2522655725479126,
"learning_rate": 0.00019678394353676203,
"loss": 0.7934,
"step": 308
},
{
"epoch": 0.692436974789916,
"grad_norm": 0.2722029387950897,
"learning_rate": 0.000196734547528919,
"loss": 0.8775,
"step": 309
},
{
"epoch": 0.6946778711484594,
"grad_norm": 0.26442402601242065,
"learning_rate": 0.00019668478136052774,
"loss": 0.8696,
"step": 310
},
{
"epoch": 0.6969187675070028,
"grad_norm": 0.26388052105903625,
"learning_rate": 0.00019663464522202162,
"loss": 0.8569,
"step": 311
},
{
"epoch": 0.6991596638655462,
"grad_norm": 0.26141613721847534,
"learning_rate": 0.00019658413930524952,
"loss": 0.8485,
"step": 312
},
{
"epoch": 0.7014005602240896,
"grad_norm": 0.23456323146820068,
"learning_rate": 0.00019653326380347533,
"loss": 0.8913,
"step": 313
},
{
"epoch": 0.7036414565826331,
"grad_norm": 0.24977454543113708,
"learning_rate": 0.00019648201891137723,
"loss": 0.8729,
"step": 314
},
{
"epoch": 0.7058823529411765,
"grad_norm": 0.22706039249897003,
"learning_rate": 0.0001964304048250469,
"loss": 0.7885,
"step": 315
},
{
"epoch": 0.7081232492997199,
"grad_norm": 0.24601607024669647,
"learning_rate": 0.00019637842174198867,
"loss": 0.8387,
"step": 316
},
{
"epoch": 0.7103641456582633,
"grad_norm": 0.2545998990535736,
"learning_rate": 0.000196326069861119,
"loss": 0.8114,
"step": 317
},
{
"epoch": 0.7126050420168067,
"grad_norm": 0.24375082552433014,
"learning_rate": 0.00019627334938276546,
"loss": 0.843,
"step": 318
},
{
"epoch": 0.7148459383753502,
"grad_norm": 0.2519029378890991,
"learning_rate": 0.00019622026050866614,
"loss": 0.9088,
"step": 319
},
{
"epoch": 0.7170868347338936,
"grad_norm": 0.27405208349227905,
"learning_rate": 0.0001961668034419688,
"loss": 0.958,
"step": 320
},
{
"epoch": 0.719327731092437,
"grad_norm": 0.25491318106651306,
"learning_rate": 0.0001961129783872301,
"loss": 0.7692,
"step": 321
},
{
"epoch": 0.7215686274509804,
"grad_norm": 0.250347375869751,
"learning_rate": 0.00019605878555041485,
"loss": 0.8321,
"step": 322
},
{
"epoch": 0.7238095238095238,
"grad_norm": 0.24250715970993042,
"learning_rate": 0.00019600422513889516,
"loss": 0.7832,
"step": 323
},
{
"epoch": 0.7260504201680672,
"grad_norm": 0.2599019706249237,
"learning_rate": 0.00019594929736144976,
"loss": 0.8736,
"step": 324
},
{
"epoch": 0.7282913165266106,
"grad_norm": 0.25766611099243164,
"learning_rate": 0.00019589400242826305,
"loss": 0.7419,
"step": 325
},
{
"epoch": 0.730532212885154,
"grad_norm": 0.254658967256546,
"learning_rate": 0.00019583834055092445,
"loss": 0.9058,
"step": 326
},
{
"epoch": 0.7327731092436974,
"grad_norm": 0.2731349468231201,
"learning_rate": 0.00019578231194242743,
"loss": 0.9683,
"step": 327
},
{
"epoch": 0.735014005602241,
"grad_norm": 0.22153563797473907,
"learning_rate": 0.00019572591681716887,
"loss": 0.8146,
"step": 328
},
{
"epoch": 0.7372549019607844,
"grad_norm": 0.24707463383674622,
"learning_rate": 0.00019566915539094803,
"loss": 0.8204,
"step": 329
},
{
"epoch": 0.7394957983193278,
"grad_norm": 0.2541884779930115,
"learning_rate": 0.00019561202788096597,
"loss": 0.7638,
"step": 330
},
{
"epoch": 0.7417366946778712,
"grad_norm": 0.2314218282699585,
"learning_rate": 0.00019555453450582452,
"loss": 0.7502,
"step": 331
},
{
"epoch": 0.7439775910364146,
"grad_norm": 0.24334260821342468,
"learning_rate": 0.00019549667548552556,
"loss": 0.912,
"step": 332
},
{
"epoch": 0.746218487394958,
"grad_norm": 0.24473878741264343,
"learning_rate": 0.00019543845104147,
"loss": 0.7558,
"step": 333
},
{
"epoch": 0.7484593837535014,
"grad_norm": 0.25403597950935364,
"learning_rate": 0.00019537986139645726,
"loss": 0.8406,
"step": 334
},
{
"epoch": 0.7507002801120448,
"grad_norm": 0.24638508260250092,
"learning_rate": 0.0001953209067746841,
"loss": 0.736,
"step": 335
},
{
"epoch": 0.7529411764705882,
"grad_norm": 0.25604501366615295,
"learning_rate": 0.00019526158740174393,
"loss": 0.867,
"step": 336
},
{
"epoch": 0.7551820728291316,
"grad_norm": 0.2617231011390686,
"learning_rate": 0.00019520190350462584,
"loss": 0.8654,
"step": 337
},
{
"epoch": 0.757422969187675,
"grad_norm": 0.2620690166950226,
"learning_rate": 0.0001951418553117139,
"loss": 0.8428,
"step": 338
},
{
"epoch": 0.7596638655462185,
"grad_norm": 0.25454917550086975,
"learning_rate": 0.0001950814430527861,
"loss": 0.8676,
"step": 339
},
{
"epoch": 0.7619047619047619,
"grad_norm": 0.26449549198150635,
"learning_rate": 0.00019502066695901358,
"loss": 0.8709,
"step": 340
},
{
"epoch": 0.7641456582633053,
"grad_norm": 0.24709953367710114,
"learning_rate": 0.0001949595272629597,
"loss": 0.8312,
"step": 341
},
{
"epoch": 0.7663865546218488,
"grad_norm": 0.2593318819999695,
"learning_rate": 0.00019489802419857917,
"loss": 0.8274,
"step": 342
},
{
"epoch": 0.7686274509803922,
"grad_norm": 0.2448015958070755,
"learning_rate": 0.00019483615800121716,
"loss": 0.8226,
"step": 343
},
{
"epoch": 0.7708683473389356,
"grad_norm": 0.26445460319519043,
"learning_rate": 0.00019477392890760839,
"loss": 0.8492,
"step": 344
},
{
"epoch": 0.773109243697479,
"grad_norm": 0.2573208808898926,
"learning_rate": 0.00019471133715587622,
"loss": 0.8294,
"step": 345
},
{
"epoch": 0.7753501400560224,
"grad_norm": 0.26753151416778564,
"learning_rate": 0.00019464838298553173,
"loss": 0.8446,
"step": 346
},
{
"epoch": 0.7775910364145658,
"grad_norm": 0.2667967975139618,
"learning_rate": 0.00019458506663747285,
"loss": 0.8388,
"step": 347
},
{
"epoch": 0.7798319327731092,
"grad_norm": 0.25934740900993347,
"learning_rate": 0.00019452138835398332,
"loss": 0.8149,
"step": 348
},
{
"epoch": 0.7820728291316527,
"grad_norm": 0.2525850832462311,
"learning_rate": 0.00019445734837873202,
"loss": 0.8325,
"step": 349
},
{
"epoch": 0.7843137254901961,
"grad_norm": 0.2837854325771332,
"learning_rate": 0.00019439294695677167,
"loss": 0.8795,
"step": 350
},
{
"epoch": 0.7865546218487395,
"grad_norm": 0.26006975769996643,
"learning_rate": 0.00019432818433453818,
"loss": 0.8267,
"step": 351
},
{
"epoch": 0.7887955182072829,
"grad_norm": 0.24872632324695587,
"learning_rate": 0.00019426306075984965,
"loss": 0.8553,
"step": 352
},
{
"epoch": 0.7910364145658263,
"grad_norm": 0.23633168637752533,
"learning_rate": 0.00019419757648190533,
"loss": 0.8544,
"step": 353
},
{
"epoch": 0.7932773109243697,
"grad_norm": 0.2425261288881302,
"learning_rate": 0.00019413173175128473,
"loss": 0.8645,
"step": 354
},
{
"epoch": 0.7955182072829131,
"grad_norm": 0.23861053586006165,
"learning_rate": 0.00019406552681994663,
"loss": 0.8363,
"step": 355
},
{
"epoch": 0.7977591036414566,
"grad_norm": 0.24729354679584503,
"learning_rate": 0.00019399896194122822,
"loss": 0.8401,
"step": 356
},
{
"epoch": 0.8,
"grad_norm": 0.23853589594364166,
"learning_rate": 0.000193932037369844,
"loss": 0.8237,
"step": 357
},
{
"epoch": 0.8022408963585435,
"grad_norm": 0.2654939293861389,
"learning_rate": 0.00019386475336188484,
"loss": 0.7491,
"step": 358
},
{
"epoch": 0.8044817927170869,
"grad_norm": 0.27639004588127136,
"learning_rate": 0.000193797110174817,
"loss": 0.8755,
"step": 359
},
{
"epoch": 0.8067226890756303,
"grad_norm": 0.23675435781478882,
"learning_rate": 0.00019372910806748125,
"loss": 0.8292,
"step": 360
},
{
"epoch": 0.8089635854341737,
"grad_norm": 0.2371446043252945,
"learning_rate": 0.0001936607473000917,
"loss": 0.809,
"step": 361
},
{
"epoch": 0.8112044817927171,
"grad_norm": 0.2529555559158325,
"learning_rate": 0.0001935920281342349,
"loss": 0.8403,
"step": 362
},
{
"epoch": 0.8134453781512605,
"grad_norm": 0.2568047344684601,
"learning_rate": 0.00019352295083286896,
"loss": 0.7848,
"step": 363
},
{
"epoch": 0.8156862745098039,
"grad_norm": 0.2613024413585663,
"learning_rate": 0.0001934535156603222,
"loss": 0.844,
"step": 364
},
{
"epoch": 0.8179271708683473,
"grad_norm": 0.24079741537570953,
"learning_rate": 0.0001933837228822925,
"loss": 0.8029,
"step": 365
},
{
"epoch": 0.8201680672268907,
"grad_norm": 0.24890519678592682,
"learning_rate": 0.0001933135727658462,
"loss": 0.8712,
"step": 366
},
{
"epoch": 0.8224089635854341,
"grad_norm": 0.25198838114738464,
"learning_rate": 0.00019324306557941682,
"loss": 0.8368,
"step": 367
},
{
"epoch": 0.8246498599439775,
"grad_norm": 0.2470208704471588,
"learning_rate": 0.0001931722015928044,
"loss": 0.8896,
"step": 368
},
{
"epoch": 0.826890756302521,
"grad_norm": 0.25046589970588684,
"learning_rate": 0.00019310098107717418,
"loss": 0.7602,
"step": 369
},
{
"epoch": 0.8291316526610645,
"grad_norm": 0.28005844354629517,
"learning_rate": 0.0001930294043050558,
"loss": 0.9443,
"step": 370
},
{
"epoch": 0.8313725490196079,
"grad_norm": 0.6102829575538635,
"learning_rate": 0.00019295747155034202,
"loss": 0.911,
"step": 371
},
{
"epoch": 0.8336134453781513,
"grad_norm": 0.2472289353609085,
"learning_rate": 0.0001928851830882879,
"loss": 0.8271,
"step": 372
},
{
"epoch": 0.8358543417366947,
"grad_norm": 0.24470177292823792,
"learning_rate": 0.0001928125391955095,
"loss": 0.7837,
"step": 373
},
{
"epoch": 0.8380952380952381,
"grad_norm": 0.2458019107580185,
"learning_rate": 0.00019273954014998308,
"loss": 0.8496,
"step": 374
},
{
"epoch": 0.8403361344537815,
"grad_norm": 0.249577596783638,
"learning_rate": 0.00019266618623104385,
"loss": 0.7729,
"step": 375
},
{
"epoch": 0.8425770308123249,
"grad_norm": 0.2461290806531906,
"learning_rate": 0.000192592477719385,
"loss": 0.7955,
"step": 376
},
{
"epoch": 0.8448179271708683,
"grad_norm": 0.27021750807762146,
"learning_rate": 0.00019251841489705655,
"loss": 0.9384,
"step": 377
},
{
"epoch": 0.8470588235294118,
"grad_norm": 0.2547335624694824,
"learning_rate": 0.00019244399804746435,
"loss": 0.8254,
"step": 378
},
{
"epoch": 0.8492997198879552,
"grad_norm": 0.2620961368083954,
"learning_rate": 0.0001923692274553689,
"loss": 0.8587,
"step": 379
},
{
"epoch": 0.8515406162464986,
"grad_norm": 0.25756919384002686,
"learning_rate": 0.0001922941034068844,
"loss": 0.7906,
"step": 380
},
{
"epoch": 0.853781512605042,
"grad_norm": 0.25694355368614197,
"learning_rate": 0.0001922186261894775,
"loss": 0.7887,
"step": 381
},
{
"epoch": 0.8560224089635854,
"grad_norm": 0.2538197636604309,
"learning_rate": 0.0001921427960919663,
"loss": 0.8861,
"step": 382
},
{
"epoch": 0.8582633053221288,
"grad_norm": 0.2463122308254242,
"learning_rate": 0.00019206661340451925,
"loss": 0.8175,
"step": 383
},
{
"epoch": 0.8605042016806723,
"grad_norm": 0.23905646800994873,
"learning_rate": 0.00019199007841865396,
"loss": 0.7914,
"step": 384
},
{
"epoch": 0.8627450980392157,
"grad_norm": 0.2590409517288208,
"learning_rate": 0.0001919131914272361,
"loss": 0.8385,
"step": 385
},
{
"epoch": 0.8649859943977591,
"grad_norm": 0.2520500719547272,
"learning_rate": 0.00019183595272447842,
"loss": 0.7961,
"step": 386
},
{
"epoch": 0.8672268907563025,
"grad_norm": 0.2568177878856659,
"learning_rate": 0.00019175836260593938,
"loss": 0.8051,
"step": 387
},
{
"epoch": 0.869467787114846,
"grad_norm": 0.27684271335601807,
"learning_rate": 0.00019168042136852228,
"loss": 0.8953,
"step": 388
},
{
"epoch": 0.8717086834733894,
"grad_norm": 0.26281246542930603,
"learning_rate": 0.0001916021293104739,
"loss": 0.8602,
"step": 389
},
{
"epoch": 0.8739495798319328,
"grad_norm": 0.29199978709220886,
"learning_rate": 0.00019152348673138353,
"loss": 0.9776,
"step": 390
},
{
"epoch": 0.8761904761904762,
"grad_norm": 0.25914543867111206,
"learning_rate": 0.0001914444939321817,
"loss": 0.8394,
"step": 391
},
{
"epoch": 0.8784313725490196,
"grad_norm": 0.2472960352897644,
"learning_rate": 0.0001913651512151391,
"loss": 0.7995,
"step": 392
},
{
"epoch": 0.880672268907563,
"grad_norm": 0.24599507451057434,
"learning_rate": 0.00019128545888386536,
"loss": 0.7801,
"step": 393
},
{
"epoch": 0.8829131652661064,
"grad_norm": 0.23776014149188995,
"learning_rate": 0.00019120541724330803,
"loss": 0.8068,
"step": 394
},
{
"epoch": 0.8851540616246498,
"grad_norm": 0.27298274636268616,
"learning_rate": 0.0001911250265997512,
"loss": 0.9139,
"step": 395
},
{
"epoch": 0.8873949579831932,
"grad_norm": 0.2653828561306,
"learning_rate": 0.0001910442872608145,
"loss": 0.8917,
"step": 396
},
{
"epoch": 0.8896358543417366,
"grad_norm": 0.25794684886932373,
"learning_rate": 0.00019096319953545185,
"loss": 0.819,
"step": 397
},
{
"epoch": 0.8918767507002802,
"grad_norm": 0.24696165323257446,
"learning_rate": 0.0001908817637339503,
"loss": 0.7535,
"step": 398
},
{
"epoch": 0.8941176470588236,
"grad_norm": 0.2529006898403168,
"learning_rate": 0.00019079998016792885,
"loss": 0.7379,
"step": 399
},
{
"epoch": 0.896358543417367,
"grad_norm": 0.24874147772789001,
"learning_rate": 0.00019071784915033717,
"loss": 0.8973,
"step": 400
},
{
"epoch": 0.8985994397759104,
"grad_norm": 0.27747422456741333,
"learning_rate": 0.00019063537099545455,
"loss": 0.8685,
"step": 401
},
{
"epoch": 0.9008403361344538,
"grad_norm": 0.25184211134910583,
"learning_rate": 0.00019055254601888866,
"loss": 0.7937,
"step": 402
},
{
"epoch": 0.9030812324929972,
"grad_norm": 0.2565883994102478,
"learning_rate": 0.00019046937453757413,
"loss": 0.8677,
"step": 403
},
{
"epoch": 0.9053221288515406,
"grad_norm": 0.24833756685256958,
"learning_rate": 0.00019038585686977167,
"loss": 0.8777,
"step": 404
},
{
"epoch": 0.907563025210084,
"grad_norm": 0.24929295480251312,
"learning_rate": 0.00019030199333506666,
"loss": 0.8167,
"step": 405
},
{
"epoch": 0.9098039215686274,
"grad_norm": 0.24500809609889984,
"learning_rate": 0.00019021778425436795,
"loss": 0.8675,
"step": 406
},
{
"epoch": 0.9120448179271708,
"grad_norm": 0.25895681977272034,
"learning_rate": 0.0001901332299499066,
"loss": 0.8718,
"step": 407
},
{
"epoch": 0.9142857142857143,
"grad_norm": 0.22836971282958984,
"learning_rate": 0.00019004833074523478,
"loss": 0.8602,
"step": 408
},
{
"epoch": 0.9165266106442577,
"grad_norm": 0.26318010687828064,
"learning_rate": 0.00018996308696522433,
"loss": 0.838,
"step": 409
},
{
"epoch": 0.9187675070028011,
"grad_norm": 0.24389663338661194,
"learning_rate": 0.00018987749893606575,
"loss": 0.7798,
"step": 410
},
{
"epoch": 0.9210084033613445,
"grad_norm": 0.26448702812194824,
"learning_rate": 0.0001897915669852667,
"loss": 0.8053,
"step": 411
},
{
"epoch": 0.923249299719888,
"grad_norm": 0.25245535373687744,
"learning_rate": 0.000189705291441651,
"loss": 0.7915,
"step": 412
},
{
"epoch": 0.9254901960784314,
"grad_norm": 0.25623124837875366,
"learning_rate": 0.00018961867263535715,
"loss": 0.9167,
"step": 413
},
{
"epoch": 0.9277310924369748,
"grad_norm": 0.25271835923194885,
"learning_rate": 0.00018953171089783723,
"loss": 0.8663,
"step": 414
},
{
"epoch": 0.9299719887955182,
"grad_norm": 0.24757803976535797,
"learning_rate": 0.00018944440656185556,
"loss": 0.8411,
"step": 415
},
{
"epoch": 0.9322128851540616,
"grad_norm": 0.23879915475845337,
"learning_rate": 0.00018935675996148738,
"loss": 0.8071,
"step": 416
},
{
"epoch": 0.934453781512605,
"grad_norm": 0.3934721350669861,
"learning_rate": 0.0001892687714321177,
"loss": 0.7911,
"step": 417
},
{
"epoch": 0.9366946778711485,
"grad_norm": 0.27968111634254456,
"learning_rate": 0.00018918044131043985,
"loss": 0.8452,
"step": 418
},
{
"epoch": 0.9389355742296919,
"grad_norm": 0.2701101303100586,
"learning_rate": 0.00018909176993445442,
"loss": 0.8723,
"step": 419
},
{
"epoch": 0.9411764705882353,
"grad_norm": 0.2606533169746399,
"learning_rate": 0.00018900275764346768,
"loss": 0.7908,
"step": 420
},
{
"epoch": 0.9434173669467787,
"grad_norm": 0.2672193944454193,
"learning_rate": 0.00018891340477809055,
"loss": 0.9491,
"step": 421
},
{
"epoch": 0.9456582633053221,
"grad_norm": 0.2676648795604706,
"learning_rate": 0.00018882371168023706,
"loss": 0.8352,
"step": 422
},
{
"epoch": 0.9478991596638655,
"grad_norm": 0.265023410320282,
"learning_rate": 0.0001887336786931233,
"loss": 0.7894,
"step": 423
},
{
"epoch": 0.9501400560224089,
"grad_norm": 0.24014434218406677,
"learning_rate": 0.00018864330616126586,
"loss": 0.8394,
"step": 424
},
{
"epoch": 0.9523809523809523,
"grad_norm": 0.24389366805553436,
"learning_rate": 0.00018855259443048067,
"loss": 0.7857,
"step": 425
},
{
"epoch": 0.9546218487394958,
"grad_norm": 0.2471270114183426,
"learning_rate": 0.00018846154384788162,
"loss": 0.8576,
"step": 426
},
{
"epoch": 0.9568627450980393,
"grad_norm": 0.2577723264694214,
"learning_rate": 0.00018837015476187916,
"loss": 0.8377,
"step": 427
},
{
"epoch": 0.9591036414565827,
"grad_norm": 0.3224787414073944,
"learning_rate": 0.00018827842752217917,
"loss": 0.8801,
"step": 428
},
{
"epoch": 0.9613445378151261,
"grad_norm": 0.25494757294654846,
"learning_rate": 0.00018818636247978145,
"loss": 0.8173,
"step": 429
},
{
"epoch": 0.9635854341736695,
"grad_norm": 0.24220331013202667,
"learning_rate": 0.00018809395998697833,
"loss": 0.7747,
"step": 430
},
{
"epoch": 0.9658263305322129,
"grad_norm": 0.2741996645927429,
"learning_rate": 0.00018800122039735358,
"loss": 0.8636,
"step": 431
},
{
"epoch": 0.9680672268907563,
"grad_norm": 0.2558667063713074,
"learning_rate": 0.0001879081440657807,
"loss": 0.8456,
"step": 432
},
{
"epoch": 0.9703081232492997,
"grad_norm": 0.25905126333236694,
"learning_rate": 0.00018781473134842197,
"loss": 0.7961,
"step": 433
},
{
"epoch": 0.9725490196078431,
"grad_norm": 0.23688547313213348,
"learning_rate": 0.00018772098260272667,
"loss": 0.7555,
"step": 434
},
{
"epoch": 0.9747899159663865,
"grad_norm": 0.2701859176158905,
"learning_rate": 0.00018762689818743007,
"loss": 0.9353,
"step": 435
},
{
"epoch": 0.9770308123249299,
"grad_norm": 0.24914319813251495,
"learning_rate": 0.00018753247846255174,
"loss": 0.7678,
"step": 436
},
{
"epoch": 0.9792717086834734,
"grad_norm": 0.26182475686073303,
"learning_rate": 0.00018743772378939448,
"loss": 0.8374,
"step": 437
},
{
"epoch": 0.9815126050420168,
"grad_norm": 0.2661891281604767,
"learning_rate": 0.00018734263453054273,
"loss": 0.8816,
"step": 438
},
{
"epoch": 0.9837535014005602,
"grad_norm": 0.24797801673412323,
"learning_rate": 0.0001872472110498612,
"loss": 0.8069,
"step": 439
},
{
"epoch": 0.9859943977591037,
"grad_norm": 0.24370808899402618,
"learning_rate": 0.0001871514537124936,
"loss": 0.7405,
"step": 440
},
{
"epoch": 0.9882352941176471,
"grad_norm": 0.2744685709476471,
"learning_rate": 0.00018705536288486118,
"loss": 0.7706,
"step": 441
},
{
"epoch": 0.9904761904761905,
"grad_norm": 0.265438437461853,
"learning_rate": 0.0001869589389346611,
"loss": 0.8649,
"step": 442
},
{
"epoch": 0.9927170868347339,
"grad_norm": 0.24661187827587128,
"learning_rate": 0.0001868621822308655,
"loss": 0.8138,
"step": 443
},
{
"epoch": 0.9949579831932773,
"grad_norm": 0.2580495774745941,
"learning_rate": 0.00018676509314371974,
"loss": 0.7765,
"step": 444
},
{
"epoch": 0.9971988795518207,
"grad_norm": 0.2546556293964386,
"learning_rate": 0.00018666767204474094,
"loss": 0.8873,
"step": 445
},
{
"epoch": 0.9994397759103641,
"grad_norm": 0.25944411754608154,
"learning_rate": 0.00018656991930671686,
"loss": 0.8651,
"step": 446
}
],
"logging_steps": 1,
"max_steps": 1784,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 446,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 2.310622116052992e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}